ison with our approach. To realize a supplementary
comparison., we retrieved the results of (Huang et al.,
2019).
6.1 External Validity
Actually, our approach can be evaluated only on Java
open source projects. So, we conducted our exper-
iment only on four Java open-source projects that
have extensively been used in previous studies. This
can influence the generalizability of our results. To
mitigate this threat, further studies are required to
analyze our approach even to more datasets from
other types of projects whether proprietary software
or commercial one written in other programming lan-
guages. Other threats are related to the suitability
of our performance metrics to evaluate our JIT-DP
model. However, we use F1 and PofB20 which are
applied by past software engineering studies to anal-
yse various prediction techniques (noa, 2020; Xuan
et al., 2015).
7 CONCLUSION AND FUTURE
WORKS
This paper proposes an end-to-end deep learning
framework for just-in-time defect prediction to au-
tomatically learn expressive features from the set
of code changes.We conduct evaluations on four
open-source projects.The experiment results proved
that our approach improves significantly the exist-
ing work DBN- based features and CBS+ on average
of 20.86 and 34.1 in F1, respectively in the task of
within-project defect prediction. Besides, it improves
the cross-defect prediction technique DBN-CPP and
CBS+ on average of 32.85 and 39.95 respectively in
F1. Also, our approach can outperform it under the
effort-aware evaluation context.
In the future, we would like to extend our eval-
uation to other open source and commercial projects
in order to reduce the threats to external validity. In
addition, we plan to make our framework applica-
ble to other open-source projects written in differ-
ent languages besides Java language, such as Python,
C/C++, etc.
REFERENCES
(2020). Multi-objective Cross-Project Defect Prediction -
IEEE Conference Publication.
Barnett, J. G., Gathuru, C. K., Soldano, L. S., and McIn-
tosh, S. (2016). The relationship between commit
message detail and defect proneness in java projects
on github. In 2016 IEEE/ACM 13th Working Confer-
ence on Mining Software Repositories (MSR), pages
496–499. IEEE.
Dam, H. K., Pham, T., Ng, S. W., Tran, T., Grundy,
J., Ghose, A., Kim, T., and Kim, C.-J. (2018). A
deep tree-based model for software defect prediction.
arXiv:1802.00921 [cs]. arXiv: 1802.00921.
D’Ambros, M., Lanza, M., and Robbes, R. (2010). An ex-
tensive comparison of bug prediction approaches. In
2010 7th IEEE Working Conference on Mining Soft-
ware Repositories (MSR 2010), pages 31–41. IEEE.
D’Ambros, M., Lanza, M., and Robbes, R. (2012). Evaluat-
ing defect prediction approaches: a benchmark and an
extensive comparison. Empirical Software Engineer-
ing, 17(4):531–577. Publisher: Springer.
Fan, Y., Xia, X., da Costa, D. A., Lo, D., Hassan, A. E., and
Li, S. (2019). The Impact of Changes Mislabeled by
SZZ on Just-in-Time Defect Prediction. IEEE Trans-
actions on Software Engineering.
Ferreira, F., Silva, L. L., and Valente, M. T. (2019). Soft-
ware Engineering Meets Deep Learning: A Literature
Review. arXiv preprint arXiv:1909.11436.
Hoang, T., Khanh Dam, H., Kamei, Y., Lo, D., and
Ubayashi, N. (2019). DeepJIT: An End-to-End Deep
Learning Framework for Just-in-Time Defect Predic-
tion. In 2019 IEEE/ACM 16th International Confer-
ence on Mining Software Repositories (MSR), pages
34–45.
Huang, Q., Xia, X., and Lo, D. (2019). Revisiting super-
vised and unsupervised models for effort-aware just-
in-time defect prediction. Empirical Software Engi-
neering, 24(5):2823–2862.
Jiang, T., Tan, L., and Kim, S. (2013). Personalized de-
fect prediction. In 2013 28th IEEE/ACM Interna-
tional Conference on Automated Software Engineer-
ing (ASE), pages 279–289. Ieee.
Kamei, Y., Fukushima, T., McIntosh, S., Yamashita,
K., Ubayashi, N., and Hassan, A. E. (2016).
Studying just-in-time defect prediction using cross-
project models. Empirical Software Engineering,
21(5):2072–2106.
Kamei, Y., Matsumoto, S., Monden, A., Matsumoto, K.-i.,
Adams, B., and Hassan, A. E. (2010). Revisiting com-
mon bug prediction findings using effort-aware mod-
els. In 2010 IEEE International Conference on Soft-
ware Maintenance, pages 1–10. IEEE.
Kamei, Y., Shihab, E., Adams, B., Hassan, A. E., Mockus,
A., Sinha, A., and Ubayashi, N. (2012). A large-
scale empirical study of just-in-time quality assur-
ance. IEEE Transactions on Software Engineering,
39(6):757–773. Publisher: IEEE.
Kim, S., Whitehead Jr, E. J., and Zhang, Y. (2008). Classi-
fying software changes: Clean or buggy? IEEE Trans-
actions on Software Engineering, 34(2):181–196.
Li, J., He, P., Zhu, J., and Lyu, M. R. (2017). Software
defect prediction via convolutional neural network.
In 2017 IEEE International Conference on Software
Quality, Reliability and Security (QRS), pages 318–
328. IEEE.
ENASE 2022 - 17th International Conference on Evaluation of Novel Approaches to Software Engineering
136