Table 5: The AUC score for each model using the SRN Val-
idation Dataset (SRN) and our User Labeled Test Dataset
(UL).
SRN UL
Logistic Regression 0.962 0.729
SRN 0.955 0.634
HAN 0.932 0.666
HAVAN (HAN w/ VAT) 0.939 0.707
tention could be presented to a user. It also seems ad-
vantageous to leverage the vast number of unlabeled
examples with semi-supervised learning to classify is-
sues as security-related. Still, for our approach, it is
important to note that the number of labeled relevant
security examples is relatively few in comparison to
the full unlabeled dataset. To enable the use of more
aggressive SSL-methods, there is a need to acquire
more labeled examples.
REFERENCES
Chen, Y., Santosa, A. E., Yi, A. M., Sharma, A., Sharma,
A., and Lo, D. (2020). A machine learning approach
for vulnerability curation. In Proceedings of the 17th
International Conference on Mining Software Repos-
itories, MSR ’20, page 32–42. Association for Com-
puting Machinery.
Goodfellow, I. J., Shlens, J., and Szegedy, C. (2014).
Explaining and harnessing adversarial examples.
arXiv:1412.6572.
Halko, N., Martinsson, P. G., and Tropp, J. A. (2011).
Finding structure with randomness: Probabilistic al-
gorithms for constructing approximate matrix decom-
positions. SIAM Review, 53(2):217–288.
Hochreiter, S. and Schmidhuber, J. (1997). Long short-term
memory. Neural computation, 9:1735–80.
Khurana, D., Koli, A., Khatter, K., and Singh, S. (2017).
Natural language processing: State of the art, current
trends and challenges. arXiv:1708.05148.
Miyato, T., Dai, A. M., and Goodfellow, I. (2017). Adver-
sarial training methods for semi-supervised text clas-
sification. arXiv:1605.07725.
Miyato, T., Maeda, S.-i., Koyama, M., Nakae, K., and Ishii,
S. (2015). Distributional smoothing with virtual ad-
versarial training. arXiv:1507.00677.
Nigam, K. and Ghani, R. (2000). Analyzing the effective-
ness and applicability of co-training. In Proceedings
of the Ninth International Conference on Information
and Knowledge Management, CIKM ’00, page 86–93.
Association for Computing Machinery.
Palacio, D. N., McCrystal, D., Moran, K., Bernal-C
´
ardenas,
C., Poshyvanyk, D., and Shenefiel, C. (2019). Learn-
ing to identify security-related issues using convolu-
tional neural networks. In 2019 IEEE International
Conference on Software Maintenance and Evolution
(ICSME), pages 140–144.
SEMERU-Lab (2021). Securereqnet. https://github.com/
WM-SEMERU/SecureReqNet.
Synopsys (2020). 2020 open source security and
risk analysis. https://www.synopsys.com/
software-integrity/resources/analyst-reports/
open-source-security-risk-analysis.html.
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones,
L., Gomez, A. N., Kaiser, Ł., and Polosukhin, I.
(2017). Attention is all you need. In Advances in
neural information processing systems, pages 5998–
6008.
Yang, Z., Yang, D., Dyer, C., He, X., Smola, A., and Hovy,
E. (2016). Hierarchical attention networks for docu-
ment classification. In Proceedings of the 2016 Con-
ference of the North American Chapter of the Asso-
ciation for Computational Linguistics: Human Lan-
guage Technologies, pages 1480–1489. Association
for Computational Linguistics.
Zhou, Y. and Sharma, A. (2017). Automated identification
of security issues from commit messages and bug re-
ports. In Proceedings of the 2017 11th Joint Meeting
on Foundations of Software Engineering, ESEC/FSE
2017, page 914–919. Association for Computing Ma-
chinery.
Zou, D., Deng, Z., Li, Z., and Jin, H. (2018). Automatically
identifying security bug reports via multitype features
analysis. In Susilo, W. and Yang, G., editors, Infor-
mation Security and Privacy, pages 619–633, Cham.
Springer.
APPENDIX
A Annotation Guidelines
An annotation policy was established in order to make
the annotation process more efficient and to favor re-
peatability and reproducibility. All data in the User
Labeled Test Dataset was annotated by one of the au-
thors with knowledge in the field of cybersecurity, a
condition that must be met in order to adequately label
data as relating to cybersecurity. Some data was an-
notated by multiple parties and compared in the cases
of mismatch to ensure the annotations were similar.
Many issues were ambiguous and unclear, making
it important to create a policy. The annotation guide-
line was used to establish a unified labeling method.
It was updated regularly during the annotation phase
whenever a new kind of case arose. The categories
do not discriminate between questions, warnings, or
other discussions about a certain topic. The text is
annotated as the most severe category that accurately
describes it. The priority goes from Vuln being high-
est to Safe being lowest.
Vuln: Presence of known exploits, user-reported vul-
nerabilities.
ICISSP 2022 - 8th International Conference on Information Systems Security and Privacy
94