
of Computational and Experimental Science and En-
gineering, 10(3).
Heiding, F., Schneier, B., Vishwanath, A., Bernstein, J., and
Park, P. S. (2024). Devising and detecting phishing
emails using large language models. IEEE Access,
12:42131–42146.
Interaction Design Foundation - IxDF (2016). Three com-
mon problems in enterprise system user experience.
https://www.interaction-design.org/literature/article/
three-common-problems-in-enterprise-system-user-
experience.
ISO/IEC (2011). Systems and software engineering —-
Systems and software Quality Requirements and
Evaluation (SQuaRE) -— System and software qual-
ity models. Standard, ISO/IEC.
Jamal, S., Wimmer, H., and Sarker, I. H. (2024). An im-
proved transformer-based model for detecting phish-
ing, spam and ham emails: A large language model
approach. SECURITY AND PRIVACY, 7(5):e402.
Josten, M. and Weis, T. (2024). Investigating the Effec-
tiveness of Bayesian Spam Filters in Detecting LLM-
modified Spam Mails. 10.48550/arXiv.2408.14293.
Khalil, M. and Er, E. (2023). Will ChatGPT Get
You Caught? Rethinking of Plagiarism Detection.
In Learning and Collaboration Technologies, pages
475–487. Springer Nature Switzerland.
Kirchenbauer, J., Geiping, J., Wen, Y., Katz, J., Miers, I.,
and Goldstein, T. (2023). A watermark for large lan-
guage models. In Proceedings of the 40th Interna-
tional Conference on Machine Learning, volume 202,
pages 17061–17084.
Koide, T., Fukushi, N., Nakano, H., and Chiba, D.
(2024). ChatSpamDetector: Leveraging Large Lan-
guage Models for Effective Phishing Email Detection.
10.48550/arXiv.2402.18093.
Kreps, S., McCain, R. M., and Brundage, M. (2022). All
the news that’s fit to fabricate: Ai-generated text as a
tool of media misinformation. Journal of Experimen-
tal Political Science, 9(1):104–117.
Krishna, K., Song, Y., Karpinska, M., Wieting, J., and
Iyyer, M. (2023). Paraphrasing evades detectors of ai-
generated text, but retrieval is an effective defense. In
Advances in Neural Information Processing Systems,
volume 36, pages 27469–27500.
Li, T., Das, S., Lee, H.-P. H., Wang, D., Yao, B., and Zhang,
Z. (2024). Human-centered privacy research in the age
of large language models. In Extended Abstracts of
the CHI Conference on Human Factors in Computing
Systems, CHI EA ’24.
Meier, R. (2024). LLM-Aided Social Media Influence Op-
erations. In Large Language Models in Cybersecu-
rity: Threats, Exposure and Mitigation, pages 105–
112. Springer Nature Switzerland.
Mihai, I.-C. (2023). Editorial: The transformative im-
pact of artificial intelligence on cybersecurity. Inter-
national Journal of Information Security and Cyber-
crime, 12(1):9–10.
Mitchell, E., Lee, Y., Khazatsky, A., Manning, C. D., and
Finn, C. (2023). DetectGPT: Zero-shot machine-
generated text detection using probability curvature.
In Proceedings of the 40th International Conference
on Machine Learning, volume 202.
Nguyen, T. N. and Choo, R. (2021). Human-in-the-
loop xai-enabled vulnerability detection, investiga-
tion, and mitigation. In 2021 36th IEEE/ACM Interna-
tional Conference on Automated Software Engineer-
ing (ASE), pages 1210–1212.
OpenAI (2024). Hello GPT-4o | OpenAI. https://
openai.com/index/hello-gpt-4o/, 2024.11.07.
Orenstrakh, M. S., Karnalim, O., Su
´
arez, C. A., and Liut,
M. (2024). Detecting LLM-Generated Text in Com-
puting Education: Comparative Study for ChatGPT
Cases. In 2024 IEEE 48th Annual Computers, Soft-
ware, and Applications Conference, pages 121–126.
Otieno, D. O., Siami Namin, A., and Jones, K. S. (2023).
The application of the bert transformer model for
phishing email classification. In 2023 IEEE 47th An-
nual Computers, Software, and Applications Confer-
ence (COMPSAC), pages 1303–1310.
Roy, S. S., Thota, P., Naragam, K. V., and Nilizadeh, S.
(2024). From Chatbots to Phishbots?: Phishing Scam
Generation in Commercial Large Language Models.
In 2024 IEEE Symposium on Security and Privacy
(SP), pages 36–54. IEEE Computer Society.
Shimada, H. and Kimura, M. (2024). A method for dis-
tinguishing model generated text and human written
text. Journal of Advances in Information Technology,
15:714–722.
Singh, S., Cornell, K., and Vaishnav, L. (2025). The hid-
den dangers of publicly accessible llms: A case study
on gab ai. In Digital Forensics and Cyber Crime.
Springer Nature Switzerland. To be published.
Sun, Y., He, J., Cui, L., Lei, S., and Lu, C.-T. (2024).
Exploring the deceptive power of llm-generated fake
news: A study of real-world detection challenges.
10.48550/arXiv.2403.18249.
Wilner, A., Jeffery, A., Lalor, J., Matthews, K., Robinson,
K., Rosolska, A., and Yorgoro, C. (2019). On the so-
cial science of ransomware: Technology, security, and
society. Comparative Strategy, 38(4):347–370.
Wu, J., Guo, J., and Hooi, B. (2024). Fake News in Sheep’s
Clothing: Robust Fake News Detection Against LLM-
Empowered Style Attacks. In Proceedings of the 30th
ACM SIGKDD Conference on Knowledge Discovery
and Data Mining, KDD ’24, pages 3367–3378.
Yu, Z., Liu, X., Liang, S., Cameron, Z., Xiao, C., and
Zhang, N. (2024). Don’t listen to me: Understand-
ing and exploring jailbreak prompts of large lan-
guage models. In 33rd USENIX Security Symposium
(USENIX Security 24), pages 4675–4692.
Zellers, R., Holtzman, A., Rashkin, H., Bisk, Y., Farhadi,
A., Roesner, F., and Choi, Y. (2019). Defending
against neural fake news. In Advances in Neural In-
formation Processing Systems, volume 32.
C¸ etin, O., Ekmekcioglu, E., Arief, B., and Hernandez-
Castro, J. (2024). An Empirical Evaluation of Large
Language Models in Static Code Analysis for PHP
Vulnerability Detection. JUCS - Journal of Univer-
sal Computer Science, 30(9):1163–1183.
Navigating the Security Challenges of LLMs: Positioning Target-Side Defenses and Identifying Research Gaps
247