Large language models (LLMs) have dramatically reshaped the field of natural language processing, presenting groundbreaking advancements in many areas, from chatbots to content creation. However, with the increasing adoption of these sophisticated models, it is crucial to scrutinize the vulnerabilities associated with their training and inference stages. This comprehensive analysis highlights the critical threats and inefficiencies inherent to these processes and emphasizes the need for robust countermeasures. This paper presents an extensive study of training and inference time vulnerabilities in Large Language Models (LLMs), specifically focusing on poisoning, backdoor, paraphrasing, and spoofing attacks. We introduce novel evaluation frameworks and detection mechanisms for each attack type. Our experimental results across multiple attack vectors demonstrate varying degrees of model susceptibility and reveal critical security implications. The proposed defensive mechanisms showcase impressive model performance, highlighted by consistent successful evaluation outcomes.
Large Language Models (LLMs) Cyber-security Adversarial attacks Critical vulnerabilities
Birincil Dil | İngilizce |
---|---|
Konular | Bilgisayar Sistem Yazılımı |
Bölüm | Research Articles |
Yazarlar | |
Erken Görünüm Tarihi | 29 Haziran 2025 |
Yayımlanma Tarihi | 29 Haziran 2025 |
Gönderilme Tarihi | 9 Şubat 2025 |
Kabul Tarihi | 17 Mart 2025 |
Yayımlandığı Sayı | Yıl 2025 Cilt: 6 Sayı: 1 |