A systematic literature Review for Transformer-based Software Vulnerability detection
80 studies analyzed reveal transformers beat traditional ML for code security...
A team of researchers led by Fiza Naseer has published a comprehensive systematic literature review (SLR) on transformer-based software vulnerability detection, analyzing 80 studies published between 2021 and 2025. The review follows Kitchenham's SLR guidelines and examines how transformer architectures—including encoder-only (e.g., BERT), decoder-only (e.g., GPT), and encoder-decoder models—are being applied to identify security flaws in source code, logs, and smart contracts. The study reveals that transformers outperform traditional machine learning and deep learning approaches due to their superior contextual modeling and representation learning capabilities.
The review identifies several critical technical challenges that remain unaddressed. Data imbalance across vulnerability types, lack of model interpretability, scalability limitations for large codebases, and poor generalization across programming languages are highlighted as key obstacles. The authors provide a consolidated resource for researchers and practitioners, mapping prevalent benchmarks, evaluation metrics, and reference models. This SLR serves as a foundational guide for developing more reliable, precise, and interpretable transformer-based vulnerability detection systems, particularly for high-stakes domains like healthcare, government, and finance.
- 80 studies analyzed across encoder, decoder, and combined transformer architectures
- Key challenges include data imbalance, interpretability, scalability, and cross-language generalization
- Transformers applied to source code, logs, and smart contracts for vulnerability detection
Why It Matters
This SLR provides a roadmap for building more reliable AI security tools for critical software systems.