Description

Identifying system vulnerabilities is a perpetual cat-and-mouse game between software developers and security researchers. As software ecosystems grow in complexity and scale, the discovery of zero-day vulnerabilities has become an increasingly arduous and resource-intensive task. Historically, vulnerability detection has relied on manual code audits and automated tools like static analyzers and fuzzers. However, traditional static analysis often suffers from high false-positive rates, while dynamic fuzzing struggles to reach deep code paths or understand complex semantic logic. These legacy methods are frequently unable to capture the nuanced developer intent, often missing subtle bugs that do not follow well-defined, known-vulnerability patterns or signatures.

To address these limitations, Large Language Models (LLMs) have been integrated into the vulnerability discovery pipeline to provide semantic awareness and reasoning capabilities. These models can analyze source code, binary instructions, and execution traces to identify potential flaws across various software entities—including kernels, drivers, and network protocols. By leveraging the vast knowledge embedded in LLMs, security tools can now predict vulnerabilities by understanding the "context" of code rather than just its syntax. However, applying LLMs to system-level security faces hurdles such as limited context windows and the difficulty of verifying model-generated reports. Despite these challenges, LLM-assisted detection is a rapidly expanding field, generally categorized into three research directions: prompt-engineered static analysis, LLM-guided fuzzing, and neural-symbolic hybrid detection models.

We are working on enhancing the precision of LLM-based vulnerability detection to make these tools more reliable for large-scale enterprise software. Our research focuses on utilizing retrieval-augmented generation (RAG) and specialized fine-tuning to reduce false positives and improve the "exploitability" assessment of discovered bugs. For example, we have developed a collaborative LLM framework that cross-references potential vulnerabilities with historical CVE data and real-world execution constraints. This methodology allows the detection system to prioritize high-risk flaws while providing developers with actionable insights and automated patching suggestions to secure the system.

Collaborators

Alumni