kth.sePublications KTH
Change search
Link to record
Permanent link

Direct link
Publications (5 of 5) Show all publications
Cipollone, D., Wang, C., Scazzariello, M., Ferlin, S., Izadi, M., Kostic, D. & Chiesa, M. (2025). Automating the Detection of Code Vulnerabilities by Analyzing GitHub Issues. In: Proceedings of IEEE/ACM International Workshop on Large Language Models for Code 2025, LLM4Code 2025: . Paper presented at 2025 IEEE/ACM International Workshop on Large Language Models for Code, LLM4Code 2025, Ottawa, ON, Canada, May 3, 2025. Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Automating the Detection of Code Vulnerabilities by Analyzing GitHub Issues
Show others...
2025 (English)In: Proceedings of IEEE/ACM International Workshop on Large Language Models for Code 2025, LLM4Code 2025, Institute of Electrical and Electronics Engineers (IEEE), 2025Conference paper, Published paper (Refereed)
Abstract [en]

In today’s digital landscape, the importance of timely and accurate vulnerability detection has significantly increased. This paper presents a novel approach that leverages transformer-based models and machine learning techniques to automate the identification of software vulnerabilities by analyzing GitHub issues. We introduce a new dataset specifically designed for classifying GitHub issues relevant to vulnerability detection. We then examine various classification techniques to determine their effectiveness. The results demonstrate the potential of this approach for real-world application in early vulnerability detection, which could substantially reduce the window of exploitation for software vulnerabilities. This research makes a key contribution to the field by providing a scalable and computationally efficient framework for automated detection, enabling the prevention of compromised software usage before official notifications. This work has the potential to enhance the security of open-source software ecosystems.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Vulnerability Detection, Transformer-based Models, Large Language Models, LLMs, Embedding Models
National Category
Computer Systems Computer Sciences Computer Vision and Learning Systems
Identifiers
urn:nbn:se:kth:diva-374904 (URN)10.1109/LLM4Code66737.2025.00010 (DOI)001554529600006 ()2-s2.0-105009082420 (Scopus ID)979-8-3315-2615-3 (ISBN)
Conference
2025 IEEE/ACM International Workshop on Large Language Models for Code, LLM4Code 2025, Ottawa, ON, Canada, May 3, 2025
Projects
Digital Futures
Funder
Knut and Alice Wallenberg FoundationVinnova, 2023-03003Swedish Research Council, 2021-0421
Note

Part of ISBN 979-8-3315-2615-3

QC 20260108

Available from: 2026-01-07 Created: 2026-01-07 Last updated: 2026-01-08Bibliographically approved
Puccioni, L., Farshin, A., Scazzariello, M., Wang, C., Chiesa, M. & Kostic, D. (2025). Deriving Coding-Specific Sub-Models from LLMs using Resource-Efficient Pruning. In: Proceedings of IEEE/ACM International Workshop on Large Language Models for Code 2025, LLM4Code 2025: . Paper presented at 2025 IEEE/ACM International Workshop on Large Language Models for Code, LLM4Code 2025, Ottawa, ON, Canada, May 3, 2025. Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Deriving Coding-Specific Sub-Models from LLMs using Resource-Efficient Pruning
Show others...
2025 (English)In: Proceedings of IEEE/ACM International Workshop on Large Language Models for Code 2025, LLM4Code 2025, Institute of Electrical and Electronics Engineers (IEEE), 2025Conference paper, Published paper (Refereed)
Abstract [en]

Large Language Models (LLMs) have demonstrated their exceptional performance in various complex code generation tasks. However, their broader adoption is limited by significant computational demands and high resource requirements, particularly memory and processing power. To mitigate such requirements, model pruning techniques are used to create more compact models with significantly fewer parameters. However, current approaches do not focus on the efficient extraction of programming-language-specific sub-models. In this work, we explore the idea of efficiently deriving coding-specific sub-models through unstructured pruning (i.e., Wanda). We investigate the impact of different domain-specific calibration datasets on pruning outcomes across three distinct domains and extend our analysis to extracting four language-specific sub-models: Python, Java, C++, and JavaScript. We demonstrate that it is possible to efficiently extract programming-language-specific sub-models using appropriate calibration datasets while maintaining acceptable accuracy w.r.t. full models. We are also the first to provide analytical evidence that domain-specific tasks activate distinct regions within LLMs, supporting the creation of specialized sub-models through unstructured pruning. We believe that this work has significant potential to enhance LLM accessibility for coding by reducing computational requirements to enable local execution on consumer-grade hardware, and supporting faster inference times critical for real-time development feedback.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Large Language Models, LLMs, pruning, code
National Category
Computer Systems Computer Sciences Computer Vision and Learning Systems
Identifiers
urn:nbn:se:kth:diva-374905 (URN)10.1109/LLM4Code66737.2025.00028 (DOI)001554529600024 ()2-s2.0-105009110881 (Scopus ID)
Conference
2025 IEEE/ACM International Workshop on Large Language Models for Code, LLM4Code 2025, Ottawa, ON, Canada, May 3, 2025
Projects
Digital Futures
Funder
Knut and Alice Wallenberg FoundationVinnova, 2023-03003Swedish Research Council, 2021-0421
Note

Part of ISBN 979-8-3315-2615-3

QC 20260108

Available from: 2026-01-07 Created: 2026-01-07 Last updated: 2026-01-08Bibliographically approved
Wang, C., Scazzariello, M. & Chiesa, M. (2025). From Scientific Texts to Verifiable Code: Automating the Process with Transformers. In: 2025 IEEE/ACM International Workshop On Large Language Models For Code, LLM4CODE: . Paper presented at 2025 International Workshop on Large Language Models for Code-LLM4Code, MAY 03, 2025, Ottawa, CANADA (pp. 213-216). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>From Scientific Texts to Verifiable Code: Automating the Process with Transformers
2025 (English)In: 2025 IEEE/ACM International Workshop On Large Language Models For Code, LLM4CODE, Institute of Electrical and Electronics Engineers (IEEE) , 2025, p. 213-216Conference paper, Published paper (Refereed)
Abstract [en]

Despite the vast body of research literature proposing algorithms with formal guarantees, the amount of verifiable code in today's systems remains minimal. This discrepancy stems from the inherent difficulty of verifying code, particularly due to the time-consuming nature and strict formalism of proof details that formal verification tools require. However, the emergence of Transformers in Large Language Models presents a promising solution to this challenge. In this position paper, we believe that Transformers have the potential to read research papers that propose algorithms with formal proofs and translate these proofs into verifiable code. We leverage Transformers to first build a formal structure of the proof using the original text from the paper, and then to handle the tedious, low-level aspects of proofs that are often omitted by humans. We argue that this approach can significantly reduce the barrier to formal verification. The above idea of reading papers to write verifiable code opens new avenues for automating the verification of complex systems, enabling a future where formally verified algorithms from academic research can more seamlessly transition into real-world software systems, thereby improving code reliability and security.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-375481 (URN)10.1109/LLM4Code66737.2025.00032 (DOI)001554529600028 ()2-s2.0-105009035671 (Scopus ID)
Conference
2025 International Workshop on Large Language Models for Code-LLM4Code, MAY 03, 2025, Ottawa, CANADA
Note

Part of ISBN 979-8-3315-2616-0; 979-8-3315-2615-3

QC 20260126

Available from: 2026-01-26 Created: 2026-01-26 Last updated: 2026-01-27Bibliographically approved
Wang, C., Scazzariello, M., Farshin, A., Ferlin, S., Kostic, D. & Chiesa, M. (2024). NetConfEval: Can LLMs Facilitate Network Configuration?. Proceedings of the ACM on Networking, 2(CoNEXT2), Article ID 7.
Open this publication in new window or tab >>NetConfEval: Can LLMs Facilitate Network Configuration?
Show others...
2024 (English)In: Proceedings of the ACM on Networking, ISSN 2834-5509, Vol. 2, no CoNEXT2, article id 7Article in journal (Refereed) Published
Abstract [en]

This paper explores opportunities to utilize Large Language Models (LLMs) to make network configuration human-friendly, simplifying the configuration of network devices & development of routing algorithms and minimizing errors. We design a set of benchmarks (NetConfEval) to examine the effectiveness of different models in facilitating and automating network configuration. More specifically, we focus on the scenarios where LLMs translate high-level policies, requirements, and descriptions (i.e., specified in natural language) into low-level network configurations & Python code. NetConfEval considers four tasks that could potentially facilitate network configuration, such as (i) generating high-level requirements into a formal specification format, (ii) generating API/function calls from high-level requirements, (iii) developing routing algorithms based on high-level descriptions, and (iv) generating low-level configuration for existing and new protocols based on input documentation. Learning from the results of our study, we propose a set of principles to design LLM-based systems to configure networks. Finally, we present two GPT-4-based prototypes to (i) automatically configure P4-enabled devices from a set of high-level requirements and (ii) integrate LLMs into existing network synthesizers.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2024
Keywords
benchmark, code generation, function calling, large language models (llms), network configuration, network synthesizer, p4, rag, routing algorithms
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-357124 (URN)10.1145/3656296 (DOI)
Projects
Digital Futures
Funder
Vinnova, 2023-03003EU, European Research Council, 770889Swedish Research Council, 2021-0421
Note

QC 20241211

Available from: 2024-12-04 Created: 2024-12-04 Last updated: 2024-12-11Bibliographically approved
Wang, C., Scazzariello, M., Farshin, A., Kostic, D. & Chiesa, M.Making Network Configuration Human Friendly.
Open this publication in new window or tab >>Making Network Configuration Human Friendly
Show others...
(English)Manuscript (preprint) (Other academic)
Abstract [en]

This paper explores opportunities to utilize Large Language Models (LLMs) to make network configuration human-friendly, simplifying the configuration of network devices and minimizing errors. We examine the effectiveness of these models in translating high-level policies and requirements (i.e., specified in natural language) into low-level network APIs, which requires understanding the hardware and protocols. More specifically, we propose NETBUDDY for generating network configurations from scratch and modifying them at runtime. NETBUDDY splits the generation of network configurations into fine-grained steps and relies on self-healing code-generation approaches to better take advantage of the full potential of LLMs. We first thoroughly examine the challenges of using these models to produce a fully functional & correct configuration, and then evaluate the feasibility of realizing NETBUDDY by building a proof-of-concept solution using GPT-4 to translate a set of high-level requirements into P4 and BGP configurations and run them using the Kathará network emulator.

Keywords
Network Configuration, Large Language Models (LLMs), GPT-4, P4, BGP, Kathará
National Category
Computer Systems Communication Systems
Research subject
Computer Science; Information and Communication Technology
Identifiers
urn:nbn:se:kth:diva-336534 (URN)
Projects
Digital Futures
Funder
Swedish Research Council, 2021-04212
Note

QC 20230913

Available from: 2023-09-12 Created: 2023-09-12 Last updated: 2023-09-13Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0009-0000-4604-1180

Search in DiVA

Show all publications