kth.sePublications KTH
Change search
Link to record
Permanent link

Direct link
Scazzariello, MarianoORCID iD iconorcid.org/0000-0002-9780-873X
Publications (10 of 13) Show all publications
Cipollone, D., Wang, C., Scazzariello, M., Ferlin, S., Izadi, M., Kostic, D. & Chiesa, M. (2025). Automating the Detection of Code Vulnerabilities by Analyzing GitHub Issues. In: Proceedings of IEEE/ACM International Workshop on Large Language Models for Code 2025, LLM4Code 2025: . Paper presented at 2025 IEEE/ACM International Workshop on Large Language Models for Code, LLM4Code 2025, Ottawa, ON, Canada, May 3, 2025. Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Automating the Detection of Code Vulnerabilities by Analyzing GitHub Issues
Show others...
2025 (English)In: Proceedings of IEEE/ACM International Workshop on Large Language Models for Code 2025, LLM4Code 2025, Institute of Electrical and Electronics Engineers (IEEE), 2025Conference paper, Published paper (Refereed)
Abstract [en]

In today’s digital landscape, the importance of timely and accurate vulnerability detection has significantly increased. This paper presents a novel approach that leverages transformer-based models and machine learning techniques to automate the identification of software vulnerabilities by analyzing GitHub issues. We introduce a new dataset specifically designed for classifying GitHub issues relevant to vulnerability detection. We then examine various classification techniques to determine their effectiveness. The results demonstrate the potential of this approach for real-world application in early vulnerability detection, which could substantially reduce the window of exploitation for software vulnerabilities. This research makes a key contribution to the field by providing a scalable and computationally efficient framework for automated detection, enabling the prevention of compromised software usage before official notifications. This work has the potential to enhance the security of open-source software ecosystems.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Vulnerability Detection, Transformer-based Models, Large Language Models, LLMs, Embedding Models
National Category
Computer Systems Computer Sciences Computer Vision and Learning Systems
Identifiers
urn:nbn:se:kth:diva-374904 (URN)10.1109/LLM4Code66737.2025.00010 (DOI)001554529600006 ()2-s2.0-105009082420 (Scopus ID)979-8-3315-2615-3 (ISBN)
Conference
2025 IEEE/ACM International Workshop on Large Language Models for Code, LLM4Code 2025, Ottawa, ON, Canada, May 3, 2025
Projects
Digital Futures
Funder
Knut and Alice Wallenberg FoundationVinnova, 2023-03003Swedish Research Council, 2021-0421
Note

Part of ISBN 979-8-3315-2615-3

QC 20260108

Available from: 2026-01-07 Created: 2026-01-07 Last updated: 2026-01-08Bibliographically approved
Puccioni, L., Farshin, A., Scazzariello, M., Wang, C., Chiesa, M. & Kostic, D. (2025). Deriving Coding-Specific Sub-Models from LLMs using Resource-Efficient Pruning. In: Proceedings of IEEE/ACM International Workshop on Large Language Models for Code 2025, LLM4Code 2025: . Paper presented at 2025 IEEE/ACM International Workshop on Large Language Models for Code, LLM4Code 2025, Ottawa, ON, Canada, May 3, 2025. Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Deriving Coding-Specific Sub-Models from LLMs using Resource-Efficient Pruning
Show others...
2025 (English)In: Proceedings of IEEE/ACM International Workshop on Large Language Models for Code 2025, LLM4Code 2025, Institute of Electrical and Electronics Engineers (IEEE), 2025Conference paper, Published paper (Refereed)
Abstract [en]

Large Language Models (LLMs) have demonstrated their exceptional performance in various complex code generation tasks. However, their broader adoption is limited by significant computational demands and high resource requirements, particularly memory and processing power. To mitigate such requirements, model pruning techniques are used to create more compact models with significantly fewer parameters. However, current approaches do not focus on the efficient extraction of programming-language-specific sub-models. In this work, we explore the idea of efficiently deriving coding-specific sub-models through unstructured pruning (i.e., Wanda). We investigate the impact of different domain-specific calibration datasets on pruning outcomes across three distinct domains and extend our analysis to extracting four language-specific sub-models: Python, Java, C++, and JavaScript. We demonstrate that it is possible to efficiently extract programming-language-specific sub-models using appropriate calibration datasets while maintaining acceptable accuracy w.r.t. full models. We are also the first to provide analytical evidence that domain-specific tasks activate distinct regions within LLMs, supporting the creation of specialized sub-models through unstructured pruning. We believe that this work has significant potential to enhance LLM accessibility for coding by reducing computational requirements to enable local execution on consumer-grade hardware, and supporting faster inference times critical for real-time development feedback.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Large Language Models, LLMs, pruning, code
National Category
Computer Systems Computer Sciences Computer Vision and Learning Systems
Identifiers
urn:nbn:se:kth:diva-374905 (URN)10.1109/LLM4Code66737.2025.00028 (DOI)001554529600024 ()2-s2.0-105009110881 (Scopus ID)
Conference
2025 IEEE/ACM International Workshop on Large Language Models for Code, LLM4Code 2025, Ottawa, ON, Canada, May 3, 2025
Projects
Digital Futures
Funder
Knut and Alice Wallenberg FoundationVinnova, 2023-03003Swedish Research Council, 2021-0421
Note

Part of ISBN 979-8-3315-2615-3

QC 20260108

Available from: 2026-01-07 Created: 2026-01-07 Last updated: 2026-01-08Bibliographically approved
Wang, C., Scazzariello, M. & Chiesa, M. (2025). From Scientific Texts to Verifiable Code: Automating the Process with Transformers. In: 2025 IEEE/ACM International Workshop On Large Language Models For Code, LLM4CODE: . Paper presented at 2025 International Workshop on Large Language Models for Code-LLM4Code, MAY 03, 2025, Ottawa, CANADA (pp. 213-216). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>From Scientific Texts to Verifiable Code: Automating the Process with Transformers
2025 (English)In: 2025 IEEE/ACM International Workshop On Large Language Models For Code, LLM4CODE, Institute of Electrical and Electronics Engineers (IEEE) , 2025, p. 213-216Conference paper, Published paper (Refereed)
Abstract [en]

Despite the vast body of research literature proposing algorithms with formal guarantees, the amount of verifiable code in today's systems remains minimal. This discrepancy stems from the inherent difficulty of verifying code, particularly due to the time-consuming nature and strict formalism of proof details that formal verification tools require. However, the emergence of Transformers in Large Language Models presents a promising solution to this challenge. In this position paper, we believe that Transformers have the potential to read research papers that propose algorithms with formal proofs and translate these proofs into verifiable code. We leverage Transformers to first build a formal structure of the proof using the original text from the paper, and then to handle the tedious, low-level aspects of proofs that are often omitted by humans. We argue that this approach can significantly reduce the barrier to formal verification. The above idea of reading papers to write verifiable code opens new avenues for automating the verification of complex systems, enabling a future where formally verified algorithms from academic research can more seamlessly transition into real-world software systems, thereby improving code reliability and security.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-375481 (URN)10.1109/LLM4Code66737.2025.00032 (DOI)001554529600028 ()2-s2.0-105009035671 (Scopus ID)
Conference
2025 International Workshop on Large Language Models for Code-LLM4Code, MAY 03, 2025, Ottawa, CANADA
Note

Part of ISBN 979-8-3315-2616-0; 979-8-3315-2615-3

QC 20260126

Available from: 2026-01-26 Created: 2026-01-26 Last updated: 2026-01-27Bibliographically approved
Polverini, M., Cianfrani, A., Caiazzi, T., Scazzariello, M., Abdelsalam, A., Filsfils, C. & Camarillo, P. (2024). Achieving Best-path Selection at Line Rate through the SRv6 Live-Live Behavior. In: Proceedings of IEEE/IFIP Network Operations and Management Symposium 2024, NOMS 2024: . Paper presented at 2024 IEEE/IFIP Network Operations and Management Symposium, NOMS 2024, Seoul, Korea, May 6 2024 - May 10 2024. Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Achieving Best-path Selection at Line Rate through the SRv6 Live-Live Behavior
Show others...
2024 (English)In: Proceedings of IEEE/IFIP Network Operations and Management Symposium 2024, NOMS 2024, Institute of Electrical and Electronics Engineers (IEEE) , 2024Conference paper, Published paper (Refereed)
Abstract [en]

The network programming model of the Segment Routing (SRv6) architecture offers the possibility to define new functions aiming at improving the network performance. In this paper we introduce SRv6 Live-Live, a new behavior for the SRv6 data plane. SRv6 Live-Live is based on two primitives: i) traffic duplication, performed at the ingress node, and ii) the traffic de-duplication, executed at the egress node. The proposed behavior is suitable for the service provisioning of traffic flows having stringent requirements in terms of reliability, low delay and high throughput. Our preliminary performance evaluation, conducted in an emulated environment and realized by using a prototype implementation based on P4, shows that SRv6 Live-Live enhances the performance of the selected traffic flows in challenging network scenarios, characterized by high level of packet corruption/loss and large values of bandwidth-delay products.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2024
Keywords
network programmability, P4, SRv6 behavior
National Category
Communication Systems Computer Sciences
Identifiers
urn:nbn:se:kth:diva-351007 (URN)10.1109/NOMS59830.2024.10575338 (DOI)001270140300091 ()2-s2.0-85198336593 (Scopus ID)
Conference
2024 IEEE/IFIP Network Operations and Management Symposium, NOMS 2024, Seoul, Korea, May 6 2024 - May 10 2024
Note

Part of ISBN 9798350327939

QC 20240725

Available from: 2024-07-24 Created: 2024-07-24 Last updated: 2024-10-01Bibliographically approved
Scazzariello, M., Caiazzi, T. & Chiesa, M. (2024). Deliberately Congesting a Switch for Better Network Functions Performance. In: 2024 IEEE 32nd International Conference on Network Protocols, ICNP 2024: . Paper presented at 32nd IEEE International Conference on Network Protocols, ICNP 2024, Charleroi, Belgium, Oct 28 2024 - Oct 31 2024. Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Deliberately Congesting a Switch for Better Network Functions Performance
2024 (English)In: 2024 IEEE 32nd International Conference on Network Protocols, ICNP 2024, Institute of Electrical and Electronics Engineers (IEEE) , 2024Conference paper, Published paper (Refereed)
Abstract [en]

Traditional wisdom suggests maintaining minimal occupancy in the port queues of network devices to prevent packet delays or drops en route to their destination. In this paper, however, we explore the unconventional idea of deliberately congesting the queues of a network device to enhance the performance of a Network Function (NF) deployment. The key intuition behind this approach is to utilize the existing memory available in the switch queues to store packet payloads while their headers are processed on an external NF processor. We present two techniques for congesting a port on a switch: i) self-clocking packet recirculation, which recirculates packets within the switch to automatically achieve the correct queuing delay, and ii) a proportional controller using multicast forwarding, which adjusts the rate of packet forwarding based on the level of congestion. We evaluate our approaches both in simulations and a prototype.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2024
National Category
Telecommunications Communication Systems
Identifiers
urn:nbn:se:kth:diva-360560 (URN)10.1109/ICNP61940.2024.10858581 (DOI)001434856900079 ()2-s2.0-85218065506 (Scopus ID)
Conference
32nd IEEE International Conference on Network Protocols, ICNP 2024, Charleroi, Belgium, Oct 28 2024 - Oct 31 2024
Note

Part of ISBN 9798350351712

QC 20250226

Available from: 2025-02-26 Created: 2025-02-26 Last updated: 2025-12-08Bibliographically approved
Ghasemirahni, H., Farshin, A., Scazzariello, M., Chiesa, M. & Kostic, D. (2024). Deploying Stateful Network Functions Efficiently using Large Language Models. In: EuroMLSys 2024 - Proceedings of the 2024 4th Workshop on Machine Learning and Systems: . Paper presented at 4th Workshop on Machine Learning and Systems, EuroMLSys 2024, held in conjunction with ACM EuroSys 2024, Athens, Greece, Apr 22 2024 (pp. 28-38). Association for Computing Machinery (ACM)
Open this publication in new window or tab >>Deploying Stateful Network Functions Efficiently using Large Language Models
Show others...
2024 (English)In: EuroMLSys 2024 - Proceedings of the 2024 4th Workshop on Machine Learning and Systems, Association for Computing Machinery (ACM) , 2024, p. 28-38Conference paper, Published paper (Refereed)
Abstract [en]

Stateful network functions are increasingly used in data centers. However, their scalability remains a significant challenge since parallelizing packet processing across multiple cores requires careful configuration t o avoid compromising the application’s semantics or performance. This challenge is particularly important when deploying multiple stateful functions on multi-core servers. This paper proposes FlowMage, a system that leverages Large Language Models (LLMs) to perform code analysis and extract essential information from stateful network functions (NFs) prior to their deployment on a server. FlowMage uses this data to find an efficient configuration of an NF chain that maximizes performance while preserving the semantics of the NF chain. Our evaluation shows that, utilizing GPT-4, FlowMage is able to find and apply optimized configuration when deploying stateful NFs chain on a server, resulting in significant p erformance i mprovement (up t o 1 1×) in comparison to the default configuration of the system.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2024
Keywords
Intra-Server Load Balancing, LLMs, RSS Configuration, Stateful Network Functions, Static Code Analysis
National Category
Computer Systems Communication Systems
Research subject
Computer Science
Identifiers
urn:nbn:se:kth:diva-346539 (URN)10.1145/3642970.3655836 (DOI)001221134800004 ()2-s2.0-85192276579 (Scopus ID)
Conference
4th Workshop on Machine Learning and Systems, EuroMLSys 2024, held in conjunction with ACM EuroSys 2024, Athens, Greece, Apr 22 2024
Projects
ULTRA
Funder
EU, Horizon 2020, 770889
Note

Part of ISBN 979-840070541-0

QC 20240520

Available from: 2024-05-16 Created: 2024-05-16 Last updated: 2024-12-06Bibliographically approved
Ghasemirahni, H., Farshin, A., Scazzariello, M., Maguire Jr., G. Q., Kostic, D. & Chiesa, M. (2024). FAJITA: Stateful Packet Processing at 100 Million pps. Proceedings of the ACM on Networking, 2(CoNEXT3), 1-22
Open this publication in new window or tab >>FAJITA: Stateful Packet Processing at 100 Million pps
Show others...
2024 (English)In: Proceedings of the ACM on Networking, E-ISSN 2834-5509, Vol. 2, no CoNEXT3, p. 1-22Article in journal (Refereed) Published
Abstract [en]

Data centers increasingly utilize commodity servers to deploy low-latency Network Functions (NFs). However, the emergence of multi-hundred-gigabit-per-second network interface cards (NICs) has drastically increased the performance expected from commodity servers. Additionally, recently introduced systems that store packet payloads in temporary off-CPU locations (e.g., programmable switches, NICs, and RDMA servers) further increase the load on NF servers, making packet processing even more challenging. This paper demonstrates existing bottlenecks and challenges of state-of-the-art stateful packet processing frameworks and proposes a system, called FAJITA, to tackle these challenges & accelerate stateful packet processing on commodity hardware. FAJITA proposes an optimized processing pipeline for stateful network functions to minimize memory accesses and overcome the overheads of accessing shared data structures while ensuring efficient batch processing at every stage of the pipeline. Furthermore, FAJITA provides a performant architecture to deploy high-performance network functions service chains containing stateful elements with different state granularities. FAJITA improves the throughput and latency of high-speed stateful network functions by ~2.43x compared to the most performant state-of-the-art solutions, enabling commodity hardware to process up to ~178 Million 64-B packets per second (pps) using 16 cores.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2024
Keywords
packet processing frameworks, stateful network functions
National Category
Communication Systems Computer Systems
Research subject
Computer Science
Identifiers
urn:nbn:se:kth:diva-357087 (URN)10.1145/3676861 (DOI)
Projects
ULTRA
Funder
EU, Horizon 2020, 770889Swedish Research Council, 2021-04212Vinnova, 2023-03003
Note

QC 20241206

Available from: 2024-12-04 Created: 2024-12-04 Last updated: 2024-12-06Bibliographically approved
Wang, C., Scazzariello, M., Farshin, A., Ferlin, S., Kostic, D. & Chiesa, M. (2024). NetConfEval: Can LLMs Facilitate Network Configuration?. Proceedings of the ACM on Networking, 2(CoNEXT2), Article ID 7.
Open this publication in new window or tab >>NetConfEval: Can LLMs Facilitate Network Configuration?
Show others...
2024 (English)In: Proceedings of the ACM on Networking, ISSN 2834-5509, Vol. 2, no CoNEXT2, article id 7Article in journal (Refereed) Published
Abstract [en]

This paper explores opportunities to utilize Large Language Models (LLMs) to make network configuration human-friendly, simplifying the configuration of network devices & development of routing algorithms and minimizing errors. We design a set of benchmarks (NetConfEval) to examine the effectiveness of different models in facilitating and automating network configuration. More specifically, we focus on the scenarios where LLMs translate high-level policies, requirements, and descriptions (i.e., specified in natural language) into low-level network configurations & Python code. NetConfEval considers four tasks that could potentially facilitate network configuration, such as (i) generating high-level requirements into a formal specification format, (ii) generating API/function calls from high-level requirements, (iii) developing routing algorithms based on high-level descriptions, and (iv) generating low-level configuration for existing and new protocols based on input documentation. Learning from the results of our study, we propose a set of principles to design LLM-based systems to configure networks. Finally, we present two GPT-4-based prototypes to (i) automatically configure P4-enabled devices from a set of high-level requirements and (ii) integrate LLMs into existing network synthesizers.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2024
Keywords
benchmark, code generation, function calling, large language models (llms), network configuration, network synthesizer, p4, rag, routing algorithms
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-357124 (URN)10.1145/3656296 (DOI)
Projects
Digital Futures
Funder
Vinnova, 2023-03003EU, European Research Council, 770889Swedish Research Council, 2021-0421
Note

QC 20241211

Available from: 2024-12-04 Created: 2024-12-04 Last updated: 2024-12-11Bibliographically approved
Girondi, M., Scazzariello, M., Maguire Jr., G. Q. & Kostic, D. (2024). Toward GPU-centric Networking on Commodity Hardware. In: 7th International Workshop on Edge Systems, Analytics and Networking (EdgeSys 2024),  April 22, 2024, Athens, Greece: . Paper presented at 7th International Workshop on Edge Systems, Analytics and Networking (EdgeSys 2024), April 22, 2024, Athens, Greece . New York: ACM Digital Library
Open this publication in new window or tab >>Toward GPU-centric Networking on Commodity Hardware
2024 (English)In: 7th International Workshop on Edge Systems, Analytics and Networking (EdgeSys 2024),  April 22, 2024, Athens, Greece, New York: ACM Digital Library, 2024Conference paper, Published paper (Refereed)
Abstract [en]

GPUs are emerging as the most popular accelerator for many applications, powering the core of machine learning applications. In networked GPU-accelerated applications input & output data typically traverse the CPU and the OS network stack multiple times, getting copied across the system’s main memory. These transfers increase application latency and require expensive CPU cycles, reducing the system’s efficiency, and increasing the overall response times. These inefficiencies become of greater importance in latency-bounded deployments, or with high throughput, where copy times could quickly inflate the response time of modern GPUs.We leverage the efficiency and kernel-bypass benefits of RDMA to transfer data in and out of GPUs without using any CPU cycles or synchronization. We demonstrate the ability of modern GPUs to saturate a 100-Gbps link, and evaluate the network processing timein the context of an inference serving application.

Place, publisher, year, edition, pages
New York: ACM Digital Library, 2024
Keywords
GPUs, Commodity Hardware, Inference Serving, RDMA
National Category
Communication Systems Computer Systems
Research subject
Computer Science
Identifiers
urn:nbn:se:kth:diva-345624 (URN)10.1145/3642968.3654820 (DOI)001234771200008 ()2-s2.0-85192024363 (Scopus ID)
Conference
7th International Workshop on Edge Systems, Analytics and Networking (EdgeSys 2024), April 22, 2024, Athens, Greece 
Note

QC 20240415

Part of ISBN 979-8-4007-0539-7

Available from: 2024-04-15 Created: 2024-04-15 Last updated: 2024-08-28Bibliographically approved
Scazzariello, M., Caiazzi, T., Ghasemirahni, H., Barbette, T., Kostic, D. & Chiesa, M. (2023). A High-Speed Stateful Packet Processing Approach for Tbps Programmable Switches. In: 20th USENIX Symposium on Networked Systems Designand Implementation (NSDI ’23): . Paper presented at NSDI'23 - 20th USENIX Symposium on Networked Systems Design and Implementation, April 17–19, 2023, Boston, MA, USA (pp. 1237-1255). The USENIX Association
Open this publication in new window or tab >>A High-Speed Stateful Packet Processing Approach for Tbps Programmable Switches
Show others...
2023 (English)In: 20th USENIX Symposium on Networked Systems Designand Implementation (NSDI ’23), The USENIX Association , 2023, p. 1237-1255Conference paper, Oral presentation with published abstract (Refereed)
Abstract [en]

High-speed ASIC switches hold great promise for offloading complex packet processing pipelines directly in the highspeed data-plane. Yet, a large variety of today’s packet processing pipelines, including stateful network functions andpacket schedulers, require storing some (or all the) packetsfor short amount of times in a programmatic manner. Such aprogrammable buffer feature is missing on today’s high-speedASIC switches.In this work, we present RIBOSOME, a system that extendsprogrammable switches with external memory (to store packets) and external general-purpose packet processing devicessuch as CPUs or FPGAs (to perform stateful operations). Astoday’s packet processing devices are bottlenecked by theirnetwork interface speeds, RIBOSOME carefully transmits onlythe relevant bits to these devices. RIBOSOME leverages sparebandwidth from any directly connected servers to store theincoming payloads through RDMA. Our evaluation showsthat RIBOSOME can process 300G of traffic through a stateful packet processing pipeline (e.g., firewall, load balancer,packet scheduler) by running the pipeline logic on a singleserver equipped with one 100G interface.

Place, publisher, year, edition, pages
The USENIX Association, 2023
National Category
Computer Systems Communication Systems
Identifiers
urn:nbn:se:kth:diva-326619 (URN)001066630000065 ()2-s2.0-85159326513 (Scopus ID)
Conference
NSDI'23 - 20th USENIX Symposium on Networked Systems Design and Implementation, April 17–19, 2023, Boston, MA, USA
Funder
Swedish Research Council, 2021-04212EU, European Research Council, 770889
Note

Part of proceedings ISBN 978-1-939133-33-5

QC 20230807

Available from: 2023-05-07 Created: 2023-05-07 Last updated: 2023-10-16Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0002-9780-873X

Search in DiVA

Show all publications