kth.sePublications KTH
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Deriving Coding-Specific Sub-Models from LLMs using Resource-Efficient Pruning
Spotify.
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Software and Computer systems, SCS, Network Systems Laboratory (NS Lab).ORCID iD: 0000-0001-5083-4052
RISE Research Institutes of Sweden.ORCID iD: 0000-0002-9780-873X
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Software and Computer systems, SCS.ORCID iD: 0009-0000-4604-1180
Show others and affiliations
2025 (English)In: Proceedings of IEEE/ACM International Workshop on Large Language Models for Code 2025, LLM4Code 2025, Institute of Electrical and Electronics Engineers (IEEE), 2025Conference paper, Published paper (Refereed)
Abstract [en]

Large Language Models (LLMs) have demonstrated their exceptional performance in various complex code generation tasks. However, their broader adoption is limited by significant computational demands and high resource requirements, particularly memory and processing power. To mitigate such requirements, model pruning techniques are used to create more compact models with significantly fewer parameters. However, current approaches do not focus on the efficient extraction of programming-language-specific sub-models. In this work, we explore the idea of efficiently deriving coding-specific sub-models through unstructured pruning (i.e., Wanda). We investigate the impact of different domain-specific calibration datasets on pruning outcomes across three distinct domains and extend our analysis to extracting four language-specific sub-models: Python, Java, C++, and JavaScript. We demonstrate that it is possible to efficiently extract programming-language-specific sub-models using appropriate calibration datasets while maintaining acceptable accuracy w.r.t. full models. We are also the first to provide analytical evidence that domain-specific tasks activate distinct regions within LLMs, supporting the creation of specialized sub-models through unstructured pruning. We believe that this work has significant potential to enhance LLM accessibility for coding by reducing computational requirements to enable local execution on consumer-grade hardware, and supporting faster inference times critical for real-time development feedback.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025.
Keywords [en]
Large Language Models, LLMs, pruning, code
National Category
Computer Systems Computer Sciences Computer Vision and Learning Systems
Identifiers
URN: urn:nbn:se:kth:diva-374905DOI: 10.1109/LLM4Code66737.2025.00028ISI: 001554529600024Scopus ID: 2-s2.0-105009110881OAI: oai:DiVA.org:kth-374905DiVA, id: diva2:2025739
Conference
2025 IEEE/ACM International Workshop on Large Language Models for Code, LLM4Code 2025, Ottawa, ON, Canada, May 3, 2025
Projects
Digital Futures
Funder
Knut and Alice Wallenberg FoundationVinnova, 2023-03003Swedish Research Council, 2021-0421
Note

Part of ISBN 979-8-3315-2615-3

QC 20260108

Available from: 2026-01-07 Created: 2026-01-07 Last updated: 2026-01-08Bibliographically approved

Open Access in DiVA

fulltext(5715 kB)35 downloads
File information
File name FULLTEXT01.pdfFile size 5715 kBChecksum SHA-512
cb28f4c2e4adef2b26b94c91f240b5439cc6c4a9bed46c81a75c55c0cfcc4f962075e4c8e91b94d4bcec24ae016e680474e74b84c3a463f59d79f4f5798e5021
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopus

Authority records

Farshin, AlirezaScazzariello, MarianoWang, ChangjieChiesa, MarcoKostic, Dejan

Search in DiVA

By author/editor
Farshin, AlirezaScazzariello, MarianoWang, ChangjieChiesa, MarcoKostic, Dejan
By organisation
Network Systems Laboratory (NS Lab)Software and Computer systems, SCS
Computer SystemsComputer SciencesComputer Vision and Learning Systems

Search outside of DiVA

GoogleGoogle Scholar
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 5104 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf