Öppna denna publikation i ny flik eller fönster >>Visa övriga...
2025 (Engelska)Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]
In Machine Learning (ML) and Deep Learning (DL) research, ablation studies are typically performed to provide insights into the individual contribution of different building blocks and components of an ML/DL system (e.g., a deep neural network), as well as to justify that certain additions or modifications to an existing ML/DL system can result in the proposed improved performance. Although dedicated frameworks for performing ablation studies have been introduced in recent years, conducting such experiments is still associated with requiring tedious, redundant work, typically involving maintaining redundant and nearly identical versions of code that correspond to different ablation trials. Inspired by the recent promising performance of Large Language Models (LLMs) in the generation and analysis of ML/DL code, in this paper we discuss the potential of LLMs as facilitators of ablation study experiments for scientific research projects that involve or deal with ML and DL models. We first discuss the different ways in which LLMs can be utilized for ablation studies and then present the prototype of a tool called AblationMage, that leverages LLMs to semi-automate the overall process of conducting ablation study experiments. We showcase the usability of AblationMage as a tool through three experiments, including one in which we reproduce the ablation studies from a recently published applied DL paper.
Ort, förlag, år, upplaga, sidor
ACM Digital Library, 2025
Nyckelord
Ablation Studies, Deep Learning, Feature Ablation, Model Ablation, Large Language Models
Nationell ämneskategori
Datavetenskap (datalogi)
Forskningsämne
Datalogi; Datalogi
Identifikatorer
urn:nbn:se:kth:diva-360719 (URN)10.1145/3721146.3721957 (DOI)001477868300025 ()2-s2.0-105003634645 (Scopus ID)
Konferens
The 5th Workshop on Machine Learning and Systems (EuroMLSys), co-located with the 20th European Conference on Computer Systems (EuroSys)
Forskningsfinansiär
Vinnova, 2016–05193
Anmärkning
QC 20250303
2025-02-282025-02-282025-07-01