Utilizing Large Language Models for Machine Learning Explainability

Alexandros Vassiliades, Nikolaos Polatidis, Stamatios Samaras, Sotiris Diplaris, Ignacio Cabrera Martin, Yannis Manolopoulos, Stefanos Vrochidis, Ioannis Kompatsiaris

Research output: Chapter in Book/Conference proceeding with ISSN or ISBNConference contribution with ISSN or ISBNpeer-review

Abstract

This study explores the explainability capabilities of large language models (LLMs), when employed to autonomously generate machine learning (ML) solutions. We examine two classification tasks: (i) a binary classification problem focused on predicting driver alertness states, and (ii) a multilabel classification problem based on the yeast dataset. Three state-of-the-art LLMs (i.e. OpenAI GPT, Anthropic Claude, and DeepSeek) are prompted to design training pipelines for four common classifiers: Random Forest, XGBoost, Multilayer Perceptron, and Long Short-Term Memory networks. The generated models are evaluated in terms of predictive performance (recall, precision, and F1-score)and explainability using SHAP (SHapley Additive exPlanations). Specifically, we measure Average SHAP Fidelity (Mean Squared Error between SHAP approximations and model outputs) and Average SHAP Sparsity (number of features deemed influential). The results reveal that LLMs
are capable of producing effective and interpretable models, achieving high fidelity and consistent sparsity, highlighting their potential as automated tools for interpretable ML pipeline generation. Importantly, our findings indicate that the quality of LLM-generated pipelines closely approximates that of manually engineered solutions both in predictive accuracy and explainability, suggesting that LLMs can reliably support or even partially automate explainable ML workflows with minimal loss in interpretive depth.
Original languageEnglish
Title of host publicationTHE 17TH INTERNATIONAL CONFERENCE ON MANAGEMENT OF DIGITAL ECOSYSTEMS
Subtitle of host publicationMEDES 2025
PublisherSpringer
Number of pages14
Publication statusAccepted/In press - 26 Sept 2025
EventThe 17th International Conference on Management of Digital Ecosystems - Ho Chi Minh City, Viet Nam
Duration: 24 Nov 202526 Nov 2025
https://conferences.sigappfr.org/medes2025/

Publication series

NameCommunications in Computer and Information Science
PublisherSpringer
ISSN (Print)1865-0929
ISSN (Electronic)1865-0937

Conference

ConferenceThe 17th International Conference on Management of Digital Ecosystems
Abbreviated titleMEDES 2025
Country/TerritoryViet Nam
CityHo Chi Minh City
Period24/11/2526/11/25
Internet address

Bibliographical note

Not Yet Published

Fingerprint

Dive into the research topics of 'Utilizing Large Language Models for Machine Learning Explainability'. Together they form a unique fingerprint.

Cite this