DiagHW: A Compact LLM for Hardware Failure Diagnosis via a Novel Knowledge Distillation Pipeline
Resumo
Hardware failure diagnosis from textual user-reported issues presents significant challenges due to the ambiguous and non-technical nature of these reports. While Large Language Models show promise in this domain, state-of-the-art models with billions of parameters pose practical limitations for deployment on low-power consumer devices. This work introduces a novel knowledge distillation pipeline for hardware failure detection, leveraging a large teacher model to generate synthetic training data for a smaller student model. Then, we present the DiagHW model, a compact 1.2B parameter fine-tuned LLaMA-3.2-1b-instruct model, which achieves diagnostic accuracy comparable to much larger models (up to 72B parameters). Our extensive experimental validation involved 32,414 inferences across 28 baseline models and 10 fine-tuned variants, processing a total of 19,692,210 tokens for these evaluations, complemented by 105,925,876 tokens processed during the fine-tuning stages.
Publicado
29/09/2025
Como Citar
CAMINHA, Carlos; SILVA, Maria de Lourdes M.; CHAVES, Iago C.; BRITO, Felipe T.; FARIAS, Victor A. E.; MACHADO, Javam C..
DiagHW: A Compact LLM for Hardware Failure Diagnosis via a Novel Knowledge Distillation Pipeline. In: BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS), 35. , 2025, Fortaleza/CE.
Anais [...].
Porto Alegre: Sociedade Brasileira de Computação,
2025
.
p. 393-407.
ISSN 2643-6264.
