Go Back Research Article March, 2025

INTERPRETABLE MACHINE LEARNING FOR ROOT CAUSE ANALYSIS IN CHIP VALIDATION FAILURES ACROSS MULTI-DIE SYSTEMS

Abstract

As multi-die systems become prevalent in modern semiconductor architectures, validation complexities have increased significantly. Root cause analysis (RCA) of failures in such complex systems demands interpretable methods that ensure engineers can understand and act upon the outcomes. This paper explores how interpretable machine learning (ML) techniques—particularly decision trees, LIME, and SHAP—can be applied to RCA for validation failures in multi-die chips. We demonstrate how model explainability can support engineers in isolating faults more efficiently than traditional rule-based or black-box ML models.

Keywords

interpretable machine learning root cause analysis chip validation multi-die systems shap lime system-on-chip (soc) hardware debugging
Document Preview
Download PDF
Details
Volume 3
Issue 1
Pages 1-8
ISSN 6542-5231