Knowledge Injection XAI
Published:
Goal
To develop a distributed framework that can predict out-of-distribution (OOD) robustness of models using only Explainable AI (XAI) metrics extracted from clean data, eliminating the need for corrupted test sets during evaluation.
Technical Approach
- Built a Spark pipeline integrating DINOv2 as the backbone vision encoder with LoRA (Low-Rank Adaptation) fine-tuning and XGBoost for robustness forecasting.
- Extracted XAI metrics from clean data to predict how models would perform under distribution shifts and corruptions.
- Compared low-rank vs. high-rank adapter configurations to analyze their impact on corruption resistance.
Key Metrics & Results
- Achieved 0.739 ROC-AUC on the OOD robustness prediction task.
- Demonstrated that low-rank adapters resist corruptions 19% better than high-rank ones, providing actionable insights for model design.
- The distributed Spark pipeline enables scalable evaluation across large model collections.
Tech Stack
- Python, PyTorch, Apache Spark, DINOv2, LoRA, XGBoost
