Deep Domain Adaptation under Label Scarcity
Amar Prakash Azad, Dinesh Garg, et al.
CODS-COMAD 2021
As machine learning algorithms make inroads into our lives and society, calls are increasing from multiple stakeholders for these algorithms to explain their outputs. Moreover, these stakeholders, whether they be government regulators, affected citizens, domain experts, or developers, present different requirements for explanations. To address these needs, we introduce AI Explainability 3601, an open-source software toolkit featuring eight diverse state-of-the-art explainability methods, two evaluation metrics, and an extensible software architecture that organizes these methods according to their use in the AI modeling pipeline. Additionally, we have implemented enhancements to bring research innovations closer to consumers of explanations, ranging from simplified, accessible versions of algorithms to guidance material to help users navigate the space of explanations along with tutorials and an interactive web demo to introduce AI explainability to practitioners. Together, our toolkit can help improve transparency of machine learning models and provides a platform to integrate new explainability techniques as they are developed.
Amar Prakash Azad, Dinesh Garg, et al.
CODS-COMAD 2021
Sijia Liu, Pin-Yu Chen, et al.
IEEE TSP
Heshan Fernando, Lisha Chen, et al.
ICASSP 2024
Surya Shravan Kumar Sajja, Nupur Aggarwal, et al.
CODS-COMAD 2021