Skater is a unified framework to enable Model Interpretation for all forms of the model to help one build an Interpretable machine learning system often needed for real-world use-cases(** we are actively working towards to enabling faithful interpretability for all forms models). It is an open-source python library designed to demystify the learned structures of a black box model both globally(inference on the basis of a complete data set) and locally(inference about an individual prediction). The concept of model interpretability in the field of machine learning is still new, largely subjective, and, at times, controversial. Model interpretation is the ability to explain and validate the decisions of a predictive model to enable fairness, accountability, and transparency in algorithmic decision-making. The library has embraced object-oriented and functional programming paradigms as deemed necessary to provide scalability and concurrency while keeping code brevity in mind.
Features
- Post hoc interpretation
- Natively interpretable models
- Skater provides the ability to interpret the model in both ways
- Model agnostic Partial Dependence Plots
- Local Interpretable Model Explanation(LIME)
- Layer-wise Relevance Propagation (e-LRP): image