Difference between revisions of "Explainable / Interpretable AI"

From
Jump to: navigation, search
Line 8: Line 8:
 
[http://www.google.com/search?q=Explainable+Artificial+Intelligence+deep+machine+learning+ML ...Google search]
 
[http://www.google.com/search?q=Explainable+Artificial+Intelligence+deep+machine+learning+ML ...Google search]
  
 +
* [http://pythondata.com/local-interpretable-model-agnostic-explanations-lime-python/ Local Interpretable Model-agnostic Explanations (LIME) in Python]
 +
* [http://github.com/slundberg/shap Shapley Additive Explanations (SHAP)]
 
* [http://medium.com/civis-analytics/demystifying-black-box-models-with-shap-value-analysis-3e20b536fc80 Demystifying Black-Box Models with SHAP Value Analysis | Peter Cooman]
 
* [http://medium.com/civis-analytics/demystifying-black-box-models-with-shap-value-analysis-3e20b536fc80 Demystifying Black-Box Models with SHAP Value Analysis | Peter Cooman]
* [http://pythondata.com/local-interpretable-model-agnostic-explanations-lime-python/ Local Interpretable Model-agnostic Explanations (LIME) in Python]
+
* [[Python#ELI5|ELI5]] debug machine learning classifiers and explain their predictions & inspect black-box models
* [http://github.com/slundberg/shap Shapley Additive Explanations (SHAP)]  
 
 
* [[AI Verification and Validation]]
 
* [[AI Verification and Validation]]
 
* [[Journey to Singularity]]
 
* [[Journey to Singularity]]

Revision as of 20:00, 23 July 2019

Youtube search... ...Google search

AI system produces results with an account of the path the system took to derive the solution/prediction - transparency of interpretation, rationale and justification. 'If you have a good causal model of the world you are dealing with, you can generalize even in unfamiliar situations. That’s crucial. We humans are able to project ourselves into situations that are very different from our day-to-day experience. Machines are not, because they don’t have these causal models. We can hand-craft them but that’s not enough. We need machines that can discover causal models. To some extend it’s never going to be perfect. We don’t have a perfect causal model of the reality, that’s why we make a lot of mistakes. But we are much better off at doing this than other animals.' Yoshua Benjio


SHAP (SHapley Additive exPlanations)

a unified approach to explain the output of any machine learning model. SHAP connects game theory with local explanations, uniting several previous methods [1-7] and representing the only possible consistent and locally accurate additive feature attribution method based on expectations (see our papers for details).

shap_diagram.png