skip to main content
10.1145/3077257.3077271acmconferencesArticle/Chapter ViewAbstractPublication PagesmodConference Proceedingsconference-collections
research-article

PALM: Machine Learning Explanations For Iterative Debugging

Published: 14 May 2017 Publication History

Abstract

When a Deep Neural Network makes a misprediction, it can be challenging for a developer to understand why. While there are many models for interpretability in terms of predictive features, it may be more natural to isolate a small set of training examples that have the greatest influence on the prediction. However, it is often the case that every training example contributes to a prediction in some way but with varying degrees of responsibility. We present Partition Aware Local Model (PALM), which is a tool that learns and summarizes this responsibility structure to aide machine learning debugging. PALM approximates a complex model (e.g., a deep neural network) using a two-part surrogate model: a meta-model that partitions the training data, and a set of sub-models that approximate the patterns within each partition. These sub-models can be arbitrarily complex to capture intricate local patterns. However, the meta-model is constrained to be a decision tree. This way the user can examine the structure of the meta-model, determine whether the rules match intuition, and link problematic test examples to responsible training data efficiently. Queries to PALM are nearly 30x faster than nearest neighbor queries for identifying relevant data, which is a key property for interactive applications.

References

[1]
Keystone ML. http://keystone-ml.org/.
[2]
Tensor Flow. https://www.tensorflow.org/.
[3]
Peter Bailis, Edward Gan, Samuel Madden, Deepak Narayanan, Kexin Rong, and Sahaana Suri. 2017. Macrobase: Prioritizing attention in fast data. SIGMOD.
[4]
James Cheney, Laura Chiticariu, Wang-Chiew Tan, and others. 2009. Provenance in databases: Why, how, and where. Foundations and Trends® in Databases 1, 4 (2009), 379--474.
[5]
Andrew Crotty, Alex Galakatos, and Tim Kraska. 2014. Tupleware: Distributed Machine Learning on Small Clusters. In IEEE Data Eng. Bull. http://sites.computer.org/debull/A14sept/p63.pdf
[6]
Roy Fox*, Sanjay Krishnan*, Ken Goldberg, and Ion Stoica. 2017. Multi-Layer Deep Option Discovery. In Under Review ICML.
[7]
Tameru Hailesilassie. 2016. Rule Extraction Algorithm for Deep Neural Networks: A Review. arXiv preprint arXiv:1610.05267 (2016).
[8]
Joseph M. Hellerstein, Christopher Re, Florian Schoppmann, Daisy Zhe Wang, Eugene Fratkin, Aleksander Gorajek, Kee Siong Ng, Caleb Welton, Xixuan Feng, Kun Li, and Arun Kumar. 2012. The MADlib Analytics Library or MAD Skills, the SQL. In VLDB. http://vldb.org/pvldb/vol5/p1700_joehellerstein_vldb2012.pdf
[9]
Jinkyu Kim and John Canny. 2017. Interpretable Learning for Self-Driving Cars by Visualizing Causal Attention. arXiv preprint arXiv:1703.10631 (2017).
[10]
Tim Kraska, Ameet Talwalkar, John C. Duchi, Rean Griffith, Michael J. Franklin, and Michael I. Jordan. 2013. MLbase: A Distributed Machine-learning System. In CIDR. http://www.cidrdb.org/cidr2013/Papers/CIDR13_Paper118.pdf
[11]
Sanjay Krishnan, Animesh Garg, Richard Liaw, Lauren Miller, Florian T. Pokorny, and Ken Goldberg. 2016. HIRL: Hierarchical Inverse Reinforcement Learning for Long-Horizon Tasks with Delayed Rewards. CoRR abs/1604.06508 (2016). http://arxiv.org/abs/1604.06508
[12]
Sanjay Krishnan, Jiannan Wang, Eugene Wu, Michael J. Franklin, and Ken Goldberg. 2016. ActiveClean: Interactive Data Cleaning For Statistical Modeling. PVLDB 9, 12 (2016), 948--959. http://www.vldb.org/pvldb/vol9/p948-krishnan.pdf
[13]
Tao Lei, Regina Barzilay, and Tommi Jaakkola. 2016. Rationalizing neural predictions. arXiv preprint arXiv:1606.04155 (2016).
[14]
Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. 2016. Why Should I Trust You?: Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, 1135--1144.
[15]
Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. 2013. Intriguing properties of neural networks. arXiv preprint arXiv:1312.6199 (2013).
[16]
Jessica Taylor, Eliezer Yudkowsky, Patrick LaVictoire, and Andrew Critch. 2016. Alignment for advanced machine learning systems. Technical Report. Technical Report 20161, MIRI.
[17]
Eugene Wu and Samuel Madden. 2013. Scorpion: Explaining Away Outliers in Aggregate Queries. In VLDB. http://www.vldb.org/pvldb/vol6/p553-wu.pdf

Cited By

View all
  • (2024)Explainable AI for CybersecurityAdvances in Explainable AI Applications for Smart Cities10.4018/978-1-6684-6361-1.ch002(31-97)Online publication date: 18-Jan-2024
  • (2024)"We Have No Idea How Models will Behave in Production until Production": How Engineers Operationalize Machine LearningProceedings of the ACM on Human-Computer Interaction10.1145/36536978:CSCW1(1-34)Online publication date: 26-Apr-2024
  • (2024)An interpretable machine learning methodology to generate interaction effect hypotheses from complex datasetsDecision Sciences10.1111/deci.12642Online publication date: 13-Aug-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
HILDA '17: Proceedings of the 2nd Workshop on Human-In-the-Loop Data Analytics
May 2017
89 pages
ISBN:9781450350297
DOI:10.1145/3077257
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 14 May 2017

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

SIGMOD/PODS'17
Sponsor:

Acceptance Rates

Overall Acceptance Rate 28 of 56 submissions, 50%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)100
  • Downloads (Last 6 weeks)6
Reflects downloads up to 23 Sep 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Explainable AI for CybersecurityAdvances in Explainable AI Applications for Smart Cities10.4018/978-1-6684-6361-1.ch002(31-97)Online publication date: 18-Jan-2024
  • (2024)"We Have No Idea How Models will Behave in Production until Production": How Engineers Operationalize Machine LearningProceedings of the ACM on Human-Computer Interaction10.1145/36536978:CSCW1(1-34)Online publication date: 26-Apr-2024
  • (2024)An interpretable machine learning methodology to generate interaction effect hypotheses from complex datasetsDecision Sciences10.1111/deci.12642Online publication date: 13-Aug-2024
  • (2024)A Survey on Verification and Validation, Testing and Evaluations of Neurosymbolic Artificial IntelligenceIEEE Transactions on Artificial Intelligence10.1109/TAI.2024.33517985:8(3765-3779)Online publication date: Aug-2024
  • (2024)Trading-Off Interpretability and Accuracy in Medical Applications: A Study Toward Optimal Explainability of Hoeffding Trees2024 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE)10.1109/FUZZ-IEEE60900.2024.10611982(1-10)Online publication date: 30-Jun-2024
  • (2024)Interpretability Index Based on Balanced Volumes for Transparent Models and Agnostic Explainers2024 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE)10.1109/FUZZ-IEEE60900.2024.10611937(1-10)Online publication date: 30-Jun-2024
  • (2024)Survey on Explainable AI: Techniques, challenges and open issuesExpert Systems with Applications10.1016/j.eswa.2024.124710255(124710)Online publication date: Dec-2024
  • (2024)When debugging encounters artificial intelligence: state of the art and open challengesScience China Information Sciences10.1007/s11432-022-3803-967:4Online publication date: 21-Feb-2024
  • (2024)Understanding via exemplification in XAI: how explaining image classification benefits from exemplarsAI & SOCIETY10.1007/s00146-023-01837-4Online publication date: 27-Jan-2024
  • (2024)Interpretability and Explainability of Machine Learning Models: Achievements and ChallengesInternational Conference on Biomedical and Health Informatics 202210.1007/978-3-031-59216-4_9(81-94)Online publication date: 30-Apr-2024
  • Show More Cited By

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media