Interactive machine teaching: a human-centered approach to building machine-learned models

ABSTRACT Modern systems can augment people’s capabilities by using machine-learned models to surface intelligent behaviors. Unfortunately, building these models remains challenging and beyond the reach of non-machine learning experts. We describe interactive machine teaching (IMT) and its potential to simplify the creation of machine-learned models. One of the key characteristics of IMT is its iterative process in which the human-in-the-loop takes the role of a teacher teaching a machine how to perform a task. We explore alternative learning theories as potential theoretical foundations for IMT, the intrinsic human capabilities related to teaching, and how IMT systems might leverage them. We argue that IMT processes that enable people to leverage these capabilities have a variety of benefits, including making machine learning methods accessible to subject-matter experts and the creation of semantic and debuggable machine learning (ML) models. We present an integrated teaching environment (ITE) that embodies principles from IMT, and use it as a design probe to observe how non-ML experts do IMT and as the basis of a system that helps us study how to guide teachers. We explore and highlight the benefits and challenges of IMT systems. We conclude by outlining six research challenges to advance the field of IMT.

[1]  Soroush Ghorashi,et al.  Using Expert Patterns in Assisted Interactive Machine Learning: A Study in Machine Teaching , 2019, INTERACT.

[2]  Eric Rosenbaum,et al.  Scratch: programming for all , 2009, Commun. ACM.

[3]  Timothy C. Au Random Forests, Decision Trees, and Categorical Predictors: The "Absent Levels" Problem , 2017, J. Mach. Learn. Res..

[4]  Steven M. Drucker,et al.  Gamut: A Design Probe to Understand How Data Scientists Understand Machine Learning Models , 2019, CHI.

[5]  Randal S. Olson,et al.  Automating Biomedical Data Science Through Tree-Based Pipeline Optimization , 2016, EvoApplications.

[6]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[7]  Zachary Chase Lipton The mythos of model interpretability , 2016, ACM Queue.

[8]  Cynthia Rudin,et al.  Optimized Risk Scores , 2017, KDD.

[9]  L. Ungar,et al.  MediBoost: a Patient Stratification Tool for Interpretable Decision Making in the Era of Precision Medicine , 2016, Scientific Reports.

[10]  Been Kim,et al.  Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.

[11]  Gabriel J. Brostow,et al.  Becoming the expert - interactive multi-class machine teaching , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Jerry Alan Fails,et al.  Interactive machine learning , 2003, IUI '03.

[13]  Karen M. Feigh,et al.  Effect of Interaction Design on the Human Experience with Interactive Reinforcement Learning , 2019, Conference on Designing Interactive Systems.

[14]  Saleema Amershi,et al.  Designing for effective end-user interaction with machine learning , 2011, UIST '11 Adjunct.

[15]  Ian H. Witten,et al.  Interactive machine learning: letting users build classifiers , 2002, Int. J. Hum. Comput. Stud..

[16]  Markus Nowak,et al.  Applying Radical Constructivism to Machine Learning: A Pilot Study in Assistive Robotics , 2018 .

[17]  Enrico Bertini,et al.  The Exploratory Labeling Assistant: Mixed-Initiative Label Curation with Large Document Collections , 2018, UIST.

[18]  Carla E. Brodley,et al.  Deploying an interactive machine learning system in an evidence-based practice center: abstrackr , 2012, IHI '12.

[19]  Chris Russell,et al.  Explaining Explanations in AI , 2018, FAT.

[20]  Mordechai Ben-Ari,et al.  Constructivism in computer science education , 1998, SIGCSE '98.

[21]  Sumit Basu,et al.  Learning to generalize for complex selection tasks , 2009, IUI.

[22]  Patrice Y. Simard,et al.  AnchorViz: Facilitating Semantic Data Exploration and Concept Discovery for Interactive Machine Learning , 2019, ACM Trans. Interact. Intell. Syst..

[23]  Maya Cakmak,et al.  Power to the People: The Role of Humans in Interactive Machine Learning , 2014, AI Mag..

[24]  Barry J. Wadsworth Piaget's Theory of Cognitive and Affective Development: Foundations of Constructivism , 2003 .

[25]  Randal S. Olson,et al.  Evaluation of a Tree-based Pipeline Optimization Tool for Automating Data Science , 2016, GECCO.

[26]  David Weinberger,et al.  Accountability of AI Under the Law: The Role of Explanation , 2017, ArXiv.

[27]  Daniel S. Weld,et al.  The challenge of crafting intelligible intelligence , 2018, Commun. ACM.

[28]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[29]  James A. Landay,et al.  Gestalt: integrated support for implementation and analysis in machine learning , 2010, UIST.

[30]  Burr Settles,et al.  Active Learning , 2012, Synthesis Lectures on Artificial Intelligence and Machine Learning.

[31]  Desney S. Tan,et al.  CueFlik: interactive concept learning in image search , 2008, CHI.

[32]  A. Amsel,et al.  Behaviorism, Neobehaviorism, and Cognitivism in Learning Theory: Historical and Contemporary Perspectives , 1988 .

[33]  Sandra Zilles,et al.  An Overview of Machine Teaching , 2018, ArXiv.

[34]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[35]  Lalana Kagal,et al.  Explaining Explanations: An Overview of Interpretability of Machine Learning , 2018, 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA).

[36]  Shachar Lovett,et al.  Active Classification with Comparison Queries , 2017, 2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS).

[37]  Julio Delgado,et al.  Elastic Machine Learning Algorithms in Amazon SageMaker , 2020, SIGMOD Conference.

[38]  Advait Sarkar,et al.  Constructivist Design for Interactive Machine Learning , 2016, CHI Extended Abstracts.

[39]  David Maxwell Chickering,et al.  Machine Teaching: A New Paradigm for Building Machine Learning Systems , 2017, ArXiv.

[40]  D. Sculley,et al.  Hidden Technical Debt in Machine Learning Systems , 2015, NIPS.

[41]  Johannes Gehrke,et al.  Intelligible Models for HealthCare: Predicting Pneumonia Risk and Hospital 30-day Readmission , 2015, KDD.

[42]  Marcia J. Bates,et al.  Indexing and Access for Digital Libraries and the Internet: Human, Database, and Domain Factors , 1998, J. Am. Soc. Inf. Sci..

[43]  Christopher Meek A Characterization of Prediction Errors , 2016, ArXiv.

[44]  Saul Greenberg,et al.  Prototyping an intelligent agent through Wizard of Oz , 1993, INTERCHI.

[45]  Weng-Keen Wong,et al.  Principles of Explanatory Debugging to Personalize Interactive Machine Learning , 2015, IUI.

[46]  Karen M. Feigh,et al.  Interaction Algorithm Effect on Human Experience with Reinforcement Learning , 2018, ACM Transactions on Human-Robot Interaction.

[47]  Kalyan Veeramachaneni,et al.  Deep feature synthesis: Towards automating data science endeavors , 2015, 2015 IEEE International Conference on Data Science and Advanced Analytics (DSAA).

[48]  Per Ola Kristensson,et al.  A Review of User Interface Design for Interactive Machine Learning , 2018, ACM Trans. Interact. Intell. Syst..

[49]  Patrice Y. Simard,et al.  Analysis of a Design Pattern for Teaching with Features and Labels , 2016, ArXiv.

[50]  Fred P. Brooks,et al.  The Mythical Man-Month , 1975, Reliable Software.

[51]  Christopher Ré,et al.  Snorkel: Rapid Training Data Creation with Weak Supervision , 2017, Proc. VLDB Endow..

[52]  Maya Cakmak,et al.  Eliciting good teaching from humans for machine learners , 2014, Artif. Intell..

[53]  Grant Potter,et al.  Machine Learning for Kids , 2018 .

[54]  Jure Leskovec,et al.  Interpretable Decision Sets: A Joint Framework for Description and Prediction , 2016, KDD.

[55]  Seth Flaxman,et al.  European Union Regulations on Algorithmic Decision-Making and a "Right to Explanation" , 2016, AI Mag..

[56]  David Maxwell Chickering,et al.  ModelTracker: Redesigning Performance Analysis Tools for Machine Learning , 2015, CHI.

[57]  Ben Shneiderman,et al.  Readings in information visualization - using vision to think , 1999 .

[58]  Andrea Lockerd Thomaz,et al.  Robot Learning from Human Teachers , 2014, Robot Learning from Human Teachers.

[59]  Andrew McCallum,et al.  Confidence Estimation for Information Extraction , 2004, NAACL.

[60]  Yichong Xu,et al.  Noise-Tolerant Interactive Learning Using Pairwise Comparisons , 2017, NIPS.

[61]  Lars Schmidt-Thieme,et al.  Beyond Manual Tuning of Hyperparameters , 2015, KI - Künstliche Intelligenz.

[62]  Xiaojin Zhu,et al.  Machine Teaching: An Inverse Problem to Machine Learning and an Approach Toward Optimal Education , 2015, AAAI.

[63]  Michael S. Bernstein,et al.  Break It Down: A Comparison of Macro- and Microtasks , 2015, CHI.

[64]  Cynthia Rudin,et al.  Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead , 2018, Nature Machine Intelligence.

[65]  David Maxwell Chickering,et al.  Interactive Semantic Featuring for Text Classification , 2016, ArXiv.

[66]  Qian Yang,et al.  Grounding Interactive Machine Learning Tool Design in How Non-Experts Actually Build Models , 2018, Conference on Designing Interactive Systems.

[67]  Brian E. Granger,et al.  IPython: A System for Interactive Scientific Computing , 2007, Computing in Science & Engineering.

[68]  Todd Kulesza,et al.  Structured labeling for facilitating concept evolution in machine learning , 2014, CHI.