Browse > Article
http://dx.doi.org/10.7236/IJASC.2020.9.2.90

Understanding Interactive and Explainable Feedback for Supporting Non-Experts with Data Preparation for Building a Deep Learning Model  

Kim, Yeonji (Department of Computer Science and Engineering, Ewha Womans University)
Lee, Kyungyeon (Department of Computer Science and Engineering, Ewha Womans University)
Oh, Uran (Department of Computer Science and Engineering, Ewha Womans University)
Publication Information
International journal of advanced smart convergence / v.9, no.2, 2020 , pp. 90-104 More about this Journal
Abstract
It is difficult for non-experts to build machine learning (ML) models at the level that satisfies their needs. Deep learning models are even more challenging because it is unclear how to improve the model, and a trial-and-error approach is not feasible since training these models are time-consuming. To assist these novice users, we examined how interactive and explainable feedback while training a deep learning network can contribute to model performance and users' satisfaction, focusing on the data preparation process. We conducted a user study with 31 participants without expertise, where they were asked to improve the accuracy of a deep learning model, varying feedback conditions. While no significant performance gain was observed, we identified potential barriers during the process and found that interactive and explainable feedback provide complementary benefits for improving users' understanding of ML. We conclude with implications for designing an interface for building ML models for novice users.
Keywords
End-user Machine Learning; Interactivity; Explainability;
Citations & Related Records
연도 인용수 순위
  • Reference
1 Papanastasopoulos, Z., Samala, R. K., Chan, H. P., Hadjiiski, L., Paramagul, C., Helvie, M. A., & Neal, C. H. (2020, March). Explainable AI for medical imaging: deep-learning CNN ensemble for classification of estrogen receptor status from breast MRI. In Medical Imaging 2020: Computer-Aided Diagnosis (Vol. 11314, p. 113140Z). International Society for Optics and Photonics.
2 Kashyap, S., Karargyris, A., Wu, J., Gur, Y., Sharma, A., Wong, K. C., ... & Syeda-Mahmood, T. (2020, April). Looking in the Right Place for Anomalies: Explainable Ai Through Automatic Location Learning. In 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI) (pp. 1125-1129). IEEE.
3 Ramos, G., Meek, C., Simard, P., Suh, J., & Ghorashi, S. (2020). Interactive machine teaching: a humancentered approach to building machine-learned models. Human-Computer Interaction, 1-39.
4 Ishibashi, T., Nakao, Y., & Sugano, Y. (2020, March). Investigating audio data visualization for interactive sound recognition. In Proceedings of the 25th International Conference on Intelligent User Interfaces (pp. 67-77).
5 S. Amershi, J. Fogarty, A. Kapoor, and D. Tan, "Examining multiple potential models in end-user interactive concept learning," in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM, pp. 1357-1360, 2010. DOI: https://doi.org/10.1145/1753326.1753531
6 S. Amershi, J. Fogarty, A. Kapoor, and D. Tan, "Effective end-user interaction with machine learning," in Twenty- Fifth AAAI Conference on Artificial Intelligence, 2011. DOI:
7 R. Fiebrink, P. R. Cook, and D. Trueman, "Human model evaluation in interactive supervised learning," in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 147-156, 2011. DOI: https://doi.org/10.1145/1978942.1978965
8 T. Hitron, Y. Orlev, I. Wald, A. Shamir, H. Erel, and O. Zuckerman, "Can children understand machine learning concepts?: The effect of uncovering black boxes," in Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, ser. CHI '19. New York, NY, USA: ACM, pp. 415:1-415:11, 2019. DOI: http://doi.acm.org/10.1145/3290605.3300645
9 A. Holzinger, "Interactive machine learning for health informatics: when do we need the human-in-the-loop?" Brain Informatics, vol. 3, no. 2, pp. 119-131, 2016. DOI: https://doi.org/10.1007/s40708-016-0042-6   DOI
10 J. A. Fails and D. R. Olsen Jr, "Interactive machine learning," in Proceedings of the 8th international conference on Intelligent user interfaces. ACM, pp. 39-45, 2003. DOI: https://doi.org/10.1145/604045.604056
11 K. Patel, N. Bancroft, S. M. Drucker, J. Fogarty, A. J. Ko, and J. Landay, "Gestalt: integrated support for implementation and analysis in machine learning," in Proceedings of the 23nd annual ACM symposium on User interface software and technology. ACM, pp. 37-46, 2010. DOI: https://doi.org/10.1145/1866029.1866038
12 J. Fogarty, D. Tan, A. Kapoor, and S. Winder, "Cueflik: interactive concept learning in image search," in Proceedings of the sigchi conference on human factors in computing systems. ACM, pp. 29-38, 2008. DOI: https://doi.org/10.1145/1357054.1357061
13 D. Gunning, "Explainable artificial intelligence (xai)," Defense Advanced Research Projects Agency (DARPA), nd Web, vol. 2, 2017. https://www.darpa.mil/attachments/XAIProgramUpdate.pdf
14 Q. Yang, J. Zimmerman, A. Steinfeld, and A. Tomasic, "Planning adaptive mobile experiences when wireframing," in Proceedings of the 2016 ACM Conference on Designing Interactive Systems, pp. 565-576, 2016. DOI: https://doi.org/10.1145/2901790.2901858
15 S. Amershi, M. Cakmak, W. B. Knox, and T. Kulesza, "Power to the people: The role of humans in interactive machine learning," AI Magazine, vol. 35, no. 4, pp. 105-120, 2014. DOI: https://doi.org/10.1609/aimag.v35i4.2513   DOI
16 S. L. Rosenthal and A. K. Dey, "Towards maximizing the accuracy of human-labeled sensor data," in Proceedings of the 15th international conference on Intelligent user interfaces. ACM, pp. 259-268, 2010. DOI: https://doi.org/10.1145/1719970.1720006
17 C. J. Cai, E. Reif, N. Hegde, J. Hipp, B. Kim, D. Smilkov, M. Wattenberg, F. Viegas, G. S. Corrado, M. C. Stumpe et al., "Human-centered tools for coping with imperfect algorithms during medical decision-making," in Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. ACM, p. 4, 2019. DOI: https://doi.org/10.1145/3290605.3300234
18 M. Choi, C. Park, S. Yang, Y. Kim, J. Choo, and S. R. Hong, "Aila: Attentive interactive labeling assistant for document classification through attention-based deep neural networks," in Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. ACM, p. 230, 2019. DOI: https://doi.org/10.1145/3290605.3300460
19 G. Litjens, T. Kooi, B. E. Bejnordi, A. A. A. Setio, F. Ciompi, M. Ghafoorian, J. A. Van Der Laak, B. Van Ginneken, and C. I. Sanchez, "A survey on deep learning in medical image analysis," Medical image analysis, vol. 42, pp. 60-88, 2017. DOI: https://doi.org/10.1016/j.media.2017.07.005   DOI
20 G. Litjens, C. I. Sanchez, N. Timofeeva, M. Hermsen, I. Nagtegaal, I. Kovacs, C. Hulsbergen-Van De Kaa, P. Bult, B. Van Ginneken, and J. Van Der Laak, "Deep learning as a tool for increased accuracy and efficiency of histopathological diagnosis," Scientific reports, vol. 6, p.26286, 2016. DOI: https://doi.org/10.1038/srep26286   DOI
21 R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, "Grad-cam: Visual explanations from deep networks via gradient-based localization," in Proceedings of the IEEE International Conference on Computer Vision, pp. 618-626, 2017. DOI: https://doi.org/10.1109/iccv.2017.74
22 J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, "Imagenet: A large-scale hierarchical image database," in 2009 IEEE conference on computer vision and pattern recognition. IEEE, pp. 248-255, 2009. DOI: https://doi.org/10.1109/cvpr.2009.5206848
23 X. Meng, J. Bradley, B. Yavuz, E. Sparks, S. Venkataraman, D. Liu, J. Freeman, D. Tsai, M. Amde, S. Owen et al., "Mllib: Machine learning in apache spark," The Journal of Machine Learning Research, vol. 17, no. 1, pp. 1235-1241, 2016. https://dl.acm.org/doi/abs/10.5555/2946645.2946679
24 Q. Yang, J. Suh, N.-C. Chen, and G. Ramos, "Grounding interactive machine learning tool design in how nonexperts actually build models," in Proceedings of the 2018 Designing Interactive Systems Conference. ACM, pp. 573-584, 2018. DOI: https://doi.org/10.1145/3196709.3196729
25 T. Kulesza, M. Burnett, W.-K. Wong, and S. Stumpf, "Principles of explanatory debugging to personalize interactive machine learning," in Proceedings of the 20th international conference on intelligent user interfaces. ACM, pp. 126-137, 2015. DOI: https://doi.org/10.1145/2678025.2701399
26 M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, and L.-C. Chen, "Mobilenetv2: Inverted residuals and linear bottlenecks," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510-4520, 2018. DOI: https://doi.org/10.1109/cvpr.2018.00474
27 D. Frossard, "Vgg in tensorflow," VGG in TensorFlow. Davi Frossard, 2017.
28 S. G. Hart and L. E. Staveland, "Development of nasa-tlx (task load index): Results of empirical and theoretical research," in Human Mental Workload, ser. Advances in Psychology, P. A. Hancock and N. Meshkati, Eds. North-Holland, vol. 52, pp. 139-183, 1988. DOI: https://doi.org/10.1016/s0166-4115(08)62386-9
29 T. Kulesza, S. Stumpf, M. Burnett, and I. Kwan, "Tell me more?: the effects of mental model soundness on personalizing an intelligent agent," in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM, pp. 1-10, 2012. DOI: https://doi.org/10.1145/2207676.2207678
30 W. B. Knox and P. Stone, "Reinforcement learning from human reward: Discounting in episodic tasks," in 2012 IEEE RO-MAN: The 21st IEEE International Symposium on Robot and Human Interactive Communication. IEEE, pp. 878-885, 2012. DOI: https://doi.org/10.1109/roman.2012.6343862