A look under the hood of the Interactive Deep Learning Enterprise (No-IDLE)
Daniel Sonntag, Michael Barz, Thiago Gouvêa
TL;DR
The paper outlines the No-IDLE initiative, a prototype for interactive deep learning that foregrounds human-in-the-loop design, multimodal interaction, and explainability to extend DL reach to non-experts. It centers on an interactive photo book use case to study combined NLP, MMI, ML, and HCI components, including gaze-driven feedback, entity-aware captioning, and mixed-initiative learning. Key contributions include a detailed blueprint for an end-to-end interactive DL system, methodologies for incremental model updates via explanatory feedback, and a VR-enabled evaluation plan to measure usability, learning efficiency, and user experience. The work aims to create a scalable testbed that informs broader AI deployment in domains like healthcare and manufacturing, with future directions involving integration with large language models such as ChatGPT.
Abstract
This DFKI technical report presents the anatomy of the No-IDLE prototype system (funded by the German Federal Ministry of Education and Research) that provides not only basic and fundamental research in interactive machine learning, but also reveals deeper insights into users' behaviours, needs, and goals. Machine learning and deep learning should become accessible to millions of end users. No-IDLE's goals and scienfific challenges centre around the desire to increase the reach of interactive deep learning solutions for non-experts in machine learning. One of the key innovations described in this technical report is a methodology for interactive machine learning combined with multimodal interaction which will become central when we start interacting with semi-intelligent machines in the upcoming area of neural networks and large language models.
