In Machine Learning, the behavior of the computer agent is expected to improve over time in order to increase its usefulness to the end users. Traditional supervised techniques have made considerable progress by inducing a generalized function from examples, that are annotated by specialists (or by end users) before the learning process. While this approach can be succeeded in some applications, the absence of interaction between machines and specialists during the machine learning process leaves many important questions unanswered, compromising the usefulness of the solutions to many applications: how to minimize human effort with maximum efficacy in machine learning? can the machines learn from their errors? can the specialists understand the behavior of the machines, explain their actions, and trust on their decisions? what can machines and specialists learn from their interaction? This lecture is concerned with techniques to address such questions in the context of image annotation.
Image annotation consists of assigning one or multiple labels per image in order to make a decision or to support a human decision about a problem (e.g., the medical diagnosis). The pipeline for image annotation involves extraction, characterization, and classification of the content of interest, named samples. Samples may be pixels, regions of connected pixels with similar color and texture patterns (superpixels), connected components with known shapes (objects), or regions around objects (subimages). In any case, sample extraction is a fundamental problem that often requires object (semantic) segmentation. Nevertheless, interactive segmentation methods are rarely designed to improve from errors. Sample characterization aims at learning image features, usually based on the knowledge of the specialists (the handcrafted features) or based on a reasonable amount of previously extracted and annotated samples. The second strategy is not feasible when specialists are required to manually extract and annotate samples, raising two important questions: can feature learning techniques succeed from small labeled training sets? Can the specialist interact in feature learning to cope with the absence of labeled data, to improve the process, and to better understand the correlation between features and problem? Once the feature space is defined, the choice of key samples for label supervision is paramount in the design of the classifier. However, active learning techniques usually simulate user interaction during the process, disregarding the need for efficiency and interactive response times.
Sample extraction has been investigated as a separated task from characterization and classification, and the last two have also been investigated as a single operation. Indeed, their separation is important, with the specialist being part of the learning loop in all three steps, and the integration of their results in a same system is paramount for effective and efficient interactive machine learning.
The lecture proposes a methodology to address the problem, presents previous and underdevelopment work, and concludes with our still modest experience in what specialists and machines can learn from each other.
Alexandre Xavier Falcão is full professor at the Institute of Computing, University of Campinas, Campinas, SP, Brazil. He received a B.Sc. in Electrical Engineering from the Federal University of Pernambuco, Recife, PE, Brazil, in 1988. He has worked in biomedical image processing, visualization and analysis since 1991. In 1993, he received a M.Sc. in Electrical Engineering from the University of Campinas, Campinas, SP, Brazil. During 1994-1996, he worked with the Medical Image Processing Group at the Department of Radiology, University of Pennsylvania, PA, USA, on interactive image segmentation for his doctorate. He got his doctorate in Electrical Engineering from the University of Campinas in 1996. In 1997, he worked in a project for Globo TV at a research center, CPqD-TELEBRAS in Campinas, developing methods for video quality assessment. His experience as professor of Computer Science and Engineering started in 1998 at the University of Campinas. His main research interests include image/video processing, visualization, and analysis; graph algorithms and dynamic programming; image annotation, organization, and retrieval; machine learning and pattern recognition; and image analysis applications in Biology, Medicine, Biometrics, Geology, and Agriculture.