ABSTRACT
Information and communication technologies supply data every day at incredibly increasing rate, however, almost all of the accumulated data are unlabeled and obtaining their labels is expensive and time-consuming. Among the raw data, selecting and labeling some samples expected to be more informative than others can enhance machines without high cost. This process is called selective sampling, essential part of active learning. So far, most researches have concentrated on classical uncertainty measures to acquire informative data, which is related to "exploitation" process of learning. However, when the initial labeled dataset is too small or biased, the early stage model can be unreliable and its decision boundary would be over-fitted to the initial data. Moreover, the obtained data by the exploitation strategy may exacerbate the model further. We introduced "exploration" strategy as well as "exploitation" strategy. In this paper, we employ Self-Organizing Maps (SOM), one of neural networks to estimate and explore data distribution. For exploitation, margin sampling is applied to the classifier, neural network with soft-max output layer. The effectiveness proposed methods are demonstrated on ILSVRC-2011 image classification task based on features extracted from well-trained Convolutional Neural Networks (CNN). Active learning with exploration strategy shows its potential by stabilizing the early stage model and reducing the classification error rate, and finally making it to be high-quality models.
- COHN, David; ATLAS, Les; LADNER, Richard. Improving generalization with active learning. Machine Learning, 1994, 15.2: 201--22 Google ScholarDigital Library
- SCHOHN, Greg; COHN, David. Less is more: Active learning with support vector machines. In: ICML. 2000. p. 839--846. Google ScholarDigital Library
- Xu, Zhao, et al. "Representative sampling for text classification using support vector machines." ECIR. Vol. 3. 2003. Google ScholarDigital Library
- FREUND, Yoav, et al. Selective sampling using the query by committee algorithm. Machine learning, 1997, 28.2--3: 133--168. Google ScholarDigital Library
- IYENGAR, Vijay S.; APTE, Chidanand; ZHANG, Tong. Active learning using adaptive resampling. In: Proceedings of the sixth ACM SIGKDD international conference on Knowledge discovery and data mining. ACM, 2000. p. 91--98. Google ScholarDigital Library
- LINDENBAUM, Michael; MARKOVITCH, Shaul; RUSAKOV, Dmitry. Selective sampling for nearest neighbor classifiers. Machine learning, 2004, 54.2: 125--152. Google ScholarDigital Library
- ROY, Nicholas; MCCALLUM, Andrew. Toward optimal active learning through monte carlo estimation of error reduction. ICML, Williamstown, 2001. Google ScholarDigital Library
- T. Scheffer, C. Decomain, and S. Wrobel. Active hidden Markov models for information extraction. In Proceedings of the International Conference on Advances in Intelligent Data Analysis (CAIDA), pages 309--318. Springer-Verlag, 2001. Google ScholarDigital Library
Index Terms
- Active Learning for Large-scale Object Classification: from Exploration to Exploitation
Recommendations
An active learning-based SVM multi-class classification model
Traditional multi-class classification models are based on labeled data and are not applicable to unlabeled data. To overcome this limitation, this paper presents a multi-class classification model that is based on active learning and support vector ...
Combining active learning and semi-supervised for improving learning performance
ISABEL '11: Proceedings of the 4th International Symposium on Applied Sciences in Biomedical and Communication TechnologiesIn many learning tasks, there are abundant unlabeled samples but the number of labeled training samples is limited, because labeling the samples requires the efforts of human annotators and expertise. There are three major techniques for labeling the ...
Semi-supervised Dictionary Active Learning for Pattern Classification
Pattern Recognition and Computer VisionAbstractGathering labeled data is one of the most time-consuming and expensive tasks in supervised machine learning. In practical applications, there are usually quite limited labeled training samples but abundant unlabeled data that is easy to collect. ...
Comments