Abstract
In real applications of inductive learning, labeled instances are often deficient. The countermeasure is either to ask experts to label informative instances in active learning, or to borrow useful information from abundant labeled instances in the source domain in transfer learning. Due to the high cost of querying experts, it is promising to integrate the two methodologies into a more robust and reliable classification framework to compensate the disadvantages of both methods. Recently, a few research studies have been investigated to integrate the two methods together, which is called transfer active learning. However, when there exist unrelated domains which have different distributions or label assignments, an inevitable problem named negative transfer will happen which leads to degenerated performance. Also, how to avoid selecting unconcerned samples to query is still an open question. To tackle these issues, we propose a hybrid algorithm for active learning with the help of transfer learning by adopting a divergence measure to measure the similarities between different domains, so that the negative effects can be alleviated. To avoid querying irrelevant instances, we also present an adaptive strategy that is able to eliminate unnecessary instances in the input space and models in the model space. Extensive experiments on both synthetic and real data sets show that our algorithm is able to query less instances and converges faster than the state-of-the-art methods.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Harpale, A., Yang, Y.: Active Learning for Multi-Task Adaptive Filtering. In: Proceeding of the 27th International Conference of Machine Learning (ICML 2010), pp. 431–438 (2010)
McCallum, A., Nigam, K.: Employing EM in Pool-based Active Learning for Text Classification. In: Proceeding of the 15th International Conference of Machine Learning (ICML 1998), pp. 350–358 (1998)
Settles, B.: Active Learning Literature Survey. Technical Report 1648, University of Wisconsin–Madison (2010)
Cohn, D., Ladner, R., Waibel, A.: Improving Generalization with Active Learning. In: Machine Learning, pp. 201–221 (1994)
Lewis, D.D., Gale, W.A.: A Sequential Algorithm for Training Text Classifiers. In: Proceedings of the 17th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 1994), pp. 3–12 (1994)
Pereira, F., Tishby, N., Lee, L.: Distributional Clustering of English Words. In: Proceeding of the 31st Annual Meeting of the Association for Computational Linguistics (ACL 1993), pp. 183–190 (1993)
Li, H., Shi, Y., Chen, M.Y., Hauptmann, A.G., Xiong, Z.: Hybrid Active Learning for Cross-domain Video Concept Detection. In: Proceedings of the International Conference on Multimedia (MM 2010 ), pp. 1003–1006 (2010)
Shao, H., Suzuki, E.: Feature-based Inductive Transfer Learning through Minimum Encoding. In: Proceeding of the SIAM International Conference on Data Mining (SDM 2011), pp. 259–270 (2011)
Seung, H.S., Opper, M., Sompolinsky, H.: Query by Committee. In: Computational Learning Theory, pp. 287–294 (1992)
Dagan, I., Engelson, S.P.: Committee-Based Sampling For Training Probabilistic Classifiers. In: Proceeding of the 23rd International Conference on Machine Learning (ICML 2006), pp. 150–157 (1995)
Muslea, I., Minton, S., Knoblock, C.A.: Active Semi-Supervised Learning = Robust Multi-View Learning. In: Proceeding of the 19th International Conference on Machine Learning (ICML 2002), pp. 435–442 (2002)
Lin, J.: Divergence Measures based on the Shannon Entropy. IEEE Transactions on Information theory 37, 145–151 (1991)
Church, K.W., Gale, W.A.: A Comparison of the Enhanced Good-Turing and Deleted Estimation Methods for Estimating Probabilities of English Bigrams. Computer Speech and Language 5, 19–54 (1991)
Li, L., Jin, X., Pan, S., Sun, J.: Multi-Domain Active Learning for Text Classification. In: Proceeding of the 18th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD 2012 (2012)
Balcan, M.F., Beygelzimer, A., Langford, J.: Agnostic Active Learning. In: Proceeding of the 23rd International Conference on Machine Learning (ICML 2006), pp. 65–72 (2006)
Rosenstein, M.T., Marx, Z., Kaelbling, L.P., Dietterich, T.G.: To Transfer or not to Transfer. In: NIPS 2005 Workshop, Inductive Transfer: 10 Years Later (2005)
Caruana, R.: Multitask Learning. Machine Learning 28, 41–75 (1997)
Reichart, R., Tomanek, K., Hahn, U.: Multi-task Active Learning for Linguistic Annotations. In: Annual Meeting of the Association for Computational Linguistics (ACL 2008), pp. 861–869 (2008)
Raj, S., Ghosh, J., Crawford, M.M.: An Active Learning Approach to Knowledge Transfer for Hyperspectral Data Analysis. In: Proceeding of the IEEE International Conference on Geoscience and Remote Sensing Symposium (IGARSS 2006), pp. 541–544 (2006)
Dai, W., Yang, Q., Xue, G.R., Yu, Y.: Boosting for Transfer Learning. In: Proceeding of the 24th International Conference of Machine Learning (ICML 2007), pp. 193–200 (2007)
Shi, X., Fan, W., Ren, J.: Actively Transfer Domain Knowledge. In: Daelemans, W., Goethals, B., Morik, K. (eds.) ECML PKDD 2008, Part II. LNCS (LNAI), vol. 5212, pp. 342–357. Springer, Heidelberg (2008)
Zhang, Y.: Multi-Task Active Learning with Output Constraints. In: Proceeding of the 24th AAAI Conference on Artificial Intelligence, AAAI 2010 (2010)
Zhu, Z., Zhu, X., Ye, Y., Guo, Y.-F., Xue, X.: Transfer Active Learning. In: Proceeding of the 20th International Conference on Information and Knowledge Management (CIKM 2011), pp. 2169–2172 (2011)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Shao, H., Tong, B., Suzuki, E. (2012). Query by Committee in a Heterogeneous Environment. In: Zhou, S., Zhang, S., Karypis, G. (eds) Advanced Data Mining and Applications. ADMA 2012. Lecture Notes in Computer Science(), vol 7713. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-35527-1_16
Download citation
DOI: https://doi.org/10.1007/978-3-642-35527-1_16
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-35526-4
Online ISBN: 978-3-642-35527-1
eBook Packages: Computer ScienceComputer Science (R0)