Abstract
A common assumption in supervised learning is that the input points in the training set follow the same probability distribution as the input points used for testing. However, this assumption is not satisfied, for example, when the outside of training region is extrapolated. The situation where the training input points and test input points follow different distributions is called the covariate shift. Under the covariate shift, standard machine learning techniques such as empirical risk minimization or cross-validation do not work well since their unbiasedness is no longer maintained. In this paper, we propose a new method called importance-weighted cross-validation, which is still unbiased even under the covariate shift. The usefulness of our proposed method is successfully tested on toy data and furthermore demonstrated in the brain-computer interface, where strong non-stationarity effects can be seen between calibration and feedback sessions.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Agostinelli, C.: Robust model selection by cross-validation via weighted likelihood methodology. Technical Report 1999.37, Dipartimento di Scienze Statistiche, Università di Padova (1999)
Blankertz, B., Dornhege, G., Krauledat, M., Müller, K.-R., Curio, G.: The Berlin brain-computer interface: Report from the feedback sessions. Technical Report 1, Fraunhofer FIRST (2005)
Duda, R.O., Hart, P.E., Stor, D.G.: Pattern Classification. Wiley, New York (2001)
Heckman, J.J.: Sample selection bias as a specification error. Econometrica 47(1), 153–162 (1979)
Schölkopf, B., Smola, A.J.: Learning with Kernels. MIT Press, Cambridge (2002)
Shimodaira, H.: Improving predictive inference under covariate shift by weighting the log-likelihood function. Journal of Statistical Planning and Inference 90(2), 227–244 (2000)
Sugiyama, M.: Active learning in approximately linear regression based on conditional expectation of generalization error. Journal of Machine Learning Research 7, 141–166 (2006)
Sugiyama, M., Blankertz, B., Krauledat, M., Dornhege, G., Müller, K.-R.: Importance-weighted cross-validation for covariate shift. Technical Report TR06-0002, Department of Computer Science, Tokyo Institute of Technology (February 2006)
Vapnik, V.N.: Statistical Learning Theory. Wiley, New York (1998)
Vidaurre, C., Schlögl, A., Cabeza, R., Pfurtscheller, G.: About adaptive classifiers for brain computer interfaces. Biomedizinische Technik 49(1), 85–86 (2004)
Wolpaw, J.R., Birbaumer, N., McFarland, D.J., Pfurtscheller, G., Vaughan, T.M.: Brain-computer interfaces for communication and control. Clinical Neurophysiology 113(6), 767–791 (2002)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Sugiyama, M., Blankertz, B., Krauledat, M., Dornhege, G., Müller, KR. (2006). Importance-Weighted Cross-Validation for Covariate Shift. In: Franke, K., Müller, KR., Nickolay, B., Schäfer, R. (eds) Pattern Recognition. DAGM 2006. Lecture Notes in Computer Science, vol 4174. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11861898_36
Download citation
DOI: https://doi.org/10.1007/11861898_36
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-44412-1
Online ISBN: 978-3-540-44414-5
eBook Packages: Computer ScienceComputer Science (R0)