Loading [a11y]/accessibility-menu.js
Minimax Active Learning Via Minimal Model Capacity | IEEE Conference Publication | IEEE Xplore

Minimax Active Learning Via Minimal Model Capacity


Abstract:

Active learning is a form of machine learning which combines supervised learning and feedback to minimize the training set size, subject to low generalization errors. Sin...Show More

Abstract:

Active learning is a form of machine learning which combines supervised learning and feedback to minimize the training set size, subject to low generalization errors. Since direct optimization of the generalization error is difficult, many heuristics have been developed which lack a firm theoretical foundation. In this paper, a new information theoretic criterion is proposed based on a minimax log-loss regret formulation of the active learning problem. In the first part of this paper, a Redundancy Capacity theorem for active learning is derived along with an optimal learner. Building on this, a new active learning criterion is proposed which naturally induces an exploration - exploitation trade-off in feature selection. In the second part, the linear separator hypotheses class with additive label noise is considered and a low complexity algorithm is proposed which optimizes the active learning criterion from the first part. This greedy algorithm is based on the Posterior Matching scheme for communication with feedback and is shown that for BSC and BEC label noise, the proposed information theoretic criterion decays at an exponential rate.
Date of Conference: 13-16 October 2019
Date Added to IEEE Xplore: 05 December 2019
ISBN Information:
Print on Demand(PoD) ISSN: 1551-2541
Conference Location: Pittsburgh, PA, USA

Contact IEEE to Subscribe

References

References is not available for this document.