skip to main content
10.1145/2487575.2487674acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article

SVMpAUCtight: a new support vector method for optimizing partial AUC based on a tight convex upper bound

Published: 11 August 2013 Publication History

Abstract

The area under the ROC curve (AUC) is a well known performance measure in machine learning and data mining. In an increasing number of applications, however, ranging from ranking applications to a variety of important bioinformatics applications, performance is measured in terms of the partial area under the ROC curve between two specified false positive rates. In recent work, we proposed a structural SVM based approach for optimizing this performance measure (Narasimhan and Agarwal, 2013). In this paper, we develop a new support vector method, SVMpAUCtight, that optimizes a tighter convex upper bound on the partial AUC loss, which leads to both improved accuracy and reduced computational complexity. In particular, by rewriting the empirical partial AUC risk as a maximum over subsets of negative instances, we derive a new formulation, where a modified form of the earlier optimization objective is evaluated on each of these subsets, leading to a tighter hinge relaxation on the partial AUC loss. As with our previous method, the resulting optimization problem can be solved using a cutting-plane algorithm, but the new method has better run time guarantees. We also discuss a projected subgradient method for solving this problem, which offers additional computational savings in certain settings. We demonstrate on a wide variety of bioinformatics tasks, ranging from protein-protein interaction prediction to drug discovery tasks, that the proposed method does, in many cases, perform significantly better on the partial AUC measure than the previous structural SVM approach. In addition, we also develop extensions of our method to learn sparse and group sparse models, often of interest in biological applications.

References

[1]
S. Agarwal. The Infinite Push: A new support vector ranking algorithm that directly optimizes accuracy at the absolute top of the list. In SDM, 2011.
[2]
S. Agarwal, D. Dugar, and S. Sengupta. Ranking chemical structures for drug discovery: A new machine learning approach. Journal of Chemical Information and Modeling, 50(5):716--731, 2010.
[3]
S. Agarwal, T. Graepel, R. Herbrich, S. Har-Peled, and D. Roth. Generalization bounds for the area under the ROC curve. Journal of Machine Learning Research, 6:393--425, 2005.
[4]
P. Balamurugan, S. Shevade, and T. Babu. Sequential alternating proximal method for scalable sparse structural SVMs. In ICDM, 2012.
[5]
D. P. Bertsekas. Nonlinear Programming. Athena Scientific, Belmont, MA, 1999.
[6]
S. Boyd, L. Xiao, and A. Mutapcic. Subgradient methods. http://www.stanford.edu/class/ee392o/subgrad_method.pdf, 2003.
[7]
T. P. Conrads, M. Zhou, E. F. I. Petricoin, L. Liotta, and T. D. Veenstra. Cancer diagnosis using proteomic patterns. Expert Rev. Mol. Diagn., (3):411--420, 2003.
[8]
C. Cortes and M. Mohri. AUC optimization vs. error rate minimization. In NIPS, 2004.
[9]
L. E. Dodd and M. S. Pepe. Partial AUC estimation and regression. Biometrics, 59(3):614--623, 2003.
[10]
J. Duchi. Elastic net projections, http://www.cs.berkeley.edu/ jduchi/projects/proj_elastic_net.pdf, 2009.
[11]
J. Duchi, S. Shalev-Shwartz, Y. Singer, and T. Chandra. Efficient projections onto the l1-ball for learning in high dimensions. In ICML, 2008.
[12]
T. R. Golub. et al. Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. Science, 286(5439):531--537, 1999.
[13]
T. Joachims. A support vector method for multivariate performance measures. In ICML, 2005.
[14]
T. Joachims. Training linear SVMs in linear time. In KDD, 2006.
[15]
T. Joachims, T. Finley, and C.-N. J. Yu. Cutting-plane training of structural SVMs. Machine Learning, 77(1):27--59, 2009.
[16]
R. N. Jorissen and M. K. Gilson. Virtual screening of molecular databases using a support vector machine. Journal of Chemical Information and Modeling, 45:549--561, 2005.
[17]
O. Komori and S. Eguchi. A boosting method for maximizing the partial area under the ROC curve. BMC Bioinformatics, 11:314, 2010.
[18]
H. Narasimhan and S. Agarwal. A structural SVM based approach for optimizing partial AUC. In ICML, 2013.
[19]
H. Narasimhan and S. Agarwal. SVM pAUC tight: A new support vector method for optimizing partial AUC based on a tight convex upper bound (full version). http://clweb.csa.iisc.ernet.in/harikrishna/Papers/SVMpAUC-tight/kdd-full.pdf, 2013.
[20]
M. S. Pepe and M. L. Thompson. Combining diagnostic test results to increase accuracy. Biostatistics, 1(2):123--140, 2000.
[21]
Y. Qi, Z. Bar-Joseph, and J. Klein-seetharaman. Evaluation of different biological data and computational classification methods for use in protein interaction prediction. Proteins, 63:490--500, 2006.
[22]
A. Rakotomamonjy. Sparse support vector infinite push. In ICML, 2012.
[23]
R. B. Rao, O. Yakhnenko, and B. Krishnapuram. KDD Cup 2008 and the workshop on mining medical data. SIGKDD Explor. Newsletter, 10(2):34--38, 2008.
[24]
M. T. Ricamato and F. Tortorella. Partial AUC maximization in a linear combination of dichotomizers. Pattern Recognition, 44(10--11):2669--2677, 2011.
[25]
C. Rudin. The p-norm push: A simple convex ranking algorithm that concentrates at the top of the list. Journal of Machine Learning Research, 10:2233--2271, 2009.
[26]
Y. Singer and N. Srebro. Pegasos: Primal estimated sub-gradient solver for SVM. In ICML, 2007.
[27]
T. Takenouchi, O. Komori, and S. Eguchi. An extension of the receiver operating characteristic curve and AUC-optimal classification. Neural Computation, 24(10):2789--2824, 2012.
[28]
R. Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society, Series B, 58:267--288, 1994.
[29]
E. van den Berg, M. Schmidt, M. P. Friedlander, and K. Murphy. Group sparsity via linear-time projection. Technical Report TR-2008-09, UBC, 2008.
[30]
Z. Wang and Y.-C. Chang. Marker selection via maximizing the partial area under the ROC curve of linear risk scores. Biostatistics, 12(2):369--385, 2011.
[31]
Z. Wang and J. Shawe-Taylor. Large-margin structured prediction via linear programming. In AISTATS, 2009.
[32]
S.-H. Wu, K.-P. Lin, C.-M. Chen, and M.-S. Chen. Asymmetric support vector machines: low false-positive learning under the user tolerance. In KDD, 2008.
[33]
M. Yuan, M. Yuan, Y. Lin, and Y. Lin. Model selection and estimation in regression with grouped variables. Journal of the Royal Statistical Society, Series B, 68:49--67, 2006.
[34]
J. Zhu, E. P. Xing, and B. Zhang. Primal sparse max-margin markov networks. In KDD, 2009.
[35]
H. Zou and T. Hastie. Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society, Series B, 67:301--320, 2005.

Cited By

View all
  • (2024)Weakly Supervised AUC Optimization: A Unified Partial AUC ApproachIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2024.335781446:7(4780-4795)Online publication date: Jul-2024
  • (2024)Malicious Log Detection Using Machine Learning to Maximize the Partial AUC2024 IEEE 21st Consumer Communications & Networking Conference (CCNC)10.1109/CCNC51664.2024.10454779(339-344)Online publication date: 6-Jan-2024
  • (2023)Rank-based Decomposable Losses in Machine Learning: A SurveyIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2023.3296062(1-20)Online publication date: 2023
  • Show More Cited By

Index Terms

  1. SVMpAUCtight: a new support vector method for optimizing partial AUC based on a tight convex upper bound

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    KDD '13: Proceedings of the 19th ACM SIGKDD international conference on Knowledge discovery and data mining
    August 2013
    1534 pages
    ISBN:9781450321747
    DOI:10.1145/2487575
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 11 August 2013

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. cutting-plane method
    2. partial auc
    3. roc curve
    4. svm

    Qualifiers

    • Research-article

    Conference

    KDD' 13
    Sponsor:

    Acceptance Rates

    KDD '13 Paper Acceptance Rate 125 of 726 submissions, 17%;
    Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

    Upcoming Conference

    KDD '25

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)18
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 28 Feb 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Weakly Supervised AUC Optimization: A Unified Partial AUC ApproachIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2024.335781446:7(4780-4795)Online publication date: Jul-2024
    • (2024)Malicious Log Detection Using Machine Learning to Maximize the Partial AUC2024 IEEE 21st Consumer Communications & Networking Conference (CCNC)10.1109/CCNC51664.2024.10454779(339-344)Online publication date: 6-Jan-2024
    • (2023)Rank-based Decomposable Losses in Machine Learning: A SurveyIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2023.3296062(1-20)Online publication date: 2023
    • (2023)Optimizing Two-Way Partial AUC With an End-to-End FrameworkIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2022.318531145:8(10228-10246)Online publication date: Aug-2023
    • (2022)Large-scale optimization of partial AUC in a range of false positive ratesProceedings of the 36th International Conference on Neural Information Processing Systems10.5555/3600270.3602535(31239-31253)Online publication date: 28-Nov-2022
    • (2022)AUC Maximization in the Era of Big Data and AI: A SurveyACM Computing Surveys10.1145/355472955:8(1-37)Online publication date: 3-Aug-2022
    • (2022)Optimizing Partial Area Under the Top-K Curve: Theory and PracticeIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2022.3199970(1-27)Online publication date: 2022
    • (2022)Learning With Multiclass AUC: Theory and AlgorithmsIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2021.310112544:11(7747-7763)Online publication date: 1-Nov-2022
    • (2022)Transfer Anomaly Detection for Maximizing the Partial AUC2022 International Joint Conference on Neural Networks (IJCNN)10.1109/IJCNN55064.2022.9892011(1-8)Online publication date: 18-Jul-2022
    • (2021)Identifying Benchmarks for Failure Prediction in Industry 4.0Informatics10.3390/informatics80400688:4(68)Online publication date: 30-Sep-2021
    • Show More Cited By

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media