skip to main content
10.1145/2733373.2806325acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
short-paper

Discriminative Light Unsupervised Learning Network for Image Representation and Classification

Published: 13 October 2015 Publication History

Abstract

This paper proposes a discriminative light unsupervised learning network (DLUN) to counter the image classification challenge. Compared with the traditional convolutional networks learning filters by the time-consuming stochastic gradient descent, DLUN learns the filter bank from diverse image patches with the classical K-means, which significantly reduces the training complexity while maintains the high discriminative ability. Besides, we design a new pooling strategy named voting pooling which considers the contribution difference of the adjacent activations. In the output layer, DLUN computes histograms in the size-changed dense sliding windows, followed by a max pooling operation on histogram bins at different scales to obtain the most competitive features. The classification performance on two widely used benchmarks verifies that DLUN is competitive among some state-of-the-arts.

References

[1]
T.-H. Chan, K. Jia, S. Gao, J. Lu, Z. Zeng, and Y. Ma. Pcanet: A simple deep learning baseline for image classification? In arXiv:1404.3606v2., 2014.
[2]
A. Coates, H. Lee, and A. Y. Ng. An analysis of single-layer networks in unsupervised feature learning. In AISTATS 14, 2011.
[3]
A. Coates and A. Y. Ng. The importance of encoding versus training with sparse coding and vector quantization. In ICML, 2011.
[4]
K. He, X. Zhang, S. Ren, and J. Sun. Spatial pyramid pooling in deep convolutional networks for visual recognition. In ECCV, 2014.
[5]
Y. Jia, C. Huang, and T. Darrell. Beyond spatial pyramids: Receptive field learning for pooled image features. In CVPR, 2012.
[6]
Y. Jia, O. Vinyals, and T. Darrell. Pooling-invariant image feature learning. In arXiv:1302.5056v1, 2013.
[7]
A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural network. In NIPS, 2012.
[8]
D. Li, L. Yang, X.-S. Hua, and H.-J. Zhang. Large-scale robust visual codebook construction. In ACM MM, 2010.
[9]
Y. Liang, L. Dong, S. Xie, N. Lv, and Z. Xu. Compact feature based clustering for large-scale image retrieval. In ICME, 2014.
[10]
K. Sohn and H. Lee. Learning invariant representations with local transformations. In ICML, 2012.
[11]
M. Thom and G. Palm. Sparse activity and sparse connectivity in supervised learning. Journal of Machine Learning Research, 14(3):1091--1143, 2013.
[12]
L. Wan, M. Zeiler, S. Zhang, Y. LeCun, and R. Fergus. Regularization of neural networks using dropconnect. In ICML, 2013.
[13]
K. Yu, Y. Lin, and J. Lafferty. Learning image representations from the pixel level via hierarchical sparse coding. In CVPR, 2011.

Cited By

View all
  • (2019)Automatic Post-Disaster Damage Mapping Using Deep-Learning Techniques for Change Detection: Case Study of the Tohoku TsunamiRemote Sensing10.3390/rs1109112311:9(1123)Online publication date: 10-May-2019
  • (2019)Bangla language modeling algorithm for automatic recognition of hand-sign-spelled Bangla sign languageFrontiers of Computer Science10.1007/s11704-018-7253-314:3Online publication date: 7-Dec-2019
  • (2017)E-GrabCutFrontiers of Computer Science: Selected Publications from Chinese Universities10.1007/s11704-016-5558-711:4(649-660)Online publication date: 1-Aug-2017

Index Terms

  1. Discriminative Light Unsupervised Learning Network for Image Representation and Classification

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    MM '15: Proceedings of the 23rd ACM international conference on Multimedia
    October 2015
    1402 pages
    ISBN:9781450334594
    DOI:10.1145/2733373
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 13 October 2015

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. image classification
    2. image representation
    3. unsupervised learning network

    Qualifiers

    • Short-paper

    Funding Sources

    Conference

    MM '15
    Sponsor:
    MM '15: ACM Multimedia Conference
    October 26 - 30, 2015
    Brisbane, Australia

    Acceptance Rates

    MM '15 Paper Acceptance Rate 56 of 252 submissions, 22%;
    Overall Acceptance Rate 2,145 of 8,556 submissions, 25%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)3
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 03 Mar 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2019)Automatic Post-Disaster Damage Mapping Using Deep-Learning Techniques for Change Detection: Case Study of the Tohoku TsunamiRemote Sensing10.3390/rs1109112311:9(1123)Online publication date: 10-May-2019
    • (2019)Bangla language modeling algorithm for automatic recognition of hand-sign-spelled Bangla sign languageFrontiers of Computer Science10.1007/s11704-018-7253-314:3Online publication date: 7-Dec-2019
    • (2017)E-GrabCutFrontiers of Computer Science: Selected Publications from Chinese Universities10.1007/s11704-016-5558-711:4(649-660)Online publication date: 1-Aug-2017

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media