As a guest user you are not logged in or recognized by your IP address. You have
access to the Front Matter, Abstracts, Author Index, Subject Index and the full
text of Open Access publications.
Deep learning techniques are the current state-of-the-art in image classification techniques, but they have the drawback of requiring a large number of labelled training images. In this context, Visual Dictionaries (Codebooks or Bag of Visual Words – BoVW) are a very interesting alternative to Deep Learning models whenever we have a low number of labelled training images. However, the methods usually extract interest points from each image independently, which then leads to unrelated interest points among the classes and they build one visual dictionary in an unsupervised fashion. In this work, we present: 1) the use of class-specific superpixel segmentation to define interest points that are common for each class, 2) the use of Convolutional Neural Networks (CNNs) with random weights to cope with the absence of labeled data and extract more representative local features, and 3) the construction of specialized visual dictionaries for each class. We conduct experiments to show that our method can outperform a CNN trained from a small set of labelled images and can be equivalent to a CNN with pre-trained features. Also, we show that our method is better than other traditional BoVW approaches.
This website uses cookies
We use cookies to provide you with the best possible experience. They also allow us to analyze user behavior in order to constantly improve the website for you. Info about the privacy policy of IOS Press.
This website uses cookies
We use cookies to provide you with the best possible experience. They also allow us to analyze user behavior in order to constantly improve the website for you. Info about the privacy policy of IOS Press.