In this paper we propose a new and effective scheme for applying frequent itemset mining to image classification tasks. We refer to the new set of obtained patterns as Frequent Local Histograms or FLHs. During the construction of the FLHs, we pay special attention to keep all the local histogram information during the mining process and to select the most relevant reduced set of FLH patterns for classification. The careful choice of the visual primitives and some proposed extensions to exploit other visual cues such as colour or global spatial information allow us to build powerful bag-of-FLH-based image representations. We show that these bag-of-FLHs are more discriminative than traditional bag-of-words and yield state-of-the art results on various image classification benchmarks.
Fig :Histogram mining process
In order to avoid information loss and to make patterns more discriminative we directly mine the local histograms. An example of local histogram mining is shown below.
Table :Sample local histogram data. Transaction ID 1,2 and 3 are generated from image I_1 and transaction ID 4, 5 and 6 are generated from image I_2.
Table :Histogram patterns of length 1, 2 and 3 are shown respectively.
Using a support threshold of 2 we select set of closed patterns.
Table :Candidate frequent local histograms or FLHs.
Fig :The difference between BOW histogram and FLH histogram for image I_1 and image I_2. (Left BOW and right FLH)
There are three key elements for the success of FLH patterns.
Fig :Effect of neighbourhood size,dictionary size and SIFT feature size on Oxford-Flower-17 dataset.
Fig :Effect of neighbourhood size on several datasets.
Relevant pattern mining is used to select the most suitable set of patterns for classification.
Fig : Some relevant patterns on Oxford-Flower-17 dataset.
Dataset | FLH | FLH+BOW | GRID-FLH |
GRAZ-Person | 94.0 | 95.0 | 95.8 |
GRAZ-Bike | 89.2 | 90.1 | 91.4 |
Table : Results using GRAZ-01.
Dataset | FLH | FLH+BOW | GRID-FLH |
Oxford Flower Using ColorName | 72.1 | 74.4 | 74.8 |
Oxford Flower Using SIFT | 92.5 | 92.7 | 92.9 |
Oxford Flower Using ColorName+SIFT | 92.5 | 92.7 | 92.9 |
Dataset | FLH_C+FLH_S | FLH_CS | FLH_CS+BOW |
Oxford Flower Using ColorName+SIFT | 93.0 | 93.4 | 94.5 |
Table : Results using Oxford-Flower17.
Dataset | FLH | FLH+BOW | GRID-FLH |
15-Scenes | 70.4 | 83.0 | 86.2 |
Table : Results using 15-Scences.
Class | ![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
|
FLH | 67.9 | 70.6 | 41.0 | 54.6 | 64.9 | 60.9 | 85.8 | 56.6 | 59.6 | 40.0 | |
FLH+BOW | 69.2 | 73.0 | 42.7 | 56.3 | 64.9 | 60.9 | 86.6 | 58.9 | 63.3 | 41.8 | |
Class | ![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
Map |
FLH | 64.7 | 47.3 | 56.6 | 65.7 | 80.7 | 46.3 | 41.8 | 54.6 | 71.0 | 77.6 | 60.4 |
FLH+BOW | 74.3 | 48.4 | 61.8 | 68.4 | 81.2 | 48.5 | 41.8 | 60.4 | 72.1 | 80.8 | 62.8 |
Table : PASCAL-VOC2007 results.
Mining Mid-level Features for Image Classification
The authors acknowledge the support of the IBBT Impact project Beeldcanon and the FP7 ERC Starting Grant 240530 COGNIMUND **.