19
Pattern Recognition 76 (2018) 50–68 Contents lists available at ScienceDirect Pattern Recognition journal homepage: www.elsevier.com/locate/patcog Color texture description with novel local binary patterns for effective image retrieval Chandan Singh a,, Ekta Walia b , Kanwal Preet Kaur a a Department of Computer Science, Punjabi University, Patiala 147002, India b Department of Computer Science, University of Saskatchewan, Canada a r t i c l e i n f o Article history: Received 22 June 2016 Revised 31 July 2017 Accepted 16 October 2017 Available online 17 October 2017 Keywords: Local binary pattern (LBP) Local binary pattern for color images (LBPC) Local binary pattern of hue component (LBPH) Local color texture Image retrieval a b s t r a c t We propose a novel local color texture descriptor called local binary pattern for color images (LBPC). The proposed descriptor uses a plane to threshold color pixels in the neighborhood of a local window into two categories. To boost the discriminative power of the proposed LBPC operator, local binary patterns of the hue component in the HSI color space, called the local binary pattern of the hue (LBPH) is derived. Further, LBPC, LBPH are fused to derive LBPC+LBPH which when combined with the color histogram (CH) of the hue component results in an effective image retrieval method LBPC+LBPH+CH. The uniform patterns of the two proposed descriptors ULBPC and ULBPH are combined to yield another low dimension local color descriptor ULBPC+ULBPH+CH which provides a good tradeoff between retrieval accuracy and speed. Detailed experiments conducted on Wang, Holidays, Corel-5K and Corel-10K datasets demonstrate that the proposed low dimension descriptors LBPC+LBPH+CH, and ULBPC+ULBPH+CH outperform the state-of-the-art color texture descriptors in terms of retrieval accuracy and speed. © 2017 Elsevier Ltd. All rights reserved. 1. Introduction Content-based image retrieval (CBIR) is one of the active re- search areas in the field of pattern recognition and artificial intel- ligence. Even after more than three decades of extensive research, the quest for more-and-more effective CBIR systems has attracted many researchers to develop systems which provide high retrieval rates within low retrieval time. The early CBIR systems were fo- cused on the grayscale images, but with the widespread use of color images over Internet and the advantage of a color attribute for discrimination purpose, color information is being incorporated to enhance the performance of retrieval systems. A retrieval sys- tem provides a user with a way to access, browse and fetch im- ages efficiently from databases. These databases are used in a vari- ety of fields including information security, biometric system (iris, fingerprint, and face matching), biodiversity, digital library, crime prevention, medical imaging, historical archives, video surveillance, human-computer interaction, etc. The CBIR systems depend heavily on two steps: feature extrac- tion and feature matching. Feature extraction is the most impor- tant step as it requires an image to be represented by highly dis- criminative features with small variations among the features of intra-class images and high variations among inter-class images. Corresponding author. E-mail addresses: [email protected] (C. Singh), [email protected] (E. Walia), [email protected] (K.P. Kaur). The features are desired to be robust to geometric and photometric changes such as translation, rotation, scale, occlusion, illumination, viewpoint, etc. Broadly, there are two categories of feature representation: (1) global or holistic approach and (2) local approach. A global ap- proach extracts features from the whole image ignoring the local characteristics and spatial relationship between pixels. They are computationally efficient and robust to image noise. The global methods suffer from their inadequacies to handle some of the is- sues related to occlusion, view points and illumination changes, and local characteristics of image shape. These issues are well ad- dressed by the local feature extraction methods which extract fea- tures from local regions of an image. These local regions can be as simple as partitions of an image or are selected through key points. The most common global feature extraction techniques based on color include the MPEG-7 feature sets such as color histograms, dominant color descriptor, scalable color descriptor, and color lay- out descriptor [1]. The color histogram features have been ob- served to yield very good performance; they are invariant to trans- lation and rotation and can be made scale invariant after normal- ization by the image size. Texture based global feature extraction methods include gray level co-occurrence matrices [2], Tamura tex- ture features [3], Markov random field model [4] and Gabor fil- tering [5]. A comparative performance analysis of these texture features on the retrieval of texture images establishes the supe- riority of Gabor filters over others [6]. Effect of different Gabor filter parameters on texture image retrieval has been studied in https://doi.org/10.1016/j.patcog.2017.10.021 0031-3203/© 2017 Elsevier Ltd. All rights reserved.

Color texture description with novel local binary patterns

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Color texture description with novel local binary patterns

Pattern Recognition 76 (2018) 50–68

Contents lists available at ScienceDirect

Pattern Recognition

journal homepage: www.elsevier.com/locate/patcog

Color texture description with novel local binary patterns for effective

image retrieval

Chandan Singh

a , ∗, Ekta Walia

b , Kanwal Preet Kaur a

a Department of Computer Science, Punjabi University, Patiala 147002, India b Department of Computer Science, University of Saskatchewan, Canada

a r t i c l e i n f o

Article history:

Received 22 June 2016

Revised 31 July 2017

Accepted 16 October 2017

Available online 17 October 2017

Keywords:

Local binary pattern (LBP)

Local binary pattern for color images (LBPC)

Local binary pattern of hue component

(LBPH)

Local color texture

Image retrieval

a b s t r a c t

We propose a novel local color texture descriptor called local binary pattern for color images (LBPC). The

proposed descriptor uses a plane to threshold color pixels in the neighborhood of a local window into

two categories. To boost the discriminative power of the proposed LBPC operator, local binary patterns of

the hue component in the HSI color space, called the local binary pattern of the hue (LBPH) is derived.

Further, LBPC, LBPH are fused to derive LBPC + LBPH which when combined with the color histogram

(CH) of the hue component results in an effective image retrieval method LBPC + LBPH + CH. The uniform

patterns of the two proposed descriptors ULBPC and ULBPH are combined to yield another low dimension

local color descriptor ULBPC + ULBPH + CH which provides a good tradeoff between retrieval accuracy and

speed. Detailed experiments conducted on Wang, Holidays, Corel- 5K and Corel- 10K datasets demonstrate

that the proposed low dimension descriptors LBPC + LBPH + CH, and ULBPC + ULBPH + CH outperform the

state-of-the-art color texture descriptors in terms of retrieval accuracy and speed.

© 2017 Elsevier Ltd. All rights reserved.

T

c

v

g

p

c

c

m

s

a

d

t

s

o

d

o

s

l

i

m

t

1. Introduction

Content-based image retrieval (CBIR) is one of the active re-

search areas in the field of pattern recognition and artificial intel-

ligence. Even after more than three decades of extensive research,

the quest for more-and-more effective CBIR systems has attracted

many researchers to develop systems which provide high retrieval

rates within low retrieval time. The early CBIR systems were fo-

cused on the grayscale images, but with the widespread use of

color images over Internet and the advantage of a color attribute

for discrimination purpose, color information is being incorporated

to enhance the performance of retrieval systems. A retrieval sys-

tem provides a user with a way to access, browse and fetch im-

ages efficiently from databases. These databases are used in a vari-

ety of fields including information security, biometric system (iris,

fingerprint, and face matching), biodiversity, digital library, crime

prevention, medical imaging, historical archives, video surveillance,

human-computer interaction, etc.

The CBIR systems depend heavily on two steps: feature extrac-

tion and feature matching. Feature extraction is the most impor-

tant step as it requires an image to be represented by highly dis-

criminative features with small variations among the features of

intra-class images and high variations among inter-class images.

∗ Corresponding author.

E-mail addresses: [email protected] (C. Singh), [email protected] (E.

Walia), [email protected] (K.P. Kaur).

t

f

r

fi

https://doi.org/10.1016/j.patcog.2017.10.021

0031-3203/© 2017 Elsevier Ltd. All rights reserved.

he features are desired to be robust to geometric and photometric

hanges such as translation, rotation, scale, occlusion, illumination,

iewpoint, etc.

Broadly, there are two categories of feature representation: (1)

lobal or holistic approach and (2) local approach. A global ap-

roach extracts features from the whole image ignoring the local

haracteristics and spatial relationship between pixels. They are

omputationally efficient and robust to image noise. The global

ethods suffer from their inadequacies to handle some of the is-

ues related to occlusion, view points and illumination changes,

nd local characteristics of image shape. These issues are well ad-

ressed by the local feature extraction methods which extract fea-

ures from local regions of an image. These local regions can be as

imple as partitions of an image or are selected through key points.

The most common global feature extraction techniques based

n color include the MPEG-7 feature sets such as color histograms,

ominant color descriptor, scalable color descriptor, and color lay-

ut descriptor [1] . The color histogram features have been ob-

erved to yield very good performance; they are invariant to trans-

ation and rotation and can be made scale invariant after normal-

zation by the image size. Texture based global feature extraction

ethods include gray level co-occurrence matrices [2] , Tamura tex-

ure features [3] , Markov random field model [4] and Gabor fil-

ering [5] . A comparative performance analysis of these texture

eatures on the retrieval of texture images establishes the supe-

iority of Gabor filters over others [6] . Effect of different Gabor

lter parameters on texture image retrieval has been studied in

Page 2: Color texture description with novel local binary patterns

C. Singh et al. / Pattern Recognition 76 (2018) 50–68 51

[

f

T

m

e

t

o

t

v

w

i

b

s

r

p

t

m

c

r

t

o

[

R

e

t

g

g

c

t

t

f

t

e

c

[

p

p

a

v

c

f

c

A

i

v

h

i

i

s

p

a

c

e

I

c

t

m

f

p

o

r

n

Y

o

b

g

[

r

c

p

d

a

c

t

t

O

i

c

r

m

Q

o

i

t

l

t

t

t

d

b

r

p

s

l

t

a

g

t

i

t

p

t

p

l

t

o

a

f

f

i

w

r

u

F

c

a

f

p

t

d

t

T

[

h

w

o

u

t

7] . Rotation and scale invariant optimum Gabor filter parameters

or texture image retrieval have been derived by Han and Ma [8] .

hey have observed that the optimum filter parameters provide

uch superior performance than the conventional filter param-

ters. Bianconi and Fernandez [9] have also improved Gabor fil-

er parameters to provide better texture classification results. One

f the problems associated with Gabor filtering is that its fea-

ure extraction process is very slow [7] . Shape features also pro-

ide powerful information for image retrieval. Shape matching is a

ell-researched area with many shape representation and match-

ng techniques [10] . Shape features are generally used in two ways:

oundary-based and region-based. In the boundary-based repre-

entation of an image, the boundary or contour of an object is

equired to be extracted, while in the region-based techniques all

ixels of an image take part in the computation of the shape fea-

ures [10,11] .

The global feature-based methods have been the mainstay of

any CBIR systems. However, their inability to deal with many

omplex issues related to geometric deformations and photomet-

ic changes have necessitated the extensive use of local feature ex-

raction methods, which can resolve these issues effectively. Some

f the effective local feature extraction methods are LBP [12] , SIFT

13] , PCA-SIFT [14] , GLOH [15] , SURF [16] , HOG [17] , DAISY [18] ,

ank-SIFT [19] , BRIEF [20] , ORB [21] , WLD [22] , and many more.

Color and texture provide important information for deriving

ffective f eatures there by ensuring high performance of the re-

rieval systems. The classical texture features were derived for the

rayscale images. Among the various local texture descriptors for

rayscale images, SIFT has proven to be the most effective and suc-

essful descriptor in the state-of-the-art recognition and classifica-

ion system [23] . To capture the texture of color images, it is ex-

ended to several variants e.g. color SIFT [24] . The color SIFT has

urther been evaluated against several color descriptors and found

o outperform them. However, color SIFT is computation intensive

specially when the size of the image or size of the database in-

reases. The local binary pattern (LBP), developed by Ojala et al.

12] is also an effective texture descriptor which is found to be

owerful and successful in many pattern recognition and com-

uter vision applications. It captures local texture features, which

re invariant to illumination changes. It is simple, fast and pro-

ides strong discriminative power as compared to many other lo-

al texture descriptors. The LBP operator has been used success-

ully in texture classification [25-28] , face recognition [29-31] , fa-

ial expression recognition [32-35] , and image retrieval [36,37] , etc.

survey paper [38] provides a comprehensive discussion on facial

mage analysis using the classical LBP operator and several of its

ariants and cites 159 papers in this area.

Most of the works on classical LBP operator and its variants

ave been developed for gray scale image processing. The increas-

ng demand for color images over Internet and their ever increas-

ng use for many practical applications have motivated the re-

earchers to develop descriptors which can represent color texture

attern as effectively as the LBP operator does for gray scale im-

ges. A natural extension of the gray scale LBP operator is to pro-

ess each channel of a color image as a gray scale image. This strat-

gy was used by Mäenpää et al. [39] for color texture description.

n their multispectral LBP (MSLBP), they use three channels of a

olor image and six sets of LBPs opponent color to capture spa-

ial correlation between the two channels of the color spectra. The

ethod is effective. However, it results in a very high dimensional

eature vector. Later, Mäenpää and Pietikäinen [40] conducted ex-

eriments on color texture and observed that instead of taking six

pponent color components only three pairs are sufficient for rep-

esenting cross correlation features between the three color chan-

els. Choi et al. [41] derived LBP histograms for each channel in

C b C r color space and applied PCA to reduce the dimensionality

f the feature vector. They observed that their method performed

etter in face recognition application than the LBP operator on

ray face images, which were derived from color images. Lee et al.

42] derived local color vector binary patterns (LCVBPs) for face

ecognition problem. The LCVBP operator consists of two parts:

olor norm patterns and color angular patterns. The color angular

attern captures discriminative features of two color spectra and

erives spatial correlation of local color texture. The LCVBP oper-

tor is very effective as compared to the LBP of individual color

hannels. To reduce the dimension of LBP operator and apply it

o color images, Zhu et al. [43] proposed an orthogonal combina-

ion of local binary patterns (RGB-OC-LBP). Their proposed RGB-

C-LBP operator performs better than the color SIFT operator in

mage matching, object recognition, and scene classification appli-

ations. Recently, Lan et al. [44] have introduced quaternion local

anking binary pattern (QLRBP), which combines the color infor-

ation provided by multispectral channels in color images. The

LRBP operator is derived using quaternionic representation (QR)

f color images. The QLRBP can handle all color channels directly

n the quaternionic domain and represents color texture without

reating the color channels separately. Li et al. [45] have developed

ocal similarity pattern (CLSP) for representing the color image as

he co-occurrence of its image pixel color quantization informa-

ion and the local color image textural information. In an attempt

o utilize the cross channel information, Dubey et al. [46] have

eveloped two sets of patterns, multichannel adder and decoder-

ased LBPs (MDLBPs) and observed that the latter provides better

etrieval performance than the former. The performance of their

roposed approach is better than the other similar LBP-based de-

criptors, but the size of the feature vectors for both methods is

arge.

In this paper, we propose an operator called local binary pat-

ern for color images (LBPC), which derives texture patterns for

color image similar to the way LBP operator derives texture for

ray scale images. For this purpose, we treat a color pixel as a vec-

or having m -components and form a hyperplane. The hyperplane

s used as a boundary to threshold and partition color pixels into

wo classes. A color pixel in a 3 × 3 neighborhood of the current

ixel is assigned a value 1 if it lies on or above the plane, and

he value 0, if it is below the plane. Thus, the proposed operator

rovides spatial relationship among color pixels, which represents

ocal texture features. We compute histograms of the binary pat-

erns obtained from color images in a way akin to the histograms

f LBP operator for gray scale images. If we use 8-neighborhood of

color pixel, then we have 256 histogram bins, which are used as

eatures representing local texture patterns of color images. We re-

er this operator as the LBPC operator. The dimensionality of LBPC

s reduced by deriving uniform patterns [12] with 59 bins. Since

e are investigating effective descriptors for color images, a natu-

al question arises as what is the retrieval performance when we

se local binary patterns of the color component of color images.

or the purpose of representing local binary patterns of the color

omponents, the H component in the HSI color model [47] is an

ppropriate choice as compared to all other color models. There-

ore, in our second proposed method, we derive the local binary

atterns of H component of HSI color model. We refer this descrip-

or as local binary pattern of the hue (H) component (LBPH). The

iscriminative power of LBPC is enhanced by fusing the LBPC fea-

ures with LBPH features. This approach is referred as LBPC + LBPH.

he color histograms are among the best color image descriptors

4 8,4 9] and hence, in our proposed third approach, we derive color

istogram (CH) of H channel in the HSI color model and fuse it

ith LBPC + LBPH. This approach is referred as LBPC + LBPH + CH. In

rder to reduce the dimension of LBPC + LBPH + CH, we derive the

niform patterns of LBPC and LBPH and fuse them to the CH fea-

ures, which yields an effective low dimension local color texture

Page 3: Color texture description with novel local binary patterns

52 C. Singh et al. / Pattern Recognition 76 (2018) 50–68

w

n

g

s

λ

θ

s

a

g

w

u

t

f

t

2

o

a

b

h

o

4

t

r

O

a

O

w

S

a

(

t

r

t

c

t

o

t

t

f

f

f

2

t

p

s

c

descriptor called ULBPC + ULBPH + CH whose performance is compa-

rable to LBPC + LBPH + CH. All existing state-of-the-art methods and

proposed methods have been implemented to analyze and com-

pare their image retrieval performance. The feature vector dimen-

sion plays an important role in analyzing retrieval speed of a fea-

ture extraction method. Therefore, the dimension of each method

is taken into account while performing their comparative analysis.

The rest of the paper is organized as follows. In Section 2 , we

provide an overview of some of the existing state-of-the-art local

binary pattern-based descriptors for color images. The proposed

operator LBPC has been derived in Section 3 . Discussion on the

parameters used by LBPC operator is presented in Section 4 . In

Section 5 , we discuss the LBPH operator and the color histogram

(CH) features and their fusion with the LBPC features. Section 6 ex-

plains the various similarity measures and performance parameters

used in image retrieval applications. Detailed experimental analy-

sis of retrieval performance and computation time on various color

image databases is presented in Section 7 . Section 8 concludes the

paper.

2. Related work

In this section, we present an overview of the closely related

works which pertain to the LBP-like methods developed for the

color images to utilize the benefits of cross-correlation among the

color channels. Each method has its own merits and demerits as

explained below.

2.1. Multispectral local binary pattern (MSLBP)

Mäenpää et al. [39] use three channels of a color image in the

RGB color space and six pairs of the opponent colors. The oppo-

nent colors are used to obtain color texture features of two col-

ors, which represent cross-correlation between color values as well

as spatial relationships between them. The three LBP feature vec-

tors are obtained in a way similar to the gray scale LBP features

by treating each of the channels of an RGB image as a gray scale

image [12] . The six opponent LBP feature vectors are derived as

follows:

MSLB P ( i, j ) ( x c , y c ) =

P −1 ∑

p=0

S( v i ( x p , y p ) − v j ( x c , y c ) ) × 2

p , (1)

where ( i, j ) ∈ {(1, 2), (2, 3), (3, 1), (2, 1), (3, 2), (1, 3)}, and

S (v i ( x p , y p ) −v j ( x c , y c )

)=

{1 , i f v i ( x p , y p ) − v j ( x c , y c ) ≥ 0 ,

0 , otherwise.

(2)

Here, v j ( x c , y c ) is the intensity value of the center pixel of a

3 × 3 window from the j th color component image, v i ( x p , y p ) is the

intensity value of the p th neighborhood pixel from the i th color

component image. The total number of MSLBP features is 2304,

which is the size of the feature vector obtained after the concate-

nation of nine LBP operators. The approach provides high recogni-

tion rate. However, the size of the feature vector is too large which

slows down the retrieval speed.

2.2. Local color vector binary pattern (LCVBP)

Local color vector binary pattern [42] consists of color norm bi-

nary patterns and color angular binary patterns. For color norm

binary patterns, we take the norm of all color pixel values as fol-

lows:

I ( x, y ) = ‖

I ( x, y ) ‖

=

r 2 ( x, y ) + g 2 ( x, y ) + b 2 ( x, y ) , (3)

here r ( x, y ), g ( x, y ), b ( x, y ) are the pixel values of R, G, B compo-

ents at ( x, y ) coordinates. Thus, a color image is converted into a

ray scale image I ( x, y ) from which LBP feature vector is obtained.

The color angular features value θ ( i, j ) between the i th and j th

pectral bands is computed by

( i, j ) ( x, y ) =

v i ( x, y )

v j ( x, y ) + ε, ( i, j ) ∈ { ( 1 , 2 ) , ( 2 , 3 ) , ( 1 , 3 ) } (4)

( i, j ) ( x, y ) = ta n

−1 (λ( i, j ) ( x, y )

). (5)

Here, ε is a small parameter to avoid the denominator to as-

ume zero value.

The images λ( i, j ) ( x, y ) are calculated between R and G, G and B,

nd B and R components. The three images λ( i, j ) ( x, y ) are treated as

ray images and their three LBP feature vectors are computed in a

ay similar to the LBP operator for gray images. The full LCVBP

ses 4 × 256 = 1024 features. To reduce the number of features,

he authors have proposed uniform LCVBP for which the size of

eature vector is 4 × 59 = 236. This method provides a good

radeoff between the retrieval performance and retrieval time.

.3. Orthogonal combination of local binary pattern (OC-LBP)

The OC-LBP operator [43] reduces the dimensionality of the LBP

perator. It consists of two operators OC-LBP 1 and OC-LBP 2, which

re derived by splitting 8-neighborhood pixels into 2 sets of neigh-

orhoods each comprising 4 pixels. The first set contains the four

orizontal and vertical neighbors of the center pixel and the sec-

nd set contains the 4 diagonal neighbors. Since each set contains

pixels, the total number of patterns in each set is 16. Thus, the

wo sets contain 32 patterns of the OC-LBP operator, which are de-

ived as follows.

C − LB P 1 ( x c , y c ) =

3 ∑

p=0

S( I 2 p − I c ) × 2

p , (6)

nd

C − LB P 2 ( x c , y c ) =

3 ∑

p=0

S( I 2 p+1 − I c ) × 2

p , (7)

here

( z ) =

{1 , i f z ≥ 0 ,

0 otherwise. (8)

The values I c and I p represent intensity values of a gray im-

ge or of a component image in a color space at the center pixel

x c , y c ) and at a neighborhood pixel ( x p , y p ), respectively. In [43] ,

he OC-LBP operator has been derived on six color models, which

epresent various illumination changes. It has been observed that

he RGB-OC-LBP operator provides overall best results in the object

lassification problem. The RGB-OC-LBP operator is applied on the

hree component images in the RGB color space by treating each

f the component image as a gray image. Thus, there are 96 fea-

ures derived by the RGB-OC-LBP operator. The main purpose of

his method is to reduce the dimensionality of the feature vector

rom 768 LBP features of the component images to 96 features, 32

eatures each for the three component images. The retrieval per-

ormance is, however, low.

.4. Quaternionic local ranking binary pattern (QLRBP)

Lan et al. [44] have used quaternionic representation (QR) of

he color images to obtain color texture features. Similar to the

rocedure followed for gray scale images to obtain LBP, they con-

ider a 3 × 3 neighborhood of a color pixel. By using a reference

olor pixel ( r ′ , g ′ , b ′ ) and a color pixel ( r, g, b ) in the window, they

Page 4: Color texture description with novel local binary patterns

C. Singh et al. / Pattern Recognition 76 (2018) 50–68 53

d

q

o

d

b

θ

w

p

p

a

o

a

o

t

f

i

θ

a

7

l

m

2

p

i

W

t

d

c

u

w

t

t

i

L

w

b

w

t

m

d

C

w

t

t

t

2

b

r

[

c

o

L

t

1

i

m

m

m

{

e

l

l

t

m

m

m

l

l

f

d

m

erive an operator called QLRBP. They use Clifford translation of

uaternionic (CTQ) ranking function, which is complex. The phase

f the function is used for the ranking purpose in a 3 × 3 win-

ow. The CTQ phase between two color pixels ( r, g, b ) and ( r ′ , g ′ ,

′ ) is given by

( x, y ) = ta n

−1

( gb ′ − bg ′ ) 2 + ( br ′ − rb ′ ) 2 + ( r g ′ − gr ′ ) 2

−( r r ′ + g g ′ + bb ′ ) , (9)

here r = r( x, y ) , g = g( x, y ) , and b = b( x, y ) . The angles are com-

uted for all 9 pixels in a 3 × 3 window. One may note that the

hase angle θ ( x, y ) is the angle between two vectors I = ( r, g, b )

nd I ′ = ( r ′ , g ′ , b ′ ) . Therefore, the resulting QLRBP operator is an

perator which is based on the angle between a reference pixel I ′ nd a pixel I of the color image. Therefore, the QLRBP features are

btained by considering the angle image θ ( x, y ) in a way similar to

he LBP features of a gray scale image. In order to provide the dif-

erent weights to different color channels, the weighted L 1 -phase

s used by Lan et al. [44] as follows.

( x, y ) = ta n

−1 α1 | gb ′ − bg ′ | + α2 | br ′ − rb ′ | + α3 | r g ′ − gr ′ | −( r r ′ + g g ′ + bb ′ )

· (10)

By setting I ′ to a given value, 256 features are obtained. The

uthors have used three sets of I ′ to drive a feature vector of size

68. Since the reference vector I ′ is set for the whole image, no

ocal information is exploited fully to derive high retrieval perfor-

ance.

.5. Completed local similarity pattern (CLSP)

The CLSP method [45] consists of two phases. In the first

hase, a color image is quantized into S number of clusters us-

ng the K -means algorithm, creating a dictionary of color words

= { w 1 , w 2 , . . . w S } . A cluster word represents its center w i =( r i , g i , b i ) , i = 1 , . . . , S. Next, for each pixel p of the color image,

he Euclidean distance from each cluster center is obtained, i.e.

i 2 (p) = ‖ I p − w i ‖ 2 2

, i = 1 , 2 , . . . , S.

The nearest K 1 distances are selected and the weights u i ( p ) are

omputed as follows.

i ( p ) = u i ( x, y ) =

exp

(−βd i

2 ( p )

)∑ K 1

i =1 exp

(−βd i

2 ( p )

) , i ∈ K 1 − N N ( p ) , (11)

here K 1 − N N (p) denotes a set of K 1 -nearest neighbors of p ob-

ained by using Euclidean distance, and β is a smoothing parame-

er which is taken as 2.

The second phase of CLSP method determines the local similar-

ty pattern (LSP) in a 3 × 3 window as follows.

SP ( x c , y c ) =

P−1 ∑

p=0

b P × 2

p , (12)

here,

P =

{1 , p ∈ K 2 − N N ( x c , y c )

0 otherwise, (13)

here K 2 − N N ( x c , y c ) denotes the K 2 -nearest neighbors of the cen-

er pixel ( x c , y c ) in a 3 × 3 window using the following distance

etric

P =

I p − I c ‖

I p + I c ‖

, p = 0 , 1 , . . . , P − 1 . (14)

Finally, the CLSP histograms are obtained by using the equation:

LSP ( s, t ) =

M−1 ∑

x =0

N−1 ∑

y =0

u s ( x, y ) δ( LSP ( x, y ) = t ) , (15)

here M × N represents the size of the image, and s = 1 , 2 , . . . , S,

= 1 , 2 , . . . , 2 P . In particular, in [45] S = 10 , and P = 8 . The size of

he feature vector is S × 2 P = 2560 . Despite the large size of its fea-

ure vector, the performance of the CLSP operator is not very high.

.6. Multichannel decoded linear binary pattern

Recently, in a new approach, two multichannel decoded local

inary patterns (MDLBP) have been derived for color images to

epresent the cross-channel correlation of the color components

46] . The mentioned approach works as follows. Let the color

hannels be represented by R, G, and B. In the first stage, LBPs

f each channel using 8 neighborhood configuration are derived.

et LBP i t ( x, y ) , i = 0 , 1 , . . . , 7 , represent the i th bit of a string of

he LBP of the t th component at a pixel location ( x , y ). Here, t = , 2 , 3 , and these values correspond to the R , G , and B component

mages. In the first stage, the multichannel adder maLBP i (x, y ) , and

ulti-channel decoder mdLB P i ( x, y ) , i = 0 , 1 , . . . , 7 , are derived as:

aLB P i ( x, y ) =

3 ∑

t=1

LBP i t ( x, y ) , i = 0 , 1 , . . . , 7 , (16)

dLB P i ( x, y ) =

3 ∑

t=1

2

t−1 × LBP i t ( x, y ) , i = 0 , 1 , . . . , 7 . (17)

It is noted that maLBP i (x, y ) ∈ { 0 , 1 , 2 , 3 } , and mdLB P i ( x, y ) ∈ 0 , 1 , . . . , 7 } . Next, four multichannel adder bins maAM

k (x, y ) and

ight multichannel decoder bins mdDM

k (x, y ) are obtained as fol-

ows:

maA M

k ( x, y ) ← 0 , k = 0 , 1 , 2 , 3 , 4 ; mdD M

k ( x, y ) ← 0 ,

k = 0 , 1 , . . . , 7 , (18)

k ← maLB P i ( x, y ) ; maA M

k ( x, y ) ← maA M

k ( x, y ) + 2

i ,

i = 0 , 1 , . . . 7 , (19)

k ← md LB P i ( x, y ) ; md D M

k ( x, y ) ← md D M

k ( x, y ) + 2

i ,

i = 0 , 1 , . . . 7 . (20)

The histogram bins contain values between 0 and 255 at a pixel

ocation ( x, y ), represented by l . For the whole image, the 4 mul-

ichannel adder maAH

k (l) and 8 multichannel decoder histograms

dDM

k (l) are obtained using:

aA H

k ( l ) ← 0 , k = 0 , 1 , 2 , 3 , l = 0 , 1 , . . . , 255 , (21)

dD H

k ( l ) ← 0 , k = 0 , 1 , . . . , 7 , l = 0 , 1 , . . . , 255 , (22)

← maA M

k ( x, y ) ; maA H

k ( l ) ← maA H

k ( l ) + 1 , k = 0 , 1 , 2 , 3 ,

(23)

← md D M

k ( x, y ) ; md D H

k ( l ) ← md D H

k ( l ) + 1 , k = 0 , 1 , . . . 7 ,

(24)

or x = 0 , 1 , . . . , M − 1 , and y = 0 , 1 , . . . , N − 1 .

There are 4 × 256 multichannel adder histograms and 8 × 256

ecoder histograms. Although it provides good recognition perfor-

ance, the size of feature vectors is very high.

Page 5: Color texture description with novel local binary patterns

54 C. Singh et al. / Pattern Recognition 76 (2018) 50–68

8 neighbors(a)

(P=8, R=1)(b)

(P=12, R=1.5)(c)

(P=8, R=2)(d)

(P=16, R=3)(e)

Fig. 1. 8 neighbors and circularly symmetric neighbor sets for different ( P, R ): (a) 8 neighbors, (b) P = 8 , R = 1 , (C) P = 12 , R = 1 . 5 , (d) P = 8 , R = 2 , and (e) P = 16 , R = 3 .

a

a

d

t

o

u

t

b

4

a

e

u

t

w

t

t

a

i

T

a

s

b

n

w

n

u

s

o

t

q

3. The proposed local binary patterns for color images (LBPC)

3.1. The classical LBP operator

The classical LBP operator is defined for the gray scale images.

The general form of the operator for a circularly symmetric neigh-

bor set of P members on a circle of radius R and center ( x c , y c ),

denoted by LBP P, R ( x c , y c ) is defined as:

LB P P,R ( x c , y c ) =

P−1 ∑

p=0

S ( I ( x p , y p ) − I ( x c , y c ) ) × 2

p , (25)

where

S ( I ( x p , y p ) − I ( x c , y c ) ) =

{1 i f I ( x p , y p ) − I ( x c , y c ) ≥ 0

0 otherwise, (26)

and I ( x, y ) represents the intensity at pixel location ( x, y ). In its

simplest form, 8 neighborhood pixels of a pixel at ( x, y ) are consid-

ered, i.e., P = 8 , and R = 1 . This form provides a fast approach for

the computation of the LBP without involving a circular symmetry

of the members which needs interpolation of intensity. The gen-

eral case of the LBP P, R ( x c , y c ) specifies P equally-spaced locations

on a circle of radius R and center ( x c , y c ) whose coordinates are

determined by

x p = x c + Rcos ( 2 π p/P ) ,

y p = y c + Rsin ( 2 π p/P ) , (27)

p = 0 , 1 , . . . , P − 1 . The intensity values at locations which do not

fall at the original positions are interpolated using bilinear inter-

polation. Fig. 1 (a) depicts the 8 neighborhood configuration and

Fig. 1 (b)–(e) depict the circularly symmetric neighborhoods with

(P, R ) ∈ { (8 , 1) , (12 , 1 . 5) , (8 , 2) , (16 , 3) } in the increasing order of

the radius. The parameter P determines the angular space between

members and the parameters R determines the size of the win-

dow. The original locations of the pixels are shown by open bullets

nd the P neighborhoods by the solid bullets. An original pixel lies

t the cross-section of the horizontal and vertical lines drawn by

ashed lines. The purpose of using such types of grid is to show

he locations of the neighbors which fall on the original positions

f the pixels of the image. For example, for 8 neighborhood config-

ration shown in Fig. 1 (a), all 8 neighbors fall on the original loca-

ions whereas for P = 8 , R = 1 , and P = 8 , R = 2 , of the 8 neigh-

ors, 4 neighbors fall on the original locations and the remaining

neighbors are interpolated. For P = 12 , R = 1 . 5 , all 12 neighbors

re interpolated. The two types of grid used here show the pix-

ls which take part to interpolate a neighbor shown by gray shade

sing the bilinear interpolation which requires 4 pixels for the in-

erpolation process.

The LBP operator derives binary patterns, called LBP patterns,

hose values lie between 0 and 2 P−1 . The LBP patterns are ob-

ained for each pixel of an image and the features are obtained in

he form of the histograms of the LBP patterns. These histograms

re used to represent the texture of a gray scale image. The higher

s the value of P , the larger is the size of the histogram patterns.

o reduce the size of the feature vector “uniform” binary patterns

re used. Let LBP P, R ( x, y ) denote the LBP value at pixel ( x, y ). Let

denote the string of the binary values. Clearly, | s | = P . Out of 2 P

inary patterns only those patterns are termed as “uniform” (de-

oted by ULBP ) which satisfy the following condition:

P

i =1

| s i − s i −1 | + | s 0 − s P | ≤ 2 (28)

here s i is the i th bit of the string. All other patterns are termed

on-uniform and placed into a single group. The total number of

niform patterns is P ( P − 1 ) + 3 which is much smaller than 2 P ,

pecially when P is large. For example, for P = 8 , the total number

f the LBP and ULBP patterns are 256, and 59, respectively.

The classical LBP operator for gray scale images cannot be ex-

ended to color images because a color pixel represents a vector

uantity with ( R , G , B ) components whereas a gray scale pixel is

Page 6: Color texture description with novel local binary patterns

C. Singh et al. / Pattern Recognition 76 (2018) 50–68 55

a

E

3

t

m

c

t

a

W

m

c

a

c

fi

i

s

a

p

n

o

n

w

a

w

s

g

a

o

p

p

v

t

m

R

n

w

H

t

E

a

t

w

e

a

g

e

s

w

s

r

L

S

I

h

t

n

t

t

s

p

p

4

L

R

t

c

d

c

c

s

4

4

t

a

t

t

e

r

i

r

t

o

4

l

R

B

(

4

c

scalar quantity. Therefore, a comparison of the type given by

q. (26) cannot be made for color pixels to get a binary string.

.2. The proposed LBPC operator

The proposed LBPC operator is based on the concept of parti-

ioning or thresholding color pixels using a hyper-plane in m di-

ensions. Since m = 3 in the present case, the hyper-plane be-

omes an Euclidean plane in 3-D color space. There are other

hresholding alternatives, such as a hyper-sphere, a hyper-cube or

hyper-ellipsoid, but they are not as effective as the hyper-plane.

e have come to this conclusion after conducting detailed experi-

ents using these thresholding alternatives.

The task ahead is to derive a thresholding plane in 3-D

olor space which is performed as follows. Let a vector I( x, y ) =( r( x, y ) , g( x, y ) , b( x, y ) ) or simply, I = ( r, g, b ) be used to represent

color pixel in the R, G, B color space. Thus, the number of color

omponents m is 3. We define a local window of size ( 2 R + 1 ) ×( 2 R + 1 ) , R ≥ 1 , centered at a pixel c with the color vector I c =( r c , g c , b c ) . Let I p = ( r p , g p , b p ) be a neighborhood pixel p . We de-

ne a color plane L in the color space. A plane is derived by defin-

ng the normal to the plane and a reference point in the color

pace. Let the normal to the plane be denoted by n = ( n 1 , n 2 , n 3 )

nd the reference point R o = ( r o , g o , b o ) , then the equation of the

lane with normal vector n and reference point R o is given by:

· ( I p − R o ) = 0 , (29)

r,

1 ( r − r o ) + n 2 ( g − g o ) + n 3 ( b − b o ) = 0 . (30)

hich is the result of the dot product between the vector n

nd the vector formed by joining R o = ( r o , g o , b o ) with I p = ( r, g, b )

hich yields the vector I p − R o . The color plane divides the color

pace into two classes. All pixels that lie on or above the plane are

rouped in one class and all other pixels that lie below the plane

re grouped in another class.

We can select the normal vector n in many ways but an obvi-

us choice is a line joining the pure black pixel (0, 0, 0) and the

ure white pixel (1, 1, 1). This represents gray intensity line and all

rimary colors R, G, and B are present in equal proportion. An ob-

ious choice for the reference point would be the center pixel I c =( r c , g c , b c ) of the square neighborhood. With these values of the

wo parameters, we define a plane called color plane, which is nor-

al to the line joining (0, 0, 0) and (1, 1, 1) and passes through

o = I c = ( r c , g c , b c ) . This facilitates the process of thresholding the

eighborhood pixels I p , p = 0 , 1 , . . . .P − 1 into two classes: those

hich lie on or above-plane and those which lie below-plane.

ere, P is the total number of neighborhood pixels. A decision to

his effect can be taken by evaluating the following expression.

( I p ) = E ( r, g, b ) = n 1 ( r p − r c ) + n 2 ( g p − g c ) + n 3 ( b p − b c ) , (31)

A color pixel I p = ( r p , g p , b p ) is above or on the plane, if E ( I p ) ≥ 0

nd it is below the plane if E ( I p ) < 0 . Therefore, the color pixels in

he neighborhood of the center pixel are divided into two classes

ith a well-defined mechanism. This can be thought of a natural

xtension of the local binary patterns (LBP) of the gray scale im-

ges which are obtained by thresholding the gray values about the

ray value of the center pixel ( x c , y c ) of a local window. In fact, the

xpression in the right hand side in Eq. (31) reduces to an expres-

ion which is used for the classical LBP operator for a gray image if

e set g = b = 0 , and the R -component image is treated as repre-

enting a gray image where r o = r c and n 1 = 0. Mathematically, we

epresent the LBPC as follows.

BP C ( x c , y c ) =

P−1 ∑

p=0

S ( I p ) × 2

p , (32)

( I p ) =

{1 ,

0 ,

E ( I p ) ≥ 0 ,

E ( I p ) < 0 . (33)

The expression E ( I p ) is evaluated using Eq. (31) , where I p = p ( x, y ) , p = 0 , . . . , P − 1 , represents a color pixel in the neighbor-

ood of the pixel I c = I c ( x, y ). For the 8-neighborhood P = 8.

Fig. 2 displays color images taken from Corel-5k dataset [50] in

he first column, columns 2 to 4 display LBP images of the compo-

ent images of R, G, and B, column 5 displays the LBP images of

he intensity value I = ( R + G + B ) / 3 and the last column displays

he LBPC image. It is observed that the LBPC image not only pre-

erves the shape of an object, but it also provides a dense texture

attern as compared to the texture patterns obtained by the com-

onent images and the intensity image.

. Discussion on the parameters used by the LBPC operator

There are three parameters that determine the performance of

BPC: the normal to the plane n = ( n 1 , n 2 , n 3 ) , the reference point

o = ( r o , g o , b o ) through which the plane passes, and the size of

he circularly symmetric neighborhood window( P, R ). A common

hoice for the size of the neighborhood is a 3 × 3 window with

( P, R ) = ( 8 , 1 ) . There are other values of ( P, R ) such as ( P, R ) =( 8 , 2 ) , ( 12 , 1 . 5 ) , ( 16 , 3 ) , and so on, with the neighborhood win-

ow size ( 2 R + 1 ) × ( 2 R + 1 ) . We have shown in Section 3.1 the

onstruction and configuration of some of the neighborhoods. One

an also refer [25,31] for the construction of high order circularly

ymmetric neighborhoods.

.1. Normal to the plane

.1.1. Global natural normal

A most appropriate choice for the direction of the normal to

he plane is parallel to the gray line joining pure black pixel (0,0,0)

nd pure white pixel (1,1,1). This choice of the normal vector leads

o n 1 = n 2 = n 3 . This provides us an isometric plane whose orien-

ation with respect to R, G, B axis is the same. In this case, the

quation of the plane given by Eq. (30) reduces to

+ g + b − ( r o , g o , b o ) = 0 . (34)

The thresholding operation given by Eq. (30) compares the two

ntensity values I = (r + g + b)/3 and I o = ( r o + g o + b o ) / 3 , at I p and I o ,

espectively, which does not contain color information. Therefore,

he contribution of the color texture for the derivation of the LBPC

perators becomes ineffective.

.1.2. Global average normal

We find the global centroid of the color components as fol-

ows.

a =

1

MN

M−1 ∑

x =0

N−1 ∑

y =0

r(x, y ) , G a =

1

MN

M−1 ∑

x =0

N−1 ∑

y =0

g(x, y ) ,

a =

1

MN

M−1 ∑

x =0

N−1 ∑

y =0

b(x, y ) . (35)

The unit normal vector is set to n =R a √

R a 2 + G a 2 + B a 2

, G a √

R a 2 + G a 2 + B a 2

, B a √

R a 2 + G a 2 + B a 2

).

.1.3. Local average normal

The neutral normal n 1 = n 2 = n 3 and the global normal n =( R a , G a , B a ) do not provide the local characteristics of the spatial

orrelation of the color pixels. In order to extract the local features,

Page 7: Color texture description with novel local binary patterns

56 C. Singh et al. / Pattern Recognition 76 (2018) 50–68

Fig. 2. Original color images and their LBP and LBPC images: (a) original color image, (b) to (e) LBP of R,G,B, and I components, and (f) LBPC image. (For interpretation of

the references to color in this figure legend, the reader is referred to the web version of this article.)

4

c

m

L

l

r

r

R

n

4

we derive local normal to the plane. A local normal is obtained af-

ter averaging the color components in the local window of size

( 2 R + 1) × ( 2 R + 1 ) :

R l =

1

( 2 R + 1 ) 2

2 R ∑

x =0

2 R ∑

y =0

r ( x, y ) G l =

1

( 2 R + 1 ) 2

2 R ∑

x =0

2 R ∑

y =0

g ( x, y ) ,

B l =

1

( 2 R + 1 ) 2

2 R ∑

x =0

2 R ∑

y =0

b ( x, y ) . (36)

The unit normal vector to the plane is set to:

n =

(

R l √

R l 2 + G l

2 + B l 2 ,

G l √

R l 2 + G l

2 + B l 2 ,

B l √

R l 2 + G l

2 + B l 2

)

.

4.1.4. Center normal

Another important method to represent local characteristics is

to use the normal to the plane whose direction ratios are in

proportion to the color components of the center pixel. Let I c =( r c , g c , b c ) represent the center pixel. If n 1 = r c , n 2 = g c , and n 3 =b c , then the unit normal to the plane is defined as center normal.

The unit normal vector is given by

n =

(

r c √

2 2 2 ,

g c √

2 2 2 ,

b c √

2 2 2

)

.

r c + g c + b c r c + g c + b c r c + g c + b c f

.1.5. Mean normal

The minimum and the maximum of the components of the

olor vectors in the neighborhood window are obtained, and their

ean values are assigned to the components of the normal vector.

ike the local normal and the center normal, it also represents the

ocal characteristics of the color image.

min = min ( r p ) , g min = min ( g p ) , b min = min ( b p ) , 0 ≤ p ≤ P − 1 ,

(37)

max = max ( r p ) , g max = max ( g p ) , b max = max ( b p ) , 0 ≤ p ≤ P − 1 ,

(38)

m

= ( r min + r max ) / 2 , G m

= ( g min + g max ) / 2 , B m

= ( b min + b max ) / 2 ,

(39)

The unit mean normal vector is given by

=

(

R m √

R m

2 + G m

2 + B m

2 ,

G m √

R m

2 + G m

2 + B m

2 ,

B m √

R m

2 + G m

2 + B m

2

)

.

.2. Reference point

The color vector at the center of the window is the most pre-

erred choice for the reference point, i.e. ( r o , g o , b o ) = ( r c , g c , b c ) .

Page 8: Color texture description with novel local binary patterns

C. Singh et al. / Pattern Recognition 76 (2018) 50–68 57

O

o

g

a

w

u

p

5

h

o

h

i

c

t

h

c

R

H

w

θ

b

t

t

l

L

L

w

S

T

a

c

v

h

i

n

f

n

v

6

6

f

m

f

a

t

e

a

O

s

n

d

t

p

t

m

f

f

(

o

C

D

E

D

w

D

S

D

6

q

s

a

P

w

p

a

P

P

ther choices for the reference point are taken as the average

f pixel values in the local window ( r o , g o , b o ) = ( R l , G l , B l ) or the

lobal average ( r o , g o , b o ) = ( R a , G a , B a ) or a neutral color, such

s ( r o , g o , b o ) = ( 0 . 5 , 0 . 5 , 0 . 5 ) and the mean value ( r o , g o , b o ) =( R m

, G m

, B m

) . Among the various choices for the reference point,

e find that the center pixel ( r c , g c , b c ) and the average color val-

es of local window ( R l , G l , B l ) are the most effective reference

oints.

. Local binary pattern of the hue component (LBPH) and color

istogram (CH) and their fusion with LBPC features

The color is a powerful image descriptor. Many objects are rec-

gnized solely by their colors. The HSI color model provides the

ue (H) component, which represents the color texture of a color

mage. Therefore, to segment an image based on color, the hue

omponent is a natural choice. In addition to using RGB color space

o derive LBPC of a color image, we use color histogram (CH) of

ue component and derive the LBP of the hue image, which is

alled LBPH. The hue image is derived from the color image in the

GB color space using the following color conversion formula [47] .

( x, y ) =

{θ ( x, y ) i f b ( x, y ) ≤ g ( x, y ) ,

2 π − θ i f b ( x, y ) > g ( x, y ) , (40)

here

( x, y )

= co s −1

⎧ ⎨

1 2 [ ( r ( x, y ) − g ( x, y ) ) + ( r ( x, y ) − b ( x, y ) ) ] [

( r ( x, y ) −g ( x, y ) ) 2 + ( r ( x, y ) −b ( x, y ) ) ( g ( x, y ) −b ( x, y ) )

] 1 2

⎫ ⎬

.

(41)

The function H ( x, y ) is treated like a gray image whose local

inary patterns are obtained in a way akin to the local binary pat-

erns of gray image. These features are referred to as LBPH fea-

ures. The procedure for deriving LBPH features is explained as fol-

ows.

Let ( x c , y c ) be the center of a ( 2 R + 1 ) × ( 2 R + 1 ) window. The

BPH features are obtained as

BP H( x c , y c ) =

P−1 ∑

p=0

S( Q p ) × 2

p , (42)

here

( Q p ) =

{1

0

i f H( x p , y p ) ≥ H( x c , y c ) , otherwise.

(43)

Here ( x p , y p ) is the coordinate of a neighborhood pixel.

Color histograms (CH) are very effective image descriptors [48] .

herefore, in our proposed methods, we include CH features, which

re derived as follows.

Let n be the total number of color bins and h i denote the i th

olor bin for i = 0 , 1 , . . . .n − 1 , which is initialized to zero. The

alue of the i th bin is updated as:

i ← 0 , i = 0 , 1 , 2 , . . . , 2

P − 1 . (44)

= int ( nθ ( x, y ) / 2 π) , h i ← h i + 1 . (45)

f or x = 0 , 1 , . . . , M − 1 , y = 0 , 1 , . . . , N − 1 .

Here, it is assumed that θ ( x, y ) ∈ [0, 2 π ]. All histogram bins are

ormalized by dividing them by the size of the image before their

usion. The three feature vectors LBPC, LBPH, and CH, of size n c ,

h , and n , respectively, are fused together to form a single feature

ector of size n c + n h + n whose components are represented by

{ LBP C [ 0 ] , LBP C [ 1 ] , . . . , LBP C [ n c − 1 ] , LBP H [ 0 ] , LBP H [ 1 ] ,

. . . , LBP H [ n h − 1 ] , CH [ 0 ] , CH [ 1 ] , . . . , CH [ n − 1 ] } .

. Similarity measures and performance parameters

.1. Similarity measures

The performance of a retrieval system depends not only on ef-

ective features but also on strong similarity measures or distance

etrics. There are several similarity measures, which have been

ound to be very successful in image retrieval systems. Since we

re dealing with histogram-based feature vectors, our choice for

he similarity measures will be based on this aspect. Some of the

ffective similarity measures for histogram-based feature vectors

re Chi-square, Canberra, extended-Canberra, and square-chord.

ther commonly used distance measures such as histogram inter-

ection, L 1 -norm, L 2 -norm, cos-correlation, Jeffrey distance, etc. are

ot as effective as the former ones. In [51] , the extended-Canberra

istance has been proposed and compared with several other dis-

ance measures and it has been found to outperform them. In this

aper, we carry out performance comparison analysis with respect

o these similarity measures and find out an effective distance

easure which provides the overall best retrieval performance. The

our distance measures considered in this paper are explained as

ollows.

Let F q

i and F t

i represent the i th feature components of the query

probe) ‘ q ’ and database (gallery) ‘t’ images, respectively. The size

f the feature vector is L . The four distance measures are given by:

anberra distance:

CD ( q, t ) =

L −1 ∑

i =0

∣∣F q i − F t

i

∣∣F q

i + F t

i

. (46)

xtended-Canberra distance :

ECD ( q, t ) =

L −1 ∑

i =0

∣∣F q i − F t

i

∣∣(F q

i + μq

)+ (F t

i + μt )

, (47)

here μq =

1 L

∑ L −1 i =0 F

q i

and μt =

1 L

∑ L −1 i =0 F

t i

. χ2 - distance :

Chi ( q, t ) =

L −1 ∑

i =0

(F q

i − F t

i

)2

F q i

+ F t i

. (48)

quare-Chord distance :

SC ( q, t ) =

L −1 ∑

i =0

(√

F q i

−√

F t i

)2

. (49)

.2. Performance parameters

In our experiments, each image in the database is used as a

uery image. The performance of an image retrieval system is mea-

ured using precision P ( N ) and recall R ( N ) for retrieving top N im-

ges, which are defined in [51] .

( N ) =

I N N

; R ( N ) =

I N M

, (50)

here I N is the number of relevant images retrieved from top N

ositions and M is the total number of images in the dataset that

re similar to query image. The average precision of a single query

(q ) is the mean of all precision values P (n ) , n = 1 , 2 , . . . , N, i.e.

( q ) =

1

N

N ∑

n =1

P ( n ) . (51)

Page 9: Color texture description with novel local binary patterns

58 C. Singh et al. / Pattern Recognition 76 (2018) 50–68

r

c

c

M

o

a

v

1

a

p

t

o

2

7

w

c

o

o

d

r

d

i

a

t

r

e

i

c

1

c

m

1

s

a

e

7

w

p

m

W

C

o

c

t

a

p

m

n

s

U

T

t

t

b

l

The mean average precision ( mAP ) is the mean of the average

scores over all queries Q:

mAP =

1

Q

Q ∑

q =1

P ( q ) . (52)

The mAP measure contains both the precision and recall infor-

mation and represents the entire ranking [52] .

The P − R graph is not an appropriate measure when the num-

ber of relevant images in each class is variable. The bull’s eye per-

formance ( BEP ) is a better performance measure [53] which is de-

fined for a query image q as:

BEP ( q ) =

I q

M

, (53)

where M is the total number of relevant images in the database

corresponding to the query image q and I q represents the num-

ber of relevant images among the top 2 M retrievals. Clearly, I q ≤ M ,

even if the top 2 M images are searched in the database for the im-

ages relevant to the query image. The average BEP(mBEP) value for

all query images Q is defined as:

mBE P =

1

Q

Q ∑

q =1

BE P ( q ) . (54)

7. Experimental results

This section provides several experimental results to demon-

strate the effectiveness of the proposed methods and compare

their results with the closely related existing seven color texture

operators- LBP of component images, MSLBP, LCVBP, RGB-OC-LBP,

QLRBP, CLSP, and MDLBP. Out of the two MDLBP operators, the

decoder operator performs better than the adder operator [46] .

Hence, we consider the decoder operator for the performance com-

parison. The LBP of the component images is simply an exten-

sion of the LBP of images for component images R, G, and B of

a color image. We also choose Gabor filters because Gabor fil-

ters are among the most effective texture descriptors [5-9] which

have been applied for texture image retrieval for gray scale im-

ages. A design strategy adopted by Manjunath and Ma [5] and

Han and Ma [8] to derive Gabor filters provides better results over

its other implementations. The strategy developed by them en-

sures that the contours of half-peak magnitude support of the fil-

ter responses in the frequency domain touch each other while the

other strategies allow them to overlap. Thus, the Gaussian stan-

dard deviations in 2-dimensions are dependent on the other filter

parameters. These Gabor filter parameters are: maximum center

frequency U h = 0 . 38 , minimum center frequency U l = 0 . 05 , num-

ber of scales S = 4 , and number of orientations O = 4 . The size of

the filter is set to W = 11 . We conducted detailed experiments for

image retrieval and found that their design strategy and filter pa-

rameters yielding better results than the other implementations.

We apply Gabor filters using these parameters on the individual

components of color images and concatenate these features. For

deriving Gabor features, MATLAB functions have been used after

setting the Gabor parameters and modifying the code for the com-

putation of the Gaussian standard deviations in accordance with

works of Manjunath and Ma [5] and Han and Ma [8] . There are

16 Gabor filtered images for each component providing a total of

48 images. The mean and the standard deviation of the magnitude

of the filtered Gabor images have been derived to yield a feature

set of size 96 (16 images per component × 3 components × 2 fea-

tures per image). We have implemented our proposed and exist-

ing six methods in Visual C ++ 6.0 under Microsoft Windows en-

vironment on a PC with 2.50 GHZ CPU and 8GB main memory.

The methods LBP, MSLBP, LCVBP, RGB-OC-LBP, and MDLBP do not

equire any other parameter. The LBP operator is applied on each

omponent image, therefore, the size of LBP feature vector for a

olor image is 768. The size of MSLBP, LCVBP, RGB-OC-LBP, and

DLBP feature vector is 2304, 236, 96, and 2048, respectively. The

perator QLRBP requires three weight parameters α1, α2, and α3,

nd the value of the reference vector I ′ = ( r ′ , g ′ , b ′ ) . We set these

alues same as given by Lan et al. [44] . These values are: α1 = , α2 = 1 , and α3 = 1 . There are three reference vectors I ′ 1 =

( 0 . 9922 , 0 . 0857 , 0 . 090 ) , I ′ 2 = ( 0 . 0912 , 0 . 9908 , 0 . 0999 ) , and I ′ 3

=( 0 . 0852 , 0 . 0855 , 0 . 9927 ) . The three reference quaternions create

feature vector of size 768. The CLSP operator [45] requires three

arameters: size of visual words dictionary S which is set to 10 and

he nearest neighborhood parameters K 1 = 6 , and K 2 = 3 . The CLSP

perator creates a feature vector whose size is S × 256, which is

560.

.1. Datasets

Wang or SIMPLIcity [54] : It is a subset of Corel image database,

hich contains 10 0 0 color images, which are divided into 10

lasses of 100 images each. Each class contains images with res-

lution either 256 × 384 pixels or 384 × 256 pixels. The 10 classes

f Wang image database are: African people, beach, building, bus,

inosaur, elephant, flower, horse, glacier, and food.

Holidays [55] : It contains 500 image groups, and each of which

epresents a distinct scene in all 1491 personal holiday photos un-

ergoing various transformations such as rotation, viewpoint and

llumination changes, blurring, etc. The number of photos in an im-

ge group is variable. The dataset contains a large variety of scene

ypes such as natural, man-made, water and fire effects, etc. The

esolution of the images is very high (2448 × 3204), and for our

xperiments, we scale them to the size of 128 × 128 using bicubic

nterpolation of MATLAB library.

Corel- 5K [50] : This dataset contains 50 0 0 images and covers 50

ategories of images. Every category contains 100 images of size

92 × 128 or 128 × 192 pixels in JPEG format. The dataset Corel- 5K

ontains images including diverse contents such as tiger, mountain,

ushroom, fort, ocean, car, ticket, etc.

Corel- 10K [50] : This dataset contains 10,0 0 0 images and covers

00 categories of images. Every category contains 100 images of

ize 192 × 128 or 128 × 192 pixels in JPEG format. It consists of im-

ges such as cat, rose, sunset, duck, train, musical instrument, fish,

agle, judo-karate, etc.

.2. Selection of parameters for LBPC

There are three parameters to be fixed for the LBPC operator:

indow size, normal to the plane and reference point. To find the

roper window size for LBPC, we present the results of experi-

ents for retrieval performance using different window sizes on

ang and Holidays datasets. Wang dataset is a representative of

orel- 5K and Corel- 10K datasets. Holidays contains variable number

f images in the classes. Thus, the presentation of the results will

over the complete spectrum of our experimental setup. To reduce

he size of the presented data, we take the normal to the plane n

s the “local average normal” and the intensity values of the center

ixels as the reference point R O of the plane. The retrieval perfor-

ance for these two datasets are shown in Tables 1 , and 2 for a

umber of combinations of ( P, R ). For P = 8 and 12, we have pre-

ented results for LBPC and ULBPC. For P = 16 , only the values of

LBPC are shown to avoid large feature size of LBPC which is 2 16 .

he criteria for the selection of the best ( P, R ) values include re-

rieval performance and retrieval speed. After comparing the re-

rieval performance, the combination (8, 2) provides the overall

est results. The performance of the combination (8, 1) is much

ess than (8, 2). Both these combinations have the same retrieval

Page 10: Color texture description with novel local binary patterns

C. Singh et al. / Pattern Recognition 76 (2018) 50–68 59

Table 1

Mean average precision (mAP) in percent for top 100 images, N = 100 , retrieved by LBPC and ULBPC on Wang dataset

for various neighborhood pixels (P), and radius (R).

(P, R) Method No. of features Distance measures Average

Chi-square Canberra Extended-Canberra Square-chord

(8,1) LBPC 256 54.31 53.85 57.45 54.33 54.98

ULBPC 59 52.61 55.77 55.36 52.62 54.09

(8,2) LBPC 256 55.49 56.27 58.05 55.50 56.32

ULBPC 59 53.28 55.68 55.43 53.29 54.42

(8,3) LBPC 256 54.22 54.60 56.19 54.21 54.66

ULBPC 59 52.20 53.85 54.10 52.18 53.08

(12,1.5) LBPC 4096 55.36 56.99 57.43 55.32 56.27

ULBPC 135 55.21 56.93 57.37 55.24 56.19

(12, 2) LBPC 4096 55.02 54.37 57.97 55.07 55.60

ULBPC 135 54.08 55.81 56.19 54.13 55.05

(16,2) ULBPC 243 55.15 56.56 57.21 55.22 56.03

(16,3) ULBPC 243 55.18 56.58 57.24 55.25 56.06

Table 2

Average bull’s eye performance (mBEP) in percent obtained by LBPC and ULBPC on Holidays dataset for various neigh-

borhood pixels (P), and radius (R).

(P, R) Method No. of features Distance measures

Chi-square Canberra Extended-Canberra Square-chord Average

(8,1) LBPC 256 58.60 58.459 60.48 58.52 59.01

ULBPC 59 57.07 58.33 57.87 57.00 57.57

(8,2) LBPC 256 58.29 59.46 60.60 58.11 59.11

ULBPC 59 56.30 57.15 57.35 56.34 56.79

(8,3) LBPC 256 56.54 57.84 58.52 56.25 57.29

ULBPC 59 54.72 55.75 55.92 54.53 55.23

(12,1.5) LBPC 4096 58.20 59.75 59.96 58.34 59.06

ULBPC 135 58.11 59.71 59.57 58.10 58.87

(12,2) LBPC 4096 57.43 59.32 58.11 58.23 58.27

ULBPC 135 57.37 58.66 58.67 57.26 57.99

(16,2) ULBPC 243 58.02 59.34 59.81 57.73 58.72

(16,3) ULBPC 243 57.70 57.02 57.98 57.38 57.52

t

i

p

t

l

u

t

s

a

p

m

i

n

c

t

t

t

i

p

7

t

R

c

m

t

c

t

a

a

f

a

L

a

t

t

i

i

f

L

L

v

o

w

t

s

M

L

a

h

T

s

r

t

ime because out of the 8 neighbors, 4 neighbors are interpolated

n both the cases. Also, the feature size is the same. Although, the

erformance of some of the ULBPCs of P = 12 , and 16 is higher

han the ULBPC of P = 8 , the size of their feature vectors is much

arger than P = 8 . Therefore, in our all foregoing experiments we

se P = 8 , and R = 2 .

The LBPC operator derives local texture for color images after

hresholding the color pixels using a plane in the local window of

ize 3 × 3. A plane requires two parameters: normal to the plane

nd a reference point. The performance of the LBPC operator de-

ends on the optimal choices of these parameters.

As discussed in Section 4 , we have several choices for the nor-

al to the plane and the reference point. The normal to the plane

ncludes the vector representing gray line using normal vector n =( 1 , 1 , 1 ) , the global average normal n = ( R a , G a , B a ) , local average

ormal n = ( R l , G l , B l ), mean normal n = ( R m

, G m

, B m

), and the

enter normal n = ( R c , G c , B c ) .

Among the various choices for the reference point R o , the cen-

er pixel R o = ( r c , g c , b c ) , is intuitively the best choice, because

he plane should pass through the center pixel about which the

hresholding operation is being performed. However, during exper-

mental analysis we observed that the local average normal, R o =( R l , G l , B l ) , taken as a reference point, also provides competitive

erformance.

.3. Comparison with existing methods

The performance of the proposed methods is compared with

he nine approaches: Gabor filtering, LBP, ULBP, MSLBP, LCVBP,

GB-OC-LBP, QLRBP, CLSP, and MDLBP using the mean average pre-

ision ( mAP ). The number of features used by all methods is also

entioned in the results. In order to analyze the effectiveness of

he “uniform” local binary patterns, it is also applied to the three

omponent images yielding the method ULBP. The uniform pat-

erns are also extended to LBPC and LBPH, resulting in the oper-

tors ULBPC and ULBPH, respectively. The proposed methods are

pplied separately as well in four combinations by fusing their

eatures. The individual methods are LBPC, LBPH, ULBPC, ULBPH,

nd CH and four combinations are LBPC + LBPH, ULBPC + ULBPH,

BPC + LBPH + CH, and ULBPC + ULBPH + CH. These nine methods are

nalyzed to compare their relative retrieval performance in order

o explore the best methods providing high accuracy and low re-

rieval time. It is important to note here that the LCVBP operator

s based on the “uniform” patterns, which provides very compet-

tive results in comparison to the full LCVBP operator using 256

eatures for each of the four combinations. The dimension of the

CVBP operator is 236 as compared to the dimension of the full

CVBP operator, whose size is 1024. Therefore, there is an ad-

antage of speed by using low dimensional feature vector with-

ut much loss of retrieval accuracy. A similar trend is observed

hile using LBPC and LPBH operators, which will be discussed in

he following experimental analysis. Therefore, we compare the re-

ults of 18 approaches. These approaches are: Gabor, LBP, ULBP,

SLBP, LCVBP, RGB-OC-LBP, QLRBP, CLSP, MDLBP, LBPC, ULBPC,

BPH, ULBPH, CH, LBPC + LBPH, ULBPC + ULBPH, LBPC + LBPH + CH,

nd ULBPC + ULBPH + CH.

The mAP values obtained by the various methods for top one

undred images ( N = 100 ) on the Wang dataset are presented in

able 3 . For convenience, we mark the top six methods with bold

tyle and also number them from 1 to 6 in the order of their

etrieval performance. We chose six methods because in most of

he cases our 4 proposed methods turn out to be on the top

Page 11: Color texture description with novel local binary patterns

60 C. Singh et al. / Pattern Recognition 76 (2018) 50–68

Table 3

Mean average precision ( mAP ) in percent for N = 100 obtained by various approaches in RGB color space on Wang dataset using

various distance measures.

Method No. of features Chi-square Canberra Extended-Canberra Square chord

Existing LBP [12] 3 × 256 = 768 55.28 53.42 56.93 55.33

ULBP [12] 3 × 59 = 177 53.34 56.63 54.19 53.37

MSLBP [39] 9 × 256 = 2304 59.86 56.14 60.62 (5) 59.86

LCVBP [42] 4 × 59 = 236 53.42 57.19 56.83 53.44

RGB-OC-LBP [43] 3 × 32 = 96 47.90 51.18 49.39 47.93

QLRBP [44] 3 × 256 = 768 53.47 52.79 56.03 53.50

CLSP [45] 10 × 256 = 2560 48.37 43.25 45.84 48.05

GABOR [8] 96 58.86 59.54 59.53 58.91

MDLBP [46] 8 × 256 = 2048 59.58 58.15 60.82 (4) 59.58

Proposed LBPC 256 55.49 56.27 58.05 55.05

ULBPC 59 53.28 55.68 55.43 53.29

LBPH 256 49.24 47.57 50.72 49.31

ULBPH 59 47.30 47.77 48.75 47.33

CH 30 47.49 41.84 48.37 47.73

LBPC + LBPH 2 × 256 = 512 58.80 59.09 61.92 (3) 58.82

ULBPC + ULBPH 2 × 59 = 118 56.90 59.16 59.61 (6) 56.89

LBPC + LBPH + CH 2 × 256 + 30 = 542 57.06 61.12 65.16 (1) 56.12

ULBPC + ULBPH + CH 2 × 59 + 30 = 148 56.33 60.97 63.59 (2) 55.46

Average mAP – 53.99 54.32 56.21 53.88

a

C

(

i

i

v

U

t

A

m

a

t

0

L

v

t

and two more methods have been included from the existing

methods. All database images are used as query images. It is ob-

served from the table that the proposed approach LBPC + LBPH + CH

achieves the highest mAP value of 65.16% using the extended-

Canberra distance, followed by the same approach with uniform

patterns, ULBPC + ULBPH + CH, with mAP value of 63.59%. There is

not much difference between the two proposed approaches while

the latter uses only 148 features as compared to 542 features used

by the former approach. The third best mAP value is achieved by

LBPC + LBPH which is 61.92%. The next best mAP value is achieved

by MDLBP which is 60.82% using extended-Canberra. The fifth

best value of mAP is achieved by MSLBP method yielding mAP

value 60.62%. Among the nine state-of-the-art-methods LBP, ULBP,

MSLBP, LCVBP, RGB-OC-LBP, QLRBP, CLSP, GABOR, MDLBP, the per-

formance of MDLBP is the best which provides mAP of 60.82%, fol-

lowed by MSLBP, GABOR, LCVBP, LBP, ULBP, QLRBP, RGB-OC-LBP,

30

40

50

60

70

80

10 20 30 40 50 60 70Recall

Pre

cisi

on

(P) (%

)

Fig. 3. P-R curves for the seven existing methods MSLBP, GABOR, LCVBP, LBP, ULBP, QLRBP

for N = 1 to 100 on Wang dataset.

nd lastly CLSP, yielding mAP values of 60.62% (using extended-

anberra), 59.54% (using Canberra), 57.19% (using Canberra), 56.93%

using extended-Canberra), 56.63% (using Canberra), 56.03% (us-

ng extended-Canberra), 51.18% (using Canberra), and 48.37% (us-

ng chi-square), respectively. The proposed operator LBPC pro-

ides mAP value of 58.05%, which is much more than the LBP,

LBP, LCVBP, RGB-OC-LBP,QLRBP, and CLSP but slightly less than

he other three approaches such as MDLBP, MSLBP, and Gabor.

lthough MDLBP achieves highest mAP among the nine existing

ethods, it has slower retrieval speed (refer Section 7.4 for time

nalysis). When the feature dimension of LBPC is reduced from 256

o 59 by using ULBPC method, the reduction in mAP value is only

.67% while the feature dimension is reduced to 59 from 256. The

BPH and CH provide complementary features, leading to increased

alues of mAP when fused with LBPC or ULBPC. The performance of

he method LBPC + LBPH which yields mAP of 61.92% is much bet-

80 90 100

LBPC+LBPH+CH (Proposed)ULBPC+ULBPH+CH (Proposed)MSLBPGABORLCVBPLBPULBPQLRBPMDLBP

, and MDLBP, and two proposed methods, LBPC + LBPH + CH, and ULBPC + ULBPH + CH

Page 12: Color texture description with novel local binary patterns

C. Singh et al. / Pattern Recognition 76 (2018) 50–68 61

Table 4

Class-wise mean average precision ( mAP ) in percent for N = 100 obtained by

LBPC + LBPH + CH and MDLBP [46] in RGB color space on Wang dataset using

extended-Canberra distance measure.

Class LBPC + LBPH + CH (Proposed) MDLBP [46] Difference

African people 61.52 55.75 5.77

Beach 43.36 45.21 −1.85

Building 54.57 56.81 −2.24

Bus 87.90 80.79 7.11

Dinosaur 98.88 97.10 1.78

Elephant 44.54 41.53 3.01

Flower 83.08 80.08 3.00

Horse 81.59 61.61 19.98

Glacier 39.89 33.8 6.09

Food 56.31 55.54 0.77

Average 65.16 60.82 4.34

t

m

f

f

m

u

b

b

U

p

t

a

m

C

t

e

t

C

5

v

t

s

n

a

a

G

t

r

G

s

a

a

o

p

c

m

n

i

i

n

o

n

a

m

A

v

d

p

o

i

t

p

t

t

M

r

t

M

t

M

a

f

c

t

o

w

p

a

i

p

U

v

C

w

m

t

er than the performance of their individual operators which yield

AP of 58.05% and 50.72%, respectively. When CH features are

used with LBPC + LBPH to obtain LBPC + LBPH + CH method, the per-

ormance of the resultant method is significantly increased to yield

AP of 65.16% from 61.92%. The method ULBPC + ULBPH + CH which

ses uniform patterns for LBPC and LBPH, yields mAP of 63.59%

y using only 148 features. Therefore, when a tradeoff is required

etween retrieval accuracy and retrieval speed, then the method

LBPC + ULBPH + CH is the first choice among all existing and pro-

osed approaches. Thus, the proposed method LBPC + LBPH + CH is

he best local descriptor for color texture images, followed by

nother proposed method ULBPC + ULBPH + CH that uses low di-

ension feature vector. Lastly, the performance of the extended-

anberra distance measure is overall the best which is reflected in

he average mAP values shown in the last row of Table 3 . The av-

rage mAP values are computed over all methods for a given dis-

ance measure. The average mAP obtained by extended-Canberra,

anberra, chi-square, and square-chord, are 56.21%, 54.32%, 53.99%,

3.88%, respectively. It is seen in the table that out of 18 best mAP

alues, the extended-Canberra provides the best results for thir-

een methods, while the Canberra and chi-square distance mea-

ures provide best results for four and one method, respectively.

The precision versus recall values is plotted in Fig. 3 for

ine methods (2 proposed, 7 existing) which provide the over-

ll best results. The two proposed methods are LBPC + LBPH + CH,

nd ULBPC + ULBPH + CH and the seven existing methods are MSLBP,

ABOR, LCVBP, LBP, ULBP, QLRBP, and MDLBP. It is observed from

he graph that the proposed methods consistently provide the best

etrieval rates for all values of recall followed by MDLBP, MSLBP,

ABOR, LCVBP, LBP, ULBP, and QLRBP. The trend of precision is the

ame for all values of recall N = 1 to 100 .

Table 5

Number of relevant images retrieved by MDLBP (Method A) and proposed method L

N Class

African people Beach Building Bus Dinosaur

A B A B A B A B A B

1 100 100 100 100 100 100 100 100 100 100

2 85 89 62 61 83 81 95 100 98 99

3 82 77 56 57 81 81 97 98 99 100

4 75 75 58 53 72 80 96 99 99 99

5 71 79 59 54 77 71 98 99 98 99

6 67 75 57 50 76 71 94 97 99 100

7 70 75 47 47 72 73 95 98 98 100

8 71 74 45 43 67 66 94 97 98 100

9 71 68 56 39 71 67 96 95 98 100

10 68 71 49 49 69 65 93 96 98 100

11 64 75 54 49 64 66 90 98 100 99

12 65 67 47 51 64 66 89 97 99 100

To analyze the comparative performance of the best methods

mong the proposed (LBPC + LBPH + CH) and the existing (MDLBP)

pproaches, we present the mAP values for each of the ten classes

f the Wang dataset in Table 4 for N = 100 . The performance of the

roposed method LBPC + LBPH + CH is significantly higher for the

lass Horse which is 81.59% as compared to MDLBP which yields an

AP of 61.61%, a value lower by 19.98%. The difference is also sig-

ificant for four other classes Bus, Glacier , and African people , which

s 7.11%, 6.09%, and 5.77%, respectively. We analyzed the objects

n this class and observed that many variations exist such as the

umber of horses which varies from one to six, view angle, texture

f the grass (background) and color of the enclosure (barrier). The

umbers of images having one, two, three, four, five and six horses

re 10, 73, 9, 3, 4, and 1. The other classes for which our proposed

ethod provides significantly better results are Bus, Glacier , and

frican people . Like the class Horse , these classes have a number of

ariations. Other classes do not have many variations in images. A

eeper analysis of the number of the retrieved relevant images was

erformed to compare the performance for each value of N , and

ut of 100 values of N , the results for N = 1 to 12 are presented

n Table 5 . The table shows the number of the relevant images re-

rieved by MDLBP (Method A) and LBPC + LBPH + CH (Method B) at

ositions N = 1 , 2 , . . . , 12 , for each of the classes. It is shown in the

able that for the class Horse the number of relevant images ob-

ained by Method B is significantly higher than those obtained by

ethod A, except for N = 7 , for which Method A provides 3 more

elevant images than Method B. We can observe similar trends for

he other three classes. Bus, Glacier , and African people for which

ethod B provides significantly better retrieval performance. On

he other hand, the class Building shows slightly better results for

ethod A. For other classes there is not very significant difference,

lthough Method B performs slightly better than Method A. Thus,

or out of ten classes, Method B provides better results for eight

lasses, while Method A provides slightly better results for only

wo classes. Our proposed method shows significantly better result

n one of the eight classes, Horse , which contains lots of variations

ithin the objects in the images of this class. The proposed ap-

roach is thus more robust to intra-class variations in the images

s compared to the existing methods.

The retrieval results on the Corel- 5K dataset are presented

n Table 6 . The maximum value of mAP is achieved by the

roposed method LBPC + LBPH + CH which is 43.81%, followed by

LBPC + ULBPH + CH yielding mAP of 41.68%. The next largest mAP

alue is achieved by MDLBP which is 39.99% using extended-

anberra. The fourth largest value of mAP is obtained by MSLBP

hich is 39.95%. Another proposed method LBPC + LBPH achieves

AP of 38.34%. The performance of other existing methods in

he decreasing mAP values is: LCVBP (38.14%), GABOR (37.18%),

BPC + LBPH + CH (Method B) for each value of N( N = 1 to 12 ) on Wang dataset.

Elephant Flower Horse Glacier Food

A B A B A B A B A B

100 100 100 100 100 100 100 100 100 100

78 85 98 94 94 99 53 62 80 81

80 80 95 95 93 97 36 53 80 72

62 75 92 95 92 95 37 50 70 75

60 67 92 93 90 94 35 49 75 78

51 50 92 91 89 91 33 54 68 68

53 57 92 92 95 92 35 44 61 73

49 49 92 91 87 94 33 42 68 70

47 51 91 92 83 94 34 48 65 60

36 47 86 92 82 87 32 44 67 65

36 42 88 94 82 90 35 43 63 64

44 47 93 88 74 88 33 46 60 64

Page 13: Color texture description with novel local binary patterns

62 C. Singh et al. / Pattern Recognition 76 (2018) 50–68

Table 6

Mean average precision ( mAP ) in percent for N = 100 obtained by various approaches in RGB color space on Corel-5k dataset

using various distance measures.

Method No. of features Chi-square Canberra Extended-Canberra Square chord

Existing LBP [12] 3 × 256 = 768 35.38 32.57 35.75 35.41

ULBP [12] 3 × 59 = 177 32.87 34.27 34.26 32.88

MSLBP [39] 9 × 256 = 2304 35.53 33.26 39.95 (4) 35.33

LCVBP [42] 4 × 59 = 236 36.39 38.14 (6) 37.95 36.40

RGB-OC-LBP [43] 3 × 32 = 96 27.74 28.69 28.12 27.74

QLRBP [44] 3 × 256 = 768 35.21 31.72 36.54 35.22

CLSP [45] 10 × 256 = 2560 29.45 28.96 29.67 30.77

GABOR [8] 96 37.18 36.40 36.95 37.12

MDLBP 8 × 256 = 2048 37.89 34.93 39.99 (3) 38.08

Proposed LBPC 256 31.87 31.41 34.08 32.81

ULBPC 59 27.79 28.08 28.28 27.78

LBPH 256 26.95 24.07 28.23 27.02

ULBPH 59 24.43 24.95 26.16 24.43

CH 30 25.20 23.24 25.91 25.47

LBPC + LBPH 2 × 256 = 512 35.88 35.15 38.34 (5) 35.93

ULBPC + ULBPH 2 × 59 = 118 33.50 34.57 37.60 32.51

LBPC + LBPH + CH 2 × 256 + 30 = 542 34.76 40.36 43.81 (1) 34.30

ULBPC + ULBPH + CH 2 × 59 + 30 = 148 33.49 39.21 41.68 (2) 33.08

Average mAP – 32.31 32.22 34.63 32.35

Fig. 4. P-R curves for the seven existing methods MSLBP, LCVBP, GABOR, QLRBP, LBP, ULBP, and MDLBP and two proposed methods, LBPC + LBPH + CH, and ULBPC + ULBPH + CH

for N = 1 to 100 on Corel- 5K dataset.

s

f

3

r

F

i

t

d

e

s

i

e

t

t

o

t

i

QLRBP (36.54%), LBP (35.75%), ULBP (34.27%), CLSP (30.77%),

and RGB-OC-LBP (28.69%). Interestingly, the difference between

mAP values obtained by the proposed methods LBPC + LBPH + CH

and ULBPC + ULBPH + CH is only 2.13% which is insignificant,

whereas the difference in the size of feature vectors used by

these two proposed approaches is very large. The former method

uses 542 features as compared to less than one-third i.e. 148

features used by the latter. When we look for a good trade-

off between the retrieval accuracy and speed among the exist-

ing methods, then LCVBP is a good choice among the existing

methods, which has 236 features and it achieves mAP value of

38.14%. Among the proposed methods, ULBPC + ULBPH + CH achieves

mAP values of 41.68% with only 148 features. Thus, the proposed

method ULBPC + ULBPH + CH provides mAP values with only 148

features which is larger by 3.54% as compared to LCVBP which

uses 236 features. Out of the 18 best mAP values across the four

distance measures, extended-Canberra provides 13 best values, fol-

lowed by Canberra with 3 best values, chi-square with 1 value and

quare-chord with 1 value. The average mAP values of all methods

or extended-Canberra, square-chord, chi-square, and Canberra, are

4.63%, 32.35%, 32.31%, and 32.22%, respectively. The precision and

ecall values for the nine prominent methods have been plotted in

ig. 4 . It is seen that the proposed methods outperform the exist-

ng methods for all values of recall.

The retrieval results for Corel- 10K are presented in Table 7 . The

rend in the values of mAP is similar to that for Wang and Corel- 5K

atasets. The proposed method LBPC + LBPH + CH obtains the high-

st value of mAP, which is 36.99%, and its “uniform pattern” ver-

ion ULBPC + ULBPH + CH achieves the next highest value 34.32%. It

s important to note here that while there is a marginal differ-

nce of 2.67% in the mAP values of these two proposed method,

he feature size of the latter (148) is very small as compared

o the feature size of the former (542). The third largest value

f mAP is achieved by MDLBP, which is 33.97% with 2048 fea-

ures. The fourth largest value is achieved by LBPC + LBPH, which

s 33.50% by using 512 features. Among the existing methods,

Page 14: Color texture description with novel local binary patterns

C. Singh et al. / Pattern Recognition 76 (2018) 50–68 63

Table 7

Mean average precision ( mA P) in percent for N = 100 obtained by various approaches in RGB color space on Corel- 10K dataset

using various distance measures.

Method No. of features Chi-square Canberra Extended-Canberra Square-chord

Existing LBP [12] 3 × 256 = 768 28.97 26.71 29.33 28.98

ULBP [12] 3 × 59 = 177 26.96 28.03 28.08 26.97

MSLBP [39] 9 × 256 = 2304 28.99 27.08 31.69 (5) 28.86

LCVBP [42] 4 × 59 = 236 29.25 29.62 29.72 29.26

RGB-OC-LBP [43] 3 × 32 = 96 23.10 23.74 23.33 23.11

QLRBP [44] 3 × 256 = 768 26.40 23.99 27.64 26.39

CLSP [45] 10 × 256 = 2560 24.64 23.71 22.53 24.64

GABOR [8] 96 28.06 27.39 29.68 28.065

MDLBP [46] 8 × 256 = 2048 31.83 29.43 33.97 (3) 33.23

Proposed LBPC 256 24.45 24.37 27.25 25.89

ULBPC 59 21.27 22.40 22.50 21.38

LBPH 256 20.94 19.46 21.98 21.27

ULBPH 59 19.37 19.97 20.13 19.09

CH 30 18.63 17.79 19.18 18.89

LBPC + LBPH 2 × 256 = 512 31.30 32.20 33.50 (4) 31.37

ULBPC + ULBPH 2 × 59 = 118 28.98 29.40 31.02 (6) 27.88

LBPC + LBPH + CH 2 × 256 + 30 = 542 31.79 32.37 36.99 (1) 34.08

ULBPC + ULBPH + CH 2 × 59 + 30 = 148 28.09 32.04 34.32 (2) 26.17

Average BEP – 26.28 26.09 27.94 26.42

10

20

30

40

50

10 20 30 40 50 60 70 80 90 100

LBPC+LBPH+CH (Proposed)

ULBPC+ULBPH+CH(Proposed)

MSLBP

LCVBP

GABOR

LBP

ULBP

QLRBP

MDLBP

Recall

Pre

cisi

on

(P)

Fig. 5. P-R curves for the seven existing methods MSLBP, LCVBP, GABOR, LBP, ULBP, QLRBP, and MDLBP, and two proposed methods, LBPC + LBPH + CH, and ULBPC + ULBPH + CH

for N = 1 to 100 on Corel- 10K dataset.

M

L

R

p

g

w

L

C

d

m

p

b

r

N

s

t

s

m

p

T

i

m

i

p

b

L

U

(

C

t

M

U

l

o

M

t

i

a

a

6

SLBP achieves the second highest mAP value 31.69%, followed by

CVBP (29.72%), GABOR (29.68%), LBP(29.33%), ULBP(28.09%), QL-

BP (27.6 4%), CLSP(24.6 4%), and RGB-OC-LBP (23.74%). The pro-

osed method LBPC + LBPH yields mAP value 33.50% which is

reater than the value of MSLBP. The other proposed methods

ith their decreasing mAP values are: ULBPC + ULBPH (31.02%),

BPC (27.25%), ULBPC (22.50%), LBPH(21.98%), ULBPH (20.13%) and

H(19.18%). Once again the performance of the extended-Canberra

istance measure is much better than the other three distance

easures, which provides an average mAP of 27.94% as com-

ared to the average mAP of 26.42%, 26.28%, and 26.09% provided

y the square-chord, Chi-square, and Canberra distance measures,

espectively. Fig. 5 shows the precision versus recall values for

= 1 to 100 . The proposed methods yield higher values of preci-

ion for all values of recall.

The trend of retrieval performance of all methods with respect

o the size of the three ( Wang, Corel- 5K, Corel- 10K) datasets is con-

istent. As the size of the dataset increases, there is a drop in the

AP values for all methods. The relative drop is almost the same.

The results of the retrieval performance of the existing and

roposed methods on Holidays dataset are presented in Table 8 .

his dataset presents a variety of objects. The number of relevant

mages in a class is variable. Therefore, the retrieval accuracy is

easured in terms of average value of BEP ( mBEP ). It is shown

n the table that the highest value of mBEP is obtained by the

roposed method LBPC + LBPH + CH which is 63.25%, followed

y ULBPC + ULBPC + CH(62.97%), MDLBP (62.46%), MSLBP(61.92%),

BPC + LBPH (61.75%), LBP(61.63%), LBPC(60.48%), LCVBP (60.17%),

LBP(59.19%), ULBPC + ULBPH(58.66%), CLSP(57.86%), ULBPC

57.35%), GABOR (55.34%), QLRBP (55.21%), RGB-OC-LBP(54.15%),

H(51.90%), LBPH(44.38%), and lastly by ULBPH(42.67%). Among

he existing approaches the best value of mBEP is achieved by

DLBP (62.46%), MSLBP(61.92%), LBP(61.63%), LCVBP(60.17%),

LBP(59.19%), CLSP(57.86%), GABOR (55.34%), QLRBP(55.21%), and

astly by RGB-OC-LBP(54.15%). When we compare the performance

f the proposed LBPC method with the existing ten methods, then

DLBP outperforms LBPC marginally. However, it has 2048 fea-

ures as compared to only 256 features used by LBPC. Thus, when

t comes to having a good tradeoff between retrieval accuracy

nd speed, then only three methods ULBPC + ULBPH + CH, LBPC,

nd LCVBP, stand out as the winners with mBEP values of 62.97%,

0.48%, and 60.17%, respectively. The numbers of features used

Page 15: Color texture description with novel local binary patterns

64 C. Singh et al. / Pattern Recognition 76 (2018) 50–68

Table 8

Average bull’s eye performance ( mBEP ) in percent for obtained by various approaches in RGB color space on Holidays dataset

using various distance measures.

Method No. of features Chi-square Canberra Extended-Canberra Square chord

Existing LBP [12] 3 × 256 = 768 59.63 58.95 61.63 (6) 59.52

ULBP [12] 3 × 59 = 177 57.85 59.12 59.19 57.80

MSLBP [39] 9 × 256 = 2304 59.80 59.62 61.92 (4) 59.79

LCVBP [42] 4 × 59 = 236 59.13 59.67 60.17 54.04

RGB-OC-LBP [43] 3 × 32 = 96 53.87 54.15 54.09 53.87

QLRBP [44] 3 × 256 = 768 55.21 53.27 54.80 55.11

CLSP [45] 10 × 256 = 2560 57.38 57.10 57.86 57.42

GABOR [8] 96 55.28 54.78 54.74 55.34

MDLBP [46] 8 × 256 = 2048 60.81 57.99 62.46 (3) 60.86

Proposed LBPC 256 58.29 59.46 60.48 58.11

ULBPC 59 56.30 57.15 57.35 56.34

LBPH 256 44.32 42.64 44.23 44.38

ULBPH 59 42.43 41.30 42.67 42.39

CH 30 51.14 51.11 51.90 51.17

LBPC + LBPH 2 × 256 = 512 59.48 59.51 61.75 (5) 59.46

ULBPC + ULBPH 2 × 59 = 118 57.25 57.15 58.66 57.16

LBPC + LBPH + CH 2 × 256 + 30 = 542 61.32 60.66 63.25 (1) 61.24

ULBPC + ULBPH + CH 2 × 59 + 30 = 148 60.36 60.06 62.97 (2) 60.17

Average mBEP – 56.10 55.76 57.23 55.79

Table 9

Time taken (seconds) by various methods for feature extraction and image retrieval for top 100 images ( N = 100) on Corel- 5K and Corel- 10K

datasets.

Sr. no. Methods Number of features Feature extraction time (s) Retrieval time(s) Total time(s)

Corel-5K Corel-10K Corel-5K Corel-10K

a) LBP [12] 768 0.099 0.668 1.885 0.767 1.984

b) ULBP [12] 177 0.046 0.293 0.796 0.339 0.842

c) MSLBP [39] 2304 0.731 1.757 4.074 2.488 4.805

d) LCVBP [42] 236 0.051 0.312 0.967 0.363 1.018

e) RGB-OC-LBP [43] 96 0.027 0.203 0.639 0.23 0.666

f) QLRBP [44] 768 0.092 0.668 1.885 0.76 1.977

g) CLSP [45] 2560 0.529 2.273 5.092 2.802 5.621

h) GABOR [8] 96 0.765 0.203 0.639 0.968 1.404

i) MDLBP [46] 2048 0.212 1.556 2.126 1.768 2.338

j) LBPC 256 0.038 0.330 0.977 0.368 1.015

k) ULBPC 59 0.015 0.170 0.593 0.185 0.608

l) LBPH 256 0.040 0.330 0.977 0.37 1.017

m) ULBPH 59 0.016 0.170 0.593 0.186 0.609

n) CH 30 0.014 0.154 0.550 0.168 0.564

o) LBPC + LBPH 512 0.078 0.540 1.427 0.618 1.505

p) ULBPC + ULBPH 118 0.031 0.254 0.702 0.285 0.733

q) LBPC + LBPH + CH 542 0.092 0.596 1.663 0.688 1.755

r) ULBPC + ULBPH + CH 148 0.045 0.283 0.843 0.328 0.888

p

t

v

f

0

T

i

t

i

m

o

r

d

M

C

s

p

o

U

i

1

t

by these methods are 148, 256 and 236, respectively. Among the

various distance measures, the performance of extended-Canberra

is the best with an average mBEP of 57.23%, followed by chi-square

(56.10%), square-chord (55.79%), and Canberra (55.76%). When we

derive the best mBEP value for a method with respect to different

measures, then extended-Canberra distance provides the best

results for all eighteen methods.

7.4. Time analysis

The total time taken by a method for the complete retrieval

process consists of three parts: feature extraction, distance com-

putation D ( q, t ) for a query image q and a database image t , and

sorting of the top N images from a database of M images. The

computation of D ( q, t ) depends on the size L of the feature vec-

tor. The order of time complexity of this step is O ( L ). We use a

simple sorting algorithm (bubble sort) with time complexity of

O ( NM ) to retrieve N top images from M database images. This

remains the same for all the methods but varies amongst the

two datasets as they have different number of database images.

We club the times for distance computation and sorting and call

it retrieval time. For the purpose of the qualitative analysis, we

resent the time for feature extraction, retrieval time and the to-

al time for all 18 methods in Table 9 . The size of the feature

ector is also given. It is observed that CH takes the least time

or feature extraction and image retrieval, which is 0.014 (s) and

.154 (s) for Corel- 5K, and 0.014 (s) and 0.550 (s) for Corel- 10K.

he time taken for feature extraction is the same for all images

n these databases because their size is the same. For the quan-

itative analysis, we present the time analysis results in graph-

cal form in Fig. 6 showing the number of features for all 18

ethods and their feature extraction and the total time (in sec-

nds) for Corel- 5K and Corel- 10K datasets. The methods are ar-

anged in the increasing order of the total time taken for Corel- 10K

ataset. The longest time is taken by CLSP followed by MSLBP, and

DLBP which is 5.621(s), 4.805(s), and 2.338(s) respectively, for

orel- 10K. All these three methods have very high feature dimen-

ions which are 2560, 2304, and 2048, respectively. All our pro-

osed methods are much lower in dimensions (i.e., approximately

ne-fourth for LBPC + LBPH + CH/LBPC + LBPH, and one-thirteenth for

LBPC + ULBPH + CH/ULBPC + ULBPH) and thus take much less time

n comparison to them. LBPC + LBPH, and LBPC + LBPH + CH take

.505 (s), and 1.755 (s) for Corel- 10K and are faster by more than

wice over the average speed of these existing methods . Also, for

Page 16: Color texture description with novel local binary patterns

C. Singh et al. / Pattern Recognition 76 (2018) 50–68 65

Fig. 6. Time taken for feature extraction and the total retrieval time in (seconds) taken for Corel- 5K and Corel- 10K datasets.

C

a

a

1

l

A

1

U

d

b

f

t

t

s

U

a

s

H

a

C

t

e

orel- 10K, ULBPC + ULBPH, and ULBPC + ULBPH + CH take 0.733 (s)

nd 0.888(s) and are approximately five times faster than the aver-

ge performance of these existing methods .

Amongst the methods with feature dimension between 512 and

024, our proposed methods (LBPC + LBPH, LBPC + LBPH + CH) take

ess time as compared to LBP [12] , QLRBP[44] and MDLBP[46].

lso, amongst the methods with feature dimension between

18 and 256, the proposed methods i.e. ULBPC + ULBPH and

LBPC + ULBPH + CH, take less time than LCVBP [42] for both

atasets. However, their retrieval performance is comparable or

etter than the LCVBP method which takes 0.363 (s), and 1.018 (s)

or Corel -5K and Corel- 10K datasets which is the median of the to-

al time for all the methods presented here. They also take less

ime as compared to Gabor [8] which has only 96 features.

It may be recalled that MDLBP provides the next best re-

ults after our two proposed approaches (LBPC + LBPH + CH and

LBPC + ULBPH + CH). It may also be noted that among the existing

pproaches, MDLBP provides the best retrieval results. Its ranking

tands at third for Wang, Corel- 5K , Corel- 10K, and Holidays datasets.

owever, regarding time efficiency, it is the slowest after CSLSP

nd MSLBP with 2.338(s) for feature extraction and retrieval on

orel- 10K. Thus, the proposed approaches are not only best in

he retrieval performance, but also they are computationally very

fficient.

Page 17: Color texture description with novel local binary patterns

66 C. Singh et al. / Pattern Recognition 76 (2018) 50–68

8. Conclusion

The proposed local binary patterns for color images (LBPC) for

deriving the color texture patterns are very effective in color im-

age retrieval problem. We use a plane in 3D color space using RGB

color model and threshold the color pixels in a circularly sym-

metric neighborhood of a pixel with equally-spaced members P

and radius R . The values P = 8 , and R = 2 are observed to pro-

vide very good retrieval accuracy and speed in comparison to

any other combination of P and R . The color pixels, which lie

on or above the plane, are assigned a value 1 and those which

lie below the plane are assigned the value 0. The retrieval accu-

racy is further enhanced by deriving the local binary patterns of

the hue (H) component of the HSI color space, which is called

LBPH. The color histogram (CH) is an effective descriptor of color

images and has been used in the proposed methods to enhance

the retrieval performance. The performance of these color descrip-

tors has been analyzed individually as well as in their combined

forms after fusing their features. Since all approaches are based

on normalized histograms, the fusion of these features is a sim-

ple concatenation process. Therefore, equal weightage is given to

all modalities. The uniform patterns of LBPC and LBPH are also

derived. Thus, we have analyzed five different modalities: LBPC,

ULBPC, LBPH, ULBPH, CH, and their four combinations: LBPC + LBPH,

ULBPC + ULBPH, LBPC + LBPH + CH, and ULBPC + ULBPH + CH. Detailed

experimental analysis for retrieval performance and computation

time reveals that the proposed method LBPC + LBPH + CH achieves

the highest value of mAP across all datasets in comparison to the

existing best local color texture descriptor-multichannel decoded

local binary pattern (MDLBP). A deep analysis reveals that the pro-

posed descriptor is very effective when intra-class variation among

images is very high. The proposed method LBPC + LBPH + CH not

only provides the best mAP values but it also provides fast retrieval

speed because it uses very small number of features (542) as com-

pared to the large number of features used by the MDLBP (2048).

Another proposed method ULBPC + ULBPH + CH with only 148 fea-

tures provides retrieval accuracy comparable to LBPC + LBPH + CH,

but with very low retrieval time. In fact, the retrieval accuracy of

ULBPC + ULBPH + CH is slightly lower than LBPC + LBPH + CH. It pro-

vides much better retrieval results at low computational cost as

compared to the best existing method MDLBP that has very high

feature dimension (2048) which results in high computation cost.

The experimental results also reveal that the extended-Canberra

distance measure outperforms all prominent distance measures to

provide the highest mAP values for most of the methods across all

datasets.

Acknowledgements

We appreciate the suggestions and comments given by the

anonymous reviewers to raise the standard of the paper. The au-

thors are grateful to the University Grants Commission (UGC), New

Delhi, India, for providing financial grants for the major research

project entitled, “Development of efficient techniques for feature

extraction and classification for invariant pattern matching and

computer vision applications”, vide its File No.: 43-275/2014(SR).

One of the authors (Kanwal Preet Kaur) is thankful to the Univer-

sity Grant Commission (UGC) and the Ministry of Minority Affairs

(MOMA), Govt. of India, for providing Maulana Azad National Fel-

lowship (F1-17.1/2013-14/MANF-2013-14-SIK-PUN-22020) for car-

rying out the research work leading to the Ph.D. degree .

References

[1] B.S. Manjunath , J.-R. Ohm , V.V. Vasudevan , A. Yamada , Color and texture de-

scriptor, IEEE Trans. Circ. Syst. Video Technol. 11 (6) (2001) 703–715 .

[2] R.M. Haralick , K. Shangmugam , Textural feature for image classification, IEEETrans. Syst. Man Cybern. (6) (1973) 610–621 SMC-3 .

[3] H. Tamura , S. Mori , T. Yamawaki , Texture features corresponding to visual per-ception, IEEE Trans. Syst. Man Cybern. 8 (6) (1978) 460–473 .

[4] G. Cross , A. Jain , Markov random field texture models, IEEE Trans. Pattern Anal.Mach. Intell. 5 (1) (1983) 25–39 .

[5] B.S. Manjunath , W.Y. Ma. , Texture features for browsing and retrieval of imagedata, IEEE Trans. Pattern Anal. Mach. Intell. 18 (8) (1996) 837–842 .

[6] P. Howarth , S. Rüger , Evaluation of texture features for content-based image

retrieval, in: Proceedings of International conference Image and Video Re-trieval, 2004, pp. 326–334 .

[7] L. Chen , G. Lu , D. Zhang , Effects of different Gabor filters parameters on im-age retrieval by texture, in: Proceedings of IEEE 10th InternationalMultimedia

Modelling Conference, 2004, pp. 273–278 . [8] J. Han , K-K. Ma , Rotation-invariant and scale-invariant Gabor features for tex-

ture image retrieval, Image Vis. Comput. 25 (9) (2007) 1474–1481 .

[9] F. Bianconi , A. Fernández , Evaluation of the effects of Gabor filter parameterson texture classification, Pattern Recognit. 40 (12) (2007) 3325–3335 .

[10] D.S. Zhang , G. Lu , Review of shape representation and description techniques,Pattern Recognit. 37 (2004) 1–19 .

[11] J. Vogel , B. Schiele , Performance evaluation and optimization for content-basedimage retrieval, Pattern Recognit. 39 (5) (2006) 897–909 .

[12] T. Ojala , M. Pietikäinen , T. Mäenpää, Multiresolution gray-scale and rotation

invariant texture classification with local binary patterns, IEEE Trans. PatternAnal. Mach. Intell. 24 (7) (2002) 971–987 .

[13] D.G. Lowe , Distinctive image features from scale-invariant keypoints, Int. J.Comput. Vis. 60 (2) (2004) 91–110 .

[14] Y. Ke , R. Sukthankar , PCA-SIFT: a more distinctive representation for local im-age descriptors, in: IEEE Conference on Computer Vision and Pattern Recogni-

tion (CVPR), 2004, pp. 506–513 .

[15] K. Mikolajczyk , C. Schmid , A performance evaluation of local descriptors, IEEETrans. Pattern Anal. Mach. Intell. 27 (10) (2005) 1615–1630 .

[16] H. Bay , A. Ess , T. Tuytelaars , L.J.V. Gool , SURF: speeded-up robust features,Comput. Vis. Image Understand. 110 (3) (2008) 346–359 .

[17] N. Dalal , B. Triggs , Histograms of oriented gradients for human detection, in:Proceedings of the IEEE Conference on Computer Vision and Pattern Recogni-

tion (CVPR), 2005, pp. 886–893 .

[18] E. Tola , V. Lepetit , P. Fua , DAISY: an efficient dense descriptor applied to wide-baseline stereo, IEEE Trans. Pattern Anal. Mach. Intell. 32 (5) (2010) 815–830 .

[19] B. Li , R. Xiao , Z. Li , Z. Cai , B-L. Lu , L. Zhang , Rank-SIFT: learning to rank repeat-able local interest points, in: IEEE Conference on Computer Vision and Pattern

Recognition (CVPR), 2011, pp. 1737–1744 . [20] M. Calonder , V. Lepetit , M.O. Zuysal , T. Trzcinski , C. Strecha , P. Fua , Brief: com-

puting a local binary descriptor very fast, IEEE Trans. Pattern Anal. Mach. Intell.

34 (7) (2012) 1281–1298 . [21] E. Rublee , V. Rabaud , K. Konolige , GR. Bradski , ORB: an efficient alternative to

SIFT or SURF, in: Proceedings of the IEEE International Conference on Com-puter Vision (ICCV), 2011, pp. 2564–2571 .

[22] J. Chen , S. Shan , C. He , G. Zhao , M. Pietikäinen , X. Chen , W. Gao , WLD: a ro-bust local image descriptor, IEEE Trans. Pattern Anal. Mach. Intell. 32 (9) (2010)

1705–1720 . [23] L.J.V.Gool Everingham , C.K.I. Williams , J.M. Winn , A. Zisserman , The Pascal vi-

sual object classes (VOC) challenge, Int. J. Comput. Vis. 88 (2) (2010) 303–338 .

[24] G.J. Burghouts , J.-M. Geusebroek , Performance evaluation of local color invari-ants, Comput. Vis. Image Understand. 113 (1) (2009) 48–62 .

[25] T. Mäenpää, M. Pietikäinen , T. Ojala , Texture classification by multi-predicatelocal binary pattern operators, in: Proceedings of the International Conference

on Pattern Recognition (ICPR), 20 0 0, pp. 3951–3954 . [26] T. Mäenpää, T. Ojala , M. Pietikäinen , M. Soriano , Robust texture classification

by subsets of local binary patterns, in: Proceedings of the International Con-

ference on Pattern Recognition (ICPR), 20 0 0, pp. 3947–3950 . [27] M. Pietikäinen , T. Mäenpää, V. Jaakko , Color texture classification with color

histograms and local binary patterns, in: Workshop on Texture Analysis in Ma-chine Vision, 2002, pp. 109–112 .

[28] Y. Guo , Z. Guoying , M. PietikäInen , Discriminative features for texture descrip-tion, Pattern Recognit. 45 (10) (2012) 3834–3843 .

[29] T. Ahonen , A. Hadid , M. Pietikäinen , Face recognition with local binary pat-

terns, in: Proceedings of the European Conference on Computer Vision (ECCV),2004, pp. 469–481 .

[30] T. Ahonen , A. Hadid , M. Pietikäinen , in: Face Recognition With Local BinaryPatterns Lecture Notes in Computer Science, 3021, Springer-Verlag, Berlin Hei-

delberg New York, 2004, pp. 469–481 . [31] T. Ahonen , A. Hadid , M. Pietikäinen , Face description with local binary pat-

terns: application to face recognition, IEEE Trans. Pattern Anal. Mach. Intell. 28

(12) (2006) 2037–2041 . [32] C. Shan , S. Gong , P.W. McOwan , Robust facial expression recognition using lo-

cal binary patterns, in: Proceedings of IEEE International Conference of ImageProcessing, 2, 2005, pp. 370–373 .

[33] Z. Guoying , M. Pietikainen , Dynamic texture recognition using local binary pat-terns with an application to facial expressions, IEEE Trans. Pattern Anal. Mach.

Intell. 29 (6) (2007) 915–928 .

[34] C. Shan , S. Gong , P.W. McOwan , Facial expression recognition based on lo-cal binary patterns: a comprehensive study, Image Vis. Comput. 27 (6) (2009)

803–816 . [35] S. Zhang , X. Zhao , B. Lei , Facial expression recognition based on local binary

Page 18: Color texture description with novel local binary patterns

C. Singh et al. / Pattern Recognition 76 (2018) 50–68 67

[

[

[

[

[

[

[

[

[

[

[

[

[

[

[

[

patterns and local fisher discriminant analysis, WSEAS Trans. Sig. Process. 8(1) (2012) 21–31 .

36] V. Takala , T. Ahonen , M. Pietikäinen , Block-based methods for image retrievalusing local binary patterns, in: Proceedings of Scandinavian Conference on Im-

age Analysis, 2005, pp. 882–891 . [37] O.A.B. Penatti , E. Valle , R.S. Torres , Comparative study of global color and tex-

ture descriptors for web image retrieval, J. Vis. Commun. Image Represent. 23(2) (2012) 359–380 .

38] D. Huang , S. Caifeng , A. Mohsen , W. Yunhong , C. Liming , Local binary patterns

and its application to facial image analysis: a survey, IEEE Trans. Syst. ManCybern. Part C 41 (6) (2011) 765–781 .

39] T. Mäenpää, M. Pietikäinen , J. Viertola , Separating color and pattern informa-tion for color texture discrimination, in: Proceedings of 16th International Con-

ference on Pattern Recognition, vol. 1, 2002, pp. 668–671 . 40] T. Mäenpää, M. Pietikäinen , Classification with color and texture: jointly or

separately? Pattern Recognit. 37 (8) (2004) 1629–1640 .

[41] J.Y. Choi , K.N. Plataniotis , Y.M. Ro , Using local binary pattern features for facerecognition, in: Proceedings of 17th IEEE International Conference on Image

Processing (ICIF 2010) Sept. 26–29, Hong Kong, 2010, pp. 4541–4544 . 42] S.H. Lee , J.Y. Choi , Y.M. Ro , K.N. Plataniotis , Local color vector binary patterns

from multichannel face images for face recognition, IEEE Trans. Image Process.21 (4) (2012) 2347–2353 .

43] C. Zhu , C.-H. Bichot , L. Chen , Image region description using orthogonal com-

bination of local binary patterns enhanced with color information, PatternRecognit. 46 (2013) 1949–1963 .

44] R. Lan , Z. Yicong , Y.T. Yuan , Quaternionic local ranking binary pattern: a localdescriptor of color images, IEEE Trans. Image Process. 25 (2) (2015) 566–579 .

45] J. Li , N. Sang , C. Gao , Completed local similarity pattern for color image recog-nition, Neurocomputing 182 (2016) 111–117 .

46] S.R. Dubey , S.K. Singh , R.K. Singh , Multichannel decoded local binary patternsfor content-based image retrieval, IEEE Trans. Image Process. 25 (9) (2016)

4018–4032 . [47] R.C. Gonzalez , R.E. Woods , Digital Image Processing, third ed., Prentice- Hall,

2007 . 48] M. Swain , D. Ballard , Color indexing, Int. J. Comput. Vis. 7 (1) (1991) 11–32 .

49] M. Flickner , H. Sawhney , W. Niblack , J. Ashley , Q. Huang , B. Dom , P. Yanker ,

Query by image and video content: the QBIC system, Computer 28 (9) (1995)23–32 .

50] G-H Liu, Image Rank Machine Based on Visual Attention Mechanism < http://www.ci.gxnu.edu.cn/cbir/Dataset.aspx >.

[51] G-H Liu , J-Y Yang , Content-based image retrieval using color difference his-togram, Pattern Recognit. 46 (2013) 188–198 .

52] T. Deselaers , D. Keysers , H. Ney , Features for image retrieval: an experimental

comparison, Inf. Retrieval 11 (2) (2008) 77–107 . 53] S. Li , M-C Lee , C.-M. Pun , Complex Zernike moments feature for shape-based

image retrieval, IEEE Trans. Syst. Man Cybern. Part A 39 (1) (2009) 227–237 . 54] J.Z. Wang , J. Li , G. Wiederhold , Simplicity: semantics-sensitive integrated

matching for picture libraries, IEEE Trans. Pattern Anal. Mach. Intell. 23 (9)(2001) 947–963 .

55] H. Jegou , M. Douze , C. Schmid , Hamming embedding and weak geometric con-

sistency for large scale image search, in: Proceedings of European Conferenceon Computer Vision (ECCV), 2008, pp. 304–317 .

Page 19: Color texture description with novel local binary patterns

68 C. Singh et al. / Pattern Recognition 76 (2018) 50–68

.Sc. degree in mathematics from Kumaon University, Nainital, India, in 1975 and 1977,

m Indian Institute of Technology, Kanpur, India, in 1982. He joined M/S Jyoti Ltd., Baroda, tiala, India, in 1987. In the year 1994, he joined the Department of Computer Science at

year 1995. Dr. Singh has also served as Dean, Faculty of Engineering and Technology,

e has worked in many diverse areas such as Fluid Dynamics, Finite Element Analysis, years of teaching/research experience. For the last 25 years, he has been working in

ise Removal, Image Superresolution and Medical Image Segmentation. He has published more than 44 papers in various national and international conferences.

ersity, and Masters as well as Ph.D. from Punjabi University, Patiala, India. She served

elhi in 1998, and as lecturer and senior lecturer in the NITTTR, Chandigarh, India for unjabi University, M.M. University, and South Asian University, in India. At present, she is

ty of Saskatchewan, Canada, as a Professional research associate. Currently, she is working applications . Her research interests include Digital Image Watermarking, Content-Based

lysis. She has a number of papers in international journals and conference publications

any reputed image processing journals and conferences. She has also chaired sessions in

computer applications in 2007 from Panjab University, Chandigarh, India, and a post-

stitute of Management, Jalandhar, India, in 2010. She started her career as an assistant 0. She is currently pursuing Ph.D. degree in computer science from Punjabi University,

sing and Content-Based Image Retrieval.

Chandan Singh received B.Sc. degree in Science and M

respectively and a Ph.D. degree in applied mathematics froIndia, in 1982, and later Thapar Corporate R&D Centre, Pa

Punjabi University, Patiala. He became professor in the

Dean, Faculty of Physical Sciences, and Dean, Research. HOptimization and Numerical Analysis. He has almost 40

Pattern Recognition, Image Retrieval, Face Recognition, Nomore than 84 papers in various International journals and

Ekta Walia received B.Sc. degree from Kurukshetra Univ

as a software consultant with DCM DataSystems, New Dseven years. From 2007 to 2014, she served as faculty in P

associated with Department of Computer Science, Universion medical image analysis and computer-aided diagnosis

Image Retrieval, Face Recognition and Medical Image Ana

in these areas. She has been on the reviewing board of mInternational Conferences of repute.

Kanwal Preet Kaur received an undergraduate degree in

graduate degree in computer applications from Lovely Inprofessor in Guru Nanak College, Budhlada, India, in 201

Patiala, India. Her research interests include Image Proces