review on generic object recognition techniques: challenges and opportunities

30
http://www.iaeme.com/IJARET/index.asp 104 [email protected] International Journal of Advanced Research in Engineering and Technology (IJARET) Volume 6, Issue 12, Dec 2015, pp. 104-133, Article ID: IJARET_06_12_010 Available online at http://www.iaeme.com/IJARET/issues.asp?JType=IJARET&VType=6&IType=12 ISSN Print: 0976-6480 and ISSN Online: 0976-6499 © IAEME Publication REVIEW ON GENERIC OBJECT RECOGNITION TECHNIQUES: CHALLENGES AND OPPORTUNITIES Prof. Deepika Shukla Comp. Science and Engineering Department, Institute of Technology, Nirma University, Ahmedabad, India Apurva Desai Department of Computer Science and Information Technology, VNSGU, Surat India ABSTRACT Recognizing objects automatically from an image is a fundamental step for many real-world computer vision applications. It is the task of identifying an instance of object in an image or video sequence without or least human intervention and assistance. In-spite of very high complexity, human beings perform this task with very less effort and even in the state of least attention. Little effort is needed for the humans to recognize huge number of and various categories of objects in images, though ‘object’ in the image may be different with respect to size / scale, viewpoint, position or orientation. We are even able to recognize the objects from an image, when they are only partially visible or present against cluttered background. Not only this, the recognition can be for specific instance of object or object category/class. When the task is done for classes of the object it is known as Generic object recognition or object-class detection or category-level object recognition. It has been found that over the years many techniques have evolved for recognizing object classes from images, but any automated object recognition system till date has not gained this capability fully at par with human beings. This very fact makes recognition of objects from an image, the most basic and fundamental challenge in the field of computer vision research. The purpose of this study is to give an overview and categorization of the approaches used in the literature for the purpose of Generic Object Recognition and various technical advancements achieved in the field. Mostly the survey focusses on the leading work since year 2000. We have discussed the challenges that the field is currently facing. We have also made an attempt to suggest future research directions in the area of Generic Object Recognition. Finally we conclude the study with a hope that in

Upload: iaeme-publication

Post on 27-Jan-2017

96 views

Category:

Engineering


1 download

TRANSCRIPT

http://www.iaeme.com/IJARET/index.asp 104 [email protected]

International Journal of Advanced Research in Engineering and Technology

(IJARET) Volume 6, Issue 12, Dec 2015, pp. 104-133, Article ID: IJARET_06_12_010

Available online at

http://www.iaeme.com/IJARET/issues.asp?JType=IJARET&VType=6&IType=12

ISSN Print: 0976-6480 and ISSN Online: 0976-6499

© IAEME Publication

REVIEW ON GENERIC OBJECT

RECOGNITION TECHNIQUES:

CHALLENGES AND OPPORTUNITIES

Prof. Deepika Shukla

Comp. Science and Engineering Department,

Institute of Technology, Nirma University, Ahmedabad, India

Apurva Desai

Department of Computer Science and Information Technology,

VNSGU, Surat India

ABSTRACT

Recognizing objects automatically from an image is a fundamental step for

many real-world computer vision applications. It is the task of identifying an

instance of object in an image or video sequence without or least human

intervention and assistance. In-spite of very high complexity, human beings

perform this task with very less effort and even in the state of least attention.

Little effort is needed for the humans to recognize huge number of and various

categories of objects in images, though ‘object’ in the image may be different

with respect to size / scale, viewpoint, position or orientation. We are even

able to recognize the objects from an image, when they are only partially

visible or present against cluttered background. Not only this, the recognition

can be for specific instance of object or object category/class. When the task is

done for classes of the object it is known as Generic object recognition or

object-class detection or category-level object recognition. It has been found

that over the years many techniques have evolved for recognizing object

classes from images, but any automated object recognition system till date has

not gained this capability fully at par with human beings. This very fact makes

recognition of objects from an image, the most basic and fundamental

challenge in the field of computer vision research. The purpose of this study is

to give an overview and categorization of the approaches used in the literature

for the purpose of Generic Object Recognition and various technical

advancements achieved in the field. Mostly the survey focusses on the leading

work since year 2000.

We have discussed the challenges that the field is currently facing. We

have also made an attempt to suggest future research directions in the area of

Generic Object Recognition. Finally we conclude the study with a hope that in

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 105 [email protected]

near future more sophisticated object class recognition systems would be

developed in an efficient and cost effective manner.

Key words: Object Recognition, Generic Object Recognition, Object class

Recognition, Scene Understanding, Scene categorization, Image Analysis,

Computer Vision, Machine Vision, Scene Analysis, Image Analysis.

Cite this Article: Prof. Deepika Shukla and Apurva Desai, Review on

Generic Object Recognition Techniques: Challenges and Opportunities.

International Journal of Advanced Research in Engineering and Technology,

6(12), 2015, pp. 104-133.

http://www.iaeme.com/IJARET/issues.asp?JType=IJARET&VType=6&IType=12

1. INTRODUCTION

Automated recognition of objects in images is a critical and fundamental step for

many real-world computer vision applications. It is the task of finding a given object

in an image or video sequence without or least human intervention/ assistance. As we

know, very little effort is required at our part to detect and recognize huge number of

classes of objects in images though image of the object may be different with respect

to size / scale, viewpoint position or orientation. Human beings are able to recognize

the objects from an image even when they are only partially visible or present against

cluttered background. Also, the ability to generalize from examples and categories

objects, events, scenes, and places is one of the core capabilities of the human visual

system; For human being this is a mundane activity, but imbibing these capabilities in

machine, has still proved to be significantly challenging task for computer vision

systems in general.

The reason behind this may root to the fact that “Automatic Object Recognition”

requires understanding of human visual perception and so becomes a

multidisciplinary research area involving knowledge and expertise of fields like

optics, psychology, pattern recognition, artificial intelligence, machine learning and

most importantly cognitive science which in itself needs sophisticated concepts and

tools from mathematics as well as computer science [1].

Object recognition is a dominant field of research in the computer vision as well

as image analysis applications and even the simplest machine vision task cannot be

solved without the help of recognition. The fact can be evidenced by the existence of

vast volume of research conducted in this area over the past three decades. The

statement can be substantiated by the fact that, if one just gives, “objects recognition

from images” as the search string on ieeexplore.org, gets more than 20000 results. So,

from the substantial volume of current literature existing on the topic, we can also say

that “Object Recognition” field is closely tied to and is part and parcel of computer

vision research.

This paper reviews most of the leading state-of-the-art researches performed in the

area of Generic object recognition. But more specifically, it is focused to get the

insight into following Research Questions pertaining to the topic of Generic Object

Recognition.

What are the generic object recognition techniques and approaches drawn by the

literature?

What different representation techniques are used for object representation?

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 106 [email protected]

Which feature detection and extraction methods are used by most of the prominent

researchers on the topic?

Which classification/learning technique has been used in the classification stage of the

object recognition pipeline?

The rest of this paper is organized as follows. Section 2 introduces and explains

the problem of Generic object recognition which can be considered as specific subset

of object recognition problem. Section 3 concentrates on the challenges that the field

of object recognition faces in general and Generic Object Recognition in particular.

Section 4 discusses the vast literature existing for the topic area. Section 5 manifests

roadmap to future research areas and directions. Section 6 finally sought conclusion

of the study.

2. GENERIC OBJECT RECOGNITION PROBLEM

The problem of object recognition can be viewed as a classification or labelling

problem where models/representation of known objects are available to the system

and when a novel image is given, the system has to predict the class of the object[s]

present in the Image. Formally, it can be stated as, given an image containing one or

more objects of interest (and background) and a set of labels corresponding to a set of

models known to the system, the system should assign correct labels to regions, or a

set of regions, in the image. i. e. Object recognition systems should assign a high level

definition of an object based on the image data, that is represented.

Oftentimes, the task of object recognition is considered as broadly comprising of three

sub-tasks;

Object detection: Detecting whether an instance of the object category is present in

the image or not.

Localization: To give the location of object category. Drawing a bounding box

surrounding the object instance is most prominently used in literature to show the

result of localization.

Visual category recognition: To recognize and label the class/category of object

present in the image.

Moreover, the image being presented to the object recognition framework for the

purpose of recognizing objects from it may have single instance of some class of

object or may have multiple instances of single class or multiple instances of multiple

classes. Therefore, the object recognition approaches at the top-most level can broadly

be categorized to follow top-down, bottom-up or hybrid approach. And within that it

can be for recognizing specific or generic object. So basically image-based object

recognition can be stated as; Given a database of objects and an image, determine

what, if any of the object[s] are present in the image. Thus the problem of object class

recognition can be considered as an instance of supervised classification.

Another dimension along which, the task of object recognition can be categorized

is: First, Where a specific object to be recognized is known to the system and the

system is trained for that specific object category only. For example, Face recognition

, pedestrian recognition Second, Generic object recognition system. Generic object

recognition means that the computer recognizes objects from images by their general

name [2] or common name. Figure-1 shows an instance of Generic Object

Recognition. Generic object recognition has been also referred as object-class

detection or category-level object recognition in literature [14] which aims at

recognizing the class to which the object present in the image belongs. The images

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 107 [email protected]

can have single instance of a class, multiple instances of same class or multiple

instances of multiple classes. When categorization of multiple objects of multiple

classes in an image is done, it is known as scene categorization.

Figure 1 Generic Object Recognition

2.1. Architecture of the object recognition system

The current vision systems can said to be consisting of activities as shown in Figure -

2.

Figure 2 Activities involved in a typical vision system

Any recognition system would involve these or some subset of these activities in

its life cycle. In general, after image acquisition stage, image is pre-processed for

performing noise removal and some kind of enhancement. The pre-processing stage is

followed by feature extraction and description/representation stage which then are

passed for recognition. In the representation stage, the objects can be represented as 2-

D or 3-D. Figure-3 shows the general architecture of object recognition system.

Object Recognition task is affected by several factors and can differ according to

various aspects as shown in Figure-4. It shows the categorization of aspects in which

the work is going on, in the field of object recognition. The approaches may differ on

the basis of form and representation of objects, Matching schemes, Image Formation

Model, Type of Features, Type of Image and type of data suited for categorization.

Once we studied various aspects we figured out that these approaches mainly differ in

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 108 [email protected]

the object representation method based on type of features or the classification

approach adopted by the method in the recognition phase.

As the factor changes it can easily be observed that the approach changes

substantially but basically these approaches broadly follows three paradigms for

formulating and attempting the solution to the problem of Object Recognition from an

image; Bottom-up, Top-down, Hybrid paradigm [103].

Figure 3 Generic Architecture of Object Recognition System

Bottom-upIt can also be considered as Image analysis from its low level data and is

based on image segmentation techniques. It considers the raw image data which is

available in the form as it is acquired. Boundaries of the homogeneous regions are

extracted by performing non-purposive segmentation without prior knowledge about

properties of individual object classes. No attempt is made to make any prior

assumptions related to what these objects are. Fixed set of attributes are used to

characterize these regions and objects are linked together to characterize the scene

itself. However, without some additional information, purely bottom-up approaches

have so far been unable to yield figure-ground segmentations of sufficient quality for

object categorization [Leibe & Shiele] till 2009, then after many approaches have

been developed [85, 86, 88, 89] which uses bottom-up segmentation methods as

discussed in [82] and [85] and have achieved remarkable results which will be

discussed in detail later in literature review section of the paper.

Top-down: This is Image Analysis from the Semantic level data. Contradictory to

earlier approach, this methodology proceeds with an assumption that the image does

contain a particular object. If the problem is of scene categorization, it assumes that it

is a particular type of scene. The system will attempt to verify the existence of a

hypothesized object. Purposive segmentation may be performed or specialized ways

are used to represent the object.

Hybrid: Combination of the earlier two paradigms are used [61],[79] in this kind of

approach.

3. KEY CHALLENGES

3.1. Challenges overview

As stated earlier, the problem of Object recognition in general and Generic Object

Recognition in particular faces various challenges.

(I) The appearance of an object in the image can have a large range of variation due

to:

1. Viewpoint changes

2. Scale, Orientation and Shape changes (e.g., non-rigid objects)

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 109 [email protected]

3. Photometric effects (scene illumination etc.)

4. Scene/Background clutter (therefore objects may be occluded)

(II) Different views of the same object can give rise to widely different images.

(III) Large number of object categories existing in real-world and these categories

may have very less inter-class variation.

Figure 4 Factors affecting the task of object recognition

3.2. Description

Object recognition can be considered as yet another data processing task, so data is

given the highest priority thus acquisition should be considered as most important

step. In recent years, with the advent of high quality camera and other image

capturing devices, we can collect a huge amount of data (images) in various forms

like intensity images, range images and also from various sources like web but the

major problem that computer vision research community is facing today is scarcity of

accurately and precisely labelled image examples. As stated earlier, object recognition

problem can be considered essentially, a supervised classification task, and for that to

work successfully there remains the need of labelled images examples. The problem

becomes more gruesome for the reason that the task is labour intensive. Also due to

the non-availability of human experts which can do image annotation task efficiently

and accurately the task becomes more challenging.

‘Feature Extraction’ is the next crucial step in the pipeline of generic object

recognition. Assuming that the data is available, the feature extraction becomes the

most important stage of the entire object recognition framework. If, suitable features

of right dimensions are not extracted, this phase can become the bottle neck of the

recognition pipeline. Though recently many sophisticated approaches have been

developed and are existing in the literature ,but they are not sufficient to describe

every object , so feature extraction becomes too object specific and varies as either

viewpoint, size and illumination conditions of the image capturing varies. Thus

representing images by effective features is crucial to the performance of various

image analysis tasks. Features can be low-level (colour, texture, Intensity), middle-

level (Image Patches) or High-level (objects, textually annotated objects). Figure-5

shows a probable classification of different kinds of features.

Choosing and deploying an appropriate classifier is the next important step of the

pipeline. The classifier can be linear or a non-linear one. Various classifiers like

Byesian classifier, SVM, decision trees, Neural Networks etc. are utilized in literature

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 110 [email protected]

for the purpose of classification each possessing its own benefits and drawbacks. One

important inherent issue related to the classification stage is; scalability of the

classifier. As number of existing object categories is too large in real world and many

visual features are required to model each category thus forcing the system to have

huge volume of training data to model variety of category classifiers. In order to keep

scalability manageable, a linear classifier is commonly utilized, but its classification

performance is inferior to the nonlinear one whereas the non-linear classifiers are

more computation intensive. To remedy this defect of the linear classifiers, a design

of rich image feature set, (which is after all, a key factor in the success of the image

recognition system) per object class is required so that the system can distinctly

recognize objects from images possessing inter-class and intra-class variation as

shown in Figure-6. Additionally, classifier has to be updated continuously because

even if it is trained once for a category/class of object and previously unseen instances

of object emerge or appearance of the object evolves, the earlier trained classifier will

not give correct results. This kind of flexibility and resilience to change is inherently

expected from any object recognition framework.

Figure 5 Classification of Image Features

Figure 6 Images of different instances of object (Dog) in varied imaging conditions.

Intra-class appearance variations refer to the appearance differences among

different objects of the same class [14]. Intra-class appearance variation may be either

due to difference in colour, shapes and sizes of the object’s instance or due to

difference in imaging conditions. For example Image of the same object taken at

different time of the day, in different seasons or at different places, with different

devices and different viewpoints, will be entirely be different. In addition to intra-

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 111 [email protected]

class appearance variations, Generic Object Recognition system has to efficiently and

distinctly handle inter-class appearance variations also; which in many cases would be

very less as shown in Figure 7. For example, object recognition system should be

capable of distinctly recognize between a donkey and a horse or a Horse and a Mule.

Figure 7 Images of Horses and Donkeys with very small inter-class appearance

variation: Lower row is images of horses (adapted from [14])

The performance of generic object recognition framework is generally judged

upon criteria like robustness against noise, invariance to basic geometric

transformations, invariance to illumination and viewpoint changes and its ability to

handle the number and different types of objects, ability to handle intra-class and

inter-class variations, recognize objects in presence of clutter or complicated

background and also to be able to recognize the object even if it is partially occluded

accurately and efficiently. These requirements are expected implicitly and must be

present in any framework for object recognition and as a result these issues can be

considered as key challenges for the field of generic object recognition.

4. LITERATURE REVIEW

4.1. Overview

The object recognition pipeline, as stated in the earlier section, consists of the key

tasks like Image acquisition, Pre-processing, Feature Extraction, Feature

representation/Feature description and Classification. However Image acquisition and

Pre-processing phase falls out of the scope of this study. Although, most of the related

work surveyed and cited here focusses on one or the other phase of this pipeline, our

main focus in this study is on feature extraction and description techniques and also to

obtain the answers to the research questions put up at the beginning of this

manuscript. Although, various groups of researchers in the literature have attempted

to survey and review the work in the field of computer vision but either they are

related to some specific object like a survey on face recognition is presented in

[108,109] or various descriptors have been compared and surveyed [14,45,48] or a

separate survey is presented in [114] on object recognition using deep neural

networks. That is one particular aspect of the topic is explored and related literature

review is presented while discussing their core work. Periodically comprehensive

surveys on generic object recognition [14, 15] have been published in the past but

looking to the rapid pace of achievements in the field, it seems natural to survey the

most recent developments and object recognition techniques available in the

literature. In this study, we have mostly tried to review the work done in the field

since year 2000 but more emphasis is given on surveying the work done after 2011.

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 112 [email protected]

The rationale behind this is that most of the surveys and papers mostly talk

extensively of the approaches before 2011. But, looking at the pace of technical

advancements in the field, lot of approaches have emerged since 2011 which demands

detailed reportage and covering them is the basic motive of this review. Therefore, the

study is also aimed at presenting the survey in such a way which should help to gain

an insight into this field of research. Also as noted in introduction section that the task

of object recognition is considered as broadly comprising of three sub-tasks: Object

detection, Object localization and object classification, but in this manuscript we have

studied the approaches of generic object recognition which is the highest level of task

in the object categorization subtasks; i.e to categorize the class of object in the image,

object detection is inherently performed and many a times they needs to be localized

also. Due to this reason we have not segregated the approaches on the basis of

detection, localization or categorization.

4.2. Features and Feature Descriptors

The foundations of the field can be traced back to 1950s and 1960s, when early work

was done in very simplistic domains [1]. The world was modelled as being composed

of blocks defined by the coordinates of their vertices and edge information. The

“block image” represented areas of uniform brightness in the image and the edges of

blocks were located in the areas of intensity discontinuity. But very soon it was

realised that, it is not an ideal way to represent the complicated information presented

in the image. Since then various strategies are being developed for the task of object

recognition with an emphasis on feature extraction stage and in the usage of novel and

efficient type of feature descriptor.

Object recognition can be classified into various broad categories. These include

model-based approaches, shape-based approaches and appearance-based approaches.

Model-based approaches try to represent objects as set of various three dimensional

objects [1, 12, 13] like generalized cylinders, cones, cubes, cuboids spheres etc.

Shape-based approaches [13, 19, 20, 21, 52, 53] represent the objects by shape

primitives like boundary fragments, contours, shapelets, etc In contrast, for

appearance-based models only the appearance is used, which is usually captured by

different two-dimensional views of the object-of-interest. So, it can be observed

easily, whatever be the representation method object representation takes the centre

stage in the entire object recognition pipeline. And in turn the problem of object class

recognition reduces to the generating an efficient representation of object which can

detect, localise and identify the class of object discriminatively and repetitively.

As stated earlier extracting and describing features efficiently, of the objects from

the images, decides the fate and success of a typical object recognition system. In a

generic object recognition or categorization system, the relevant features or

descriptors from a characteristic point, patch or region of an image are often obtained

by different approaches. As shown in Figure-5, the features at the top most level can

be divided into two categories global and local wherein the former characterizes the

image as a whole whereas latter represents some local information in form of pixel,

patch or region. Yet another direction along which many researchers have tried to classify features is structural and statistical. Although, there are various classifications

for features but there exists significant overlap among these classes. For example,

local features can be structural as well as statistical. These features are often

combined to form various descriptors especially region level descriptors are formed

by combining colour, texture and other such low-level features.

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 113 [email protected]

As far as pixel level features are concerned, they are regarded as low level

information of the image and are directly computed from the grayscale value of the

pixels individually and generally used to build more sophisticated patch-level or

region level descriptors. We now briefly discuss some of the best performing

descriptors proposed and utilized over the years. This is not meant to be an exhaustive

discussion of the existing approaches, but rather to provide a sample of some

relatively successful and widely used approaches over the years.

4.2.1. Appearance-Based Object Representation

Local Scale-Invariant Features (SIFT) [3][4] introduced by Lowe, is regarded as one

of the most popular patch-level feature descriptors reported in literature. Feature

identified are shown to be completely invariant to basic geometric transformation and

partially invariant to illumination changes and occlusion. SIFT features proved more

successful, as they do not depend on exact grey level distribution within an image

patch, instead use general configuration of image gradient [60]. This was considered

as one of the prominent approach in the area of object recognition and the work is

considered as milestone in the research of object recognition, computer vision and

other image analysis problems. However, as the descriptors are appearance based, and

may produce poor result especially if the object does not have enough information of

its texture features. The SIFT is applied for the problem of object recognition in many

works. Two such usages are mentioned in [3] and [4]. In various other work [2, 39,

42, 75, 110, 111, 112], some kind of improvisation has been achieved by combining

other features along with SIFT or using other filters than Gaussian [110]. The

dimension of key-points obtained when SIFT is applied is relatively large in number,

hence resulting into high dimensional data. This drawback was realised by authors in

[5] and SIFT was extended as PCA-SIFT , where Principal Component Analysis is

applied to normalized gradient patch resulting into lesser dimensional descriptor.

PCA-SIFT yields 36-dimensional descriptor which is fast for computation and

matching but are less distinctive [6] while descriptor introduced by Mikolajczyk and

Schmid[45] namely GLOH (Gradient Location-Oriented Histogram) is another

variant over SIFT which proved to be more distinctive with the same dimension[6].

Also, a colour image-based SIFT has been demonstrated in [75], wherein in place of

intensity gradients, colour gradients are used in Gaussian framework.

As mentioned earlier, high dimensionality of the descriptor is the major limitation

of SIFT, another effective patch-level descriptor SURF (Speeded-Up Robust

Features) is proposed in [6] by Bay et al. The authors have made use of integral

images which results into yielding not only faster but distinctive and repeatable

features. The authors based their descriptor on Hessian matrix but uses very basic

approximation. Moreover only 64 dimensions are used which is much less than

SIFT’s 128 dimensional vector. Though one can argue that PCA-SIFT results in only

36-dimensional vector but at the same time it loses the distinctiveness whereas SURF

has been proved more distinctive and repeatable.

Another level at which features descriptors are generated in numerous papers is at

region level. Dalal and Triggs [32, 33, 34] used grids of locally normalised

Histograms of Oriented Gradients (HOG) as descriptors for object detection in static

images. The technique counts occurrences of gradient orientation in localized portions

of an image. The detector window is tiled with a grid of overlapping blocks in which

Histogram of Oriented Gradient feature vectors are extracted. Detector thus presented

is contrast-based which makes it robust to small changes in image contour locations

and directions, and significant changes in image illumination and colour, while

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 114 [email protected]

remaining highly discriminative for overall visual form. In work of Dalal and Triggs

[32, 33,34] is aimed at detection of humans in particular, but also proved effective in

detecting other object classes from images. HOG descriptor has proved very efficient

descriptor for representing structured objects. For example, It has outperformed all

other descriptors in pedestrian detection from videos and images. Inspired by HOG

[32] Bosch et al[36] proposed a novel descriptor called PHOG (pyramid of HOG).

The idea was to represent local image shape and its spatial layout, together with a

spatial pyramid kernel of Bag of Features (BoF) [25,26]. Each image is divided into a

sequence of increasingly finer spatial grids by repeatedly doubling the number of

divisions in each axis direction (like a quadtree). The number of points in each grid

cell is then recorded. HOG vector is computed for each grid cell at each pyramid

resolution level. The final PHOG descriptor for the image is a concatenation of all the

HOG vectors. This concatenated HOG vector is then normalized to ensure that texture

rich or images with more edges are not weighted more strongly than others. Another

descriptor which is built on the idea of histogram of gradients (HOG) is CoHOG (Co-

occurance histograms of gradients) proposed in [37]. CoHOG can express shapes in

more detail than HOG as CoHOG are histograms whose basic units are pairs of

gradient orientations. Histogram is referred as co-occurrence matrix. Due to this

pairing, the vocabulary size increases resulting into more specific expression of shape

of object in the image. The usage of higher dimensional matrix makes CoHOG

powerful in terms of its discriminative power but at the same time becomes highly

computation intensive.

Bag of Features and visual codebook based approaches

The approach is inspired by BoW (Bag of Words) approach which was first proposed

in 1997 by [38] for describing the textual data for the purpose various text analysis

tasks. It is used to represent a text document or a sentence written in natural language,

as set of words, not taking into consideration its grammar or the order in which these

words occur in the original text. The frequency of occurrence of each word is

calculated and then used for various language processing tasks. The analogous term

BoF( Bag of Features), is used to represent the approach. Similar to BoW model, here

image is represented as order less collections of local features of Image. Similar terms

like Bag of Keypoints (BoK), Bag of Visual Words(BoVW) by various researchers is

used in their works. The method is based on vector quantization of affine invariant

descriptors of image patches [39]. A bag of keypoints corresponds to a histogram of

the number of occurrences of particular image patterns in a given image. The method

uses clustering to obtain quite high-dimensional feature vectors for a classifier. As

construction of codebook is done in the BoF approach, at many times it is also

referred as codebook-based approach. The method includes following main steps.

Detection of image patches for computation of patch descriptors

Computing patch descriptors for these patches. These descriptors can be any feature

invariant descriptors like SIFT [3,4] or any variant of it or any other lower level

descriptor like Harris-affine [43], MSER.

Construction of a visual codebook/vocabulary/dictionary by assigning patch

descriptors to predetermined clusters (a vocabulary) with a vector quantization

algorithm that groups similar features together. For determining clusters instances of

usage of several clustering techniques are available. However, more frequently k-

means clustering is applied. [39]. Whereas Hierarchical k-means clustering is adopted

by [49] and mean-shift by authors in [35].

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 115 [email protected]

Generating a histogram of number of occurrences of particular patches assigned to

each cluster. The size of the resulting histogram equals the size of the codebook and

hence the number of clusters obtained from the clustering technique [40].

Treating the bag of features as a feature vector and using a classifier to classify the

respective image patch. A distance measure is required when comparing two term

vectors for similarity but this measure operates in the term vector space as opposed to

the feature space.

There are two reasons why the bag-of-features image representation (BoF) proved

to be popular for indexing and categorization applications. First, this representation

benefits from powerful local descriptors, such as the SIFT descriptor and Second,

these vector representations can be compared with standard distances, and

subsequently be used by robust classification methods such as support vector

machines [50]. Also, the codebook model-based approaches, while ignoring any

structural aspect in vision, provide state-of-the-art performances on current datasets

[40]. The discriminative power of such a visual codebook determines the quality of

the codebook model, whereas the size of the codebook controls the complexity of the

model. The codebook-based approaches are considered as simple and efficient, and

also can be made robust to clutter, occlusion, viewpoint change, and even non-rigid

deformations [26, 25]. Inspite of being one of the popular and successful approaches,

we find that BoF and visual codebook generation approach has also got certain

limitations. As BoF expresses the image as appearance frequency histograms of visual

words by quantizing SIFT like features, location information and the geometric

relationship between key-points are lost. Also as vector quantization is involved so

inherently loss of information occurs. Also due to loss of geometric relation between

the features, localization of the object is not possible.

To overcome the limitation of orderless representation of objects, several researchers have proposed approaches to augment bag of features with global spatial

relations in a way that significantly, at one end improves classification performance

while at the other end remain simple and computationally efficient so that can be

applied for the real-world applications [27]. Authors in [27] have demonstrated that

bag of feature description of the image can be extended to spatial pyramids so that the

spatial location information of the features can be retained. To generate these spatial

pyramids, the input image is partitioned into increasingly fine sub-regions.

Histograms of local features are computed over these sub-regions. The histograms are

further concatenated to generate the final features. This representation is combined

with a kernel-based pyramid matching scheme proposed by [24] that efficiently

computes approximate global geometric correspondence between sets of features in

two images. While the spatial pyramid representation sacrifices the geometric

invariance properties of bags of features, it compensates for this loss with increased

discriminative power derived from the global spatial information.

Similarly in [2], to overcome the problem inherent to BoF approach, graph is

constructed by connecting SIFT key-points with lines. As a result, the key-points

maintain their relationship, and then structural representation with location

information is achieved. Since graph representation is not suitable for statistical work,

the graph is embedded into a vector space according to the graph edit distance. As a

result, authors achieved recognition accuracy compared to the conventional method in

their experiments on PASCAL VOC and Caltech-101 datasets. So, the basic idea to

achieve the improvement in BoF approach is to somehow incorporate the spatial

location information of features in BoF features so that the method can not only be

used for recognition but can also be successfully applied for object localization. The

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 116 [email protected]

authors in [47] have achieved improvisation by adding binary signatures to the

descriptors, First, Hamming Embedding (HE) of the SIFT descriptors; analogical to

hamming distance and second integrate a weak geometric consistency (WGC) check

within the inverted file system which penalizes the descriptors that are not consistent

in terms of angle and scale. In this way geometrical information is incorporated in the

index with very large datasets. But at the same time both Hamming Embedding (HE)

and WGC require to store additional information, hence memory requirement of index

increases.

The visual codebook approach has been used by several other researches in

slightly different way. For example, Liebe et al in [7,8,9] have adopted a two staged

approach In first stage a codebook of local appearances is learnt which contains

information, which local structures may appear on objects of the target category.

Next an Implicit Shape Model (ISM) that specifies where on the object the codebook

entries may occur. To create the codebook the authors have adopted the method

presented in [17] by Agarwal and Roth. From a variety of images, 25 x 25 pixel

patches are extracted with the Harris interest point detector. These patches are

clustered using agglomerative clustering to generate a compact cluster. These

codebook entries are used to define implicit shape model of the objects. The approach

do not try to create and define a separate model for all possible shapes an object can

take rather define shapes of an object in terms of patches that are consistent in local

appearances. Due to this concept, less number of training examples are needed to

learn object’s probable shapes. A second time codebook entries are scanned and all

those entries are activated whose similarity is above a certain chosen threshold. The

threshold chosen would be same as the threshold used during clustering performed in

the first step. While in recognition stage generalized Hough transform is performed

for identifying possible object centre.

GIST: Humans can recognize the gist of a novel image in a single glance,

independent of its complexity [69], by considering them in a “holistic” manner, while

overlooking most of the details of the constituent objects. Intuitively, GIST

summarizes the gradient information (scales and orientations) for different parts of an

image, which provides a rough description (the gist) of the scene. Input image is

divided into non-overlapping regions. The region is then further divided into sub-

regions and then Gradient Orientation histogram is computed for these sub-regions.

The GIST descriptor for a region is formed by concatenating these Gradient

Orientation histograms for all sub-regions of a region. The approach is more

prevalently used for scene understanding purpose. Approaches based on GIST cannot

be considered as an alternative to image analysis based on local feature based

approaches but can be considered as an additional support for recognition problems

by helping to constrain the local feature based image analysis. In [72,73] short binary

codes are used to compress local GIST descriptors and demonstrated that the

approach works on millions of images obtained from internet without sacrificing the

recognition accuracy and effectiveness.

4.2.2. Shape-Base Approaches

Many approaches based on intensity, colour gradient of the image patches or region is

discussed in the previous part of this paper. Although as noted, these descriptors are

very powerful and have shown to perform object recognition with remarkable

effectiveness. Still there may be a case where two object class exist with same colour

and texture with only difference in shape or for the classes where the appearance is

very much variable in every instance of the object. Such objects cannot be represented

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 117 [email protected]

with only colour, intensity based features alone. For example, if we consider fruit

class, raw mango and capsicum are of green colour, but having entirely different

shape. Recognition community also very soon understood that across the exemplars

that belong to a category, shape is a more invariant property than appearance. As a

result, the majority of recognition systems from the mid-1960s to the late 1990s

attempted to extract shape features, typically beginning with the extraction of edges,

at occluding boundaries and surface discontinuities, edges capture shape information

So, shape is yet another important cue which can be used to generate a discriminative

representation of objects. To compute the shape of the object, different authors have

taken different methods. Shape cues are frequently captured and described at the

region level for object class recognition or detection using contour or boundary

fragments [19], shapelets, edgelets [20], shockgrphs etc. Another area of research, as

far as object’s shape-based detection is concerned, is how to set up the

correspondence between shape extracted from training and test image i.e ; How to say

two shapes are matching [52,53]. One of the limitations of shape-based object

description is that it cannot capture intra-class variations in very discriminative way.

For example, a Zebra cannot be differentiated from a Horse. Often shape-based cues

are combined with other appearance based object cues.

4.2.3 Part-Based approaches

Object as 3D volumetric parts

Earliest attempts at solving the object recognition problem used high level 3D parts

based objects, such as generalized cylinders (Binford) and other deformable objects,

such as geons (Biederman [13] )and superquadrics (Pentland) [79]. The common

characteristic among all of them is that they all based on symmetry; a physical

regularity in our world which is exploited by our human visual system. However in

practice it becomes too complex to extract such parts efficiently and in an inexpensive

manner. But once extracted they are more semantically nearer to description of the

image content. Such parts would be limited in number, as compared to the approaches

where low-level features and mid-level features are used to describe the object.

Although the methods based on low-level and mid-level features score on their

simplicity, ease of extraction and attractive invariance properties; but have proved to

be weak in expressing high-level semantic information of the image. The above noted

facts had made object representation using 3D volumetric parts had achieved lot of

attention in the decade of 70s and 80s. The detailed coverage of the topic is out of the

idea of this study but the works of Binford and Nevatia [115] can be explored for

further information related to the concept.

Recognition based on parts

In Part-based object recognition approaches, object is modeled as a set of

geometrically constrained set of various parts of the image where each part has a

distinctive appearance and spatial position. In such approaches, shape is represented

by the mutual position of the parts [22]. Using such features it is determined whether

an instance of object of interest exist in the image or not and if at all it exist where it

exist in the image. Various methods exist in literature which differs on how these

parts are detected, how their position could be represented and what should be the

ideal number of parts to represent an image. Generally these parameters are tuned to

the requirement of the approach. In [22] Objects are modelled as flexible

constellations of parts. A probabilistic representation (which in this case the authors

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 118 [email protected]

have used Gaussian), is used for all aspects of the object like shape, appearance,

occlusion and relative scale. To learn and model the object category, first regions and

their scales are detected. Once the regions are identified, they are cropped from the

image and rescaled to the size of a small typically 11×11 pixel patch and then

parameters of the above densities are estimated from these regions, such that the

model gives a maximum-likelihood description of the training data. To detect the

features, a histogram is generated of the intensities in a circular region of some radius.

This is done for each point on the image. The entropy of this histogram is calculated

and local maxima of this histogram are considered as scale of the region. The N-

regions with highest saliency over the image provide the features for learning and

recognition. To reduce the dimension of the feature set, PCA has been used.

Deformable part based approach

Deformable Part Models constitute the state of the art for sliding-window object

detection [99]. The DPM’s are inspired by pictorial structure representation

introduced in [91] by Fischler and Elschlager where an object is modelled by a

collection of parts arranged in a deformable configuration [92]. To represent visual

properties of the object small picture segments are used whereas the deformable

configuration is captured by spring-like connections between these visual picture

segments. An energy function is computed by computing match cost for each part and

deformation cost for each pair of connected parts and this energy function is

minimized to find the best match of model with in an image. The effectiveness of

pictorial representation in case of image matching demonstrated in [91] is due to the

fact that the representation is simple. In addition the representation possesses wide

general applicability as it is not dependent on any particular scheme to model the

appearance of the parts so can be used to represent quite generic objects. But at the

other end, the model suffers from certain very critical limitations. Too many

parameters are involved in the construction of the model thus the energy minimization

function solving becomes very computation intensive. Also the best match is only

found likewise, if the image consist multiple instances of the same object, they would

not be detected by the pictorial representation given by [91]. The issues in pictorial

representation are aptly handled by Felzenswalb in pioneering work reported in [92].

Pictorial representation proposed by Fischler and Elschlager constructs the

representation which can be viewed as graph whereas Felzenswalb and Huttenlocher

used tree representation realising that many objects in real-world can be represented

by using a tree structure especially when the object to be modelled are human beings,

animals. Using this improvisation finding best match model to an image can be

computed in polynomial time. The approach demands that the graph which is

generated to represent the object be acyclic and function dij(li , lj) measuring the

degree of deformation of the model when part vi is placed at location li and part vj is

placed at location lj needs to be a Mahalanobis distance between transformed

locations. DPM’s are impressive way of object representation. While deformable

models can capture significant variations in appearance, a single deformable model is

often not expressive enough to represent a rich set of object category [93]. It can also be noted that in practice simple models generally outperform approaches using

deformable part based representation. The reason being the simpler models can be

trained easily whereas it is more difficult to train more sophisticated models like

DPM. Authors in [93] illustrates that a deformable part-based model represents an

object by a low-resolution root filter and a set of higher-resolution part filters

arranged in a flexible spatial configuration. The flexible spatial configuration helps to

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 119 [email protected]

model the visual appearance at multiple scales. The approach has achieved benchmark

results in the PASCAL object detection challenges. The approach basically uses HOG

(Histogram of gradients) [32] using star-structured part-based model defined by a

filter similar to filter used in [32] and set of part-based filters and deformation models.

The model presented by [101] is effective for shallow structures consisting of at

most two layers, but as the number of layers in the structure increases, it becomes

difficult to scale the model without incorporating and tuning additional parameters.

Yullie et al in [106] have extended the model discussed in [101]. In this paper authors

have proposed that description of object class using several templates from different

viewpoints. Each template is represented as a tree-structure consisting of three layers.

The first layer represents entire image. The second layer divides the image into 9 sub-

images and third layer divides each sub-image of second layer into four sub-images

making third layer of 36 sub-images.

The approach used by Dalal and Triggs [32], to detect pedestrians, fails in

presence of articulation whereas [93, 94, 95, 96] allows an intermediate layer of parts

that can now be shifted with respect to each other making overall model deformable

and in this way achieves generalization. But such approaches do not work when it is a

question of extracting human pose from images. In [102] Bourdev and Malik have

introduced ‘Poselets’ ; parts that are tightly clustered in both appearance and

configuration; for detection and pose estimation of in image consisting human body.

Whereas, in [79] Pablo et al have unified the approaches presented by Dalal and

Triggs [32], Felzenswalb [95] and Bourdev and Malik [102] into a single recognition

framework and tries to take the benefit of each approach. The region-based object

descriptors are used to perform purposive semantic segmentation and subsequently

their outputs are combined and hence performance is achieved.

4.2.4. Recent Approaches and Advancement

We have discussed many approaches with their benefits and limitations in the earlier

sections. One thing can also be noted that all those approaches to object recognition

make essential use of machine learning methods. Most current machine learning

methods work well because of human-designed representations and inputs features.

Early conventional approaches involve hand-crafted features for object representation

and look for these features in image. To do this the programmer was required to have

a deep knowledge of the data and would laboriously engineer each one the feature

detection algorithms [114]. There have been big improvements in image analysis over

the last few years due to the adoption of deep learning neural networks to solve vision

problems. Fig-8 shows schematically the difference between traditional vision

systems and recent deep neural network based system.

Neural Nets for Object Recognition: Neural Networks have been used in object

recognition systems since decades. Neural Nets implement a classification approach.

Their attraction lies in their ability to partition the feature space using nonlinear

boundaries for classes. Earlier Neural Networks were used as classifier only in the

classification stage of Object recognition pipeline (Figure 8), but only recently, with

the progress in vision research and the increase in computational power, neural

networks are utilized for automatic feature learning( from the raw data of the image)

as well as classification also. LeCunn [123] in 1989 demonstrated an algorithm to

train Neural Networks in supervised way and proved applications like hand-written

digit recognition performs remarkably and is benefitted from it. Since then

Convolutional Neural network are being used by many research communities.

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 120 [email protected]

Convolutional Neural Networks are different than conventional approaches like BoF,

DPM (Deformable Part Model).This difference is due to two very important reasons.

First, they are deep architecture whereas the conventional approaches were shallow

architectures. And second they doesn’t need to have prior knowledge of data of

image. Deep learning neural networks made it possible to learn features in an

unsupervised manner directly from data instead of handcrafting them explicitly. The

approach has helped vision tasks particularly object recognition greatly thereby

enabling effective capturing of low-level as well as middle level cues of object to be

recognized. As a result Deep learning Neural Networks have brought huge

improvements in the performance of image analysis results, over the last few years.

What makes deep architectures achieve such a good result?

Conventional neural nets used 1 to 2 layers of neurons whereas Deep Neural

Network” is one class of neural nets that uses deep architectures with 2 to 6 layers of

neurons stacked on top of each other. As a result DNN can learn more complex

models easily without the need of hand-designed features. DNN’s have shown good

results on ImageNet dataset [126]. On the test data authors achieved top-1 and top-5

error rates of 37.5% and 17%. Their neural network consisted of 650,000 neurons and

had 5 convolutional layers and learnt 60 million parameters in ILSVRC 2010.

Like every other approach Deep architecture also has got certain limitations.

Needs very sophisticated hardware and also image of fixed size typically 224 x 224.

Contains huge number of parameters to be trained so computation intensive

When trained using Gradient descent, the gradient does not trickle down to the lower

layers; so the sub-optimal sets of weights are obtained [114].

Various modifications to DNN’s have been suggested in the literature to

overcome these limitations. To overcome the constraint of fixed sized images required

by deep neural networks, several efficient pooling strategies are proposed. In [113],

network is equipped with spatial pooling strategy (SPP-net). SPP-net can generate

fixed length representation irrespective of image scale and size. Spatial pyramid

pooling is based on spatial pyramid matching [24] which in turn is an extension of

BoF [26] approach. Another improvement achieved is RNN (Recursive Neural

Network) [130] used for scene classification. The method predicts tree structure for

scene images.

Figure 8(a). Block diagram representing typical traditional object recognition system

Various competitive challenges and Datasets

Here in this section we present some of the challenges that computer vision

community organize annually to invite, evaluate and report the innovative approaches

developed by research groups all across the world. These challenges serve the purpose

of setting a common platform for researchers to present their work and compete with

each other in the area of Object detection, localization and categorization. These

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 121 [email protected]

challenges also provide dataset of sample images so that the approaches can be

evaluated for all possible image content and variations in image capturing conditions.

Figure 8(b). Block diagram of Deep Architectures (Image taken from:

ufldl.stanford.edu/eccv10-tutorial/eccv10-tutorial_part4.ppt)

Early approaches for object recognition used very small set of images to evaluate

their algorithmic work. But with the advent of sophisticated world-wide-web, large

number of annotated images is readily available in public as well as private

repositories. To harness the benefit of such repositories datasets are created and made

available to the research community. As mentioned earlier these challenges is an

effort to bring together the research community together in a framework of

competition so that best approaches in computer vision can be evaluated and

publicized. These challenges consist of two components. First, a publicly available

dataset with ground truth annotations with standardised evaluation software and

second a competition and workshop [119]. To review these challenges, we first

discuss the datasets made available by these competitions along with certain other

widely used datasets.

Datasets: No research is possible in any research area [30] without appropriate

datasets. The same fact applies to object recognition and computer vision research

also. Appropriate datasets are needed for all stages of recognition research; may be for

learning visual models of objects and scene categories, detecting and localizing

instances of these models in images, and evaluating the performance of recognition

algorithms. Work mentioned in [30] reviews existing Image datasets from the point of

expectation, challenges and limitations. Datasets ideally should offer vide range of

image variability and should be sufficiently challenging so that algorithms can be

evaluated. One of the major limitations in creating such datasets is that images are to

be annotated. This task of annotating has to be done by human experts and turns out

to be mammoth task considering huge number of objects existing in the real world

which are to be recognized for various application and it is not so easy to get the

human experts accomplishing this effectively, correctly and efficiently. A wonderful

approach of automatic dataset collection using web is mentioned in [66], using an

object recognition techniques in incremental method. The images present on web are

used to learn the model in a robust way. Another solution for getting annotated

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 122 [email protected]

training examples can be by crowdsourcing but the most common error that an

untrained annotator is susceptible to is a failure to consider a relevant class as a

possible label because they are unaware of its existence.

Now we discuss certain most prevalent datasets.

Caltech-101 & Caltech-256: Caltech-101 is a collection of pictures of objects

belonging to 101 categories collected by Fei-Fei et al [64] in 2003. About 40 to 800

images exist per category. Most categories have about 50 images. Most images have

little or no clutter. The objects tend to be centered in each image and in stereotypical

pose. In comparison to Caltech-101, Caltech-256 is collection of 256 categories of

objects. Total 30608 images are present. Fig-9 compares Caltech-101 and Caltech-

256.

Figure 9 (Courtesy: http://www.vision.caltech.edu/Image_Datasets/Caltech256/details.html)

TRECVID: TRECVID organizes competition every year and for evaluating the

performance, releases dataset consisting of video shots. The goal of the conference

series is to encourage research in information retrieval by providing a large test

collection, uniform scoring procedures, and a forum for organizations interested in

comparing their results. Annotation is not provided by the organizers.

LabelMe: LabelMe [74] is a publically available annotated image database open for

public contribution. The dataset is provided with annotation tool, so that anyone can

annotate any image. As images are annotated by experts as well as casual users,

cannot be relied for obtaining test set whereas huge quantity of training images can be

obtained.

COIL-20 & COIL-100: Coil-20 and Coil-100 is a database of grayscale images of 20

and 100 categories of objects respectively [120]. Different poses of objects were

generated by placing the objects on rotating turn table and images were captured at

angular displacement of 5 degrees generating 72 views per object. It consists of 720

unprocessed images of 10 object categories. 1440 size normalized images are also

provided.

Microsoft COCO: Common Objects in Context database is a large-scale database of

images that addresses three core research problems in scene understanding: detecting

non-iconic views of objects, contextual reasoning between objects and the precise 2D

localization of objects [122]. Contextual knowledge can be helpful to boost all the

components of the object recognition framework. The dataset is provided to support

object recognition based on the context in which they lie in the scene. The dataset

consist 91 common object categories from which 82 of them having more than 5000

labeled instances. In total dataset have 2,500,000 labeled instances in 328000 images.

The dataset consists of less object categories but very high number of instances per

category that differentiates it with other popular large-scale datasets like PASCAL

VOC and ImageNet dataset which would be discussed in following sections.

The PASCAL Visual Object Classes Challenge

The PASCAL VOC challenge was first time organized in the year 2005. Since then

up to 2012, every year this challenge was organized annually. The challenge basically

consists of two components. A dataset consisting of 1000 images related to objects of

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 123 [email protected]

20 categories; obtained from Flickr web-site; was made available publicly and

competition involving object classification, detection, segmentation, action

classification and person layout. Everingham et al has reviewed PASCAL VOC in

[119]. The objects were fully annotated for each of the objects. Not that in 2005, such

a rich dataset was released. Only dataset consisting of four categories (motorbikes,

bicycles, cars, and people) was made available in 2005, but every year organizers kept

on enriching it and finally in 2011 of 1000 images were released. To assess different

methods bootstrapping of ROC curve is used.

The evaluation technique is used in a number of different ways: to simply judge

the variability for a given method, to compare the relative strength of two methods, or

to look at rank ranges in order to get an overall sense of all methods in a competition

[119].

ImageNet Large Scale Visual Recognition Challenge (ILSVRC): ILSVRC was

first organized in 2010 and since then, the event is organized annually. ILSVRC is

one of the most prestigious series of competition and workshop in computer vision

community to evaluate the performance of all contemporary approaches developed by

various researchers. The challenge from various aspects is nicely reviewed in [118].

Similar to PASCAL VOC, ILSVRC also provides a huge collection of annotated

images under the name ImageNet by Deng. et al.

ImageNet Dataset: ImageNet is an image database organized according to

the WordNet hierarchy in which each node of the hierarchy is depicted by hundreds

and thousands of images. Currently, there exist over five hundred images per node

[121]. In ImageNet, on an average 1000 images to illustrate each synset has been

provided. Images of each concept are quality-controlled and human-annotated.

ILSVRC has dimension very high as compared to PASCAL VOC [121]. As per 2010

data it is organized in form of 12 subtrees with 5247 synsets and 3.2 million images in

total. As ImageNet organization is inspired by WordNet structure and there are

around 80,000 noun synset in WordNet, Similarly ImageNet also aims at providing

nearly all the majority of the 80,000 synsets of WordNet with an average of 500- 1000

clean and full resolution images. To evaluate the approaches effective strategy of

bootstrapping used by PASCAL VOC is employed in ILSVR challenge series also. In

Table 2, we present the comparison between PASCAL VOC challenge dataset and

ILSVRC challenge as per year 2012 as referred from [101].

Table 2 Comparison of PASCAL VOC and ILSVRC as per [101]

Aspect for comparison PASCAL VOC ILSVRC

Diversity of Object classes Objects are only one class

label for e.g Boat for all types

of boat, be it lifeboat, fireboat

Objects are further refined to

subcategories for for e.g boat

is just not boat but lifeboat,

gondola

Chance Performance of

Localization(CPL)

8.8% on validation set for 20

categories

20.8% for all 1000 categories

Average Object Scale /class 0.241 0.358

Average Number of Instances per

class

1.69 1.59

Clutter per class when clutter is

computed as No. of boxes.

129.96 106.98

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 124 [email protected]

In addition to all these, various other datasets are used in literature like GRAZ-01

by Opelt and Pinz, which contains four types of images: bikes, people, background

with no bikes, background with no people. INRIA (people) dataset used by Dalal and

Triggs in their work in [32, 33], MNIST – Dataset of handwritten digits, ImageCLEF,

INRIA (Horses,cars), TinyImages dataset by Torralba et al , ETH-80 etc. CIFAR-10

and CIFAR-100 set has 6000 examples of each of 10 classes and the CIFAR-100 set

has 600 examples of each of 100 non-overlapping classes [125]. The list that we have

considered is not exhaustive but exemplary. For an exhaustive list [127] can further be

explored.

5. FUTURE RESEARCH DIRECTIONS

Object recognition is one of the most exciting research area in the field of computer

vision. Today need is to develop systems which are, computationally efficient and at

the same time cost effective. We suggest some of the future research directions which

can be explored and in turn be incorporated in recognition systems. We attempt to

suggest these directions at algorithmic level or at product level; many of which can be

at present considered as an idea which may need knowledge base from

multidisciplinary fields.

Currently, deep learning is the current state-of-the-art in object recognition and has

produced promising results but they suffer from certain serious limitations of being

resource intensive. So, in absence of sophisticated hardware DNN’s cannot be

adopted for object recognition. In such cases enhancing the performance of

conventional feature extraction techniques on shallow architectures can be helpful.

New approaches are in need which require shallow architectures and are still

efficient. Also it is realised that DNN’s have not shown very impressive results in the

task of object localization, this area can further be explored.

It is also to be understood, how learning of features take place in Convolutional

neural networks? What makes deep architectures giving so high accuracy of

recognition?

Due to advent of mobile and other hand-held devices with very nice image capturing

abilities, algorithms for mobile and other hand-held devices are in great need.

Although considerable work exists in literature related to action recognition systems

and a complete line of research is going on in this direction as the area in itself

involves many and varied issues and research problems Products can be developed

involving action and activity recognition from videos.

Computer vision techniques can be good method to generate assistive technology for

blind people. For example, products can be developed which sees the surrounding

and generates a natural language description of the scene and can be given as output

in spoken form. This will help to understand the surroundings and will help blind

people to navigate.

Research in the area of understanding videos from its content already started, but still

in its infancy, Generic object recognition also paves the path for research in areas like

emotion recognition which will actually enable us to recognize the meaning of the

content in the video.

Robotics is one another important field which can benefit from active object

recognition. Today’s robots are able to work only in well -structured and constrained

environments. Whereas, the requirement is to develop robots which can learn, adapt

and execute their tasks in real human environments.

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 125 [email protected]

Almost every device has a camera and devices are now powerful enough to record

and process live video. These videos can be exploited for real-time applications. How

do we organize and personalize all of this content for common man?

New performance evaluation techniques are needed.

Many rich datasets have been generated and made available to public like ImageNet,

PASCAL VOC by computer vision research community. Although these datasets

holds huge number of images pertaining to various categories, but if we were to reach

to the level of near human vision capabilities in terms of flexibility and dynamism,

then these datasets are to be enriched. Novel ways of labelling huge number of

unlabeled image data should be found so that images annotated with ground truth can

be generated and can be made available publically.

6. CONCLUSION

From the literature available on the subject it was found that the demand for the

efficient Generic object recognition system is increasing very fast as the spectrum of

applications in which the object recognition is needed is very wide and rich. One

major problem in process of Generic Object Recognition is that, categories available

in the real world are varied and huge in number. Due to this fact, the training of the

recognition system for such a large number of categories and classes becomes a

challenging task however sophisticated, the approach may be. Also in such kind of

system the property of plasticity i.e the system should be able to gradually train itself

for unseen categories, is expected which further adds to the complexity of the system.

Such systems can be developed which should be flexible enough to train themselves

for new classes of objects. Another important issue with generic object recognition

system is with the feature extraction and description phase. In most of the approaches,

the number of features obtained is too large and are handcrafted. This very critical

limitation has been overcome by deep architectures which in turn have exploited

sophisticated hardware accelerations evolved recently. Approaches are needed which

make the entire set up cost effective requiring fewer resources. Ideally, it is desirable

that, the recognition task should be performed at semantic level which will result into

near human vision systems.

One of the key objectives behind this survey was to get the answers of the

research questions identified by us and mentioned in Section 1. From the literature

surveyed it can be deduced that, the earlier work related to generic object recognition

were putting more weightage on feature extraction stage and type of features, whereas

the later works were giving more prominence to type of classifier used. Also, recent

approaches are learning features directly from the image data. This can be regarded as

very striking innovation achieved by the vision community. Now the ways are needed

which can bring enhancement to these approaches. The above efforts can also be

extended for 3D images and also for videos.

As a result of this study and from the referred material, a general remark can also

be made about the kind of work that is done in the field. Most of the papers before

2008, mainly present novel ways of modelling the object class. i.e they emphasize on

novel ways of feature detection and descript tion. However, work presented and

published in the recent past since 2011, with the advent of sophisticated hardware,

more emphasis is given on handling more categories accurately and efficiently.

In this paper a current scenario of generic object recognition is portrayed in brief

with a hope that in near future, such an object recognition system will be developed

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 126 [email protected]

which would be capable of performing vision task similar to the human vision system

least possible effort and in a cost effective manner.

REFERENCES

[1] Bennamoun, Mohammed, and George J. Mamic. Object recognition:

fundamentals and case studies. Springer Science & Business Media, 2002.

[2] Takahiro Hori, Tetsuya Takiguchi, Yasuo Ariki. Generic Object Recognition

Using Graph Embedding into a Vector Space, American Journal of Software

Engineering and Applications. Vol. 2, No. 1, 2013, pp. 13-18.

[3] David. G. Lowe, “Object Recognition from Local Scale-Invariant Features”,

Proc. Of the International Conference on Computer Vision, /corfu.(Sept-

1999)

[4] David.G.Lowe,” Distinctive Image Features from Scale-Invariant

Keypoints”, 2004.

[5] Ke, Yan, and Rahul Sukthankar. "PCA-SIFT: A more distinctive

representation for local image descriptors." Computer Vision and Pattern

Recognition, 2004. CVPR 2004. Proceedings of the 2004 IEEE Computer

Society Conference on. Vol. 2. IEEE, 2004.

[6] Bay, Herbert, Tinne Tuytelaars, and Luc Van Gool. "Surf: Speeded up robust

features." Computer Vision–ECCV 2006. Springer Berlin Heidelberg, 2006.

404-417.

[7] B. Leibe, A. Leonardis, and B. Schiele. Combined Object Categorization and

Segmentation with an Implicit Shape Model”, In ECCV04. Workshop on

Stat. Learning in Computer Vision, pages 17–32, May 2004.

[8] Alexander Thomas, Vittorio Ferrari, Bastian Leibe, Tinne Tuytelaars, Bernt

Schiele, Luc Van Gool”, Towards Multi-View Object Class Detection”,

Proceedings of the 2006 IEEE Computer Society Conference on Computer

Vision and Pattern Recognition (CVPR’06)

[9] Bastian Leibe, Aleˇs Leonardis, and Bernt Schiele,” Robust Object Detection

with Interleaved Categorization and Segmentation”, IJCV, (2008) 77,259-

289.

[10] Jia, Menglei , Li, Hua, Xie, Xing , Chen, Zheng Ma, Wei-ying, “Automatic

Classification Of Objects Within IMAGES” United states Microsoft

Corporation (Redmond, WA, US)20080037877

http://www.freepatentsonline.com/y2008/0037877.html

[11] Pisipati, Radha Krishna (Hyderabad, IN), Syed, Shahanaz (Guntur, IN),

Jonna, Kishore (Proddatur, IN), Bandyopadhyay, Subhadip (Kolkata, IN),

Narayan, Rudra Narayan (Jemadeipur, IN) 2014. Systems And Methods

For Multi-Dimensional Object Detection United States 20140029852

http://www.freepatentsonline.com/y2014/0029852.html

[12] Besl, Paul J., and Ramesh C. Jain. "Three-dimensional object recognition."

ACM Computing Surveys (CSUR) 17.1 (1985): 75-145.

[13] Irving Biederman. Recognition-by-components: A theory of human image

understanding. Psychological Review, 94(2):115-147, 1987.

[14] Zhang, Xin, et al. "Object class detection: A survey." ACM Computing

Surveys (CSUR) 46.1 (2013): 10.

[15] Andreopoulos, Alexander, and John K. Tsotsos. "50 Years of object

recognition: Directions forward." Computer Vision and Image Understanding

117.8 (2013): 827-891.

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 127 [email protected]

[16] Roth, Peter M., and Martin Winter. "Survey of appearance-based methods for

object recognition." Inst. for Computer Graphics and Vision, Graz University

of Technology, Austria, Technical Report ICGTR0108 (ICG-TR-01/08)

(2008).

[17] Agarwal, Shivani, Aatif Awan, and Dan Roth. "Learning to detect objects in

images via a sparse, part-based representation." Pattern Analysis and

Machine Intelligence, IEEE Transactions on 26.11 (2004): 1475-1490.

[18] Fergus, Robert, Pietro Perona, and Andrew Zisserman. "Object class

recognition by unsupervised scale-invariant learning." Computer Vision and

Pattern Recognition, 2003. Proceedings. 2003 IEEE Computer Society

Conference on. Vol. 2. IEEE, 2003.

[19] A. Opelt, A. Pinz, and A. Zisserman. A Boundary-Fragment Model For

Object Detection. In Proc. ECCV, volume 2, pp 575–588, May 2006.

[20] Wu, Bo, and Ramakant Nevatia. "Detection of multiple, partially occluded

humans in a single image by bayesian combination of edgelet part

detectors."Computer Vision, 2005. ICCV 2005. Tenth IEEE International

Conference on. Vol. 1. IEEE, 2005.

[21] Andreas Opelt, Axel Pinz,Andrew Zisserman,” Incremental Learning Of

Object Detectors Using A Visual Shape Alphabet”, Proceedings of the 2006

IEEE Computer Society Conference on Computer Vision and Pattern

Recognition (CVPR’06)

[22] Andreas Opelt, Axel Pinz ,Andrew Zisserman,,“Fusing shape and appearance

information for object category detection”, 2006 - eprints.pascal-network.org

[23] Andreas Opelt, Axel Pinz ,Andrew Zisserman, “Learning an Alphabet of

Shape and Appearance for Multi-Class Object Detection”, IJCV (2008) 80:

16–44

[24] Grauman, Kristen, and Trevor Darrell. "Pyramid match kernel and related

techniques." U.S. Patent No. 7,949,186. 24 May 2011.

[25] Zhang, Jianguo, et al. "Local features and kernels for classification of texture

and object categories: A comprehensive study." International journal of

computer vision 73.2 (2007): 213-238.

[26] Lazebnik, Svetlana, Cordelia Schmid, and Jean Ponce. "Beyond bags of

features: Spatial pyramid matching for recognizing natural scene categories."

Computer Vision and Pattern Recognition, 2006 IEEE Computer Society

Conference on. Vol. 2. IEEE, 2006.

[27] Lazebnik, Svetlana, Cordelia Schmid, and Jean Ponce. "Spatial pyramid

matching." Object Categorization: Computer and Human Vision Perspectives

3 (2009): 4.

[28] Svetlana Lazebnik, Cordelia Schmid, and Jean Ponce. “A Discriminative

Framework for Texture and Object Recognition Using Local Image Features.

In Toward Category-Level Object Recognition. Springer-Verlag Lecture

Notes in Computer Science, J. Ponce, M. Hebert, C. Schmid, and A.

Zisserman (eds.), 2006.

[29] Lazebnik, Svetlana. "Local, semi-local and global models for texture, object

and scene recognition." (2006).

[30] J. Ponce, T. L. Berg, M. Everingham, D. A. Forsyth, M. Hebert, S. Lazebnik,

M. Marszalek, C. Schmid, B. C. Russell, A. Torralba, C. K. I. Williams, J.

Zhang, and A. Zisserman. “Dataset Issues in Object Recognition” In Toward

Category-Level Object Recognition. Springer-Verlag Lecture Notes in

Computer Science, J. Ponce, M. Hebert, C. Schmid, and A. Zisserman (eds.),

2006.

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 128 [email protected]

[31] Dorko, Gyuri, and Cordelia Schmid. "Object class recognition using

discriminative local features." (2005): 22.

[32] Dalal, N. And Triggs, B. 2005. Histograms of Oriented Gradients for Human

Detection. In Proceedings of theIEEE Conference on Computer Vision and

Pattern Recognition (CVPR’05).

[33] Dalal,N.,Triggs, B., And Schmid,C. 2006. Human Detection Using Oriented

Histograms Of Flow And Appearance. In Proceedings of the European

Conference on Computer Vision (ECCV’06).

[34] Dalal, Navneet. Finding people in images and videos. Diss. Institut National

Polytechnique de Grenoble-INPG, 2006.

[35] Jurie, Frederic, and Bill Triggs. "Creating efficient codebooks for visual

recognition." Computer Vision, 2005. ICCV 2005. Tenth IEEE International

Conference on. Vol. 1. IEEE, 2005.

[36] Bosch Anna, Andrew Zisserman, and Xavier Munoz. "Representing shape

with a spatial pyramid kernel." Proceedings of the 6th ACM international

conference on Image and video retrieval. ACM, 2007.

[37] Watanabe, Tomoki, Satoshi Ito, and Kentaro Yokoi. "Co-occurrence

histograms of oriented gradients for pedestrian detection." Advances in

Image and Video Technology. Springer Berlin Heidelberg, 2009. 37-47.

[38] JOACHIMS, T. 1997. A probabilistic analysis of the rocchio algorithm with

tfidf for text categorization. In Proceedings of the International Conference

on Machine Learning (ICML’97)

[39] Csurka, Gabriella, et al. "Visual categorization with bags of keypoints."

Workshop on statistical learning in computer vision, ECCV. Vol. 1. No. 1-22.

2004.

[40] Ramanan, Amirthalingam, and Mahesan Niranjan. "A review of codebook

models in patch-based visual object recognition." Journal of Signal

Processing Systems 68.3 (2012): 333-352.

[41] K. Mikolajczyk, C.Schmid,” Indexing based on scale invariant interest

points”, International Conference on Computer Vision (ICCV '01) 1 (2001)

525—531.

[42] K.Mikolajczyk A. Zisserman C. Schmid,” Shape recognition with edge-based

features”, British Machine Vision Conference (BMVC '03) 2 (2003) 779—

788.

[43] K. Mikolajczyk and C. Schmid. “Scale and affine invariant interest point

detectors”. Int. J. Comput. Vision, 60(1):63–86, 2004.

[44] K. Mikolajczyk1, T. Tuytelaars2, C. Schmid4, A. Zisserman ,”A Comparison

of Affine Region Detectors”, International Journal of Computer Vision 65,

1/2 (2005) 43—72

[45] K.Mikolajczyk and C.Schmid,”A performance evaluation of local

descriptors”, IEEE Transactions on Pattern Analysis and Machine

Intelligence 27, 10 (2005) 1615—1630.

[46] Douze, Matthijs, et al. "Evaluation of gist descriptors for web-scale image

search." Proceedings of the ACM International Conference on Image and

Video Retrieval. ACM, 2009.

[47] Jégou, Hervé, Matthijs Douze, and Cordelia Schmid. "Improving bag-of-

features for large scale image search." International Journal of Computer

Vision 87.3 (2010): 316-336.

[48] Tuytelaars, Tinne, and Krystian Mikolajczyk. "Local invariant feature

detectors: a survey." Foundations and Trends® in Computer Graphics and

Vision 3.3 (2008): 177-280.

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 129 [email protected]

[49] K. Mikolajczyk, Bastian Leibe, Bernt Schiele,” Multiple Object Class

Detection with a Generative Model”, Computer Vision and Pattern

Recognition, 2006 IEEE Computer Society Conference on vol-1,pp 26 – 36

[50] Jégou, Hervé, et al. "Aggregating local descriptors into a compact image

representation." Computer Vision and Pattern Recognition (CVPR), 2010

IEEE Conference on. IEEE, 2010.

[51] Wengert, Christian, Matthijs Douze, and Hervé Jégou. "Bag-of-colors for

improved image search." Proceedings of the 19th ACM international

conference on Multimedia. ACM, 2011.

[52] Belongie, S., Malik, J., And Puzicha, J. 2001. “Matching shapes”, In

Proceedings of the IEEE International Conference on Computer Vision

(ICCV’01).

[53] Belongie, Serge, Jitendra Malik, and Jan Puzicha. "Shape matching and

object recognition using shape contexts." Pattern Analysis and Machine

Intelligence, IEEE Transactions on 24.4 (2002): 509-522

[54] Andras Ferencz, Erik G. Learned-Miller, Jitendra Malik,” Building a

Classification Cascade for Visual Identification from One Example”,

Proceedings of the Tenth IEEE International Conference on Computer Vision

(ICCV’05).

[55] Hao Zhang Alexander C. Berg Michael Maire Jitendra Malik,”SVM-

KNN:Discriminative Nearest Neighbor Classification for Visual

Recognition”, Proceedings of the 2006 IEEE Computer Society Conference

on Computer Vision and Pattern Recognition (CVPR’06)

[56] Bjorn Ommer Jitendra Malik,”Multi-Scale Object Detection by Clustering

Lines”, 2009 IEEE 12th International Conference on Computer Vision

(ICCV)pp 484-491

[57] Subhransu Maji, Jitendra Malik,”Object Detection using a Max-Margin

Hough Transform”,IEEE 2009, pp 1038-1045 .

[58] Vidal-Naquet, Michel, and Shimon Ullman. "Object Recognition with

Informative Features and Linear Classification." ICCV. Vol. 3. 2003

[59] Fergus, Robert, Pietro Perona, and Andrew Zisserman. "Object class

recognition by unsupervised scale-invariant learning." Computer Vision and

Pattern Recognition, 2003. Proceedings. 2003 IEEE Computer Society

Conference on. Vol. 2. IEEE, 2003.

[60] Boris Epshtein Shimon Ullman,” Identifying Semantically Equivalent Object

Fragments”, Proceedings of the 2005 IEEE Computer Society Conference on

Computer Vision and Pattern Recognition (CVPR’05)

[61] Eran Borenstein and Shimon Ullman,” Combined Top-Down/Bottom-Up

Segmentation”, IEEE Transactions On Pattern Analysis And Machine

Intelligence, Vol. 30, No. 12, December 2008, pp 2109-2125

[62] L. Fei-Fei, R. Fergus, and P. Perona,”Learning generative visual models from

few training examples: An incremental bayesian approach tested on 101

object categories.” In Proc. CVPR Workshop on Generative-Model Based

Vision, 2004.

[63] Fei-Fei, Li, Rob Fergus, and Pietro Perona. "Learning generative visual

models from few training examples: An incremental bayesian approach tested

on 101 object categories." Computer Vision and Image Understanding 106.1

(2007): 59-70.

[64] www.vision.caltech.edu/Image_Datasets/Caltech101/Caltech101.html

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 130 [email protected]

[65] Fei-Fei, Li, Robert Fergus, and Pietro Perona. "One-shot learning of object

categories." Pattern Analysis and Machine Intelligence, IEEE Transactions

on28.4 (2006): 594-611.

[66] Li-Jia Li, Gang Wang and Li Fei-Fei,” OPTIMOL: Automatic Online Picture

Collection via Incremental Model Learning”, 2007 IEEE

[67] Hao Su, Min Sun, Li Fei-Fei,Silvio Savarese,”Learning a Dense Multi-View

Representation For Detection, Viewpoint Classification And Synthesis Of

Object Categories”, 2009 IEEE 12th International Conference on Computer

Vision (ICCV)

[68] Bangpeng Yao, Li Fei-Fei,” Recognizing Human-Object Interactions in Still

Images by Modeling Mutual Context of Object and Human Pose in Human-

Object Interaction Activities”, Ieee Transactions on Pattern Analysis and

Machine Intelligence, Vol. 34, No. 9, September 2012, pp 1691-1703

[69] Oliva, Aude, and Antonio Torralba. "Building the gist of a scene: The role of

global image features in recognition." Progress in brain research 155 (2006):

23-36.

[70] Antonio Torralba, Kevin P. Murphy and William T. Freeman, “Sharing

Visual Features for Multiclass and Multiview Object Detection”, April 2004.

[71] Antonio Torralba , Kevin P. Murphy, William T. Freeman,” Sharing features:

efficient boosting procedures for multiclass object detection”

[72] Oliva, Aude, and Antonio Torralba. "Modeling the shape of the scene: A

holistic representation of the spatial envelope." International journal of

computer vision 42.3 (2001): 145-175.

[73] Torralba, Antonio, Robert Fergus, and Yair Weiss. "Small codes and large

image databases for recognition." Computer Vision and Pattern Recognition,

2008. CVPR 2008. IEEE Conference on. IEEE, 2008.

[74] BC Russell, A Torralba, KP Murphy,“LabelMe: a database and web-based

tool for image annotation”,International journal of Computer Vision,

2008(77) – Springer, pp-157-173.

[75] Taha H. Rassem, Bee Ee Khoo,” Object Class Recognition using

Combination of Color SIFT Descriptors”, 2011 IEEE

[76] Gyuri Dork_o, Cordelia Schmid,” Object Class Recognition Using

Discriminative Local Features”,Technical Report

[77] Gy. Dorko and C. Schmid. Selection of scale-invariant parts for object class

recognition”. In Proceedings of the Ninth IEEE International Conference on

Computer Vision (ICCV’03), pages 634–639, 2003.

[78] Thomas Serre, Lior Wolf, Stanley Bileschi, Maximilian Riesenhuber, and

Tomaso Poggio,” Robust Object Recognition with Cortex-Like

Mechanisms”, IEEE TRANSACTIONS ON PATTERN ANALYSIS AND

MACHINE INTELLIGENCE, VOL. 29, NO. 3, MARCH 2007 pp 411-426.

[79] B. Mayurathan, A. Ramanan, S. Mahesan & U.A.J. Pinidiyaarachchi,”

Speeded-up and Compact Visual Codebook for Object Recognition”,

International Journal of Image Processing (IJIP), Volume (7): Issue (1): 2013

pp 31-50

[80] Rafael C. Gonzalez and Richard E. Woods. Digital Image Processing.

Prentice-Hall Inc., 2002.

[81] R. Fergus, L. Fei-Fei, P. Perona, and A. Zisserman, “Learning object

Categories from google’s Image Search,” Computer Vision, 2005. ICCV’

2005. Tenth IEEE International Conference on, 2005.

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 131 [email protected]

[82] Joao Carreira and Cristian Sminchisescu, “Constrained Parametric Min-Cuts

for Automatic Object Segmentation”, Computer Vision and Pattern

Recognition (CVPR), 2010 IEEE Conference pp 3241-3248

[83] Van de Sande, K. E. A., Gevers, T., & Snoek, C. G. M. (2008). Evaluation of

color descriptors for object and scene recognition. In Proceedings of the IEEE

conference on computer vision and pattern recognition (CVPR’08)

[84] Uijlings, Jasper RR, et al. "Selective search for object

recognition."International journal of computer vision 104.2 (2013): 154-171.

[85] Van de Sande, Koen EA, et al. "Segmentation as selective search for object

recognition." Computer Vision (ICCV), 2011 IEEE International Conference

on. IEEE, 2011. (selective search) (reviewerI and III)

[86] Fuxin Li_ and Joao Carreira_ and Cristian Sminchisescu, “Object

Recognition as Ranking Holistic Figure-Ground Hypotheses” , Computer

Vision and Pattern Recognition (CVPR), 2010 IEEE Conference , pp 1712 –

1719.

[87] Girshick, Ross, et al. "Rich feature hierarchies for accurate object detection

and semantic segmentation." Computer Vision and Pattern Recognition

(CVPR), 2014 IEEE Conference on. IEEE, 2014.

[88] Carreira, Joao, et al. "Semantic segmentation with second-order pooling."

Computer Vision–ECCV 2012. Springer Berlin Heidelberg, 2012. 430-443.

[89] Carreira, Joao, and Cristian Sminchisescu. "CPMC: Automatic object

segmentation using constrained parametric min-cuts." Pattern Analysis and

Machine Intelligence, IEEE Transactions on 34.7 (2012): 1312-1328.

[90] Li, Fuxin, Joao Carreira, and Cristian Sminchisescu. "Object recognition as

ranking holistic figure-ground hypotheses." Computer Vision and Pattern

Recognition (CVPR), 2010 IEEE Conference on. IEEE, 2010

[91] Fischler, Martin A., and Robert A. Elschlager. "The representation and

matching of pictorial structures." IEEE Transactions on Computers 22.1

(1973): 67-92.

[92] Felzenszwalb, Pedro F., and Daniel P. Huttenlocher. "Pictorial structures for

object recognition." International Journal of Computer Vision 61.1 (2005):

55-79.

[93] Felzenszwalb, Pedro F., et al. "Object detection with discriminatively trained

part-based models." Pattern Analysis and Machine Intelligence, IEEE

Transactions on 32.9 (2010): 1627-1645.

[94] Girshick, Ross B., Pedro F. Felzenszwalb, and D. McAllester.

"Discriminatively trained deformable part models, release 5." (2012).

[95] Felzenszwalb, Pedro, David McAllester, and Deva Ramanan. "A

discriminatively trained, multiscale, deformable part model." Computer

Vision and Pattern Recognition, 2008. CVPR 2008. IEEE Conference on.

IEEE, 2008.

[96] Felzenszwalb, Pedro F., Ross B. Girshick, and David McAllester. "Cascade

object detection with deformable part models." Computer vision and pattern

recognition (CVPR), 2010 IEEE conference on. IEEE, 2010.

[97] Ferrari, Vittorio, Frederic Jurie, and Cordelia Schmid. "Accurate object

detection with deformable shape models learnt from images." Computer

Vision and Pattern Recognition, 2007. CVPR'07. IEEE Conference on. IEEE,

2007.

[98] Pentland, Alex P. "Automatic extraction of deformable part models."

International Journal of Computer Vision 4.2 (1990): 107-126.

Prof. Deepika Shukla and Apurva Desai

http://www.iaeme.com/IJARET/index.asp 132 [email protected]

[99] Pandey, Megha, and Svetlana Lazebnik. "Scene recognition and weakly

supervised object localization with deformable part-based models." Computer

Vision (ICCV), 2011 IEEE International Conference on. IEEE, 2011.

[100] Ren, Xiaofeng, and Deva Ramanan. "Histograms of sparse codes for object

detection." Computer Vision and Pattern Recognition (CVPR), 2013 IEEE

Conference on. IEEE, 2013.

[101] Yang, Yi, and Deva Ramanan. "Articulated pose estimation with flexible

mixtures-of-parts." Computer Vision and Pattern Recognition (CVPR), 2011

IEEE Conference on. IEEE, 2011.

[102] Bourdev, Lubomir, and Jitendra Malik. "Poselets: Body part detectors trained

using 3d human pose annotations." Computer Vision, 2009 IEEE 12th

International Conference on. IEEE, 2009.

[103] Arbeláez, Pablo, et al. "Semantic segmentation using regions and parts."

Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference

on. IEEE, 2012.

[104] Bourdev, Lubomir, et al. "Detecting people using mutually consistent poselet

activations." Computer Vision–ECCV 2010. Springer Berlin Heidelberg,

2010. 168-181.

[105] Arbeláez, Pablo, Bharath Hariharan, Chunhui Gu, Saurabh Gupta, Lubomir

Bourdev, and Jitendra Malik. "Semantic segmentation using regions and

parts." In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE

Conference on, pp. 3378-3385. IEEE, 2012.

[106] Zhu, Long, et al. "Latent hierarchical structural learning for object detection."

Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference

on. IEEE, 2010.

[107] Zeiler, Matthew D., and Rob Fergus. "Visualizing and understanding

convolutional networks." Computer Vision–ECCV 2014. Springer

International Publishing, 2014. 818-833.

[108] Zhao, Wenyi, et al. "Face recognition: A literature survey." Acm Computing

Surveys (CSUR) 35.4 (2003): 399-458.

[109] Yang, Ming-Hsuan, David Kriegman, and Narendra Ahuja. "Detecting faces

in images: A survey." Pattern Analysis and Machine Intelligence, IEEE

Transactions on 24.1 (2002): 34-58.

[110] T.yamazaki,T.Fujikawa,J.Katto,”Improving the performance of SIFT using

Bilateral Filter and its Application to Generic Object Recognition.”, ICASSP

2012, IEEE , pp 945 – 948.

[111] Chiu, Liang-Chi, et al. "Fast SIFT Design For Real-Time Visual Feature

Extraction." Image Processing, IEEE Transactions on 22.8 (2013): 3158-

3167.

[112] Kamencay, Patrik, et al. "Feature extraction for object recognition using

PCA-KNN with application to medical image analysis." Telecommunications

and Signal Processing (TSP), 2013 36th International Conference on. IEEE,

2013.

[113] He, Kaiming, et al. "Spatial pyramid pooling in deep convolutional networks

for visual recognition."arXiv preprint arXiv: 1406.4729 (2014).

[114] Goyal, Soren, and Paul Benjamin. "Object Recognition Using Deep Neural

Networks: A Survey." arXiv preprint arXiv: 1412.3684 (2014).

[115] Nevatia, Ramakant, and Thomas O. Binford. "Description and recognition of

curved objects." Artificial Intelligence 8.1 (1977): 77-98.

Review on Generic Object Recognition Techniques: Challenges and Opportunities

http://www.iaeme.com/IJARET/index.asp 133 [email protected]

[116] Fidler, Sanja, Marko Boben, and Ales Leonardis. "Learning a hierarchical

compositional shape vocabulary for multi-class object representation." arXiv

preprint arXiv: 1408.5516 (2014).

[117] Lee, Tom, Sanja Fidler, and Sven Dickinson. "Multi-cue mid-level

grouping."

[118] Russakovsky, Olga, et al. "Imagenet large scale visual recognition

challenge." arXiv preprint arXiv: 1409.0575 (2014).

[119] Everingham, Mark, et al. "The pascal visual object classes challenge: A

retrospective." International Journal of Computer Vision 111.1 (2014): 98-

136.Fei

[120] Nene, Sameer A., Shree K. Nayar, and Hiroshi Murase. Columbia object

image library (COIL-20). Technical Report CUCS-005-96, 1996.

[121] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, L. Fei-Fei, ImageNet: a large-

scale hierarchical image database, IEEE Computer Vision and Pattern

Recognition, 2009. <http://www.image-net.org/>.

[122] Lin, Tsung-Yi, et al. "Microsoft COCO: Common objects in

context." Computer Vision–ECCV 2014. Springer International Publishing,

2014. 740-755.

[123] LeCun, Yann, et al. "Backpropagation applied to handwritten zip code

recognition." Neural computation 1.4 (1989): 541-551.

[124] Humphrey, Eric J., Juan Pablo Bello, and Yann LeCun. "Moving Beyond

Feature Design: Deep Architectures and Automatic Feature Learning in

Music Informatics." ISMIR. 2012.

[125] Krizhevsky, Alex, and Geoffrey Hinton. "Learning multiple layers of features

from tiny images." Computer Science Department, University of Toronto,

Tech. Rep 1.4 (2009): 7.

[126] Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton. "Imagenet

classification with deep convolutional neural networks." Advances in neural

information processing systems. 2012.

[127] http://riemenschneider.hayko.at/vision/dataset/index.php as referred on 12th

April 2015

[128] http://image-net.org/challenges/LSVRC/2012/analysis/as referred on 12th

April 2015

Zeiler, Matthew D., and Rob Fergus. "Visualizing and understanding

convolutional networks." Computer Vision–ECCV 2014. Springer

International Publishing, 2014. 818-833.

[129] Bengio, Yoshua, et al. "Greedy layer-wise training of deep

networks."Advances in neural information processing systems 19 (2007):

153.

[130] Mrs. Manisha Bhisekar and Prof. Prajakta Deshmane, Image Retrieval and

Face Recognition Techniques: Literature Survey. International Journal of

Electronics and communication Engineering and Technology, 5(1), 2014, pp.

52-58.

[131] Yoel E. Almeida, Ashray S. Bhandare and Aishwary P. Nipane, Computer

Vision Based Adaptive Lighting Solutions for Smart and Efficient System.

International Journal of Computer Engineering and Technology, 6(3), 2015,

pp. 01-11.

[132] Socher, Richard, et al. Parsing natural scenes and natural language with

recursive neural networks. Proceedings of the 28th international conference

on machine learning (ICML-11). 2011.