[go: up one dir, main page]

WO2015030689A2 - A tool and method for robust, scale and orientation invariant object detection and classification - Google Patents

A tool and method for robust, scale and orientation invariant object detection and classification Download PDF

Info

Publication number
WO2015030689A2
WO2015030689A2 PCT/TR2014/000318 TR2014000318W WO2015030689A2 WO 2015030689 A2 WO2015030689 A2 WO 2015030689A2 TR 2014000318 W TR2014000318 W TR 2014000318W WO 2015030689 A2 WO2015030689 A2 WO 2015030689A2
Authority
WO
WIPO (PCT)
Prior art keywords
hog
unit
mlp
camera
hog feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/TR2014/000318
Other languages
French (fr)
Other versions
WO2015030689A3 (en
Inventor
Halis ALTUN
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Publication of WO2015030689A2 publication Critical patent/WO2015030689A2/en
Publication of WO2015030689A3 publication Critical patent/WO2015030689A3/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/50Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis

Definitions

  • the invention relates to shape detection method and system and more particularly to automatic shape detection method and system that detects the shape of an object in an image acquired by a camera.
  • Object detection is a method which is used in various industrial application and therefore a plenty of methods have been proposed for a solution to this problem.
  • a method should be robust against to noise and illumination effects, and should be invariant to scale and orientation of the object, due to possible various poses and position of the object in front of a camera.
  • various approaches have been advised to tackle these problems.
  • Recognizing an object automatically in an image using features related to its shape is mostly a fundamental step in the automatic detection and classification of the objects. Therefore shape detection plays an important role and has an intensive usage in various applications.
  • the adverse effects of illumination and noise are alleviated using edge information.
  • a normalization scheme is proposed to tackle with the problems related to different sizes of the object in an image. This will be a case for example if the distance of the object from the camera is not fixed.
  • an alignment scheme is proposed to get rid of the orientation problem due to different position of the object in an image.
  • HOG Histogram of Oriented Gradient
  • AMDF average magnitude difference function
  • the obtained HOG features present a waveform which presents a circular periodicity within 0-360 degree.
  • the invention uses an AMDF module, which is first time in literature, to find out the rotation angle present in the image of the object by comparing the current waveform (i.e HOG feature vector) to the waveforms obtained from the original template objects and stored in memory.
  • a score which indicates the similarity between the original waveforms and present waveform of the given object, and the rotation angle between two of them are provided by the AMDF module. Based on these information, an alignment stage is utilized to remove the rotation from object shape before directing the features of the object to the recognition module.
  • Multilayer Pereptron type Artificial Neural Network (ANN) is proposed as a classifier in the recognition module. ANN is trained using the original reference waveforms belongs to the shapes to be determined.
  • the present invention provides an object detection and classification method and system.
  • the invention requires acquiring an image by a camera, wherein the object will be detected automatically by the proposed method.
  • the image will go under a preprocessing stage to provide smoothed version of the acquired image.
  • the smoothed image will be used to extract feature vectors based on the edge information.
  • the obtained feature vector will be robust against illumination and noise.
  • the feature vectors will further go under a process to make them robust against scale and orientation of object in the image.
  • the present invention eliminates the adverse effect of illumination and noise by relying on the edge information. Furthermore, in order to alleviate the problems due to orientation the present invention suggest to use AMDF as an indication of the degree of rotation of the object in the image; and in order to alleviate the problems due to scale and size variation, the present invention suggest to use a normalization of HOG vectors based on an average value of HOG feature vectors.
  • a new method is advised for alignment of HOG feature vectors in order to correctly classify the object, which might lays in different positions and orientation in front of a camera.
  • the stored HOG vectors which belong to the pre-determined objects, and which is already obtained and stored in the memory, is compared to the HOG vector, which is extracted from the current input image using AMDF module.
  • the AMDF module will provide a score as a degree of similarity between the current HOG feature vector and the HOG feature vectors stored in the memory. It is easy to show that the orientation of the object in the current image will introduce a shift on the HOG feature vectors.
  • HOG feature vector is sent to the recognition unit which consists of artificial neural network as a classifier.
  • a multilayer perceptron (MLP) neural network will classify the aligned HOG feature vector.
  • MLP NN multilayer perceptron neural network
  • the number of output of MLP NN unit will equal to the number of shapes to be detected. If the present shape is recognized successfully the corresponding output of the MLP NN will be activated. For example, in the case of successful recognition, if the present object belong to the third class, the third output of MLP NN will be activated while the rest of the output will be de-activated.
  • the detected label information and the location information of the object will be sent to a robot which will perform the defined operation on the detected object accordingly.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Length Measuring Devices By Optical Means (AREA)

Abstract

The invention relates to shape detection method and system and more particularly to automatic shape detection method and system that detects the shape of an object in an image acquired by a camera.

Description

DESCRIPTION
A TOOL AND METHOD FOR ROBUST, SCALE AND ORIENTATION
INVARIANT OBJECT DETECTION AND CLASSIFICATION TECHNICAL FIELD
The invention relates to shape detection method and system and more particularly to automatic shape detection method and system that detects the shape of an object in an image acquired by a camera. PRIOR TECHNIQUE
Object detection is a method which is used in various industrial application and therefore a plenty of methods have been proposed for a solution to this problem. For a successful solution, a method should be robust against to noise and illumination effects, and should be invariant to scale and orientation of the object, due to possible various poses and position of the object in front of a camera. In literature, various approaches have been advised to tackle these problems. In the present invention, we propose new robust and fast approach.
BRIEF SUMMARY OF THE INNOVATION
Recognizing an object automatically in an image using features related to its shape is mostly a fundamental step in the automatic detection and classification of the objects. Therefore shape detection plays an important role and has an intensive usage in various applications.
In the invention, the adverse effects of illumination and noise are alleviated using edge information. Furthermore, a normalization scheme is proposed to tackle with the problems related to different sizes of the object in an image. This will be a case for example if the distance of the object from the camera is not fixed. Also an alignment scheme is proposed to get rid of the orientation problem due to different position of the object in an image.
In the invention, Histogram of Oriented Gradient (HOG) algorithm is proposed to extract shape related features using edge information. The average magnitude difference function (AMDF) is employed to obtain rotation invariant shape detection algorithm. AMDF is a dissimilarity function which commonly used in Speech Analysis to provide a similarity between waveforms. In the invention, the obtained HOG features present a waveform which presents a circular periodicity within 0-360 degree. The invention uses an AMDF module, which is first time in literature, to find out the rotation angle present in the image of the object by comparing the current waveform (i.e HOG feature vector) to the waveforms obtained from the original template objects and stored in memory. A score which indicates the similarity between the original waveforms and present waveform of the given object, and the rotation angle between two of them are provided by the AMDF module. Based on these information, an alignment stage is utilized to remove the rotation from object shape before directing the features of the object to the recognition module. In the invention, Multilayer Pereptron type Artificial Neural Network (ANN) is proposed as a classifier in the recognition module. ANN is trained using the original reference waveforms belongs to the shapes to be determined.
DETAILED DESCRIPTION OF THE INNOVATION
The present invention provides an object detection and classification method and system. The invention requires acquiring an image by a camera, wherein the object will be detected automatically by the proposed method. The image will go under a preprocessing stage to provide smoothed version of the acquired image. The smoothed image will be used to extract feature vectors based on the edge information. The obtained feature vector will be robust against illumination and noise. Also the feature vectors will further go under a process to make them robust against scale and orientation of object in the image.
In the proposed methods, HOG features are obtained from the current edge information which would produced by one of the edge operators available in literature.
The present invention eliminates the adverse effect of illumination and noise by relying on the edge information. Furthermore, in order to alleviate the problems due to orientation the present invention suggest to use AMDF as an indication of the degree of rotation of the object in the image; and in order to alleviate the problems due to scale and size variation, the present invention suggest to use a normalization of HOG vectors based on an average value of HOG feature vectors.
In the present invention, a new method is advised for alignment of HOG feature vectors in order to correctly classify the object, which might lays in different positions and orientation in front of a camera. In the proposed method, the stored HOG vectors, which belong to the pre-determined objects, and which is already obtained and stored in the memory, is compared to the HOG vector, which is extracted from the current input image using AMDF module. The AMDF module will provide a score as a degree of similarity between the current HOG feature vector and the HOG feature vectors stored in the memory. It is easy to show that the orientation of the object in the current image will introduce a shift on the HOG feature vectors. If a method would be able to remove this shift on the HOG feature vector, then a high score of similarity will be obtained between the current HOG feature vector and one of the template HOG feature vectors. On the operation, AMDF will introduce a shift on the current HOG feature, between 0 and 360, and then calculate the degree of similarity for each steps. Therefore the highest degree of similarity will occurs at the point if the same amount of shift is introduced, which will eventually indicate the degree of shift on the current HOG feature vector, which should be removed to have an orientation invariant method. After the alignment process, HOG feature vector is sent to the recognition unit which consists of artificial neural network as a classifier.
In the next step a multilayer perceptron (MLP) neural network (NN) will classify the aligned HOG feature vector. As MLP NN is trained with various samples of the object, a high success score will be obtained even under noise, occlusion and deformation of the shape of the object. The number of output of MLP NN unit will equal to the number of shapes to be detected. If the present shape is recognized successfully the corresponding output of the MLP NN will be activated. For example, in the case of successful recognition, if the present object belong to the third class, the third output of MLP NN will be activated while the rest of the output will be de-activated. The detected label information and the location information of the object will be sent to a robot which will perform the defined operation on the detected object accordingly.

Claims

1. An object detection and classification system based on the shape of the object which compares a set of features stored in the memory with the object in an image taken by a camera; and consist of
- The Histogram of Oriented Gradient (HOG) unit which extracts oriented gradients from the image taken by a camera
- The average magnitude difference function (AMDF) unit which find the degree of dissimilarity between two HOG feature vectors
- The multilayer perceptron (MLP) neural network (NN) classifier unit which detects and classifies
2. The Histogram of Oriented Gradient (HOG) mentioned in claim 1 wherein a normalized HOG feature vectors by the HOG unit is extracted in a way that alleviates the problems
- caused by the fact that the size of the object in the capture image could be different from its original size due to the non- fixed distance between the object and the the camera; or
- caused by the fact that the objects might be produced with different scale and size.
3. The average magnitude difference function (AMDF) mentioned in claim 1 wherein the degree of dissimilarity is calculated by AMDF unit using
- The HOG features of the object in the current image acquired by the camera which will be extracted and provided by the preceding HOG unit
- The set of pre-stored HOG feature vectors which belongs to the each of reference objects
4. The average magnitude difference function (AMDF) mentioned in claim 1 wherein the degree of dissimilarity is calculated by AMDF unit in a way that provides a maximum similarity possible between two HOG vectors by removing the shift in the feature vector extracted from the image of the object due to the orientation of the object in the image.
5. The multilayer perceptron (MLP) neural network (NN) mentioned in claim 1 wherein the multilayer perceptron (MLP) neural network (NN) classifier unit has as many output as number of the possible classes and each of output identifies and labels the corresponding class of the object detected.
6. The multilayer perceptron (MLP) neural network (NN) mentioned in claim 1 wherein the multilayer perceptron (MLP) neural network (NN) classifier unit will set only one of its output an active output, which indicates the detected class of the object, and the rest of the outputs will be deactivated.
7. The method for robust, scale and orientation invariant object detection and classification which operates using the steps of:
- capturing the image of the object by a camera
- extracting the HOG feature vectors from the edge information obtained by one of the available edge detection algorithm
- alleviating the effect of the illumination and noise by using the edge information obtained by one of the available edge detection algorithm
- normalizing the HOG feature vectors in order to remove the problems caused by the fact that the objects might be produced with different scale and size or the distance between the camera and the object is not fixed.
- finding the amount of shift caused by orientation of the object by calculating the degree of similarity between two HOG feature vectors by the AMDF unit; one HOG feature vector that is obtained from the image acquired by the camera and one HOG feature vector which stored in the memory available on the system for each of reference objects
- aligning the HOG feature vector by removing the present shift and sending the aligned HOG feature vector to the multilayer perceptron (MLP) neural network (NN) classifier unit setting only the corresponding output of the multilayer perceptron (MLP) neural network (NN) classifier unit as an active output in order to indicate the detected object
Sending the label of the detected object and the location information to a robot which performs pre-defined operation on the detected objects accordingly.
PCT/TR2014/000318 2013-08-27 2014-08-27 A tool and method for robust, scale and orientation invariant object detection and classification Ceased WO2015030689A2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
TR2013/10136 2013-08-27
TR201310136 2013-08-27

Publications (2)

Publication Number Publication Date
WO2015030689A2 true WO2015030689A2 (en) 2015-03-05
WO2015030689A3 WO2015030689A3 (en) 2015-04-23

Family

ID=52146638

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/TR2014/000318 Ceased WO2015030689A2 (en) 2013-08-27 2014-08-27 A tool and method for robust, scale and orientation invariant object detection and classification

Country Status (1)

Country Link
WO (1) WO2015030689A2 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017161167A1 (en) * 2016-03-18 2017-09-21 President And Fellows Of Harvard College Automatically classifying animal behavior
CN108647665A (en) * 2018-05-18 2018-10-12 西安电子科技大学 Vehicle real-time detection method of taking photo by plane based on deep learning
CN108960310A (en) * 2018-06-25 2018-12-07 北京普惠三农科技有限公司 A kind of agricultural pest recognition methods based on artificial intelligence
US11020025B2 (en) 2015-10-14 2021-06-01 President And Fellows Of Harvard College Automatically classifying animal behavior
US11263444B2 (en) 2012-05-10 2022-03-01 President And Fellows Of Harvard College System and method for automatically discovering, characterizing, classifying and semi-automatically labeling animal behavior and quantitative phenotyping of behaviors in animals

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5786495B2 (en) * 2011-06-30 2015-09-30 富士通株式会社 Image recognition apparatus, image recognition method, and computer program for image recognition

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
None

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11263444B2 (en) 2012-05-10 2022-03-01 President And Fellows Of Harvard College System and method for automatically discovering, characterizing, classifying and semi-automatically labeling animal behavior and quantitative phenotyping of behaviors in animals
US11020025B2 (en) 2015-10-14 2021-06-01 President And Fellows Of Harvard College Automatically classifying animal behavior
US11622702B2 (en) 2015-10-14 2023-04-11 President And Fellows Of Harvard College Automatically classifying animal behavior
US11944429B2 (en) 2015-10-14 2024-04-02 President And Fellows Of Harvard College Automatically classifying animal behavior
WO2017161167A1 (en) * 2016-03-18 2017-09-21 President And Fellows Of Harvard College Automatically classifying animal behavior
US10909691B2 (en) 2016-03-18 2021-02-02 President And Fellows Of Harvard College Automatically classifying animal behavior
US11669976B2 (en) 2016-03-18 2023-06-06 President And Fellows Of Harvard College Automatically classifying animal behavior
CN108647665A (en) * 2018-05-18 2018-10-12 西安电子科技大学 Vehicle real-time detection method of taking photo by plane based on deep learning
CN108960310A (en) * 2018-06-25 2018-12-07 北京普惠三农科技有限公司 A kind of agricultural pest recognition methods based on artificial intelligence

Also Published As

Publication number Publication date
WO2015030689A3 (en) 2015-04-23

Similar Documents

Publication Publication Date Title
CN108235770B (en) Image recognition method and cloud system
CN111428731B (en) Multi-category identification positioning method, device and equipment based on machine vision
CN103824091B (en) A kind of licence plate recognition method for intelligent transportation system
Nalepa et al. Fast and accurate hand shape classification
Mccullagh Face detection by using haar cascade classifier
WO2015165365A1 (en) Facial recognition method and system
US10311287B2 (en) Face recognition system and method
US8090151B2 (en) Face feature point detection apparatus and method of the same
WO2015030689A2 (en) A tool and method for robust, scale and orientation invariant object detection and classification
Muralidharan et al. Combining local and global feature for object recognition using SVM-KNN
KR20170108339A (en) Method for recognizing plural object in image
CN109558773B (en) Information identification method and device and electronic equipment
Khurana et al. Static hand gestures recognition system using shape based features
CN109426769A (en) The iris identification method and iris authentication system of face auxiliary
Hamouz et al. Affine-invariant face detection and localization using gmm-based feature detector and enhanced appearance model
Shitole et al. Recognition of handwritten devanagari characters using linear discriminant analysis
Kumari et al. A machine learning algorithm for automatic number plate recognition
Shariat et al. A new adaptive segmental matching measure for human activity recognition
Bousmaha et al. YOLOv7-face: A real-time face detector
JP5171362B2 (en) Strong classifier learning apparatus and method, and strong classifier learning program
Ahilapriyadharshini et al. Weber local descriptor based object recognition
Zainuddin et al. The Performance of Face Recognition Using the Combination of Viola-Jones, Local Binary Pattern Histogram and Euclidean Distance
Vijayarani et al. Facial feature extraction based on FPD and GLCM algorithms
CN105574494A (en) Multi-classifier posture identification method and method
Singh et al. Detecting face region in binary image

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14819095

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205 DATED 02/06/2016)

122 Ep: pct application non-entry in european phase

Ref document number: 14819095

Country of ref document: EP

Kind code of ref document: A2