You are on page 1of 4

2016 World Conference on Futuristic Trends in Research and Innovation for Social Welfare (WCFTR’16)

Object Recognition for Blind people Using


Portable Camera
Vikky Mohane Prof. Chetan Gode
Dept. of Electronics & Tel., YCCE, Dept. of Electronics & Tel., YCCE,
Nagpur, India. Nagpur, India.
vikkymohane@gamil.com chetan_gode@rediffmail.com

Abstract- For blind persons it is very essential to recognize The task of finding and identifying objects is still a
a product of their daily use so we implied a method to identify challenge for computer vision systems. It is noted that
product in their everyday routine by use of camera. To approach used for object recognition is feature extraction.
separate a object from un-necessary background, movement
based technique is used to spot object of concern from the
video by instructing person to shake the object that gives
A. Features extraction
moving object area by K-means clustering background When we used features instead of raw pixel data
separation technique. After detection of moving area feature classification becomes easier [10]. Visual substitution
extraction SIFT technique is used to take out key points of that systems generally exploit a single camera to capture image
product and that key-point is matched with the data base data. Further processing for Recognition is then processed
images if it matches efficiently then it identify a product and based on different features extracted from that data.
produce name of that product to blind persons in speech. Features extraction is the process by which interesting parts
Instead of OCR we used SIFT because in OCR orientation of of an image are detected and represented for further
camera must be towards label of product and for blind person processing. It is important step in most computer vision and
it is very tough to locate label efficiently whereas in SIFT
product orientation is not a big deal as key-points can be
image processing solutions because it marks the transition
matched from any position. from pictorial to non pictorial (alphanumerical, usually
quantitative) data representation [11]. Types of features
Index terms-Blind, Feature extraction, OCR, SIFT. extracted from image depend on the type of image (e.g.
binary, gray-level, or color), the level of granularity
I. INTRODUCTION (complete image or individual regions) desired, and the
According to the World Health Organization, 314 context of the application. The features which are extracted
million people in the world are visually impaired and out of are usually represented in an alphanumerical way for further
them 15% blind. 13% of visually impaired people reside in processing. The actual representation based on the technique
developed countries. Recently, several electronic assisting used.
devices [2], [3], [4], [5], [6], [7], [8] & [9] developed for
these persons but they cannot handle product recognition B. Features extraction based SIFT
efficiently. For example, moveable bar code readers used to Scale Invariant Features Transform approach [12] was
assists visually impaired people to recognize different designed to match pictures or objects of a different scene.
products but to find the position of the bar code and to point The features are partially invariant to illumination and 3D
the bar code reader at the bar code efficiently is not easy for camera viewpoint and invariant to scale and rotation. There
blind people. are two important steps to extract SIFT features: localization
The majority of existing systems are for mobility and key-points by extreme scale space detection and generation
obstacle detection but still blind people face big problem to of descriptors of points of interest. SIFT algorithm is applied
identify of their environment and its components [10]. to object recognition. First, key-points are extracted from a
Identifying and locating objects is the need for mobility of reference image and stored in the database. Then objects in
blind people by travel assistance and aid to navigation. For the target image are compared with the stored references
this reason, we were interested in the evaluation of an based on Euclidean distance between features vectors. The
efficient algorithm for fast and Robust computer vision to object is identified by the matching between key-points.
identify and locate objects in a video scene .We aim to In principle, SIFT works like a multi-scale corner
describe our method [10] to identify the nearby object for detector with sub-pixel positioning accuracy and a rotation-
blind people by proposing an approach for analyzing video invariant features descriptor attached to each candidate point
scenes based on feature extraction. [1].
SIFT features carry with the location of key-points,
II. OBJECTS RECOGNITION USING FEATURES scale, and orientation. Lowe uses a four-dimensional
EXTRACTION similarity transformation as the original Hough banning
structure, thus each bin denotes a particular location for the
object center, scale, and in-plane rotation. Each matching

978-1-4673-9214-3/16/$31.00 © 2016 IEEE


2016 World Conference on Futuristic Trends in Research and Innovation for Social Welfare (WCFTR’16)
feature votes for the nearest 24 bins and peaks in the The basic effect of the operator on a binary image is to
transform are then selected for a more careful affine motion gradually broaden the boundaries of regions of foreground
fit [13]. pixels. Thus areas of foreground pixels widen in size while
holes within those regions will reduce in size. After dilation,
III. ALGORITHM OF PROPOSED SYSTEM: to separate object from un-necessary background, movement
based technique is used to identify object of concern from
The figure above shows the algorithm of our proposed the video by instructing person to shake the object that gives
system our proposed system use a lone camera to capture moving object area by K-means clustering background
images of the view in front of the user. Different features separation technique. In this technique n observations are
are then extracted from those images and objects in the separated into k clusters in which each observation belongs
scene are recognized by comparing these features with to the cluster with the nearest mean, allocated as element of
those of known database objects. the cluster. After foreground and background separation of
Before applying SIFT algorithm to the image, an image, in segmented filled image filling holes in binary
preprocessing of image takes place. It includes various steps image takes place. The shapes of input image are expanded
are as shown in Figure 1. by the dilation operation using structuring element and
In image acquisition, the camera captures scenes which after dilation noise is removed by removing small objects
contains region of interest in the form of images. RGB that is noise then with the processing of segmented noise
information is extracted from an image to apply proper removed image, processed Image is generated. Features are
median filtering on an image. To eliminate noise median extracted from the processed image which is used for
filtering is used. As a pre-processing step median filtering matching with data base images. If key points are matched
is used for better results of advance processing. Median then identify the name of product by auditory system
filtering conserves edges while eliminate noise and hence otherwise detected as unknown object by our system.
it has significant application in digital image processing.
The dilation is performed over medium filtered image. IV. EXPERIMENTAL RESULTS

In our system we first check the algorithm over images


to check efficiency of the algorithm and then we apply it on
video scene to recognize objects from video scene. To
extract features SIFT method is used. In Figure 2 image
acquisition takes place from the camera.

Fig. 2. Input image from camera

Fig. 1. Our proposed approach for object recognition for blind people
2016 World Conference on Futuristic Trends in Research and Innovation for Social Welfare (WCFTR’16)
Data base images shown in Figure 3, after
preprocessing and foreground and background separation
SIFT algorithm is performed and key-point of input image is
drawn and that is matched with the features of database
image after successful matching image recognition takes
place. Distinct features can be find by SIFT even in the
small objects for efficient recognition of object we have to
improve the rate of object recognition. For efficient
matching we will use SIFT in object recognition.
In the above operation 564 key-points found, object
detected as bourn vita and processing time is 5.35 ms
similarly various objects are recognized using same
algorithm efficiently.

V. CONCLUSION
In recent years, some solutions have been devised to
help blind people or visual impaired in recognizing objects
in their environment But it is not that efficient to recognize
surrounding objects our purpose is to provide a robust and
easy system for blind people to recognize their surrounding
objects. Our proposed system uses a single camera to
capture images of the view in front of the user. Different
features are then extracted from those images and objects in
the scene are recognized by comparing these features with
those of known database objects. We have evaluated SIFT
Fig. 3 . Database images algorithm for feature extraction. Our future work consists
on to increase the object recognition rate, to use our system
for different application like obstacle detection and to find
distance information between camera and object.

REFERENCES

[1] HanenJabnoun, FaouziBenzarti,and Hamid Amiri


“Object recognition for blind people based on features
extraction”In INTERNATIONAL IMAGE PROCESSING
APPLICATIONS AND SYSTEMS CONFERENCE 2014
[2] B. Durette, N. Louveton, D. Alleysson, and J. H´erault.
”Visuo-auditorysensory substitution for mobility
assistance: testing The VIBE”. InWorkshop on Computer
Vision Applications for the Visually Impaired,Marseille,
France, 2008.
Fig. 4. Processed image over which sift performed [3] A. F. R. Hern´andez et al. ”Computer Solutions on
Sensory Substitution for Sensory Disabled People”. In
Proceedings of the 8th WSEAS International Conference
on Computational Intelligence, Man-machine Systems
and Cybernetics, pages 134–138, 2009.
[4] H. Tang and D. J. Beebe. ”An oral tactile interface for
blind navigation”. IEEE Trans Neural SystRehabilEng,
14(1):116–123, March 2006.
[5] M. Auvray, S. Hanneton, and J. K. O’Regan. ”Learning
to perceive with a visuo - auditory substitution system:
Localisation and object recognition with ’The vOICe’ ”.
Perception, 36:416–430, 2007.
[6] S. Kammoun et al. ”Navigation and space perception
assistance for the visually impaired: The NAVIG
project”. IRBM, Numrospcial ANRTECSAN, 33(2):182–
189, April 2012.
[7] F.G. Brian Katz et al. ”NAVIG: Guidance system for the
Fig. 5. SIFT operation visuallyimpaired using virtual augmented reality”.
Technology and Disability, 24(2):163–178, June 2012.
2016 World Conference on Futuristic Trends in Research and Innovation for Social Welfare (WCFTR’16)
[8] S. Abboud, S. Hanassy, S. Levy-Tzedek, S. description”. In International Conferenceof Soft
Maidenbaum, and A. Amedi. ”EyeMusic: Introducing a Computing and Pattern Recognition. Tunsia, pages 300–
”visual” colorful experience for the blind using auditory 305,August 2014.
sensory substitution”. Restorative neurology and [11] L. Juan and O. Gwon. ”A Comparison of SIFT, PCA-
neuroscience, January 7 2014. SIFT and SURF”.International Journal of Image
[9] B.D.C. Martnez, O.O. Vergara-Villegas, V.G.C. Snchez, Processing (IJIP), 3(4):143–152, 2009
H. de JessOchoaDomnguez, and L.O. Maynez. ”Visual [12] G.D. Lowe. ”Distinctive Image Features from Scale-
Perception Substitutionby the Auditory Sense”. In Invariant Keypoints”.Int. J. Comput. Vision, 60(2):91–
BeniaminoMurgante, Osvaldo Gervasi, Andrs Iglesias, 110, November 2004.
David Taniar, and Bernady O. Apduhan, editors, ICCSA [13] R. Szeliski. ”Computer Vision: Algorithms and
(2), volume 6783 of Lecture Notes in Computer Science, Applications”. Springer-Verlag New York, Inc., New
pages 522–533. Springer, 2011. York, NY, USA, 2010.
[10] H. Jabnoun, F. Benzarti, and H. Amiri. ”Visual
substitution system for blind people based on SIFT

You might also like