Professional Documents
Culture Documents
Realtime Visual Recognition in Deep Convolutional Neural Networks
Realtime Visual Recognition in Deep Convolutional Neural Networks
Collecting the data is one task and making that data useful is an-other
vital task.
Data collected from various means will be in an unorganized format and
there may be lot of null values, in-valid data values and unwanted data.
Cleaning all these data and replacing them with appropriate or
approximate data and removing null and missing data and replacing
them with some fixed alternate values are the basic steps in pre
processing of data.
Even data collected may contain completely garbage values. It may not
be in exact format or way that is meant to be. All such cases must be
verified and replaced with alternate values to make data meaning
meaningful and useful for further processing. Data must be kept in a
organized format.
Yolo Algorithm
Yolo is an algorithm that uses convolutional neural
networks for object detection.
In comparison to recognition algorithms, a detection
algorithm does not only predict class labels, but detects
locations of objects as well.
The algorithm divides the image into grids and runs the
image classification and localization algorithm
(discussed under object localization) on each of the grid
cells. For example, we have an input image of size 256 ×
256. We place a 3 × 3 grid on the image .
References
1.P. F. Felzenszwalb, R. B. Girshick, D. McAllester, and D. Ramanan, “Object detection with
discriminatively trained part-based models,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 32, no. 9, pp.
1627 1645, Sep. 2010.
2. B. Leibe, A. Leonardis, and B. Schiele, “Robust object detection with interleaved categorization and
segmentation,” Int. J. Comput. Vis., vol. 77, nos. 1-3, pp. 259-289, May 2008.
3. J. Zhang, M. Marszalek, S. Lazebnik, and C. Schmid, “Local features and kernels for classi cation of
texture and object categories: A comprehensive study,” in Proc. Conf. Comput. Vis. Pattern
Recognit.Workshop (CVPRW), Jun. 2006, p. 13.
4. P. Viola and M. Jones, “Rapid object detection using a boosted cascade of simple features,” in Proc.
IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. (CVPR), vol. 1. Dec. 2001, pp. 511-518.
5. M. Weber, M. Welling, and P. Perona, “Towards automatic discovery of object categories,” in Proc.
IEEE Conf. Comput. Vis. Pattern Recognit., vol. 2. Jun. 2000, pp. 101-108.
6. A. Ayvaci and S. Soatto, “Detachable object detection: Segmentation and depth ordering from short-
baseline video,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 34, no. 10, pp. 1942 1951, Oct. 2012.
7. D. Liu, M.-L. Shyu, Q. Zhu, and S.-C. Chen, “Moving object detection under object occlusion
situations in video sequences,'' in Proc. IEEE Int. Symp. Multimedia (ISM), Dec. 2011, pp. 271-278.
8. J. Kim, G. Ye, and D. Kim, ``Moving object detection under free-moving camera,'' in Proc. 17th IEEE
Int. Conf. Image Process. (ICIP), Sep. 2010, pp. 4669-4672.
9.B. Qi, M. Ghazal, and A. Amer, ``Robust global
motion estimation oriented to video object
segmentation,'' IEEE Trans. Image Process., vol. 17,
no. 6, pp. 958 967, Jun. 2008.
10. S. Kumar and M. Hebert, ``A hierarchical eld
framework for unified context-based classi cation,'' in
Proc. 10th IEEE Int. Conf. Comput. Vis. (ICCV), vol.
2. Oct. 2005, pp. 1284-1291.