Professional Documents
Culture Documents
Abstract—We develop a novel camera-based computer vision are identical in size and color, and inaccessible to people who
technology to automatically recognize banknotes to assist visu- are blind or significantly visually impaired. Although a federal
ally impaired people. Our banknote recognition system is robust judge has ruled that the U.S. Treasury Department is violating
and effective with the following features: 1) high accuracy: high
true recognition rate and low false recognition rate; 2) robust- the law by failing to design and issue currency that is readily dis-
ness: handles a variety of currency designs and bills in various tinguishable to blind and visually impaired people, in November
conditions; 3) high efficiency: recognizes banknotes quickly; and 2006, the Treasury Department argued that any change would
4) ease of use: helps blind users to aim the target for image cap- be costly, with estimates ranging from $75 million for equip-
ture. To make the system robust to a variety of conditions includ- ment and $9 million in annual expenses to punch holes, to $178
ing occlusion, rotation, scaling, cluttered background, illumination
change, viewpoint variation, and worn or wrinkled bills, we pro- million in one-time charges and $50 million in annual expenses
pose a component-based framework by using speeded up robust to print different-size bills [20]. Some latest redesigned money
features (SURF). Furthermore, we employ the spatial relationship has been issued with an enlarged color number for visually im-
of matched SURF features to detect if there is a bill in the camera paired people. However, it may take years to issue bills with
view. This process largely alleviates false recognition and can guide additional blind-friendly changes.
the user to correctly aim at the bill to be recognized. The robust-
ness and generalizability of the proposed system are evaluated on According to the American Foundation for the Blind [1], one
a dataset including both positive images (with U.S. banknotes) and way that a blind person can identify paper currency is to fold
negative images (no U.S. banknotes) collected under a variety of each denomination in different ways. The recommendation for
conditions. The proposed algorithm achieves 100% true recogni- folding some currency is to fold $5 bills lengthwise, $10 bills
tion rate and 0% false recognition rate. Our banknote recognition widthwise, $20 bills are folded twice, and $1 bills may remain
system is also tested by blind users.
unfolded and put into a wallet. Although the idea of folding
Index Terms—Banknote recognition, blind and visually the bills is good, it needs others’ help to organize bills. Recent
impaired, component based, computer vision, speeded up robust advances in theories, sensors, and embedded computing hold
feature (SURF).
the promise to enable computer vision technique to address
their needs. Although a number of studies on camera-based bill
I. INTRODUCTION recognition have been published in pieces of literature [3], [9],
ORLD Health Organization approximates that there [10], [13], [16]–[18], [31]–[33], some of them are restricted to
W were 161 million visually impaired people around the
world in 2002, which is about 2.6% of the total population.
specific and standard environment. For example, the whole bill
must be visible without occlusion, wrinkles, etc. Furthermore,
Among these statistics, 124 million had low vision and 37 mil- none of the existing methods provides the analysis of recogni-
lion were blind [15]. Visually impaired people face a number tion of negative images (i.e., images do not contain banknotes)
of challenges when interacting with the environments because and aiming guidance for blind users. An automatic banknote
so much information is encoded visually in daily life. One spe- recognition system for blind people should be able to recognize
cific difficulty that a blind person would encounter is to know banknotes in a wide variety of real-world environments, such as
the value of the currency or bill he or she is holding. Currently, occlusions, cluttered background, varied illumination, different
printed denominations of U.S. currency are $1, $2, $5, $10, $20, viewpoints, and worn or wrinkled bills, as well as to provide the
$50, and $100. With a few recent exceptions, all of the banknotes feedback of aiming guidance.
The extraction of sufficient, stable, and distinctive monetary
features is significant for accuracy and robustness of a banknote
recognition algorithm. Recent developments of interest points
Manuscript received June 6, 2011; revised September 17, 2011; accepted detectors and descriptors in computer vision, such as scale in-
November 26, 2011. Date of publication January 18, 2012; date of current ver- variant feature transform (SIFT) [19] and speeded up robust
sion October 12, 2012. This work was supported in part by the National Science
Foundation under Grant IIS-0957016, by the Emerging Frontiers in Research feature (SURF) [2], enable the extraction of sufficient, stable,
and Innovation under Grant 1137172, and by the National Institutes of Health and distinctive recognition of monetary characteristics by using
under Grant 1R21EY020990. This paper was recommended by Associate Editor local image features. The sophisticated designs of local im-
L. Zhang.
The authors are with The City College of New York, New York, NY age features make them robust to geometric and photometric
10031 USA (e-mail: fhasanu00@ccny.cuny.edu; xyang02@ccny.cuny.edu; variations, such as rotation, scaling, viewpoint variation, and
ytian@ccny.cuny.edu). illumination change. These properties make them well adapted
Color versions of one or more of the figures in this paper are available online
at http://ieeexplore.ieee.org. to detect objects appearing with different scales and orientations
Digital Object Identifier 10.1109/TSMCC.2011.2178120 in images.
TABLE I
CONFUSION MATRIX TO MATCH RESULTS OF LOCAL IMAGE
FEATURES EXTRACTED BY SURF
TABLE II
PROCEDURE OF BANKNOTE RECOGNITION
Fig. 6. Computation of purity value in each cluster. The blue points matched
to reference region a are the dominant points in cluster S i . The purity value of
each cluster is the number of dominant points with respect to the total number
of points in this cluster.
A. Experiment Setup
The banknote dataset is collected from a wide variety of
circumstances. Fig. 7 demonstrates four sample images from
Fig. 7. Training and testing dataset of banknote images taken under differ-
each condition. The samples from the first row to the sixth row ent conditions and negative images: first row (partial occlusion), second row
correspond to the conditions of partial occlusion, cluttered back- (cluttered background), third row (rotation), fourth row (scaling change), fifth
ground, rotation, scaling, lighting change, and wrinkling. The row (illumination change), sixth row (wrinkling), and seventh row (negative
images).
seventh row displays four sample images of negative images.
In some cases, some of the conditions are combined into one TABLE III
condition, such as scaling and cluttering. The dataset will be RECOGNITION RESULTS FOR SEVEN CATEGORIES OF BANKNOTES
released to public at our website. In real-world applications, the
user may take the banknote in his/her hand which can corre-
spond to the “both occluded and wrinkled” case. The user may
also take a negative image with no bill due to the aiming prob-
lem. The banknote dataset that is presented in our experiment
is more challenging than that from other banknote recognition
papers. For example, the dataset in [32] was collected by us-
ing a scanner to scan the bills that were taken under restricted
or standard conditions. Therefore, our dataset generalizing the
conditions of taking banknote images is more challenging and
more closely approximates the real-world environment. In the
experiments, we randomly select 140 positive images and 50
negative images as training set, which are used to obtain the au-
tomatic thresholds for reference regions. The remaining positive the proposed algorithm achieves 100% true recognition accu-
and negative images are used as a testing set. racy for all seven categories and 0% false recognition rate.
Table III shows the evaluation results for each category. The
experimental results have shown the effectiveness of the mon-
B. Experimental Results and Discussion
etary features extracted by SURF and our component-based
Each category of bill images covers all of the conditions framework for banknote recognition.
of partial occlusion, cluttered background, rotation, scaling Although our algorithm is evaluated in a more challenging
change, and illumination change, as well as wrinkling. In the dataset, our algorithm achieves much better recognition results
recognition experiments that are evaluated in our testing dataset, and outperforms the existing banknote recognition algorithms.
1028 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART C: APPLICATIONS AND REVIEWS, VOL. 42, NO. 6, NOVEMBER 2012
our system treats images in both cases as negative recognitions [12] M. Inaba, N. Katoh, and H. Imai, “Applications of weighted Voronoi
and sends feedback to blind subjects to guide users to adjust diagrams and randomization to variance-based K-clustering,” in Proc.
ACM 10th Annu. Symp. Comput. Geom., 1994.
their image capture. Since blind users can be collaborative, they [13] N. Jahangir and A. Chowdhury, “Bangladeshi banknote recognition by
will be able to capture good quality images by system feedback neural network with axis symmetrical masks,” in Proc. IEEE 10th Int.
and a simple user training process. Conf. Comput. Inf. Technol., Dec. 2007, pp. 1–5.
[14] Y. Ke and R. Sukthankar, “PCA-SIFT: A more distinctive representation
for local image descriptors,” in Proc. IEEE Comput. Soc. Conf. Comput.
VII. CONCLUSION AND FUTURE WORK Vis. Pattern Recognit, 2004, pp. II-506–II-513.
[15] I. Kocur, R. Parajasegaram, and G. Pokharel, “Global data on visual im-
In this paper, we have proposed a component-based frame- pairment in the year 2002,” Bull. World Health Org., vol. 82, pp. 844–851,
2004.
work for banknote recognition using SURF. Patches with de- [16] T. Kosaka and S. Omatu, “Bill money classification by competitive learn-
scriptive features for each banknote category are selected as ing,” in IEEE Midnight-Sun Workshop Soft Comput. Methods Ind. Appl.,
reference regions to match with query images. The proposed al- 1999, pp. 5–9.
[17] T. Kosaka, S. Omatsu, and T. Fujinaka, “Bill classification by using the
gorithm has been evaluated by dataset to a variety of conditions LVQ method,” in Proc. IEEE Int. Conf. Syst., Man, Cybern., 2001, vol. 3,
including occlusion, rotation, scaling, cluttered background, il- pp. 1430–1435.
lumination change, viewpoint variation, and worn or wrinkled [18] J. Lee, S. Jeon, and H. Kim, “Distinctive point extraction and recognition
algorithm for various kinds of euro banknotes,” Int. J. Control, Autom.,
bills, and further by blind subjects. Our approach achieves 100% Syst., vol. 2, no. 2, pp. 201–206, 2004.
true recognition rate and 0% negative recognition rate. [19] D. Lowe, “Distinctive image features from scale-invariant keypoints,” Int.
In the future work, we will address the issue of motion blur J. Comput. Vis., vol. 60, pp. 91–110, 2004.
[20] C. Lupsa. (2006). Should U.S. bills be blind friendly? The Christian
in two perspectives: 1) integrate deblurring techniques [43], Science Monitor [Online]. Available: http://www.csmonitor.com/2006/
[44] into our system; 2) provide a simple training process to 1207/p17s02-lign.html
familiarize blind users with the device to reduce motion blur. Our [21] R. Manduchi, J. Coughlan, and V. Ivanchenko, “Search strategies of visu-
ally impaired persons using a camera phone wayfinding system,” presented
future work will also focus on evaluating the proposed system at the 11th Int. Conf. Comput. Helping People Spec. Needs, Linz, Austria,
to recognize banknotes of different countries and transferring 2008.
the method to mobile phones. [22] R. Manduchi, S. Kurniawan, and H. Bagherinia, “Blind guidance us-
ing mobile computer vision: A usability study,” in Proc. 12th Int. ACM
SIGACCESS Conf. Comput. Accessibility, 2010, pp. 241–242.
ACKNOWLEDGMENT [23] R. Manduchi and S. Kurniawan, “Mobility-related accidents experienced
by people with visual impairment,” Insight: Res. Practice Visual Impair-
The authors would like to thank Dr. A. Arditi for the re- ment Blindness., vol. 4, no. 2, pp. 44–54, 2011.
cruitment of blind subjects to conduct the algorithm and system [24] K. Mikolajczyk and C. Schmid, “A performance evaluation of local de-
scriptors,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 27, no. 10,
evaluation. The authors also would like to thank the anony- pp. 1615–1630, Oct. 2005.
mous reviewers for their constructive comments and insightful [25] National Research Council, Electronic Travel Aids: New Directions for
suggestions that improved the quality of this paper. Research (Working Group on Mobility Aids for the Visually Impaired and
Blind Committee on Vision), Washington, DC: National Academy Press,
1986, p. 107.
REFERENCES [26] Note Teller 2. (2011). [Online]. Available: http://www.brytech.com
[27] T. Reiff and P. Sincak, “Multi-agent sophisticated system for intelli-
[1] American Foundation for the Blind. (2011). [Online]. Available: http:// gent technologies,” IEEE 6th Int. Conf. Comput. Cybern., Nov. 2008,
www.afb.org/ pp. 37–40.
[2] H. Bay, T. Tuytellars, and L. Gool, “SURF: Speeded up robust features,” [28] C. Schimd and R. Mohr, “Local gray value invariants for image retrieval,”
in Proc. 9th Eur. Conf. Comput. Vis., 2006, pp. 404–417. IEEE Trans. Pattern Anal. Mach. Intell., vol. 19, no. 5, pp. 530–535, May
[3] S. Chae, J. Kim, and S. Pan, “A study on the Korean banknote recognition 1997.
using RGB and UV information,” Commun. Comput. Inf. Sci., vol. 56, [29] H. Shen, K. Chan, J. Coughlan, and J. Brabyn, “A mobile phone system
pp. 477–484, 2009. to find crosswalks for visually impaired pedestrians,” Technology and
[4] D. Dakopoulos and N. G. Bourbakis, “Wearable obstacle avoidance elec- Disability, vol. 20, pp. 217–224, 2008.
tronic travel aids for blind: A survey,” IEEE Trans. Syst., Man, Cybern. [30] H. Shen and J. Coughlan, “Grouping using factor graphs: An approach
C, Appl. Rev., vol. 40, no. 1, pp. 25–35, Jan. 2010. for finding text with a camera phone,” in Proc. Workshop Graph-based
[5] A. Frosini, M. Gori, and P. Priami, “A neural network-based model for Represent. Pattern Recognit, 2007, pp. 394–403.
paper currency recognition and verification,” IEEE Trans. Neural Netw., [31] F. Takeda and S. Omatu, “High speed paper currency recognition by neural
vol. 7, no. 6, pp. 1482–1490, Nov. 1996. networks,” IEEE Trans. Neural Netw., vol. 6, no. 1, pp. 73–77, Jan. 1995.
[6] N. Giudice and G. Legge, “Blind navigation and the role of technology,” [32] F. Takeda, L. Sakoobunthu, and H. Satou, “Thai banknote recognition
in The Engineering Handbook of Smart Technology for Aging, Disability, using neural network and continues learning by DSP unit,” in Proc. 7th
and Independence, A. A. Helal, M. Mokhtari, and B. Abdulrazak, Eds. Int. Conf. Knowledge-Based Intell. Inf. Eng. Syst., 2003, pp. 1169–1177.
Hoboken, NJ: Wiley, 2008. [33] F. Takeda and T. Nishikage, “Multiple kinds of paper currency recognition
[7] C. Harris and M. Stephens, “A combined corner and edge detector,” in using neural network and application for Euro currency,” in Proc. IEEE-
Proc. Fourth Alvey Vis. Conf., Manchester, U.K, 1988, pp. 147–151. INNS-ENNS Int. Joint Conf. Neural Netw., 2000, pp. 143–147.
[8] C. Harris, “Geometry from Visual Motion,” in Active Vision. Cambridge, [34] E. Tekin, J. Coughlan, and H. Shen, “Real-time detection and reading of
MA: MIT Press, 1993. LED/LCD displays for visually impaired persons,” in Proc. IEEE Work-
[9] H. Hassanpour and P. Farahabadi, “Using hidden Markov models for paper shop Appl. Comput. Vis., 2011, pp. 491–496.
currency recognition,” Expert Syst. Appl., vol. 36, pp. 10105–10111, 2009. [35] Y. Tian, C. Yi, and A. Arditi, “Improving computer vision-based indoor
[10] F. Hasanuzzaman, X. Yang, and Y. Tian, “Robust and effective component- wayfinding for blind persons with context information,” in Proc. 12th Int.
based banknote recognition by SURF features,” in Proc. 20th Annu. Wire- Conf. Comput. Helping People Spec. Needs, 2010, pp. 255–262.
less Opt. Commun. Conf., 2011, pp. 1–6. [36] Y. Tian, X. Yang, and A. Arditi, “Computer vision-based door detection
[11] B. Hodges, D. Burton, and M. Uslan, “Show me the money: Access issues: for accessibility of unfamiliar environments to blind persons,” in Proc.
An evaluation of the note teller 2 money identifier. AFB AccessWorld: 12th Int. Conf. Comput. Helping People Spec. Needs, 2010, pp. 263–270.
Technology and People Who Are Blind or Visually Impaired,” vol. 8, [37] A. Vila, N. Ferrer, J. Mantecon, D. Breton, and J. Garcia, “Development of
no. 3, May 2007. a fast and non-destructive procedure for characterizing and distinguishing
1030 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART C: APPLICATIONS AND REVIEWS, VOL. 42, NO. 6, NOVEMBER 2012
original and fake euro notes,” Analytica Chimica Acta, vol. 559, pp. 263– Xiaodong Yang received the B.S. degree from the
273, 2006. Huazhong University of Science and Technology,
[38] X. Yang, Y. Tian, C. Yi, and A. Arditi, “Context-based indoor object Wuhan, China, in 2009. He is currently working to-
detection as an aid to blind persons accessing unfamiliar environments,” ward the Ph.D. degree in electrical engineering with
Int. Conf. ACM Multimedia, 2010, pp. 1087–1090. The City College of New York, New York.
[39] S. Yuan, Y. Tian, and A. Arditi, “Clothing matching for visually impaired His current research focuses on recognition of
persons,” Technol. Disability, vol. 23, pp. 75–85, 2011. objects, texture, and scene categories. His research
[40] E. Zhang, B. Jiang, J. Duan, and Z. Bian, “Research on paper currency interests include image processing, object detection,
recognition by neural networks,” in Proc. Int. Conf. Mach. Learning and developing effective image representations for
Cybern., 2003, pp. 2193–2197. recognition.
[41] R. Huber, H. Ramoser, K. Mayer, H. Penz, and M. Rubik, “Classification
of coins using an eigenspace approach,” Pattern Recognit. Lett., vol. 26,
pp. 61–75, 2005.
[42] L. Shen, S. Jia, Z. Ji, and W. Chen, “Extracting local texture features
for image based coin recognition,” IET Image Process., vol. 5, no. 5,
pp. 394–401, Aug. 2011.
[43] W. Dong, L. Zhang, G. Shi, and X. Wu, “Image Deblurring and super-
resolution by adaptive sparse domain selection and adaptive regulariza- YingLi Tian (M’99–SM’01) received the B.S. and
tion,” IEEE Trans. Image Process., vol. 20, no. 7, pp. 1838–1857, Jul. M.S. degrees from Tianjin University, Tianjin, China,
2011. in 1987 and 1990, respectively, and the Ph.D. degree
[44] A. Beck and M. Teboulle, “Fast gradient-based algorithms for constrained from the Chinese University of Hong Kong, Hong
total variation image denoising and deblurring problems,” IEEE Trans. Kong, in 1996.
Image Process., vol. 18, no. 11, pp. 2419–2434, Nov. 2009. After holding a faculty position with the
[45] S. Shoval, J. Borenstein, and Y. Koren, “Auditory guidance with the National Laboratory of Pattern Recognition, Chinese
Navbelt-a computerized travel aid for the blind,” IEEE Trans. Syst., Man, Academy of Sciences, Beijing, China, she joined
Cybern. C, Appl. Rev., vol. 28, no. 3, pp. 459–467, Aug. 1998. Carnegie Mellon University in 1998, where she was a
[46] A. Huete, J. Victores, S. Martinez, A. Gimenez, and C. Balaguer, “Per- Postdoctoral Fellow with the Robotics Institute. From
sonal autonomy rehabilitation in home environment by a portable assistive 2001 to 2008, she was a Research Staff Member with
robot,” IEEE Trans. Syst., Man, Cybern. C, Appl. Rev., no. 99, pp. 1–10, IBM T. J. Watson Research Center. She is one of the inventors of the IBM
2011. Smart Surveillance Solutions. In September 2008, she joined the Department
[47] N. Aim and J. Arnott, “Computer-assisted conversation for nonvocal peo- of Electrical Engineering, City College of New York, New York, where she
ple using prestored texts,” IEEE Trans. Syst., Man, Cybern. C, Appl. Rev., is currently an Associate Professor. Her current research interest focuses on a
vol. 28, no. 3, pp. 318–328, Aug. 1998. wide range of computer vision problems from motion detection and analysis to
human identification, facial expression analysis, and video surveillance.