Professional Documents
Culture Documents
Senior Design Project 2
Senior Design Project 2
1
ﻣﻠﺨﺺ اﻟﺮﺳﺎﻟﺔ
ﺗﺸﻴﺮ اﻹﺣﺼﺎﺋﻴﺎت إﻟﻰ أن ﻋﺪد اﻟﻤﻨﺪرﺟﻴﻦ رﺳﻤﻴﴼ ﺗﺤﺖ ﺗﺼﻨﻴﻒ ذوي اﻹﻋﺎﻗﺔ اﻟﺒﺼﺮﻳﺔ ﻳﺼﻞ ﺗﻘﺮﻳﺒﴼ
إﻟﻰ 258ﻣﻠﻴﻮن ﺣﺎﻟﺔ ﻓﻲ اﻟﻌﺎﻟﻢ39 ،ﻣﻠﻴﻮن ﻣﻨﻬﻢ ﻓﺎﻗﺪو اﻟﺒﺼﺮ ﺗﻤﺎﻣﴼ ﺑﻴﻨﻤﺎ 248ﻣﻠﻴﻮن ﻣﻨﻬﻢ
ﻳﻤﺘﻠﻜﻮن ﻧﺴﺒﺔ ﺑﺼﺮ ﺿﺌﻴﻠﺔ ﺟﺪﴽ .ﺑﺎﻟﺘﻤﻌﻦ ﻓﻲ اﻷرﻗﺎم و ﻓﻲ اﻟﺤﺎﻻت اﻟﻤﺮﺿﻴﺔ ﻣﻦ ﺣﻮﻟﻨﺎ ﻧﺠﺪ أن
اﻟﻐﺎﻟﺒﻴﺔ اﻟﻌﻈﻤﻰ ﻫﻢ ﻣﻦ أﺻﺤﺎب اﻟﻨﻈﺮ اﻟﻀﻌﻴﻒ ،اﻟﺬﻳﻦ ﻗﺪ ﻳﻜﻮﻧﻮ ﻗﻠﻴﻠﻲ ﺣﻆ ﺑﺴﺒﺐ اﻟﺪﻣﺞ ﻏﻴﺮ
اﻟﻤﻨﺼﻒ ﻓﻲ اﻟﻤﺴﻤﻴﺎت ﻧﻈﺮﴽ ﻹﻣﻜﺎﻧﻴﺔ ﺗﻮﻓﻴﺮ ﺣﻠﻮل واﺳﻌﺔ ﻟﻤﺸﻜﻠﺘﻬﻢ اﻟﺘﻲ ﺗﺨﺘﻠﻒ ﺗﻤﺎﻣﺎ ﻋﻦ
ﻓﺎﻗﺪي اﻟﺒﺼﺮ و ﺑﺎﻟﺘﺎﻟﻲ ﺗﻮﻓﻴﺮ ﻧﻤﻂ ﻣﻦ اﻟﺤﻴﺎة أﺳﻬﻞ ﺑﻜﺜﻴﺮ ﻳﺨﺘﻠﻒ ﺗﻤﺎﻣﴼ ﻋﻤﺎ ﻫﻢ ﻋﻠﻴﻪ ،ﻣﻤﺎ
ﻳﺴﺎﻫﻢ ﻓﻲ ﺗﻌﺰﻳﺰ ﺛﻘﺘﻬﻢ ﺑﺎﻟﻨﻔﺲ و اﻹﺣﺴﺎس ﺑﺤﻴﺎة ﻧﻮﻋﴼ ﻣﺎ أﻗﺮب ﻟﻠﻄﺒﻴﻌﻴﺔ .ﻋﺪة ﺣﻠﻮل ﻇﻬﺮت و
ﺑﺸﻜﻞ ﻣﺨﺘﻠﻒ ﺗﻌﺎﻟﺞ ﺷﺘﻰ ﻣﺸﻜﻼت اﻟﺘﺄﻗﻠﻢ ،و ﻟﻜﻦ ﺣﺘﻰ اﻵن ﻻ ﻳﻮﺟﺪ ﺣﻞ ﺟﺬري ﻣﻠﻤﻮس.
ﻳﻬﺪف ﺗﺼﻤﻴﻢ NOORإﻟﻰ ﺗﺴﻬﻴﻞ ﺣﻴﺎة اﻟﻤﺴﺘﺨﺪم و زﻳﺎدة ﻛﻔﺎءﺗﻪ ،دﻗﺘﻪ وﺳﺮﻋﺔ ﺣﺮﻛﺘﻪ ﺣﻴﻦ
ﻳﻨﺘﻘﻞ ﻓﻲ ﺑﻴﺌﺔ ﻣﻌﻴﻨﺔ ﻣﺠﻬﻮﻟﺔ اﻟﻤﻌﺎﻟﻢ ﺑﺎﻟﻨﺴﺒﺔ ﻟﻪ ،ﻋﻦ ﻃﺮﻳﻖ اﺳﺘﺨﺪام ﻛﺎﻣﻴﺮا ﻣﺜﺒﺘﺔ ﻋﻠﻰ ﻫﻴﻜﻞ
اﻟﻨﻈﺎرة ﺗﻘﻮم ﺑﺘﺼﻮﻳﺮ اﻟﻤﺸﺎﻫﺪ ﺛﻢ ﻣﻌﺎﻟﺠﺘﻬﺎ ﻟﺘﻨﺎﺳﺐ اﻟﻤﺴﺘﺨﺪم و أﺧﻴﺮﴽ إرﺳﺎﻟﻬﺎ إﻟﻰ ﺷﺎﺷﺔ ﺗﻘﻊ
ﻋﻠﻰ ﻣﻘﺮﺑﺔ ﻣﻦ ﻋﻴﻦ اﻟﻤﺴﺘﺨﺪم ،و ﺑﺬﻟﻚ ﺳﺘُﺘﺎح ﻟﺬوي اﻟﺒﺼﺮ اﻟﻤﺤﺪود ﺟﺪﴽ ﻓﺮﺻﺔ اﻟﺘﻌﺮف ﻋﻠﻰ
اﻟﻌﺎﻟﻢ ﻋﻦ ﻗﺮب ﺑﺸﻜﻞ أﻛﺒﺮ .أﺛﺒﺘﺖ ﻧﺘﺎﺋﺞ اﻻﺧﺘﺒﺎرات اﻟﺘﻲ أﻗﻴﻤﺖ ﻋﻠﻰ ذوي اﻟﻨﻈﺮ اﻟﻤﺤﺪود أن
اﻻﺳﺘﺨﺪام اﻟﻤﺘﻜﺮر ل NOORو ﻣﻊ اﻟﺘﻌﻮد ﻋﻠﻴﻬﺎ ﻳﺴﺎﻋﺪ اﻟﻤﺴﺘﺨﺪم ﻓﻲ اﺟﺘﻴﺎز اﻷﺷﻴﺎء ﻣﻦ ﺣﻮﻟﻪ
دون اﻻﺻﻄﺪام ﺑﻬﺎ و ﺑﺎﻟﺘﺎﻟﻲ ﻳﻌﺰز ﻟﺪﻳﻪ اﻟﻘﺪرة ﻋﻠﻰ اﻻﻋﺘﻤﺎد ﻋﻠﻰ ﻧﻔﺴﻪ ﻓﻲ اﻟﺘﻨﻘﻞ ﻣﻦ ﻣﻜﺎن
ﻵﺧﺮ.
2
Abstract
Estimated statistics show that 285 million people around the world are considered
blind; 246 million of them remain very low vision while 39 are totally blind, and as noticed,
the majority of them are with low-vision. People under this classification may be oppressed,
knowing that they have a better chance to improve their quality of life, and therefore,
provide a better lifestyle which contributes to build up a stronger self-confidence, while the
totally blind people have much narrowed options. Many researches with different solutions
have been done, but yet, no radical solution can cure and fix the various cases. The design
of the proposed smart glass aims to facilitate the users’ life and increase their efficiency,
reliability, and mobility while moving through an unknown environment. This is done by
placing a stereo-vision camera on the body of the glass that captures the surrounding
environment, processes it, and sends it to a screen that is close to the patient’s eyes. By
that, they will have the opportunity to experience the world in an exceptional way.
Participants were able to walk through an unknown environment with different set-up
mazes independently, where the results showed that the time taken to complete a set-up
maze decreased logarithmically, number of obstacle collisions decreased linearly, and the
user's velocity increased linearly.
3
Table Of Content
Chapter 1 ............................................................................................................................................................... 9
1 Introduction ................................................................................................................................................. 9
1.1 Background .......................................................................................................................................... 9
1.2 Cases ....................................................................................................................................................... 9
1.3 Thesis objective ............................................................................................................................... 10
1.4 Thesis organization ........................................................................................................................ 11
Chapter 2 ............................................................................................................................................................. 12
2 Terminology ............................................................................................................................................... 12
2.1 Introduction ...................................................................................................................................... 12
2.2 Technical terminology .................................................................................................................. 12
2.2.1 Stereovision ................................................................................................................................ 12
2.2.2 Depth map .................................................................................................................................. 13
2.2.3 Back-illumination sensor ........................................................................................................ 14
2.3 Medical terminology ...................................................................................................................... 15
2.3.1 Retinitis Pegmentosa (RP) ..................................................................................................... 15
2.3.2 Glaucoma .................................................................................................................................... 16
2.3.3 Cataracts ...................................................................................................................................... 17
2.3.4 Diabetic Retinopathy ............................................................................................................... 18
2.3.5 Age-related macular degeneration .................................................................................... 19
Chapter 3 ............................................................................................................................................................. 21
3 Related work .............................................................................................................................................. 21
3.1 Introduction ...................................................................................................................................... 21
3.2 Ultrasonic sensor ............................................................................................................................ 21
3.3 Infrared distance sensor .............................................................................................................. 22
3.4 Assistance aids ................................................................................................................................. 22
3.4.1 Smart cane .................................................................................................................................. 23
Figure 3.2 : The Blindsport cane structure [22] ............................................................................ 23
3.4.2 Applications with sensors ...................................................................................................... 23
3.5 Technical designs in the markets ............................................................................................. 24
3.5.1 Google glass ................................................................................................................................ 24
3.5.2 OrCam ........................................................................................................................................... 25
3.5.3 Depth based smart glasses .................................................................................................... 26
Chapter 4 ............................................................................................................................................................. 28
4 Synthesis ...................................................................................................................................................... 28
4.1 Introduction ...................................................................................................................................... 28
4.2 Depth cameras .................................................................................................................................. 28
4.2.1 Orbbec .......................................................................................................................................... 29
4.2.2 Microsoft Kinect ........................................................................................................................ 31
4.2.3 Asus Xtion Pro ............................................................................................................................ 33
4.2.4 ZED ................................................................................................................................................. 34
4.2.5 Alternatives comparison ........................................................................................................ 37
4.3 Jetson TK1 .......................................................................................................................................... 37
4.4 Displays ............................................................................................................................................... 39
4.4.1 OLED .............................................................................................................................................. 39
4.4.2 TFT ................................................................................................................................................. 40
4
Choice 2: ..................................................................................................................................................... 41
4.5 Battery ................................................................................................................................................. 42
4.6 Conclusion .......................................................................................................................................... 42
Chapter 5 ............................................................................................................................................................. 43
5 Proposed design ....................................................................................................................................... 43
5.1 Proposed solution ........................................................................................................................... 43
5.2 Hardware design ............................................................................................................................. 44
5.3 Software design ............................................................................................................................... 45
5.3.1 ZED SDK ........................................................................................................................................ 46
5.4 Testing methods .............................................................................................................................. 53
5.4.1 Study 1: Obstacle avoidance test on healthy people with normal vision ............. 53
5.4.2 Study 2: Obstacle avoidance test on low-vision people .............................................. 56
Chapter 6 ............................................................................................................................................................. 58
6 Constraints and specifications ........................................................................................................... 58
6.1 Introduction ...................................................................................................................................... 58
6.2 List of constraints ........................................................................................................................... 58
6.2.1 Environmental concerns ........................................................................................................ 58
6.2.2 Ethical and social issues .......................................................................................................... 58
6.2.3 Economic issues ........................................................................................................................ 58
6.2.4 Limitations and restrictions .................................................................................................. 59
6.3 List of specifications ...................................................................................................................... 59
6.3.1 Parts list and estimated budget ........................................................................................... 59
4755 ...................................................................................................................................................................... 59
Chapter 7 ............................................................................................................................................................. 61
7 Conclusion .................................................................................................................................................. 61
7.1 Thesis summary .............................................................................................................................. 61
7.2 Difficulties .......................................................................................................................................... 61
7.3 Future work ....................................................................................................................................... 62
5
List of Tables
Table 4.1: The specifications of Orbbec Persee ...................................................................... 30
Table 4.2: Orbbec Astro Pro Specifications ............................................................................ 30
Table 4.3: Microsoft kinect specifications. ............................................................................. 32
Table 4.4: Asus Xtion specifications ........................................................................................ 33
Table 4.5: The output resolution of ZED ................................................................................. 34
Table 4.6: Physical specifications of ZED ................................................................................ 35
Table 4.7: Alternatives main differences ................................................................................ 37
Table 4.8: Specifications of Jetson TK1 .................................................................................. 38
Table 4.9: 1.66" flexible OLED specifications .......................................................................... 40
Table 4.10: TFT LCD model technical specifications [42] ........................................................ 40
Table 4.11: Adafruit display specifications ............................................................................. 42
Table 5.1: Key shortcuts to control different characteristics of depth image ........................ 48
Table 5.2: Low-vision participants details .............................................................................. 56
Table 6.1: Parts List and Estimated Budget ............................................................................ 59
6
List of figures
Figure 1.1: the block diagram of the entire system design .................................................... 10
Figure 2.1: Disparity map generation . ................................................................................... 13
Figure 2.2: Triangulation diagram .......................................................................................... 14
Figure 2.3: Normal Retina (Left) vs. Retinitis Pegmentosa (Right) ......................................... 15
Figure 2.4: The developing stage of the vision for a person with glaucoma ......................... 17
Figure 2.5: blurry and indistict image results while having a cataract .................................. 18
Figure 2.6: A person with Diabetic Retinopathy will see blurry and distorted images (left) . 19
Figure 2.7: AMD cause a blind spots and blurry image (Right) .............................................. 20
Figure 3.1: Seeedstudio Ultrasonic ........................................................................................ 21
Figure 3.2 : The Blindsport cane structure ............................................................................. 23
Figure 3.3: Prototype of gloves with sensors ......................................................................... 24
Figure 3.4: Google glass .......................................................................................................... 25
Figure 3.5: OrCam ................................................................................................................... 25
Figure 3.6 : A) Prof. Stephen Hicks mart glasses with a laptop that should be carried all the
time. B) Different mazes of a single corridor for tests purposes. C) Time, velocity and
total collisions vs. number of trials graphs. .................................................................... 27
Figure 4.1: Orbbic Persee ....................................................................................................... 29
Figure 4.2: Orbbec Astra Pro .................................................................................................. 30
Figure 4.3: A human standing at 6m distance can be clearly distinguished in the ................. 31
Figure 4.4: Microsoft Kinect sensor with the cover taken off ............................................... 31
Figure 4.5: Microsoft Kinect sensor depth image .................................................................. 32
Figure 4.6: Human face captured with Microsoft Kinect sensor. (a) Video frame, (b) depth
image, and (c) a close-up of the facial surface .............................................................. 32
Figure 4.7: Asus Xtion ............................................................................................................ 33
Figure 4.8: ZED stereovision camera ...................................................................................... 34
Figure 4.9: original image and the depth map corresponding to it using ZED SDK with
OpenCV library ............................................................................................................... 35
Figure 4.10:How a ZED camera works .................................................................................... 36
Figure 4.11: A block diagram of using ZED to generate a depth map .................................... 36
Figure 4.12: Jetson TK1 ........................................................................................................... 38
Figure 4.13L Flexible OLED .................................................................................................... 39
Figure 4.14: TFT LCD model .................................................................................................... 41
Figure 4.15: Adafruit 5" display .............................................................................................. 41
Figure 5.1: Hardware structure of NOOR ............................................................................... 44
Figure 5.2: The block diagram of the entire system design .................................................... 45
Figure 5.3: OpenCV codes flowchart ...................................................................................... 47
Figure 5.4: A combination of the left and right images, default in OpenCV ........................... 48
Figure 5.5: Left view (on the left), and right view (on the right) ............................................ 48
Figure 5.6: Overlay image of left and right sources ................................................................ 49
Figure 5.7: Depth map ............................................................................................................ 49
Figure 5.8: FULL sensing mode (right), RAW sensing mode (left) ........................................... 49
Figure 5.9: Performance mode (left), Quality mode (right) .................................................... 50
7
Figure 5.10: Left and right views in disparity .......................................................................... 50
Figure 5.11: Decrease (left) and increase (right) reliability index by 1 in RAW mode ........... 50
Figure 5.12:Decrease (left) and increase (right) reliability index by 1 in FULL mode ............. 51
Figure 5.13: depth map represented by OpenCV. .................................................................. 51
Figure 5.14: depth map represented by OpenGL ................................................................... 52
Figure 5.15: OpenGL code flowchart ...................................................................................... 52
Figure 5.16: A photo of one arrangement of the different obstacles and schematics show the
setup up of the five mazes ............................................................................................. 54
Figure 5.17: a) Number of trials vs. Collisions graph. b) Number of trials vs. Time graph. c)
Number of trials vs. Velocity graph, which belongs to the average results . .................. 55
Figure 5.18: a) Number of trials vs. Collisions graph. b) Number of trials vs. Time graph. c)
Number of trials vs. Velocity graph, which belongs to participant 2 .............................. 57
8
Chapter 1
1 Introduction
1.1 Background
According to the World Health Organization, there are around 285 million
people in the world considered blind, the majority (246 million) has low-vision and the rest
are totally blind. The statistics also mention that 82% of the blind people are above 50, and
the numbers of the visual impairments has decreased upon the last 20 years according to
the different and successful treatment methods [1].
Scientists and doctors classify the vision level into 4 important categories: the
normal vision, middle visual impairments, severe visual impairments which identify the term
“low-vision”, and finally the blindness.
During the research the focus will be on the low-vision category since they are the
majority and a lot of help can be given to them so they can have a better quality of life.
Since that no patients are identical; five main cases that lead to vision loss will be discussed
which are: Retinitis Pigmentosa, Glaucoma, Cataracts, Diabetic Retinopathy, and Age-
Related Macular degeneration [2].
A patient who’s under these categories may be able to get benefit from the design.
The focus here will be on the symptoms, once they are known, designing the glasses can be
done to enhance and improve the quality of the reflected images on their eyes in a way that
suits their conditions, but not treating them.
This chapter is organized as follows: section 1.2 briefly explains the five main cases.
Section 1.3 illustrates the main objective of the project and shows the block diagram.
Section 1.4 mentions the thesis organization.
1.2 Cases
This project focuses on five intrinsic cases for the purpose of helping patients with
these cases, however the tests were done on two cases only. The cases are summarized as
follows:
• Retinitis pigmentosa, where the patient gradually lose the night and
circumferential vision, which leads up to the tunnel vision and blurry view.
• Glaucoma, patients with this disease face an unexpected sight vision loss with hazy
view and acute head pains due to the altitude compression on the eye.
9
1.3 Thesis objective
The intention of the project is to resolve the problem of vision with the low-vision
people, who are officially classified as blind persons. Although they are considered blind,
they have different issues that can be handled with variety of solutions; one of them is to
improve the remained vision by enhancing various environment scenes. This can be
achieved by implementing a smart glass –NOOR- that depends on capturing the scenes,
transforming the scene to a depth map, and after all, reflecting the new scenes on a screen.
The significant value of the project is the improvements in the wearer’s social and
academic life. Identically, wearing NOOR helps the user to avoid the surrounding diversified
obstacles and thus the mobility increases. If such a product is released, the independency of
the users can be augmented, which will get them involved more in the society. The project’s
strategy is to use the product on people with different low-vision cases, where the low-
vision patients can communicate better with others, reach desired destinations faster and
easily, and participate in the various events. Moreover, disposing of canes make a person
looks normal.
In order for NOOR to function well, a special camera mounted on the top of the
glasses’ body reconstructs a depth map scene and a display role as the lenses of the glass
receive the reconstructed scene’s video signal with a frame rate of at least 30 fps and
presents it.
Figure 1.1 shows the block diagram of the entire system design. By the time the
camera detects the surrounding scenes, it conveys it to an integrated graphic processor unit
(GPU) to enhance the images with the help of Jetson TK1, henceforth a depth map of the
original image is generated which will be delivered to the display to fit the different
conditions.
Jetson TK1
Left
rd
source ZED SDK + 3 party library
Depth
Image Image 3D map
Detection map
Camera (Video
analysis Correction generation
stream) Right
source
Display
Figure 1.1: the block diagram of the entire system design
10
1.4 Thesis organization
The thesis is organized as follows: chapter 2, which comes after the introduction
represents some technical and medical terminologies in order to simplify the thesis for the
reader. Chapter 3 discusses the previous works done that are related to the proposed
design, chapter 4 compares between different cameras and displays and shows the
specifications of each and mentions the chosen components, chapter 5 illustrates the
proposed design, explains the final block diagram, presents tests methodologies and shows
all the results, chapter 6 lists the constraints and specifications of the system, and chapter 7
concludes the work that has been done.
11
Chapter 2
2 Terminology
2.1 Introduction
In order to simplify the important concepts of the proposed design, some of the
technical terms, methods, and medical cases should be cleared, therefore, this chapter goes
through all the techniques and the terminologies in a simple and useful way while
explaining them briefly. It is arranged as follows: section 2.2 mentions the technical
terminologies used in the thesis. Section 2.3 demonstrates the five main disease cases and
explains their symptoms and treatments.
If we have two cameras calibrated and placed side-by-side then we can obtain 3D
information from an arbitrary scene and the position of the scene’s points, this process is
called the stereovision. It is defined as the way of getting depth information about a certain
scene from two 2D images. Provided that this technique is taken from the human and
animals eyes, and it has been applied to the computers to give the devices a new vision [3].
Left and right images are taking from two cameras and are reconstructed to create
the geometry of the scene; usually there is a horizontal shift between the two images.
If the intrinsic parameters of each camera and the relative pose between the
cameras are known, and if an image point in the left camera and its corresponding point in
the right camera are measured, the absolute point position can be found by intersecting
(triangulate) the rays.
12
2.2.2 Depth map
The least difficult and most helpful method of performing depth estimations taken
from an image is the depth map. It is a 2D array where the distance information is
represented by the regular x and y axes and the compared z measurements are stored in
the pixels. Depth map contains the information that is related to the distance and the exact
location of an object or a surface on a scene. It is a 3D image that consists of gray pixels,
with the values from 0 to 255; the near the object is, the lighter the represented color,
therefore, the nearest object represents the “0” value, while the far objects is represented
by darker colors which corresponds to high values [5].
The generated depth map from the stereovision camera is supported by the
triangulation method. As known, it has two stereo cameras placed one next to the other
horizontally with a certain distance between them. The scenes captured from these
cameras have different angles of view, and the cameras have been utilized and organized in
a way that they overlap at the required distance of the object. The result is two different
viewpoints for each object in the scene [6].
Two actions should be taken into considerations in order to reach the goal of
obtaining a 3D depth map from a scene: First, analyzing the corresponding left and right
images’ point for a specific surface point that appears in the scene is a must. Secondly, to
calculate the ray intersection point of the pixels in the left and right cameras, the structure
and the geometry of the camera should be known [7].
The used technique is quite simple, the disparity map which is the final look of the
depth map, consists of many disparity values, each value has been calculated using the
difference of the left and right pixels of the same point, figure 2.1 illustrates this clearly.
Figure 2.1: Disparity map generation [7].
13
The mathematical equation (1) is needed to calculate the depth corresponding value
for each point of the scene. And the equivalent triangle diagram for equation (1) is
represented in figure 2.2.
! !∗!
Disparity = X − X = ! (1)
Figure 2.2: Triangulation diagram [8]
Mentioning that B is the separation between the two cameras, x and x' are the
separation between points in picture plane relating to the 3D scene point and their
camera's center. And f is the central length of camera. To sum up with, the depth of the
image's pixels can be obtained if B, x-x’, and f were given [8].
The front-illumination, customary approach to sense the light, works like the human
eyes' rule. The photo detectors in this situation lies behind the lens where lattice that
contains the photo components and wiring has been submitted in front in order to simplify
the assembling, however the wiring is keeping the photocathode layer from accepting all
the light data and in this case diminishing the sign that is accessible to be caught.
14
The BSI sensor has this noteworthy productivity on the grounds because the light
encroaches straightforwardly onto the photoreceptors, instead of threading its way through
a maze of hardware as on an ordinary sensor. This is accomplished by turning the sensor
over so that the inverse side of the sensor gathers the light, henceforth the term 'posterior
brightening'. As for gathering photons, is around half more than the normal customary
sensor [10].
There are more than 30 million people with Retinitis Pegmentosa around the world,
and yet, there is only one treatment that affects only 10 percent of the disease. Therefore,
the unmet need is major, especially for people with advance RP. Nowadays, many people
suffer with this disease; unfortunately, medical communities believe that there is nothing
that can be done to treat RP.
Retinitis Pegmentosa is a group of multiple diseases that involves both eyes, rather
than a simple single disease. It affects the function of the light response cells [11]. Several
researches are done in order to find what causes such a disease, but up to this date, the
only known cause is that it is a genetic disorder.
Figure 2.3: Normal Retina (Left) vs. Retinitis Pegmentosa (Right) [2]
It is very common that people with RP will be diagnosed in their early age. Symptoms
may not appear all at the same time, but may differ according to each case. Not forget to
mention that the sequence of appearing is almost the same for all cases.
The beginning of appearance starts with a low night vision, which means that the
light sensing cells that are responsible of vision progressively decay, and seeing through dim
15
light become a real challenge to them. Later on, the patients start to lose lateral vision
(tunnel vision), eventually extending into the central visual field, as if they are looking
through a straw. It is worth to highlight that they do lose their peripheral vision gradually,
by the time they reach blindness [12]. In addition to that, the central vision reflects a blurry
scene. Visual sharpness -more likely to say shade vision- might get compromised because of
going with abnormalities in the cone photoreceptor cells, which would be the answer for
the difficulty to differentiate between color levels, and visual acuity. Presently, the eyes
become more sensitive to light. The prediction of the vision loss range or how does it
proceed is very difficult when the patient has retinitis Pigmentosa. Therefore the eye
specialist monitors the retinal cells health and does some tests to define how well the
patient sees.
With RP patients, it is hard for the doctor to determine the extent of vision loss or the
speed of such process. He will keep monitoring the eye condition and with some several
tests he can recognize how well the patient will be able to see.
2.3.2 Glaucoma
Specialists define Glaucoma as group of eye disorder that affects the optic nerves,
the one responsible for carrying information from the eye to the brain. Some time it’s called
the silent disease, since there are no initial symptoms appears at the very begging stages. It
is considered a huge vision destroyer, and statistics show that it is one of the strong reasons
that cause blindness in US [13]. It is detected through an eye exam that includes examining
the vision field and acuity [14].
Scientists refer this damage to the high-pressure level in the eye. In open-angle
glaucoma, the point in your eye where the iris attaches the cornea is as open as it has to be,
yet the eye's waste canals get to be stopped up after some time, bringing on an increment
in initial eye pressure and resulting harm to the optic nerve [14]. Glaucoma may reach you if
your parents have the sickness, on the off chance that you are African-American or Latino, if
you are diabetic or have cardiovascular illness. The danger of glaucoma additionally
increment with age [13].
16
Figure 2.4: The developing stage of the vision for a person with glaucoma [13]
Type 1: Usually the optic nerves are totally damaged when the patient start noticing
the symptoms. In the most spread type of glaucoma “Chronic Glaucoma”, peripheral vision
damages without any notice from the patient , and in most cases when they reach the
doctor, around 50% of their vision is almost gone [13]. Vision acuity “sharpness of vision” is
also maintained at the late stages of the disease.
Type 2: In Acute Glaucoma, sudden rises in eye pressure could cause a blurry vision,
pain, and colored holes around the light. Losing peripheral sight gradually and feeling blind
spots should be considered also as main symptoms [14].
2.3.3 Cataracts
Cataracts, the world's heading reason for blindness, it is difficult to identify the risk
factors of cataracts. It causes impair vision due to the clouding of the lens of both eyes [15].
People with cataracts have some difficulties in reading under low light; moreover their
shade recognition may be altered [1].
As known that healthy lens is transparent; therefore lights can travel through the
lens easily. There are many reasons compile as we get older to cause cloudy areas in the
lens, that's why the lens can no longer transmit the light through the lens easily. Cataracts
have many reasons abbreviated as follows:
• Age related; Note that, the greatest risk factor of cataract is age
• Congenital
17
• Genetic (family history)
Figure 2.5: blurry and indistict image results while having a cataract [1]
Researchers predicted that the number of people who will be affected by cataracts
would increase to 30.1 million in the next 20 years.
Symptoms and side effects normally creep up numerous years after onset, generally
when the person grows up. Progressively, a great amount of the lens becomes shady and
cloudy. Cataracts often influence both eyes, but infrequently the same [14]. People with
cataract may have the accompanying symptoms:
• Blurry vision.
• The patients see small patches blur part with their field of vision.
• Difficulty in reading under low light.
• Bright lights affect the vision.
• Unclear and faded colors.
The treatment of such cases differs according to the activities that the patient is
doing. For reading amplifying glasses must be worn, and to decrease the glare, sunglasses
will be helpful. Note that, these measures are temporary; cataract will keep developing and
progressively impair the vision more [15].
Diabetes is a condition where the measure of sugar in the blood is too high that the
body can't utilize it appropriately. Diabetic Retinopathy affects up to 80 percent of all
patients who have had diabetes for 10 years or more. More than 30 % of diabetic patients
have some type of diabetic eye ailment, for the most part diabetic retinopathy (DR). It is
also the main source of preventable visual impairment among people of working age, and a
noteworthy reason for sight loss for older people. It’s the damage in the retina due to
diabetes that can in the long run lead to blindness [16].
18
Figure 2.6: A person with Diabetic Retinopathy will see blurry and distorted images (left) [16]
Diabetes occur when the body is unable to use and store sugar, elevating the sugar
level which damages the blood vessels in various organs like the heart, kidneys and eyes.
This can prompt loss of focal and perhaps peripheral vision.
In the first stage of the disease (non-proliferative Diabetic Retinopathy NPDR), there
will be no indications. Patients can have perfect visual acuity and the disease can only be
detected using funds photography that enables viewing the small blood-filled bulges in
artery walls. Fluorescein angiography can be utilized in case of low vision to see the back of
the eye showing any narrow or blocked blood vessels which indicates lack of blood flow
(retinal ischemia).
The second stage is called proliferative diabetic retinopathy (PDR), where abnormal
blood vessels form at the eye’s back burst and bleed causing blurred, fluctuating vision or
vision loss.
Age is the most common risk factor for AMD and it’s most likely to happen after age
60. And some other factors include smoking, which doubles the chance of having AMD. Also,
studies show that race is also a factor where Caucasians are more likely to have the disease.
19
Genetics also are considered as a factor of getting AMD where the risk is higher when there
is family history. There are nearly 20 effective genes, which may increase the risk of having
AMD, and more are suspected.
In its initial stages, age-related macular degeneration might not have side effects and
may be unrecognized until it advances or influences both eyes. First indication of macular
degeneration is that lines are distorted. This may elevate to a continuous loss of focal vision.
Symptoms also include straight lines begin to seem twisted and dark, hazy regions or white
out shows up in the focal point of vision and rarely a change in colors [19].
Figure 2.7: AMD cause a blind spots and blurry image (Right) [19]
20
Chapter 3
3 Related work
3.1 Introduction
Visually impaired people suffered for a long time, many ideas and techniques came
to the surface in order to simplify their lifestyles, but yet visually impaired and especially the
low vision people couldn’t manage the life in an easy way and independently as they
supposed to do.
Here is an overview of the field, several techniques and ideas that are presented
with an explanation of the advantages and disadvantages, and the reasons that make us
chose a totally different technique. This chapter discusses the previous works done to help
the visually impaired, and it is organized as follows: section 3.2 and 3.3 explains the
ultrasonic sensor and the infrared distance sensor respectively. Section 3.4 mentions some
assistance aids such as the smart cane and some applications with sensors. And finally,
section 3.5 lists some of the technical designs in the market such as the Google glass and the
orCam and depth map smart glass.
21
The Seeedstudio ultrasonic range finder (shown in figure 3.1) is a non-contact model
that measures the distance of the objects from a certain point using special but simple
criteria. It is compatible for electronic blocks, and manufactured to serve the project
modules for different industrial aspects.
The sensor works by sending out an ultrasound signals and receiving the echo after it
is reflected from the obstacles. Using the speed of sound and the time it takes to travel from
the objects, distances can be calculated accurately.
A paper published in October 22, 2011 mentioned that the sensor has a high-level
accuracy, long detection range and can identify the near obstacles easily, but one of the
disadvantages they faced was the wide beam of the sensor which makes it hard to
determine were the barriers are located exactly. The sensor may also detect what is called
“ghost echo” and it is defined as the reflected waves not by the wanted obstacle but by
unwanted objects such as walls, this may be misleading for the visually impaired person
[20].
These infrared sensors are inexpensive, consume almost no power, and can be
placed anywhere easily since it is very small, but there are many disadvantages for such
sensors. Many objects do not reflect light, so using this sensor in an application for the blind
people is risky since the sensor will not identify some of the obstacles, plus that the readings
that come out of two-colored surface at the same place may differ.
In addition, the range detected by the infrared sensor is very small, and for this blind
should be very close the obstacle to be able to avoid it while the goal of the project is to
make them be aware before reaching the objects [21].
22
3.4.1 Smart cane
The first thing comes to mind while mentioning the visually impaired is the cane that
they carry all time in order to prevent collision with the obstacles. However, scientists and
engineers work hard in order to develop smart canes with various specifications to ensure a
safe walk for the user.
A project team advised by Donn Koh in University Of Singapore came out with a
white smart cane ”The Blindspot cane” (showed in figure 3.2) that can detect the objects
using sensors hanged to the cane and as they mentioned, this white cane can ensure a safe
path for the blind. The cane helps also to locate and identify the nearby friends, using the
GPS and Wi-Fi systems, a friend who is using their smart phone, can be easily recognized
and the user will have the ability to meet them. The feedback of the location of obstacles
and nearby friend is easily transferred to the user through an audio Bluetooth earpiece.
Type of cane used differs according to the sight loss, a person who is totally blind or
his vision is too low needs a long and strong cane in order to navigate through the new
environments by themselves with minimum collisions as possible. In the other hand, a
person with poor vision but not severe damages can use a lighter and shorter cane [22]. But
yet, why choosing to carry canes when we can get rid of such devices completely?
3.4.2.1 Glasses
The Idea of these glasses is to use the sensors again which are implemented in the
body to ensure getting the maximum amount of field information. If we take Theiavision
glasses as an example, the distance information will be presented to the user through a
sound generator that is connected to the microcontroller output pins [23].
There is also a Greek student who won the local award in Google’s 2014 Science Fair;
he presented a prototype of glasses with implemented sensors to detect the environment,
23
he wrote the code for the system by himself and transfers it to the glass. As an output of the
system, he added a vibration motor to the side of the glasses instead of sound commands,
to make it easier for the blind to interact with the environment without any sound
distortion [24].
This method may perfectly suit the totally blind people, but for this project cases,
people with low vision can identify the surrounding environment by a completely different,
better and easier criteria. In addition, adding a vibration motor can harm the brain since the
blind will keep having a feedback almost all the time.
3.4.2.2 Gloves
Similar to the glasses with the sensors and vibration methods, the idea of the gloves
also came to the surface. The wrist-mounted gloves helps the user to navigate safely in a
complex environment; it has ultrasonic sensors that detects the distance of the obstacle up
to 10 ft and converts the data into a pressure applied on the wrist, the closer the wearer
comes to the object, the faster the vibrations are [25].
Google glasses are a wearable and unique headset that looks like eyeglasses that can
communicate with the wearer using voice command. The glasses contains a small screen at
the corner of the frame to allow the user to stay in touch with the surrounding world, they
24
will be able to login personal emails, take pictures, communicate with friends, answering
calls, and reach the desired places using the integrated navigation system [26].
The company tried to help the blind people also, and many experiments were held in
order to check if the glasses could serve the low-vision people or not. University of Navada
has received the funding from Google glasses’ team to develop the glass in order to fit the
blind person needs and to help to reach their destinations without any help. The proposed
navigation system is indoor, after the user indicated were to go, the voice command will be
given in order to help the user through their route [27].
But still, the developed idea requires a cane, and as we mentioned previously, we
want to get rid of any additional tools in order to make the low-vision person looks like a
person with a normal vision.
3.5.2 OrCam
The establishment of the orCam was in the 2010, the significant value of the orCam
is to help and provide independency to the visually impaired and the blinds. It is a high
technology system with many different features to facilitate everyday tasks and to help
25
them exploring the surroundings. OrCam assists the visually impaired and blinds to
understand text and read it to the wearer using a small size earpiece that does not ban the
hearing, distinguish and differentiate faces that works on memorizing. A bright small camera
is mounted on the right frame of the glass is programmed in a way to do all the features.
OrCam doesn’t need an internet to move freely outdoors, it uses the internal memory that
has the ability to remember and recall hundreds of saved matters [28].
Unlike the previously mentioned ideas and techniques, this head mounted visual
display uses the user's remaining vision. Prof Stephen Hicks and his team from the
University of Oxford came up with an idea of a head-mounted glass that uses the remaining
vision, where this smart glass has a depth camera attached to a transparent display and an
earpiece all combined to provide the user a convenient movement avoiding any obstacles.
The transparent displays were custom-made so the user will see the depth map image from
the inside and others will see the user wearing a transparent glass. The depth camera helps
the user to see nearby objects in a higher light intensity while the far objects will be unseen
clearly.
The down-side of this visual aid cannot help the user outdoors where sunlight exists,
that’s because the used depth map camera cannot give an output in a bright environment,
and moreover, the user must carry a laptop all the time in order for the camera to process
the surrounding scene and generate a corresponding depth map.
As seen from the results in figure 3.6, time spent to get familiar with the glass and walk
comfortably in an unknown environment decreases logarithmically, user's velocity increases
linearly and number of obstacles collisions decreases linearly [29].
The idea of the proposed design -NOOR- resembles Prof. Hicks works with
differences, and mainly NOOR works indoors and outdoors using ZED camera. Furthermore,
instead of allowing a user to carry a laptop while moving, Jetson TK1 is used which is a
12x12 cm CPU that can be easily programmed. Not to forget to mention that the tests
criteria in this report will follow Prof. Hick’s tests for comparison reasons.
26
Figure 3.6 : A) Prof. Stephen Hicks mart glasses with a laptop that should be carried all the time. B)
Different mazes of a single corridor for tests purposes. C) Time, velocity and total collisions vs.
number of trials graphs. [29]
27
Chapter 4
4 Synthesis
4.1 Introduction
The main idea of the adopted solution is to extend the vision sense through an
interface. The development of a glass system to those who retains low-vision needs
basically three main elements: CPU, depth camera and a display, which should be chosen
accurately to insure that the users get the best quality they can get.
This chapter is organized as follows: section 4.2 mentions the different alternatives
of the depth cameras with their specifications. Section 4.3 mentions the differences
between the two kinds of displays and shows their advantages over each other and presents
the selected one. Section 4.4 concludes the chapter by mentioning the chosen depth
camera and display.
Orbbec, Xtion, Kinect and ZED cameras are being significantly important 3D cameras.
These cameras merge two techniques of the classic computer vision with the structure light,
which are: the depth from focus and stereo. The system uses an infrared laser with a dotted
manner or uses the stereo-vision principle. Moreover, Orbbec, Xtion, Kinect and ZED are of
the most essential 3D measuring devices for objectives recognitions, spectacle
reconstructions, mapping, and robotics. The utilization of Orbbec, Xtion, Kinect and ZED
cameras in some applications other than entertainment is of an extremely great interest,
nevertheless, to ensure that its application in other areas where the objects geometric
properties are important, the technical properties of the system must be studied [29].
28
This section draws an analogy of the four cameras and compares between the
optimum results of different developers and the gotten results from the tests done on the
chosen camera.
4.2.1 Orbbec
According to the design, the depth camera used in the smart glass should give a high
quality grey-scale output, with a high frame rate per second while maintaining a small
weight and dimensions, and reflect excellent images of the surrounding environment. Such
requirements are available in several depth cameras; one of them is Orbbec.
The company developed until now three different designs, with different
specifications, this allows the designer to choose their own type to fit the project. The
outside shape differs between them, the sizes, and some minor and interior specifications,
but they almost follow the same method in analyzing and examining the field, and they give
approximately the same output [30].
The specifications of the orbbec persee make it a very special device that can be
used in many applications rather than the entertainment applications. Table 4.1 shows the
different specifications of it.
Figure 4.1: Orbbic Persee [30]
29
Table 4.1: The specifications of Orbbec Persee
Table 4.2: Orbbec Astro Pro Specifications
30
4.2.1.3 Output form
What makes the Orbbec 3D technology a good choice is the output quality. The
depth camera’s output is represented in figure 4.3. Brightness, color levels, and the amount
of the represented details, all will be adjustable so it can fit the cases and help them in a
very efficient and specific way.
Figure 4.3: A human standing at 6m distance can be clearly distinguished in the
Depth image (left) and color image (right) from the wall behind her at 8m distance [31]
Figure 4.4: Microsoft Kinect sensor with the cover taken off [32]
Figure 4.5 illustrates the depth map generated by the kinect sensor. The value of the
depth map is encoded only with grey-scale; dark pixel indicates that the point is located
closer to the camera. Where the black pixel means that there is no depth value for it, this
case could happen if the points are located too far where the values of the depth map are
31
not accurate, are located too close where the field of view of the infrared projector and
infrared camera is limited and it enters the blind region [32].
As any other cameras, Microsoft kinect has many specifications to be taken into
consideration, table 4.3 shows the different specifications.
Table 4.3: Microsoft kinect specifications.
Kinect Array specification
Angle of viewing 43ᶱV , 57ᶱH
Dimensions 15.0 x 6.1 x 4.8 cm
Default resolution VGA (640 x 480)
depth stream
Working Area Indoor
32
Xtion PRO LIVE uses infrared sensors, adaptive depth recognition, colored picture
detecting and sound stream to catch a client's ongoing movement and voice, giving an exact
tracking. Xtion also has a development solution that allows developers to have many
applications related to the moving gestures using motion-sensing technology [35].
Xtion is compatible with OPNI NITE middleware SDK which makes the motion-
sensing applications and games development much easier, which is beneficial to have an
easier programming language. One of the most important features its high distance range
that it can reach where the range starts from 0.8m up to 3.5m.
Table 4.4 shows the main different specifications of Xtion Pro Live Asus to be
compared with the other camera's specifications [36].
33
Dimensions 18 x 3.5 x 5 cm
Weight 540g
Working Area Indoor
4.2.4 ZED
ZED is a stereo camera with a small-weight depth sensor that uses the passive
stereovision technique, and the output of the system is a video with high resolution and
presented using USB 3.0.
ZED SDK is the software that helps to construct the depth scene from the side-by-
side video cameras; the graphic processing unit in the machine also helps to do that [37].
4.2.4.1 Specifications
ZED camera can work in both environments, indoor and outdoor. Other than that,
regarding the output resolution, each quality requires a certain video mode and gives
different value of fps, table 4.5 explains this more clearly.
Table 4.5: The output resolution of ZED
Resolution Frame Per Second Video Mode
4416x1242 15 2.2K
3840X1080 30 1080p
2560x720 60 720p
1280x480 120 VGA
The depth resolution is the same as the selected video resolution, the range starts
from 1.5m up to 20m, and the depth format is 32-bits with OpenCV compatible which
allows detecting and recognizing faces, motions, constructing 3D modules, and adding the
images together so they can build a high resolution scene. The lens of the cameras is a wide-
angle, made of glass and provides 110 degree as a maximum field of view with aperture
f/2.0.
34
In order to fit all the conditions and to improve the performance of the camera in
the low-light areas, people who work on ZED added a backlight illumination sensor, which is
a digital image sensor that uses a special arrangement to increase the light that enters the
camera, this technique is mostly important in the low-light environments. The sensors of
ZED are 1/27” and have been chosen perfectly with a resolution 4M per sensor [33].
Controlling the camera is one of the main specifications that any developer is looking
for. ZED team made it very easy to adjust so it fits the user needs. The resolution,
brightness, frame rate, saturation, contrast, sharpness, and white balance, all can be
controlled somehow without adding any external item. Table 4.6 describes the physical
specifications of ZED camera
Table 4.6: Physical specifications of ZED
The 3D depth map of ZED corresponds to the distance from the stereo camera, to
put it differently, near objects are reflected with the same shape but in white, each time the
object gets further away from the camera the white reflection decreases and the object
enters the grey-scale region. The black background of the picture represents objects that
are at a distance more than 20m that cannot be detected by the camera.
Figure 4.9: original image and the depth map corresponding to it using ZED SDK with OpenCV library
Writing a code using the ZED SDK software, where the original image and its output
are shown in figure 4.9, can produce the grey-scale depth map output.
35
4.2.4.3 How it works
Figure 4.10:How a ZED camera works
Figure 4.10 shows how a ZED camera works. Firstly, the ZED records the scenes using
the left and right sources of cameras in a very accurate way and sends the video scenes
through a USB 3.0 interface. Secondly, the software (ZED SDK) has an internal graphic chip
that comprehends the texture of the delivered video scene and calculates the depth map of
it in the real time. Finally, the output of the ZED SDK can be implemented for different
purposes and applications [37].
3D generation process
Left
Stereo source Image Image 3D map Depth
analysis Correction generation
map
camera
Right
source
Figure 4.11 represents the steps of generating depth map in the camera. ZED has
two cameras (left and right sources), each source captures the same scene in a different
angle to produce a full scene when collaborating, the scenes enter the 3D generation
processes, where the image analysis block detects the roll, tilt, height, focus and depth, it
sends the signal to the image correction block, where it matches the colors, and sends the
signal again to the 3D map generation where it generates a dense and a semi-dense maps.
The result of the three blocks is a 3 dimensions depth map.
36
4.2.4.5 Experimental results
Due to the different features of the ZED stereolabs camera, numerous characteristics
can be taken into considerations. As it's known that the ZED works with ZED SDK together
with a third party. The project focuses on two main libraries, which are: OpenCV and
OpenGL provided that all codes are written in C++ language.
37
in order to make the design more efficient and user friendly. Figure 4.12 shows the Jetson
TK1 and table 4.8 shows the specifications of the Jetson TK-1.
Table 4.8: Specifications of Jetson TK1
Dimensions 5" x 5" (127mm x 127mm) board
GPU NVIDIA Kepler"GK20a" GPU with 192 SM3.2 CUDA cores (up to
326 GFLOPS)
CPU NVIDIA "4-Plus-1" 2.32GHz ARM quad-core Cortex-A15 CPU with
Cortex-A15 battery-saving shadow-core
Storage 16GB fast eMMC 4.51 (routed to SDMMC4)
USB 3.0 A full-size Type-A female socket
HDMI A full-size port
Power A 12V DC barrel power jack
Camera ports 2 fast CSI-2 MIPI camera ports
LCD port LVDS and eDP Display Panel
Hardware- • CUDA 6.0 (SM3.2, roughly the same as desktop SM3.5)
accelerated APIs • OpenGL 4.4
supported: • OpenMAX IL multimedia codec including H.264, VC-1 and
VP8 through Gstreamer
• NPP (CUDA optimized NVIDIA Performance Primitives)
• OpenCV4Tegra
38
In order to get the Jetson and ZED working together there were some steps for
setting up. First of all Linux operating system must be installed, followed by the Jetson
development pack. Then, enabling USB 3.0 since it is recognized by default as USB 2.0. And
the last step is installing the ZED SDK; this enables the user to use the ZED camera using the
Jetson without the need of large computer devices. And finally, showing the depth map
using different libraries.
4.4 Displays
To get the best fitting display on the glass, looking through variety of choices was an
enormous challenge, each had advantages over others and disadvantages too. In this
section the best alternatives are highlighted and the focus is on the main differences
between them in order to be able to choose the right display that fits the specified
requirements.
4.4.1 OLED
The organic light emitting diode (OLED) is made by two separate conductors, which
are filled by progression thin films that are organic. Once an electrical current passes
through it, the OLED emits bright light. Backlights are not required for the OLED since that it
emits its own light. OLED has many advantages over LCD. It consumes a lower amount of
power; battery contrast is faster, as well as the refresh rate, and a considerably greater and
better brightness [38].
A flexible 1.66" OLED display is one of the options considered for the project; it is a
parallel interface small screen, perfect for different applications, such as: smart phones,
modern smart watches or even for wearable watches. It can easily bend to fit different kinds
of projects. Figure 4.13 shows the flexible OLED [39]. Table 4.9 shows the different
specifications of each item mentioned [39].
Figure 4.13L Flexible OLED [40]
39
Table 4.9: 1.66" flexible OLED specifications
Item Specifications
Dimensions 32.2 x 54.99 x 1.18 mm (W x H xT)
Dots Number (Resolution) 132 x 176 x RGB ( W x H x RGB)
Viewing Area 26.314 x 34.762 mm (W x H)
Weight 0.8 grams
Current Consumption – Logic 1.5 mA (Typical)
Current Consumption – OLED 37 mA (Typical)
4.4.2 TFT
To improve the quality of the image and the contrast, a thin-film-transistor LCD (TFT
LCD) is being used widely as an efficient display. It is an active-type matrix that is
implemented in a huge range of applications such as mobile monitors, TVs, GPS monitors,
digital cameras, and a lot of video games as well. Not to forget to mention that they are
available with different types, each has its own functioning system.
What makes the TFT a perfect choice for any consumer is the variety of the electrical
interfaces that can be connected to. It accepts HDMI, VGA, RGB, DVI, or DPI, each according
to the chosen TFT. To convert these types to a digital RGB while maintaining the original
resolution of the display, manufacturer implemented a converter in a control board, which
is connected to the interface [41].
Choice 1:
As a strong option, a 2” TFT LCD model with an RGB interface will perfectly fit the
design (Model no. ZP20053), it has a good resolution. Table 4.10 describes the technical
specifications and figure 4.14 shows the TFT LCD model with the dimensions [42].
Table 4.10: TFT LCD model technical specifications [42]
Resolution (RGB) 480×240
Size (cm) 4.61 × 4.096 × 0.253
Active area size (cm) 4.056 x 3.048
Interface RGB ( 8-bits )
Color Depth 16.7M
Number of pins 40
40
Figure 4.14: TFT LCD model
Choice 2:
Adafruit 5” is a display (figure 4.15) with smaller than usual HDMI screen and an
implicit touchscreen. Very simple to use, and the user can utilize such display with any PC
that has HDMI interface. The 5” display has a sufficient resolution (800x480), which makes it
very useful to run most of the programs and software, and with this efficiency it can be
implemented in a wide range of embedded projects. One of the advantages is that this
display can be run using a USB cable as long as the used CPU can support 500 mA.
Figure 4.15: Adafruit 5" display
There are two versions of such screens, touch and non-touchable screens. The
display has good specifications that match the proposed design and table 4.11 describes
them [1].
41
Table 4.11: Adafruit display specifications
Size (cm) Weight Type Power Operating
Consumption Temperature (°C)
12.6x7.6x0.7 0.106 TFT 500 mA 0 – 70
The Display has a backlight that helps to increase the brightness and makes the
screen looks better. Enabling the backlight will consume more power and thus will reduce
the battery lifetime [1].
4.5 Battery
A Li-Po battery is a shortcut of Lithium Polymer battery. The one used in this project is
a 2200mAh rechargeable battery with 3 cells, each cell needs 3.3V and the total amount of
voltage of all three cells is 11.1V, once charging, it should not exceed 12.6V, and once
discharging, it should not be less than 9.9V. Fully charging the battery from 9.9V to 12.6V
needs around an hour and works for 6 hours, if the same battery is connected to Jetson TK1
and the display it takes more power therefore it discharges faster, where it takes around 2
hours to be fully discharged.
4.6 Conclusion
After a long study and comparing many alternatives a conclusion can be drawn, the
depth camera which will be used in the design is the ZED because of its specifications that
meet the desired requirements, the Jetson TK1 will be added to the system to play the role
of the CPU, and the display that will play the role of the lens will be the 5” Adafruit display
due to the availability of the screen in the market, where it is going to replace the two small
displays in the previous proposed design.
42
Chapter 5
5 Proposed design
5.1 Proposed solution
Among all the mentioned solutions, we adopted the idea of smart glasses, where the
user will move independently without any additional assistants such as canes or guide dogs.
The goal of the project is to help the low-vision user to avoid the obstacles surrounding
them in an unknown environment, and to be able to interact with different environments by
improving the user’s mobility. The main idea of the proposed solution is to extend the vision
sense through an interface.
The development of a glass system to those who retain low-vision needs basically
three main elements, which have been chosen accurately to insure that the users get the
best quality they can get.
In the first place, after analyzing the available sources in the market, differentiating
between varieties of choices was a big challenge. After all, the best camera that matches the
required specifications is ZED; it is carefully studied in the depth cameras section. The
general principle of the camera is to deliver an accurate perception in grey-scale 3D
mapping in real time using stereovision technique. Locating two stereo cameras that scan a
wide field of the environment with accurate distance measurements and send it to an
internal processor that cooperates with the Jetson TK1 processor to transfer the scenes into
a depth map format.
Not only the cameras and the Jetson TK1, but also a display is essential to play the
role of the glasses’ lenses. After investigating the specifications of many displays, the chosen
display is a TFT display that blocks the outcome light with the help of the glass’s body and
perfectly fits the shape of the structure design.
Equally important, the used imaging techniques will play the major role in the
design. In order to simplify the user’s life, few modifications must be done to the original
scene. By writing a certain programming code, the brightness, contrast, saturation,
sharpness, and white balance adjustments can be controlled, which will result in a new and
better image that will be reconstructed to suit the patient’s condition.
The design will allow the low-vision persons to experience the world around them
for the first time, where it aims to help the low-vision people navigate independently
through different environments while minimizing the number of collisions.
43
5.2 Hardware design
The glass’s hardware consists of two main parts and some subsidiary parts: the ZED
camera, a single TFT display, a box that contains Jetson TK1 and the system batteries, a
sponge supporting pads and the plastic body of the camera.
The zed camera plays a huge part in the system design where it is the main receiver
for the surrounding scene, wire connection connects the output of the zed camera to Jetson
TK1, which is located in the box where the battery and all necessary wires are also placed. A
hallow sponge is placed between the users eyes and the displays to provide maximum level
of comfort to the user, the sponge will block the user from seeing any outer scenes so there
will be no interference between what the user sees directly and what he/she sees in the
displays.
Figure 5.1: Hardware structure of NOOR
44
Jetson TK1
Left
rd
source ZED SDK + 3 party library
Depth
Image Image 3D map
Detection map
Camera (Video
analysis Correction generation
stream) Right
source
Display
Figure 5.2: The block diagram of the entire system design
The detailed block diagram of the system is shown in figure 5.2. Firstly, the camera’s
function is to detect different scenes and to help to process them in the same time; the left
and right stereo cameras do detecting, and the ZED SDK with the help of a third party library
such as OpenCV and OpenGL do processing. The main processing operations are handled in
Jetson TK1 since it has a very powerful GPU. In the processing stage, an image enhancement
code is written to the integrated GPU to change the properties of image and to calculate the
depth map by using the triangulation method. The end of this point transforms the image
into a depth map image. Later on, the image will be conveyed to the display using an HDMI
cable.
A box is added to system as mentioned to hide the Jetson TK1 and the battery. The
content of the box functions as the heart of the system; where it contains the most
significant components that allows the system to operate. Firstly, Jetson TK1 is placed in the
box to maximize the protection level, and according to the fact that it is unpractical to let
the users carry the Jetson TK1 while they are moving through an environment, therefore,
hiding it in the box is more convenient. Secondly, the Li-Po battery is also placed in the box
but is separated from the Jetson TK1 except a small hole that connects the charging wire
between them.
45
language is C++ language, and the code is written using a special program called ZED SDK
with a third party library.
To be able to use the ZED SDK, the user’s operating system must match specific
requirements, with the following configuration, note that the main reason of choosing the
Jetson TK1 is that all the specifications are included in it:
• Dual-core 2,3GHz
• 4GB of RAM
• NVIDIA GPU with compute capabilities >2.0
• CUDA 6.0
• USB port 3.0
• Windows 7, 8, 8.1(64 bits) Ubuntu 14.04, L4T21.3/4
For having a high-resolution footage while recording the video, user needs 250 MB/s
at least as a transfer speed [37].
Due to the different features of the ZED stereolabs camera, numerous characteristics
can be taken into considerations. As it's known that the ZED works with ZED SDK together
with a third party. The project will focus on two main libraries, which are: OpenCV and
OpenGL provided that all codes are written in C++ language.
Three codes were written, each one of them gives different quality of output. Due to
the differences in the users medical situations, and after testing the glasses on the chosen
cases, the results proved that there is no fixed code for the glasses and any code of these
three may be the chosen one, depending on the user itself.
46
5.3.1.1.1 ZED with OpenCV
ZED with OpenCV demonstrates how to link the SDK with the library of the OpenCV.
A written code using OpenCV library can either conceive the depth or the disparity map.
Each camera of the ZED captures the same scene but in a slightly different angle, and the
default situation is a normal colored image.
Two codes has been written to represent the depth map with OpenCV library, each
one of them handles different situations, and figure 5.3 illustrates the flow chart of both
codes.
Yes
No
Detect image size
Yes
Choosing wanted
resolution
Yes
No
Get right and left
image buffers
Yes
Measure disparity
map and retrieve
depth map
Figure 5.3: OpenCV codes flowchart
47
5.3.1.1.1.1 Code 1
A simple switch statement in the code can show different views using keyboard
shortcuts. The results and the shortcuts are shown in table 5.1.
Table 5.1: Key shortcuts to control different characteristics of depth image
Main Hotkeys Display Hotkeys
b: Decrease reliability index by 1 0: Left View
Figure 5.4: A combination of the left and right images, default in OpenCV
Figure 5.5: Left view (on the left), and right view (on the right)
Figure 5.5 shows that the left and right cameras capture the same scene but with
different angles as the human eyes work. The overlay image shown in figure 5.4 combines
both the right and the left images.
48
Pressing on 'd' key toggles between the two maps, depth and disparity. Disparity is a
2D image that refers to the separation between two relating points in the left and right
pictures of a stereo pair, while depth is a 3D image contains the depth values for different
pixels.
Correspondingly, sensing modes can vary between FULL and RAW to grab the image
as well as the depth. RAW sensing mode is adequate for collision avoidance and navigation
applications, while FULL mode is targeted at augmented reality and computational imaging
where a fully dense depth map is needed.
Figure 5.8: FULL sensing mode (right), RAW sensing mode (left)
49
Similarly, ZED can be initialized in a computation mode: performance, quality or
neither. Performance mode is the fastest, it needs less GPU memory, and the resulted
disparity map is less detailed, while the quality mode gives a better quality with a more
detailed disparity map.
Decreasing and increasing reliability indices by 1 can also be done in both RAW and
FULL mode.
Figure 5.11: Decrease (left) and increase (right) reliability index by 1 in RAW mode
50
Figure 5.12:Decrease (left) and increase (right) reliability index by 1 in FULL mode
Using ZED with OpenCV gives good results with high resolutions and image quality
except that while video casting in real time, a noticed delay was detected in version 0.9.2,
the duration of the delay was around 1.5 seconds, which is much higher than the expected
delay and can’t serve the project purposes perfectly, but in version 0.9.3 the delay has
reduced by almost 70% where the delay became unnoticed.
5.3.1.1.1.2 Code 2
In this approach delay is around 1 second compared with code 1,the frame rates is
much better and gives continuous video stream, but the quality of the depth map is poor.
51
Figure 5.14: depth map represented by OpenGL
Yes
Configure window
size
Yes
Yes
No
Draw image texture
in left and right part
Yes
Measure disparity
map and retrieve
depth map
52
5.4 Testing methods
Briefly, the system is a glass that bocks the outer environment and shows the scene
in a grey-scale depth map on a 5 inch screen that roles as the lens of the glass to help the
low-vision people. Tests were made on two categories: healthy people with normal vision,
and people with low-vision. The purpose of testing on normal people is to insure that the
system works perfectly and they can avoid the different obstacles while moving through an
environment, and on low-vision people to insure that the system can help them walking
independently with no assistants in a unique way. All participants: normal and low vision
people were asked to get used to the glass before using it, this is done by giving them a
minute or two to scan the environment before walking. The test environment was a narrow
corridor with different obstacles (chairs, vases, small tables and others), each participant
has walked in the same corridor for number of times with different obstacles arrangements.
5.4.1 Study 1: Obstacle avoidance test on healthy people with normal vision
Tests were done on five healthy persons with normal vision. The test had two
intrinsic purposes: foremost to define if the users are able to avoid the different obstacles
around them in a narrow maze, comparatively to measure the rate at which the users could
figure out how to utilize the unfamiliar screen. Users where directed on how to use the
system by allowing them to relax and scan the environment before moving around, else,
there was no other training or preparing for the test. Participants were requested to move
through the prepared maze that is a narrow corridor (approximately 9x2 meters) with
different obstacles without crashing or colliding the obstacle edges or hitting walls.
Rearranging the places of the obstacles created ten different mazes, and the users finished
each maze in a random route. Number of trials versus time to complete a maze, number of
collisions, and the user velocity were recorded.
All participants were able to complete all the ten mazes; figure 5.17 shows the
relationship between the number of trials versus time, collisions and velocity.
53
Figure 5.16: A photo of one arrangement of the different obstacles and schematics show the setup
up of the five mazes
54
Trial Vs. number of collisions
2
Collisions 1.5
y = -0.0824x + 1.0533
1 R² = 0.30461
0.5
0
0 2 4 6 8 10 12
Trial
30 R² = 0.748
20
10
0
0 2 4 6 8 10 12
Trial
30
20
y = 1.9145x + 14.504
10 R² = 0.87186
0
0 2 4 6 8 10 12
Trial
Figure 5.17: a) Number of trials vs. Collisions graph. b) Number of trials vs. Time graph. c) Number of
trials vs. Velocity graph, which belongs to the average results .
55
The average time took both participants to complete maze 1 is 39.3 seconds while
only an average of 14.7 seconds took them to complete maze 10, as seen from the trials
versus time graph, the time is decreasing logarithmically by an average of R²=0.748.
Similarly, number of collisions in the graph decreased from an average of 1.6 collisions in
the first trial to 0.2 collisions is the last trial; at the same time the velocity of the participants
has increased almost linearly from an average of 13.74 m/min to 36.73 m/min.
With the co-operation of Emirates Association of the Visually Impaired (EAVI), Noor
has been tested on three healthy participants with different visual cases. Two of them were
diagnosed with Retinitis Pegmentosa, and one with Glaucoma, note that the two patients
with RP have different levels of vision. Table 5.2 illustrates the details of the participant
cases.
Before starting the tests on the participants, they were asked to get familiar with
the glass by scanning the surrounding environment, and few instructions were given to
them in order to teach them how to use NOOR. Participant 2 felt comfortable while using
the glass, she was able to sense the environment clearly and avoid different obstacles after
several trials, although participant 3 is diagnosed with the same disability but the glass could
not improve her movement rate, since she felt more comfortable while moving through an
environment without using NOOR. Participant 1 is diagnosed with glaucoma, but yet, the
glass could not serve his needs. The level of disabilities of patients is not identical, therefore,
two persons with the same case may have different levels and by that may interact with
NOOR in a different way. Perhaps, if NOOR was tested on a larger number of people with
low vision the results may prove that NOOR can serve and help other cases too.
56
Number of trials Vs. collisions
3
2
collisions
0 y = -0.6x + 2.8
0 1 2 3 R² = 0.9
4 5 6
-1
Trial
60
Time (sec)
40 y = -9.1x + 82.9
R² = 0.93549
20
0
0 1 2 3 4 5 6
Trial
12 y = 1.577x + 4.471
10 R² = 0.89485
8
6
4
2
0
0 1 2 3 4 5 6
Trial
Figure 5.18: a) Number of trials vs. Collisions graph. b) Number of trials vs. Time graph. c) Number of
trials vs. Velocity graph, which belongs to participant 2
57
Chapter 6
Obviously, the design will not cause harms for humans and animals since it is
rechargeable. The surrounding environment may be affected if and only if the Lithium
battery was deposed. Some ignorant people miss the right way of how to get rid of
batteries, which leads to environmental damages caused by the incorrect disposal method,
these materials infiltrate to the soil and underground water, subsequently harm the plants,
animals, and humans. Therefore, the ideal method is to use rechargeable systems.
Releasing the product in the market will help the society in different ways, for example:
• The product will help a huge number of people with low vision, and cut off their
permanent social issues.
• The product will improve the level of communication between the user and others
surrounding them.
Although the idea of the product is new in the Arabian Gulf area, in the belief, people
in UAE will accept it no matter how the glass will look like, since this product has great
humanitarian aspects.
While choosing the design hardware and components, cost has been taken into
considerations; it has been minimized as much as possible to make the product available for
58
almost everyone. The body of NOOR with the camera and the displays costs around 4600
AED.
There are different kinds of assisting aids to help the visually impaired, but yet the smart
glass is not available for the Arabian Gulf area, therefore, our design will be the first of its
kind to be produced here.
The product will provide a new kind of assistance to those who retain low vision, they
will be more comfortable with the product since using the smart glass will terminate the
cane or any other assistance aids.
Every product must have limitations and restrictions that limit the effectiveness of it,
and NOOR also has limitations such as:
• The proposed design is not applied for all low-vision cases, only the ones mentioned
in the medical terminology section.
• The proposed design does not detect transparent objects since that the laser beam
goes through it and does not have a reflection that can be caught.
• The proposed design does not detect objects that are at a distance more than 15m.
• If a zoom lens has been applied, the resolution decreases.
• The project does not operate in dark environments.
59
Table 6.2:Power dessipation
Estimated power
Parts
dissipation
ZED stereo camera 380mA x 5V = 1900 mW
Adafruit display 250mA x 5V = 1250 mW
60
Chapter 7
7 Conclusion
7.1 Thesis summary
As has been noted, the majority of the blind people around the world return low-
vision. This thesis aims to discuss a procedure that provides assistance to low-vision people
while moving around an unknown environment using the wearable obstacle detection
smart glass "NOOR". The original scene of the surrounding environment is caught by a
stereovision camera, which passes through the processing stage that happens in Jetson TK1
with a co-operation with ZED camera. In this stage, a depth map is constructed from the
original scene using the pixel triangulation method. The luminance, brightness, sharpness,
edge detection and other enhancements have been moderated to fit the medical condition
needs. The Jetsons’s outcome video signal is carried out to the TFT display, which plays the
role of the glasses’ lens.
To ensure that NOOR functions well two studies have been performed. The first
study included healthy people with normal vision to prove that the user needs time to be
familiar with the glass and the rate of learning for each user differs. Study 1 also proved that
by using NOOR, the user would be able to differentiate the closer obstacles and therefore
avoid them. Study 2 included people with low vision, and the results showed that the depth
map representation is very helpful for some of them, and with NOOR they were able to get
rid of the cane or any assistance aids.
Generally speaking, the substantial purpose of this project is to promote the life’s
different aspects, and to improve the ability to avoid obstacles. It is important to realize that
the glass’s function is to consolidate the scene, not to recover the vision.
7.2 Difficulties
As any other implementation process, NOOR’s execution procedure phase faced many
pitfalls that can be summarized as follows:
1. The project’s interest of area tends to computer engineering field more than
Electrical and Electronics engineering field, and thus, setting up the environment
obtained the biggest effort.
2. The main part of the project is to present the scene as a depth map, this process do
not occur easily, several codes and several techniques were tested to ensure a good
quality for the user.
61
3. The noticed delay between reality and virtual scenes took time to be reduced to an
unnoticed delay concerning that the project works in real time, and the changes
made in the environment should be reflected as fast as possible to the user.
4. JetsonTK1 is a new product and it is difficult to find a suitable display that can be
attached to it and fits the requirements.
62
References
[1] B. Mandracchia, A. Finizo, P. Ferraro. Cataract vision mimicked by means of protein denaturation In
egg albumen. Institute of Applied Sciences & Intelligent Systems, Naples, Italy.
[2] Usskamp V, Duebel J, et al. (2010). Genetic Reactivation of Cone Photoreceptors Restores Visual
Responses in Retinitis Pigmentosa
[3] stereo vision. USA: William Hood, 2015.
[4] Wikipedia, 'Computer stereo vision', 2015. [Online]. Available:
https://en.wikipedia.org/wiki/Computer_stereo_vision. [Accessed: 25- Nov- 2015].
[5] Wikipedia, 'Depth map', 2015. [Online]. Available: https://en.wikipedia.org/wiki/Depth_map.
[Accessed: 27- Nov- 2015].
[6] I-art3d.com, 'About Depth Map', 2015. [Online]. Available: http://www.i-
art3d.com/Eng/About_Depth.htm. [Accessed: 26- Nov- 2015].
[7] A. vorbehalten, Obtaining Depth Information from Stereo Images, 1st ed. Dimbacher: Ensenso and IDS
Imaging Development Systems GmbH, 2012, pp. 2-6.
[8] Docs.opencv.org, 'Depth Map from Stereo Images', 2015. [Online]. Available:
http://docs.opencv.org/master/dd/d53/tutorial_py_depthmap.html#gsc.tab=0. [Accessed: 01- Dec-
2015].
[9] Wikipedia, 'Back-illuminated sensor', 2015. [Online]. Available: https://en.wikipedia.org/wiki/Back-
illuminated_sensor. [Accessed: 26- Nov- 2015].
[10] B. Newman, 'BSI sensor (backside illumination): Advanced technology guide', What Digital Camera,
2015. [Online]. Available: http://www.whatdigitalcamera.com/technology_guides/bsi-sensor-backside-
illumination-advanced-technology-guide-60726. [Accessed: 27- Nov- 2015].
[11] G. J. Farrard, P. F. Kenna, P. Humphries. (2002). "NEW EMBO MEMBer's REVIEW: On the genetics of
retinitis pigmentosa and on mutation-independent approaches to therapeutic intervention". The EMBO
Journal 21 (5): 857–64.
[12] M. Haddrill,( 2015, July). Retinitis Pigmentosa [Online]. Available:
http://www.allaboutvision.com/conditions/retinapigment.htm
[13] “Glaucoma in twenty first century”, in The Glaucoma Book, 1st ed. New York (2010). ch.1, pp.03-21.
[14] National Eye Institution. Facts About Glaucoma[Online]. Available:
https://nei.nih.gov/health/glaucoma/glaucoma_facts
[15] Hodge, W. G., Whitcher, J. P., & Satariano, W. (1994). Risk factors for age-related cataracts.
Epidemiologic reviews, 17(2), 336-346.
[16] Klein, Ronald, and Barbara EK Klein. "Diabetic eye disease." The Lancet350.9072 (1997): 197-204.
[17] Wills Eye Institution. A Patient guide to Diabetic Retinopathy. Available:
https://www.willseye.org/sites/all/files/WE_Service%20Brochure_Diabetic_FINAL%201110_0.pdf
[18] Leibowitz, Howard M., et al. "The Framingham Eye Study monograph: an ophthalmological and
epidemiological study of cataract, glaucoma, diabetic retinopathy, macular degeneration, and visual
acuity in a general population of 2631 adults, 1973-1975." Survey of ophthalmology 24.Suppl (1979): 335-
610.
[19] Bird, A. C., et al. "An international classification and grading system for age-related maculopathy and
age-related macular degeneration." Survey of ophthalmology 39.5 (1995): 367-374.
[20] Diomin V. (2011). Low-cost sensor data processing of autonomous mobile robot guide. Available:
http://robotics.bstu.by/mwiki/images/e/ec/Demin-OWD-2011.pdf
[21] Aabloo, A. (2014). THEIAVISION: ULTRASONIC GLASSES FOR PEOPLE WITH VISUAL IMPAIRMENT.
[Online] Available: http://www.tuit.ut.ee/sites/default/files/tuit/arvutitehnika-thesis-bsc-2014-soo-
sander-text-20140528.pdf
[22] CNET, 'BlindSpot: Smart cane concept looks to future - CNET', 2011. [Online]. Available:
http://www.cnet.com/news/blindspot-smart-cane-concept-looks-to-future/. [Accessed: 18- Sep- 2015].
[23] VICE, 'This Greek Teen Invented a Pair of Smartglasses for the Blind | VICE | United Kingdom', 2015.
[Online]. Available: http://www.vice.com/en_uk/read/angelos-getsis-smartglasses-for-the-blind-876.
[Accessed: 12- Sep- 2015].
[24] NMANewsDirect, Wrist-mounted glove helps blind navigate. 2013. [online] Available:
https://www.youtube.com/watch?v=VNa-7n6APDE
63
[25] Discovertechnologies.com, 'Google Glass for Indoor Navigation', 2014. [Online]. Available:
http://discovertechnologies.com/blog/google-glass-indoor-navigation/. [Accessed: 04- Oct- 2015].
[26] Washington Post, 'Everything you need to know about Google Glass', 2015. [Online]. Available:
https://www.washingtonpost.com/news/the-switch/wp/2014/02/27/everything-you-need-to-know-
about-google-glass/. [Accessed: 04- Oct- 2015].
[27] OrCam, 'OrCam - See for Yourself', 2015. [Online]. Available: http://www.orcam.com/. [Accessed:
27- Nov- 2015].
[28] S. Hicks, I. Wilson, L. Muhammed, J. Worsfold, S. Downes and C. Kennard, 'A Depth-Based Head-
Mounted Visual Display to Aid Navigation in Partially Sighted Individuals',PLOS ONE, vol. 8, no. 7, p.
e67695, 2013
[29] Ndcn.ox.ac.uk, 'Stephen Hicks — NDCN', 2015. [Online]. Available:
http://www.ndcn.ox.ac.uk/team/stephen-hicks. [Accessed: 07- Dec- 2015].
[30] Orbbec3d.com, 'Products | Orbbec', 2015. [Online]. Available: https://orbbec3d.com/products/.
[Accessed: 06- Oct- 2015].
[31] Indiegogo, 'CLICK HERE to support Orbbec Persee: World's First 3D Camera-Computer', 2015.
[Online]. Available: https://www.indiegogo.com/projects/orbbec-persee-world-s-first-3d-camera-
computer#/. [Accessed: 06- Oct- 2015].
[32] Wikipedia, 'Kinect', 2015. [Online]. Available: https://en.wikipedia.org/wiki/Kinect. [Accessed: 06-
Oct- 2015].
[33] Zhang, Z. (2012). Microsoft kinect sensor and its effect. MultiMedia, IEEE,19(2), 4-10.
[34] Gonzalez-Jorge, H., Riveiro, B., Vazquez-Fernandez, E., Martínez-Sánchez, J., & Arias, P. (2013).
Metrological evaluation of microsoft kinect and asus xtion sensors. Measurments, 46(6), 1800-1806.
[35] Asus.com, 'Xtion PRO LIVE - Overview', 2015. [Online]. Available: https://www.asus.com/us/3D-
Sensor/Xtion_PRO_LIVE/. [Accessed: 09- Oct- 2015].
[36]Amazon.com, 'Amazon.com: Asus Xtion PRO LIVE, 90IW0122-B01UA: Computers & Accessories',
2015. [Online]. Available: http://www.amazon.com/Asus-Xtion-PRO-LIVE-90IW0122-
B01UA/dp/B00CET0ZBO. [Accessed: 09- Nov- 2015].
[37] Stereolabs.com, 'ZED - Stereo Camera for Depth Sensing', 2015. [Online]. Available:
https://www.stereolabs.com/. [Accessed: 17- Nov- 2015].
[34] Oled-info.com, 'OLED introduction and basic OLED information | OLED-Info', 2015. [Online].
Available: http://www.oled-info.com/introduction. [Accessed: 06- Nov- 2015].
[35] w. 1.66 " '1.66 "flexible Oled Lcd Screen/tft Small Lcd Display/for Smart Phone,Smart
Watch,Wearable Watch - Buy Oled Lcd Screen,Flexible Lcd Screen,Tft Small Lcd Display Product on
Alibaba.com', www.alibaba.com, 2015. [Online]. Available: http://www.alibaba.com/product-detail/1-
66-flexible-OLED-lcd-screen_60044435510.html?spm=a2700.7724857.29.53.UHwlUu. [Accessed: 06-
Nov- 2015].
[36] 2015. [Online]. Available:
http://file:///C:/Users/User/Downloads/developing%20file%20%20NVK176SC001G-C-
0000%20for1.66%20inch.pdf. [Accessed: 06- Nov- 2015].
[37] Appliancedesign.com, 'Selecting the Best LCD Display for Your Product: OLED vs TFT', 2012. [Online].
Available: http://www.appliancedesign.com/articles/93121-selecting-the-best-lcd-display-for-your-
product-oled-vs-tft. [Accessed: 09- Nov- 2015].
[38] Amgoo.com, 'LCD Or OLED? What Is The Difference Between Smartphone Screen Types?', 2015.
[Online]. Available: http://www.amgoo.com/smartphone-blog/lcd-or-oled-what-is-the-difference-
between-smartphone-screen-types. [Accessed: 14- Nov- 2015].
[39] www.alibaba.com, '2 Inch Tft Lcd Module', 2015. [Online]. Available:
http://www.alibaba.com/product-detail/resolution-480-240-with-RGB-
interface_60374800113.html?spm=a2700.7724857.29.240.NQ5a6e. [Accessed: 19- Nov- 2015].
[40] Ebay.com, 'Electronics, Cars, Fashion, Collectibles, Coupons and More | eBay', 2015. [Online].
Available: http://www.ebay.com. [Accessed: 20- Nov- 2015].
[38] Oled-info.com, 'OLED introduction and basic OLED information | OLED-Info', 2015. [Online].
Available: http://www.oled-info.com/introduction. [Accessed: 06- Nov- 2015].
[39] w. 1.66 " '1.66 "flexible Oled Lcd Screen/tft Small Lcd Display/for Smart Phone,Smart
Watch,Wearable Watch - Buy Oled Lcd Screen,Flexible Lcd Screen,Tft Small Lcd Display Product on
Alibaba.com', www.alibaba.com, 2015. [Online]. Available: http://www.alibaba.com/product-detail/1-
66-flexible-OLED-lcd-screen_60044435510.html?spm=a2700.7724857.29.53.UHwlUu. [Accessed: 06-
Nov- 2015].
64
[40] 2015. [Online]. Available:
http://file:///C:/Users/User/Downloads/developing%20file%20%20NVK176SC001G-C-
0000%20for1.66%20inch.pdf. [Accessed: 06- Nov- 2015].
[41] Appliancedesign.com, 'Selecting the Best LCD Display for Your Product: OLED vs TFT', 2012. [Online].
Available: http://www.appliancedesign.com/articles/93121-selecting-the-best-lcd-display-for-your-
product-oled-vs-tft. [Accessed: 09- Nov- 2015].
[42] Amgoo.com, 'LCD Or OLED? What Is The Difference Between Smartphone Screen Types?', 2015.
[Online]. Available: http://www.amgoo.com/smartphone-blog/lcd-or-oled-what-is-the-difference-
between-smartphone-screen-types. [Accessed: 14- Nov- 2015].
[43] www.alibaba.com, '2 Inch Tft Lcd Module', 2015. [Online]. Available:
http://www.alibaba.com/product-detail/resolution-480-240-with-RGB-
interface_60374800113.html?spm=a2700.7724857.29.240.NQ5a6e. [Accessed: 19- Nov- 2015].
[44] Ebay.com, 'Electronics, Cars, Fashion, Collectibles, Coupons and More | eBay', 2015. [Online].
Available: http://www.ebay.com. [Accessed: 20- Nov- 2015].
65
Appendices:
1- System setup
2- LCD datasheet
3- Jetson TK1 datasheet
66
System Setup
67
LCD datasheet
68
69