Professional Documents
Culture Documents
MATHS Report
MATHS Report
Submitted by
Hanisha - 1RV22CS
2. Objectives
3. Methodology
5. Implementations
8. Conclusion
INTRODUCTION
Hardware Components:
Frame: Holds all the components together.
Pair of Glasses: Lenses held by the frame, potentially translucent.
Input Unit: Allows user interaction, including buttons for taking
photos and selecting languages.
Camera Unit: Captures images of text that need translation.
Projection Device: Rotatably coupled to the frame, used to
project translated text.
Software Components:
LINK:https://www.ripublication.com/ijaer18/ijaerv13n9_90.pdf
1.Glove-based approaches
Glove-based approaches involve wearing a sensor glove,
simplifying the segmentation process, while vision-based
approaches use image processing algorithms to detect
and track hand signs and facial expressions.
2. Vision-based approaches
Vision-based methods are preferred due to their ease of
use and absence of additional hardware requirements,
but they may have accuracy issues that need to be
addressed.
HARDWARE COMPONENTS
● Frame, pair of glasses
● Raspberry Pi 4 Model B
● Spoken Input : USB or I2S microphone
● Camera unit: To detect text and sign language
● Memory unit: language database
● Speakers to output the translated audio
SOFTWARE COMPONENTS
WORKING
1.Initialising:
Upon startup, the Raspberry Pi boots up and
initializes all necessary hardware components,
including the camera module, microphone,
speaker/headphones, and display module.
2.Input Acquisition:
The camera module captures real-time video
feed of the user's hand gestures. The
microphone captures spoken language input
from the user.
3.Gesture Recognition:
The captured video frames are processed using
image processing algorithms implemented on
the Raspberry Pi.Image segmentation
techniques are applied to isolate the hand
region from the background.Feature extraction
algorithms, such as contour detection and
keypoint extraction, are used to identify
relevant hand gestures.
4.Speech Recognition:
The captured spoken language input from the
microphone is processed using speech
recognition software libraries or services.The
speech recognition algorithms convert the
spoken language into text format, which serves
as the input for the translation process.
5.Language Translation:
The recognized sign language gestures and the
transcribed spoken language text are input into
the translation system. Language translation
software, such as Google Translate API or
Microsoft Translator, translates the spoken
language text into the desired target
language.For sign language translation, a
database of sign language gestures mapped to
corresponding spoken language translations is
used to translate recognized gestures into text
format.
6.Output Generation:
The translated text, both from the spoken
language and sign language inputs, is
displayed on the wearable display module.
7.User Interaction:The user interacts with the
system through an app, adjusting modes and
receiving real-time feedback for a tailored
translation experience.
8.Feedback and Optimization:The system
provides feedback to the user in the form of visual
and auditory cues, confirming successful
translation and providing assistance in case of
errors or misunderstandings.
IMPLEMENTATION
Image:
Camera: