Professional Documents
Culture Documents
Under Supervision of
Project Proposal
1. INTRODUCTION:
In today's tech-driven world, Android devices are ubiquitous, comprising a wide range
of smartphones and tablets. Android apps have become indispensable in enhancing the
functionality of these devices.
However, not all users have access to certain advanced features . Applications
integrating OCR, translation, and voice recognition functionalities have a rich history
that can be traced with the development of Optical Character Recognition technology,
initially used for reading characters from scanned documents. The concept of language
translation applications gained significant traction with the emergence of software
tools like Google Translate, which marked a significant advancement in facilitating
cross-lingual communication. Furthermore, the integration of voice recognition
capabilities led to the evolution of comprehensive Android applications that could
transcribe spoken language, ultimately revolutionizing the way people communicate
and interact globally.
2. MOTIVATION:
The creation of this app was motivated by a profound belief in the power of
communication and technology to dismantle language barriers, fostering a global
community where individuals of diverse linguistic backgrounds can seamlessly
connect and exchange ideas. Our mission was fueled by the frustration arising from
communication gaps in our increasingly interconnected world, driving us to empower
people and enrich their lives through a comprehensive tool that effortlessly converts
images to text, transcribes speech, and translates on-screen content. Our app not only
embodies technological innovation but also represents our dedication to promoting
2
Proposal Document
Aim:
Develop an Android application that seamlessly bridges language barriers,
allowing users to effortlessly understand and communicate in different
languages through multiple integrated functionalities under one umbrella.
Objectives:
Objectives for achieving this aim of our android app will consist of three core
features:
1. Image To Text Feature: will allow users to capture text from images using
Optical Character Recognition (OCR), enabling them to edit, translate, and
share the extracted text.In this feature we will extract the data of different
languages like English, Spanish, hindi, Chinese and others.
2. Speech To Text Feature: will facilitate spoken word transcription into text for
real-time translation and language understanding.
3. On-Screen Translation Feature: will introduce an innovative on-screen
cursor that can translate text from any location on the device's screen by just
placing the cursor on expected text, making it easy for users to comprehend
and communicate across different languages. User just have to move cursor to
the text and this feature will convert it into any language of user’s choice.
3
Proposal Document
4. PROBLEM STATEMENT:
To overcome language barriers, enhance communication, and improve
accessibility for diverse linguistic groups such as students, and professionals
through different features.
5. METHODOLOGY:
The development of the "All in One Translator" Android app will involve the
following key methodologies:
4
Proposal Document
Development Tools:
Android Studio: We will need Android Studio as the primary integrated
development environment (IDE) for building Android application.
Java: The app is developed using the Java programming language. Java is the
primary language for Android app development.
we'll need an OCR library or API to extract text from images and may require
image processing libraries or Android's built-in image processing capabilities.
To convert spoken words into written text, we'll need a speech recognition
library or API.
To provide real-time translation, we will use web scrapping technique that
supports various languages. We also need “allow over the app “ permission
that allows us to get the access over the screen while using another app .
Android Mobile Device: The app will run on Android devices, necessitating
testing and debugging on these devices.
Distribution Of Modules:
On Screen Translator:
In our Final Year Project (FYP), Sania Naveed will take charge of the On-Screen
Translator feature. She will be responsible for developing this feature, which allows
users to translate text instantly on their device screens. Users can perform translations
conveniently by simply hovering the cursor over the text, whether they are browsing
the web or engaged in chat conversations, anytime and anywhere.
Image to Text:
Sehar Zafar will be responsible for the Image to Text feature. She will be actively
involved in its development. This feature allows users to select or capture an image,
after which our app extracts all English text from the image. Subsequently, users have
two options: they can either translate the text into their desired language or create a
PDF document containing the extracted text.
5
Proposal Document
Speech to Text:
Tayyaba Riaz will be responsible for speech to text feature. This feature enables users
to speak in any language, with our app converting their voice into text. Users are then
presented with two options: they can either translate the text or create a PDF document
containing it. Additionally, Tayyaba will oversee the overall app design and the
implementation of smaller features, such as language selection for translation.
6. REFERENCES:
(PDF) Android application to convert speech to text and ... - researchgate. (n.d.-c).
https://www.researchgate.net/publication/332113272_Android_Application_to_Conve
rt_Speech_to_Text_and_Text_to_Speech
Publication, I. (2020, June 11). Text extraction from images using OCR. International
Journal for Research in Applied Science and Engineering Technology.
https://www.academia.edu/43313847/Text_Extraction_from_Images_Using_OCR
6
Proposal Document