You are on page 1of 4

SPEECH SYNTHESIZER SYSTEM

HMR Institute of Technology & Management


Hamidpur, Delhi

(Major Project)

TOPIC:

Speech
Synthesizer
System
Submitted By:

Ishaan prakesh (0151337208)


Jatin kataria (0131337208)
Nitin gupta (0091337208)

1|Page
SPEECH SYNTHESIZER SYSTEM

Objective:

The Speech synthesis project aimed at developing a gesture interface for driving
(“conducting”) a speech synthesis system. Four real-time gesture controlled
synthesis systems have been developed. For these “Synthesizers” are based on
formant synthesis and they include refined voice source components. One of them
is based on an augmented LF model (including an aperiodic component), the other
one is based on a Causal/Anticausal Linear Model of the voice source (CALM)
also augmented with an aperiodic component. All these systems are controlled by
various gesture devices. Informal testing and public demonstrations showed that
very natural and expressive synthetic voices can be produced in real time by some
combination of input devices/synthesis system.

Abstract
Speech synthesis is the artificial production of human speech. A hardware system
used for the purpose of speech synthesis is called a speech synthesizer. It can be
implemented in the hardware or software level. Example can be a text
to speech converter (TTS) which converts the input text into the speech. Some
other systems convert the input symbols or representations or phonetic symbols
into speech. The pieces of recorded speech stored in a database can be
concatenated to produce speech as we see in the railway station  announcement 
system where the pieces of speech are assembled to produce a
complete announcement. Entire words or sentences stored will produce the
maximum clarity in the case of real world applications. 

A (TTS) system consists of mainly two parts, a language processing module and a
signal processing module. In languages such as English, language processing is a
major part. All TTS systems existing allow users to specify what is to be spoken,
but do not give any control on how it has to spoken. A signal processing module
then will bring out this speech by making appropriate variations to the sound
database. It will then be possible for our program to sing/speak in a fashion that
one desires

2|Page
SPEECH SYNTHESIZER SYSTEM

Requirements

To use the Java Speech API, a user must have certain minimum software and
hardware available. The following is a broad sample of requirements. The
individual requirements of speech synthesizers and speech recognizers can vary
greatly and users should check product requirements closely.

 Speech software: A JSAPI-compliant speech recognizer or synthesizer is


required.
 System requirements: most desktop speech recognizers and some speech
synthesizers require relatively powerful computers to run effectively. Check
the minimum and recommended requirements for CPU, memory and disk
space when purchasing a speech product.

 Audio Hardware: Speech synthesizers require audio output. Speech


recognizers require audio input. Most desktop and laptop computers now
sold have satisfactory audio support. Most dictation systems perform better
with good quality sound cards.

 Microphone: Desktop speech recognition systems get audio input through a


microphone. Some recognizers, especially dictation systems, are sensitive to
the microphone and most recognition products recommend particular
microphones. Headset microphones usually provide best performance,
especially in noisy environments. Table-top microphones can be used in
some environments for some applications.
 Technical Details:

Software Requirements-
Platform Used : Java (JDK 1.6)

Framework : Synthesizer (Free TTS)

Hardware Requirements-
 Processor: Intel P-IV CPU 1.60 GHz.
 Minimum 1 GB RAM.
 Minimum 40 GB Hard disk.
 Wi-Fi Intranet in college premises.

3|Page
SPEECH SYNTHESIZER SYSTEM

 Separate system for being server for the application.

  Design Goals for the Java Speech API

Along with the other Java Media APIs, the Java Speech API lets developers
incorporate advanced user interfaces into Java applications. The design goals for
the Java Speech API included:

 Provide support for speech synthesizers and for both command-and-control


and dictation speech recognizers.

 Provide a robust cross-platform, cross-vendor interface to speech synthesis


and speech recognition.

 Enable access to state-of-the-art speech technology.

 Support integration with other capabilities of the Java platform, including


the suite of Java Media APIs.

 Be simple, compact and easy to learn.

References:
1. http://java.sun.com/products/java-media/speech/
2. http://tcts.fpms.ac.be/synthesis/maxmbrola/
3. http://www.disc2.dk/tools/SGsurvey/

4|Page

You might also like