SlideShare a Scribd company logo
IOSR Journal of Electronics and Communication Engineering (IOSR-JECE)
e-ISSN: 2278-2834,p- ISSN: 2278-8735.Volume 10, Issue 6, Ver. III (Nov - Dec .2015), PP 77-94
www.iosrjournals.org
DOI: 10.9790/2834-10637794 www.iosrjournals.org 77 | Page
Synthesized Speech using a small Microcontroller
Lalit Bhanwrela1
, Mohit Khamele2
Department of Electronics and Communication IES College of Technology
Bhopal, INDIA
Department of Electronics and Communication IES College of Technology
Bhopal, INDIA
Abstract: Speech was the natural form of human communication. Automatic speech signal generation on
computers is commonly called speech synthesis. Speech synthesizer can be created by concatenating pieces of
recorded audio speech that was stored in a database memory. Text-to-speech synthesis was an important object
in speech interface, which allows low-bandwidth text to supply a user with easy to gain the information. Speech
synthesis system has wide range of applications in our daily life. Speech synthesis plays an important role in
signal processing. In this paper a Speech synthesized system is designed which is based on microcontroller, a
project that performs several commands as defined for the synthesis of the speech signal.
Index Terms: Microcontoller, Control System
I. Introduction
Embedded systems play an important role in our day to day life. Due to limited works this systems can
be highly optimized by particular needs [27]. Some of the embedded systems applications were controlling the
devices with some amount of intelligence built-in [1].
Speech synthesis system has wide range of applications in our daily life. Speech synthesis plays an
important role in signal processing. The natural, intelligible synthetic voices in the expanding field of speech
technology had a lot of demand [4]. Speech was the natural form of human communication. Automatic speech
signal generation on computers is commonly called speech synthesis. Speech synthesis was the artificial
production of human speech [2]. Speech synthesizer can be created by concatenating pieces of recorded audio
speech that was stored in a database memory. Text-to-speech synthesis was an important object in speech
interface, which allows low-bandwidth text to supply a user with easy to gain the information. In speech synthesis
field, a number of research programmes have taken place during the last decade [13]. In 1984 an industry expert
Mr Drinkwater said that about the synthesis “synthesis is going to be the next bam-burning technology” [2]. Now
a day‟s synthesis speech was useful for many applications. In 1999 an American business magnate said that “In a
few years personal computers will talk back to us” [2].In the past (Allen et al., 1987), the text to speech synthesis
was done from a complete knowledge of the acoustics of the language being targeted.
For getting text to speech synthesis different techniques were used [5]. Concatenate method was the
simplest method for achieving the synthesis speech. By using this method the speech was generated by linking
pre recorded speech segments to build syllables, phrases, or words [2]. the applicable criteria that follow.
The block diagram for the system was shown in the following figure 1.
INPUT OUTPUT
Figure 1: Block diagram of the system
From the above figure the system will have four main sections; the microcontroller of the system would
take the commands from user and control the converter. The link between the user interface (computer) and the
microcontroller was two ways because after taking the some commands from user the system would also respond
with proper feedback. After processing the inputs the output will reflect through the speaker controlled by the
microcontroller which was illustrated by the figure 1.
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 78 | Page
II. Problem Analysis
This Paper was highlighted the requirements for developing the system. These requirements were
categorized into two major sections those were hardware requirements and software requirements. Based on the
project deliverables the requirements of the system were given below and clearly explained. The block box
testing‟s were also mentioned in the ending of this Paper.
2.1) Hardware requirements of the system
This project “synthesized speech using a small microcontroller” was a type embedded system so it
would be a combination of electronic hardware and software code. The following figure was show the system
would have electronic hardware capable of control the ISD board as shown in Fig. 2.1 (a).The following figure
was show the system would have electronic hardware capable of speech synthesizer which can controlled by the
microcontroller. The block diagram for synthesized speech using microcontroller was shown in the Figure
2.1(b)
Fig. 2.1 (a) Block diagram for audio record/playback system
Fig. 2.1 (b) Text to speech synthesis system block diagram
For this system a synthesizer was used for converting the text to speech.
2.1.1) Speech synthesizer
A synthesizer was required for getting the speech from text; it should be controlled by microcontroller.
The selection of the synthesizer would also be a challenging task there were many kind of synthesizers available
in the market also there were some techniques for hardware as well as software techniques were there for getting
speech from text. Software approach was good but in implementation quiet complex so hardware approach was
followed in this project. The synthesizer was required for this project which would capable to read the English
words, numbers and mathematical functions. The synthesizer chip should be light weight and communicate with
the host processor. It should have built in rule database for converting the text to speech.
2.1.2) Microcontroller
Microcontroller was the heart of this project or one can say it was the master as it controls the other
devices so, the selection of the microcontroller would also be a challenging task. Microcontrollers were
available from Microchip‟s PIC with a large range of size and features. In this project the priority was given to
microcontroller which one can control the operation of the ISD board and synthesizing the speech within that
microcontroller. If the microcontroller was smaller and light weight it would be better to keep the PCB small.
Microcontroller with internal features like internal oscillator, interrupts would be better. Microcontroller should
have the I/O pins for connecting to external modules. Before making any selection the size and power efficiency
could be considered.
2.1.3) Speaker
It was a device which converts electrical signal in to sound. Speaker was commonly used to hear
sounds. In this project the speaker should be used for hearing the synthesized speech in the output. A speaker
was required for this project which would be small, light weight and it should have low impedance, low power
ratings.
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 79 | Page
2.2) Speech synthesis methods:
Speech synthesis methods were used for converting the speech from the text. A synthesis method
which was used in this project, it should be simple and easy to implement the system.
2.3) Serial communication to establish a user interface
A module would be required to establish a user interface to control the system. The system should be
user control so it would be serially communicated with the host PC. RS232 and MM232r were used for the
serial communication.
2.4) Software requirements of the system
Software program was necessary for microcontroller to control the system. MPLAB was the program
development tool which can enable the user to write the program for the PIC microcontroller. MPLAB provided
the C18 compiler for writing the program. The program was supposed to write for control the ISD board as well
as to generate the synthesized speech.
2.4.1) Programming language, development and debugging tool
There were two popular languages (Assembly and embedded C) available for eight bit
microcontrollers, any language can be used for this system. Embedded C was quite similar to the normal C
language but only the difference was in embedded C programming the initializations and some special functions
were used, on the other hand the assembly instruction set was different to the manufacture of the microcontroller
from one to another. It was necessary before writing the program to know about the register set (SFR) of that
microcontroller. A program could be written in their own way.
The software tool MPLAB along the C18 compiler was providing the assembler and compiler.
Microcontroller did not understand assembly of the C program which was stored in the ROM of the PIC as a
.hex file. The assembler and compiler can converts the instructions to the machine code. The MPLAB was also
available the debugging features. By using the MBLAB ICD2 programmer and debugger kit the converted
hexadecimal file can be loaded into the microcontroller.
2.5) Black box testing
The output of the system was depending on the corresponding input. For this system the input would be user
commands and the output would be speaker. The black box testing should be done for the system.
I. The system should be user interface with the system.
II. The ISD board should be controlled by the microcontroller.
III. The speech should be synthesized within a microcontroller.
Fig. 2.5 Black box testing of the system
III. Proposed Problem Solutions
Based on the hardware and software requirements as mentioned before Paper, the solutions were given
for both hardware and software. According to the deliverables mentioned in the previous Paper, the first task
was to control audio record/playback ISD board and second one was generate the synthesized speech within the
microcontroller. The hardware and software solution were given below.
3.1) PIC18F1320 Microcontroller
PIC18F1320 was an 8 bit microcontroller developed by Microchip Company and it was power efficient
because used Nano watt technology. This microcontroller was used in this project because it has external
oscillator and many other features [30]. PIC18F microcontroller support enhanced USART module and supports
RS485, RS232 and LIN1.2 which has auto baud rate and also auto wake up start bit. The microcontroller
PIC188F1320 has four crystal modules (LP, XT, HS and HSPLL) and high computational performance. It could
be works at high speed up to 16-40MHZ. So PIC18F1320 (Fig 3.1) was used rather than PIC16F series (it works
at low speed). For more information about the PIC the references 30 and 27 were studied. It had two ports A and
B and port B pins were used in this project. In this project the microcontroller was used to control the audio
record/playback device and generating the synthesized speech.
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 80 | Page
Figure 3.1: PIC18F1320
3.2) Synthesis method
Speech synthesis technologies can be broadly categorized into three categories. They were given below [29]:
 Concatenate synthesis
 Formant synthesis
 Articulator synthesis
3.2.1) concatenate speech synthesis:
In this technology the speech can be generated by linking pre-recorded speech segments to make
syllables, phrases, or dip hones. Waveform segments were store in a database. It was the simplest method for
speech synthesis [14].
3.2.2) Formant synthesis:
Formants were defined as “resonant frequencies that occur at the main resonant areas of the vocal tract
for a given sound”. It consists of artificial reconstruction of the formant characteristics to be produced.
3.2.3) Articulator synthesis:
Articulators were speech organs. It was based on computational models of the human vocal tract and it
was a most natural sounding speech synthesis. But it was complex [4].
Finally concatenate method was identified, which was used for converting the speech from text [2]. In
this project for synthesizing the speech a synthesizer was used. This synthesizer chip follows the concatenate
method because it was a simplest method compare to other method.
3.3) TTS256 Synthesizer chip
Figure 3.3: TTS256 speech synthesizer chip
It was very hard to find out the synthesizer for this task because most of the synthesizers (WTS701,
SP0256 chips) were not available in the market. Finally after a long searching TTS256 chip was available in
speechchips.com. TTS256 was an 8 bit microprocessor and it was programmed with built in 600 letter-to-sound
rule database. TTS256 speech chip was a 28 pin DIP and compatible with PIC, BASIC stamp, OOPic [33]. It
can generate the speech code from ASCII text and it could be read English words, numbers and mathematical
functions. For serial communication parameters it was set the baud rate 9600. The datasheet mentioned in the
ref.[33] was studied. The host microcontroller can send up to 128 characters to the TTS256 chip via serial port
on 18th
pin (RX). The text which was entered by user was spoken when the TTS receives a new line character or
the maximum number of characters (128) has been entered.
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 81 | Page
3.4) Speaker
Figure 3.4: Speaker
The speaker selected in this project, which was having the following feature.
1) It was small and light weight
2) It had low power rating (0.5W) and Impedance (8 ohm)
3.5) MM232R module
It was a single chip asynchronous serial data transfer interface. In MM232R module FT232RQ chip
device was used and it can convert USB to RS232 / RS422 / RS485 converters. RS232 and MM232 were the
two solutions for serial communication. MM232R was support for 7 or 8 data bits and one or two stop bits. The
data transfer rates from range of 300 baud to 3 mega baud (RS422 / RS485) and 300 baud to 1 mega baud
(RS232). It can support for bus powered, high bus powered and self powered USB configurations. RS232 was
also used for serial data transfer [33]. If RS232 was used the MAX232 IC should be used for voltage levels, then
the cost of the PCB will increase so MM232 was used instead of RS232. The signal voltage levels of the
MM232 were in the range from 3.3v to 5v.
Figure 3.3: MM232R
3.6) Program language
Embedded C was used for the programming the PIC. MPLAB was used for writing the program of the
control system as well as speech synthesis system. MPLAB IDE was windows based Integrated Development
Environment for programming the microcontrollers. The MPLAB provided the ability to the users to create and
edit the source code in the editor window. In MPLAB IDE software tool the C18compiler was available to
assemble, compile and link the source code [31]. MPLAB ICD 2 debugger and programmer kit was used for
load the hexadecimal code to the PIC18F 1320 microcontroller.
Figure 3.6: MPLAB program structure
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 82 | Page
Flow chart for control system
The operation of the audio record and playback system was explained in the form of flow chart. The flow chart
for control system was shown in the following figure.
Figure 3.7: Flow chart for control PCB
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 83 | Page
Figure 3.7: Flow chart for Speech Synthesis System
IV. Problem Implementation
4.1 Hardware implementation of the control system
The solutions for getting the hardware and software code of the system were explained in the previous
Paper were finally implemented and explained in this Paper, for some of the hardware problems like
stabilization more than one solution were given. The outcomes of some solutions were necessary to discuss at
this at this as it was necessary to stick to one approach from here onward. According to the deliverables
mentioned in the first Paper the primary aim was control the ISD board and second one was to generate
synthesis speech system. The description of the hardware and software for these two tasks were separately given
below. The software programs were developed and listed in the appendix for both audio record/playback system
and synthesized speech system. Before implementing the control system to control the ISD board, the basic
operation of the ISD 1420P device was explained below.
Figure 4.1: ISD1420P PCB
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 84 | Page
4.1.1) Working of the audio record/playback device (ISD1420p board):
The system was configured with microphone, on-chip oscillator, speaker amplifier, smoothing filter,
and two push buttons. The PCB circuit device which was given by the supervisor was shown in the following
figure 4.1.1(b). The ISD1420P device was used for audio record/playback system. The following operating
sequence was demonstrated the working of chip device.
For recording the audio data: The device which was having two push buttons, one for recoding the data and
another was for play backing. When push button (REC) was pressed, then the REC signal LOW initiated a
record cycle from the starting of the data space [28]. When recording process was beginning the red LED was
glowing until the data space had been filled. Once the data space was filled then the recording was stopped. The
power of the device was automatically down then the REC signal was pulled to high.
For playback the audio data:
When the push button (PLAY) was pressed then active low PLAY signal initiated a playback cycle
from the starting of the data space which was recorded earlier. If the device was reached to EOM marker then
automatically power was down. While the process was going on the green LED was glowing until and unless
the EOM marker reached. [28]
4.1.2) Design of PCB for controlling of the Audio record/playback system
Without using of the PLAY and RECORD buttons the ISD1420 device should be operated. For this a
control PCB was designed. The schematic diagram for controlling the audio record and playback device system
was designed in Proteus ISIS. The Proteus ISIS professional provided the libraries for constructing the
schematic designs. From the libraries picked the PIC18F1320, resistors (10KΩ and 330Ω), LED and capacitor
then constructed the circuit for control PCB for controlling the ISD kit.
Figure: 4.2 Schematic diagram of Control PCB
The components PIC18F1320, resistors, LED and capacitor were available in the Proteus ISIS library
but the component MM232R was not available in the list of Proteus ISIS so four pin connectors were used. The
10K resistor used for avoiding the PIC reset and LED was used for power indication. Some additional elements
were also selected like 2 pin Molex connectors for power supply as well as for ISD board.
The following PCB was the implementation of the schematic discussed in the previous Paper. The PCB
was designed for controlling the audio record/playback device in Proteus Ares software. The PCB layout of the
audio record/playback system was shown in the following figure 4.1.2(b).
RA0/AN0
1
RA1/AN1/LVDIN
2
RA4/T0CKI
3
RA5/MCLR/VPP
4
RA2/AN2/VREF-
6
RA3/AN3/VREF+
7
RB0/AN4/INT0
8
RB1/AN5/TX/CK/INT1
9
RB4/AN6/RX/DT/KBI0
10
RB5/PGM/KBI1
11
RB6/PGC/T1OSO/T1CKI/P1C/KBI2
12
RB7/PGD/T1OSI/P1D/KBI3
13
RA6/OSC2/CLKO
15
RA7/OSC1/CLKI
16
RB2/P1B/INT2
17
RB3/CCP1/P1A
18
U1
PIC18F1320
PACKAGE=DIL18
VDD=GND
VSS=+5V
C2
10u
GND
+5V
GND
Speech Synthesizer, Venu Gopal, MSc Embedded Systems
R1
10k
1
2
3
4
8
7
6
5
J1
CONN-DIL8
1
2
3
4
J2
CONN-SIL4
1
2
3
4
J3
CONN-SIL4
+5V
1
2
5V SUPPLY TO BOARD
CONN-H2
1
2
ISD BOARD SWITCH CONNECTOR
CONN-H2
D1
DIODE-LED
R2
330R
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 85 | Page
Figure 4.1.2(b): PCB layout of the control pcb system.
Operation of the control system
For controlling of the ISD1420 record/playback device the control PCB was designed. MM232R was
used for sending data and receiving data [32]. The 5th
pin of the MM232R was connected to 10th
pin (RB4) of
the PIC18F1320 for receiving the data and 7th
pin of the MM232R was connected to the 9th
pin (RB1) of the PIC
for transmitting the data from the PIC. MM232R pins 2 and 3 were connected to 5v power supply. Based on the
command received from the user via MM232R the PIC will control the operation of record and playback device.
The RB2 pin of the PIC was connected to the 27th
pin (active low REC) ISD board and RB3 pin of the PIC was
connected to the 23rd
pin (active low PLAY) ISD board. Whenever the user enter the letter „r‟ on the
HyperTerminal the MM232R send this signal to the RX of the PIC. Then the PIC send an active low signal to
the ISD board, this active low signal (REC) was enable the pin then the will ready for recording. At the other
hand the PIC was send the ISD device status (START RECORDING) to the MM232R, this message will
display on the HyperTerminal window. Same way if the user entered the letter „p‟, the PIC sends an active low
signal to the ISD board then the device start to play. If any letter (except „r‟ or „p‟) the PIC was send the status
as „wrong command‟ to the MM232R.
4.1.4) Testing of the control system
The following hardware tests were done for the PCB.
1) For checking short circuit connectivity of the PCB initially the printed circuit board was tested without
placing components.
2) After placing the components these individual pins were tested through digital CRO.
3) The power of the components was tested by using the multi meter.
HyperTerminal was used for creating a user interface to control the audio record/playback device.
MM232R communication protocol was used for serial communication from host to the microcontroller for
sending and receiving the data. For creating a serial communication, the required HyperTerminal settings were
mentioned in the following figure 4.1.4(a).
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 86 | Page
Figure 4.1.4(a): HyperTerminal window settings
Software implementation for control PCB
The PIC18F1320 microcontroller was control the operation of the audio record/playback. For this the
PIC should be programmed, so the PIC C program was written by using MPLAB. To write the PIC C program
in the MPLAB first selected the new project wizard option, and then PIC18F1320 was selected for
microcontroller and C18 compiler was added. The complete PIC C program of control PCB was listed in
Appendix. The PIC C program was written in a file save with extension of .C and then this file was added to the
source file [27]. For clear explanation the whole program was divided as small sections. As the software code
was going to be in embedded C so the initial sections of every C program have a header file followed by some
declarations, definitions then the main and other functions. The MPLAB libraries provided all the header files
[31].
Figure 4.2(a): Error free code for control system
Hence the program for control system was explained clearly and this PIC C program was compiled by using
C18 compiler. The build succeeded program for the control system was shown in the following figure 4.2(a).
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 87 | Page
Figure 4.2(b): PIC programmed by using ICD2 kit
With the help of ICD2 kit the PIC was loaded to target device. The PIC was programmed on the ICD2 kit
compatibility board. This process was shown in the following figure 4.2(b).
Using the MPLAB ICD2 debugger and programmer kit the program successfully loaded into the PIC18F1320.
After completion of this loading process in MPLAB window the status of the ICD 2 kit was displayed and it was
shown in the figure 4.2(c).
Figure 4.2(c): Successful loaded of the program
After loading the program into the PIC then the microcontroller placed on the control PCB. The complete
hardware connections for control the audio record/playback system was shown in the following figure 4.2(d).
4.2 Hardware implementation for synthesis speech system
The main aim of the project was to generate the synthesis speech within a small microcontroller. The
schematic diagram for synthesis speech system was designed in Proteus ISIS. The Proteus ISIS professional
provided the libraries for constructing the schematic designs. The main PCB was could handle the both systems
those were it could control the ISD board and it could be generate the synthesis speech within a microcontroller.
The main PCB would reduce the cost of the cost and PCB area. The schematic diagram of the speech synthesis
system was shown in the following figure 4.3(a).
To develop the system the required components were selected by using Proteus ISIS professional
software. The component PIC18F1320, resistors, LED and capacitor were available in the Proteus ISIS library
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 88 | Page
but the components MM232R and TTS256 synthesizer chip were not available in the list of Proteus ISIS so
collection of four pin serial in line connectors were used for MM232R and four serial in line connectors were
used for construct the TT256 chip. The 10K resistor used for avoiding the PIC reset and LED was used for
power indication. Some additional elements were also selected like 2 pin Molex connectors for power supply as
well as for ISD board.
Figure 4.2(d): Complete hardware connections of the audio record/playback system
Figure 4.3(a): Schematic design for synthesis speech system
Pin allocation for microcontroller
Detail Pin No.
Indicator Pin B5
MM232R Rx, TTS256 Rx Pin B4 (Rx)
MM232R Tx Pin B1 (Tx)
Recording Pin B2
Playing Pin B3
RA0/AN0
1
RA1/AN1/LVDIN
2
RA4/T0CKI
3
RA5/MCLR/VPP
4
RA2/AN2/VREF-
6
RA3/AN3/VREF+
7
RB0/AN4/INT0
8
RB1/AN5/TX/CK/INT1
9
RB4/AN6/RX/DT/KBI0
10
RB5/PGM/KBI1
11
RB6/PGC/T1OSO/T1CKI/P1C/KBI2
12
RB7/PGD/T1OSI/P1D/KBI3
13
RA6/OSC2/CLKO
15
RA7/OSC1/CLKI
16
RB2/P1B/INT2
17
RB3/CCP1/P1A
18
U1
PIC18F1320
PACKAGE=DIL18
VDD=+5V
VSS=GND
C2
10u
GND
+5V
GND
Speech Synthesizer, Venu Gopal, MSc Embedded Systems
R1
10k
1
2
3
4
8
7
6
5
MM232 B
CONN-DIL8
1
2
3
4
MM232 A
CONN-SIL4
1
2
3
4
MM232 C
CONN-SIL4
+5V
1
2
5V SUPPLYTO BOARD
CONN-H2
1
2
ISDBOARDSWITCHCONNECTOR
CONN-H2
D1
DIODE-LED
R2
330R
1
2
SPKCONNECTION
CONN-SIL2
1
2
3
4
5
6
7
TTS B
CONN-SIL71
2
3
4
5
6
7
TTS A
CONN-SIL7
1
2
3
4
5
6
7
TTS C
CONN-SIL7
1
2
3
4
5
6
7
TTS D
CONN-SIL7
D2
DIODE-LED
R3
330R
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 89 | Page
Software implementation for Speech synthesis system
The PIC18F1320 microcontroller was used for generating the synthesis speech system. For this the PIC
should be programmed, so the PIC C program was written by using MPLAB. To write the PIC C program in the
MPLAB, first selected the new project wizard option, and then PIC18F1320 was selected for microcontroller
and C18 compiler was added. The complete PIC C program of the speech synthesis system was listed in
Appendix B. The PIC C program was written in a file, save it with extension of .C and then this file was added
to the source file. For clear explanation the whole program was divided as small sections. As the software code
was going to be in embedded C so the initial sections of every C program have a header file followed by some
declarations, definitions then the main and other functions. The MPLAB libraries provided all the header files.
Figure 4.4(a): Error free program
The above program was loaded in to the PIC18F 1320 microcontroller with the help of MPLAB ICD2
debugger and programmer kit provided by the university. The following figure 33 was show the message after
loading the program in to the PIC.
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 90 | Page
Figure 4.4(b): Successful loading of the program
Working of the speech synthesis system
HyperTerminal was used for creating a user interface to control the ISD board and synthesis speech
system. MM232R communication protocol was used for serial communication from host to the microcontroller
for sending and receiving the data. For creating a serial communication, the required HyperTerminal settings
were mentioned in the following figure. After loading the program into the PIC then the microcontroller placed
on the synthesis speech system PCB. The complete hardware connections for the speech synthesis system were
shown in the following figure 4.5(b).
MM232R serial communication module was used for user interface. The user could write the text on
hyper terminal window, the corresponding speech will be produced by the speaker. Two LEDs were used in this
system one was for power indication (+5V) and second one indicates whether the program was loaded in to the
PIC microcontroller or not. The resistor 10K ohm was used for avoid the reset of the PIC.
Figure 4.5(a): Hyper terminal settings for serial communication
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 91 | Page
The final PCB would do the both tasks. The data which was entered on the hyper terminal will send to
the 10th
pin (RX) of the PIC18F1320 microcontroller then the received data will sent to the 18th
pin (RX) of the
TTS256 synthesizer chip. The TTS256 synthesizer chip can convert the data which was sent by microcontroller
to speech. The 24th
pin of the TTS256 chip was connected to the speaker, 14th
pin was connected to ground and
28th
pin was connected to 5V power supply. The text which was entered by user was spoken when the TTS256
chip receives a new line character or the maximum number of characters (128) has been entered. By using this
PCB the ISD board was also controlled. If the user entered the letter „r‟ on the hyper terminal, the
microcontroller will send the active low signal to the ISD board then the ISD board start recording the voice.
Simultaneously when the user entered the letter „p‟ on the hyper terminal, the microcontroller will send the
active low signal to the ISD board then device could start playing the sound which was previously recorded in
the ISD chip.
V. Results
In this Paper the system outcomes were explained clearly. Initially the control PCB was designed for
control the audio record/playback device (ISD board). The main task of the project synthesis speech systems
results were given and explained in the last section.
5.1) Results of the control system
Based on the command which was entered by the user on the hyper terminal the microcontroller
controls the ISD board. When the user entered command „r‟ on hyper terminal then the microcontroller send the
active low signal to the ISD board the system starts recording. The microcontroller sends the status of the ISD
board to the 7th
pin (RX) of the MM232R and it was displayed on the hyper terminal. The red LED of the ISD
board blink until the system recording the voice. The following figures (5.1(a) and 5.1(b)) were show the
process of the system.
Figure 5.1(a): Computer received the data from the PIC when the system starts recording
Figure 5.1(b): ISD board when recording the voice
Same way when the user entered the command „p‟ on hyper terminal then the microcontroller send the
active low signal to the ISD board the system starts recording. The microcontroller sends the status of the ISD
board to the 7th
pin (RX) of the MM232R and it was displayed on the hyper terminal. The green LED of the ISD
board blink until the system would play the voice which was stored previously in the ISD1420P. The following
figures (5.1(c) and 5.1(d)) were show the process of the system. When the user entered wrong command then
the PIC would send message to the computer and it was displayed on the hyper terminal as “wrong command”.
„r‟ entered by the user
Status of the ISD board
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 92 | Page
Figure 5.1(c): Computer received the data from the PIC when the system start play
Outcome for the speech synthesis system
The final PCB which was the main PCB, designed for the controlling of the ISD board and synthesis speech
system. The following figure which was shows the two LED indicators, one was for power indication and
another one for checking the PIC programmed or not.
Figure 5.1(d): ISD board when playing the voice
Figure 5.2: The speech synthesis system
In the above figure show the LED power indicator was blinking so the 5V power was coming to circuit.
Second one was also blinking so the program was also loaded into the PIC. According to the software program
the LED blinks up to 55msec but the LED was continuously blinking. Also when the user trying to enter any
character on the hyper terminal window and it was not displayed and also the output speech was not coming
from the speaker. The system could not work because the software might not match with the hardware
connections of the system or might be problem in the software code. Many changes were done in case of
modifying the software code as well as in hardware some connections was changed.
VI. Discussion On Results
Well the results of the complete system were as expected, but in sometimes it was not picking up the
commands entered by the user through the hyper terminal. There was some problem with the playback of the
sound possible causes may be compatibility issues of TTS IC. The Embedded C program was logical right and
„p‟ entered by the user
Status of the ISD board
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 93 | Page
there were no compiling errors. There was no RJ45 connector on the target board so it was not possible to get
the live debug simulation on the MPLAB.
The sound of the speaker was not that good may be because of the quality of the device, but on the
other hand it was also not possible to attach a bid speaker directly to the TTS IC as it cannot supply the required
source current. Behavior of complete system was good system was capable of doing task to some extent for
what it was designed to. With this technical discussion for this project comes to an end.
VII. Conclusion
After going through this task it was evident that to develop complete system even if its looks simple at
first look lots of effort and dedication is required. Various fields required to be combined like good software
knowledge, electronics and analytical skills. Sometimes a simple appearing task may take a long time to work
accordingly, as happened while tried to achieve a proper communication between microcontroller and the TTS
IC. Every element is required to be looked after. Analyzing why something fails or did not worked as required
helps a lot and makes the vision clear brings success closer and enhances knowledge.
The work on the project was quite satisfactory, but still there were some shortcomings like sometimes
system was not working at all, so some changes are suggested in software and electronic design. Another major
addition that could be done in future design is the addition of on board debugging mechanism to attach the ICD
kit to the main board this could be easily done by adding a RJ connector. Various features like audio filter
circuit which would only allow the audible frequencies to pass to the speaker could be added along with some
amplifier and volume control.
In the embedded program additional commands could be added like for volume control, the duration
for which TTS is required to be in play or record mode. Some features like to monitor the errors on the system
could be added and if any error occurs then could be displayed on the hyper terminal, and even intimate the user
by producing speech from the TTS. There is no limit as such to the additions that could be done in the future to
this system as the developed project is just the basic one so many enhancements are possible.
References
[1]. Juang, B.H. Tsuhan Chen, The past, present, and future of speech processing Volume 15, Issue 3, ISSN: 1053-5888, Pages 24-48.
[2]. Henton, Caroline, Challenges and Rewards in Using Parametric or Concatenative Speech synthesis, ISSN: 1381-2416.
[3]. Kim, Jongkuk - Hahn, Hernsoo – Yoon, Uei-Joong-Bae, Myungjin, On a pitch alteration for speech synthesis systems, year 2009,
volume 50, issue 4, pages 435-446, ISSN:0929-6212.
[4]. Rahier, M.C. , Defraeye, P.J. Guebels, P.-P, Patovan, B, A 3 Âľm NMOS high performance LPC speech synthesizer chip, Volume
18, Issue 3, Pages 349-359, ISSN:0018-9200.
[5]. Henton, Caroline, Challenges and rewards in using parametric or concatenative speech synthesis, Year
[6]. 2002, Volume 5, Issue 2, Pages 117-131, ISSN: 1381-2416.
[7]. Holfelder, Wieland, Interactive remote recording and playback of multicast videoconferences, Year 1998, Volume 21, Issue 15,
Pages 1285-1294, ISSN: 0140-3664.
[8]. Kim, Jong Kuk-Hahn, Hern Soo-Bae, Myung Jin, On a speech multiple system implementation foe synthesis, year 2009, volume 49,
issue 4, ISSN: 0929-6212.
[9]. Hilt, V. Mauve, M. Vogel, J. Effelsberg, W, Recording and playing back interactive media streams, Date Oct 2005, volume 7 issue
5,pages 960-971, ISSN: 1520-9210
[10]. A. K. Rath and P.K. Meher, Design of a merged DSP microcontroller for embedded systems using discrete orthogonal transform,
Year 2006, pages 388-394, ISSN: 1549-3636.
[11]. Schroeder, E.F. platte, H. –J. Krahe, D., ‘MSC’ :stereo audio coding with CD-quality and 256 kBit/sec, Date: Nov.1987, Volume
CE-33, Issue 4, Pages 512-519, ISSN: 0098-3063.
[12]. Geoff Jackson, Peter Holzmann, A Single-chip Text-to-Speech Synthesis Device Utilizing Analog Nonvolatile Multilevel Flash
Storage, Issue Date: NOV.2002, ISSN: 0018-9200.
[13]. Liu, Fu-Hua-Gu, Liang-Gao, Machael, Applications of Language Modelling in Speech-to-Speech Translation table of content, Year-
2004, ISSN: 1381-2416.
[14]. Lee, C, -H; Jung, S.-K. ; Kang, H. –G., Applying a speaker-Dependent speech compression Technique to concatenative TTS
synthesizers, Year- 2007, ISSN: 1558-7916.
[15]. Henton, Caroline, Challenges and Rewards in using Parametric or concatenative speech synthesis, Year- 2002, Issn: 1381-2416.
[16]. Goldsmith, John, Dealing with Prosody in a tex-to-speech system table of content, Year- 1999, ISSN: 1381-2416.
[17]. Sotiris Karabetsos, Pirros Tsiakoulis, Embedded unit selection Text-to-speech synthesis, Year- 2009, ISSN: 0098-3063.
[18]. Kim, Jong kuk-Hahn, Hern Soo- Bae, Myung Jin, On a speech Multiple system Implementation for speech synthesis table of
content, Year- 2009, ISSN: 0929-6212.
[19]. Taylor & Francis, Technical notes on Audio Recording, Year- 2000, ISSN: 0835-1813.
[20]. El-Imam, Yousif A. –Don, Zuraida Mohammed, Text-to-Speech conversion of standard Malay, Year- 1999, ISSN: 1381-2416.
[21]. Steve Heath, Embedded System Design, Second Edition, ISBN 0-7506-5546-1.
[22]. Holzmann, P. Jackson, G.B. Raina, A. Hung-chuan pai Ming-Bing Chang Awsare, S.V. Engh, L.D. Kao,O.C. Palmer, C.R. Chun-
Mai Liu Kordesch, A.V. Ken su Hemming,M. An analog and digital record, playback and processing system using a coarse-fine
programming method, Date 18-20 April 2001, pages 192-195, ISBN: 0-7803-6412-0.
[23]. Embedded C by Michale J Pont, Pearson International, ISBN-0173290532, year of publication 2002
[24]. Rice, S.V. Bailey, S.M, An integrated environment for audio recording,editing, and retrieval, Date Dec 2006, ISBN: 0-7695-2746-
9.
[25]. Kumar, K. Raja Ramaiah, P. Seetha, DSP and Microcontroller based speech processor for auditory prosthesis, Date 20-23 Dec
2006, pages 518-522, ISBN: 1-4244-0716-8.
Synthesized Speech using a small Microcontroller
DOI: 10.9790/2834-10637794 www.iosrjournals.org 94 | Page
[26]. Thiang, Implementation of speech recognition on MCS51 microcontroller for controlling wheelchair, Date : 25-28 Nov 2007, pages
1193-1198, ISBN: 978-1-4244-1355-3.
[27]. Sarathy, K.P. Ramakrishnan, A.G, A research bed for unit selection based text to speech synthesis, Date 15-19 Dec 2008, pages
229-232, ISBN: 978-1-4244-3471-8.
[28]. PIC Microcontroller and Embedded Systems by Mazidi, Muhammad Ali, Pearson International, ISBN- 0136009026, Year of
publication 2007.

More Related Content

What's hot

iCognos: Cognitive Stimulation through te
iCognos: Cognitive Stimulation through teiCognos: Cognitive Stimulation through te
iCognos: Cognitive Stimulation through te
Sergi Torrellas
 
Template abstract_book
Template  abstract_bookTemplate  abstract_book
Template abstract_book
Madhuri Yellapu
 
CDR2(Sajjad Tarahomi)
CDR2(Sajjad Tarahomi)CDR2(Sajjad Tarahomi)
CDR2(Sajjad Tarahomi)
Sajjad Tarahomi
 

What's hot (16)

mobile control robot using DTMF module
mobile control robot using DTMF modulemobile control robot using DTMF module
mobile control robot using DTMF module
 
Trends and Implications in Embedded Systems Development
Trends and Implications in Embedded Systems DevelopmentTrends and Implications in Embedded Systems Development
Trends and Implications in Embedded Systems Development
 
iCognos: Cognitive Stimulation through te
iCognos: Cognitive Stimulation through teiCognos: Cognitive Stimulation through te
iCognos: Cognitive Stimulation through te
 
Pen doc
Pen docPen doc
Pen doc
 
IRJET- Voice Assisted Text Reading and Google Home Smart Socket Control Syste...
IRJET- Voice Assisted Text Reading and Google Home Smart Socket Control Syste...IRJET- Voice Assisted Text Reading and Google Home Smart Socket Control Syste...
IRJET- Voice Assisted Text Reading and Google Home Smart Socket Control Syste...
 
Embedded systems ppt
Embedded systems pptEmbedded systems ppt
Embedded systems ppt
 
IRJET - Mirroring of Source and Sink Devices in Android Screen Casting
IRJET - Mirroring of Source and Sink Devices in Android Screen CastingIRJET - Mirroring of Source and Sink Devices in Android Screen Casting
IRJET - Mirroring of Source and Sink Devices in Android Screen Casting
 
Embedded systems: Future perspectives
Embedded systems: Future perspectivesEmbedded systems: Future perspectives
Embedded systems: Future perspectives
 
Home automation using dtmf
Home automation using dtmfHome automation using dtmf
Home automation using dtmf
 
Embedded ARM Control Robotic Arm using BoaWeb server – a Survey
Embedded ARM Control Robotic Arm using BoaWeb server – a SurveyEmbedded ARM Control Robotic Arm using BoaWeb server – a Survey
Embedded ARM Control Robotic Arm using BoaWeb server – a Survey
 
Template abstract_book
Template  abstract_bookTemplate  abstract_book
Template abstract_book
 
Embedded systems The Past Present and the Future
Embedded systems The Past Present and the FutureEmbedded systems The Past Present and the Future
Embedded systems The Past Present and the Future
 
CDR2(Sajjad Tarahomi)
CDR2(Sajjad Tarahomi)CDR2(Sajjad Tarahomi)
CDR2(Sajjad Tarahomi)
 
Introduction To Embedded Systems
Introduction To Embedded SystemsIntroduction To Embedded Systems
Introduction To Embedded Systems
 
Comparison of Cinepak, Intel, Microsoft Video and Indeo Codec for Video Compr...
Comparison of Cinepak, Intel, Microsoft Video and Indeo Codec for Video Compr...Comparison of Cinepak, Intel, Microsoft Video and Indeo Codec for Video Compr...
Comparison of Cinepak, Intel, Microsoft Video and Indeo Codec for Video Compr...
 
Introduction to Embedded Systems I: Chapter 2 (1st portion)
Introduction to Embedded Systems I: Chapter 2 (1st portion)Introduction to Embedded Systems I: Chapter 2 (1st portion)
Introduction to Embedded Systems I: Chapter 2 (1st portion)
 

Viewers also liked

Viewers also liked (20)

D0441318
D0441318D0441318
D0441318
 
A011120104
A011120104A011120104
A011120104
 
S01821121124
S01821121124S01821121124
S01821121124
 
F017652530
F017652530F017652530
F017652530
 
E017642734
E017642734E017642734
E017642734
 
D017612529
D017612529D017612529
D017612529
 
C010111117
C010111117C010111117
C010111117
 
Simulation based Evaluation of a Simple Channel Distribution Scheme for MANETs
Simulation based Evaluation of a Simple Channel Distribution Scheme for MANETsSimulation based Evaluation of a Simple Channel Distribution Scheme for MANETs
Simulation based Evaluation of a Simple Channel Distribution Scheme for MANETs
 
O017148084
O017148084O017148084
O017148084
 
Model-based Approach of Controller Design for a FOPTD System and its Real Tim...
Model-based Approach of Controller Design for a FOPTD System and its Real Tim...Model-based Approach of Controller Design for a FOPTD System and its Real Tim...
Model-based Approach of Controller Design for a FOPTD System and its Real Tim...
 
Qualitative Chemistry Education: The Role of the Teacher
Qualitative Chemistry Education: The Role of the TeacherQualitative Chemistry Education: The Role of the Teacher
Qualitative Chemistry Education: The Role of the Teacher
 
C018111623
C018111623C018111623
C018111623
 
E010224043
E010224043E010224043
E010224043
 
Recent Developments and Analysis of Electromagnetic Metamaterial with all of ...
Recent Developments and Analysis of Electromagnetic Metamaterial with all of ...Recent Developments and Analysis of Electromagnetic Metamaterial with all of ...
Recent Developments and Analysis of Electromagnetic Metamaterial with all of ...
 
Cash Flow Valuation Mode Lin Discrete Time
Cash Flow Valuation Mode Lin Discrete TimeCash Flow Valuation Mode Lin Discrete Time
Cash Flow Valuation Mode Lin Discrete Time
 
Nutritional Analysis of Edible Wild Fruit (Zizyphus Jujuba Mill.) Used By Rur...
Nutritional Analysis of Edible Wild Fruit (Zizyphus Jujuba Mill.) Used By Rur...Nutritional Analysis of Edible Wild Fruit (Zizyphus Jujuba Mill.) Used By Rur...
Nutritional Analysis of Edible Wild Fruit (Zizyphus Jujuba Mill.) Used By Rur...
 
Improvement of Congestion window and Link utilization of High Speed Protocols...
Improvement of Congestion window and Link utilization of High Speed Protocols...Improvement of Congestion window and Link utilization of High Speed Protocols...
Improvement of Congestion window and Link utilization of High Speed Protocols...
 
E0432933
E0432933E0432933
E0432933
 
Cryptography On Android Message Application Using Look Up Table And Dynamic ...
Cryptography On Android Message Application Using Look  Up Table And Dynamic ...Cryptography On Android Message Application Using Look  Up Table And Dynamic ...
Cryptography On Android Message Application Using Look Up Table And Dynamic ...
 
C1303061622
C1303061622C1303061622
C1303061622
 

Similar to N010637794

Iirdem design and implementation of finger writing in air by using open cv (c...
Iirdem design and implementation of finger writing in air by using open cv (c...Iirdem design and implementation of finger writing in air by using open cv (c...
Iirdem design and implementation of finger writing in air by using open cv (c...
Iaetsd Iaetsd
 
Voice Controlled Intelligent Remote for Consumer Electronics
Voice Controlled Intelligent Remote for Consumer ElectronicsVoice Controlled Intelligent Remote for Consumer Electronics
Voice Controlled Intelligent Remote for Consumer Electronics
NITIN DESAI
 
Bq4301381388
Bq4301381388Bq4301381388
Bq4301381388
IJERA Editor
 
Module 2 3
Module 2 3Module 2 3
Module 2 3
ryanette
 
2nd ARM Developer Day - mbed Workshop - ARM
2nd ARM Developer Day - mbed Workshop - ARM2nd ARM Developer Day - mbed Workshop - ARM
2nd ARM Developer Day - mbed Workshop - ARM
Antonio Mondragon
 

Similar to N010637794 (20)

Iirdem design and implementation of finger writing in air by using open cv (c...
Iirdem design and implementation of finger writing in air by using open cv (c...Iirdem design and implementation of finger writing in air by using open cv (c...
Iirdem design and implementation of finger writing in air by using open cv (c...
 
An introduction to digital signal processors 1
An introduction to digital signal processors 1An introduction to digital signal processors 1
An introduction to digital signal processors 1
 
Desktop assistant
Desktop assistant Desktop assistant
Desktop assistant
 
Automatic Irrigation System Project Report
Automatic Irrigation System Project ReportAutomatic Irrigation System Project Report
Automatic Irrigation System Project Report
 
Voice Controlled Intelligent Remote for Consumer Electronics
Voice Controlled Intelligent Remote for Consumer ElectronicsVoice Controlled Intelligent Remote for Consumer Electronics
Voice Controlled Intelligent Remote for Consumer Electronics
 
IDE Code Compiler for the physically challenged (Deaf, Blind & Mute)
IDE Code Compiler for the physically challenged (Deaf, Blind & Mute)IDE Code Compiler for the physically challenged (Deaf, Blind & Mute)
IDE Code Compiler for the physically challenged (Deaf, Blind & Mute)
 
Automatic Control of Instruments Using Efficient Speech Recognition Algorithm
Automatic Control of Instruments Using Efficient Speech Recognition AlgorithmAutomatic Control of Instruments Using Efficient Speech Recognition Algorithm
Automatic Control of Instruments Using Efficient Speech Recognition Algorithm
 
Design and Development of ARM9 Based Embedded Web Server
Design and Development of ARM9 Based Embedded Web ServerDesign and Development of ARM9 Based Embedded Web Server
Design and Development of ARM9 Based Embedded Web Server
 
Bq4301381388
Bq4301381388Bq4301381388
Bq4301381388
 
Embedded systems
Embedded systemsEmbedded systems
Embedded systems
 
Web-Based Lighting Automation System
Web-Based Lighting Automation SystemWeb-Based Lighting Automation System
Web-Based Lighting Automation System
 
Module 2 3
Module 2 3Module 2 3
Module 2 3
 
Real Time Direct Speech-to-Speech Translation
Real Time Direct Speech-to-Speech TranslationReal Time Direct Speech-to-Speech Translation
Real Time Direct Speech-to-Speech Translation
 
Speech & Web based appliance control
Speech & Web based appliance controlSpeech & Web based appliance control
Speech & Web based appliance control
 
2nd ARM Developer Day - mbed Workshop - ARM
2nd ARM Developer Day - mbed Workshop - ARM2nd ARM Developer Day - mbed Workshop - ARM
2nd ARM Developer Day - mbed Workshop - ARM
 
Ijetr042175
Ijetr042175Ijetr042175
Ijetr042175
 
Controlling Home Appliances by using Universal Remote Control System (IoT and...
Controlling Home Appliances by using Universal Remote Control System (IoT and...Controlling Home Appliances by using Universal Remote Control System (IoT and...
Controlling Home Appliances by using Universal Remote Control System (IoT and...
 
R43019698
R43019698R43019698
R43019698
 
voip
voipvoip
voip
 
Assistive device for blind , deaf and dumb.pptx
Assistive device for blind , deaf and dumb.pptxAssistive device for blind , deaf and dumb.pptx
Assistive device for blind , deaf and dumb.pptx
 

More from IOSR Journals

More from IOSR Journals (20)

A011140104
A011140104A011140104
A011140104
 
M0111397100
M0111397100M0111397100
M0111397100
 
L011138596
L011138596L011138596
L011138596
 
K011138084
K011138084K011138084
K011138084
 
J011137479
J011137479J011137479
J011137479
 
I011136673
I011136673I011136673
I011136673
 
G011134454
G011134454G011134454
G011134454
 
H011135565
H011135565H011135565
H011135565
 
F011134043
F011134043F011134043
F011134043
 
E011133639
E011133639E011133639
E011133639
 
D011132635
D011132635D011132635
D011132635
 
C011131925
C011131925C011131925
C011131925
 
B011130918
B011130918B011130918
B011130918
 
A011130108
A011130108A011130108
A011130108
 
I011125160
I011125160I011125160
I011125160
 
H011124050
H011124050H011124050
H011124050
 
G011123539
G011123539G011123539
G011123539
 
F011123134
F011123134F011123134
F011123134
 
E011122530
E011122530E011122530
E011122530
 
D011121524
D011121524D011121524
D011121524
 

Recently uploaded

Search and Society: Reimagining Information Access for Radical Futures
Search and Society: Reimagining Information Access for Radical FuturesSearch and Society: Reimagining Information Access for Radical Futures
Search and Society: Reimagining Information Access for Radical Futures
Bhaskar Mitra
 

Recently uploaded (20)

To Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMsTo Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMs
 
Salesforce Adoption – Metrics, Methods, and Motivation, Antone Kom
Salesforce Adoption – Metrics, Methods, and Motivation, Antone KomSalesforce Adoption – Metrics, Methods, and Motivation, Antone Kom
Salesforce Adoption – Metrics, Methods, and Motivation, Antone Kom
 
JMeter webinar - integration with InfluxDB and Grafana
JMeter webinar - integration with InfluxDB and GrafanaJMeter webinar - integration with InfluxDB and Grafana
JMeter webinar - integration with InfluxDB and Grafana
 
AI for Every Business: Unlocking Your Product's Universal Potential by VP of ...
AI for Every Business: Unlocking Your Product's Universal Potential by VP of ...AI for Every Business: Unlocking Your Product's Universal Potential by VP of ...
AI for Every Business: Unlocking Your Product's Universal Potential by VP of ...
 
Connector Corner: Automate dynamic content and events by pushing a button
Connector Corner: Automate dynamic content and events by pushing a buttonConnector Corner: Automate dynamic content and events by pushing a button
Connector Corner: Automate dynamic content and events by pushing a button
 
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...
 
UiPath Test Automation using UiPath Test Suite series, part 1
UiPath Test Automation using UiPath Test Suite series, part 1UiPath Test Automation using UiPath Test Suite series, part 1
UiPath Test Automation using UiPath Test Suite series, part 1
 
Search and Society: Reimagining Information Access for Radical Futures
Search and Society: Reimagining Information Access for Radical FuturesSearch and Society: Reimagining Information Access for Radical Futures
Search and Society: Reimagining Information Access for Radical Futures
 
Designing Great Products: The Power of Design and Leadership by Chief Designe...
Designing Great Products: The Power of Design and Leadership by Chief Designe...Designing Great Products: The Power of Design and Leadership by Chief Designe...
Designing Great Products: The Power of Design and Leadership by Chief Designe...
 
Behind the Scenes From the Manager's Chair: Decoding the Secrets of Successfu...
Behind the Scenes From the Manager's Chair: Decoding the Secrets of Successfu...Behind the Scenes From the Manager's Chair: Decoding the Secrets of Successfu...
Behind the Scenes From the Manager's Chair: Decoding the Secrets of Successfu...
 
Mission to Decommission: Importance of Decommissioning Products to Increase E...
Mission to Decommission: Importance of Decommissioning Products to Increase E...Mission to Decommission: Importance of Decommissioning Products to Increase E...
Mission to Decommission: Importance of Decommissioning Products to Increase E...
 
UiPath Test Automation using UiPath Test Suite series, part 3
UiPath Test Automation using UiPath Test Suite series, part 3UiPath Test Automation using UiPath Test Suite series, part 3
UiPath Test Automation using UiPath Test Suite series, part 3
 
Exploring UiPath Orchestrator API: updates and limits in 2024 🚀
Exploring UiPath Orchestrator API: updates and limits in 2024 🚀Exploring UiPath Orchestrator API: updates and limits in 2024 🚀
Exploring UiPath Orchestrator API: updates and limits in 2024 🚀
 
De-mystifying Zero to One: Design Informed Techniques for Greenfield Innovati...
De-mystifying Zero to One: Design Informed Techniques for Greenfield Innovati...De-mystifying Zero to One: Design Informed Techniques for Greenfield Innovati...
De-mystifying Zero to One: Design Informed Techniques for Greenfield Innovati...
 
UiPath Test Automation using UiPath Test Suite series, part 2
UiPath Test Automation using UiPath Test Suite series, part 2UiPath Test Automation using UiPath Test Suite series, part 2
UiPath Test Automation using UiPath Test Suite series, part 2
 
Free and Effective: Making Flows Publicly Accessible, Yumi Ibrahimzade
Free and Effective: Making Flows Publicly Accessible, Yumi IbrahimzadeFree and Effective: Making Flows Publicly Accessible, Yumi Ibrahimzade
Free and Effective: Making Flows Publicly Accessible, Yumi Ibrahimzade
 
Measures in SQL (a talk at SF Distributed Systems meetup, 2024-05-22)
Measures in SQL (a talk at SF Distributed Systems meetup, 2024-05-22)Measures in SQL (a talk at SF Distributed Systems meetup, 2024-05-22)
Measures in SQL (a talk at SF Distributed Systems meetup, 2024-05-22)
 
Speed Wins: From Kafka to APIs in Minutes
Speed Wins: From Kafka to APIs in MinutesSpeed Wins: From Kafka to APIs in Minutes
Speed Wins: From Kafka to APIs in Minutes
 
IoT Analytics Company Presentation May 2024
IoT Analytics Company Presentation May 2024IoT Analytics Company Presentation May 2024
IoT Analytics Company Presentation May 2024
 
Custom Approval Process: A New Perspective, Pavel Hrbacek & Anindya Halder
Custom Approval Process: A New Perspective, Pavel Hrbacek & Anindya HalderCustom Approval Process: A New Perspective, Pavel Hrbacek & Anindya Halder
Custom Approval Process: A New Perspective, Pavel Hrbacek & Anindya Halder
 

N010637794

  • 1. IOSR Journal of Electronics and Communication Engineering (IOSR-JECE) e-ISSN: 2278-2834,p- ISSN: 2278-8735.Volume 10, Issue 6, Ver. III (Nov - Dec .2015), PP 77-94 www.iosrjournals.org DOI: 10.9790/2834-10637794 www.iosrjournals.org 77 | Page Synthesized Speech using a small Microcontroller Lalit Bhanwrela1 , Mohit Khamele2 Department of Electronics and Communication IES College of Technology Bhopal, INDIA Department of Electronics and Communication IES College of Technology Bhopal, INDIA Abstract: Speech was the natural form of human communication. Automatic speech signal generation on computers is commonly called speech synthesis. Speech synthesizer can be created by concatenating pieces of recorded audio speech that was stored in a database memory. Text-to-speech synthesis was an important object in speech interface, which allows low-bandwidth text to supply a user with easy to gain the information. Speech synthesis system has wide range of applications in our daily life. Speech synthesis plays an important role in signal processing. In this paper a Speech synthesized system is designed which is based on microcontroller, a project that performs several commands as defined for the synthesis of the speech signal. Index Terms: Microcontoller, Control System I. Introduction Embedded systems play an important role in our day to day life. Due to limited works this systems can be highly optimized by particular needs [27]. Some of the embedded systems applications were controlling the devices with some amount of intelligence built-in [1]. Speech synthesis system has wide range of applications in our daily life. Speech synthesis plays an important role in signal processing. The natural, intelligible synthetic voices in the expanding field of speech technology had a lot of demand [4]. Speech was the natural form of human communication. Automatic speech signal generation on computers is commonly called speech synthesis. Speech synthesis was the artificial production of human speech [2]. Speech synthesizer can be created by concatenating pieces of recorded audio speech that was stored in a database memory. Text-to-speech synthesis was an important object in speech interface, which allows low-bandwidth text to supply a user with easy to gain the information. In speech synthesis field, a number of research programmes have taken place during the last decade [13]. In 1984 an industry expert Mr Drinkwater said that about the synthesis “synthesis is going to be the next bam-burning technology” [2]. Now a day‟s synthesis speech was useful for many applications. In 1999 an American business magnate said that “In a few years personal computers will talk back to us” [2].In the past (Allen et al., 1987), the text to speech synthesis was done from a complete knowledge of the acoustics of the language being targeted. For getting text to speech synthesis different techniques were used [5]. Concatenate method was the simplest method for achieving the synthesis speech. By using this method the speech was generated by linking pre recorded speech segments to build syllables, phrases, or words [2]. the applicable criteria that follow. The block diagram for the system was shown in the following figure 1. INPUT OUTPUT Figure 1: Block diagram of the system From the above figure the system will have four main sections; the microcontroller of the system would take the commands from user and control the converter. The link between the user interface (computer) and the microcontroller was two ways because after taking the some commands from user the system would also respond with proper feedback. After processing the inputs the output will reflect through the speaker controlled by the microcontroller which was illustrated by the figure 1.
  • 2. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 78 | Page II. Problem Analysis This Paper was highlighted the requirements for developing the system. These requirements were categorized into two major sections those were hardware requirements and software requirements. Based on the project deliverables the requirements of the system were given below and clearly explained. The block box testing‟s were also mentioned in the ending of this Paper. 2.1) Hardware requirements of the system This project “synthesized speech using a small microcontroller” was a type embedded system so it would be a combination of electronic hardware and software code. The following figure was show the system would have electronic hardware capable of control the ISD board as shown in Fig. 2.1 (a).The following figure was show the system would have electronic hardware capable of speech synthesizer which can controlled by the microcontroller. The block diagram for synthesized speech using microcontroller was shown in the Figure 2.1(b) Fig. 2.1 (a) Block diagram for audio record/playback system Fig. 2.1 (b) Text to speech synthesis system block diagram For this system a synthesizer was used for converting the text to speech. 2.1.1) Speech synthesizer A synthesizer was required for getting the speech from text; it should be controlled by microcontroller. The selection of the synthesizer would also be a challenging task there were many kind of synthesizers available in the market also there were some techniques for hardware as well as software techniques were there for getting speech from text. Software approach was good but in implementation quiet complex so hardware approach was followed in this project. The synthesizer was required for this project which would capable to read the English words, numbers and mathematical functions. The synthesizer chip should be light weight and communicate with the host processor. It should have built in rule database for converting the text to speech. 2.1.2) Microcontroller Microcontroller was the heart of this project or one can say it was the master as it controls the other devices so, the selection of the microcontroller would also be a challenging task. Microcontrollers were available from Microchip‟s PIC with a large range of size and features. In this project the priority was given to microcontroller which one can control the operation of the ISD board and synthesizing the speech within that microcontroller. If the microcontroller was smaller and light weight it would be better to keep the PCB small. Microcontroller with internal features like internal oscillator, interrupts would be better. Microcontroller should have the I/O pins for connecting to external modules. Before making any selection the size and power efficiency could be considered. 2.1.3) Speaker It was a device which converts electrical signal in to sound. Speaker was commonly used to hear sounds. In this project the speaker should be used for hearing the synthesized speech in the output. A speaker was required for this project which would be small, light weight and it should have low impedance, low power ratings.
  • 3. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 79 | Page 2.2) Speech synthesis methods: Speech synthesis methods were used for converting the speech from the text. A synthesis method which was used in this project, it should be simple and easy to implement the system. 2.3) Serial communication to establish a user interface A module would be required to establish a user interface to control the system. The system should be user control so it would be serially communicated with the host PC. RS232 and MM232r were used for the serial communication. 2.4) Software requirements of the system Software program was necessary for microcontroller to control the system. MPLAB was the program development tool which can enable the user to write the program for the PIC microcontroller. MPLAB provided the C18 compiler for writing the program. The program was supposed to write for control the ISD board as well as to generate the synthesized speech. 2.4.1) Programming language, development and debugging tool There were two popular languages (Assembly and embedded C) available for eight bit microcontrollers, any language can be used for this system. Embedded C was quite similar to the normal C language but only the difference was in embedded C programming the initializations and some special functions were used, on the other hand the assembly instruction set was different to the manufacture of the microcontroller from one to another. It was necessary before writing the program to know about the register set (SFR) of that microcontroller. A program could be written in their own way. The software tool MPLAB along the C18 compiler was providing the assembler and compiler. Microcontroller did not understand assembly of the C program which was stored in the ROM of the PIC as a .hex file. The assembler and compiler can converts the instructions to the machine code. The MPLAB was also available the debugging features. By using the MBLAB ICD2 programmer and debugger kit the converted hexadecimal file can be loaded into the microcontroller. 2.5) Black box testing The output of the system was depending on the corresponding input. For this system the input would be user commands and the output would be speaker. The black box testing should be done for the system. I. The system should be user interface with the system. II. The ISD board should be controlled by the microcontroller. III. The speech should be synthesized within a microcontroller. Fig. 2.5 Black box testing of the system III. Proposed Problem Solutions Based on the hardware and software requirements as mentioned before Paper, the solutions were given for both hardware and software. According to the deliverables mentioned in the previous Paper, the first task was to control audio record/playback ISD board and second one was generate the synthesized speech within the microcontroller. The hardware and software solution were given below. 3.1) PIC18F1320 Microcontroller PIC18F1320 was an 8 bit microcontroller developed by Microchip Company and it was power efficient because used Nano watt technology. This microcontroller was used in this project because it has external oscillator and many other features [30]. PIC18F microcontroller support enhanced USART module and supports RS485, RS232 and LIN1.2 which has auto baud rate and also auto wake up start bit. The microcontroller PIC188F1320 has four crystal modules (LP, XT, HS and HSPLL) and high computational performance. It could be works at high speed up to 16-40MHZ. So PIC18F1320 (Fig 3.1) was used rather than PIC16F series (it works at low speed). For more information about the PIC the references 30 and 27 were studied. It had two ports A and B and port B pins were used in this project. In this project the microcontroller was used to control the audio record/playback device and generating the synthesized speech.
  • 4. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 80 | Page Figure 3.1: PIC18F1320 3.2) Synthesis method Speech synthesis technologies can be broadly categorized into three categories. They were given below [29]:  Concatenate synthesis  Formant synthesis  Articulator synthesis 3.2.1) concatenate speech synthesis: In this technology the speech can be generated by linking pre-recorded speech segments to make syllables, phrases, or dip hones. Waveform segments were store in a database. It was the simplest method for speech synthesis [14]. 3.2.2) Formant synthesis: Formants were defined as “resonant frequencies that occur at the main resonant areas of the vocal tract for a given sound”. It consists of artificial reconstruction of the formant characteristics to be produced. 3.2.3) Articulator synthesis: Articulators were speech organs. It was based on computational models of the human vocal tract and it was a most natural sounding speech synthesis. But it was complex [4]. Finally concatenate method was identified, which was used for converting the speech from text [2]. In this project for synthesizing the speech a synthesizer was used. This synthesizer chip follows the concatenate method because it was a simplest method compare to other method. 3.3) TTS256 Synthesizer chip Figure 3.3: TTS256 speech synthesizer chip It was very hard to find out the synthesizer for this task because most of the synthesizers (WTS701, SP0256 chips) were not available in the market. Finally after a long searching TTS256 chip was available in speechchips.com. TTS256 was an 8 bit microprocessor and it was programmed with built in 600 letter-to-sound rule database. TTS256 speech chip was a 28 pin DIP and compatible with PIC, BASIC stamp, OOPic [33]. It can generate the speech code from ASCII text and it could be read English words, numbers and mathematical functions. For serial communication parameters it was set the baud rate 9600. The datasheet mentioned in the ref.[33] was studied. The host microcontroller can send up to 128 characters to the TTS256 chip via serial port on 18th pin (RX). The text which was entered by user was spoken when the TTS receives a new line character or the maximum number of characters (128) has been entered.
  • 5. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 81 | Page 3.4) Speaker Figure 3.4: Speaker The speaker selected in this project, which was having the following feature. 1) It was small and light weight 2) It had low power rating (0.5W) and Impedance (8 ohm) 3.5) MM232R module It was a single chip asynchronous serial data transfer interface. In MM232R module FT232RQ chip device was used and it can convert USB to RS232 / RS422 / RS485 converters. RS232 and MM232 were the two solutions for serial communication. MM232R was support for 7 or 8 data bits and one or two stop bits. The data transfer rates from range of 300 baud to 3 mega baud (RS422 / RS485) and 300 baud to 1 mega baud (RS232). It can support for bus powered, high bus powered and self powered USB configurations. RS232 was also used for serial data transfer [33]. If RS232 was used the MAX232 IC should be used for voltage levels, then the cost of the PCB will increase so MM232 was used instead of RS232. The signal voltage levels of the MM232 were in the range from 3.3v to 5v. Figure 3.3: MM232R 3.6) Program language Embedded C was used for the programming the PIC. MPLAB was used for writing the program of the control system as well as speech synthesis system. MPLAB IDE was windows based Integrated Development Environment for programming the microcontrollers. The MPLAB provided the ability to the users to create and edit the source code in the editor window. In MPLAB IDE software tool the C18compiler was available to assemble, compile and link the source code [31]. MPLAB ICD 2 debugger and programmer kit was used for load the hexadecimal code to the PIC18F 1320 microcontroller. Figure 3.6: MPLAB program structure
  • 6. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 82 | Page Flow chart for control system The operation of the audio record and playback system was explained in the form of flow chart. The flow chart for control system was shown in the following figure. Figure 3.7: Flow chart for control PCB
  • 7. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 83 | Page Figure 3.7: Flow chart for Speech Synthesis System IV. Problem Implementation 4.1 Hardware implementation of the control system The solutions for getting the hardware and software code of the system were explained in the previous Paper were finally implemented and explained in this Paper, for some of the hardware problems like stabilization more than one solution were given. The outcomes of some solutions were necessary to discuss at this at this as it was necessary to stick to one approach from here onward. According to the deliverables mentioned in the first Paper the primary aim was control the ISD board and second one was to generate synthesis speech system. The description of the hardware and software for these two tasks were separately given below. The software programs were developed and listed in the appendix for both audio record/playback system and synthesized speech system. Before implementing the control system to control the ISD board, the basic operation of the ISD 1420P device was explained below. Figure 4.1: ISD1420P PCB
  • 8. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 84 | Page 4.1.1) Working of the audio record/playback device (ISD1420p board): The system was configured with microphone, on-chip oscillator, speaker amplifier, smoothing filter, and two push buttons. The PCB circuit device which was given by the supervisor was shown in the following figure 4.1.1(b). The ISD1420P device was used for audio record/playback system. The following operating sequence was demonstrated the working of chip device. For recording the audio data: The device which was having two push buttons, one for recoding the data and another was for play backing. When push button (REC) was pressed, then the REC signal LOW initiated a record cycle from the starting of the data space [28]. When recording process was beginning the red LED was glowing until the data space had been filled. Once the data space was filled then the recording was stopped. The power of the device was automatically down then the REC signal was pulled to high. For playback the audio data: When the push button (PLAY) was pressed then active low PLAY signal initiated a playback cycle from the starting of the data space which was recorded earlier. If the device was reached to EOM marker then automatically power was down. While the process was going on the green LED was glowing until and unless the EOM marker reached. [28] 4.1.2) Design of PCB for controlling of the Audio record/playback system Without using of the PLAY and RECORD buttons the ISD1420 device should be operated. For this a control PCB was designed. The schematic diagram for controlling the audio record and playback device system was designed in Proteus ISIS. The Proteus ISIS professional provided the libraries for constructing the schematic designs. From the libraries picked the PIC18F1320, resistors (10KΩ and 330Ω), LED and capacitor then constructed the circuit for control PCB for controlling the ISD kit. Figure: 4.2 Schematic diagram of Control PCB The components PIC18F1320, resistors, LED and capacitor were available in the Proteus ISIS library but the component MM232R was not available in the list of Proteus ISIS so four pin connectors were used. The 10K resistor used for avoiding the PIC reset and LED was used for power indication. Some additional elements were also selected like 2 pin Molex connectors for power supply as well as for ISD board. The following PCB was the implementation of the schematic discussed in the previous Paper. The PCB was designed for controlling the audio record/playback device in Proteus Ares software. The PCB layout of the audio record/playback system was shown in the following figure 4.1.2(b). RA0/AN0 1 RA1/AN1/LVDIN 2 RA4/T0CKI 3 RA5/MCLR/VPP 4 RA2/AN2/VREF- 6 RA3/AN3/VREF+ 7 RB0/AN4/INT0 8 RB1/AN5/TX/CK/INT1 9 RB4/AN6/RX/DT/KBI0 10 RB5/PGM/KBI1 11 RB6/PGC/T1OSO/T1CKI/P1C/KBI2 12 RB7/PGD/T1OSI/P1D/KBI3 13 RA6/OSC2/CLKO 15 RA7/OSC1/CLKI 16 RB2/P1B/INT2 17 RB3/CCP1/P1A 18 U1 PIC18F1320 PACKAGE=DIL18 VDD=GND VSS=+5V C2 10u GND +5V GND Speech Synthesizer, Venu Gopal, MSc Embedded Systems R1 10k 1 2 3 4 8 7 6 5 J1 CONN-DIL8 1 2 3 4 J2 CONN-SIL4 1 2 3 4 J3 CONN-SIL4 +5V 1 2 5V SUPPLY TO BOARD CONN-H2 1 2 ISD BOARD SWITCH CONNECTOR CONN-H2 D1 DIODE-LED R2 330R
  • 9. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 85 | Page Figure 4.1.2(b): PCB layout of the control pcb system. Operation of the control system For controlling of the ISD1420 record/playback device the control PCB was designed. MM232R was used for sending data and receiving data [32]. The 5th pin of the MM232R was connected to 10th pin (RB4) of the PIC18F1320 for receiving the data and 7th pin of the MM232R was connected to the 9th pin (RB1) of the PIC for transmitting the data from the PIC. MM232R pins 2 and 3 were connected to 5v power supply. Based on the command received from the user via MM232R the PIC will control the operation of record and playback device. The RB2 pin of the PIC was connected to the 27th pin (active low REC) ISD board and RB3 pin of the PIC was connected to the 23rd pin (active low PLAY) ISD board. Whenever the user enter the letter „r‟ on the HyperTerminal the MM232R send this signal to the RX of the PIC. Then the PIC send an active low signal to the ISD board, this active low signal (REC) was enable the pin then the will ready for recording. At the other hand the PIC was send the ISD device status (START RECORDING) to the MM232R, this message will display on the HyperTerminal window. Same way if the user entered the letter „p‟, the PIC sends an active low signal to the ISD board then the device start to play. If any letter (except „r‟ or „p‟) the PIC was send the status as „wrong command‟ to the MM232R. 4.1.4) Testing of the control system The following hardware tests were done for the PCB. 1) For checking short circuit connectivity of the PCB initially the printed circuit board was tested without placing components. 2) After placing the components these individual pins were tested through digital CRO. 3) The power of the components was tested by using the multi meter. HyperTerminal was used for creating a user interface to control the audio record/playback device. MM232R communication protocol was used for serial communication from host to the microcontroller for sending and receiving the data. For creating a serial communication, the required HyperTerminal settings were mentioned in the following figure 4.1.4(a).
  • 10. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 86 | Page Figure 4.1.4(a): HyperTerminal window settings Software implementation for control PCB The PIC18F1320 microcontroller was control the operation of the audio record/playback. For this the PIC should be programmed, so the PIC C program was written by using MPLAB. To write the PIC C program in the MPLAB first selected the new project wizard option, and then PIC18F1320 was selected for microcontroller and C18 compiler was added. The complete PIC C program of control PCB was listed in Appendix. The PIC C program was written in a file save with extension of .C and then this file was added to the source file [27]. For clear explanation the whole program was divided as small sections. As the software code was going to be in embedded C so the initial sections of every C program have a header file followed by some declarations, definitions then the main and other functions. The MPLAB libraries provided all the header files [31]. Figure 4.2(a): Error free code for control system Hence the program for control system was explained clearly and this PIC C program was compiled by using C18 compiler. The build succeeded program for the control system was shown in the following figure 4.2(a).
  • 11. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 87 | Page Figure 4.2(b): PIC programmed by using ICD2 kit With the help of ICD2 kit the PIC was loaded to target device. The PIC was programmed on the ICD2 kit compatibility board. This process was shown in the following figure 4.2(b). Using the MPLAB ICD2 debugger and programmer kit the program successfully loaded into the PIC18F1320. After completion of this loading process in MPLAB window the status of the ICD 2 kit was displayed and it was shown in the figure 4.2(c). Figure 4.2(c): Successful loaded of the program After loading the program into the PIC then the microcontroller placed on the control PCB. The complete hardware connections for control the audio record/playback system was shown in the following figure 4.2(d). 4.2 Hardware implementation for synthesis speech system The main aim of the project was to generate the synthesis speech within a small microcontroller. The schematic diagram for synthesis speech system was designed in Proteus ISIS. The Proteus ISIS professional provided the libraries for constructing the schematic designs. The main PCB was could handle the both systems those were it could control the ISD board and it could be generate the synthesis speech within a microcontroller. The main PCB would reduce the cost of the cost and PCB area. The schematic diagram of the speech synthesis system was shown in the following figure 4.3(a). To develop the system the required components were selected by using Proteus ISIS professional software. The component PIC18F1320, resistors, LED and capacitor were available in the Proteus ISIS library
  • 12. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 88 | Page but the components MM232R and TTS256 synthesizer chip were not available in the list of Proteus ISIS so collection of four pin serial in line connectors were used for MM232R and four serial in line connectors were used for construct the TT256 chip. The 10K resistor used for avoiding the PIC reset and LED was used for power indication. Some additional elements were also selected like 2 pin Molex connectors for power supply as well as for ISD board. Figure 4.2(d): Complete hardware connections of the audio record/playback system Figure 4.3(a): Schematic design for synthesis speech system Pin allocation for microcontroller Detail Pin No. Indicator Pin B5 MM232R Rx, TTS256 Rx Pin B4 (Rx) MM232R Tx Pin B1 (Tx) Recording Pin B2 Playing Pin B3 RA0/AN0 1 RA1/AN1/LVDIN 2 RA4/T0CKI 3 RA5/MCLR/VPP 4 RA2/AN2/VREF- 6 RA3/AN3/VREF+ 7 RB0/AN4/INT0 8 RB1/AN5/TX/CK/INT1 9 RB4/AN6/RX/DT/KBI0 10 RB5/PGM/KBI1 11 RB6/PGC/T1OSO/T1CKI/P1C/KBI2 12 RB7/PGD/T1OSI/P1D/KBI3 13 RA6/OSC2/CLKO 15 RA7/OSC1/CLKI 16 RB2/P1B/INT2 17 RB3/CCP1/P1A 18 U1 PIC18F1320 PACKAGE=DIL18 VDD=+5V VSS=GND C2 10u GND +5V GND Speech Synthesizer, Venu Gopal, MSc Embedded Systems R1 10k 1 2 3 4 8 7 6 5 MM232 B CONN-DIL8 1 2 3 4 MM232 A CONN-SIL4 1 2 3 4 MM232 C CONN-SIL4 +5V 1 2 5V SUPPLYTO BOARD CONN-H2 1 2 ISDBOARDSWITCHCONNECTOR CONN-H2 D1 DIODE-LED R2 330R 1 2 SPKCONNECTION CONN-SIL2 1 2 3 4 5 6 7 TTS B CONN-SIL71 2 3 4 5 6 7 TTS A CONN-SIL7 1 2 3 4 5 6 7 TTS C CONN-SIL7 1 2 3 4 5 6 7 TTS D CONN-SIL7 D2 DIODE-LED R3 330R
  • 13. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 89 | Page Software implementation for Speech synthesis system The PIC18F1320 microcontroller was used for generating the synthesis speech system. For this the PIC should be programmed, so the PIC C program was written by using MPLAB. To write the PIC C program in the MPLAB, first selected the new project wizard option, and then PIC18F1320 was selected for microcontroller and C18 compiler was added. The complete PIC C program of the speech synthesis system was listed in Appendix B. The PIC C program was written in a file, save it with extension of .C and then this file was added to the source file. For clear explanation the whole program was divided as small sections. As the software code was going to be in embedded C so the initial sections of every C program have a header file followed by some declarations, definitions then the main and other functions. The MPLAB libraries provided all the header files. Figure 4.4(a): Error free program The above program was loaded in to the PIC18F 1320 microcontroller with the help of MPLAB ICD2 debugger and programmer kit provided by the university. The following figure 33 was show the message after loading the program in to the PIC.
  • 14. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 90 | Page Figure 4.4(b): Successful loading of the program Working of the speech synthesis system HyperTerminal was used for creating a user interface to control the ISD board and synthesis speech system. MM232R communication protocol was used for serial communication from host to the microcontroller for sending and receiving the data. For creating a serial communication, the required HyperTerminal settings were mentioned in the following figure. After loading the program into the PIC then the microcontroller placed on the synthesis speech system PCB. The complete hardware connections for the speech synthesis system were shown in the following figure 4.5(b). MM232R serial communication module was used for user interface. The user could write the text on hyper terminal window, the corresponding speech will be produced by the speaker. Two LEDs were used in this system one was for power indication (+5V) and second one indicates whether the program was loaded in to the PIC microcontroller or not. The resistor 10K ohm was used for avoid the reset of the PIC. Figure 4.5(a): Hyper terminal settings for serial communication
  • 15. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 91 | Page The final PCB would do the both tasks. The data which was entered on the hyper terminal will send to the 10th pin (RX) of the PIC18F1320 microcontroller then the received data will sent to the 18th pin (RX) of the TTS256 synthesizer chip. The TTS256 synthesizer chip can convert the data which was sent by microcontroller to speech. The 24th pin of the TTS256 chip was connected to the speaker, 14th pin was connected to ground and 28th pin was connected to 5V power supply. The text which was entered by user was spoken when the TTS256 chip receives a new line character or the maximum number of characters (128) has been entered. By using this PCB the ISD board was also controlled. If the user entered the letter „r‟ on the hyper terminal, the microcontroller will send the active low signal to the ISD board then the ISD board start recording the voice. Simultaneously when the user entered the letter „p‟ on the hyper terminal, the microcontroller will send the active low signal to the ISD board then device could start playing the sound which was previously recorded in the ISD chip. V. Results In this Paper the system outcomes were explained clearly. Initially the control PCB was designed for control the audio record/playback device (ISD board). The main task of the project synthesis speech systems results were given and explained in the last section. 5.1) Results of the control system Based on the command which was entered by the user on the hyper terminal the microcontroller controls the ISD board. When the user entered command „r‟ on hyper terminal then the microcontroller send the active low signal to the ISD board the system starts recording. The microcontroller sends the status of the ISD board to the 7th pin (RX) of the MM232R and it was displayed on the hyper terminal. The red LED of the ISD board blink until the system recording the voice. The following figures (5.1(a) and 5.1(b)) were show the process of the system. Figure 5.1(a): Computer received the data from the PIC when the system starts recording Figure 5.1(b): ISD board when recording the voice Same way when the user entered the command „p‟ on hyper terminal then the microcontroller send the active low signal to the ISD board the system starts recording. The microcontroller sends the status of the ISD board to the 7th pin (RX) of the MM232R and it was displayed on the hyper terminal. The green LED of the ISD board blink until the system would play the voice which was stored previously in the ISD1420P. The following figures (5.1(c) and 5.1(d)) were show the process of the system. When the user entered wrong command then the PIC would send message to the computer and it was displayed on the hyper terminal as “wrong command”. „r‟ entered by the user Status of the ISD board
  • 16. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 92 | Page Figure 5.1(c): Computer received the data from the PIC when the system start play Outcome for the speech synthesis system The final PCB which was the main PCB, designed for the controlling of the ISD board and synthesis speech system. The following figure which was shows the two LED indicators, one was for power indication and another one for checking the PIC programmed or not. Figure 5.1(d): ISD board when playing the voice Figure 5.2: The speech synthesis system In the above figure show the LED power indicator was blinking so the 5V power was coming to circuit. Second one was also blinking so the program was also loaded into the PIC. According to the software program the LED blinks up to 55msec but the LED was continuously blinking. Also when the user trying to enter any character on the hyper terminal window and it was not displayed and also the output speech was not coming from the speaker. The system could not work because the software might not match with the hardware connections of the system or might be problem in the software code. Many changes were done in case of modifying the software code as well as in hardware some connections was changed. VI. Discussion On Results Well the results of the complete system were as expected, but in sometimes it was not picking up the commands entered by the user through the hyper terminal. There was some problem with the playback of the sound possible causes may be compatibility issues of TTS IC. The Embedded C program was logical right and „p‟ entered by the user Status of the ISD board
  • 17. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 93 | Page there were no compiling errors. There was no RJ45 connector on the target board so it was not possible to get the live debug simulation on the MPLAB. The sound of the speaker was not that good may be because of the quality of the device, but on the other hand it was also not possible to attach a bid speaker directly to the TTS IC as it cannot supply the required source current. Behavior of complete system was good system was capable of doing task to some extent for what it was designed to. With this technical discussion for this project comes to an end. VII. Conclusion After going through this task it was evident that to develop complete system even if its looks simple at first look lots of effort and dedication is required. Various fields required to be combined like good software knowledge, electronics and analytical skills. Sometimes a simple appearing task may take a long time to work accordingly, as happened while tried to achieve a proper communication between microcontroller and the TTS IC. Every element is required to be looked after. Analyzing why something fails or did not worked as required helps a lot and makes the vision clear brings success closer and enhances knowledge. The work on the project was quite satisfactory, but still there were some shortcomings like sometimes system was not working at all, so some changes are suggested in software and electronic design. Another major addition that could be done in future design is the addition of on board debugging mechanism to attach the ICD kit to the main board this could be easily done by adding a RJ connector. Various features like audio filter circuit which would only allow the audible frequencies to pass to the speaker could be added along with some amplifier and volume control. In the embedded program additional commands could be added like for volume control, the duration for which TTS is required to be in play or record mode. Some features like to monitor the errors on the system could be added and if any error occurs then could be displayed on the hyper terminal, and even intimate the user by producing speech from the TTS. There is no limit as such to the additions that could be done in the future to this system as the developed project is just the basic one so many enhancements are possible. References [1]. Juang, B.H. Tsuhan Chen, The past, present, and future of speech processing Volume 15, Issue 3, ISSN: 1053-5888, Pages 24-48. [2]. Henton, Caroline, Challenges and Rewards in Using Parametric or Concatenative Speech synthesis, ISSN: 1381-2416. [3]. Kim, Jongkuk - Hahn, Hernsoo – Yoon, Uei-Joong-Bae, Myungjin, On a pitch alteration for speech synthesis systems, year 2009, volume 50, issue 4, pages 435-446, ISSN:0929-6212. [4]. Rahier, M.C. , Defraeye, P.J. Guebels, P.-P, Patovan, B, A 3 Âľm NMOS high performance LPC speech synthesizer chip, Volume 18, Issue 3, Pages 349-359, ISSN:0018-9200. [5]. Henton, Caroline, Challenges and rewards in using parametric or concatenative speech synthesis, Year [6]. 2002, Volume 5, Issue 2, Pages 117-131, ISSN: 1381-2416. [7]. Holfelder, Wieland, Interactive remote recording and playback of multicast videoconferences, Year 1998, Volume 21, Issue 15, Pages 1285-1294, ISSN: 0140-3664. [8]. Kim, Jong Kuk-Hahn, Hern Soo-Bae, Myung Jin, On a speech multiple system implementation foe synthesis, year 2009, volume 49, issue 4, ISSN: 0929-6212. [9]. Hilt, V. Mauve, M. Vogel, J. Effelsberg, W, Recording and playing back interactive media streams, Date Oct 2005, volume 7 issue 5,pages 960-971, ISSN: 1520-9210 [10]. A. K. Rath and P.K. Meher, Design of a merged DSP microcontroller for embedded systems using discrete orthogonal transform, Year 2006, pages 388-394, ISSN: 1549-3636. [11]. Schroeder, E.F. platte, H. –J. Krahe, D., ‘MSC’ :stereo audio coding with CD-quality and 256 kBit/sec, Date: Nov.1987, Volume CE-33, Issue 4, Pages 512-519, ISSN: 0098-3063. [12]. Geoff Jackson, Peter Holzmann, A Single-chip Text-to-Speech Synthesis Device Utilizing Analog Nonvolatile Multilevel Flash Storage, Issue Date: NOV.2002, ISSN: 0018-9200. [13]. Liu, Fu-Hua-Gu, Liang-Gao, Machael, Applications of Language Modelling in Speech-to-Speech Translation table of content, Year- 2004, ISSN: 1381-2416. [14]. Lee, C, -H; Jung, S.-K. ; Kang, H. –G., Applying a speaker-Dependent speech compression Technique to concatenative TTS synthesizers, Year- 2007, ISSN: 1558-7916. [15]. Henton, Caroline, Challenges and Rewards in using Parametric or concatenative speech synthesis, Year- 2002, Issn: 1381-2416. [16]. Goldsmith, John, Dealing with Prosody in a tex-to-speech system table of content, Year- 1999, ISSN: 1381-2416. [17]. Sotiris Karabetsos, Pirros Tsiakoulis, Embedded unit selection Text-to-speech synthesis, Year- 2009, ISSN: 0098-3063. [18]. Kim, Jong kuk-Hahn, Hern Soo- Bae, Myung Jin, On a speech Multiple system Implementation for speech synthesis table of content, Year- 2009, ISSN: 0929-6212. [19]. Taylor & Francis, Technical notes on Audio Recording, Year- 2000, ISSN: 0835-1813. [20]. El-Imam, Yousif A. –Don, Zuraida Mohammed, Text-to-Speech conversion of standard Malay, Year- 1999, ISSN: 1381-2416. [21]. Steve Heath, Embedded System Design, Second Edition, ISBN 0-7506-5546-1. [22]. Holzmann, P. Jackson, G.B. Raina, A. Hung-chuan pai Ming-Bing Chang Awsare, S.V. Engh, L.D. Kao,O.C. Palmer, C.R. Chun- Mai Liu Kordesch, A.V. Ken su Hemming,M. An analog and digital record, playback and processing system using a coarse-fine programming method, Date 18-20 April 2001, pages 192-195, ISBN: 0-7803-6412-0. [23]. Embedded C by Michale J Pont, Pearson International, ISBN-0173290532, year of publication 2002 [24]. Rice, S.V. Bailey, S.M, An integrated environment for audio recording,editing, and retrieval, Date Dec 2006, ISBN: 0-7695-2746- 9. [25]. Kumar, K. Raja Ramaiah, P. Seetha, DSP and Microcontroller based speech processor for auditory prosthesis, Date 20-23 Dec 2006, pages 518-522, ISBN: 1-4244-0716-8.
  • 18. Synthesized Speech using a small Microcontroller DOI: 10.9790/2834-10637794 www.iosrjournals.org 94 | Page [26]. Thiang, Implementation of speech recognition on MCS51 microcontroller for controlling wheelchair, Date : 25-28 Nov 2007, pages 1193-1198, ISBN: 978-1-4244-1355-3. [27]. Sarathy, K.P. Ramakrishnan, A.G, A research bed for unit selection based text to speech synthesis, Date 15-19 Dec 2008, pages 229-232, ISBN: 978-1-4244-3471-8. [28]. PIC Microcontroller and Embedded Systems by Mazidi, Muhammad Ali, Pearson International, ISBN- 0136009026, Year of publication 2007.