We describe in detail a Grapheme-to-Phoneme (G2P) converter required for the development of a good quality
Marathi Text-to-Speech (TTS) system. The Festival and Festvox framework is chosen for developing the
Marathi TTS system. Since Festival does not provide complete language processing support specie to various
languages, it needs to be augmented to facilitate the development of TTS systems in certain new languages.
Because of this, a generic G2P converter has been developed. In the customized Marathi G2P converter, we
have handled schwa deletion and compound word extraction. In the experiments carried out to test the Marathi
G2P on a text segment of 2485 words, 91.47% word phonetisation accuracy is obtained. This Marathi G2P has
been used for phonetising large text corpora which in turn is used in designing an inventory of phonetically rich
sentences. The sentences ensured a good coverage of the phonetically valid di-phones using only 1.3% of the
complete text corpora.
Punjabi to Hindi Transliteration System for Proper Nouns Using Hybrid ApproachIJERA Editor
The language is an effective medium for the communication that conveys the ideas and expression of the human
mind. There are more than 5000 languages in the world for the communication. To know all these languages is
not a solution for problems due to the language barrier in communication. In this multilingual world with the
huge amount of information exchanged between various regions and in different languages in digitized format,
it has become necessary to find an automated process to convert from one language to another. Natural
Language Processing (NLP) is one of the hot areas of research that explores how computers can be utilizing to
understand and manipulate natural language text or speech. In the Proposed system a Hybrid approach to
transliterate the proper nouns from Punjabi to Hindi is developed. Hybrid approach in the proposed system is a
combination of Direct Mapping, Rule based approach and Statistical Machine Translation approach (SMT).
Proposed system is tested on various proper nouns from different domains and accuracy of the proposed system
is very good.
MORPHOLOGICAL ANALYZER USING THE BILSTM MODEL ONLY FOR JAPANESE HIRAGANA SENT...kevig
This study proposes a method to develop neural models of the morphological analyzer for Japanese Hiragana sentences using the Bi-LSTM CRF model. Morphological analysis is a technique that divides text data into words and assigns information such as parts of speech. In Japanese natural language processing systems, this technique plays an essential role in downstream applications because the Japanese language does not have word delimiters between words. Hiragana is a type of Japanese phonogramic characters, which is used for texts for children or people who cannot read Chinese characters. Morphological analysis of Hiragana sentences is more difficult than that of ordinary Japanese sentences because there is less information for dividing. For morphological analysis of Hiragana sentences, we demonstrated the effectiveness of fine-tuning using a model based on ordinary Japanese text and examined the influence of training data on texts of various genres.
A ROBUST THREE-STAGE HYBRID FRAMEWORK FOR ENGLISH TO BANGLA TRANSLITERATIONkevig
Phonetic typing using the English alphabet has become widely popular nowadays for social media and chat services. As a result, a text containing various English and Bangla words and phrases has become increasingly common. Existing transliteration tools display poor performance for such texts. This paper proposes a robust Three-stage Hybrid Transliteration (THT) framework that can transliterate both English words and phonetic typed Bangla words satisfactorily. This is achieved by adopting a hybrid approach of dictionary-based and rule-based techniques. Experimental results confirm superiority of THT as it significantly outperforms the benchmark transliteration tool.
INTEGRATION OF PHONOTACTIC FEATURES FOR LANGUAGE IDENTIFICATION ON CODE-SWITC...kevig
In this paper, phoneme sequences are used as language information to perform code-switched language
identification (LID). With the one-pass recognition system, the spoken sounds are converted into
phonetically arranged sequences of sounds. The acoustic models are robust enough to handle multiple
languages when emulating multiple hidden Markov models (HMMs). To determine the phoneme similarity
among our target languages, we reported two methods of phoneme mapping. Statistical phoneme-based
bigram language models (LM) are integrated into speech decoding to eliminate possible phone
mismatches. The supervised support vector machine (SVM) is used to learn to recognize the phonetic
information of mixed-language speech based on recognized phone sequences. As the back-end decision is
taken by an SVM, the likelihood scores of segments with monolingual phone occurrence are used to
classify language identity. The speech corpus was tested on Sepedi and English languages that are often
mixed. Our system is evaluated by measuring both the ASR performance and the LID performance
separately. The systems have obtained a promising ASR accuracy with data-driven phone merging
approach modelled using 16 Gaussian mixtures per state. In code-switched speech and monolingual
speech segments respectively, the proposed systems achieved an acceptable ASR and LID accuracy.
Quality estimation of machine translation outputs through stemmingijcsa
Machine Translation is the challenging problem for Indian languages. Every day we can see some machine
translators being developed , but getting a high quality automatic translation is still a very distant dream .
The correct translated sentence for Hindi language is rarely found. In this paper, we are emphasizing on
English-Hindi language pair, so in order to preserve the correct MT output we present a ranking system,
which employs some machine learning techniques and morphological features. In ranking no human
intervention is required. We have also validated our results by comparing it with human ranking.
EXTRACTING LINGUISTIC SPEECH PATTERNS OF JAPANESE FICTIONAL CHARACTERS USING ...kevig
This study extracted and analyzed the linguistic speech patterns that characterize Japanese anime or game characters. Conventional morphological analyzers, such as MeCab, segment words with high performance, but they are unable to segment broken expressions or utterance endings that are not listed in the dictionary, which often appears in lines of anime or game characters. To overcome this challenge, we propose segmenting lines of Japanese anime or game characters using subword units that were proposed mainly for deep learning, and extracting frequently occurring strings to obtain expressions that characterize their utterances. We analyzed the subword units weighted by TF/IDF according to gender, age, and each anime character and show that they are linguistic speech patterns that are specific for each feature. Additionally, a classification experiment shows that the model with subword units outperformed that with the conventional method.
Punjabi to Hindi Transliteration System for Proper Nouns Using Hybrid ApproachIJERA Editor
The language is an effective medium for the communication that conveys the ideas and expression of the human
mind. There are more than 5000 languages in the world for the communication. To know all these languages is
not a solution for problems due to the language barrier in communication. In this multilingual world with the
huge amount of information exchanged between various regions and in different languages in digitized format,
it has become necessary to find an automated process to convert from one language to another. Natural
Language Processing (NLP) is one of the hot areas of research that explores how computers can be utilizing to
understand and manipulate natural language text or speech. In the Proposed system a Hybrid approach to
transliterate the proper nouns from Punjabi to Hindi is developed. Hybrid approach in the proposed system is a
combination of Direct Mapping, Rule based approach and Statistical Machine Translation approach (SMT).
Proposed system is tested on various proper nouns from different domains and accuracy of the proposed system
is very good.
MORPHOLOGICAL ANALYZER USING THE BILSTM MODEL ONLY FOR JAPANESE HIRAGANA SENT...kevig
This study proposes a method to develop neural models of the morphological analyzer for Japanese Hiragana sentences using the Bi-LSTM CRF model. Morphological analysis is a technique that divides text data into words and assigns information such as parts of speech. In Japanese natural language processing systems, this technique plays an essential role in downstream applications because the Japanese language does not have word delimiters between words. Hiragana is a type of Japanese phonogramic characters, which is used for texts for children or people who cannot read Chinese characters. Morphological analysis of Hiragana sentences is more difficult than that of ordinary Japanese sentences because there is less information for dividing. For morphological analysis of Hiragana sentences, we demonstrated the effectiveness of fine-tuning using a model based on ordinary Japanese text and examined the influence of training data on texts of various genres.
A ROBUST THREE-STAGE HYBRID FRAMEWORK FOR ENGLISH TO BANGLA TRANSLITERATIONkevig
Phonetic typing using the English alphabet has become widely popular nowadays for social media and chat services. As a result, a text containing various English and Bangla words and phrases has become increasingly common. Existing transliteration tools display poor performance for such texts. This paper proposes a robust Three-stage Hybrid Transliteration (THT) framework that can transliterate both English words and phonetic typed Bangla words satisfactorily. This is achieved by adopting a hybrid approach of dictionary-based and rule-based techniques. Experimental results confirm superiority of THT as it significantly outperforms the benchmark transliteration tool.
INTEGRATION OF PHONOTACTIC FEATURES FOR LANGUAGE IDENTIFICATION ON CODE-SWITC...kevig
In this paper, phoneme sequences are used as language information to perform code-switched language
identification (LID). With the one-pass recognition system, the spoken sounds are converted into
phonetically arranged sequences of sounds. The acoustic models are robust enough to handle multiple
languages when emulating multiple hidden Markov models (HMMs). To determine the phoneme similarity
among our target languages, we reported two methods of phoneme mapping. Statistical phoneme-based
bigram language models (LM) are integrated into speech decoding to eliminate possible phone
mismatches. The supervised support vector machine (SVM) is used to learn to recognize the phonetic
information of mixed-language speech based on recognized phone sequences. As the back-end decision is
taken by an SVM, the likelihood scores of segments with monolingual phone occurrence are used to
classify language identity. The speech corpus was tested on Sepedi and English languages that are often
mixed. Our system is evaluated by measuring both the ASR performance and the LID performance
separately. The systems have obtained a promising ASR accuracy with data-driven phone merging
approach modelled using 16 Gaussian mixtures per state. In code-switched speech and monolingual
speech segments respectively, the proposed systems achieved an acceptable ASR and LID accuracy.
Quality estimation of machine translation outputs through stemmingijcsa
Machine Translation is the challenging problem for Indian languages. Every day we can see some machine
translators being developed , but getting a high quality automatic translation is still a very distant dream .
The correct translated sentence for Hindi language is rarely found. In this paper, we are emphasizing on
English-Hindi language pair, so in order to preserve the correct MT output we present a ranking system,
which employs some machine learning techniques and morphological features. In ranking no human
intervention is required. We have also validated our results by comparing it with human ranking.
EXTRACTING LINGUISTIC SPEECH PATTERNS OF JAPANESE FICTIONAL CHARACTERS USING ...kevig
This study extracted and analyzed the linguistic speech patterns that characterize Japanese anime or game characters. Conventional morphological analyzers, such as MeCab, segment words with high performance, but they are unable to segment broken expressions or utterance endings that are not listed in the dictionary, which often appears in lines of anime or game characters. To overcome this challenge, we propose segmenting lines of Japanese anime or game characters using subword units that were proposed mainly for deep learning, and extracting frequently occurring strings to obtain expressions that characterize their utterances. We analyzed the subword units weighted by TF/IDF according to gender, age, and each anime character and show that they are linguistic speech patterns that are specific for each feature. Additionally, a classification experiment shows that the model with subword units outperformed that with the conventional method.
Tamil-English Document Translation Using Statistical Machine Translation Appr...baskaran_md
The Paper presents a new method for translating a text document from Tamil to English. Our method is based on the Statistical Machine Translation Approach, combined with the Morphological Analysis, due to the fact that Tamil is a highly-inflected language. This paper presents a slight modification in SMT to make the approach more efficient and effective, and the experimental results have proven the method to be speed and accurate in the translation process.
A New Approach to Parts of Speech Tagging in Malayalamijcsit
Parts-of-speech tagging is the process of labeling each word in a sentence. A tag mentions the word’s
usage in the sentence. Usually, these tags indicate syntactic classification like noun or verb, and sometimes
include additional information, with case markers (number, gender etc) and tense markers. A large number
of current language processing systems use a parts-of-speech tagger for pre-processing.
There are mainly two approaches usually followed in Parts of Speech Tagging. Those are Rule based
Approach and Stochastic Approach. Rule based Approach use predefined handwritten rules. This is the
oldest approach and it use lexicon or dictionary for reference. Stochastic Approach use probabilistic and
statistical information to assign tag to words. It use large corpus, so that Time complexity and Space
complexity is high whereas Rule base approach has less complexity for both Time and Space. Stochastic
Approach is the widely used one nowadays because of its accuracy.
Malayalam is a Dravidian family of languages, inflectional with suffixes with the root word forms. The
currently used Algorithms are efficient Machine Learning Algorithms but these are not built for
Malayalam. So it affects the accuracy of the result of Malayalam POS Tagging.
My proposed Approach use Dictionary entries along with adjacent tag information. This algorithm use
Multithreaded Technology. Here tagging done with the probability of the occurrence of the sentence
structure along with the dictionary entry.
ADVANCEMENTS ON NLP APPLICATIONS FOR MANIPURI LANGUAGEijnlc
Manipuri is both a minority and morphologically rich language with genetic features similar to Tibeto Burman languages. It has Subject-Object-Verb (SOV) order, agglutinative verb morphology and is monosyllabic. Morphology and syntax are not clearly distinguished in this language. Natural Language
Processing (NLP) is a useful research field of computer science that deals with processing of a large amount of natural language corpus. The NLP applications encompass E-Dictionary, Morphological Analyzer, Reduplicated Multi-Word Expression (RMWE), Named Entity Recognition (NER), Part of Speech
(POS) Tagging, Machine Translation (MT), Word Net, Word Sense Disambiguation (WSD) etc. In this paper, we present a study on the advancements in NLP applications for Manipuri language, at the same time presenting a comparison table of the approaches and techniques adopted and the results obtained of each of the applications followed by a detail discussion of each work.
Survey on Indian CLIR and MT systems in Marathi LanguageEditor IJCATR
Cross Language Information Retrieval (CLIR) deals with retrieving relevant information stored in a language different from
the language of user’s query. This helps users to express the information need in their native languages. Machine translation based (MTbased)
approach of CLIR uses existing machine translation techniques to provide automatic translation of queries. This paper covers the
research work done in CLIR and MT systems for Marathi language in India.
ADVANCEMENTS ON NLP APPLICATIONS FOR MANIPURI LANGUAGEkevig
Manipuri is both a minority and morphologically rich language with genetic features similar to Tibeto Burman languages. It has Subject-Object-Verb (SOV) order, agglutinative verb morphology and ismonosyllabic. Morphology and syntax are not clearly distinguished in this language. Natural Language Processing (NLP) is a useful research field of computer science that deals with processing of a large amount of natural language corpus. The NLP applications encompass E-Dictionary, Morphological
Analyzer, Reduplicated Multi-Word Expression (RMWE), Named Entity Recognition (NER), Part of Speech (POS) Tagging, Machine Translation (MT), Word Net, Word Sense Disambiguation (WSD) etc. In this paper, we present a study on the advancements in NLP applications for Manipuri language, at the same time presenting a comparison table of the approaches and techniques adopted and the results obtained of each of the applications followed by a detail discussion of each work.
A Marathi Hidden-Markov Model Based Speech Synthesis Systemiosrjce
IOSR journal of VLSI and Signal Processing (IOSRJVSP) is a double blind peer reviewed International Journal that publishes articles which contribute new results in all areas of VLSI Design & Signal Processing. The goal of this journal is to bring together researchers and practitioners from academia and industry to focus on advanced VLSI Design & Signal Processing concepts and establishing new collaborations in these areas.
Design and realization of microelectronic systems using VLSI/ULSI technologies require close collaboration among scientists and engineers in the fields of systems architecture, logic and circuit design, chips and wafer fabrication, packaging, testing and systems applications. Generation of specifications, design and verification must be performed at all abstraction levels, including the system, register-transfer, logic, circuit, transistor and process levels
S URVEY O N M ACHINE T RANSLITERATION A ND M ACHINE L EARNING M ODELSijnlc
Globalization and growth of Internet users truly demands for almost all internet based applications to
support
l
oca
l l
anguages. Support
of
l
oca
l
l
anguages can be
given in all internet based applications by
means of Machine Transliteration
and
Machine Translation
.
This paper provides the thorough survey on
machine transliteration models and machine learning
approaches
used for machine transliteration
over the
period
of more than two decades
for internationally used languages as well as Indian languages.
Survey
shows that linguistic approach provides better results for the closely related languages and probability
based statistical approaches are good when one of the
languages is phonetic and other is non
-
phonetic.
B
etter accuracy can be achieved only by using Hybrid and Combined models.
Phrase Identification is one of the most critical and widely studied in Natural Language processing (NLP) tasks. Verb Phrase Identification within a sentence is very useful for a variety of application on NLP. One of the core enabling technologies required in NLP applications is a Morphological Analysis. This paper presents the Myanmar Verb Phrase Identification and Translation Algorithm and develops a Markov Model with Morphological Analysis. The system is based on Rule-Based Maximum Matching Approach. In Machine Translation, Large amount of information is needed to guide the translation process. Myanmar Language is inflected language and there are very few creations and researches of Lexicon in Myanmar, comparing to other language such as English, French and Czech etc. Therefore, this system is proposed Myanmar Verb Phrase identification and translation model based on Syntactic Structure and Morphology of Myanmar Language by using Myanmar- English bilingual lexicon. Markov Model is also used to reformulate the translation probability of Phrase pairs. Experiment results showed that proposed system can improve translation quality by applying morphological analysis on Myanmar Language.
Hybrid part of-speech tagger for non-vocalized arabic textijnlc
Part of speech tagging (POS tagging) has a crucial role in different fields of natural language processing
(NLP) including Speech Recognition, Natural Language Parsing, Information Retrieval and Multi Words
Term Extraction. This paper proposes an efficient and accurate POS Tagging technique for Arabic
language using hybrid approach. Due to the ambiguity issue, Arabic Rule-Based method suffers from
misclassified and unanalyzed words. To overcome these two problems, we propose a Hidden Markov
Model (HMM) integrated with Arabic Rule-Based method. Our POS tagger generates a set of three POS
tags: Noun, Verb, and Particle. The proposed technique uses the different contextual information of the
words with a variety of the features which are helpful to predict the various POS classes. To evaluate its
accuracy, the proposed method has been trained and tested with two corpora: the Holy Quran Corpus and
Kalimat Corpus for undiacritized Classical Arabic language. The experiment results demonstrate the
efficiency of our method for Arabic POS Tagging. In fact, the obtained accuracies rates are 97.6%, 96.8%
and 94.4% for respectively our Hybrid Tagger, HMM Tagger and for the Rule-Based Tagger with Holy
Quran Corpus. And for Kalimat Corpus we obtained 94.60%, 97.40% and 98% for respectively Rule-Based
Tagger, HMM Tagger and our Hybrid Tagger.
A Novel Approach for Rule Based Translation of English to Marathiaciijournal
This paper presents a design for rule-based machine translation system for English to Marathi language pair. The machine translation system will take input script as English sentence and parse with the help of Stanford parser. The Stanford parser will be used for main purposes on the source side processing, in the machine translation system. English to Marathi Bilingual dictionary is going to be created. The system will take the parsed output and separate the source text word by word and searches for their corresponding target words in the bilingual dictionary. The hand coded rules are written for Marathi inflections and also reordering rules are there. After applying the reordering rules, English sentence will be syntactically reordered to suit Marathi language
IMPROVING RULE-BASED METHOD FOR ARABIC POS TAGGING USING HMM TECHNIQUEcsandit
Part-of-speech (POS) tagger plays an important role in Natural Language Applications like
Speech Recognition, Natural Language Parsing, Information Retrieval and Multi Words Term
Extraction. This study proposes a building of an efficient and accurate POS Tagging technique
for Arabic language using statistical approach. Arabic Rule-Based method suffers from
misclassified and unanalyzed words due to the ambiguity issue. To overcome these two
problems, we propose a Hidden Markov Model (HMM) integrated with Arabic Rule-Based
method. Our POS tagger generates a set of 4 POS tags: Noun, Verb, Particle, and Quranic
Initial (INL). The proposed technique uses the different contextual information of the words with
a variety of the features which are helpful to predict the various POS classes. To evaluate its
accuracy, the proposed method has been trained and tested with the Holy Quran Corpus
containing 77 430 terms for undiacritized Classical Arabic language. The experiment results
demonstrate the efficiency of our method for Arabic POS Tagging. The obtained accuracies are
97.6% and 94.4% for respectively our method and for the Rule based tagger method.
Tamil-English Document Translation Using Statistical Machine Translation Appr...baskaran_md
The Paper presents a new method for translating a text document from Tamil to English. Our method is based on the Statistical Machine Translation Approach, combined with the Morphological Analysis, due to the fact that Tamil is a highly-inflected language. This paper presents a slight modification in SMT to make the approach more efficient and effective, and the experimental results have proven the method to be speed and accurate in the translation process.
A New Approach to Parts of Speech Tagging in Malayalamijcsit
Parts-of-speech tagging is the process of labeling each word in a sentence. A tag mentions the word’s
usage in the sentence. Usually, these tags indicate syntactic classification like noun or verb, and sometimes
include additional information, with case markers (number, gender etc) and tense markers. A large number
of current language processing systems use a parts-of-speech tagger for pre-processing.
There are mainly two approaches usually followed in Parts of Speech Tagging. Those are Rule based
Approach and Stochastic Approach. Rule based Approach use predefined handwritten rules. This is the
oldest approach and it use lexicon or dictionary for reference. Stochastic Approach use probabilistic and
statistical information to assign tag to words. It use large corpus, so that Time complexity and Space
complexity is high whereas Rule base approach has less complexity for both Time and Space. Stochastic
Approach is the widely used one nowadays because of its accuracy.
Malayalam is a Dravidian family of languages, inflectional with suffixes with the root word forms. The
currently used Algorithms are efficient Machine Learning Algorithms but these are not built for
Malayalam. So it affects the accuracy of the result of Malayalam POS Tagging.
My proposed Approach use Dictionary entries along with adjacent tag information. This algorithm use
Multithreaded Technology. Here tagging done with the probability of the occurrence of the sentence
structure along with the dictionary entry.
ADVANCEMENTS ON NLP APPLICATIONS FOR MANIPURI LANGUAGEijnlc
Manipuri is both a minority and morphologically rich language with genetic features similar to Tibeto Burman languages. It has Subject-Object-Verb (SOV) order, agglutinative verb morphology and is monosyllabic. Morphology and syntax are not clearly distinguished in this language. Natural Language
Processing (NLP) is a useful research field of computer science that deals with processing of a large amount of natural language corpus. The NLP applications encompass E-Dictionary, Morphological Analyzer, Reduplicated Multi-Word Expression (RMWE), Named Entity Recognition (NER), Part of Speech
(POS) Tagging, Machine Translation (MT), Word Net, Word Sense Disambiguation (WSD) etc. In this paper, we present a study on the advancements in NLP applications for Manipuri language, at the same time presenting a comparison table of the approaches and techniques adopted and the results obtained of each of the applications followed by a detail discussion of each work.
Survey on Indian CLIR and MT systems in Marathi LanguageEditor IJCATR
Cross Language Information Retrieval (CLIR) deals with retrieving relevant information stored in a language different from
the language of user’s query. This helps users to express the information need in their native languages. Machine translation based (MTbased)
approach of CLIR uses existing machine translation techniques to provide automatic translation of queries. This paper covers the
research work done in CLIR and MT systems for Marathi language in India.
ADVANCEMENTS ON NLP APPLICATIONS FOR MANIPURI LANGUAGEkevig
Manipuri is both a minority and morphologically rich language with genetic features similar to Tibeto Burman languages. It has Subject-Object-Verb (SOV) order, agglutinative verb morphology and ismonosyllabic. Morphology and syntax are not clearly distinguished in this language. Natural Language Processing (NLP) is a useful research field of computer science that deals with processing of a large amount of natural language corpus. The NLP applications encompass E-Dictionary, Morphological
Analyzer, Reduplicated Multi-Word Expression (RMWE), Named Entity Recognition (NER), Part of Speech (POS) Tagging, Machine Translation (MT), Word Net, Word Sense Disambiguation (WSD) etc. In this paper, we present a study on the advancements in NLP applications for Manipuri language, at the same time presenting a comparison table of the approaches and techniques adopted and the results obtained of each of the applications followed by a detail discussion of each work.
A Marathi Hidden-Markov Model Based Speech Synthesis Systemiosrjce
IOSR journal of VLSI and Signal Processing (IOSRJVSP) is a double blind peer reviewed International Journal that publishes articles which contribute new results in all areas of VLSI Design & Signal Processing. The goal of this journal is to bring together researchers and practitioners from academia and industry to focus on advanced VLSI Design & Signal Processing concepts and establishing new collaborations in these areas.
Design and realization of microelectronic systems using VLSI/ULSI technologies require close collaboration among scientists and engineers in the fields of systems architecture, logic and circuit design, chips and wafer fabrication, packaging, testing and systems applications. Generation of specifications, design and verification must be performed at all abstraction levels, including the system, register-transfer, logic, circuit, transistor and process levels
S URVEY O N M ACHINE T RANSLITERATION A ND M ACHINE L EARNING M ODELSijnlc
Globalization and growth of Internet users truly demands for almost all internet based applications to
support
l
oca
l l
anguages. Support
of
l
oca
l
l
anguages can be
given in all internet based applications by
means of Machine Transliteration
and
Machine Translation
.
This paper provides the thorough survey on
machine transliteration models and machine learning
approaches
used for machine transliteration
over the
period
of more than two decades
for internationally used languages as well as Indian languages.
Survey
shows that linguistic approach provides better results for the closely related languages and probability
based statistical approaches are good when one of the
languages is phonetic and other is non
-
phonetic.
B
etter accuracy can be achieved only by using Hybrid and Combined models.
Phrase Identification is one of the most critical and widely studied in Natural Language processing (NLP) tasks. Verb Phrase Identification within a sentence is very useful for a variety of application on NLP. One of the core enabling technologies required in NLP applications is a Morphological Analysis. This paper presents the Myanmar Verb Phrase Identification and Translation Algorithm and develops a Markov Model with Morphological Analysis. The system is based on Rule-Based Maximum Matching Approach. In Machine Translation, Large amount of information is needed to guide the translation process. Myanmar Language is inflected language and there are very few creations and researches of Lexicon in Myanmar, comparing to other language such as English, French and Czech etc. Therefore, this system is proposed Myanmar Verb Phrase identification and translation model based on Syntactic Structure and Morphology of Myanmar Language by using Myanmar- English bilingual lexicon. Markov Model is also used to reformulate the translation probability of Phrase pairs. Experiment results showed that proposed system can improve translation quality by applying morphological analysis on Myanmar Language.
Hybrid part of-speech tagger for non-vocalized arabic textijnlc
Part of speech tagging (POS tagging) has a crucial role in different fields of natural language processing
(NLP) including Speech Recognition, Natural Language Parsing, Information Retrieval and Multi Words
Term Extraction. This paper proposes an efficient and accurate POS Tagging technique for Arabic
language using hybrid approach. Due to the ambiguity issue, Arabic Rule-Based method suffers from
misclassified and unanalyzed words. To overcome these two problems, we propose a Hidden Markov
Model (HMM) integrated with Arabic Rule-Based method. Our POS tagger generates a set of three POS
tags: Noun, Verb, and Particle. The proposed technique uses the different contextual information of the
words with a variety of the features which are helpful to predict the various POS classes. To evaluate its
accuracy, the proposed method has been trained and tested with two corpora: the Holy Quran Corpus and
Kalimat Corpus for undiacritized Classical Arabic language. The experiment results demonstrate the
efficiency of our method for Arabic POS Tagging. In fact, the obtained accuracies rates are 97.6%, 96.8%
and 94.4% for respectively our Hybrid Tagger, HMM Tagger and for the Rule-Based Tagger with Holy
Quran Corpus. And for Kalimat Corpus we obtained 94.60%, 97.40% and 98% for respectively Rule-Based
Tagger, HMM Tagger and our Hybrid Tagger.
A Novel Approach for Rule Based Translation of English to Marathiaciijournal
This paper presents a design for rule-based machine translation system for English to Marathi language pair. The machine translation system will take input script as English sentence and parse with the help of Stanford parser. The Stanford parser will be used for main purposes on the source side processing, in the machine translation system. English to Marathi Bilingual dictionary is going to be created. The system will take the parsed output and separate the source text word by word and searches for their corresponding target words in the bilingual dictionary. The hand coded rules are written for Marathi inflections and also reordering rules are there. After applying the reordering rules, English sentence will be syntactically reordered to suit Marathi language
IMPROVING RULE-BASED METHOD FOR ARABIC POS TAGGING USING HMM TECHNIQUEcsandit
Part-of-speech (POS) tagger plays an important role in Natural Language Applications like
Speech Recognition, Natural Language Parsing, Information Retrieval and Multi Words Term
Extraction. This study proposes a building of an efficient and accurate POS Tagging technique
for Arabic language using statistical approach. Arabic Rule-Based method suffers from
misclassified and unanalyzed words due to the ambiguity issue. To overcome these two
problems, we propose a Hidden Markov Model (HMM) integrated with Arabic Rule-Based
method. Our POS tagger generates a set of 4 POS tags: Noun, Verb, Particle, and Quranic
Initial (INL). The proposed technique uses the different contextual information of the words with
a variety of the features which are helpful to predict the various POS classes. To evaluate its
accuracy, the proposed method has been trained and tested with the Holy Quran Corpus
containing 77 430 terms for undiacritized Classical Arabic language. The experiment results
demonstrate the efficiency of our method for Arabic POS Tagging. The obtained accuracies are
97.6% and 94.4% for respectively our method and for the Rule based tagger method.
A COMPREHENSIVE ANALYSIS OF STEMMERS AVAILABLE FOR INDIC LANGUAGES ijnlc
Stemming is the process of term conflation. It conflates all the word variants to a common form called as stem. It plays significant role in numerous Natural Language Processing (NLP) applications like morphological analysis, parsing, document summarization, text classification, part-of-speech tagging, question-answering system, machine translation, word sense disambiguation, information retrieval (IR), etc. Each of these tasks requires some pre-processing to be done. Stemming is one of the important building blocks for all these applications. This paper, presents an overview of various stemming techniques, evaluation criteria for stemmers and various existing stemmers for Indic languages.
Approach To Build A Marathi Text-To-Speech System Using Concatenative Synthes...IJERA Editor
Marathi is one of the oldest languages in India. This research paper describes the development of Marathi Textto-
Speech System (TTS). In Marathi TTS the input is Marathi text in Unicode. The voices are sampled from real
recorded speech. The objective of a text to speech system is to convert an arbitrary text into its corresponding
spoken waveform. Speech synthesis is a process of building machinery that can generate human-like speech
from any text input to imitate human speakers. Text processing and speech generation are two main components
of a text to speech system. To build a natural sounding speech synthesis system, it is essential that text
processing component produce an appropriate sequence of phonemic units. Generation of sequence of phonetic
units for a given standard word is referred to as letter to phoneme rule or text to phoneme rule. The
complexity of these rules and their derivation depends upon the nature of the language. The quality of a speech
synthesizer is judged by its closeness to the natural human voice and understandability. In this research paper we
described an approach to build a Marathi TTS system using concatenative synthesis method with syllable as a
basic unit of concatenation.
Design and Development of a Malayalam to English Translator- A Transfer Based...Waqas Tariq
This paper describes a transfer based scheme for translating Malayalam, a Dravidian language, to English. This system inputs Malayalam sentences and outputs equivalent English sentences. The system comprises of a preprocessor for splitting the compound words, a morphological parser for context disambiguation and chunking, a syntactic structure transfer module and a bilingual dictionary. All the modules are morpheme based to reduce dictionary size. The system does not rely on a stochastic approach and it is based on a rule-based architecture along with various linguistic knowledge components of both Malayalam and English. The system uses two sets of rules: rules for Malayalam morphology and rules for syntactic structure transfer from Malayalam to English. The system is designed using artificial intelligence techniques.
Implementation of Text To Speech for Marathi Language Using Transcriptions Co...IJERA Editor
This research paper presents the approach towards converting text to speech using new methodology. The text to
speech conversion system enables user to enter text in Marathi and as output it gets sound. The paper presents
the steps followed for converting text to speech for Marathi language and the algorithm used for it. The focus of
this paper is based on the tokenisation process and the orthographic representation of the text that shows the
mapping of letter to sound using the description of language’s phonetics. Here the main focus is on the text to
IPA transcription concept. It is in fact, a system that translates text to IPA transcription which is the primary
stage for text to speech conversion. The whole procedure for converting text to speech involves a great deal of
time as it’s not an easy task and requires efforts.
Implementation of Marathi Language Speech Databases for Large Dictionaryiosrjce
IOSR journal of VLSI and Signal Processing (IOSRJVSP) is a double blind peer reviewed International Journal that publishes articles which contribute new results in all areas of VLSI Design & Signal Processing. The goal of this journal is to bring together researchers and practitioners from academia and industry to focus on advanced VLSI Design & Signal Processing concepts and establishing new collaborations in these areas.
Design and realization of microelectronic systems using VLSI/ULSI technologies require close collaboration among scientists and engineers in the fields of systems architecture, logic and circuit design, chips and wafer fabrication, packaging, testing and systems applications. Generation of specifications, design and verification must be performed at all abstraction levels, including the system, register-transfer, logic, circuit, transistor and process levels
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
HANDLING CHALLENGES IN RULE BASED MACHINE TRANSLATION FROM MARATHI TO ENGLISHijnlc
Machine translation is being carried out by the researchers from quite a long time. However, it is still a
dream to materialize flawless Machine Translator and the small numbers of researchers has focussed at
translating Marathi Text to English. Perfect Machine Translation Systems have not yet been fully built
owing to the fact that languages differ syntactically as well as morphologically. Majority of the researchers
have opted for Statistical Machine translation whereas in this paper we have addressed the challenges of
Rule based Machine Translation. The paper describes the major divergences observed in language
Marathi and English and many challenges encountered while attempting to build machine translation
system form Marathi to English using rule based approach and rules to handle these challenges. As there
are exceptions to the rules and limit to the feasibility of maintaining knowledgebase, the practical machine
translation from Marathi to English is a complex task.
A Context-based Numeral Reading Technique for Text to Speech Systems IJECEIAES
This paper presents a novel technique for context based numeral reading in Indian language text to speech systems. The model uses a set of rules to determine the context of the numeral pronunciation and is being integrated with the waveform concatenation technique to produce speech out of the input text in Indian languages. For this purpose, the three Indian languages Odia, Hindi and Bengali are considered. To analyze the performance of the proposed technique, a set of experiments are performed considering different context of numeral pronunciations and the results are compared with existing syllable-based technique. The results obtained from different experiments shows the effectiveness of the proposed technique in producing intelligible speech out of the entered text utterances compared to the existing technique even with very less storage and execution time.
IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...ijnlc
Machine Translation for Indian languages is an emerging research area. Transliteration is one such module that we design while designing a translation system. Transliteration means mapping of source language text into the target language. Simple mapping decreases the efficiency of overall translation system. We propose the use of stemming and part-of-speech tagging for transliteration. The effectiveness of translation can be improved if we use part-of-speech tagging and stemming assisted transliteration. We have shown that much of the content in Gujarati gets transliterated while being processed for translation to Hindi language.
Dynamic Construction of Telugu Speech Corpus for Voice Enabled Text EditorWaqas Tariq
In recent decades speech interactive systems have gained increasing importance. Performance of an ASR system mainly depends on the availability of large corpus of speech. The conventional method of building a large vocabulary speech recognizer for any language uses a top-down approach to speech. This approach requires large speech corpus with sentence or phoneme level transcription of the speech utterances. The transcriptions must also include different speech order so that the recognizer can build models for all the sounds present. But, for Telugu language, because of its complex nature, a very large, well annotated speech database is very difficult to build. It is very difficult, if not impossible, to cover all the words of any Indian language, where each word may have thousands and millions of word forms. A significant part of grammar that is handled by syntax in English (and other similar languages) is handled within morphology in Telugu. Phrases including several words (that is, tokens) in English would be mapped on to a single word in Telugu.Telugu language is phonetic in nature in addition to rich in morphology. That is why the speech technology developed for English cannot be applied to Telugu language. This paper highlights the work carried out in an attempt to build a voice enabled text editor with capability of automatic term suggestion. Main claim of the paper is the recognition enhancement process developed by us for suitability of highly inflecting, rich morphological languages. This method results in increased speech recognition accuracy with very much reduction in corpus size. It also adapts Telugu words to the database dynamically, resulting in growth of the corpus.
Machine Translation Approaches and Design AspectsIOSR Journals
Machine translation is a sub-field of computational linguistics that investigates the use of software to
translate text or speech from one natural language to another. On a basic level, MT performs simple
substitution of words in one natural language for words in another, but that alone usually cannot produce a
good translation of a text, because recognition of whole phrases and their closest counterparts in the target
language is needed. The paper focuses on Example Based Machine Translation (EBMT) system that translates
sentences from English to Hindi. Development of a machine translation (MT) system typically demands a large
volume of computational resources. For example, rule based MT systems require extraction of syntactic and
semantic knowledge in the form of rules, statistics-based MT systems require huge parallel corpus containing
sentences in the source languages and their translations in target language. Requirement of such computational
resources is much less in respect of EBMT. This makes development of EBMT systems for English to Hindi
translation feasible, where availability of large-scale computational resources is still scarce. Example based
machine translation relies on the database for its translation. The frequency of word occurrence is important for
translation.
Marathi Text-To-Speech Synthesis using Natural Language Processingiosrjce
IOSR journal of VLSI and Signal Processing (IOSRJVSP) is a double blind peer reviewed International Journal that publishes articles which contribute new results in all areas of VLSI Design & Signal Processing. The goal of this journal is to bring together researchers and practitioners from academia and industry to focus on advanced VLSI Design & Signal Processing concepts and establishing new collaborations in these areas.
Design and realization of microelectronic systems using VLSI/ULSI technologies require close collaboration among scientists and engineers in the fields of systems architecture, logic and circuit design, chips and wafer fabrication, packaging, testing and systems applications. Generation of specifications, design and verification must be performed at all abstraction levels, including the system, register-transfer, logic, circuit, transistor and process levels
Interpretation of Sadhu into Cholit Bhasha by Cataloguing and Translation Systemijtsrd
Sadhu and Cholit bhasha are two significant Bangladeshi languages. Sadhu was functional in ancient era and had Sanskrit components but in present era cholit took its place. There are many formal and legal paper works present in Sadhu language which direly need to be translated in Cholit because its more favorable and speaker friendly. Therefore, this paper dealt with this issue by familiarizing the current era with Sadhu by creating a software. Different sentences were chosen and final data set was obtained by Principal Component Analysis PCA . MATLAB and Python are used for different machine learning algorithms. Most work is being done using Scikit Learn and MATLAB machine learning toolbox. It was found that Linear Discriminant Analysis LDA functions best. Speed prediction was also done and values were determined through graphs. It was inferred that this categorizer efficiently translated all Sadhu words to Cholit precisely and in well structured way. Therefore, Sadhu will not remain a complex language in this decade. Nakib Aman Turzo | Pritom Sarker | Biplob Kumar "Interpretation of Sadhu into Cholit Bhasha by Cataloguing and Translation System" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-3 , April 2020, URL: https://www.ijtsrd.com/papers/ijtsrd30792.pdf Paper Url :https://www.ijtsrd.com/engineering/computer-engineering/30792/interpretation-of-sadhu-into-cholit-bhasha-by-cataloguing-and-translation-system/nakib-aman-turzo
A Novel Approach for Rule Based Translation of English to Marathiaciijournal
This paper presents a design for rule-based machine translation system for English to Marathi language pair. The machine translation system will take input script as English sentence and parse with the help of Stanford parser. The Stanford parser will be used for main purposes on the source side processing, in the machine translation system. English to Marathi Bilingual dictionary is going to be created. The system will take the parsed output and separate the source text word by word and searches for their corresponding target words in the bilingual dictionary. The hand coded rules are written for Marathi inflections and also reordering rules are there. After applying the reordering rules, English sentence will be syntactically reordered to suit Marathi language
A Novel Approach for Rule Based Translation of English to Marathiaciijournal
This paper presents a design for rule-based machine translation system for English to Marathi language
pair. The machine translation system will take input script as English sentence and parse with the help of
Stanford parser. The Stanford parser will be used for main purposes on the source side processing, in the
machine translation system. English to Marathi Bilingual dictionary is going to be created. The system will
take the parsed output and separate the source text word by word and searches for their corresponding
target words in the bilingual dictionary. The hand coded rules are written for Marathi inflections and also
reordering rules are there. After applying the reordering rules, English sentence will be syntactically
reordered to suit Marathi language.
Similar to Grapheme-To-Phoneme Tools for the Marathi Speech Synthesis (20)
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)MdTanvirMahtab2
This presentation is about the working procedure of Shahjalal Fertilizer Company Limited (SFCL). A Govt. owned Company of Bangladesh Chemical Industries Corporation under Ministry of Industries.
Welcome to WIPAC Monthly the magazine brought to you by the LinkedIn Group Water Industry Process Automation & Control.
In this month's edition, along with this month's industry news to celebrate the 13 years since the group was created we have articles including
A case study of the used of Advanced Process Control at the Wastewater Treatment works at Lleida in Spain
A look back on an article on smart wastewater networks in order to see how the industry has measured up in the interim around the adoption of Digital Transformation in the Water Industry.
CFD Simulation of By-pass Flow in a HRSG module by R&R Consult.pptxR&R Consult
CFD analysis is incredibly effective at solving mysteries and improving the performance of complex systems!
Here's a great example: At a large natural gas-fired power plant, where they use waste heat to generate steam and energy, they were puzzled that their boiler wasn't producing as much steam as expected.
R&R and Tetra Engineering Group Inc. were asked to solve the issue with reduced steam production.
An inspection had shown that a significant amount of hot flue gas was bypassing the boiler tubes, where the heat was supposed to be transferred.
R&R Consult conducted a CFD analysis, which revealed that 6.3% of the flue gas was bypassing the boiler tubes without transferring heat. The analysis also showed that the flue gas was instead being directed along the sides of the boiler and between the modules that were supposed to capture the heat. This was the cause of the reduced performance.
Based on our results, Tetra Engineering installed covering plates to reduce the bypass flow. This improved the boiler's performance and increased electricity production.
It is always satisfying when we can help solve complex challenges like this. Do your systems also need a check-up or optimization? Give us a call!
Work done in cooperation with James Malloy and David Moelling from Tetra Engineering.
More examples of our work https://www.r-r-consult.dk/en/cases-en/
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdffxintegritypublishin
Advancements in technology unveil a myriad of electrical and electronic breakthroughs geared towards efficiently harnessing limited resources to meet human energy demands. The optimization of hybrid solar PV panels and pumped hydro energy supply systems plays a pivotal role in utilizing natural resources effectively. This initiative not only benefits humanity but also fosters environmental sustainability. The study investigated the design optimization of these hybrid systems, focusing on understanding solar radiation patterns, identifying geographical influences on solar radiation, formulating a mathematical model for system optimization, and determining the optimal configuration of PV panels and pumped hydro storage. Through a comparative analysis approach and eight weeks of data collection, the study addressed key research questions related to solar radiation patterns and optimal system design. The findings highlighted regions with heightened solar radiation levels, showcasing substantial potential for power generation and emphasizing the system's efficiency. Optimizing system design significantly boosted power generation, promoted renewable energy utilization, and enhanced energy storage capacity. The study underscored the benefits of optimizing hybrid solar PV panels and pumped hydro energy supply systems for sustainable energy usage. Optimizing the design of solar PV panels and pumped hydro energy supply systems as examined across diverse climatic conditions in a developing country, not only enhances power generation but also improves the integration of renewable energy sources and boosts energy storage capacities, particularly beneficial for less economically prosperous regions. Additionally, the study provides valuable insights for advancing energy research in economically viable areas. Recommendations included conducting site-specific assessments, utilizing advanced modeling tools, implementing regular maintenance protocols, and enhancing communication among system components.
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...Amil Baba Dawood bangali
Contact with Dawood Bhai Just call on +92322-6382012 and we'll help you. We'll solve all your problems within 12 to 24 hours and with 101% guarantee and with astrology systematic. If you want to take any personal or professional advice then also you can call us on +92322-6382012 , ONLINE LOVE PROBLEM & Other all types of Daily Life Problem's.Then CALL or WHATSAPP us on +92322-6382012 and Get all these problems solutions here by Amil Baba DAWOOD BANGALI
#vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore#blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #blackmagicforlove #blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #Amilbabainuk #amilbabainspain #amilbabaindubai #Amilbabainnorway #amilbabainkrachi #amilbabainlahore #amilbabaingujranwalan #amilbabainislamabad
Grapheme-To-Phoneme Tools for the Marathi Speech Synthesis
1. Sangramsing Kayte et al. Int. Journal of Engineering Research and Applications www.ijera.com
ISSN: 2248-9622, Vol. 5, Issue 11, (Part - 4) November 2015, pp.87-92
www.ijera.com 87 | P a g e
Grapheme-To-Phoneme Tools for the Marathi Speech Synthesis
Sangramsing N. Kayte1
, Monica Mundada1
, Dr. Charansing N. Kayte
2
,
Dr.Bharti Gawali*
1,3
Department of Computer Science and Information Technology Dr. Babasaheb Ambedkar Marathwada
University, Aurangabad
2
Department of Digital and Cyber Forensic, Aurangabad, Maharashtra
ABSTRACT
We describe in detail a Grapheme-to-Phoneme (G2P) converter required for the development of a good quality
Marathi Text-to-Speech (TTS) system. The Festival and Festvox framework is chosen for developing the
Marathi TTS system. Since Festival does not provide complete language processing support specie to various
languages, it needs to be augmented to facilitate the development of TTS systems in certain new languages.
Because of this, a generic G2P converter has been developed. In the customized Marathi G2P converter, we
have handled schwa deletion and compound word extraction. In the experiments carried out to test the Marathi
G2P on a text segment of 2485 words, 91.47% word phonetisation accuracy is obtained. This Marathi G2P has
been used for phonetising large text corpora which in turn is used in designing an inventory of phonetically rich
sentences. The sentences ensured a good coverage of the phonetically valid di-phones using only 1.3% of the
complete text corpora.
Keywords: Grapheme-to-Phoneme (G2P), TTS, Festival, festvox, di-phone, ICT.
I. INTRODUCTION
Marathi is an Indo-Aryan language spoken
predominantly by Marathi people of Maharashtra. It
is the official language and co-official language in
Maharashtra and Goa states of Western India
respectively, and is one of the 23 official languages
of India. There were 73 million speakers in 2001;
Marathi ranks 19th in the list of most spoken
languages in the world. Marathi has the fourth largest
number of native speakers in India. Marathi has some
of the oldest literature of all modern Indo-Aryan
languages, dating from about 900 AD. The major
dialects of Marathi are Standard Marathi and the
Varhadi dialect. There are other related languages
such as Khandeshi, Dangi, Vadvali and Samavedi.
Malvani Konkani has been heavily influenced by
Marathi varieties, only a very small percentage of
Indians use English as a means of communication.
This fact, coupled with the prevalent low literacy
rates make the use of conventional user interfaces
difficult in India. Spoken language interfaces enabled
with Text-to-Speech synthesis have the potential to
make information and other ICT based services
accessible to a large proportion of the
population[1][2][3][18].
However, good quality Marathi TTS systems
that can be used for real time deployment are not
available. Though a number of research prototypes of
Indian language TTS systems have been developed
[2][3], none of these are of quality that can be
compared to commercial grade TTS systems in
languages like English, German and French. The
foremost reason for this is that developing a TTS
system in a new language needs inputs for resolving
language specific issues requiring close collaboration
between linguists and technologists. Large amount of
annotated data is required for developing language-
processing modules like parts of speech (POS)
taggers, syntactic parsers, and intonation and duration
models. This is a resource intensive task, which is
prohibitive to academic researchers in emerging
economies like India [3][17][18][19].
In this paper, we describe the effort at HP Labs
India to develop a Hindi TTS system based on the
open source TTS framework, Festival [4]
[17][18][19]. This effort is a part of the Local
Language Speech Technology Initiative (LLSTI) [2],
which facilitates collaboration between motivated
groups around the world, by enabling sharing of
tools, expertise, and support and trainingfor TTS
development in local languages. LLSTI aims to
develop a TTS framework around Festival that will
allow for rapid development of TTS systems in any
language.
The focus of this paper is on a Marathi G2P
converter and on the design of phonetically balanced
sentences. Certain linguistic characteristics of the
Marathi language, like a large phone set and the
schwa deletion issue, pose problems for developing
unit selection inventories and Grapheme-to Phoneme
converters for a Marathi TTS systems. Section 3
describes the modules and algorithms developed at
HP Labs India to overcome these problems. In
Section 4, the design of phonetically rich Marathi
RESEARCH ARTICLE OPEN ACCESS
2. Sangramsing Kayte et al. Int. Journal of Engineering Research and Applications www.ijera.com
ISSN: 2248-9622, Vol. 5, Issue 11, (Part - 4) November 2015, pp.87-92
www.ijera.com 88 | P a g e
sentences is presented. Though the effort is focused
on Marathi, attention has been paid to make these
modules as language independent as possible and
scalable to other languages. The following section
outlines the reason for choosing Festival as the base
for developing such a system [2].
II. FESTIVAL FRAMEWORK
The Festival framework has been used
extensively by the research community in speech
synthesis. It has been chosen for implementing the
Marathi TTS System because of its flexible &
modular architecture, ease of configuration, and the
ability to add new external modules. However, TTS
system development for a new language requires
substantial amount of work, especially in the text
processing modules.
The language processing modules in Festival are
not adequate for certain languages and the reliance on
Scheme as a scripting language makes it difficult for
linguists to incorporate the necessary language
specific changes within Festival. Thus, we need new
tools or modules to be plugged into Festival [4].
III. TEXT ANALYSIS MODULES
CREATED BY HP LABS INDIA
Grapheme-to-Phoneme (G2P) Conversion
In a TTS system, the G2P module converts the
normalized orthographic text input into the
underlying linguistic and phonetic representation.
G2P conversion, therefore is the most basic step in a
TTS system. It is preferable that the phoneme/phone
sequence is also appropriately tagged with intonation
markups.
Marathi, like most Maharashtra languages, uses a
syllabic script that is largely phonetic in nature. Thus,
in most cases, there is a one-to-one mapping between
graphemes and the corresponding phones. Special
ligatures are used to denote nasalization and homo-
organic nasals. The Marathi phoneset consists of 86
phones due to the fact that it has a large set of distinct
phonemes with a relatively small number of
allophones. For example, the stop system in Marathi
shows a four way distinction, viz., voiced vs.
unvoiced, aspirated vs. unaspirated, for five places of
articulation. There is also a phonological distinction
between retroflex and alveolar, aspirated and non-
aspirated taps. Also, each of the vowels has a
phonologically distinct nasalized counterpart
[17][18][19].
3.1.1. G2P Converter Architecture
Fig. 1. G2P Converter Architecture
As part of the LLSTI initiative [4][16], a generic
G2P conversion system has been developed at HP
Labs India. The system consists of a rule processing
engine which is language independent. Language
specific information is fed into the system in the form
of lexicon, rules and mapping. The architecture of the
G2P is shown in Fig 1. This G2P framework, has
then been customized for Marathi. The default
character to phone mapping is defined in the mapping
file. The format of the mapping is shown below and
explained subsequently [4][5].
CharacterThe orthographic representation of the
character.
Type of the character, e.g. C (for Consonant), V
(for Vowel) etc.
Class:-The class to which the character belongs.
These class labels can be effectively used to write a
rule representing a broad set of characters.
Phoneme The default phonetic representation
of the Specific contexts are matched using rules. The
system triggers the rule that best fits the current
context. The rule format is shown below:
..... .....
Class label of the character as defined in the
character phone mapping. Together these s represent
the context that is being matched. Action
specification node. Each such node has the form:
Action Type: Pos: Phoneme Str
Action Type This field specifies the type of this
action node. Possible values are K (Keep), R
(Replace), I (Insert) and A (Append).
Pos the index of the character which is covered by
the context of the rule (1 Pos m). Phoneme Str
Phoneme string used by this action node.
The G2P first looks, for each input word, in the
exception lexicon. For the Marathi G2P, this lexicon
is the phonetic compound word lexicon which is
generated using the algorithm described. If the word
is present in the lexicon, the phonetic transcription of
the input word is taken from the lexicon itself.
Otherwise, the G2P applies the rules on this word and
produces the phonetic transcription.
Though Hindi phonology makes it easier to
devise letter to-sound rules, there are certain
exceptions to this. Some of these exceptions can be
handled by straight-forward rules and some cannot.
Schwa deletion is one such problem encountered in
3. Sangramsing Kayte et al. Int. Journal of Engineering Research and Applications www.ijera.com
ISSN: 2248-9622, Vol. 5, Issue 11, (Part - 4) November 2015, pp.87-92
www.ijera.com 89 | P a g e
Marathi G2P conversion. This problem arises with
the phonetization of words containing the schwa
vowel. Depending on certain constructs, the schwa
vowel is deleted in some cases and retained in others.
The orthography, however, provides minimal cues in
recognizing these constructs. The following sub-
section explains the issue of schwa deletion.
Schwa Deletion
The consonant graphemes in Hindi are
associated with an inherent schwa vowel which is not
represented in orthography. Let us consider the word
kalam “pen”. The word is represented in orthography
by the consonants viz. /k/, /l/ and /m/. The inherent
schwa is associated with each of these consonants
and the schwa is pronounced while pronouncing this
word. As a rule of thumb in Hindi, a word ending
schwa is always deleted. Vowels other than the
schwa are explicitly represented in orthographic text
with the help of specific ligatures or matras around
the consonant.
Several phonological rules have been proposed
in the literature [4] for schwa deletion. These rules
take into account morpheme-internal as well as
across morpheme-boundary information to explain
this phenomenon. The Bell Labs Multilingual TTS
attempted schwa deletion using morpheme boundary
information [2]. The schwa is deleted in the context
VC(C) CV where V stands for any vowel and C
stands for any consonant. Presence of morpheme
boundary on the left of the context for the schwa
deletion block restricts application of these rules. The
schwa is also not deleted if its deletion results in the
formation of an invalid consonant cluster
[17][18][19].
Encouraging results on schwa deletion using
morpheme boundary information is reported in [6]
[7]. Morpheme boundary in a word can be detected
using a morphological analyzer. However,
unavailability of high quality Hindi morphological
analyzers prevents adopting this approach.
The following sub-sections describe how the schwa
deletion problem has been handled in the Hindi text-
to-speech system currently under development at HP
Labs India.
Dhvani Schwa Deletion Rules
A set of schwa deletion rules have been
incorporated in the Dhvani speech synthesis system
[2] [8]. These rules are syllabic and can handle schwa
deletion in independent words. Table 1 lists these
rules. Here, W(i), C, V, H, Ch and
e
stand for the
character of a word, consonant, vowel, health,
consonant-health combination or a half consonant
and schwa vowel, respectively. The rules are applied
in a word from right-to-left.
Table 1. Dhvani Schwa Deletion Rules
A modified version of the schwa deletion rules
are used in the Hindi G2P converter. These rules,
however, fail in the case of compound words, which
are highly prevalent in Hindi. Compound words are
formed by two or more words concatenated to form a
single word. To illustrate this point, let us consider
the words lok “public” and sabhA “gathering”. These
two are independent words. However, by combining
these two words, the compound word loksabhA
“lower house of the parliament” is formed.
Application of the schwa deletion rules above on this
word produces the output [l o k
e
s bh A] which is
incorrect.
This problem can be overcome if the constituents
of a compound word are analysed separately, which
requires a phonetic lexicon of compound words. Such
a lexicon is not available for Hindi [2][8]. Hence, we
have proposed an algorithm [9] to automatically
extract compound words and identify its constituent
parts from a large text corpus.
Compound Word Extraction
The proposed algorithm starts with the
assumption that independently occurring words are
valid atomic words. A trie-like [10] structure is used
to store and efficiently match the words. A potential
compound word is detected if the currently processed
word is part of a word already present in the trie or a
word in the trie is a sub-string of the current word.
For each word ( ), there are several possibilities:
Case 1: is neither in the trie nor a constituent of
any of the words currently in the trie.
Case 2: is an independent word in the trie.
Case 3: is the initial part of a compound word
currently present in the trie as an independent word.
Case 4: is the second constituent of a compound
word currently present in the trie as an independent
word.
In Case 1, the algorithm will insert into the trie.
In Case 2, no change is needed. In Case 3, the
algorithm will generate the second constituent ( ) for
each of the potential compound words where is the
4. Sangramsing Kayte et al. Int. Journal of Engineering Research and Applications www.ijera.com
ISSN: 2248-9622, Vol. 5, Issue 11, (Part - 4) November 2015, pp.87-92
www.ijera.com 90 | P a g e
first constituent. After the generation of s, the end
node corresponding to in the trie becomes a leaf
node. The algorithm check for its presence in the trie
for each of the generated s. If is present in the trie,
the algorithm marks the combination ( ) as a
compound word. Otherwise, ( ) is inserted into a
suspicion list. The algorithm in its current form does
not handle Case
4.Before processing each, the algorithm checks
for its presence in the suspicion list. If is present in
the suspicion list as the second constituent, the entry (
) is removed from the suspicion list and is marked as
a compound word.
To illustrate the algorithm, let us consider a
sample lexicon with words in the sequence
lokgAthA, loksabhA, sabhA, lok, gAtha. If both
constituents of a compound are present in the corpus,
then the order of reading the compound and its
constituent parts is irrelevant. After this the word lok
is input. The algorithm first checks it in the suspicion
list. If it is not found there, the word is matched
against the contents of the trie generating the
constituent forms gAthA and sabhA. sabhA is
already present in the trie but gAthA is not yet read
in. Hence, lok sabhA is marked as a compound word.
But since gAthA is not present in the trie, lok gAthA
is inserted into the suspicion list. After this, the
algorithm processes the word gAthA. gAthA is
present in the suspicion list as the second constituent.
Hence, lok gAthA is removed from the suspicion list
and is marked as a compound word. Then the
algorithm tries to match gAthA in the trie and since it
is not present, it is inserted into the trie.
Hence, at the end of the algorithm, the best
possible atomic word forms are present in the trie
with compounds decomposed into their constituent
parts. An improvement of 1.6% in Hindi G2P
conversion was observed as a result of incorporating
the phonetic compound word lexicon into the Hindi
G2P converter. A detailed description of the
algorithm along with the issues involved are reported
in [2][8].
G2P Results
Experiments were carried out to test the accuracy
of the Hindi G2P. Details of the converter are
presented in Table 2
Total Phones Used 86
Total Character-Phone Mappings Used 62
Total Number of Rules 84
Total Entries in Compound Word Lexicon 48428
Table 2. Customized Hindi G2P Details
To test the phonetisation accuracy of the Hindi
G2P, a text file was randomly selected from the
Emille corpus [2][4][8]. The text was phonetised
using the G2P and the phonetic output was manually
verified. Phonetisation results are presented in Table
3.
Total Words Tested 3485
Wrong Schwa Deletion 81
Word Phonetisation Accuracy 97.67 %
Table 3. Result of Marathi G2P Conversion
IV. INVENTORY DESIGN
In concatenative speech synthesis approach,
speech is generated by concatenating real or coded
speech segments. A recent trend in concatenative
synthesis approach is to use large databases of
phonetically and prosodic ally varied speech, thereby
minimizing the degradation caused by pitch and time
scale modifications to match the target specification
and to minimize the concatenation discontinuities at
segment boundaries. The quality of output speech
primarily depends on the quality of the speech corpus
from which the speech segments are obtained.
Consequently, the design of the speech corpus is a
very important and a critical issue in the data-driven
concatenative synthesis approaches. The corpus
should contain as much variation as possible both
phonetically and prosaically and at the same time the
size of the corpus should be as small as possible both
to minimize the overhead incurred in database
preparation and to facilitate efficient unit selection at
synthesis time. In our work, to generate optimal text
to excise the speech corpus, a set of phonetically rich
sentences are prepared.
The problem of selecting a set of phonetically
rich sentences from a corpus can be mapped to the
Set Covering
Problem (SCP). A survey of various algorithms
for SCP is presented in [9]. The most widely used
algorithms for SCP are greedy algorithms [2] [8]. The
performances of greedy and spitting algorithms are
comparable. We implemented greedy algorithm,
since it is marginally better and less time consuming
[10][11].
This algorithm builds the cover step-by-step. The
algorithm recursively selects sentences. The first
sentence is the sentence with the largest unit count.
All units present in the sentence are then removed
from the to-be-covered list of units. The sentence
which covers maximum number of remaining units is
selected next. This process continues until all the
units are covered or the corpus is exhausted.
The units to be covered can be assigned weights
based on their frequency in a corpus. If complete
coverage is possible, then the weight of a unit can be
made equal to the inverse of its frequency. This
makes the algorithm focus more on less frequent
units. The idea is that the high frequency units are
anyway covered en-route. However, if it is known
apriori that complete coverage is not possible, then
the weight of the unit can be made equal to its
frequency.
5. Sangramsing Kayte et al. Int. Journal of Engineering Research and Applications www.ijera.com
ISSN: 2248-9622, Vol. 5, Issue 11, (Part - 4) November 2015, pp.87-92
www.ijera.com 91 | P a g e
The Marathi G2P converter was used to
phonetize the corpus. A di-phone frequency list was
prepared by running CMU-Statistical Language
Modeling Toolkit [12] [13] on the phonetised text
corpus. An exhaustive list of 7392 diphones was
generated using the Marathi phoneset. Phonotactic
constraints of the language were not taken into
consideration while generating the diphone list. Word
beginning, word ending, sentence beginning&
sentence ending contexts were considered while
selecting the sentences. Silence-phone combinations
were also taken into account. Text selection was
carried out in three phases. Details of various corpora
used in the text selection are presented in Table 4.
4.0.6. Phase 1 & 2
Sakale Newspaper text, which is part of the
Emille corpus, was used in this step. From a total of
163,517 sentences, 665 sentences were selected
resulting in the coverage of 2404 di-phones.
In the second phase, the idea was to cover the
units not covered in the first phase. Samne newspaper
text was used in this phase. From a total of 142122
sentences, 303 sentences were selected covering 379
di-phones. Hence, total di-phone coverage at the end
of this step was 2783.
4.0.7. Phase 3
At the start of this phase, the three remaining
corpora viz. Indian info, Miscellaneous (Misc.) and
Spoken were analyzed to see their coverage of the di-
phones which were not covered after the first two
phases. The results are shown in Fig 2. It is clear
from the graph that Marathi Misc corpus had the
highest coverage. The Spoken corpus showed great
promise
Corpus Name Genre Total
Sentences
Total
Words
Lokmat Newspaper 163175 3393192
Sakale Newspaper 142122 2910613
Samne Newspaper 36057 607017
Marathi Misc. Classified 128537 1728644
Spoken Radio Text 17720 507952
Table 4.Marathi Corpora Statistics
in terms of coverage but the limited size of this
corpus restricted its usage at this stage. Hence, the
Misc corpus was used for text selection at this stage.
Another interesting observation was the formation of
the plateau with the addition of more newspaper text
which is of the same genre as that of text used in
Phases 1 & 2. This illustrates the fact that phonetic
coverage attains saturation once a significant amount
of text of one type is analyzed and additional text of
the same type is unlikely to give significant increase
in coverage.
Fig. 2. Marathi Corpora Diphone Coverage Study
Another aspect observedduringthe analysis at
this phase was that the covered-diphoneto selected-
sentence ratio was nearing one i.e. a complete
sentence was being selected to cover just one unit.
This is undesirable since the size of the unit inventory
is directly proportional to the number of selected
sentences and an unnecessarily big speech corpus
results in additional overhead in speech-segmentation
and increased search time in the unit selection stage.
Hence, instead of the complete sentence, just the
word containing the particular di-phone was included
in the cover. This strategy selected 327 unique words
resulting in a coverage of 373 di-phones [15].
Selected sentences were hand corrected to remove
typographical errors and spelling mistakes. Final
coverage results are presented in Table 5. The
selected sentences and words are just 0.3% of the
three text corpora (Webdunia, Ranchi Exp., Misc.)
which were used in text selection.
Table 5. Results of Marathi Text Selection
Total Initial Di-phones 7392
Total Sentences Analyzed 433834
Count of Selected Sentences 968
Count of Selected Words 327
Total Di-phones Covered 3156
Total Uncovered Di-phones Count 4236
As a byproduct, the final uncovered di-phone list
can be considered as phonotactically invalid Marathi
di-phones since these di-phones remained uncovered
after analyzing a large text segment of more than
400,000 lines. A separate list of words was prepared
to cover the consonant clusters in Marathi
[14][15][19].
V. CONCLUSION
An effective G2P conversion module has been
developed for Marathi. This combines our new
algorithm for automatically creating a compound
word lexicon from a text corpus with the Dhvani
6. Sangramsing Kayte et al. Int. Journal of Engineering Research and Applications www.ijera.com
ISSN: 2248-9622, Vol. 5, Issue 11, (Part - 4) November 2015, pp.87-92
www.ijera.com 92 | P a g e
schwa deletion rule base. An inventory of
phonetically rich Marathi sentences has also been
created using a greedy algorithm on Marathi text
corpora. By using a large text corpora from various
sources, we are able to extract the phonotactic
constraints in Marathi by identifying the forbidden
di-phones.
REFERENCES
[1] Sangramsing N.kayte “Marathi Isolated-Word
Automatic Speech Recognition System based
on Vector Quantization (VQ) approach” 101th
Indian Science Congress Jammu University
03th Feb to 07 Feb 2014.
[2] Sangramsing Kayte, Monica Mundada, Dr.
Charansing Kayte "Di-phone-Based
Concatenative Speech Synthesis System for
Hindi" International Journal of Advanced
Research in Computer Science and Software
Engineering -Volume 5, Issue 10, October-
2015
[3] Sangramsing Kayte, Monica Mundada, Dr.
Charansing Kayte “Di-phone-Based
Concatenative Speech Synthesis Systems for
Marathi Language” OSR Journal of VLSI and
Signal Processing (IOSR-JVSP) Volume 5,
Issue 5, Ver. I (Sep –Oct. 2015), PP 76-81e-
ISSN: 2319 –4200, p-ISSN No. : 2319 –4197
[4] Sangramsing Kayte, Monica Mundada "Study
of Marathi Phones for Synthesis of Marathi
Speech from Text" International Journal of
Emerging Research in Management
&Technology ISSN: 2278-9359 (Volume-4,
Issue-10) October 2015
[5] Sangramsing Kayte, Monica Mundada, Dr.
Charansing Kayte "A Review of Unit
Selection Speech Synthesis International
Journal of Advanced Research in Computer
Science and Software Engineering -Volume 5,
Issue 10, October-2015
[6] “Dhvani - TTS System for Indian Languages,”
(http://dhvani.sourceforge.net), 2001.
[7] Sangramsing Kayte, Monica Mundada,
Santosh Gaikwad, Bharti Gawali
"PERFORMANCE EVALUATION OF
SPEECH SYNTHESIS TECHNIQUES FOR
ENGLISH LANGUAGE " International
Congress on Information and Communication
Technology 9-10 October, 2015
[8] Deepa S.R., A.G. Ramakrishnan, Kalika Bali,
and Partha Pratim Talukdar, “Automatic
Generation of Compound Words Lexicon for
Hindi Speech Synthesis,” Language Resources
& Evaluation Conference (LREC), 2004.
[9] Jeffrey D. Ullman Alfred V. Aho, John E.
Hopcroft, Data Structure and Algorithms,
Addison-Wesley Publishing Company,
Reading, MA, 1983.
[10] Emille corpus,”(http://www.emille.
lancs.ac.uk), 2003.
[11] Monica Mundada, Bharti Gawali,
Sangramsing Kayte "Recognition and
classification of speech and its related fluency
disorders" International Journal of Computer
Science and Information Technologies
(IJCSIT)
[12] A. Caprara, M. Fischetti, and P. Toth,
“Algorithms for the Set Covering Problem,”
Tech. Rep. No. OR-98-3, DEIS-Operations
Research Group, 1998.
[13] )Monica Mundada, Sangramsing Kayte, Dr.
Bharti Gawali "Classification of Fluent and
Dysfluent Speech Using KNN Classifier"
International Journal of Advanced Research in
Computer Science and Software Engineering
Volume 4, Issue 9, September 2014
[14] Sangramsing Kayte, Dr. Bharti Gawali
“Marathi Speech Synthesis: A review”
International Journal on Recent and Innovation
Trends in Computing and Communication
ISSN: 2321-8169 Volume: 3 Issue: 6 3708 –
3711
[15] Monica Mundada, Sangramsing Kayte
“Classification of speech and its related
fluency disorders Using KNN” ISSN2231-
0096 Volume-4 Number-3 Sept 2014
[16] Sangramsing Kayte, Monica Mundada, Dr.
Charansing Kayte " Performance Calculation
of Speech Synthesis Methods for Hindi
language IOSR Journal of VLSI and Signal
Processing (IOSR-JVSP) Volume 5, Issue 6,
Ver. I (Nov -Dec. 2015), PP 13-19e-ISSN:
2319 –4200, p-ISSN No. : 2319 –4197
[17] Sangramsing Kayte, Monica Mundada, Dr.
Charansing Kayte "A Corpus-Based
Concatenative Speech Synthesis System for
Marathi" IOSR Journal of VLSI and Signal
Processing (IOSR-JVSP) Volume 5, Issue 6,
Ver. I (Nov -Dec. 2015), PP 20-26e-ISSN:
2319 –4200, p-ISSN No. : 2319 –4197
[18] Sangramsing Kayte, Monica Mundada, Dr.
Charansing Kayte "A Marathi Hidden-Markov
Model Based Speech Synthesis System" IOSR
Journal of VLSI and Signal Processing (IOSR-
JVSP) Volume 5, Issue 6, Ver. I (Nov -Dec.
2015), PP 34-39e-ISSN: 2319 –4200, p-ISSN
No. : 2319 –4197
[19] Sangramsing Kayte, Monica Mundada, Dr.
Charansing Kayte "Implementation of Marathi
Language Speech Databases for Large
Dictionary" IOSR Journal of VLSI and Signal
Processing (IOSR-JVSP) Volume 5, Issue 6,
Ver. I (Nov -Dec. 2015), PP 40-45e-ISSN:
2319 –4200, p-ISSN No. : 2319 –4197