Audio-Visual Speech Recognition Using Lip Movement for Amharic Language

dc.contributor.advisorAssabie, Yaregal (PhD)
dc.contributor.authorBelete, Befkadu
dc.date.accessioned2019-08-19T11:51:43Z
dc.date.accessioned2023-11-29T04:06:01Z
dc.date.available2019-08-19T11:51:43Z
dc.date.available2023-11-29T04:06:01Z
dc.date.issued2017-10-05
dc.description.abstractAutomatic Speech Recognition (ASR) is a technology that allows a computer to identify the words that a person speaks into a microphone or telephone and convert it to a written text. In recent years, there have been many advances in automatic speech reading system with the inclusion of visual speech features to improve recognition accuracy under noisy conditions. By identifying lip movements and characterizing their associations with speech sounds, the performance of speech recognition systems can be improved, particularly when operating in noisy environments. The aim of this study is to design and develop automatic audio-visual Amharic speech recognition using lip reading. In this study, for face and mouth detection we use Viola-Jones object recognizer called haarcascade face detection and haarcascade mouth detection respectively, after the mouth detection ROI is extracted. Extracted ROI is used as an input for visual feature extraction. DWT is used for visual feature extraction and LDA is used to reduce visual feature vector. For audio feature extraction, we use MFCC. Integration of audio and visual features are done by decision fusion. As a result of this, we used three classifiers. The first one is the HMM classifier for audio only speech recognition, the second one is HMM classifier for visual only speech recognition and the third one is CHHM for audio- visual integration. In this study, we used our own data corpus called AAVC. We evaluated our audio-visual recognition system with two different sets: speaker dependent and speaker independent. We used those two evaluation sets for both phone (vowels) and isolated word recognition. For speaker dependent dataset, we found an overall word recognition of 60.42% for visual only, 65.31% for audio only and 70.1 % for audio-visual. We also found an overall vowels (phone) recognition of 71.45% for visual only, 76.34% for audio only and 83.92 % for audio-visual speech. For speaker independent dataset, we got an overall word recognition of 61% for visual only, 63.54% for audio only and 67.08% for audio-visual. The overall vowel (phone) recognition on the speaker independent dataset is 68.04% for visual only, 71.96% for audio only and 76.79 % for audio-visual speech.en_US
dc.identifier.urihttp://etd.aau.edu.et/handle/123456789/18803
dc.language.isoenen_US
dc.publisherAddis Ababa Universityen_US
dc.subjectAmharicen_US
dc.subjectLip-Readingen_US
dc.subjectVisemesen_US
dc.subjectAppearance-Based Featureen_US
dc.subjectDwten_US
dc.subjectAavcen_US
dc.titleAudio-Visual Speech Recognition Using Lip Movement for Amharic Languageen_US
dc.typeThesisen_US

Files

Original bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
Befkadu Belete 2017.pdf
Size:
3.26 MB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.71 KB
Format:
Plain Text
Description: