Attention-based Amharic-to-Wolaita Neural Machine Translation Workineh Wogaso Gaga

dc.contributor.advisorAssabie, Yaregal (PhD)
dc.contributor.authorWogaso, Workineh
dc.date.accessioned2021-08-05T09:36:59Z
dc.date.accessioned2023-11-04T12:23:14Z
dc.date.available2021-08-05T09:36:59Z
dc.date.available2023-11-04T12:23:14Z
dc.date.issued10/5/2020
dc.description.abstractNatural language (NL) is one of the fundamental aspects of human behaviors and we can communicate all over the world through it. Natural Language Processing (NLP) is a branch of artificial intelligence that deals with the interaction between computers and humans using the NL. In the NLP world, every NL should be well understood by the machine. Machine Translation (MT) is a process by which computer software understand NL and an automatic translation of a text from one NL (source language) to another (target language). Neural Machine Translation (NMT) is a recently proposed approach to MT and has been able to gain State of The Art (SOTA) translation quality for the last few years. Unlike the traditional MT approaches, NMT aims at building a single neural network that can be jointly tuned to maximize translation performance. In this thesis we proposed Attention-based Amharic-to-Wolaita NMT. We built our system based on Encoder-Decoder architecture also called Sequence to Sequence (Seq2Seq) model by integrating a Recurrent Neural Network (RNN) and Gated Recurrent Unit (GRU). For comparison of our translation performance, we developed non-attention-based Amharic-to-Wolaita NMT. An encoder in basic (non-attention-based) Encoder-Decoder architecture encodes the complete information of the source (Amharic) sequence into a single real-valued vector called context vector which is passed to the Decoder to produce an output (Wolaita) sequence. Here, a context vector summarizes the entire input sequence into a single vector. As the length of the sentence increases, the inter-dependency of words is loosely related and it is a major drawback. The second problem of basic Encoder-Decoder model is handling of a large number of vocabulary sizes. As each word in the sentence is visited, it must be assigned a new identity number. But when the length of the corpus increases, the number used for word representation and dimension of word vector needed becomes higher. These two basic issues are solved using an attention mechanism. However, either attention-based or non-attention based NMT have not been developed for Amharic-to-Wolaita language pair. Thus, we developed attention-based NMT for Amharic-to-Wolaita translation and compare it against a non-attention-based NMT system. We used a BLEU score evaluation technique for system evaluation and we got 0.5960 BLEU score for non-attention-based system and 0.6258 BLEU score for attention-based system. Thus, attention-based system obtains up to +0.02978 BLEU improvements over non-attention-based NMT system.en_US
dc.identifier.urihttp://etd.aau.edu.et/handle/123456789/27609
dc.language.isoenen_US
dc.publisherAddis Ababa Universityen_US
dc.subjectNatural Language Processingen_US
dc.subjectMachine Translationen_US
dc.subjectNeural Machine Translationen_US
dc.subjectRecurrent Neural Networken_US
dc.subjectLanguage Modellingen_US
dc.subjectAttention-Based Encoder-Decoder Modelen_US
dc.titleAttention-based Amharic-to-Wolaita Neural Machine Translation Workineh Wogaso Gagaen_US
dc.typeThesisen_US

Files

Original bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
Workineh Wogaso 2020.pdf
Size:
1.67 MB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.71 KB
Format:
Plain Text
Description:

Collections