Repository logo

Comparing Encoder-Decoder Architectures for Neural Machine Translation: A Challenge Set Approach

dc.contributor.authorDoan, Coraline
dc.contributor.supervisorMarshman, Elizabeth
dc.date.accessioned2021-11-19T16:24:55Z
dc.date.available2021-11-19T16:24:55Z
dc.date.issued2021-11-19en_US
dc.description.abstractMachine translation (MT) as a field of research has known significant advances in recent years, with the increased interest for neural machine translation (NMT). By combining deep learning with translation, researchers have been able to deliver systems that perform better than most, if not all, of their predecessors. While the general consensus regarding NMT is that it renders higher-quality translations that are overall more idiomatic, researchers recognize that NMT systems still struggle to deal with certain classic difficulties, and that their performance may vary depending on their architecture. In this project, we implement a challenge-set based approach to the evaluation of examples of three main NMT architectures: convolutional neural network-based systems (CNN), recurrent neural network-based (RNN) systems, and attention-based systems, trained on the same data set for English to French translation. The challenge set focuses on a selection of lexical and syntactic difficulties (e.g., ambiguities) drawn from literature on human translation, machine translation, and writing for translation, and also includes variations in sentence lengths and structures that are recognized as sources of difficulties even for NMT systems. This set allows us to evaluate performance in multiple areas of difficulty for the systems overall, as well as to evaluate any differences between architectures’ performance. Through our challenge set, we found that our CNN-based system tends to reword sentences, sometimes shifting their meaning, while our RNN-based system seems to perform better when provided with a larger context, and our attention-based system seems to struggle the longer a sentence becomes.en_US
dc.identifier.urihttp://hdl.handle.net/10393/42936
dc.identifier.urihttp://dx.doi.org/10.20381/ruor-27153
dc.language.isoenen_US
dc.publisherUniversité d'Ottawa / University of Ottawaen_US
dc.rightsAttribution-NonCommercial 4.0 International*
dc.rights.urihttp://creativecommons.org/licenses/by-nc/4.0/*
dc.subjectneural machine translationen_US
dc.subjectmachine translation evaluationen_US
dc.subjectconvolutional neural networken_US
dc.subjectrecurrent neural networken_US
dc.subjectattention-based neural machine translationen_US
dc.subjectchallenge seten_US
dc.titleComparing Encoder-Decoder Architectures for Neural Machine Translation: A Challenge Set Approachen_US
dc.typeThesisen_US
thesis.degree.disciplineArtsen_US
thesis.degree.levelMastersen_US
thesis.degree.nameMAen_US
uottawa.departmentTraduction et interprétation / Translation and Interpretationen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail ImageThumbnail Image
Name:
Doan_Coraline_2021_thesis.pdf
Size:
4.73 MB
Format:
Adobe Portable Document Format
Description:

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail ImageThumbnail Image
Name:
license.txt
Size:
6.65 KB
Format:
Item-specific license agreed upon to submission
Description: