Author (Your Name)

Vivek Sah, Colby CollegeFollow

Date of Award


Document Type

Honors Thesis (Open Access)


Colby College. Computer Science Dept.


Stephanie Taylor


Machine translation is a widely researched topic in the field of Natural Language Processing and most recently, neural network models have been shown to be very effective at this task. The model, called sequence-to-sequence model, learns to map an input sequence in one language to a vector of fixed dimensionality and then map that vector to an output sequence in another language without any human intervention provided that there is enough training data. Focusing on English-French translation, in this paper, I present a way to simplify the learning process by replacing English input sentences by word-by-word translation of those sentences. I found that this approach improves the performance of a sequence-to-sequence model which is 3-layer deep and has a bidirectional LSTM encoder by more than 30% on the same dataset.


deep learning, translation, sequence to sequence, neural networks, recurrent neural networks, machine translation