Machine transformers of our model are modeled through sequence to sequence(Seq2Seq) neural architecture along with the attention mechanism. The language model leverages the likelihood of belonging to the target domain and predicts the next word. Furthermore, we explore three different scoring functions that are dot, general and concatenate and evaluate our augmented model compared to the baseline model using BLEU score as our metric.
- Improve model performance with data augmentation methods
- Explore formality style transfer datasets and models
- Compare results of different scoring variants
- Application of Attention mechanism to solve the bottleneck problem of seq2seq
- Wrote short full-paper on our findings and experiments
- Conducted baseline, augmented, ablation study experiments
- tknizer: tokenizer_formal, informal / augmented_formal, informal
- model: Data preprocessing, encoder/decoder with attention, train
- proposal
- final_report
- final_presentation (top 13 teams)