Chris Pollett >
Students > [Bio] [Blog] CS 297 Deliverables CS 298
|
CS298 ProposalDocument-Level Machine Translation with Hierarchical AttentionYu-Tang Shen (yutang.shen@sjsu.edu) Advisor: Dr. Chris Pollett Committee Members: Dr. Thomas Austin, Dr. William Andreopoulos Abstract:Current attention-based machine translation models commonly bound the input text length around 1,000 words. Although it is adequate for sentence-level translation, models that can handle longer input texts, such as documents, are needed. This project aims to implement a hierarchical attention model to meet the need. Over the course of this project involves implementing and testing two kinds of hierarchical attention model under two configurations: the default attention mechanism -- full attention and the Big bird attention mechanism will be implemented, while bi-directional translations between English and Chinese will be tested. CS297 Results
Proposed Schedule
Time allocation:
Key Deliverables:
Innovations and Challenges
References:[1] Yang, Zichao, Diyi Yang, Chris Dyer, Xiaodong He, Alex Smola, and Eduard Hovy. "Hierarchical attention networks for document classification." in Proceedings of the 2016 conference of the North American chapter of the association for computational linguistics: human language technologies. 2016. [2] Vaswani, Ashish, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. "Attention is all you need." in Advances in neural information processing systems 30. 2017. [3] Tian, Liang, Derek F. Wong, Lidia S. Chao, Paulo Quaresma, Francisco Oliveira, and Lu Yi. "UM-Corpus: A Large English-Chinese Parallel Corpus for Statistical Machine Translation," in Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14). 2014. [4] Manzil Zaheer, Guru Guruganesh, Avinava Dubey, Joshua Ainslie, Chris Alberti, Santiago Ontanon, Philip Pham, Anirudh Ravula, Qifan Wang, Li Yang, Amr Ahmed. "Big bird: Transformers for longer sequences." in Advances in neural information processing systems 33. 2020 [5] Lesly Miculicich, Dhananjay Ram, Nikolaos Pappas, and James Henderson. "Document-level neural machine translation with hierarchical attention networks." in arXiv. 2018. [6] Papineni, Kishore, Salim Roukos, Todd Ward, and Wei-Jing Zhu. "Bleu: a method for automatic evaluation of machine translation." in Proceedings of the 40th annual meeting of the Association for Computational Linguistics. 2002. [7] Pappas Nikolaos, and Andrei Popescu-Belis. "Multilingual hierarchical attention networks for document classification." in arXiv. 2017. |