Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2101.06397
Cited By
To Understand Representation of Layer-aware Sequence Encoders as Multi-order-graph
16 January 2021
Sufeng Duan
Hai Zhao
MILM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"To Understand Representation of Layer-aware Sequence Encoders as Multi-order-graph"
25 / 25 papers shown
Title
BERT, mBERT, or BiBERT? A Study on Contextualized Embeddings for Neural Machine Translation
Haoran Xu
Benjamin Van Durme
Kenton W. Murray
65
59
0
09 Sep 2021
Incorporating BERT into Parallel Sequence Decoding with Adapters
Junliang Guo
Zhirui Zhang
Linli Xu
Hao-Ran Wei
Boxing Chen
Enhong Chen
75
69
0
13 Oct 2020
FLAT: Chinese NER Using Flat-Lattice Transformer
Xiaonan Li
Hang Yan
Xipeng Qiu
Xuanjing Huang
58
396
0
24 Apr 2020
Fixed Encoder Self-Attention Patterns in Transformer-Based Machine Translation
Alessandro Raganato
Yves Scherrer
Jörg Tiedemann
73
92
0
24 Feb 2020
Incorporating BERT into Neural Machine Translation
Jinhua Zhu
Yingce Xia
Lijun Wu
Di He
Tao Qin
Wen-gang Zhou
Houqiang Li
Tie-Yan Liu
FedML
AIMat
42
359
0
17 Feb 2020
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
Zhenzhong Lan
Mingda Chen
Sebastian Goodman
Kevin Gimpel
Piyush Sharma
Radu Soricut
SSL
AIMat
356
6,449
0
26 Sep 2019
Towards Making the Most of BERT in Neural Machine Translation
Jiacheng Yang
Mingxuan Wang
Hao Zhou
Chengqi Zhao
Yong Yu
Weinan Zhang
Lei Li
CLL
46
158
0
15 Aug 2019
Head-Driven Phrase Structure Grammar Parsing on Penn Treebank
Junru Zhou
Zhao Hai
68
144
0
05 Jul 2019
Theoretical Limitations of Self-Attention in Neural Sequence Models
Michael Hahn
68
271
0
16 Jun 2019
Learning Deep Transformer Models for Machine Translation
Qiang Wang
Bei Li
Tong Xiao
Jingbo Zhu
Changliang Li
Derek F. Wong
Lidia S. Chao
74
671
0
05 Jun 2019
Syntax-Enhanced Neural Machine Translation with Syntax-Aware Word Representations
Meishan Zhang
Zhenghua Li
Guohong Fu
Min Zhang
53
55
0
08 May 2019
Convolutional Self-Attention Networks
Baosong Yang
Longyue Wang
Derek F. Wong
Lidia S. Chao
Zhaopeng Tu
51
126
0
05 Apr 2019
Modeling Recurrence for Transformer
Jie Hao
Xing Wang
Baosong Yang
Longyue Wang
Jinfeng Zhang
Zhaopeng Tu
70
85
0
05 Apr 2019
Star-Transformer
Qipeng Guo
Xipeng Qiu
Pengfei Liu
Yunfan Shao
Xiangyang Xue
Zheng Zhang
69
264
0
25 Feb 2019
Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context
Zihang Dai
Zhilin Yang
Yiming Yang
J. Carbonell
Quoc V. Le
Ruslan Salakhutdinov
VLM
216
3,726
0
09 Jan 2019
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
1.1K
7,154
0
20 Apr 2018
Non-Autoregressive Neural Machine Translation
Jiatao Gu
James Bradbury
Caiming Xiong
Victor O.K. Li
R. Socher
97
795
0
07 Nov 2017
Towards String-to-Tree Neural Machine Translation
Roee Aharoni
Yoav Goldberg
NAI
47
154
0
16 Apr 2017
What do Neural Machine Translation Models Learn about Morphology?
Yonatan Belinkov
Nadir Durrani
Fahim Dalvi
Hassan Sajjad
James R. Glass
98
414
0
11 Apr 2017
A Structured Self-attentive Sentence Embedding
Zhouhan Lin
Minwei Feng
Cicero Nogueira dos Santos
Mo Yu
Bing Xiang
Bowen Zhou
Yoshua Bengio
113
2,138
0
09 Mar 2017
Language Modeling with Gated Convolutional Networks
Yann N. Dauphin
Angela Fan
Michael Auli
David Grangier
237
2,397
0
23 Dec 2016
Layer Normalization
Jimmy Lei Ba
J. Kiros
Geoffrey E. Hinton
407
10,481
0
21 Jul 2016
Effective Approaches to Attention-based Neural Machine Translation
Thang Luong
Hieu H. Pham
Christopher D. Manning
374
7,959
0
17 Aug 2015
Sequence to Sequence Learning with Neural Networks
Ilya Sutskever
Oriol Vinyals
Quoc V. Le
AIMat
431
20,541
0
10 Sep 2014
Neural Machine Translation by Jointly Learning to Align and Translate
Dzmitry Bahdanau
Kyunghyun Cho
Yoshua Bengio
AIMat
546
27,295
0
01 Sep 2014
1