ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1808.06226
  4. Cited By
SentencePiece: A simple and language independent subword tokenizer and
  detokenizer for Neural Text Processing

SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing

19 August 2018
Taku Kudo
John Richardson
ArXiv (abs)PDFHTMLGithub (10925★)

Papers citing "SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing"

50 / 1,950 papers shown
Title
Convolutions and Self-Attention: Re-interpreting Relative Positions in
  Pre-trained Language Models
Convolutions and Self-Attention: Re-interpreting Relative Positions in Pre-trained Language Models
Tyler A. Chang
Yifan Xu
Weijian Xu
Zhuowen Tu
ViT
57
15
0
10 Jun 2021
Input Augmentation Improves Constrained Beam Search for Neural Machine
  Translation: NTT at WAT 2021
Input Augmentation Improves Constrained Beam Search for Neural Machine Translation: NTT at WAT 2021
Katsuki Chousa
Makoto Morishita
75
6
0
10 Jun 2021
Text-to-SQL in the Wild: A Naturally-Occurring Dataset Based on Stack
  Exchange Data
Text-to-SQL in the Wild: A Naturally-Occurring Dataset Based on Stack Exchange Data
Moshe Hazoom
Vibhor Malik
Ben Bogin
81
52
0
09 Jun 2021
RealTranS: End-to-End Simultaneous Speech Translation with Convolutional
  Weighted-Shrinking Transformer
RealTranS: End-to-End Simultaneous Speech Translation with Convolutional Weighted-Shrinking Transformer
Xingshan Zeng
Liangyou Li
Qun Liu
67
48
0
09 Jun 2021
SpeechBrain: A General-Purpose Speech Toolkit
SpeechBrain: A General-Purpose Speech Toolkit
Mirco Ravanelli
Titouan Parcollet
Peter William VanHarn Plantinga
Aku Rouhe
Samuele Cornell
...
William Aris
Hwidong Na
Yan Gao
R. Mori
Yoshua Bengio
126
769
0
08 Jun 2021
Diverse Pretrained Context Encodings Improve Document Translation
Diverse Pretrained Context Encodings Improve Document Translation
Domenic Donato
Lei Yu
Chris Dyer
47
16
0
07 Jun 2021
The FLORES-101 Evaluation Benchmark for Low-Resource and Multilingual
  Machine Translation
The FLORES-101 Evaluation Benchmark for Low-Resource and Multilingual Machine Translation
Naman Goyal
Cynthia Gao
Vishrav Chaudhary
Peng-Jen Chen
Guillaume Wenzek
Da Ju
Sanjan Krishnan
MarcÁurelio Ranzato
Francisco Guzman
Angela Fan
138
589
0
06 Jun 2021
How to Adapt Your Pretrained Multilingual Model to 1600 Languages
How to Adapt Your Pretrained Multilingual Model to 1600 Languages
Abteen Ebrahimi
Katharina Kann
LRMVLM
98
70
0
03 Jun 2021
Lightweight Adapter Tuning for Multilingual Speech Translation
Lightweight Adapter Tuning for Multilingual Speech Translation
Hang Le
J. Pino
Changhan Wang
Jiatao Gu
D. Schwab
Laurent Besacier
139
90
0
02 Jun 2021
Lower Perplexity is Not Always Human-Like
Lower Perplexity is Not Always Human-Like
Tatsuki Kuribayashi
Yohei Oseki
Takumi Ito
Ryo Yoshida
Masayuki Asahara
Kentaro Inui
62
77
0
02 Jun 2021
Sub-Character Tokenization for Chinese Pretrained Language Models
Sub-Character Tokenization for Chinese Pretrained Language Models
Chenglei Si
Zhengyan Zhang
Yingfa Chen
Fanchao Qi
Xiaozhi Wang
Zhiyuan Liu
Yasheng Wang
Qun Liu
Maosong Sun
53
12
0
01 Jun 2021
Multilingual Speech Translation with Unified Transformer: Huawei Noah's
  Ark Lab at IWSLT 2021
Multilingual Speech Translation with Unified Transformer: Huawei Noah's Ark Lab at IWSLT 2021
Xingshan Zeng
Liangyou Li
Qun Liu
31
2
0
01 Jun 2021
Beyond Noise: Mitigating the Impact of Fine-grained Semantic Divergences
  on Neural Machine Translation
Beyond Noise: Mitigating the Impact of Fine-grained Semantic Divergences on Neural Machine Translation
Eleftheria Briakou
Marine Carpuat
51
14
0
31 May 2021
Effective Batching for Recurrent Neural Network Grammars
Effective Batching for Recurrent Neural Network Grammars
Hiroshi Noji
Yohei Oseki
GNN
71
17
0
31 May 2021
Towards One Model to Rule All: Multilingual Strategy for Dialectal
  Code-Switching Arabic ASR
Towards One Model to Rule All: Multilingual Strategy for Dialectal Code-Switching Arabic ASR
Shammur A. Chowdhury
A. Hussein
Ahmed Abdelali
Ahmed M. Ali
78
36
0
31 May 2021
How to Split: the Effect of Word Segmentation on Gender Bias in Speech
  Translation
How to Split: the Effect of Word Segmentation on Gender Bias in Speech Translation
Marco Gaido
Beatrice Savoldi
L. Bentivogli
Matteo Negri
Marco Turchi
101
15
0
28 May 2021
ByT5: Towards a token-free future with pre-trained byte-to-byte models
ByT5: Towards a token-free future with pre-trained byte-to-byte models
Linting Xue
Aditya Barua
Noah Constant
Rami Al-Rfou
Sharan Narang
Mihir Kale
Adam Roberts
Colin Raffel
134
508
0
28 May 2021
SciFive: a text-to-text transformer model for biomedical literature
SciFive: a text-to-text transformer model for biomedical literature
Long Phan
J. Anibal
H. Tran
Shaurya Chanana
Erol Bahadroglu
Alec Peltekian
G. Altan-Bonnet
MedIm
68
151
0
28 May 2021
Extremely low-resource machine translation for closely related languages
Extremely low-resource machine translation for closely related languages
Maali Tars
Andre Tättar
Mark Fishel
37
17
0
27 May 2021
CogView: Mastering Text-to-Image Generation via Transformers
CogView: Mastering Text-to-Image Generation via Transformers
Ming Ding
Zhuoyi Yang
Wenyi Hong
Wendi Zheng
Chang Zhou
...
Junyang Lin
Xu Zou
Zhou Shao
Hongxia Yang
Jie Tang
ViTVLM
134
783
0
26 May 2021
LMMS Reloaded: Transformer-based Sense Embeddings for Disambiguation and
  Beyond
LMMS Reloaded: Transformer-based Sense Embeddings for Disambiguation and Beyond
Daniel Loureiro
A. Jorge
Jose Camacho-Collados
84
26
0
26 May 2021
Joint Optimization of Tokenization and Downstream Model
Joint Optimization of Tokenization and Downstream Model
Tatsuya Hiraoka
Sho Takase
Kei Uchiumi
Atsushi Keyaki
Naoaki Okazaki
66
17
0
26 May 2021
Focus Attention: Promoting Faithfulness and Diversity in Summarization
Focus Attention: Promoting Faithfulness and Diversity in Summarization
Rahul Aralikatte
Shashi Narayan
Joshua Maynez
S. Rothe
Ryan T. McDonald
112
46
0
25 May 2021
Learning Language Specific Sub-network for Multilingual Machine
  Translation
Learning Language Specific Sub-network for Multilingual Machine Translation
Zehui Lin
Liwei Wu
Mingxuan Wang
Lei Li
78
83
0
19 May 2021
Analysis of GraphSum's Attention Weights to Improve the Explainability
  of Multi-Document Summarization
Analysis of GraphSum's Attention Weights to Improve the Explainability of Multi-Document Summarization
M. Hickmann
Fabian Wurzberger
Megi Hoxhalli
Arne Lochner
Jessica Töllich
A. Scherp
57
1
0
19 May 2021
CoTexT: Multi-task Learning with Code-Text Transformer
CoTexT: Multi-task Learning with Code-Text Transformer
Long Phan
H. Tran
Daniel Le
Hieu Duy Nguyen
J. Anibal
Alec Peltekian
Yanfang Ye
92
136
0
18 May 2021
Exploiting Adapters for Cross-lingual Low-resource Speech Recognition
Exploiting Adapters for Cross-lingual Low-resource Speech Recognition
Wenxin Hou
Hanlin Zhu
Yidong Wang
Jindong Wang
Tao Qin
Renjun Xu
T. Shinozaki
62
65
0
18 May 2021
Do Context-Aware Translation Models Pay the Right Attention?
Do Context-Aware Translation Models Pay the Right Attention?
Kayo Yin
Patrick Fernandes
Danish Pruthi
Aditi Chaudhary
André F. T. Martins
Graham Neubig
64
34
0
14 May 2021
Which transformer architecture fits my data? A vocabulary bottleneck in
  self-attention
Which transformer architecture fits my data? A vocabulary bottleneck in self-attention
Noam Wies
Yoav Levine
Daniel Jannai
Amnon Shashua
90
20
0
09 May 2021
FastCorrect: Fast Error Correction with Edit Alignment for Automatic
  Speech Recognition
FastCorrect: Fast Error Correction with Edit Alignment for Automatic Speech Recognition
Yichong Leng
Xu Tan
Linchen Zhu
Jin Xu
Renqian Luo
Linquan Liu
Tao Qin
Xiang-Yang Li
Ed Lin
Tie-Yan Liu
KELM
109
64
0
09 May 2021
FNet: Mixing Tokens with Fourier Transforms
FNet: Mixing Tokens with Fourier Transforms
James Lee-Thorp
Joshua Ainslie
Ilya Eckstein
Santiago Ontanon
130
536
0
09 May 2021
Measuring and Increasing Context Usage in Context-Aware Machine
  Translation
Measuring and Increasing Context Usage in Context-Aware Machine Translation
Patrick Fernandes
Kayo Yin
Graham Neubig
André F. T. Martins
86
57
0
07 May 2021
Generalising Multilingual Concept-to-Text NLG with Language Agnostic
  Delexicalisation
Generalising Multilingual Concept-to-Text NLG with Language Agnostic Delexicalisation
Giulio Zhou
Gerasimos Lampouras
43
3
0
07 May 2021
Computer-Aided Design as Language
Computer-Aided Design as Language
Yaroslav Ganin
Sergey Bartunov
Yujia Li
E. Keller
Stefano Saliceti
3DV
152
95
0
06 May 2021
HerBERT: Efficiently Pretrained Transformer-based Language Model for
  Polish
HerBERT: Efficiently Pretrained Transformer-based Language Model for Polish
Robert Mroczkowski
Piotr Rybak
Alina Wróblewska
Ireneusz Gawlik
86
85
0
04 May 2021
Searchable Hidden Intermediates for End-to-End Models of Decomposable
  Sequence Tasks
Searchable Hidden Intermediates for End-to-End Models of Decomposable Sequence Tasks
Siddharth Dalmia
Brian Yan
Vikas Raunak
Florian Metze
Shinji Watanabe
89
31
0
02 May 2021
An analysis of full-size Russian complexly NER labelled corpus of
  Internet user reviews on the drugs based on deep learning and language neural
  nets
An analysis of full-size Russian complexly NER labelled corpus of Internet user reviews on the drugs based on deep learning and language neural nets
A. Sboev
S. Sboeva
I. Moloshnikov
A. Gryaznov
R. Rybka
A. Naumov
A. Selivanov
Gleb Rylkov
Viacheslav Ilyin
31
3
0
30 Apr 2021
Paraphrastic Representations at Scale
Paraphrastic Representations at Scale
John Wieting
Kevin Gimpel
Graham Neubig
Taylor Berg-Kirkpatrick
133
19
0
30 Apr 2021
Diversity-Aware Batch Active Learning for Dependency Parsing
Diversity-Aware Batch Active Learning for Dependency Parsing
Tianze Shi
Adrian Benton
Igor Malioutov
Ozan Irsoy
30
10
0
28 Apr 2021
Generating abstractive summaries of Lithuanian news articles using a
  transformer model
Generating abstractive summaries of Lithuanian news articles using a transformer model
Lukas Stankevicius
M. Lukoševičius
50
3
0
23 Apr 2021
End-to-end Speech Translation via Cross-modal Progressive Training
End-to-end Speech Translation via Cross-modal Progressive Training
Rong Ye
Mingxuan Wang
Lei Li
77
73
0
21 Apr 2021
Cross-Attention is All You Need: Adapting Pretrained Transformers for
  Machine Translation
Cross-Attention is All You Need: Adapting Pretrained Transformers for Machine Translation
Mozhdeh Gheini
Xiang Ren
Jonathan May
LRM
89
116
0
18 Apr 2021
AmericasNLI: Evaluating Zero-shot Natural Language Understanding of
  Pretrained Multilingual Models in Truly Low-resource Languages
AmericasNLI: Evaluating Zero-shot Natural Language Understanding of Pretrained Multilingual Models in Truly Low-resource Languages
Abteen Ebrahimi
Manuel Mager
Arturo Oncevay
Vishrav Chaudhary
Luis Chiruzzo
...
Graham Neubig
Alexis Palmer
Rolando A. Coto Solano
Ngoc Thang Vu
Katharina Kann
160
74
0
18 Apr 2021
A Simple and Effective Positional Encoding for Transformers
A Simple and Effective Positional Encoding for Transformers
Pu-Chin Chen
Henry Tsai
Srinadh Bhojanapalli
Hyung Won Chung
Yin-Wen Chang
Chun-Sung Ferng
118
66
0
18 Apr 2021
The Power of Scale for Parameter-Efficient Prompt Tuning
The Power of Scale for Parameter-Efficient Prompt Tuning
Brian Lester
Rami Al-Rfou
Noah Constant
VPVLM
674
4,119
0
18 Apr 2021
From Fully Trained to Fully Random Embeddings: Improving Neural Machine
  Translation with Compact Word Embedding Tables
From Fully Trained to Fully Random Embeddings: Improving Neural Machine Translation with Compact Word Embedding Tables
Krtin Kumar
Peyman Passban
Mehdi Rezagholizadeh
Yiu Sing Lau
Qun Liu
41
2
0
18 Apr 2021
Sentence Concatenation Approach to Data Augmentation for Neural Machine
  Translation
Sentence Concatenation Approach to Data Augmentation for Neural Machine Translation
Seiichiro Kondo
Kengo Hotate
Masahiro Kaneko
Mamoru Komachi
71
16
0
17 Apr 2021
"Wikily" Supervised Neural Translation Tailored to Cross-Lingual Tasks
"Wikily" Supervised Neural Translation Tailored to Cross-Lingual Tasks
Mohammad Sadegh Rasooli
Chris Callison-Burch
Derry Wijaya
CLIP
59
6
0
16 Apr 2021
AMMU : A Survey of Transformer-based Biomedical Pretrained Language
  Models
AMMU : A Survey of Transformer-based Biomedical Pretrained Language Models
Katikapalli Subramanyam Kalyan
A. Rajasekharan
S. Sangeetha
LM&MAMedIm
115
171
0
16 Apr 2021
Counter-Interference Adapter for Multilingual Machine Translation
Counter-Interference Adapter for Multilingual Machine Translation
Yaoming Zhu
Jiangtao Feng
Chengqi Zhao
Mingxuan Wang
Lei Li
66
62
0
16 Apr 2021
Previous
123...293031...373839
Next