ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2108.11193
  4. Cited By
Models In a Spelling Bee: Language Models Implicitly Learn the Character
  Composition of Tokens

Models In a Spelling Bee: Language Models Implicitly Learn the Character Composition of Tokens

25 August 2021
Itay Itzhak
Omer Levy
ArXivPDFHTML

Papers citing "Models In a Spelling Bee: Language Models Implicitly Learn the Character Composition of Tokens"

12 / 12 papers shown
Title
EXECUTE: A Multilingual Benchmark for LLM Token Understanding
EXECUTE: A Multilingual Benchmark for LLM Token Understanding
Lukas Edman
Helmut Schmid
Alexander Fraser
33
0
0
23 May 2025
Enhancing Character-Level Understanding in LLMs through Token Internal Structure Learning
Enhancing Character-Level Understanding in LLMs through Token Internal Structure Learning
Zhu Xu
Zhiqiang Zhao
Zihan Zhang
Yuchi Liu
Quanwei Shen
Fei Liu
Yu Kuang
Jian He
Conglin Liu
127
2
0
26 Nov 2024
How to Train BERT with an Academic Budget
How to Train BERT with an Academic Budget
Peter Izsak
Moshe Berchansky
Omer Levy
87
117
0
15 Apr 2021
CharacterBERT: Reconciling ELMo and BERT for Word-Level Open-Vocabulary
  Representations From Characters
CharacterBERT: Reconciling ELMo and BERT for Word-Level Open-Vocabulary Representations From Characters
Hicham El Boukkouri
Olivier Ferret
Thomas Lavergne
Hiroshi Noji
Pierre Zweigenbaum
Junichi Tsujii
101
161
0
20 Oct 2020
AraBERT: Transformer-based Model for Arabic Language Understanding
AraBERT: Transformer-based Model for Arabic Language Understanding
Wissam Antoun
Fady Baly
Hazem M. Hajj
105
969
0
28 Feb 2020
RoBERTa: A Robustly Optimized BERT Pretraining Approach
RoBERTa: A Robustly Optimized BERT Pretraining Approach
Yinhan Liu
Myle Ott
Naman Goyal
Jingfei Du
Mandar Joshi
Danqi Chen
Omer Levy
M. Lewis
Luke Zettlemoyer
Veselin Stoyanov
AIMat
639
24,431
0
26 Jul 2019
fairseq: A Fast, Extensible Toolkit for Sequence Modeling
fairseq: A Fast, Extensible Toolkit for Sequence Modeling
Myle Ott
Sergey Edunov
Alexei Baevski
Angela Fan
Sam Gross
Nathan Ng
David Grangier
Michael Auli
VLM
FaML
105
3,149
0
01 Apr 2019
Deep contextualized word representations
Deep contextualized word representations
Matthew E. Peters
Mark Neumann
Mohit Iyyer
Matt Gardner
Christopher Clark
Kenton Lee
Luke Zettlemoyer
NAI
206
11,549
0
15 Feb 2018
Neural Machine Translation of Rare Words with Subword Units
Neural Machine Translation of Rare Words with Subword Units
Rico Sennrich
Barry Haddow
Alexandra Birch
215
7,735
0
31 Aug 2015
Adam: A Method for Stochastic Optimization
Adam: A Method for Stochastic Optimization
Diederik P. Kingma
Jimmy Ba
ODL
1.8K
150,039
0
22 Dec 2014
Distributed Representations of Words and Phrases and their
  Compositionality
Distributed Representations of Words and Phrases and their Compositionality
Tomas Mikolov
Ilya Sutskever
Kai Chen
G. Corrado
J. Dean
NAI
OCL
394
33,521
0
16 Oct 2013
Efficient Estimation of Word Representations in Vector Space
Efficient Estimation of Word Representations in Vector Space
Tomas Mikolov
Kai Chen
G. Corrado
J. Dean
3DV
674
31,489
0
16 Jan 2013
1