ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1908.05620
  4. Cited By
Visualizing and Understanding the Effectiveness of BERT

Visualizing and Understanding the Effectiveness of BERT

15 August 2019
Y. Hao
Li Dong
Furu Wei
Ke Xu
ArXiv (abs)PDFHTML

Papers citing "Visualizing and Understanding the Effectiveness of BERT"

9 / 59 papers shown
Title
A Survey on Contextual Embeddings
A Survey on Contextual Embeddings
Qi Liu
Matt J. Kusner
Phil Blunsom
276
151
0
16 Mar 2020
DC-BERT: Decoupling Question and Document for Efficient Contextual
  Encoding
DC-BERT: Decoupling Question and Document for Efficient Contextual Encoding
Yuyu Zhang
Ping Nie
Xiubo Geng
Arun Ramamurthy
Le Song
Daxin Jiang
81
61
0
28 Feb 2020
A Primer in BERTology: What we know about how BERT works
A Primer in BERTology: What we know about how BERT works
Anna Rogers
Olga Kovaleva
Anna Rumshisky
OffRL
146
1,511
0
27 Feb 2020
Fixed Encoder Self-Attention Patterns in Transformer-Based Machine
  Translation
Fixed Encoder Self-Attention Patterns in Transformer-Based Machine Translation
Alessandro Raganato
Yves Scherrer
Jörg Tiedemann
102
92
0
24 Feb 2020
Sequential Latent Knowledge Selection for Knowledge-Grounded Dialogue
Sequential Latent Knowledge Selection for Knowledge-Grounded Dialogue
Byeongchang Kim
Jaewoo Ahn
Gunhee Kim
BDL
112
169
0
18 Feb 2020
SBERT-WK: A Sentence Embedding Method by Dissecting BERT-based Word
  Models
SBERT-WK: A Sentence Embedding Method by Dissecting BERT-based Word Models
Bin Wang
C.-C. Jay Kuo
50
156
0
16 Feb 2020
Are Pre-trained Language Models Aware of Phrases? Simple but Strong
  Baselines for Grammar Induction
Are Pre-trained Language Models Aware of Phrases? Simple but Strong Baselines for Grammar Induction
Taeuk Kim
Jihun Choi
Daniel Edmiston
Sang-goo Lee
70
90
0
30 Jan 2020
Emergent Properties of Finetuned Language Representation Models
Emergent Properties of Finetuned Language Representation Models
Alexandre Matton
Luke de Oliveira
SSL
40
1
0
23 Oct 2019
Speech-XLNet: Unsupervised Acoustic Model Pretraining For Self-Attention
  Networks
Speech-XLNet: Unsupervised Acoustic Model Pretraining For Self-Attention Networks
Xingcheng Song
Guangsen Wang
Zhiyong Wu
Yiheng Huang
Jane Polak Scowcroft
Dong Yu
Helen Meng
SSL
101
49
0
23 Oct 2019
Previous
12