Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2212.04755
Cited By
From Cloze to Comprehension: Retrofitting Pre-trained Masked Language Model to Pre-trained Machine Reader
9 December 2022
Weiwen Xu
Xin Li
Wenxuan Zhang
Meng Zhou
W. Lam
Luo Si
Lidong Bing
Re-assign community
ArXiv
PDF
HTML
Papers citing
"From Cloze to Comprehension: Retrofitting Pre-trained Masked Language Model to Pre-trained Machine Reader"
7 / 7 papers shown
Title
Reasons to Reject? Aligning Language Models with Judgments
Weiwen Xu
Deng Cai
Zhisong Zhang
Wai Lam
Shuming Shi
ALM
21
14
0
22 Dec 2023
Large Language Model Is Not a Good Few-shot Information Extractor, but a Good Reranker for Hard Samples!
Yubo Ma
Yixin Cao
YongChing Hong
Aixin Sun
RALM
80
135
0
15 Mar 2023
Template-free Prompt Tuning for Few-shot NER
Ruotian Ma
Xin Zhou
Tao Gui
Y. Tan
Linyang Li
Qi Zhang
Xuanjing Huang
VLM
148
178
0
28 Sep 2021
Making Pre-trained Language Models Better Few-shot Learners
Tianyu Gao
Adam Fisch
Danqi Chen
241
1,919
0
31 Dec 2020
Exploiting Cloze Questions for Few Shot Text Classification and Natural Language Inference
Timo Schick
Hinrich Schütze
258
1,588
0
21 Jan 2020
Effective Approaches to Attention-based Neural Machine Translation
Thang Luong
Hieu H. Pham
Christopher D. Manning
218
7,926
0
17 Aug 2015
Teaching Machines to Read and Comprehend
Karl Moritz Hermann
Tomás Kociský
Edward Grefenstette
L. Espeholt
W. Kay
Mustafa Suleyman
Phil Blunsom
175
3,510
0
10 Jun 2015
1