Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1909.10122
Cited By
Towards Best Experiment Design for Evaluating Dialogue System Output
23 September 2019
Sashank Santhanam
Samira Shaikh
ALM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Towards Best Experiment Design for Evaluating Dialogue System Output"
7 / 7 papers shown
Title
On the Effectiveness of Automated Metrics for Text Generation Systems
Pius von Daniken
Jan Deriu
Don Tuggener
Mark Cieliebak
21
3
0
24 Oct 2022
Achieving Reliable Human Assessment of Open-Domain Dialogue Systems
Tianbo Ji
Yvette Graham
Gareth J. F. Jones
Chenyang Lyu
Qun Liu
ALM
33
39
0
11 Mar 2022
Dynamic Human Evaluation for Relative Model Comparisons
Thórhildur Thorleiksdóttir
Cédric Renggli
Nora Hollenstein
Ce Zhang
36
2
0
15 Dec 2021
Investigating Crowdsourcing Protocols for Evaluating the Factual Consistency of Summaries
Xiangru Tang
Alexander R. Fabbri
Haoran Li
Ziming Mao
Griffin Adams
Borui Wang
Asli Celikyilmaz
Yashar Mehdad
Dragomir R. Radev
HILM
13
19
0
19 Sep 2021
All That's 'Human' Is Not Gold: Evaluating Human Evaluation of Generated Text
Elizabeth Clark
Tal August
Sofia Serrano
Nikita Haduong
Suchin Gururangan
Noah A. Smith
DeLMO
45
394
0
30 Jun 2021
Local Knowledge Powered Conversational Agents
Sashank Santhanam
Ming-Yu Liu
Raul Puri
M. Shoeybi
M. Patwary
Bryan Catanzaro
29
4
0
20 Oct 2020
Deep Reinforcement Learning for Dialogue Generation
Jiwei Li
Will Monroe
Alan Ritter
Michel Galley
Jianfeng Gao
Dan Jurafsky
214
1,327
0
05 Jun 2016
1