ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2203.10012
  4. Cited By
Report from the NSF Future Directions Workshop on Automatic Evaluation
  of Dialog: Research Directions and Challenges

Report from the NSF Future Directions Workshop on Automatic Evaluation of Dialog: Research Directions and Challenges

18 March 2022
Shikib Mehri
Jinho Choi
L. F. D’Haro
Jan Deriu
M. Eskénazi
Milica Gasic
Kallirroi Georgila
Dilek Z. Hakkani-Tür
Zekang Li
Verena Rieser
Samira Shaikh
David Traum
Yi-Ting Yeh
Zhou Yu
Yizhe Zhang
Chen Zhang
ArXivPDFHTML

Papers citing "Report from the NSF Future Directions Workshop on Automatic Evaluation of Dialog: Research Directions and Challenges"

13 / 63 papers shown
Title
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language
  Understanding
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
761
7,080
0
20 Apr 2018
RankME: Reliable Human Ratings for Natural Language Generation
RankME: Reliable Human Ratings for Natural Language Generation
Jekaterina Novikova
Ondrej Dusek
Verena Rieser
ALM
36
108
0
15 Mar 2018
Sample Efficient Deep Reinforcement Learning for Dialogue Systems with
  Large Action Spaces
Sample Efficient Deep Reinforcement Learning for Dialogue Systems with Large Action Spaces
Gellert Weisz
Paweł Budzianowski
Pei-hao Su
Milica Gasic
33
82
0
11 Feb 2018
Towards an Automatic Turing Test: Learning to Evaluate Dialogue
  Responses
Towards an Automatic Turing Test: Learning to Evaluate Dialogue Responses
Ryan J. Lowe
Michael Noseworthy
Iulian Serban
Nicolas Angelard-Gontier
Yoshua Bengio
Joelle Pineau
51
372
0
23 Aug 2017
Why We Need New Evaluation Metrics for NLG
Why We Need New Evaluation Metrics for NLG
Jekaterina Novikova
Ondrej Dusek
Amanda Cercas Curry
Verena Rieser
69
456
0
21 Jul 2017
Learning Discourse-level Diversity for Neural Dialog Models using
  Conditional Variational Autoencoders
Learning Discourse-level Diversity for Neural Dialog Models using Conditional Variational Autoencoders
Tiancheng Zhao
Ran Zhao
M. Eskénazi
47
754
0
31 Mar 2017
Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks
Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks
Chelsea Finn
Pieter Abbeel
Sergey Levine
OOD
781
11,793
0
09 Mar 2017
RUBER: An Unsupervised Method for Automatic Evaluation of Open-Domain
  Dialog Systems
RUBER: An Unsupervised Method for Automatic Evaluation of Open-Domain Dialog Systems
Chongyang Tao
Lili Mou
Dongyan Zhao
Rui Yan
55
217
0
11 Jan 2017
A Sequence-to-Sequence Model for User Simulation in Spoken Dialogue
  Systems
A Sequence-to-Sequence Model for User Simulation in Spoken Dialogue Systems
Layla El Asri
Jing He
Kaheer Suleman
92
118
0
30 Jun 2016
Learning End-to-End Goal-Oriented Dialog
Learning End-to-End Goal-Oriented Dialog
Antoine Bordes
Y-Lan Boureau
Jason Weston
74
781
0
24 May 2016
On-line Active Reward Learning for Policy Optimisation in Spoken
  Dialogue Systems
On-line Active Reward Learning for Policy Optimisation in Spoken Dialogue Systems
Pei-hao Su
Milica Gasic
N. Mrksic
L. Rojas-Barahona
Stefan Ultes
David Vandyke
Tsung-Hsien Wen
S. Young
OffRL
61
170
0
24 May 2016
How NOT To Evaluate Your Dialogue System: An Empirical Study of
  Unsupervised Evaluation Metrics for Dialogue Response Generation
How NOT To Evaluate Your Dialogue System: An Empirical Study of Unsupervised Evaluation Metrics for Dialogue Response Generation
Chia-Wei Liu
Ryan J. Lowe
Iulian Serban
Michael Noseworthy
Laurent Charlin
Joelle Pineau
94
1,292
0
25 Mar 2016
Individual and Domain Adaptation in Sentence Planning for Dialogue
Individual and Domain Adaptation in Sentence Planning for Dialogue
François Mairesse
R. Prasad
Amanda Stent
M. Walker
59
160
0
31 Oct 2011
Previous
12