ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2310.16810
21
0

Can GPT models Follow Human Summarization Guidelines? A Study for Targeted Communication Goals

25 October 2023
Yongxin Zhou
F. Ringeval
Franccois Portet
    ELM
    ALM
ArXivPDFHTML
Abstract

This study investigates the ability of GPT models (ChatGPT, GPT-4 and GPT-4o) to generate dialogue summaries that adhere to human guidelines. Our evaluation involved experimenting with various prompts to guide the models in complying with guidelines on two datasets: DialogSum (English social conversations) and DECODA (French call center interactions). Human evaluation, based on summarization guidelines, served as the primary assessment method, complemented by extensive quantitative and qualitative analyses. Our findings reveal a preference for GPT-generated summaries over those from task-specific pre-trained models and reference summaries, highlighting GPT models' ability to follow human guidelines despite occasionally producing longer outputs and exhibiting divergent lexical and structural alignment with references. The discrepancy between ROUGE, BERTScore, and human evaluation underscores the need for more reliable automatic evaluation metrics.

View on arXiv
@article{zhou2025_2310.16810,
  title={ Can GPT models Follow Human Summarization Guidelines? A Study for Targeted Communication Goals },
  author={ Yongxin Zhou and Fabien Ringeval and François Portet },
  journal={arXiv preprint arXiv:2310.16810},
  year={ 2025 }
}
Comments on this paper