Skip to content

shizhl/RADE

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 

Repository files navigation

RADE

code for paper 'Reference-Assisted Dialogue Evaluation for Open-Domain Dialogue (ACL2023)'.

Introduction

Evaluating open-domain dialogue systems is challenging for reasons such as the one-to-many problem, i.e., many appropriate responses other than just the golden response.
As of now, automatic evaluation methods need better consistency with humans, while reliable human evaluation can be time- and cost-intensive. To this end, we propose the Reference-Assisted Dialogue Evaluation (RADE) approach under the multi-task learning framework, which leverages the pre-created utterance as reference other than the gold response to relief the one-to-many problem. Specifically, RADE explicitly compares reference and the candidate response to predict their overall scores. Moreover, an auxiliary response generation task enhances prediction via a shared encoder.

Dataset

We first collect a pre-train dataset based on existing works. To support RADE, we also extend three datasets with additional rated responses other than just a golden response by human annotation.

All the datasets can be found at ./dataset folder.

Contact us

Mail to [email protected]

Citation

@inproceedings{shi2023rade,
  title={RADE: Reference-Assisted Dialogue Evaluation for Open-Domain Dialogue},
  author={Shi, Zhengliang and Sun, Weiwei and Zhang, Shuo and Zhang, Zhen and Ren, Pengjie and Ren, Zhaochun},
  booktitle={Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)},
  pages={12856--12875},
  year={2023}
}

About

dataset for paper Reference-Assisted Dialogue Evaluation for Open-Domain Dialogue.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published