MingZhong

6 models • 1 total models in database
Sort by:

unieval-fact

Towards a Unified Multi-Dimensional Evaluator for Text Generation Multi-dimensional evaluation is the dominant paradigm for human evaluation in Natural Language Generation (NLG), i.e., evaluating the generated text from multiple explainable dimensions, such as coherence and fluency. However, automatic evaluation in NLG is still dominated by similarity-based metrics (e.g., ROUGE, BLEU), but they are not sufficient to portray the difference between the advanced generation models. Therefore, we propose UniEval to bridge this gap so that a more comprehensive and fine-grained evaluation of NLG systems can be achieved. unieval-fact is the pre-trained evaluator for the factual consistency detection task. It can evaluate the model output and predict a consistency score.

4,833
3

unieval-dialog

1,347
4

unieval-sum

Towards a Unified Multi-Dimensional Evaluator for Text Generation Multi-dimensional evaluation is the dominant paradigm for human evaluation in Natural Language Generation (NLG), i.e., evaluating the generated text from multiple explainable dimensions, such as coherence and fluency. However, automatic evaluation in NLG is still dominated by similarity-based metrics (e.g., ROUGE, BLEU), but they are not sufficient to portray the difference between the advanced generation models. Therefore, we propose UniEval to bridge this gap so that a more comprehensive and fine-grained evaluation of NLG systems can be achieved. unieval-sum is the pre-trained evaluator for the text summarization task. It can evaluate the model output from four dimensions: - coherence - consistency - fluency - relevance It can also be transferred to the new dimensions and generation tasks, such as naturalness and informativeness for data-to-text.

774
5

DialogLED-base-16384

7
6

DialogLED-large-5120

5
8

unieval-intermediate

4
3