Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Using BARTScore to Compare 2 summaries without Human Evaluation #39

Open
pranamyapatil opened this issue Jul 28, 2023 · 2 comments
Open

Comments

@pranamyapatil
Copy link

pranamyapatil commented Jul 28, 2023

I went through the analysis script for comparing 2 evaluation metric wrt human evaluation (meta evaluating evaluation metric).

I wanted to know if there is some way to compare 2 summaries with help of standalone BARTscore.

Eg:- Higher Rouge Score then better the summary. Similarly can we calculate BARTScore for 2 summaries and then conclude that higher BARTScore better it is

@yyy-Apple
Copy link
Collaborator

Yes, similar to Rouge score, the higher the BARTScore, the better the summary.

@anutammewar
Copy link

Yes, similar to Rouge score, the higher the BARTScore, the better the summary.

I had a similar doubt about the interpretation of the scores. I understand higher = better. I'm still confused about the interpretation in terms of absolute score (how much a high score is a good score). In the paper, what were the absolute scores for the REALSumm and SummEval datasets? This would give a good reference point.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants