Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implement the QASPER evaluation #184

Closed
leogao2 opened this issue May 27, 2021 · 1 comment · Fixed by #264
Closed

Implement the QASPER evaluation #184

leogao2 opened this issue May 27, 2021 · 1 comment · Fixed by #264
Labels
feature request A feature that isn't implemented yet.

Comments

@leogao2
Copy link
Contributor

leogao2 commented May 27, 2021

Readers of academic research papers often read with the goal of answering specific questions. Question Answering systems that can answer those questions can make consumption of the content much more efficient. However, building such tools requires data that reflect the difficulty of the task arising from complex reasoning about claims made in multiple parts of a paper. In contrast, existing information-seeking question answering datasets usually contain questions about generic factoid-type information. We therefore present Qasper, a dataset of 5049 questions over 1585 Natural Language Processing papers. Each question is written by an NLP practitioner who read only the title and abstract of the corresponding paper, and the question seeks information present in the full text. The questions are then answered by a separate set of NLP practitioners who also provide supporting evidence to answers. We find that existing models that do well on other QA tasks do not perform well on answering these questions, underperforming humans by at least 27 F1 points when answering them from entire papers, motivating further research in document-grounded, information-seeking QA, which our dataset is designed to facilitate.

https://allenai.org/data/qasper

@leogao2 leogao2 added this to To Do, Other in Implementing Evaluations May 27, 2021
@leogao2 leogao2 moved this from To Do, Other to To do, Evaluations to Implement in Implementing Evaluations May 27, 2021
@StellaAthena StellaAthena added the feature request A feature that isn't implemented yet. label Dec 31, 2021
@StephenHogg
Copy link

Hi, can I take this one?

Implementing Evaluations automation moved this from To do, Evaluations to Implement to Done, evaluations Feb 22, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request A feature that isn't implemented yet.
Projects
No open projects
Implementing Evaluations
  
Done, evaluations
Development

Successfully merging a pull request may close this issue.

3 participants