Skip to content

a benchmark suite for testing logical reasoning abilities of prompt-based models

Notifications You must be signed in to change notification settings

csitfun/LogiEval

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

33 Commits
 
 
 
 
 
 

Repository files navigation

===This repository is now MERGED with our GLoRE benchmark.===

LogiEval: a benchmark suite for testing logical reasoning abilities of instruct-prompt large language models

Please refer to our paper "Evaluating the Logical Reasoning Ability of ChatGPT and GPT-4" if you are interested in this test suite.

This repository is based on the OpenAI Eval library. Please download the Eval package first, and put the contents in this repository Data and evals into evals/evals/registry/data/<name_of_your_eval/ and evals/evals/registry/evals/, respectively.

eg. evals/evals/registry/data/logiqa/logiqa.jsonl, evals/evals/registry/evals/logiqa.yaml

Set up

pip install evals

Eval OpenAI models

  1. export openai api key to the environment

export OPENAI_API_KEY=<your_key>

  1. run eval

oaieval <model_name> <data_name>

eg. oaieval gpt-3.5-turbo logiqa

How to cite

@misc{liu2023evaluating,
      title={Evaluating the Logical Reasoning Ability of ChatGPT and GPT-4}, 
      author={Hanmeng Liu and Ruoxi Ning and Zhiyang Teng and Jian Liu and Qiji Zhou and Yue Zhang},
      year={2023},
      eprint={2304.03439},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

About

a benchmark suite for testing logical reasoning abilities of prompt-based models

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published