Skip to content

Latest commit

 

History

History
27 lines (17 loc) · 1.36 KB

README.md

File metadata and controls

27 lines (17 loc) · 1.36 KB

Scifact-Open

This repository contains data and analysis code for the paper SciFact-Open: Towards open-domain scientific claim verification.

Table of contents

Setup

First, create a Conda environment:

conda create --name scifact-open python=3.8.5   # Create Conda env.
conda activate scifact-open
pip install -r requirements.txt                 # Install dependencies

Data

To get the data, run bash script/get_data.sh from the root of this directory. The script will download the data and populate two folders: data contains the annotations for SciFact-Open, and prediction contains model predictions for all models used for dataset creation and evaluation. More details on specific files in data.md and prediction.md.

Reproducing paper metrics

To reproduce the main results reported in Table 5 in the paper, run python script/run_eval.py. The script will evaluate the predictions from prediction/model_predictions.parqet against the data in data/claims.jsonl, and print out evaluation metrics. Note that some metrics will differ slightly (0.1 F1 or so) from the results in Table 5, which reports evaluation metrics averaged over 1,000 bootstrap-resampled versions of the dataset.