Skip to content

dwadden/scifact-open

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Scifact-Open

This repository contains data and analysis code for the paper SciFact-Open: Towards open-domain scientific claim verification.

Table of contents

Setup

First, create a Conda environment:

conda create --name scifact-open python=3.8.5   # Create Conda env.
conda activate scifact-open
pip install -r requirements.txt                 # Install dependencies

Data

To get the data, run bash script/get_data.sh from the root of this directory. The script will download the data and populate two folders: data contains the annotations for SciFact-Open, and prediction contains model predictions for all models used for dataset creation and evaluation. More details on specific files in data.md and prediction.md.

Reproducing paper metrics

To reproduce the main results reported in Table 5 in the paper, run python script/run_eval.py. The script will evaluate the predictions from prediction/model_predictions.parqet against the data in data/claims.jsonl, and print out evaluation metrics. Note that some metrics will differ slightly (0.1 F1 or so) from the results in Table 5, which reports evaluation metrics averaged over 1,000 bootstrap-resampled versions of the dataset.

About

Data and code for the SciFact-Open task

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published