Skip to content

This repository contains the data and implementation for the ACL'23 Findings paper: "Verifying Annotation Agreement without Multiple Experts: A Case Study with Gujarati SNACS"

License

Notifications You must be signed in to change notification settings

utahnlp/weak-verifiers

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

42 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

This repository contains the data and implementation for the ACL'23 Findings paper: Verifying Annotation Agreement without Multiple Experts: A Case Study with Gujarati SNACS

Setup

Supported Python Version: 3.6+
To get started with the project, follow the steps mentioned below:

  1. Clone the repository to your local working directory.
foo@bar:~$ git clone https://github.com/utahnlp/weak-verifiers.git
  1. Enter the project directory. Create a new virtual environment and activate it.
foo@bar:~$ cd weak-verifiers
foo@bar:prompts-for-structures$ python -m venv <venv_name>
foo@bar:prompts-for-structures$ source activate <venv_name>/bin/activate
(<venv_name>)foo@bar:weak-verifiers$
  1. Create necessary data and dump folders.
(<venv_name>)foo@bar:weak-verifiers$ mkdir -p models
  1. Install package requirements.
(<venv_name>)foo@bar:weak-verifiers$ pip install -r requirements.txt
  1. Install gurobipy to run DirectProbe. Install the Gurobi Optimzer (https://support.gurobi.com/hc/en-us/articles/4534161999889-How-do-I-install-Gurobi-Optimizer-). You'll need a Gurobi licence to use the optimizer. If you are in academia, you can obtain one at: https://www.gurobi.com/academia/academic-program-and-licenses/
  2. Install DirectProbe and its relevant requirements. The DirectProbe project directory should reside the in the scripts folder.

Gujarati SNACS Dataset

As a part of this paper, we contribute the first semantically annotated dataset in Gujarati which annotates adpositional and case supersenses according to the SNACS schema. Gujarati SNACS contains supersense annotations for all adpositions and case markers present in the freely available Gujarati translation by Dr. Sulbha Natraj of the popular childrens' book Le Petit Prince (The Little Prince) by Antoine de Saint-Exupéry. The translation had to be digitized and is available under data/nanakdo_rajkumar.txt.

The annotated data is available under data/annotated/ where it has been divided into chapter files. Please head on to data/README.md for details regarding the format.

Running CRA Experiments

You can find a short description on how to run CRA experiments in the README.md under the scripts sub-directory.

Running Baseline Models

You can find steps on how to train and evaluate baseline supersense disambiguation models mentioned in the paper in the README.md under the scripts sub-directory.

Citation

@inproceedings{mehta-srikumar-2023-verifying,
    title = "Verifying Annotation Agreement without Multiple Experts: A Case Study with {G}ujarati {SNACS}",
    author = "Mehta, Maitrey  and
      Srikumar, Vivek",
    booktitle = "Findings of the Association for Computational Linguistics: ACL 2023",
    month = jul,
    year = "2023",
    address = "Toronto, Canada",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2023.findings-acl.696",
    pages = "10941--10958",
    abstract = "Good datasets are a foundation of NLP research, and form the basis for training and evaluating models of language use. While creating datasets, the standard practice is to verify the annotation consistency using a committee of human annotators. This norm assumes that multiple annotators are available, which is not the case for highly specialized tasks or low-resource languages. In this paper, we ask: Can we evaluate the quality of a dataset constructed by a single human annotator? To address this question, we propose four weak verifiers to help estimate dataset quality, and outline when each may be employed.We instantiate these strategies for the task of semantic analysis of adpositions in Gujarati, a low-resource language, and show that our weak verifiers concur with a double-annotation study. As an added contribution, we also release the first dataset with semantic annotations in Gujarati along with several model baselines.",
}

Contact

Please open an issue if you find trouble running our codebase.
In case the issue is not acknowledged/addressed within a week, please email at maitrey.mehta@utah.edu

About

This repository contains the data and implementation for the ACL'23 Findings paper: "Verifying Annotation Agreement without Multiple Experts: A Case Study with Gujarati SNACS"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published