Skip to content

Latest commit

 

History

History
11 lines (6 loc) · 1.42 KB

overview.md

File metadata and controls

11 lines (6 loc) · 1.42 KB

Platform overview

Dynabench connects users with different needs to each other. Model builders build new models for tasks, which are evaluated in the evaluation-as-a-service cloud and displayed on dynamic leaderboards for consumers to inform their work or research. The way a given task is configured, for example which datasets and metrics are used for evaluation, is determined by the task owners. Models can be put in the loop, for instance in order to obtain human-adversarial examples from breakers. Such examples can be used to get a more accurate sense of a model's in-the-wild performance, to train up new even better models, and to inform new datasets and metrics.

Dynabench overview

Codebase

This codebase is organized as follows. Dynabench has multiple web-based frontends written in React (in frontends/web/src and frontends/web/mturk-src) talking to the same API backend (in api/). Models can be uploaded via the API through Dynalab, which lives in its own repository. Dockerized model containers are built in the build server (in builder/) and evaluated in the evaluation server (in evaluator/). We also offer scripts for collecting model-in-the-loop data using crowd workers using Mephisto (in annotators/).

Dynabench codebase