Skip to content

atamanovskii/can-ai-code

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Can AI Code?

A cute robot working on a laptop

A self-evaluating interview for AI coding models.

Key Ideas

  • Interview questions written by humans, test taken by AI
  • Inference scripts for all common API providers and CUDA-enabled quantization runtimes
  • Sandbox enviroment (Docker-based) for untrusted Python and NodeJS code validation
  • Evaluate effects of prompting techniques and sampling parameters on LLM coding performance
  • Evaluate LLM coding performance degradation due to quantization

News

2/01 Evaluate cognitivecomputations/dolphin-2.6-mistral-7b-dpo-laser.

1/23 Evaluate mlabonne/Beyonder-4x7B-v2 (AWQ only, FP16 was mega slow).

1/22 Re-evaluate budecosystem/code-millenials family (1B, 3B, 13B, 34B) with corrected prompts.

1/12 Evaluate deepseek-ai/deepseek-moe-16b-chat and some older WizardCoder models.

1/8 Evaluate bagel-34b-v0.2, LLaMA-Pro-8B-Instruct, SqueezeLLM quant of Mistral-Instruct and fix minor tag bug in the leaderboard app. Evaluated new GGUF quants of Mixtral including the 2.1bit quip-sharp inspired one, it does quite well on junior-v2 but falls apart on senior compared to q2k.

1/6 Fixed a number of bugs in the evaluation and extraction scripts, re-run all evaluations (scores generally go up a little bit). Introduce the senior interview, still a work in progress trying to tune the difficulty level.

Test Suites

junior-v2 is a multi-language (Python, JavaScript) suite of 12 tests created for this project to test small LLM coding performance. This project provides all necessary components to execute this evaluation.

🚧 humaneval is a Python-only suite of 164 tests created by OpenAI. This project provides template scripts to prepare and execute the humaneval interview, as well as result extraction scripts to help their evaluator. See https://github.com/openai/human-eval for more information.

Results data

All model answers and evaluation results are now included inside this repository! Install a recent release of streamlit pip install streamlit==1.23 then streamlit run app.py or streamlit run compare-app.py to run the above webapps locally.

Results HumanEval

🚧 humaneval/ development work is currently paused, there's other projects that are much further along.

See https://github.com/my-other-github-account/llm-humaneval-benchmarks and https://github.com/abacaj/code-eval for large lists of Humaneval LLM benchmark results.

Repository Structure

Interviews

  • junior-v2/*.yaml - junior coder interview questions (stable)
  • senior/*.yaml - senior coder interview questions (WIP)

Prepare

  • prompts/*.txt - LLM prompt templates for the various models
  • prepare.py - Applies templates to question turning them into language- and model-specific prompts suitable for interview

Prompts

See prompts/ for all prompts references in the leaderboard.

Interview

  • params/*.json - Sampling hyper-parameter sets (used by all interview scripts)
  • interview-*.py - Interview scripts

Parameters

See params/ for all params references in the leaderboard.

Evaluate

Compare

Interviewers: API

API Runtime Script
LiteLLM (OpenAI, etc..) interview-litellm.py
OobaBooga/KoboldCpp interview-oobabooga.py
Huggingface Inference interview-hfinference.py
Gradio (HF Spaces) interview-gradio.py

Interviewers: CUDA (Local)

Quantization Type Script Dependency
GGUF interview-llamacpp.py llamacpp or ggml binary
GPTQ (AutoGptQ) interview-cuda.py auto-gptq==0.6.0
GPTQ (ExLlama) interview-cuda.py exllama @ 3b013cd53c7d413cf99ca04c7c28dd5c95117c0d
EXL2, GPTQ (ExLlama2) interview-cuda.py exllamav2 @ 0.0.12
HQQ interview-cuda.py hqq @ 0.1.1
AWQ, FP16 (vLLM) interview-cuda.py vllm==0.3.0
CTranslate2 interview-cuda.py ctranslate2>=3.16.0
bitsandbytes interview-cuda.py bitsandbytes==0.41.3
FP16 (Transformers) interview-cuda.py transformers==4.37.2

Running on Modal

The recommended modal wrapper is interview_modal_cuda11.py which builds a CUDA11.8 based container with all the above dependencies working. An interview_modal_cuda12.py is also provided, but AutoGPTQ and CTranslate2 are not compatible.

Unfortunately the nature of Modal does not allow command-line selection of eitehr LLM model or runtime engine.

To select models, open the script and uncomment the .run_function(download...) line of choice. Note that only one model can be selected at a time. To add a new model, implement a new download... function.

To select runtime, open the script and uncomment one of the RUNTIME options. Note that for transformers you must also specify QUANT.

Question Format

A set of interview questions is a folder of .yaml files. Each Question is a top-level key:

SanityList:
    Signature: "things()"
    Input: "with no inputs"
    Output: "a list with three values: the number 5, the string 'foobar', the capital city of Spain"
    Fact: "the capital city of Spain is Madrid"
    Description: "List function, see if the model can combine input facts with internal knowledge."
    Checks:
        input_name:
            assert: "f.name"
            eq: "things"

In this example SanityList is the name of the interview question.

The first four fields are used by prepare.py to create the interview:

  • Signature is the desired function signature
  • Input describes the function inputs
  • Output describes the function outputs
  • Fact is optional and provides any context that is required to correctly perform the task

These 4 variables along with language (either python or javascript) are used to expand templates in prompts/.

The last two fields are used by evaluate.py to judge the results:

  • Description is a human-readable explanation of why this test is useful
  • Checks defines the expected behavior of the output.

Checks and the 'f' object

Each check has a name, some assert value (python code) and an expected eq value.

The f object represents the sandbox view of the function. Static analysis is performed on the function signature to extract the f.name and f.args fields, while f.call allows for function evaluation.

Output formats

All scripts output automatically named .ndjson files to the results/ directory.

Each stage outputs a super-set of fields from the stage before it, so its possible to feed eval/interview back to interview (to re-run the questions) or back to eval (to re-run the eval).

prepare

results/prepare_{interview}_{languages}_{template}.ndjson

Fields:

  • all Question fields (Signature, Input, Output, Fact, Description)
  • name
  • language
  • prompt

interview

results/interview_{interview}_{languages}_{template}_{templateout}_{params}_{model}_{timestamp}.ndjson

Fields:

  • all prepare fields
  • model
  • params
  • answer
  • runtime

eval

results/eval_{interview}_{languages}_{template}_{templateout}_{params}_{model}_{timestamp}.ndjson

Fields:

  • all eval fields
  • status
  • passed
  • total
  • checks

Roadmap / Future Work

Releases

No releases published

Packages

No packages published

Languages

  • Python 98.7%
  • Other 1.3%