Fix test_qwen2.py #23
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
name: cicd-mac | |
on: | |
workflow_dispatch: | |
push: | |
paths: | |
- "scripts/**" | |
- "native/**" | |
- "src/**" | |
- "test/**" | |
- "Makefile" | |
- ".github/trigger.txt" | |
- ".github/workflows/cicd-mac.yml" | |
jobs: | |
all-mac: | |
name: all-mac | |
strategy: | |
matrix: | |
python-version: [3.11] | |
os: [macos-latest] | |
runs-on: ${{ matrix.os }} | |
steps: | |
# https://github.com/marketplace/actions/setup-miniconda | |
- uses: conda-incubator/setup-miniconda@v3 | |
with: | |
python-version: ${{ matrix.python-version }} | |
# ------------------------------------------------------------------- | |
# Checkout llama_cpp_canister & llama_cpp_onicai_fork as nested directory | |
- name: checkout llama_cpp_canister | |
uses: actions/checkout@v4 | |
with: | |
fetch-depth: 0 | |
- name: checkout llama_cpp_onicai_fork | |
uses: actions/checkout@v4 | |
with: | |
repository: onicai/llama_cpp_onicai_fork | |
ref: onicai # Specify the branch name here | |
path: src/llama_cpp_onicai_fork | |
fetch-depth: 1 # Get just the last commit | |
submodules: 'recursive' | |
# ------------------------------------------------------------------- | |
- name: install | |
shell: bash -l {0} # activates the default conda environment ('test') | |
run: | | |
echo "Installing tool chains & dependencies" | |
pwd | |
make summary | |
make install-dfx | |
make install-python | |
make install-homebrew-mac | |
make install-jp-mac | |
- name: versions | |
shell: bash -l {0} | |
run: | | |
echo "icpp --version: $(icpp --version)" | |
echo "clang++ --version: $(clang++ --version)" | |
echo "g++ --version: $(g++ --version)" | |
#echo "wasm2wat version: $(wasm2wat --version)" | |
echo "pip version : $(pip --version)" | |
echo "python version : $(python --version)" | |
echo "jp version : $(jp --version)" | |
#echo "rustc version : $(rustc --version)" | |
echo "dfx version : $(dfx --version)" | |
echo "Ensure conda works properly" | |
conda info | |
which pip | |
which python | |
which icpp | |
- name: install-wasi-sdk | |
shell: bash -l {0} | |
run: | | |
echo "Installing wasi-sdk" | |
icpp install-wasi-sdk | |
- name: install-rust | |
shell: bash -l {0} | |
run: | | |
echo "Installing rust" | |
icpp install-rust | |
- name: build-info-cpp-wasm | |
shell: bash -l {0} | |
run: | | |
make build-info-cpp-wasm | |
- name: all-static | |
shell: bash -l {0} | |
run: | | |
make all-static | |
- name: Download tinyStories model file from HF - storiesICP42Mtok4096 | |
run: wget -nv -O ./models/storiesICP42Mtok4096.gguf "https://huggingface.co/onicai/llama_cpp_canister_models/resolve/main/storiesICP42Mtok4096.gguf?download=true" | |
- name: Create directory for Qwen models | |
run: mkdir -p ./models/Qwen/Qwen2.5-0.5B-Instruct-GGUF | |
- name: Download Qwen model | |
run: wget -nv -O ./models/Qwen/Qwen2.5-0.5B-Instruct-GGUF/qwen2.5-0.5b-instruct-q8_0.gguf https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct-GGUF/resolve/main/qwen2.5-0.5b-instruct-q8_0.gguf?download=true | |
- name: test-llm-native | |
shell: bash -l {0} | |
run: | | |
make test-llm-native | |
- name: test-llm-wasm | |
shell: bash -l {0} | |
run: | | |
make test-llm-wasm |