-
Notifications
You must be signed in to change notification settings - Fork 0
55 lines (44 loc) · 1.41 KB
/
push-dataset.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
name: Push Dataset and Finetune
on:
push:
branches:
- "dataset"
jobs:
prepare-dataset:
runs-on: ubuntu-latest
environment: huggingface
steps:
- name: Checkout repository
uses: actions/checkout@v2
- name: Set up Python
uses: actions/setup-python@v2
with:
python-version: "3.10"
- name: Install dependencies for conversion to parquet
run: |
sudo apt-get install python3-pandas
pip install pandas
pip install pyarrow
pip install datasets
pip install huggingface_hub
pip install python-decouple
- name: Add huggingface token to .env
run: |
echo "================="
echo "Add environment variables to .env file"
echo "================="
touch .env
echo "HF_TOKEN=${{ secrets.HF_TOKEN }}" >> .env
- name: Convert json to parquet and save it in huggingface
run: |
python3 parquet_dataset_generator.py
# finetune-ai-model:
# steps:
# - name: Install dependencies for creating Vast.ai instance, Create the vast.ai instance
# run: |
# pip install --upgrade vastai
# - name: Set up API key of Vast.ai
# run: |
# vastai set api-key ${{ secrets.VASTAI_TOKEN }}
# - name: Find the instance according to the requirements
# run: |