We achieved the 4th place of KDD Cup 2020 AutoGraph Track. Team members:
Zhen Wang, Daoyuan Chen, Haokun Chen, Yuexiang Xie, Yaliang Li, Bolin Ding, Wei Lin, Jingren Zhou
Alibaba Group
This repo contains both our submission and the scripts for evaluation. Our modules are placed in code_submission folder. The scripts for evaluation are provided by the competition host. Users can directly execute regression.sh.
The following explanations are copied from the README of provided starting_kit. Users can quickly pick up how to use this repo by it.
ingestion/: The code and libraries used on Codalab to run your submmission.
scoring/: The code and libraries used on Codalab to score your submmission.
code_submission/: An example of code submission you can use as template.
data/: Some sample data to test your code before you submit it.
run_local_test.py: A python script to simulate the runtime in codalab
- To make your own submission to AutoGraph challenge, you need to modify the
file
model.py
incode_submission/
, which implements your algorithm. - Test the algorithm on your local computer using Docker, in the exact same environment as on the CodaLab challenge platform. Advanced users can also run local test without Docker, if they install all the required packages.
- If you are new to docker, install docker from https://docs.docker.com/get-started/. Then, at the shell, run:
cd path/to/autograph_starting_kit/
docker run --gpus=0 -it --rm -v "$(pwd):/app/autograph" -w /app/autograph nehzux/kddcup2020:v2
The option -v "$(pwd):/app/autograph"
mounts current directory
(autograph_starting_kit/
) as /app/autograph
. If you want to mount other
directories on your disk, please replace $(pwd)
by your own directory.
The Docker image
nehzux/kddcup2020:v2
- You will then be able to run the
ingestion program
(to produce predictions) and thescoring program
(to evaluate your predictions) on toy sample data. In the AutoGraph challenge, both two programs will run in parallel to give feedback. So we provide a Python script to simulate this behavior. To test locally, run:
python run_local_test.py
If the program exits without any errors, you can find the final score from the terminal's stdout of your solution.
Also you can view the score by opening the scoring_output/scores.txt
.
The full usage is
python run_local_test.py --dataset_dir=./data/demo --code_dir=./code_submission
You can change the argument dataset_dir
to other datasets (e.g. the two
practice datasets we provide). On the other hand, you can also modify the directory containing your other sample code.
We provide 3 practice datasets for participants. They can use these datasets to:
- Do local test for their own algorithm;
- Enable meta-learning.
You may refer to challenge site for public datasets.
Unzip the zip file and you'll get datasets.
Zip the contents of code_submission
(or any folder containing
your model.py
file) without the directory structure:
cd code_submission/
zip -r mysubmission.zip *
then use the "Upload a Submission" button to make a submission to the competition page on challenge platform.
Tip: to look at what's in your submission zip file without unzipping it, you can do
unzip -l mysubmission.zip
If you run into bugs or issues when using this starting kit, please please contact us via: autograph2020@4paradigm.com