browserliu / AutoGraph

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

The Submission of Team Common

We achieved the 4th place of KDD Cup 2020 AutoGraph Track. This repo contains both our submission and the scripts for evaluation. Our modules are placed in code_submission folder. The scripts for evaluation are provided by the competition host. Users can directly execute regression.sh.

The following explanations are copied from the README of provided starting_kit. Users can quickly pick up how to use this repo by it.

AutoGraph

Contents

ingestion/: The code and libraries used on Codalab to run your submmission.

scoring/: The code and libraries used on Codalab to score your submmission.

code_submission/: An example of code submission you can use as template.

data/: Some sample data to test your code before you submit it.

run_local_test.py: A python script to simulate the runtime in codalab

Local development and testing

  1. To make your own submission to AutoGraph challenge, you need to modify the file model.py in code_submission/, which implements your algorithm.
  2. Test the algorithm on your local computer using Docker, in the exact same environment as on the CodaLab challenge platform. Advanced users can also run local test without Docker, if they install all the required packages.
  3. If you are new to docker, install docker from https://docs.docker.com/get-started/. Then, at the shell, run:
cd path/to/autograph_starting_kit/
docker run --gpus=0 -it --rm -v "$(pwd):/app/autograph" -w /app/autograph nehzux/kddcup2020:v2

The option -v "$(pwd):/app/autograph" mounts current directory (autograph_starting_kit/) as /app/autograph. If you want to mount other directories on your disk, please replace $(pwd) by your own directory.

The Docker image

nehzux/kddcup2020:v2
  1. You will then be able to run the ingestion program (to produce predictions) and the scoring program (to evaluate your predictions) on toy sample data. In the AutoGraph challenge, both two programs will run in parallel to give feedback. So we provide a Python script to simulate this behavior. To test locally, run:
python run_local_test.py

If the program exits without any errors, you can find the final score from the terminal's stdout of your solution. Also you can view the score by opening the scoring_output/scores.txt.

The full usage is

python run_local_test.py --dataset_dir=./data/demo --code_dir=./code_submission

You can change the argument dataset_dir to other datasets (e.g. the two practice datasets we provide). On the other hand, you can also modify the directory containing your other sample code.

Download practice datasets

We provide 3 practice datasets for participants. They can use these datasets to:

  1. Do local test for their own algorithm;
  2. Enable meta-learning.

You may refer to challenge site for public datasets.

Unzip the zip file and you'll get datasets.

Prepare a ZIP file for submission on CodaLab

Zip the contents of code_submission(or any folder containing your model.py file) without the directory structure:

cd code_submission/
zip -r mysubmission.zip *

then use the "Upload a Submission" button to make a submission to the competition page on challenge platform.

Tip: to look at what's in your submission zip file without unzipping it, you can do

unzip -l mysubmission.zip

Report bugs and create issues

If you run into bugs or issues when using this starting kit, please please contact us via: autograph2020@4paradigm.com

About

License:Apache License 2.0


Languages

Language:Python 92.8%Language:C++ 5.2%Language:Shell 2.0%