This repository is an environment for running experiments with algorithm distillation. Link to work in this area: arxiv.
| Tool | Version |
|---|---|
| Python | Python 3.10 |
| CUDA Toolkit | 12.8 |
| nvcc Compiler | 12.8.61 |
| CUDA Build | cuda_12.8.r12.8/compiler.35404655_0 |
Installing the CUDA toolchain is optional, and I'll leave that for you to explore.
If you only have one Python interpreter installed, you can create a virtual environment with the command:
python -m venv <your_env_name>If you have multiple Python versions, you can select the one you need with the command:
py -3.10 -m venv <your_env_name>Make sure you create the environment in the correct directory.
Windows
<your_env_name>\Scripts\activate.batLinux
source <your_env_name>/bin/activateI'd like to point out that the virtual environment in the total size is about 6 GB. This isn't a warning, just a fact. Be prepared.
pip install -r requirements.txtCreate a directory in advance to store the MLFlow server data.
<mlflow_dir_path>
├── data_local/
└── artifacts/Next, run the command.
mlflow server \
--backend-store-uri "file:///<mlflow_dir_abspath>/data_local" \
--default-artifact-root "file:///<mlflow_dir_abspath>/artifacts" \
--host localhost \
--port 5000Now we have a local server running on port 5000. You can check http://localhost:5000.
Go to the scripts folder. It contains directories with the names of the environments on which the experiments were run. Each directory contains scripts for training and evaluating models (to change hyperparameters, you need to change the values inside the script).
python scripts/<experiment_name>/<experiment_script>.py