- Create a classic personal access token in your github developer settings (it only needs the read packages permission) and paste it to .env
- Add the aws key id and secrret and the bucket name to the .env file
- Run 'make start_mlflow_ui' to init and start the docker container
- Run 'sync_mlflow_ui' to sync your local changes with the database. Deletions of any kind are disabled.
- Access the mlflow ui at localhost:4444
- Run 'make stop_mlflow_ui' to stop the docker container
- Run 'make 'remove_mlflow_ui' to delete the docker container and the image
- Datasets are stored in the folder
datasets
, which is not synced with github, but stored in s3 instead - To download all existing datasets, use
make download_datasets
- To download a specific dataset use
make download_dataset NAME=<dataset_name>
- To upload a new dataset to s3, add it to the
datasets
folder and usemake upload_datasets
- This feature is only for storing raw data. Procecced datasets are stored as artifacts and can be accessed using the mlflow ui
- Your github token must be configured to allow repo and workflow access
- Run
make deploy_to_development
to start the workflow (no protection rules) - Run
make deploy_to_production
to start the workflow (approval required, deployment is only allowed frommain
) - Check the workflow status in github actions
- Open the environment configuration
- Click
New Environment
- Enter a name (lower key preferred)
- Configure potection rules
- Set the environment secrets
AWS_ACCESS_KEY_ID
andAWS_SECRET_ACCESS_KEY
- Set the environment variables
ÀWS_DEFAULT_REGION
andBUCKET_NAME
to configure the S3 connection