- Create a classic personal access token in your github developer settings (it only needs the read packages permission) and paste it to .env
- Add the aws key id and secrret and the bucket name to the .env file
- Run
make setup_mlflow_ui
to init and start the docker container - Run
sync_mlflow_ui
to sync your local changes with the database, deletions of any kind are disabled - Access the mlflow ui at port 4444
- Run
make stop_mlflow_ui
to stop the docker container - Run
make start_mlflow_ui
to start the container again - Run make
remove_mlflow_ui
to delete the docker container and the image
- Datasets are stored in the folder
datasets
, which is not synced with github, but stored in s3 instead - To download all existing datasets, use
make download_datasets
- To download a specific dataset use
make download_dataset NAME=<dataset_name>
- To upload a new dataset to s3, add it to the
datasets
folder and usemake upload_datasets
- This feature is only for storing raw data. Procecced datasets are stored as artifacts and can be accessed using the mlflow ui
- Your github token must be configured to allow repo and workflow access
- Run
make start_dev_deployment_workflow
to start the workflow (no protection rules) - Run
make start_prod_deployment_workflow
to start the workflow (approval required, deployment is only allowed frommain
) - Check the workflow status in github actions
- Open the environment configuration
- Click
New Environment
- Enter a name (lower key preferred)
- Configure potection rules
- Set the environment secrets
AWS_ACCESS_KEY_ID
andAWS_SECRET_ACCESS_KEY
- Set the environment variables
ÀWS_DEFAULT_REGION
andBUCKET_NAME
to configure the S3 connection
- Save more artifacts
- Advanced model training code with multiple parameters
- Deploy to an automated testing environment and run tests there
- Deploy to target system
- Export the model
- Run tests with the compiles model and the implementation code