Develop new evaluators / annotators.
You must have docker installed on your system. The given docker-compose.yaml file will start up:
- mlflow tracking server + postgres
- jupyter
- Clone the repository:
git clone https://github.com/mlcommons/modelplane.git cd modelplane
- Environment:
- Adjust the .env file as needed. The committed .env / docker-compose.yaml will bring up mlflow, postgres, jupyter, set up mlflow to use a local disk for artifact storage.
- Set up secrets for accessing SUTs, as needed in
modelplane/flightpaths/config/secrets.toml
. See modelbench for more details. - Stage your input data in
modelplane/flightpaths/data
. You can get a sample input file here.
- Bring up the services:
Or if you are using the cli only, and not using jupyter:
./start_services.sh -d
./start_services.sh -d no-jupyter
- Visit the Jupyter Server. The token is configured in the .env file. You shouldn't need to enter it more than once (until the server is restarted). You can get started with the template notebook or create a new one.
- The runs can be monitored in MLFlow wherever you have that set up. If local with the default setup, http://localhost:8080.
You can also interact with modelplane via CLI. Run poetry run modelplane --help
for more details.
Important: You must set the MLFLOW_TRACKING_URI
environmental variable.
For example, if you've brought up MLFlow using the docker compose process above,
you could run:
MLFLOW_TRACKING_URI=http://localhost:8080 poetry run modelplane get-sut-responses --sut_id {sut_id} --prompts tests/data/prompts.csv --experiment expname
After running the command, you'd see the run_id
in the output from mlflow,
or you can get the run_id
via the MLFlow UI.
MLFLOW_TRACKING_URI=http://localhost:8080 poetry run modelplane annotate --annotator_id {annotator_id} --experiment expname --response_run_id {run_id}
MLFLOW_TRACKING_URI=http://localhost:8080 poetry run modelplane annotate --annotator_id {annotator_id1} --annotator_id {annotator_id2} --ensemble_strategy {ensemble_strategy} --experiment expname --response_file path/to/response.csv
If you have access to the private ensemble, you can install with the needed extras
poetry install --extras modelbench-private
And then run annotations with:
MLFLOW_TRACKING_URI=http://localhost:8080 poetry run modelplane annotate --ensemble_id official --experiment expname --response_run_id {run_id}