We are a self-hosted platform that allows AI researchers to efficiently share GPUs and schedule the execution of their scripts.
You can use GPqUeue to:
- Schedule a set of experiments across different GPUs.
- Launch training jobs with different hyperparameters.
- Track the progress of your ongoing experiments.
- Enable safe GPU sharing between multiple researchers.
Please note that this project is in an alpha version - currently it supports launching only using Docker Compose
Launch GPqUeue on your own cluster of GPUs using the following command:
./start-stag.sh
This should start a new instance of GPqUeue that is now available at localhost
.
You can then create an account and log in. That will take you to the main overview page, where you can see:
- Currently available GPUs and their status
- List of queued and completed experiments
- An option to launch a new experiment.
GPqUeue support both launching a single and multiple experiments.
To launch multiple experiments, you need to provide a yaml-like configuration file that allows to specify multiple values for some parameters:
The project is based on Python 3.9, as well as ReactJS and Redis.
Thanks a lot to the team at the Imperial College London and Royal College of Art that supervised this project as part of the "Design for Real People" module.