Console • Docs • Templates • Discord
This repo contains helpful AI/ML notebook templates. Each notebook has been coupled with the minimum GPU specs required to use them + setup scripts making a Brev template. Click the deploy badge on any notebook to deploy it.
Notebook | Description | Min. GPU | Deploy |
---|---|---|---|
AUTOMATIC1111 Stable Diffusion WebUI | Run Stable Diffusion WebUI, AUTOMATIC1111 | 1x A10G | |
ControlNet on AUTOMATIC1111 | Run ControlNet Models on Stable Diffusion WebUI | 1x A10G | |
SDXL inference with LoRA and Diffusers | Run inference using LoRA adaptors and SDXL | 1x A10G | |
Finetune BioMistral | Finetune the BioMistral model on medical data | 1x A10G | |
Deploy to Replicate | Deploy Model to Replicate | any || CPU | |
Inference on DBRX with VLLM and Gradio | Run inference on DBRX with VLLM and Gradio | 4x A100 | |
Run BioMistral | Run BioMistral | 1x A10G | |
Run Llama 2 70B | Run Llama 2 70B, or any Llama 2 Model | 4x T4 | |
Fine-tune BioMistral | A Guide to Fine-tuning BioMistral | 1x A10G | |
Fine-tune Llama 2 | A Guide to Fine-tuning Llama 2 | 1x A10G | |
Fine-tune Llama 2 - Own Data | Fine-tune Llama 2 on your own dataset | 1x A10G | |
Fine-tune Mistral | A Guide to Fine-tuning Mistral | 1x A10G | |
Fine-tune Mistral using NVIDIA NeMO and PEFT | Fine-tune Mistral using NVIDIA NeMO and PEFT | 1x A100 | |
Fine-tune Mistral - Own Data | Fine-tune Mistral on your own dataset | 1x A10G | |
Fine-tune Mixtral (8x7B MoE) | A Guide to Fine-tuning Mixtral, Mistral's 8x7B MoE | 4x T4 | |
Fine-tune Mixtral (8x7B MoE) - Own Data | A Guide to Fine-tuning Mixtral on your own dataset | 4x T4 | |
Fine-tune Phi-2 | A Guide to Fine-tuning Phi-2 | 1x A10G | |
Fine-tune Phi-2 - Own Data | Fine-tune Phi-2 on your own dataset | 1x A10G | |
GGUF Export FT Model | Export your fine-tuned model to GGUF | 1x A10G | |
Training Question/Answer models using NVIDIA NeMo | Use NeMo to train BERT, GPT, and S2S models for Q&A tasks | 1x A10G | |
Julia Install | Easily Install Julia + Notebooks | any || CPU | |
Oobabooga LLM WebUI | Run Oobabooga, the LLM WebUI (like AUTOMATIC1111) | 1x A10G | |
PDF Chatbot (OCR) | PDF Chatbot using OCR | 1x A10G | |
Use TensorRT-LLM with Mistral | Use NVIDIA TensorRT engine to run inference on Mistral-7B | 1x A10G | |
Zephyr Chatbot | Chatbot with Open Source Models | 1x A10G | |
StreamingLLM for Optimized Inference | Use StreamingLLM for infinite length input without finetuning | 1x A10G |
Brev is a dev tool that makes it really easy to code on a GPU in the cloud. Brev does 3 things: provision, configure, and connect.
Brev provisions a GPU for you. You don't have to worry about setting up a cloud account. We have solid GPU supply, but if you do have AWS or GCP, you can link them.
Brev configures your GPU with the right drivers and libraries. Use our open source tool Verb to point and click the right python and CUDA versions.
Brev.dev CLI automatically edits your ssh config so you can ssh gpu-name
or run brev open gpu-name
to open VS Code to the remote machine