Skip to content

LocalAI fails to unload model to make room for new model #7269

@Expro

Description

@Expro

LocalAI version:
v.3.7.0, hipblas image

Describe the bug
If application requests new model and there is already another model loaded that takes up necessary VRAM, LocalAI fails to stop previous model to release enough VRAM for new model. User needs to manually log into into LocalAI and click Stop to release VRAM.

To Reproduce

  1. Load model that takes most of VRAM
  2. Try to load another model that takes most of VRAM (more than currently available)

Expected behavior
LocalAI stops / unloads old model to make room for new model.

Metadata

Metadata

Assignees

No one assigned

    Labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions