I built/found it because running local LLMs often becomes messy once you have multiple models/servers — you need an easy way to launch/evict models, track GPU status, and manage simple presets (“recipes”).
vLLM Studio provides: • model lifecycle (launch/evict) • GPU/server status + health endpoints • chat UI • recipes/presets for configs
Quick start: docker compose up
Would love feedback from anyone running Local LLM setups — what features would you want in a tool like this?