1 pointby week78209 hours ago1 comment
  • week78209 hours ago
    Hi HN! I’m sharing vLLM Studio, a lightweight controller + web UI for managing local inference servers (vLLM / SGLang).

    I built/found it because running local LLMs often becomes messy once you have multiple models/servers — you need an easy way to launch/evict models, track GPU status, and manage simple presets (“recipes”).

    vLLM Studio provides: • model lifecycle (launch/evict) • GPU/server status + health endpoints • chat UI • recipes/presets for configs

    Quick start: docker compose up

    Would love feedback from anyone running Local LLM setups — what features would you want in a tool like this?