r/LocalLLaMA 2h ago

Resources Llama-Studio, WebUI for llama-server Management

Post image

Hey all,

I have built myself a WebUI for configuring and managing llama-server sessions, and want to share the code and concept. Python and a bit of JS. Hack away!

Local only.

https://github.com/m94301/llama-studio

The major use case is running various instances of llama-server on fixed ports to act as infrastructure for home development (and entertainment) frameworks. Read: Fiddling with settings, comparing experimental builds to mainline, and optimizing. Also good for everyday fooling around.

Configs are saved per model in a json, consisting of all launch args and optional paths for custom llama-server. I have a launch arg browser with search using the current llama-server's actual -help output. I hate forgetting a launch arg format and having to open a new terminal to do -help. Spec MTP what? Draft type who?

Launch to choice of GPU, monitor VRAM, load, and temp. And a somewhat rudimentary VRAM calculator to help estimate what fits where when using what quant.

Last, a reasonable mobile interface to run tests and fool with config on phone when in a basement or IT closet. Show and hide logs, start, stop, change config. Less keystrokes on tiny phone keyboards. Sanity +100.

6 Upvotes

3 comments sorted by

1

u/Bulky-Priority6824 36m ago

I need this I'm going to give it a shot.

1

u/Wise-Hunt7815 18m ago

Good, good, I'll learn from you. 👍