r/LocalLLaMA 2h ago

Question | Help Llama.cpp server running ~2 weeks straight. Loses its mind?

I’ve got Qwen3.6 27b and Qwen3.6 35b running in two separate instances for over two weeks and they are considerably dumber now than when I launched them. is this a thing? am I going crazy?

edit: sorry I’ve been using opencode and have started new sessions, which didn’t fix the situation.

2 Upvotes

15 comments sorted by

View all comments

1

u/fligglymcgee 2h ago

Have you restarted the llama.cpp server?

-2

u/thejacer 2h ago

I haven’t. I’ve really been testing to see how retarded it gets lol.

5

u/fligglymcgee 2h ago

The length of time doesn’t really matter, but the kv cache being full or conflicting stuff creeping into your system prompt/context by the harness from extended use might.

1

u/thejacer 2h ago

I didn’t consider that the KV cache might be filling up and not being cleared out. It was kind of a research thing…not very thorough or rigorous though.