Tag
1 article tagged "local-llm"
A post-mortem on building a local LLM serving layer — llama.cpp integration, model management, and where existing tools constrain research.