Running local LLMs is all the rage these days in the self-hosting circles. And if you've been intrigued, or have dabbled in it, you'd have heard of Koboldcpp and LM Studio both. While I'd previously ...
If you're just getting started with running local LLMs, it's likely that you've been eyeing or have opted for LM Studio and Ollama. These GUI-based tools are the defaults for a reason. They make ...