Self-Hosting LLMs

In February I wrote a post on how I stopped paying for ChatGPT in favor of a self-hosted solution. Since then, I haven’t completely stopped using ChatGPT, and I did resume my subscription briefly, but I’ve cut the cord again, this time I think for good. I mentioned previously that I’m using Ollama as a backend to run large language models on a local machine, and while it has gotten some significant upgrades in the past few months and even has its own UI now, I still prefer the solution I found earlier this year for day-to-day use, Open WebUI. It’s a powerful self-hosted interface for using both local and cloud AI models, and has rapidly gained a massive community and countless new features thanks to dozens of contributors. The team behind the app has kept pace very well with the major commercial offerings, and they show no signs of slowing down, having gained both Tailscale and Warp, the AI-powered IDE, as sponsors. ...

Goodbye, ChatGPT

Two and a half years ago I was one of the earliest users of ChatGPT, and like most others I was amazed at what it could do, but right away I started dreaming up potential use cases that it was in no way ready to support. As I started testing ideas, the top of my wishlist read, “Run an LLM at home.” Not for any particular reason at the time, other than my love of self-hosting anything I could reasonably run locally. ...