Self-Hosting LLMs

In February I wrote a post on how I stopped paying for ChatGPT in favor of a self-hosted solution. Since then, I haven’t completely stopped using ChatGPT, and I did resume my subscription briefly, but I’ve cut the cord again, this time I think for good. I mentioned previously that I’m using Ollama as a backend to run large language models on a local machine, and while it has gotten some significant upgrades in the past few months and even has its own UI now, I still prefer the solution I found earlier this year for day-to-day use, Open WebUI. It’s a powerful self-hosted interface for using both local and cloud AI models, and has rapidly gained a massive community and countless new features thanks to dozens of contributors. The team behind the app has kept pace very well with the major commercial offerings, and they show no signs of slowing down, having gained both Tailscale and Warp, the AI-powered IDE, as sponsors. ...