Was thinking the same honestly. GH is very sticky though, especially when you have actions and all kinds of other integrations set up. But it’s just kind of absurd at this point how many outages they have.
Are there benchmarks of this / what’s the best way to compare it against paid models? With all the rate limiting in Claude/Copilot/etc, running locally is more and more appealing.
Llama cpp is vastly superior. There was this huge bug that prevented me from using a model in ollama and it took them four months for a “vendor sync” (what they call it) which was just updating ggml which is the underpinning library used by llama cpp (same org makes both). lmstudio/lms is essentially Ollama but with llama cpp as backend. I recommend trying lmstudio since it’s the lowest friction to start
My personal Claude sub (Pro), I can burn through my limit in a couple of hours when using Opus. It's borderline unusable unless you're willing to pay for extended usage or artificially slow yourself down.
The problem is half the time you don't know you need the better model until the lesser model has made a massive mess. Then you have to do it again on the good model, wasting money. The "auto" modes don't seem to do a good job at picking a model IME.
Agreed. What network are they forgetting here? Pretty sure I've only ever seen that apply to wifi. Are they trying to say you might need to re-pair a bluetooth speaker on a date? I don't get it.
“Like carfax but for your home” is a really interesting idea. So many homes are bought with little-to-no history beyond an inspection of questionable thoroughness.
If this became the norm, somehow, it would be a really helpful tool for both buyers and sellers.
reply