This is why the bubble is deflating. “AI” boosters will cite this as an example of an LLM that doesn’t run at a loss but then, when you point out the bad answers from cheap models, they cite much more expensive models & usage patterns as being more accurate. They’re stuck with a catch 22.
add a skeleton here at some point
4 months ago