Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It's not even just the cost of finetuning. The API pricing is so low, you literally can't save money by buying a GPU and running your own LLM, no matter how many tokens you generate. It's an incredible moat for OpenAI, but something they can't provide is an LLM that doesn't talk like an annoying HR manager, which is the real use case for self-hosting.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: