This is obvious in retrospect - instead of making LLMs work better, LLM's handlers invented various techniques to make LLMs to look like they work better, one such example is summarization. Next gen LLMs then get trained on that data.
Now instead of having some answer right away, the user has to engage in discussion, which increases the cost that is sunk into the work with LLMs.
Now instead of having some answer right away, the user has to engage in discussion, which increases the cost that is sunk into the work with LLMs.