There's an economic curse on Large Language Models — the crappiest ones will be the most widely used ones.
The highest-quality models are exponentially more expensive to run, and currently are too slow for instant answers or processing large amounts of data.
Only the older/smaller/cut-down models are cheap enough to run at scale, so the biggest deployments are also the sloppiest ones.