> LLMs were not developed to be, do not function as, and are not use as data compression utilities.
Again, from a information theoretic view point, this is exactly what they are doing, how they where developed and how they function.
I don't know any serious researcher in ML that would find this claim even remotely controversial. It's really not just "a semantics game", its a part of a foundational understanding of the topic. If you want to understand LLMs from this perspective, a good place to start is with an auto-encoder which does try to learn a standard compression algorithm, the move on to more sophisticated embedding models (found in a lot of recommender systems) which try to learn an additional objective on top of minimizing reconstruction error. You'll then see that Transformers and all other major NN architectures fall out of these basic principles.
> Please, come knocking when a service provider exists that will use LLM's to compactly store your company data.
This is literally what every vectordb company does right now, as well as all "chat with your docs" type startups.