Correct me if I'm reading this wrong, but
We are normalizing dataset poisoning and obfuscation as an industry, so that companies will be able to covertly change the LLM's behavior.
More and more people use LLMs instead of algorithmic search engines (consciously or not).
What if someone injects alternate history? No Taiwan? No Holocaust?
What if the open source models that people can afford to build on will be poisoned like that, and so their dictionaries and assistants will spew propaganda?