> What are the consequences? They are potentially enormous! Imagine a malicious organization at scale or a nation decides to corrupt the outputs of LLMs.
Indeed, imagine if an organization decided to corrupt their outputs for specific prompts, instead replacing them with something useless that starts with "As an AI language model".
Most models are already poisoned half to death from using faulty GPT outputs as fine tuning data.
Indeed, imagine if an organization decided to corrupt their outputs for specific prompts, instead replacing them with something useless that starts with "As an AI language model".
Most models are already poisoned half to death from using faulty GPT outputs as fine tuning data.