Open
Description
I am suggesting a new category of Data Hazard called "AI Sourced Data". Suggested symbol : Ouroboros
These would be cases in which the data is scrapped over the internet or any other sources, which turns out to be AI-generated data. These scrapped data will then be used to train more AI models, thereby creating a negative feedback loop making worse and worse trained models.
This can be intentional in some aspects - for example : "Nightshade" - AI Poisoning for protecting Copyrights. But in many cases ,this can be oversight on training or direct malicious intent of sabotaging.
Also such models trained with 'AI sourced data', can further reinforce other data hazards such as existing bias, privacy issues, and more.