AI researchers have for years warned about model collapse , which is the degeneration of AI models after ingesting AI slop. The process effectively poisons a model with unverifiable information, but it's not to be confused with model poisoning , a serious security threat that Microsoft just published new research about. Also: More workers are using AI than ever they're also trusting it less: Inside the frustration gap While the stakes of model collapse are still...