Rogue AI Terrifies Scientists In Chilling Study
As a kind of silver lining, however, this means it’s theoretically easier to reveal poisoned AI prior to deployment.
Three further training methodologies were also involved in the study: reinforcement learning (RL), supervised fine-tuning (SFT), and adversarial training.
For those not versed in teaching nightmarish AI, RL amounts to rewarding desired behavior and, correspondingly, punishing undesirable ones.
While SFT focuses on teaching the rogue AI through a database of correct responses.
👉 For more insights, check out this resource.
👉 Discover more in this in-depth guide.