The researchers are employing a way named adversarial training to prevent ChatGPT from allowing consumers trick it into behaving badly (referred to as jailbreaking). This work pits multiple chatbots towards one another: a person chatbot plays the adversary and attacks A different chatbot by building textual content to drive it to buck its standard