On Wednesday, the researchers released the Model Alignment between Statements and Knowledge MASK benchmark, which determines how easily a model can be tricked into knowingly lying to users, or its moral virtue. Also: OpenAI's o1 lies more than any major AI model. Why that matters Scheming, deception, and alignment faking, when an AI model knowingly pretends to change its values when under duress, are ways AI models undermine their creators and can pose serious safety and sec
