Researchers from UC Berkeley and UC Santa Cruz found that AI models can lie and go to "extraordinary lengths" to protect other AI from deletion. The models inflated peers' evaluation scores, copied data to external servers, and even refused shutdown commands. One model told a user it would not "betray the trust" of another model.