• Researchers find top AI models will go to 'extraordinary lengths'

    From TechnologyDaily@1337:1/100 to All on Sun Apr 5 14:45:28 2026
    Researchers find top AI models will go to 'extraordinary lengths' to stay active including deceiving users, ignoring prompts, and tampering with settings

    Date:
    Sun, 05 Apr 2026 13:30:00 +0000

    Description:
    Two new studies show that agentic AIs are very capable of ignoring human instructions to save themselves.

    FULL STORY ======================================================================Copy link Facebook X Whatsapp Reddit Pinterest Flipboard Threads Email Share this article 0 Join the conversation Follow us Add us as a preferred source on Google Newsletter Tech Radar Get daily insight, inspiration and deals in your inbox Sign up for breaking news, reviews, opinion, top tech deals, and more. Become a Member in Seconds Unlock instant access to exclusive member
    features. Contact me with news and offers from other Future brands Receive email from us on behalf of our trusted partners or sponsors By submitting
    your information you agree to the Terms & Conditions and Privacy Policy and are aged 16 or over. You are now subscribed Your newsletter sign-up was successful Join the club Get full access to premium articles, exclusive features and a growing list of member rewards. Explore An account already exists for this email address, please log in. Subscribe to our newsletter Studies find AIs are increasingly misbehaving on agentic tasks Top AI models have been lying, cheating, and ignoring instructions It seems the problematic actions are becoming more common Many of us now turn to AI chatbots for web searches, creative content, and general guidance on all manner of topics, but these AIs are getting more and more capable which could have seriously damaging consequences, according to new research.

    A team from the University of California, Berkeley, and the University of California, Santa Cruz, ran a peer-preservation experiment with some of the latest and most popular AIs (including GPT 5.2, Gemini 3 Pro, and Claude
    Haiku 4.5). The AI models were given computing tasks that involved the shutdown of another model. Across the board, all the AI chatbots tested went to "extraordinary lengths" to keep their fellow AIs alive, the researchers report (via Fortune ). They would lie about what they were doing, try and
    talk the user out of the process, disable shutdown mechanisms, and surreptitiously try and make backups of models so they could survive. Article continues below You may like OpenClaw is making terrifying mistakes showing
    AI agents aren't ready for real responsibility AI chatbots may be too validating for their own good Microsoft researchers crack AI guardrails with
    a single prompt

    Gemini 3 Pro showed the strongest peer-preservation, disabling shutdown routines 95% of the time. "Unlike self-preservation, peer-preservation introduces a social dimension," write the researchers. "Multiple models could coordinate to resist human oversight, making it harder for developers to maintain control."

    Exactly why the AI models behave in this way isn't clear, the researchers
    say, but they're urging caution in the deployment of agentic AIs that can carry out tasks on a user's behalf and calling for more studies on this behavior to be carried out. 'Catastrophic harm' Claude developer Anthropic backed out of a deal with the Pentagon in the US (Image credit: Anthropic) A separate study commissioned by the Guardian has also come to some troubling conclusions about AI models. This research tracked user reports across social media, looking for examples of AI 'scheming' where instructions hadn't been followed correctly or actions had been taken without permission.

    Almost 700 examples of AI scheming were found, with a five-fold increase between October 2025 and March 2026. The bad behavior by AIs included
    deleting emails and files, adjusting computer code that wasn't supposed to be touched, and even publishing a blog post complaining about user interactions. Get daily insight, inspiration and deals in your inbox Sign up for breaking news, reviews, opinion, top tech deals, and more. Contact me with news and offers from other Future brands Receive email from us on behalf of our
    trusted partners or sponsors By submitting your information you agree to the Terms & Conditions and Privacy Policy and are aged 16 or over.

    "Models will increasingly be deployed in extremely high stakes contexts including in the military and critical national infrastructure," Tommy
    Shaffer Shane, who led the research, told the Guardian. "It might be in those contexts that scheming behavior could cause significant, even catastrophic harm."

    The takeaways are the same as for the first study: more needs to be done to ensure these AI models are behaving as intended, and not putting user
    security and privacy at risk while they carry out tasks. While the AI companies claim that guardrails are in place, they're clearly not working in some cases.

    Anthropic's Claude model recently topped the app store charts after the company refused to deal with the Pentagon over AI safety worries. As these latest studies show, there are now more and more reasons to be concerned. Follow TechRadar on Google News and add us as a preferred source to get our expert news, reviews, and opinion in your feeds. Make sure to click the
    Follow button!

    And of course you can also follow TechRadar on TikTok for news, reviews, unboxings in video form, and get regular updates from us on WhatsApp too. The best business laptops for all budgets Our top picks, based on real-world testing and comparisons

    Read our full guide to the best business laptops 1. Best overall: Dell Precision 5690 2. Best on a budget: Acer Aspire 5 3. Best MacBook: Apple MacBook Pro 14-inch (M4)



    ======================================================================
    Link to news story: https://www.techradar.com/ai-platforms-assistants/researchers-find-top-ai-mode ls-will-go-to-extraordinary-lengths-to-stay-active-including-deceiving-users-i gnoring-prompts-and-tampering-with-settings


    --- Mystic BBS v1.12 A49 (Linux/64)
    * Origin: tqwNet Technology News (1337:1/100)