Loading

All posts tagged in AI Safety

  • svg
    Post Image

    After facing criticism over its open and unpredictable AI tools, Quill is taking a different approach. The startup is emphasizing security and user control while building what it calls an “agentic AI” that helps users manage their workflows. This new direction aims to balance powerful AI capabilities with privacy and safety, setting it apart from

  • svg
    Post Image

    The US Department of Defense (DoD) and Anthropic are at odds over how AI should be used by the military. The disagreement has escalated to the point where Defense Secretary Pete Hegseth issued a stark warning: work with us on our terms or face being cut out from Pentagon programs. This tense standoff was highlighted

  • svg
    Post Image

    Obsidian Security, a leader in SaaS security, has achieved a major milestone by earning the ISO/IEC 42001:2023 certification. This makes it the first company worldwide to meet the new international standard for an Artificial Intelligence Management System (AIMS). The certification highlights Obsidian’s dedication to responsible AI development and strong governance practices. What the Certification Means

  • svg
    Post Image

    Anthropic has accused three Chinese AI companies of running massive campaigns to illegally extract capabilities from its Claude AI model. The company claims these campaigns involved millions of interactions and used deceptive tactics to bypass restrictions. This controversy highlights ongoing issues around AI model security and data use. How the Campaigns Worked Anthropic says the

  • svg
    Post Image

    A dangerous new malware campaign has emerged in the software world, targeting developers, CI pipelines, and AI coding platforms. This attack uses a large-scale supply chain worm that spreads through malicious npm packages. Researchers have identified it as a serious threat, calling it SANDWORM_MODE, named after environment variables embedded in its code. The malware is

  • svg
    Post Image

    Anthropic, the AI company behind the Claude language model, has come forward with serious accusations. It claims that three Chinese AI developers—DeepSeek, Moonshot, and MiniMax—have been running large-scale campaigns to steal capabilities from Claude. These campaigns involved using deceptive tactics to extract knowledge from the model without permission. The incident raises questions about how AI

  • svg
    Post Image

    In December, AWS experienced a significant outage that lasted about 13 hours. For the first time, the company confirmed that one of its AI systems was responsible for deleting and recreating a key environment. This incident has sparked questions about how we interact with AI systems and the risks involved when trusting them blindly. The

  • svg
    Post Image

    Anthropic has revealed details about aggressive efforts by foreign labs to steal capabilities from its AI system, Claude. These campaigns involve large-scale efforts to extract proprietary knowledge using deceptive tactics. Such activities pose serious threats to intellectual property and national security, especially as AI models become more powerful and widespread. Massive Campaigns to Copy AI

  • svg
    Post Image

    A recent security scare has caused concern among developers worldwide. A popular npm package was secretly modified to install a dangerous AI agent called OpenClaw on users’ machines. The update was pushed through a compromised token, and many users unknowingly ran malicious code during installation. This incident highlights the risks of supply chain attacks in

  • svg
    Post Image

    Today, Lasso Security announced a new product called Intent Deputy. It aims to improve how enterprises protect AI agents by analyzing their behavior in real time. The tool promises ultra-fast detection speeds of under 50 milliseconds and a detection accuracy of 99.83%. This breakthrough is part of Lasso’s broader vision called Intent Security, which focuses

svg To Top