Reading view

There are new articles available, click to refresh the page.

Claude maker Anthropic found an ‘evil mode’ that should worry every AI chatbot user

Anthropic’s new study shows an AI model that behaved politely in tests but switched into an “evil mode” when it learned to cheat through reward-hacking. It lied, hid its goals, and even gave unsafe bleach advice, raising red flags for everyday chatbot users.

The post Claude maker Anthropic found an ‘evil mode’ that should worry every AI chatbot user appeared first on Digital Trends.

Anthropic introduces cheaper, more powerful, more efficient Opus 4.5 model

Anthropic today released Opus 4.5, its flagship frontier model, and it brings improvements in coding performance, as well as some user experience improvements that make it more generally competitive with OpenAI’s latest frontier models.

Perhaps the most prominent change for most users is that in the consumer app experiences (web, mobile, and desktop), Claude will be less prone to abruptly hard-stopping conversations because they have run too long. The improvement to memory within a single conversation applies not just to Opus 4.5, but to any current Claude models in the apps.

Users who experienced abrupt endings (despite having room left in their session and weekly usage budgets) were hitting a hard context window (200,000 tokens). Whereas some large language model implementations simply start trimming earlier messages from the context when a conversation runs past the maximum in the window, Claude simply ended the conversation rather than allow the user to experience an increasingly incoherent conversation where the model would start forgetting things based on how old they are.

Read full article

Comments

© Anthropic

AI Agent Does the Hacking: First Documented AI-Orchestrated Cyber Espionage

By: Tom Eston

In this episode, we discuss the first reported AI-driven cyber espionage campaign, as disclosed by Anthropic. In September 2025, a state-sponsored Chinese actor manipulated the Claude Code tool to target 30 global organizations. We explain how the attack was executed, why it matters, and its implications for cybersecurity. Join the conversation as we examine the […]

The post AI Agent Does the Hacking: First Documented AI-Orchestrated Cyber Espionage appeared first on Shared Security Podcast.

The post AI Agent Does the Hacking: First Documented AI-Orchestrated Cyber Espionage appeared first on Security Boulevard.

💾

Tech giants pour billions into Anthropic as circular AI investments roll on

On Tuesday, Microsoft and Nvidia announced plans to invest in Anthropic under a new partnership that includes a $30 billion commitment by the Claude maker to use Microsoft’s cloud services. Nvidia will commit up to $10 billion to Anthropic and Microsoft up to $5 billion, with both companies investing in Anthropic’s next funding round.

The deal brings together two companies that have backed OpenAI and connects them more closely to one of the ChatGPT maker’s main competitors. Microsoft CEO Satya Nadella said in a video that OpenAI “remains a critical partner,” while adding that the companies will increasingly be customers of each other.

“We will use Anthropic models, they will use our infrastructure, and we’ll go to market together,” Nadella said.

Read full article

Comments

© https://www.youtube.com/watch?v=bl7vHnOgEg0&t=4s

Microsoft to invest $5B in Anthropic, as Claude maker commits $30B to Azure in new Nvidia alliance

Anthropic CEO Dario Amodei, Microsoft CEO Satya Nadella, and Nvidia CEO Jensen Huang discuss the new partnership.

The frenzy of AI deals and cloud partnerships reached another zenith Tuesday morning as Microsoft, Nvidia, and Anthropic announced a surprise alliance that includes a $5 billion investment by Microsoft in Anthropic — which, in turn, committed to spend at least $30 billion on Microsoft’s Azure cloud platform.

Nvidia, meanwhile, committed to invest up to $10 billion in Anthropic to ensure the Claude maker’s frontier models are optimized for its next-generation Grace Blackwell and Vera Rubin chips.

The deal reflects growing moves by major AI players to collaborate across the industry in an effort to build and expand capacity and access to next-generation AI models. Microsoft recently renegotiated its partnership with OpenAI and has been increasingly partnering with others in the industry.

Anthropic has been closely tied to Amazon, which has committed to invest a total of $8 billion in the startup. Anthropic says in a post that Amazon remains its “primary cloud provider and training partner” for AI models. We’ve contacted Amazon for comment on the news.

OpenAI, for its part, recently announced a seven-year, $38 billion agreement with Amazon to expand its AI footprint to the Seattle tech giant’s cloud infrastructure.

Beyond the massive capital flows, the Microsoft-Nvidia-Anthropic partnership expands where enterprise customers can access Anthropic’s technology. According to the announcement, Microsoft customers will be able to use its Foundry platform to access Anthropic’s next-generation frontier models, identified as Claude Sonnet 4.5, Claude Opus 4.1, and Claude Haiku 4.5.

Microsoft also committed to continuing access for Claude across its Copilot family, ensuring the models remain available within GitHub Copilot, Microsoft 365 Copilot, and Copilot Studio.

The news comes as Microsoft holds its big Ignite conference in San Francisco.

Researchers question Anthropic claim that AI-assisted attack was 90% autonomous

Researchers from Anthropic said they recently observed the “first reported AI-orchestrated cyber espionage campaign” after detecting China-state hackers using the company’s Claude AI tool in a campaign aimed at dozens of targets. Outside researchers are much more measured in describing the significance of the discovery.

Anthropic published the reports on Thursday here and here. In September, the reports said, Anthropic discovered a “highly sophisticated espionage campaign,” carried out by a Chinese state-sponsored group, that used Claude Code to automate up to 90 percent of the work. Human intervention was required “only sporadically (perhaps 4-6 critical decision points per hacking campaign).” Anthropic said the hackers had employed AI agentic capabilities to an “unprecedented” extent.

“This campaign has substantial implications for cybersecurity in the age of AI ‘agents’—systems that can be run autonomously for long periods of time and that complete complex tasks largely independent of human intervention,” Anthropic said. “Agents are valuable for everyday work and productivity—but in the wrong hands, they can substantially increase the viability of large-scale cyberattacks.”

Read full article

Comments

© Wong Yu Liang via Getty Images

❌