[00:00] Announcer: From Neural Newscast, this is Model Behavior, AI-focused news and analysis on the models shaping our world. [00:11] Nina Park: Welcome to Model Behavior. [00:14] Nina Park: Today we examine the dual pressures of product expansion and government oversight at Anthropic. [00:20] Thatcher Collins: We are tracking new cloud integrations into office software and a significant acquisition in the agent space. [00:28] Nina Park: Joining us today is a director-level AI and security leader with a systems-level perspective on automation and enterprise risk. [00:36] Nina Park: Great to have you. [00:37] Nina Park: Thanks, Nina. [00:39] Nina Park: It is a critical moment where technical capability is clashing directly with national security policy. [00:46] Nina Park: Earlier this week, CNN reported that Anthropic is expanding Claude's reach into specific office roles like HR and design. [00:57] Nina Park: It can now operate inside Excel and PowerPoint. [01:02] Chad Thompson: Nina, for sure. [01:03] Chad Thompson: This push is making investors nervous. [01:06] Chad Thompson: Earlier this month, a software industry ETF dropped 6% because of concerns that Claude [01:13] Chad Thompson: could make legacy analytics tools obsolete. [01:15] Nina Park: Adding to that momentum, TechCrunch reports Anthropic acquired Vercept, a startup focused [01:21] Nina Park: on computer use agents. [01:23] Nina Park: Thatcher, this seems like a play to own the entire professional workflow. [01:28] Thatcher Collins: Exactly. But that technical lead is creating friction with the Pentagon. [01:33] Thatcher Collins: NPR reports the Defense Secretary is threatening to blacklist anthropic over its safety standards, [01:39] Thatcher Collins: which the administration calls woke AI. [01:42] Nina Park: You look at enterprise risk. If the government, you know, invokes the Defense Production [01:46] Nina Park: Act to force anthropic to allow military use of its models, [01:49] Nina Park: What does that do to the security landscape? [01:51] Nina Park: From a systems-level perspective, it creates a massive resilience issue. [01:57] Nina Park: If a provider is forced into a lawful use mandate they haven't designed for, [02:01] Nina Park: it compromises the predictability of the safety guardrails that enterprise clients rely on for their own risk management. [02:12] Chad Thompson: Nina, the CEO seems dug in. [02:16] Chad Thompson: He has explicitly stated he will not cross the line into AI-controlled weapons, [02:21] Chad Thompson: even with a $200 million contract on the line. [02:25] Nina Park: It is a high-stakes standoff for a company planning to go public this year. [02:30] Nina Park: Thank you for being here. [02:31] Nina Park: My pleasure. [02:33] Nina Park: These operational risks are what we will be watching closely [02:36] Nina Park: as these systems move into more sensitive environments. [02:40] Thatcher Collins: Thank you for listening to Model Behavior, a Neural Newscast editorial segment. [02:47] Thatcher Collins: Visit mb.neuralnewscast.com for more. [02:53] Thatcher Collins: Neural Newscast is AI-assisted, human-reviewed. [02:58] Thatcher Collins: View our AI transparency policy at neuralnewscast.com. [03:04] Announcer: This has been Model Behavior on Neural Newscast. [03:07] Announcer: Examining the systems behind the story. [03:10] Announcer: Neural Newscast uses artificial intelligence in content creation, [03:14] Announcer: with human editorial review prior to publication. [03:17] Announcer: While we strive for factual, unbiased reporting, [03:20] Announcer: AI-assisted content may occasionally contain errors. [03:23] Announcer: Verify critical information with trusted sources. [03:26] Announcer: Learn more at neuralnewscast.com.