{"id":88392,"date":"2026-03-10T09:58:11","date_gmt":"2026-03-10T08:58:11","guid":{"rendered":"https:\/\/www.humanlevel.com\/uncategorized\/el-caso-anthropic-y-los-limites-de-la-ia"},"modified":"2026-03-10T09:58:11","modified_gmt":"2026-03-10T08:58:11","slug":"the-anthropic-case-and-the-limits-of-ai","status":"publish","type":"post","link":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai","title":{"rendered":"The Anthropic case and the limits of AI"},"content":{"rendered":"<p>The standoff between the Artificial Intelligence laboratory Anthropic, creator of the Claude model, and the United States Department of Defense (recently renamed the \u201cDepartment of War\u201d) shows us what happens when a model ceases to be just another piece of software and becomes infrastructure within sensitive environments. In this case: defense, intelligence, administration, and contractors.<\/p>\n<p>But above all, it represents the <strong>first frontal clash between commercial AI safety policies and national security operational imperatives.<\/strong><\/p>\n<p>Anthropic\u2019s categorical refusal to relax certain ethical and technical restrictions in the face of military demands has triggered a series of government reprisals, including social media posts, culminating in the company\u2019s designation as a <strong>\u201csupply chain risk.\u201d<\/strong><\/p>\n<p>Until now, the debate usually centered on whether AI worked or how much it helped. But this case is about something that will likely become increasingly critical: <strong>who is in charge, who sets the limits, and what happens when those red lines collide.<\/strong><\/p>\n<h2><strong>Context<\/strong><\/h2>\n<p>To understand the clash, one must observe two trends that had been advancing in parallel.<\/p>\n<p>On one hand, the use of assistants and models for analysis, drafting, and information classification tasks, among many other use cases, has become normalized across many administrations. In the realm of defense and intelligence, the pressure is even higher: <strong>if a technology exists that can accelerate processes or provide an advantage, <a href=\"https:\/\/www.war.gov\/News\/Releases\/Release\/Article\/4376420\/war-department-launches-ai-acceleration-strategy-to-secure-american-military-ai\/\">the temptation to integrate it as soon as possible<\/a> is clear<\/strong>, even if it is not yet mature.<\/p>\n<p>On the other hand, the companies developing these models, including Anthropic, have been defining their own safety policies. This isn&#8217;t just out of conviction, but because generative AI can be used for very different purposes depending on who holds it and what data surrounds it. That is why <strong>these laboratories are compelled to implement various limits, controls, and terms of use to mitigate risks,<\/strong> whether at a technical, legal, or even reputational level.<\/p>\n<p>And here lies the <em>heart<\/em> of the problem.<\/p>\n<ul>\n<li>Governments wanting to adopt AI rapidly and with the widest possible margin.<\/li>\n<li>AI companies imposing limits for safety, ethics, and responsibility.<\/li>\n<li>Laws and guidelines attempting to bring order, but not yet covering all scenarios at the operational level that a real deployment demands.<\/li>\n<\/ul>\n<p>When those three things do not align, conflict is almost inevitable. To this, we must add an important nuance that brings even more complexity to the matter: in AI labs, certain decisions are not only made looking at public opinion, but at their own teams and the talent that makes the product possible.<\/p>\n<h2><strong>Timeline<\/strong><\/h2>\n<p>Before the public confrontation, Claude was no longer just a tool; it had been moving closer to the defense ecosystem and classified environments for some time.<\/p>\n<p>Let\u2019s look at <a href=\"https:\/\/www.techpolicy.press\/a-timeline-of-the-anthropic-pentagon-dispute\/\">some of the key dates<\/a> to better understand the chronology:<\/p>\n<ul>\n<li><strong>2024\u20132025<\/strong>: Claude stops being just an assistant and starts entering defense and national security environments, with <a href=\"https:\/\/www.anthropic.com\/news\/anthropic-and-the-department-of-defense-to-advance-responsible-ai-in-defense-operations\">agreements and initiatives<\/a> for its use in sensitive scenarios.<\/li>\n<li><strong>January 2026:<\/strong> Tactical use of Claude in the <a href=\"https:\/\/nypost.com\/2026\/02\/14\/world-news\/ai-tool-claude-helped-capture-venezuelan-dictator-nicolas-maduro-in-us-military-raid-operation-report\/\">military incursion in Venezuela<\/a>, marking the first empirical demonstration of AI&#8217;s technical viability on a real battlefield.<\/li>\n<li><strong>February 24\u201326, 2026<\/strong>: The Pentagon demands that Anthropic accept a &#8220;use for any lawful purpose&#8221; framework. Dario Amodei (Anthropic) refuses and <a href=\"https:\/\/www.anthropic.com\/news\/statement-department-of-war\">makes his red lines public<\/a>: no relaxing of safeguards linked to mass domestic surveillance (specifically systematic, predictive, large-scale monitoring of U.S. citizens within national borders) and fully autonomous weapons.<\/li>\n<li><strong>February 27, 2026<\/strong>: The dispute escalates to the point where Trump orders federal agencies to stop using Anthropic technology, accusing the company of being operated by <a href=\"https:\/\/truthsocial.com\/@realDonaldTrump\/posts\/116144552969293195\">&#8220;radical left-wing fanatics&#8221;<\/a> who were attempting to coerce the government and force the armed forces to obey commercial terms of service over the Constitution. Even more serious for Anthropic: That same day, Secretary of Defense Hegseth urged the DoD to officially designate Anthropic as a <a href=\"https:\/\/es.euronews.com\/next\/2026\/03\/06\/ee-uu-anthropic-riesgo-cadena-suministro-uso-militar-ia\">&#8220;supply chain risk to national security.&#8221;<\/a> In a statement, he declared that as a result of this label, &#8220;effective immediately, no contractor, vendor, or partner doing business with the United States military may engage in any commercial activity with Anthropic.&#8221;<\/li>\n<li><strong>February 28, 2026:<\/strong> Sam Altman (OpenAI) announces the signing of a deal with the Pentagon to fill the void left by Anthropic with more explicit limits. This announcement triggers a 300% surge in <a href=\"https:\/\/www.euronews.com\/next\/2026\/03\/02\/cancel-chatgpt-ai-boycott-surges-after-openai-pentagon-military-deal\">ChatGPT uninstalls<\/a> in the U.S., along with numerous negative reviews. Simultaneously, Claude is rewarded with a notable increase in downloads, <a href=\"https:\/\/www.reuters.com\/technology\/anthropic-courted-pentagon-heres-why-it-walked-away-2026-03-04\/\">reaching number 1 on the U.S. App Store<\/a> on Saturday and holding the position until Monday, March 2.<\/li>\n<li><strong>March 2, 2026:<\/strong> The domino effect begins in agencies: Treasury and other bodies announce they are <a href=\"https:\/\/www.reuters.com\/business\/us-treasury-ending-all-use-anthropic-products-says-bessent-2026-03-02\/\">terminating the use of Anthropic products<\/a> in response to the presidential order.<\/li>\n<li><strong>March 6, 2026:<\/strong> The Pentagon reaffirms the national security supply chain risk label \u201ceffective immediately,\u201d and the case enters a phase of institutional pressure (with major tech companies, including <a href=\"https:\/\/es.marketscreener.com\/noticias\/grupo-de-big-tech-expresa-su-preocupaci-n-al-pent-gono-por-la-designaci-n-de-anthropic-como-riesgo-e-ce7e5fdada88f720\">Big Tech groups, expressing concern<\/a> to the Pentagon) and legal response (<a href=\"https:\/\/cnnespanol.cnn.com\/2026\/03\/09\/eeuu\/anthropic-demanda-gobierno-trump-trax\">Anthropic states it will challenge the designation)<\/a>.<\/li>\n<\/ul>\n<p><img class=\"gb-media-8d6c92ba\" title=\"donald-trump\" src=\"https:\/\/www.humanlevel.com\/wp-content\/uploads\/donald-trump.jpg\" alt=\"\" \/> <\/p>\n<h2><strong>The Core of the Conflict: What is Truly Being Debated<\/strong><\/h2>\n<p>At first glance, it might seem this is about whether AI can or cannot be used in defense. But if we scratch the surface, the real debate is something else\u2014much more uncomfortable and, likely, much more frequent from now on.<\/p>\n<p>The clash can be summarized in a very simple question: <strong>who sets the limits when a model is integrated into critical systems?<\/strong> Is it the client, because they pay and assume operational responsibility? Is it the provider, because they control access to the technology and set conditions? Or is it the law, even if it sometimes fails to reach the level of detail required for real implementation?<\/p>\n<p>The Department of Defense encourages a clear idea: if a use is legal, it should be permissible. From this approach, we can understand that the limits would already be set by laws, internal oversight, and the chain of command. Therefore, restrictions added by the provider are perceived as a brake\u2014not just on operations, but on sovereignty and response capacity.<\/p>\n<p>Anthropic, for its part, argues the opposite: even if something can be defended as legal or falls into a gray area, there are uses it does not want to enable. This is where its two publicly repeated red lines come into play: <strong>mass domestic surveillance and fully autonomous weapons.<\/strong> The company maintains that it is not just an ethical or reputational stance, but also one of safety: in certain scenarios, models are not reliable enough to operate without a human in the loop, especially when discussing decisions with irreversible consequences.<\/p>\n<p>Ultimately, what this case puts on the table is a key distinction: <strong>legal does not always equate to acceptable, and acceptable does not always equate to safe.<\/strong><\/p>\n<p>When technology moves faster than regulation, as is undoubtedly the case with AI, that middle ground becomes a swampy terrain of friction where what was once a provider&#8217;s internal policy becomes a contractual condition, and what was once a client&#8217;s operational necessity clashes with limits they do not control.<\/p>\n<p>That is why this controversy is not just an ethical discussion (which it is), but fundamentally a <strong>governance discussion: who decides, and with what authority.<\/strong><\/p>\n<h2><strong>Current Landscape: Where We Stand Right Now<\/strong><\/h2>\n<p>As of today, the board is far from a simple disagreement between a company and an institutional client. We can already see consequences.<\/p>\n<p>The first piece is the most obvious: operational rupture and forced transition.<\/p>\n<p>Following the escalation, the Pentagon formalized the \u201csupply chain risk\u201d label effective immediately, <strong>pushing agencies, contractors, and vendors to cut or rethink relationships with Anthropic,<\/strong> even though the product hasn&#8217;t technically changed overnight. This could have severe consequences for Anthropic, particularly regarding its access to the public market, and sends an uneasy signal to potential investors.<\/p>\n<p>The second piece is what makes this case a serious warning: technological dependency.<\/p>\n<p>Changing a model in environments where Claude was integrated is not immediate. Reuters, for instance, has described <a href=\"https:\/\/www.reuters.com\/technology\/palantir-faces-challenge-remove-anthropic-pentagons-ai-software-2026-03-04\/\">the case of Palantir and its Maven software<\/a>: removing Claude implies replacing it and, in some areas, rebuilding parts of the system. In other words, <strong>when the model is already infrastructure, there is an exit cost, and it must be paid.<\/strong><\/p>\n<p>And the third piece is that the gap left by a provider in such a sensitive environment rarely stays empty.<\/p>\n<p><strong>OpenAI announced its own agreement with the Department of War<\/strong> and accompanied it with a very specific stance: working with defense, yes, but with explicit limits\u2014including a later addition to state in writing that its tools will not be used for domestic surveillance of people in the U.S., nor through the use or purchase of commercial personal data. At least for public consumption.<\/p>\n<p>However, that entry has also come at a reputational cost that is still difficult to measure. This move has been interpreted by many as opportunistic to the point that Sam Altman himself acknowledged the announcement \u201cfelt opportunistic and botched\u201d due to how it was communicated, leading to a sort of boycott by some users against the company and ChatGPT that is hard to predict.<\/p>\n<p><img class=\"gb-media-df60f792\" title=\"claude\" src=\"https:\/\/www.humanlevel.com\/wp-content\/uploads\/claude.jpg\" alt=\"\" \/> <\/p>\n<h2><strong>What Might Happen From Here<\/strong><\/h2>\n<p>While we cannot know for sure what will happen and it makes no sense to make absolute predictions, we can glimpse some likely scenarios based on how negotiations, institutional pressure, and transition costs evolve.<\/p>\n<h3><strong>Minimum Agreement and Partial Return<\/strong><\/h3>\n<p>The most pragmatic move for both sides would be to close a framework that allows for a partial return to use, but with better-defined limits in writing: <strong>what exactly constitutes \u201cdomestic surveillance,\u201d what controls exist, who audits, and what responsibilities each party assumes.<\/strong> It wouldn&#8217;t really be a romantic reconciliation, but a tightly bound patch to avoid operational chaos and the cost of restructuring everything.<\/p>\n<h3><strong>Long Conflict and Phased Withdrawal<\/strong><\/h3>\n<p>If neither side yields on substance, the situation could drag on indefinitely, involving challenges, media pressure, a war of press releases, and a gradual withdrawal in practice\u2014with specific exceptions where quick changes aren&#8217;t viable. This would trigger a <strong>prolonged period of uncertainty.<\/strong> In this scenario, the wear and tear would no longer just be reputational, but also related to resources and project continuity.<\/p>\n<h3><strong>Shift Toward Reducing Dependency on a Single Provider<\/strong><\/h3>\n<p>Regardless of who wins the standoff, this case leaves a lesson for any organization (and any large company): <strong>if an AI becomes infrastructure, you must design with a Plan B in mind.<\/strong> This means more modular architectures, the ability to switch providers, and real exit plans. Not because it\u2019s ideal\u2014it isn&#8217;t\u2014but because it reduces the risk of such a conflict paralyzing operations again.<\/p>\n<h2><strong>What This Case Teaches Us (And Why It Should Matter)<\/strong><\/h2>\n<p>This entire controversy serves as a case study precisely because it doesn&#8217;t talk about hypothetical futures. It speaks of real risks that hadn&#8217;t been sufficiently considered until now and are arising as AI stops being a tool and starts functioning as infrastructure.<\/p>\n<p>If you\u2019ve made it this far, you\u2019ve likely drawn your own conclusions, but these are the clearest lessons:<\/p>\n<h3><strong>Technological Dependency Is a Risk in Itself<\/strong><\/h3>\n<p>As soon as a model is integrated into critical processes, changing it stops being a quick decision. Costs, delays, validations, team reorganization, or even system rebuilding appear. And this must be taken into account.<\/p>\n<h3><strong>The Question of Who Is in Charge Isn&#8217;t Solved by Default<\/strong><\/h3>\n<p>The client pays and assumes operational responsibility. The provider controls the technology and its conditions. And the law determines the framework. But the truth is that this framework doesn&#8217;t always reach the level of detail that a real deployment requires, and it may never fully do so. When those three elements don&#8217;t fit, conflict becomes almost inevitable, as in this case.<\/p>\n<h3><strong>Setting Limits Is Not Just About Stating Them; You Must Be Able to Uphold Them<\/strong><\/h3>\n<p>It&#8217;s one thing to publish principles and another entirely to turn them into conditions that can be applied in practice: what is specifically prohibited, how it is detected, who reviews it, what the consequences are if the line is crossed, and which party assumes responsibility.<\/p>\n<h3><strong>Ethics Appear Because Technology Is Ahead<\/strong><\/h3>\n<p>In an ideal world, everything would be perfectly regulated and every case would have its framework. In the real world, AI evolves faster than the capacity to legislate while considering operational realities. Add to this the lack of necessary operational knowledge among some legislators in this field, and you have the perfect storm. That is why ethics don&#8217;t disappear even when there are laws; they appear as a way to set limits where the norm hasn&#8217;t yet reached, or where the margin for interpretation is too wide.<\/p>\n<h3><strong>This Isn&#8217;t About Anthropic; It&#8217;s About What&#8217;s Coming<\/strong><\/h3>\n<p>This case matters because it shows us that AI is entering zones where it\u2019s not enough for it to just work. You need to know who decides, what happens if the framework changes tomorrow, and how dependent you become when you truly integrate it.<\/p>\n<p>As a user, it reminds you that your favorite tool isn&#8217;t neutral: it&#8217;s subject to agreements, limits, and decisions that can change fast.<\/p>\n<p>And as a company, it leaves a very practical lesson: <strong>if you are going to integrate AI into important processes, it\u2019s wise to do so thoughtfully\u2014thinking not just about what it brings me, but also what happens if I have to change it tomorrow.<\/strong><\/p>\n","protected":false},"excerpt":{"rendered":"<p>The Anthropic case shows the first frontal clash between commercial AI safety and operational defense&#8230;<\/p>\n","protected":false},"author":54,"featured_media":88394,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"footnotes":""},"categories":[459],"tags":[],"class_list":["post-88392","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-artificial-intelligence"],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.5 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>The Anthropic case and the limits of AI | Human Level<\/title>\n<meta name=\"description\" content=\"Standoff between Anthropic and the Pentagon: when AI becomes infrastructure in critical environments, who sets the limits?\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"The Anthropic case and the limits of AI\" \/>\n<meta property=\"og:description\" content=\"Standoff between Anthropic and the Pentagon: when AI becomes infrastructure in critical environments, who sets the limits?\" \/>\n<meta property=\"og:url\" content=\"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai\" \/>\n<meta property=\"og:site_name\" content=\"Human Level\" \/>\n<meta property=\"article:published_time\" content=\"2026-03-10T08:58:11+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/www.humanlevel.com\/wp-content\/uploads\/anthropic-social-en.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"1200\" \/>\n\t<meta property=\"og:image:height\" content=\"675\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Alberto Fern\u00e1ndez\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@HagiumSEO\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Alberto Fern\u00e1ndez\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"11 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":[\"Article\",\"BlogPosting\"],\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\\\/the-anthropic-case-and-the-limits-of-ai#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\\\/the-anthropic-case-and-the-limits-of-ai\"},\"author\":{\"name\":\"Alberto Fern\u00e1ndez\",\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en#\\\/schema\\\/person\\\/8cd46b80ecebec10e3718c9dbc9e17ec\"},\"headline\":\"The Anthropic case and the limits of AI\",\"datePublished\":\"2026-03-10T08:58:11+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\\\/the-anthropic-case-and-the-limits-of-ai\"},\"wordCount\":2232,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en#organization\"},\"image\":{\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\\\/the-anthropic-case-and-the-limits-of-ai#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/www.humanlevel.com\\\/wp-content\\\/uploads\\\/anthropic-blog.jpg\",\"articleSection\":[\"AI\"],\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\\\/the-anthropic-case-and-the-limits-of-ai#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\\\/the-anthropic-case-and-the-limits-of-ai\",\"url\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\\\/the-anthropic-case-and-the-limits-of-ai\",\"name\":\"The Anthropic case and the limits of AI | Human Level\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\\\/the-anthropic-case-and-the-limits-of-ai#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\\\/the-anthropic-case-and-the-limits-of-ai#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/www.humanlevel.com\\\/wp-content\\\/uploads\\\/anthropic-blog.jpg\",\"datePublished\":\"2026-03-10T08:58:11+00:00\",\"description\":\"Standoff between Anthropic and the Pentagon: when AI becomes infrastructure in critical environments, who sets the limits?\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\\\/the-anthropic-case-and-the-limits-of-ai#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\\\/the-anthropic-case-and-the-limits-of-ai\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\\\/the-anthropic-case-and-the-limits-of-ai#primaryimage\",\"url\":\"https:\\\/\\\/www.humanlevel.com\\\/wp-content\\\/uploads\\\/anthropic-blog.jpg\",\"contentUrl\":\"https:\\\/\\\/www.humanlevel.com\\\/wp-content\\\/uploads\\\/anthropic-blog.jpg\",\"width\":272,\"height\":272},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\\\/the-anthropic-case-and-the-limits-of-ai#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Inicio\",\"item\":\"https:\\\/\\\/www.humanlevel.com\\\/en\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Blog\",\"item\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\"},{\"@type\":\"ListItem\",\"position\":3,\"name\":\"AI\",\"item\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/blog\\\/artificial-intelligence\"},{\"@type\":\"ListItem\",\"position\":4,\"name\":\"The Anthropic case and the limits of AI\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en#website\",\"url\":\"https:\\\/\\\/www.humanlevel.com\\\/en\",\"name\":\"Human Level\",\"description\":\"Web positioning and online marketing consultant Human Level\",\"publisher\":{\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/www.humanlevel.com\\\/en?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en#organization\",\"name\":\"Human Level\",\"url\":\"https:\\\/\\\/www.humanlevel.com\\\/en\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/www.humanlevel.com\\\/wp-content\\\/uploads\\\/logohl25x3.png\",\"contentUrl\":\"https:\\\/\\\/www.humanlevel.com\\\/wp-content\\\/uploads\\\/logohl25x3.png\",\"width\":600,\"height\":93,\"caption\":\"Human Level\"},\"image\":{\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en#\\\/schema\\\/logo\\\/image\\\/\"},\"sameAs\":[\"https:\\\/\\\/www.linkedin.com\\\/company\\\/human-level-communications\",\"https:\\\/\\\/www.youtube.com\\\/user\\\/humanlevelcommunica\",\"https:\\\/\\\/bsky.app\\\/profile\\\/humanlevel.bsky.social\",\"https:\\\/\\\/instagram.com\\\/humanlevel\"]},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/en#\\\/schema\\\/person\\\/8cd46b80ecebec10e3718c9dbc9e17ec\",\"name\":\"Alberto Fern\u00e1ndez\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/www.humanlevel.com\\\/wp-content\\\/uploads\\\/1x1-alberto-fernandez-26-96x96.jpg\",\"url\":\"https:\\\/\\\/www.humanlevel.com\\\/wp-content\\\/uploads\\\/1x1-alberto-fernandez-26-96x96.jpg\",\"contentUrl\":\"https:\\\/\\\/www.humanlevel.com\\\/wp-content\\\/uploads\\\/1x1-alberto-fernandez-26-96x96.jpg\",\"caption\":\"Alberto Fern\u00e1ndez\"},\"description\":\"Alberto has been a digital enthusiast from an early age, which led him to study Computer Engineering and work as a web developer at Cronis Online. Later, he expanded his expertise in sales, marketing, and team leadership at Phone House, where he managed his own team. His constant pursuit of knowledge drove him to delve deeper into digital marketing and SEO, even completing a vocational degree in Web Application Development. After serving as Head of Web Development and SEO at 6D Visual, he now works at Human Level as a Technical SEO Consultant, consolidating over 20 years of experience.\",\"sameAs\":[\"https:\\\/\\\/es.linkedin.com\\\/in\\\/albertofernandezseo\",\"https:\\\/\\\/x.com\\\/HagiumSEO\"],\"url\":\"https:\\\/\\\/www.humanlevel.com\\\/en\\\/author\\\/alberto-fernandez\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"The Anthropic case and the limits of AI | Human Level","description":"Standoff between Anthropic and the Pentagon: when AI becomes infrastructure in critical environments, who sets the limits?","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai","og_locale":"en_US","og_type":"article","og_title":"The Anthropic case and the limits of AI","og_description":"Standoff between Anthropic and the Pentagon: when AI becomes infrastructure in critical environments, who sets the limits?","og_url":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai","og_site_name":"Human Level","article_published_time":"2026-03-10T08:58:11+00:00","og_image":[{"width":1200,"height":675,"url":"https:\/\/www.humanlevel.com\/wp-content\/uploads\/anthropic-social-en.jpg","type":"image\/jpeg"}],"author":"Alberto Fern\u00e1ndez","twitter_card":"summary_large_image","twitter_creator":"@HagiumSEO","twitter_misc":{"Written by":"Alberto Fern\u00e1ndez","Est. reading time":"11 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":["Article","BlogPosting"],"@id":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai#article","isPartOf":{"@id":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai"},"author":{"name":"Alberto Fern\u00e1ndez","@id":"https:\/\/www.humanlevel.com\/en#\/schema\/person\/8cd46b80ecebec10e3718c9dbc9e17ec"},"headline":"The Anthropic case and the limits of AI","datePublished":"2026-03-10T08:58:11+00:00","mainEntityOfPage":{"@id":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai"},"wordCount":2232,"commentCount":0,"publisher":{"@id":"https:\/\/www.humanlevel.com\/en#organization"},"image":{"@id":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai#primaryimage"},"thumbnailUrl":"https:\/\/www.humanlevel.com\/wp-content\/uploads\/anthropic-blog.jpg","articleSection":["AI"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai#respond"]}]},{"@type":"WebPage","@id":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai","url":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai","name":"The Anthropic case and the limits of AI | Human Level","isPartOf":{"@id":"https:\/\/www.humanlevel.com\/en#website"},"primaryImageOfPage":{"@id":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai#primaryimage"},"image":{"@id":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai#primaryimage"},"thumbnailUrl":"https:\/\/www.humanlevel.com\/wp-content\/uploads\/anthropic-blog.jpg","datePublished":"2026-03-10T08:58:11+00:00","description":"Standoff between Anthropic and the Pentagon: when AI becomes infrastructure in critical environments, who sets the limits?","breadcrumb":{"@id":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai#primaryimage","url":"https:\/\/www.humanlevel.com\/wp-content\/uploads\/anthropic-blog.jpg","contentUrl":"https:\/\/www.humanlevel.com\/wp-content\/uploads\/anthropic-blog.jpg","width":272,"height":272},{"@type":"BreadcrumbList","@id":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence\/the-anthropic-case-and-the-limits-of-ai#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Inicio","item":"https:\/\/www.humanlevel.com\/en"},{"@type":"ListItem","position":2,"name":"Blog","item":"https:\/\/www.humanlevel.com\/en\/blog"},{"@type":"ListItem","position":3,"name":"AI","item":"https:\/\/www.humanlevel.com\/en\/blog\/artificial-intelligence"},{"@type":"ListItem","position":4,"name":"The Anthropic case and the limits of AI"}]},{"@type":"WebSite","@id":"https:\/\/www.humanlevel.com\/en#website","url":"https:\/\/www.humanlevel.com\/en","name":"Human Level","description":"Web positioning and online marketing consultant Human Level","publisher":{"@id":"https:\/\/www.humanlevel.com\/en#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/www.humanlevel.com\/en?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/www.humanlevel.com\/en#organization","name":"Human Level","url":"https:\/\/www.humanlevel.com\/en","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.humanlevel.com\/en#\/schema\/logo\/image\/","url":"https:\/\/www.humanlevel.com\/wp-content\/uploads\/logohl25x3.png","contentUrl":"https:\/\/www.humanlevel.com\/wp-content\/uploads\/logohl25x3.png","width":600,"height":93,"caption":"Human Level"},"image":{"@id":"https:\/\/www.humanlevel.com\/en#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.linkedin.com\/company\/human-level-communications","https:\/\/www.youtube.com\/user\/humanlevelcommunica","https:\/\/bsky.app\/profile\/humanlevel.bsky.social","https:\/\/instagram.com\/humanlevel"]},{"@type":"Person","@id":"https:\/\/www.humanlevel.com\/en#\/schema\/person\/8cd46b80ecebec10e3718c9dbc9e17ec","name":"Alberto Fern\u00e1ndez","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.humanlevel.com\/wp-content\/uploads\/1x1-alberto-fernandez-26-96x96.jpg","url":"https:\/\/www.humanlevel.com\/wp-content\/uploads\/1x1-alberto-fernandez-26-96x96.jpg","contentUrl":"https:\/\/www.humanlevel.com\/wp-content\/uploads\/1x1-alberto-fernandez-26-96x96.jpg","caption":"Alberto Fern\u00e1ndez"},"description":"Alberto has been a digital enthusiast from an early age, which led him to study Computer Engineering and work as a web developer at Cronis Online. Later, he expanded his expertise in sales, marketing, and team leadership at Phone House, where he managed his own team. His constant pursuit of knowledge drove him to delve deeper into digital marketing and SEO, even completing a vocational degree in Web Application Development. After serving as Head of Web Development and SEO at 6D Visual, he now works at Human Level as a Technical SEO Consultant, consolidating over 20 years of experience.","sameAs":["https:\/\/es.linkedin.com\/in\/albertofernandezseo","https:\/\/x.com\/HagiumSEO"],"url":"https:\/\/www.humanlevel.com\/en\/author\/alberto-fernandez"}]}},"_links":{"self":[{"href":"https:\/\/www.humanlevel.com\/en\/wp-json\/wp\/v2\/posts\/88392","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.humanlevel.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.humanlevel.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.humanlevel.com\/en\/wp-json\/wp\/v2\/users\/54"}],"replies":[{"embeddable":true,"href":"https:\/\/www.humanlevel.com\/en\/wp-json\/wp\/v2\/comments?post=88392"}],"version-history":[{"count":2,"href":"https:\/\/www.humanlevel.com\/en\/wp-json\/wp\/v2\/posts\/88392\/revisions"}],"predecessor-version":[{"id":88396,"href":"https:\/\/www.humanlevel.com\/en\/wp-json\/wp\/v2\/posts\/88392\/revisions\/88396"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.humanlevel.com\/en\/wp-json\/wp\/v2\/media\/88394"}],"wp:attachment":[{"href":"https:\/\/www.humanlevel.com\/en\/wp-json\/wp\/v2\/media?parent=88392"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.humanlevel.com\/en\/wp-json\/wp\/v2\/categories?post=88392"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.humanlevel.com\/en\/wp-json\/wp\/v2\/tags?post=88392"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}