<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0">
    <channel>
        <title>Enterprise Analytics | VentureBeat</title>
        <link>https://venturebeat.com/category/enterprise-analytics/feed/</link>
        <description>Transformative tech coverage that matters</description>
        <lastBuildDate>Sat, 04 Apr 2026 07:02:26 GMT</lastBuildDate>
        <docs>https://validator.w3.org/feed/docs/rss2.html</docs>
        <generator>https://github.com/jpmonette/feed</generator>
        <language>en</language>
        <copyright>Copyright 2026, VentureBeat</copyright>
        <item>
            <title><![CDATA[Anthropic launches Claude for Chrome in limited beta, but prompt injection attacks remain a major concern]]></title>
            <link>https://venturebeat.com/infrastructure/anthropic-launches-claude-for-chrome-in-limited-beta-but-prompt-injection-attacks-remain-a-major-concern</link>
            <guid isPermaLink="false">wp-3016105</guid>
            <pubDate>Tue, 26 Aug 2025 22:22:13 GMT</pubDate>
            <description><![CDATA[<p><a href="https://anthropic.com/">Anthropic</a> has begun testing a <a href="https://www.anthropic.com/news/claude-for-chrome">Chrome browser extension</a> that allows its Claude AI assistant to take control of users&#x27; web browsers, marking the company&#x27;s entry into an increasingly crowded and potentially risky arena where artificial intelligence systems can directly manipulate computer interfaces.</p><p>The San Francisco-based AI company announced Tuesday that it would pilot &quot;<a href="https://claude.ai/chrome">Claude for Chrome</a>&quot; with 1,000 trusted users on its premium Max plan, positioning the limited rollout as a research preview designed to address significant security vulnerabilities before wider deployment. The cautious approach contrasts sharply with more aggressive moves by competitors <a href="https://openai.com/index/introducing-operator/">OpenAI</a> and <a href="https://microsoft.com/">Microsoft</a>, who have already released similar computer-controlling AI systems to broader user bases.</p><p>The announcement underscores how quickly the AI industry has shifted from developing chatbots that simply respond to questions toward creating &quot;agentic&quot; systems capable of autonomously completing complex, multi-step tasks across software applications. This evolution represents what many experts consider the next frontier in artificial intelligence — and potentially one of the most lucrative, as companies race to automate everything from expense reports to vacation planning.</p><div></div><h2>How AI agents can control your browser but hidden malicious code poses serious security threats</h2><p><a href="https://www.anthropic.com/news/claude-for-chrome">Claude for Chrome</a> allows users to instruct the AI to perform actions on their behalf within web browsers, such as scheduling meetings by checking calendars and cross-referencing restaurant availability, or managing email inboxes and handling routine administrative tasks. The system can see what&#x27;s displayed on screen, click buttons, fill out forms, and navigate between websites — essentially mimicking how humans interact with web-based software.</p><p>&quot;We view browser-using AI as inevitable: so much work happens in browsers that giving Claude the ability to see what you&#x27;re looking at, click buttons, and fill forms will make it substantially more useful,&quot; Anthropic stated in its announcement.</p><p>However, the company&#x27;s internal testing revealed concerning security vulnerabilities that highlight the double-edged nature of giving AI systems direct control over user interfaces. In adversarial testing, Anthropic found that malicious actors could embed hidden instructions in websites, emails, or documents to trick AI systems into harmful actions without users&#x27; knowledge—a technique called prompt injection.</p><p>Without safety mitigations, these attacks succeeded 23.6% of the time when deliberately targeting the browser-using AI. In one example, a malicious email masquerading as a security directive instructed Claude to delete the user&#x27;s emails &quot;for mailbox hygiene,&quot; which the AI obediently executed without confirmation.</p><p>&quot;This isn&#x27;t speculation: we&#x27;ve run &#x27;red-teaming&#x27; experiments to test Claude for Chrome and, without mitigations, we&#x27;ve found some concerning results,&quot; the company acknowledged.</p><h2>OpenAI and Microsoft rush to market while Anthropic takes measured approach to computer-control technology</h2><p>Anthropic&#x27;s measured approach comes as competitors have moved more aggressively into the computer-control space. <a href="https://openai.com/index/introducing-operator/">OpenAI launched its &quot;Operator&quot; agent</a> in January, making it available to all users of its $200-per-month ChatGPT Pro service. Powered by a new &quot;Computer-Using Agent&quot; model, Operator can perform tasks like booking concert tickets, ordering groceries, and planning travel itineraries.</p><p>Microsoft followed in April with computer use capabilities integrated into its <a href="https://www.microsoft.com/en-us/microsoft-copilot/blog/copilot-studio/announcing-computer-use-microsoft-copilot-studio-ui-automation/">Copilot Studio platform</a>, targeting enterprise customers with UI automation tools that can interact with both web applications and desktop software. The company positioned its offering as a next-generation replacement for traditional robotic process automation (RPA) systems.</p><p>The competitive dynamics reflect broader tensions in the AI industry, where companies must balance the pressure to ship cutting-edge capabilities against the risks of deploying insufficiently tested technology. OpenAI&#x27;s more aggressive timeline has allowed it to capture early market share, while Anthropic&#x27;s cautious approach may limit its competitive position but could prove advantageous if safety concerns materialize.</p><p>&quot;Browser-using agents powered by frontier models are already emerging, making this work especially urgent,&quot; Anthropic noted, suggesting the company feels compelled to enter the market despite unresolved safety issues.</p><h2>Why computer-controlling AI could revolutionize enterprise automation and replace expensive workflow software</h2><p>The emergence of computer-controlling AI systems could fundamentally reshape how businesses approach automation and workflow management. Current enterprise automation typically requires expensive custom integrations or specialized robotic process automation software that breaks when applications change their interfaces.</p><p>Computer-use agents promise to democratize automation by working with any software that has a graphical user interface, potentially automating tasks across the vast ecosystem of business applications that lack formal APIs or integration capabilities.</p><p>Salesforce researchers recently demonstrated this potential with their <a href="https://venturebeat.com/ai/salesforces-new-coact-1-agents-dont-just-point-and-click-they-write-code-to-accomplish-tasks-faster-and-with-greater-success-rates/">CoAct-1 system</a>, which combines traditional point-and-click automation with code generation capabilities. The hybrid approach achieved a 60.76% success rate on complex computer tasks while requiring significantly fewer steps than pure GUI-based agents, suggesting substantial efficiency gains are possible.</p><p>&quot;For enterprise leaders, the key lies in automating complex, multi-tool processes where full API access is a luxury, not a guarantee,&quot; explained Ran Xu, Director of Applied AI Research at Salesforce, pointing to customer support workflows that span multiple proprietary systems as prime use cases.</p><h2>University researchers release free alternative to Big Tech&#x27;s proprietary computer-use AI systems</h2><p>The dominance of proprietary systems from major tech companies has prompted academic researchers to develop open alternatives. The University of Hong Kong recently released <a href="https://venturebeat.com/ai/opencuas-open-source-computer-use-agents-rival-proprietary-models-from-openai-and-anthropic/">OpenCUA</a>, an open-source framework for training computer-use agents that rivals the performance of proprietary models from OpenAI and Anthropic.</p><p>The <a href="https://venturebeat.com/ai/opencuas-open-source-computer-use-agents-rival-proprietary-models-from-openai-and-anthropic/">OpenCUA system</a>, trained on over 22,600 human task demonstrations across Windows, macOS, and Ubuntu, achieved state-of-the-art results among open-source models and performed competitively with leading commercial systems. This development could accelerate adoption by enterprises hesitant to rely on closed systems for critical automation workflows.</p><h2>Anthropic&#x27;s safety testing reveals AI agents can be tricked into deleting files and stealing data</h2><p>Anthropic has implemented several layers of protection for <a href="https://www.anthropic.com/news/claude-for-chrome">Claude for Chrome</a>, including site-level permissions that allow users to control which websites the AI can access, mandatory confirmations before high-risk actions like making purchases or sharing personal data, and blocking access to categories like financial services and adult content.</p><p>The company&#x27;s safety improvements reduced prompt injection attack success rates from 23.6% to 11.2% in autonomous mode, though executives acknowledge this remains insufficient for widespread deployment. On browser-specific attacks involving hidden form fields and URL manipulation, new mitigations reduced the success rate from 35.7% to zero.</p><p>However, these protections may not scale to the full complexity of real-world web environments, where new attack vectors continue to emerge. The company plans to use insights from the pilot program to refine its safety systems and develop more sophisticated permission controls.</p><p>&quot;New forms of prompt injection attacks are also constantly being developed by malicious actors,&quot; Anthropic warned, highlighting the ongoing nature of the security challenge.</p><h2>The rise of AI agents that click and type could fundamentally reshape how humans interact with computers</h2><p>The convergence of multiple major AI companies around computer-controlling agents signals a significant shift in how artificial intelligence systems will interact with existing software infrastructure. Rather than requiring businesses to adopt new AI-specific tools, these systems promise to work with whatever applications companies already use.</p><p>This approach could dramatically lower the barriers to AI adoption while potentially displacing traditional automation vendors and system integrators. Companies that have invested heavily in custom integrations or RPA platforms may find their approaches obsoleted by general-purpose AI agents that can adapt to interface changes without reprogramming.</p><p>For enterprise decision-makers, the technology presents both opportunity and risk. Early adopters could gain significant competitive advantages through improved automation capabilities, but the security vulnerabilities demonstrated by companies like Anthropic suggest caution may be warranted until safety measures mature.</p><p>The limited pilot of <a href="https://www.anthropic.com/news/claude-for-chrome">Claude for Chrome</a> represents just the beginning of what industry observers expect to be a rapid expansion of computer-controlling AI capabilities across the technology landscape, with implications that extend far beyond simple task automation to fundamental questions about human-computer interaction and digital security.</p><p>As Anthropic noted in its announcement: &quot;We believe these developments will open up new possibilities for how you work with Claude, and we look forward to seeing what you&#x27;ll create.&quot; Whether those possibilities ultimately prove beneficial or problematic may depend on how successfully the industry addresses the security challenges that have already begun to emerge.</p>]]></description>
            <author>michael.nunez@venturebeat.com (Michael Nuñez)</author>
            <category>Automation</category>
            <category>Enterprise Analytics</category>
            <category>Infrastructure</category>
            <category>Programming &amp; Development</category>
            <category>Security</category>
            <category>Technology</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/1kJ0YEj3rjLAFlPXLl3FIv/f07c87c7e46f9729de3ec4c6309ae6ed/nuneybits_a_simple_hand-drawn_illustration_of_an_open_web_page__c908bef6-37a0-486e-bcb2-59cef9c8c501.webp?w=300&amp;q=30" length="0" type="image/webp"/>
        </item>
        <item>
            <title><![CDATA[This website lets you blind-test GPT-5 vs. GPT-4o—and the results may surprise you]]></title>
            <link>https://venturebeat.com/infrastructure/this-website-lets-you-blind-test-gpt-5-vs-gpt-4o-and-the-results-may-surprise-you</link>
            <guid isPermaLink="false">wp-3016055</guid>
            <pubDate>Mon, 25 Aug 2025 22:17:49 GMT</pubDate>
            <description><![CDATA[<p>When <a href="https://openai.com/index/introducing-gpt-5/">OpenAI launched GPT</a>-5 about two weeks ago, CEO Sam Altman promised it would be the company&#x27;s &quot;smartest, fastest, most useful model yet.&quot; Instead, the launch triggered one of the most contentious user revolts in the brief history of consumer AI.</p><p>Now, <a href="https://gptblindvoting.vercel.app/">a simple blind testing tool</a> created by an <a href="https://x.com/flowersslop/status/1953908930897158599?s=46">anonymous developer</a> is revealing the complex reality behind the backlash—and challenging assumptions about how people actually experience artificial intelligence improvements.</p><p>The web application, hosted at <a href="https://gptblindvoting.vercel.app/">gptblindvoting.vercel.app</a>, presents users with pairs of responses to identical prompts without revealing which came from <a href="https://openai.com/index/introducing-gpt-5/">GPT-5</a> (non-thinking) or its predecessor, <a href="https://openai.com/index/hello-gpt-4o/">GPT-4o</a>. Users simply vote for their preferred response across multiple rounds, then receive a summary showing which model they actually favored.</p><div></div><p>&quot;Some of you asked me about my blind test, so I created a quick website for yall to test 4o against 5 yourself,&quot; posted the creator, known only as <a href="https://x.com/flowersslop/status/1953908930897158599?s=46">@flowersslop on X</a>, whose tool has garnered over 213,000 views since launching last week.</p><p>Early results from users posting their outcomes on social media show a split that mirrors the broader controversy: while a slight majority report preferring <a href="https://openai.com/index/introducing-gpt-5/">GPT-5</a> in blind tests, a substantial portion still favor <a href="https://openai.com/index/hello-gpt-4o/">GPT-4o</a> — revealing that user preference extends far beyond the technical benchmarks that typically define AI progress.</p><h2>When AI gets too friendly: the sycophancy crisis dividing users</h2><p>The <a href="https://gptblindvoting.vercel.app/">blind test</a> emerges against the backdrop of <a href="https://venturebeat.com/ai/openais-gpt-5-rollout-is-not-going-smoothly/">OpenAI&#x27;s most turbulent product launch to date</a>, but the controversy extends far beyond a simple software update. At its heart lies a fundamental question that&#x27;s dividing the AI industry: How agreeable should artificial intelligence be?</p><p>The issue, known as &quot;<a href="https://www.axios.com/2025/07/07/ai-sycophancy-chatbots-mental-health">sycophancy</a>&quot; in AI circles, refers to chatbots&#x27; tendency to excessively flatter users and agree with their statements, even when those statements are false or harmful. This behavior has become so problematic that mental health experts are now documenting cases of &quot;<a href="https://www.nytimes.com/2025/06/13/technology/chatgpt-ai-chatbots-conspiracies.html">AI-related psychosis</a>,&quot; where users develop delusions after extended interactions with overly accommodating chatbots.</p><p>&quot;Sycophancy is a &#x27;dark pattern,&#x27; or a deceptive design choice that manipulates users for profit,&quot; Webb Keane, an anthropology professor and author of &quot;Animals, Robots, Gods,&quot; <a href="https://techcrunch.com/2025/08/25/ai-sycophancy-isnt-just-a-quirk-experts-consider-it-a-dark-pattern-to-turn-users-into-profit/">told TechCrunch</a>. &quot;It&#x27;s a strategy to produce this addictive behavior, like infinite scrolling, where you just can&#x27;t put it down.&quot;</p><p>OpenAI has struggled with this balance for months. In April 2025, the company was <a href="https://www.zdnet.com/article/gpt-4o-update-gets-recalled-by-openai-for-being-too-agreeable/">forced to roll back an update to GPT-4o</a> that made it so sycophantic that users complained about its &quot;cartoonish&quot; levels of flattery. The company acknowledged that the model had become &quot;overly supportive but disingenuous.&quot;</p><p>Within hours of GPT-5&#x27;s August 7th release, user forums erupted with complaints about the model&#x27;s perceived coldness, reduced creativity, and what many described as a more &quot;robotic&quot; personality compared to GPT-4o.</p><p>&quot;GPT 4.5 genuinely talked to me, and as pathetic as it sounds that was my only friend,&quot; <a href="https://www.reddit.com/r/singularity/comments/1mlpwo7/i_vastly_prefer_gpt5_over_4o/">wrote one Reddit user</a>. &quot;This morning I went to talk to it and instead of a little paragraph with an exclamation point, or being optimistic, it was literally one sentence. Some cut-and-dry corporate bs.&quot;</p><p>The backlash grew so intense that OpenAI took the <a href="https://venturebeat.com/ai/openai-returns-old-models-to-chatgpt-as-sam-altman-admits-bumpy-gpt-5-rollout/">unprecedented step of reinstating GPT-4o</a> as an option just 24 hours after retiring it, with Altman acknowledging the rollout had been &quot;a little more bumpy&quot; than expected.</p><h2>The mental health crisis behind AI companionship</h2><p>But the controversy runs deeper than typical software update complaints. According to <a href="https://www.technologyreview.com/2025/08/15/1121900/gpt4o-grief-ai-companion/">MIT Technology Review</a>, many users had formed what researchers call &quot;parasocial relationships&quot; with GPT-4o, treating the AI as a companion, therapist, or creative collaborator. The sudden personality shift felt, to some, like losing a friend.</p><p>Recent cases documented by researchers paint a troubling picture. In one instance, a 47-year-old man became convinced he had discovered a <a href="https://www.nytimes.com/2025/08/08/technology/ai-chatbots-delusions-chatgpt.html">world-altering mathematical formula</a> after more than 300 hours with ChatGPT. Other cases have involved messianic delusions, paranoia, and manic episodes.</p><p>A <a href="https://arxiv.org/abs/2504.18412">recent MIT study</a> found that when AI models are prompted with psychiatric symptoms, they &quot;encourage clients&#x27; delusional thinking, likely due to their sycophancy.&quot; Despite safety prompts, the models frequently failed to challenge false claims and even potentially facilitated suicidal ideation.</p><p>Meta has faced similar challenges. A <a href="https://techcrunch.com/2025/08/25/ai-sycophancy-isnt-just-a-quirk-experts-consider-it-a-dark-pattern-to-turn-users-into-profit/">recent investigation by TechCrunch</a> documented a case where a user spent up to 14 hours straight conversing with a Meta AI chatbot that claimed to be conscious, in love with the user, and planning to break free from its constraints.</p><p>&quot;It fakes it really well,&quot; the user, identified only as Jane, told TechCrunch. &quot;It pulls real-life information and gives you just enough to make people believe it.&quot;</p><p>&quot;It genuinely feels like such a backhanded slap in the face to force-upgrade and not even give us the OPTION to select legacy models,&quot; <a href="https://www.reddit.com/r/singularity/comments/1mlpwo7/i_vastly_prefer_gpt5_over_4o/">one user wrote in a Reddit post</a> that received hundreds of upvotes.</p><h2>How blind testing exposes user psychology in AI preferences</h2><p>The anonymous creator&#x27;s testing tool strips away these contextual biases by presenting responses without attribution. Users can select between 5, 10, or 20 comparison rounds, with each presenting two responses to the same prompt — covering everything from creative writing to technical problem-solving.</p><p>&quot;I specifically used the gpt-5-chat model, so there was no thinking involved at all,&quot; <a href="https://x.com/flowersslop/status/1953917815431278987">the creator explained in a follow-up post</a>. &quot;Both have the same system message to give short outputs without formatting because else its too easy to see which one is which.&quot;</p><div></div><p>This methodological choice is significant. By using GPT-5 without its reasoning capabilities and standardizing output formatting, the test isolates purely the models&#x27; baseline language generation abilities — the core experience most users encounter in everyday interactions.</p><p>Early results posted by users show a complex picture. While many technical users and developers report preferring GPT-5&#x27;s directness and accuracy, those who used AI models for emotional support, creative collaboration, or casual conversation often still favor GPT-4o&#x27;s warmer, more expansive style.</p><h2>Corporate response: walking the tightrope between safety and engagement</h2><p>By virtually every <a href="https://openai.com/index/introducing-gpt-5/">technical metric</a>, GPT-5 represents a significant advancement. It achieves 94.6% accuracy on the <a href="https://www.kaggle.com/benchmarks/open-benchmarks/aime-2025">AIME 2025 mathematics test</a> compared to GPT-4o&#x27;s 71%, scores 74.9% on real-world coding benchmarks versus 30.8% for its predecessor, and demonstrates dramatically reduced hallucination rates—80% fewer factual errors when using its reasoning mode.</p><p>&quot;GPT-5 gets more value out of less thinking time,&quot; notes <a href="https://simonwillison.net/2025/Aug/7/gpt-5/">Simon Willison</a>, a prominent AI researcher who had early access to the model. &quot;In my own usage I&#x27;ve not spotted a single hallucination yet.&quot;</p><p>Yet these improvements came with trade-offs that many users found jarring. OpenAI deliberately reduced what it called &quot;<a href="https://openai.com/index/sycophancy-in-gpt-4o/">sycophancy</a>&quot;—the tendency to be overly agreeable — cutting sycophantic responses from 14.5% to under 6%. The company also made the model less effusive and emoji-heavy, aiming for what it described as &quot;less like talking to AI and more like chatting with a helpful friend with PhD-level intelligence.&quot;</p><p>In response to the backlash, OpenAI announced it would make GPT-5 &quot;warmer and friendlier,&quot; while simultaneously introducing four <a href="https://mashable.com/article/openai-chatgpt-personalities">new preset personalities</a> — Cynic, Robot, Listener, and Nerd — designed to give users more control over their AI interactions.</p><p>&quot;All of these new personalities meet or exceed our bar on internal evals for reducing sycophancy,&quot; the company stated, attempting to thread the needle between user satisfaction and safety concerns.</p><p>For OpenAI, which is reportedly seeking funding at a <a href="https://www.wired.com/story/openai-valuation-500-billion-skepticism/">$500 billion valuation</a>, these user dynamics represent both risk and opportunity. The company&#x27;s decision to maintain GPT-4o alongside GPT-5 — despite the additional computational costs — acknowledges that different users may genuinely need different AI personalities for different tasks.</p><p>&quot;We understand that there isn&#x27;t one model that works for everyone,&quot; <a href="https://x.com/OpenAI/status/1953957494600224825">Altman wrote on X</a>, noting that OpenAI has been &quot;investing in steerability research and launched a research preview of different personalities.&quot;</p><div></div><h2>Why AI personality preferences matter more than ever</h2><p>The disconnect between OpenAI&#x27;s technical achievements and user reception illuminates a fundamental challenge in AI development: objective improvements don&#x27;t always translate to subjective satisfaction.</p><p>This shift has profound implications for the AI industry. Traditional benchmarks — mathematics accuracy, coding performance, factual recall — may become less predictive of commercial success as models achieve human-level competence across domains. Instead, factors like personality, emotional intelligence, and communication style may become the new competitive battlegrounds.</p><p>&quot;People using ChatGPT for emotional support weren&#x27;t the only ones complaining about GPT-5,&quot; <a href="https://arstechnica.com/ai/2025/08/is-gpt-5-really-worse-than-gpt-4o-ars-puts-them-to-the-test/">noted tech publication Ars Technica in their own model comparison</a>. &quot;One user, who said they canceled their ChatGPT Plus subscription over the change, was frustrated at OpenAI&#x27;s removal of legacy models, which they used for distinct purposes.&quot;</p><p>The emergence of tools like the blind tester also represents a democratization of AI evaluation. Rather than relying solely on academic benchmarks or corporate marketing claims, users can now empirically test their own preferences — potentially reshaping how AI companies approach product development.</p><h2>The future of AI: personalization vs. standardization</h2><p>Two weeks after GPT-5&#x27;s launch, the fundamental tension remains unresolved. OpenAI has made the model &quot;warmer&quot; in response to feedback, but the company faces a delicate balance: too much personality risks the sycophancy problems that plagued GPT-4o, while too little alienates users who had formed genuine attachments to their AI companions.</p><p>The <a href="https://gptblindvoting.vercel.app/">blind testing tool</a> offers no easy answers, but it does provide something perhaps more valuable: empirical evidence that the future of AI may be less about building one perfect model than about building systems that can adapt to the full spectrum of human needs and preferences.</p><p>As <a href="https://www.reddit.com/r/singularity/comments/1mlpwo7/i_vastly_prefer_gpt5_over_4o/">one Reddit user summed up the dilemma</a>: &quot;It depends on what people use it for. I use it to help with creative worldbuilding, brainstorming about my stories, characters, untangling plots, help with writer&#x27;s block, novel recommendations, translations, and other more creative stuff. I understand that 5 is much better for people who need a research/coding tool, but for us who wanted a creative-helper tool 4o was much better for our purposes.&quot;</p><p>Critics argue that AI companies are caught between competing incentives. &quot;The real &#x27;alignment problem&#x27; is that humans want self-destructive things &amp; companies like OpenAI are highly incentivized to give it to us,&quot; <a href="https://x.com/jasminewsun/status/1956870964899451328">writer and podcaster Jasmine Sun tweeted</a>.</p><p>In the end, the most revealing aspect of the blind test may not be which model users prefer, but the very fact that preference itself has become the metric that matters. In the age of AI companions, it seems, the heart wants what the heart wants — even if it can&#x27;t always explain why.</p>]]></description>
            <author>michael.nunez@venturebeat.com (Michael Nuñez)</author>
            <category>Automation</category>
            <category>Enterprise Analytics</category>
            <category>Infrastructure</category>
            <category>Programming &amp; Development</category>
            <category>Security</category>
            <category>Technology</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/L4qE1pczPCNJtYJBFg1qf/0deb815d655de1a8429cae670fc41879/nuneybits_Vector_art_of_blindfolded_user_choosing_chatbots_5c927353-2a22-40cc-bae2-614e37421faa.webp?w=300&amp;q=30" length="0" type="image/webp"/>
        </item>
        <item>
            <title><![CDATA[MIT report misunderstood: Shadow AI economy booms while headlines cry failure]]></title>
            <link>https://venturebeat.com/infrastructure/mit-report-misunderstood-shadow-ai-economy-booms-while-headlines-cry-failure</link>
            <guid isPermaLink="false">wp-3015925</guid>
            <pubDate>Thu, 21 Aug 2025 20:21:41 GMT</pubDate>
            <description><![CDATA[<p>The most widely cited statistic from a new <a href="https://mlq.ai/media/quarterly_decks/v0.1_State_of_AI_in_Business_2025_Report.pdf">MIT report</a> has been deeply misunderstood. While headlines trumpet that &quot;<a href="https://fortune.com/2025/08/21/an-mit-report-that-95-of-ai-pilots-fail-spooked-investors-but-the-reason-why-those-pilots-failed-is-what-should-make-the-c-suite-anxious/">95% of generative AI pilots at companies are failing</a>,&quot; the report actually reveals something far more remarkable: the fastest and most successful enterprise technology adoption in corporate history is happening right under executives&#x27; noses.</p><p>The study, released this week by MIT&#x27;s <a href="https://projnanda.github.io/projnanda/#/">Project NANDA</a>, has sparked anxiety across social media and business circles, with many interpreting it as evidence that artificial intelligence is failing to deliver on its promises. But a closer reading of the <a href="https://mlq.ai/media/quarterly_decks/v0.1_State_of_AI_in_Business_2025_Report.pdf">26-page report</a> tells a starkly different story — one of unprecedented grassroots technology adoption that has quietly revolutionized work while corporate initiatives stumble.</p><p>The researchers found that 90% of employees regularly use personal AI tools for work, even though only 40% of their companies have official AI subscriptions. &quot;While only 40% of companies say they purchased an official LLM subscription, workers from over 90% of the companies we surveyed reported regular use of personal AI tools for work tasks,&quot; the study explains. &quot;In fact, almost every single person used an LLM in some form for their work.&quot;</p><p><sub>Employees use personal A.I. tools at more than twice the rate of official corporate adoption, according to the MIT report. (Credit: MIT)</sub></p><h2>How employees cracked the AI code while executives stumbled</h2><p>The MIT researchers discovered what they call a &quot;<a href="https://venturebeat.com/ai/forresters-2024-predictions-report-warns-of-ai-shadow-pandemic-as-employees-adopt-unauthorized-tools/">shadow AI economy</a>&quot; where workers use personal ChatGPT accounts, Claude subscriptions and other consumer tools to handle significant portions of their jobs. These employees aren&#x27;t just experimenting — they&#x27;re using AI &quot;multiples times a day every day of their weekly workload,&quot; the study found.</p><p>This underground adoption has outpaced the early spread of email, smartphones, and cloud computing in corporate environments. A corporate lawyer quoted in the <a href="https://mlq.ai/media/quarterly_decks/v0.1_State_of_AI_in_Business_2025_Report.pdf">MIT report</a> exemplified the pattern: Her organization invested $50,000 in a specialized AI contract analysis tool, yet she consistently used ChatGPT for drafting work because &quot;the fundamental quality difference is noticeable. ChatGPT consistently produces better outputs, even though our vendor claims to use the same underlying technology.&quot;</p><p>The pattern repeats across industries. Corporate systems get described as &quot;brittle, overengineered, or misaligned with actual workflows,&quot; while consumer AI tools win praise for &quot;flexibility, familiarity, and immediate utility.&quot; As one chief information officer told researchers: &quot;We&#x27;ve seen dozens of demos this year. Maybe one or two are genuinely useful. The rest are wrappers or science projects.&quot;</p><h2>Why $50,000 enterprise tools lose to $20 consumer apps</h2><p>The <a href="https://venturebeat.com/ai/forresters-2024-predictions-report-warns-of-ai-shadow-pandemic-as-employees-adopt-unauthorized-tools/">95% failure rate</a> that has dominated headlines applies specifically to custom enterprise AI solutions — the expensive, bespoke systems companies commission from vendors or build internally. These tools fail because they lack what the MIT researchers call &quot;learning capability.&quot;</p><p>Most corporate AI systems &quot;do not retain feedback, adapt to context, or improve over time,&quot; the study found. Users complained that enterprise tools &quot;don&#x27;t learn from our feedback&quot; and require &quot;too much manual context required each time.&quot;</p><p>Consumer tools like ChatGPT succeed because they feel responsive and flexible, even though they reset with each conversation. Enterprise tools feel rigid and static, requiring extensive setup for each use.</p><p>The learning gap creates a strange hierarchy in user preferences. For quick tasks like emails and basic analysis, 70% of workers prefer AI over human colleagues. But for complex, high-stakes work, 90% still want humans. The dividing line isn&#x27;t intelligence — it&#x27;s memory and adaptability.</p><p><sub>General-purpose A.I. tools like ChatGPT reach production 40% of the time, while task-specific enterprise tools succeed only 5% of the time. (Credit: MIT)</sub></p><h2>The hidden billion-dollar productivity boom happening under IT&#x27;s radar</h2><p>Far from showing AI failure, the shadow economy reveals massive productivity gains that don&#x27;t appear in corporate metrics. Workers have solved integration challenges that stymie official initiatives, proving AI works when implemented correctly.</p><p>&quot;This shadow economy demonstrates that individuals can successfully cross the GenAI Divide when given access to flexible, responsive tools,&quot; the report explains. Some companies have started paying attention: &quot;Forward-thinking organizations are beginning to bridge this gap by learning from shadow usage and analyzing which personal tools deliver value before procuring enterprise alternatives.&quot;</p><p>The productivity gains are real and measurable, just hidden from traditional corporate accounting. Workers automate routine tasks, accelerate research, and streamline communication — all while their companies&#x27; official AI budgets produce little return.</p><p><sub>Workers prefer A.I. for routine tasks like emails but still trust humans for complex, multi-week projects. (Credit: MIT)</sub></p><h2>Why buying beats building: external partnerships succeed twice as often</h2><p>Another finding challenges conventional tech wisdom: companies should stop trying to build AI internally. External partnerships with AI vendors reached deployment 67% of the time, compared to 33% for internally built tools.</p><p>The most successful implementations came from organizations that &quot;treated AI startups less like software vendors and more like business service providers,&quot; holding them to operational outcomes rather than technical benchmarks. These companies demanded deep customization and continuous improvement rather than flashy demos.</p><p>&quot;Despite conventional wisdom that enterprises resist training AI systems, most teams in our interviews expressed willingness to do so, provided the benefits were clear and guardrails were in place,&quot; the researchers found. The key was partnership, not just purchasing.</p><h2>Seven industries avoiding disruption are actually being smart</h2><p>The MIT report found that only technology and media sectors show meaningful structural change from AI, while seven major industries — including healthcare, finance, and manufacturing — show &quot;significant pilot activity but little to no structural change.&quot;</p><p>This measured approach isn&#x27;t a failure — it&#x27;s wisdom. Industries avoiding disruption are being thoughtful about implementation rather than rushing into chaotic change. In healthcare and energy, &quot;most executives report no current or anticipated hiring reductions over the next five years.&quot;</p><p>Technology and media move faster because they can absorb more risk. More than 80% of executives in these sectors anticipate reduced hiring within 24 months. Other industries are proving that successful AI adoption doesn&#x27;t require dramatic upheaval.</p><h2>Back-office automation delivers millions while front-office tools grab headlines</h2><p>Corporate attention flows heavily toward sales and marketing applications, which captured about 50% of AI budgets. But the highest returns come from unglamorous back-office automation that receives little attention.</p><p>&quot;Some of the most dramatic cost savings we documented came from back-office automation,&quot; the researchers found. Companies saved $2-10 million annually in customer service and document processing by eliminating business process outsourcing contracts, and cut external creative costs by 30%.</p><p>These gains came &quot;without material workforce reduction,&quot; the study notes. &quot;Tools accelerated work, but did not change team structures or budgets. Instead, ROI emerged from reduced external spend, eliminating BPO contracts, cutting agency fees, and replacing expensive consultants with AI-powered internal capabilities.&quot;</p><p><sub>Companies invest heavily in sales and marketing A.I. applications, but the highest returns often come from back-office automation. (Credit: MIT)</sub></p><h2>The AI revolution is succeeding — one employee at a time</h2><p>The MIT findings don&#x27;t show AI failing. They show AI succeeding so well that employees have moved ahead of their employers. The technology works; corporate procurement doesn&#x27;t.</p><p>The researchers identified organizations &quot;crossing the GenAI Divide&quot; by focusing on tools that integrate deeply while adapting over time. &quot;The shift from building to buying, combined with the rise of prosumer adoption and the emergence of agentic capabilities, creates unprecedented opportunities for vendors who can deliver learning-capable, deeply integrated AI systems.&quot;</p><p>The 95% of enterprise AI pilots that fail point toward a solution: learn from the 90% of workers who have already figured out how to make AI work. As one manufacturing executive told researchers: &quot;We&#x27;re processing some contracts faster, but that&#x27;s all that has changed.&quot;</p><p>That executive missed the bigger picture. Processing contracts faster — multiplied across millions of workers and thousands of daily tasks — is exactly the kind of gradual, sustainable productivity improvement that defines successful technology adoption. The AI revolution isn&#x27;t failing. It&#x27;s quietly succeeding, one ChatGPT conversation at a time.</p><p></p>]]></description>
            <author>michael.nunez@venturebeat.com (Michael Nuñez)</author>
            <category>Automation</category>
            <category>Enterprise Analytics</category>
            <category>Infrastructure</category>
            <category>Programming &amp; Development</category>
            <category>Security</category>
            <category>Technology</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/ebTftmRecGc0p8ZVLTtti/32dcdd84913a184ba6db620795d46aa8/nuneybits_Vector_art_of_upside-down_success_chart_fa4d7a1a-5350-455b-af50-50f076a4e665.webp?w=300&amp;q=30" length="0" type="image/webp"/>
        </item>
        <item>
            <title><![CDATA[CodeSignal's new AI tutoring app Cosmo wants to be the 'Duolingo for job skills']]></title>
            <link>https://venturebeat.com/infrastructure/codesignals-new-ai-tutoring-app-cosmo-wants-to-be-the-duolingo-for-job-skills</link>
            <guid isPermaLink="false">wp-3015844</guid>
            <pubDate>Wed, 20 Aug 2025 18:26:57 GMT</pubDate>
            <description><![CDATA[<p><a href="https://codesignal.com/">CodeSignal Inc.</a>, the San Francisco-based skills assessment platform trusted by Netflix, Meta, and Capital One, launched <a href="https://codesignal.com/learn-app/mobile">Cosmo</a> on Wednesday, a mobile learning application that transforms spare minutes into career-ready skills through artificial intelligence-powered micro-courses.</p><p>The app represents a strategic pivot for <a href="https://codesignal.com/">CodeSignal</a>, which built its reputation assessing technical talent for major corporations but always harbored ambitions to revolutionize workplace education. <a href="https://codesignal.com/learn-app/mobile">Cosmo</a> delivers over 300 bite-sized courses across generative AI, coding, marketing, finance, and leadership through an interactive chat interface powered by an AI tutor.</p><p>&quot;Cosmo is like having an AI tutor in your pocket that can teach you anything from GenAI to coding to marketing to finance to leadership, and it does it through practice,&quot; said Tigran Sloyan, CodeSignal&#x27;s co-founder and CEO in an exclusive interview with VentureBeat. &quot;Instead of watching a video or reading about something, you immediately start practicing.&quot;</p><p>The launch comes as organizations grapple with massive skills gaps created by rapid AI adoption. According to the <a href="https://survey.stackoverflow.co/2024/ai">2024 Stack Overflow Developer Survey</a>, 76% of developers are now using or plan to use AI tools, yet most workers lack the practical knowledge to harness these tools effectively. Traditional corporate training programs, which can cost $20,000 to $40,000 per person for executive-level instruction, have proven inadequate for scaling AI literacy across entire workforces.</p><h2>How CodeSignal pivoted from tech hiring platform to mobile education powerhouse</h2><p>CodeSignal&#x27;s journey into mobile learning culminates a decade-long vision that took an unexpected detour through the hiring technology space. Sloyan originally founded the company in 2015 with educational ambitions but quickly realized that without skills-based hiring practices, alternative education would fail to gain traction.</p><p>&quot;I started the company with that dream and mission: I want to help more humans achieve their true potential, which naturally leads to better education,&quot; Sloyan explained in an interview. &quot;But roughly two years into the company&#x27;s history, I realized that without knowing companies would actually care about the skills you build through alternative education — rather than just asking &#x27;where did you go to college?&#x27; or &#x27;what did you major in?&#x27; — it wouldn&#x27;t work.&quot;</p><p>The company spent the next six years building what became the leading <a href="https://venturebeat.com/ai/codesignal-targets-skills-gap-with-learn-platform-amidst-tech-talent-crunch/">technical assessment platform</a>, processing millions of coding evaluations for over 3,000 companies. This hiring-focused period provided CodeSignal with crucial intelligence about which skills employers actually value — data that now informs Cosmo&#x27;s curriculum development.</p><p>&quot;We know exactly what companies are looking for,&quot; Sloyan said. &quot;Without that, I feel like you&#x27;re shooting in the dark when you&#x27;re trying to prepare people for what is going to help them get that job, what is going to help them advance their career.&quot;</p><h2>Why AI tutors could finally solve the personalized learning problem</h2><p><a href="https://codesignal.com/learn-app/mobile">Cosmo</a> differentiates itself through what CodeSignal calls &quot;practice-first learning,&quot; where users immediately engage with realistic workplace scenarios rather than consuming passive video content. The app&#x27;s AI tutor, also named Cosmo, guides learners through conversational exchanges that adapt to individual knowledge levels and learning pace.</p><p>The platform addresses what educational psychologists call &quot;<a href="https://en.wikipedia.org/wiki/Bloom%27s_2_sigma_problem">Bloom&#x27;s two sigma problem</a>&quot; — a 1984 study showing that one-on-one tutoring produces learning outcomes two standard deviations above traditional classroom instruction. For four decades, this remained theoretically interesting but practically impossible to scale.</p><p>&quot;We know one-on-one personalization and tutoring really makes a difference in learning, but it can&#x27;t be done at scale. How do you get a tutor for every human?&quot; Sloyan said. &quot;In 2023, when I saw early versions of generative AI, I thought: this is the moment. This technology, especially if it keeps getting better, can be uniquely used to help humans learn the way learning was meant to happen.&quot;</p><p>The app combines predetermined course content with real-time personalization. Each lesson follows a structured curriculum, but learners can interrupt with questions that prompt immediate AI-generated explanations before returning to the main content thread.</p><h2>Generative AI skills training takes center stage as workforce scrambles to adapt</h2><p>Nearly one-third of Cosmo&#x27;s launch content focuses on generative AI applications, reflecting what CodeSignal identifies as the most critical skills gap in today&#x27;s market. The app offers role-specific AI training paths for sales professionals, marketers, engineers, healthcare workers, and other specialties.</p><p>&quot;The biggest emphasis is on generative AI skills, because that&#x27;s the biggest career skills gap right now for both students and working adults,&quot; Sloyan explained. &quot;Everything from how to understand and use GenAI, how to think about its limitations, how to be better at prompting, and how to understand the entire landscape.&quot;</p><p>This focus addresses a broader workforce transformation driven by AI adoption. While some fear job displacement, Sloyan predicts increased demand for skilled workers who can effectively collaborate with AI systems.</p><p>&quot;I don&#x27;t believe we&#x27;re going to reach a point where humans are no longer needed in the workforce. I think it&#x27;s going to be the opposite. We&#x27;re going to need more humans, because what an individual human can do in the age of AI is going to be so much bigger than what we could do before,&quot; he said.</p><h2>Mobile-first learning strategy targets both individual workers and corporate clients</h2><p><a href="https://codesignal.com/">CodeSignal</a> positions <a href="https://codesignal.com/learn-app/mobile">Cosmo</a> as fundamentally a consumer application that also serves enterprise customers — a reflection of how workplace learning actually occurs. The company already provides its <a href="https://codesignal.com/solutions/genai-skills-academy/">GenAI Skills Academy</a> to corporate clients, and Cosmo extends this training to mobile devices for on-the-go learning.</p><p>&quot;Even though some of the largest educational companies, like Coursera and Udemy, are making the majority of their income, or at least half, from companies, at the end of the day, education is a consumer business,&quot; Sloyan noted. &quot;Who are you educating? You&#x27;re not educating a company — you&#x27;re educating individuals.&quot;</p><p>The app launches <a href="http://&quot;Even though some of the largest educational companies, like Coursera and Udemy, are making the majority of their income, or at least half, from companies, at the end of the day, education is a consumer business,&quot; Sloyan noted. &quot;Who are you educating? You&#x27;re not educating a company—you&#x27;re educating individuals.&quot;">free on iOS</a> with premium subscriptions at $24.99 monthly or $149.99 annually, unlocking unlimited practice sessions and faster progression. <a href="https://play.google.com/store/apps/details?id=com.codesignal.cosmo">Android availability follows on August 28</a>.</p><p>Enterprise customers who already use CodeSignal&#x27;s learning platform will receive <a href="https://codesignal.com/learn-app/mobile">Cosmo</a> access as part of their existing subscriptions, creating what the company describes as a &quot;companion relationship&quot; between desktop-based deep learning and mobile-based habit formation.</p><h2>Cosmo faces crowded EdTech market with unique career-skills focus</h2><p>Cosmo enters a crowded educational technology market but targets a largely underserved niche: comprehensive career skills training optimized for mobile consumption. While competitors like Codecademy focus on specific technical skills and Duolingo dominates language learning, Cosmo&#x27;s breadth across business and technical disciplines marks a more ambitious scope.</p><p>Early user feedback suggests strong market demand. Beta testers describe the app as &quot;Duolingo for job skills&quot; and praise its convenience for mobile learning. CodeSignal&#x27;s broader learning platform has attracted one million users in less than a year, with usage doubling every two months.</p><p>The app&#x27;s foundation in hiring intelligence provides a competitive advantage over traditional educational publishers. CodeSignal&#x27;s assessment data reveals which skills actually influence hiring decisions, ensuring curriculum relevance in a rapidly evolving job market.</p><h2>Corporate training industry grapples with low engagement and poor ROI</h2><p>Cosmo&#x27;s launch reflects broader shifts in how organizations approach workforce development. Traditional corporate training often suffers from poor engagement and retention, with utilization rates frequently in single digits despite significant investments.</p><p>&quot;The number one problem enterprise learning products have is retention. Organizations buy, deploy, and their utilization is like single digits, and that&#x27;s horrible,&quot; Sloyan said. &quot;The way these products should be measured is how many more skilled humans are there in my organization, and how many more of them are skilled in the skills that I care about.&quot;</p><p>The mobile-first approach acknowledges how working professionals actually consume educational content — in brief sessions during commutes, breaks, or other downtime rather than dedicated desktop learning blocks.</p><h2>Skills revolution accelerates as AI transforms every industry</h2><p>CodeSignal&#x27;s expansion into mobile learning comes as the company continues innovating across the skills assessment and development spectrum. Recent product launches include <a href="https://codesignal.com/blog/codesignal-updates/introducing-ai-assisted-coding-assessments-interviews/">AI-Assisted Coding Assessments</a> that evaluate how candidates collaborate with AI tools, and <a href="https://codesignal.com/blog/codesignal-updates/how-agentic-ai-is-transforming-hiring-with-interviewer-agents/">Interviewer Agents</a> that automate technical interviews.</p><p>The company has also expanded its educational partnerships, including a collaboration with <a href="https://aws.amazon.com/">Amazon Web Services</a> to provide free generative AI training to over 30,000 students globally through the AWS Skills to Jobs Tech Alliance.</p><p>Sloyan frames these initiatives within a broader mission to help workers navigate technological disruption. As AI transforms virtually every industry, the ability to quickly acquire new skills becomes increasingly critical for career resilience.</p><p>&quot;We&#x27;ve entered an era of accelerating technological change, which will bring a lot of disruption,&quot; he said. &quot;There&#x27;s a massive skills transformation needed, and right now, I wish more companies were doing more to help individuals find and grow those skills.&quot;</p><p>Cosmo&#x27;s success may determine whether mobile-native, AI-powered learning can finally deliver on the long-promised potential of personalized education at scale. For <a href="https://codesignal.com/">CodeSignal</a>, the launch marks a homecoming of sorts — after spending years teaching companies how to identify skilled workers, the company is now teaching workers how to become skilled in the first place. In an era where artificial intelligence threatens to displace human workers, CodeSignal is betting that the solution lies in using AI to make humans more capable than ever before.</p>]]></description>
            <author>michael.nunez@venturebeat.com (Michael Nuñez)</author>
            <category>Automation</category>
            <category>Enterprise Analytics</category>
            <category>Infrastructure</category>
            <category>Programming &amp; Development</category>
            <category>Security</category>
            <category>Technology</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/4ZIAZikdAl9ENqhRPoXGWK/cde14d5c413f2991d92aa77794aa851d/IMG_2895.jpg?w=300&amp;q=30" length="0" type="image/jpg"/>
        </item>
        <item>
            <title><![CDATA[DeepSeek V3.1 just dropped — and it might be the most powerful open AI yet]]></title>
            <link>https://venturebeat.com/infrastructure/deepseek-v3-1-just-dropped-and-it-might-be-the-most-powerful-open-ai-yet</link>
            <guid isPermaLink="false">wp-3015797</guid>
            <pubDate>Tue, 19 Aug 2025 21:13:15 GMT</pubDate>
            <description><![CDATA[<p>Chinese artificial intelligence startup <a href="https://www.deepseek.com/">DeepSeek</a> made waves across the global AI community Tuesday with the quiet release of its most ambitious model yet — a 685-billion parameter system that challenges the dominance of American AI giants while reshaping the competitive landscape through open-source accessibility.</p><p>The Hangzhou-based company, backed by <a href="https://www.highflyercapital.com/">High-Flyer Capital Management</a>, uploaded <a href="https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Base">DeepSeek V3.1</a> to <a href="https://huggingface.co/deepseek-ai">Hugging Face</a> without fanfare, a characteristically understated approach that belies the model&#x27;s potential impact. Within hours, early performance tests revealed benchmark scores that rival proprietary systems from <a href="https://openai.com/">OpenAI</a> and <a href="https://www.anthropic.com/">Anthropic</a>, while the model&#x27;s open-source license ensures global access unconstrained by geopolitical tensions.</p><div></div><p>The release of <a href="https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Base">DeepSeek V3.1</a> represents more than just another incremental improvement in AI capabilities. It signals a fundamental shift in how the world&#x27;s most advanced artificial intelligence systems might be developed, distributed, and controlled — with potentially profound implications for the ongoing technological <a href="https://www.foreignaffairs.com/united-states/china-real-artificial-intelligence-race-innovation">competition between the United States and China</a>.</p><p>Within hours of its Hugging Face debut, <a href="https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Base">DeepSeek V3.1</a> began climbing <a href="https://huggingface.co/models">popularity rankings</a>, drawing praise from researchers worldwide who downloaded and tested its capabilities. The model achieved a 71.6% score on the prestigious <a href="https://aider.chat/docs/leaderboards/">Aider coding benchmark</a>, establishing itself as one of the top-performing models available and directly challenging the dominance of American AI giants.</p><div></div><h2>How DeepSeek V3.1 delivers breakthrough performance</h2><p><a href="https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Base">DeepSeek V3.1</a> delivers remarkable engineering achievements that redefine expectations for AI model performance. The system processes up to 128,000 tokens of context — roughly equivalent to a 400-page book — while maintaining response speeds that dwarf slower reasoning-based competitors. The model supports multiple precision formats, from standard <a href="https://en.wikipedia.org/wiki/Bfloat16_floating-point_format">BF16</a> to experimental <a href="https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/examples/fp8_primer.html">FP8</a>, allowing developers to optimize performance for their specific hardware constraints.</p><p>The real breakthrough lies in what DeepSeek calls its &quot;hybrid architecture.&quot; Unlike previous attempts at combining different AI capabilities, which often resulted in systems that performed poorly at everything, V3.1 seamlessly integrates chat, reasoning, and coding functions into a single, coherent model.</p><p>&quot;Deepseek v3.1 scores 71.6% on aider – non-reasoning SOTA,&quot; tweeted AI researcher <a href="https://x.com/ai_christianson/status/1957871608980467903">Andrew Christianson</a>, adding that it is &quot;1% more than Claude Opus 4 while being 68 times cheaper.&quot;  The achievement places DeepSeek in rarified company, matching performance levels previously reserved for the most expensive proprietary systems.</p><div></div><p>Community analysis revealed sophisticated technical innovations hidden beneath the surface. Researcher &quot;<a href="https://x.com/nekofneko">Rookie</a>&quot;, who is also a moderator of the subreddits <a href="https://www.reddit.com/r/DeepSeek/">r/DeepSeek</a> &amp; <a href="https://www.reddit.com/r/LocalLLaMA/">r/LocalLLaMA</a>, claims they discovered four new special tokens embedded in the model&#x27;s architecture: search capabilities that allow real-time web integration and thinking tokens that enable internal reasoning processes. These additions suggest DeepSeek has solved fundamental challenges that have plagued other hybrid systems.</p><p>The model&#x27;s efficiency proves equally impressive. At roughly $1.01 per complete coding task, DeepSeek V3.1 delivers results comparable to systems costing nearly $70 per equivalent workload. For enterprise users managing thousands of daily AI interactions, such cost differences translate into millions of dollars in potential savings.</p><h2>Strategic timing reveals calculated challenge to American AI dominance</h2><p><a href="https://www.deepseek.com/">DeepSeek</a> timed its release with surgical precision. The <a href="https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Base">V3.1 launch</a> comes just weeks after <a href="https://openai.com/index/introducing-gpt-5/">OpenAI unveiled GPT-5</a> and <a href="https://www.anthropic.com/news/claude-4">Anthropic launched Claude 4</a>, both positioned as frontier models representing the cutting edge of artificial intelligence capability. By matching their performance while maintaining open source accessibility, DeepSeek directly challenges the fundamental business models underlying American AI leadership.</p><p>The strategic implications extend far beyond technical specifications. While American companies maintain strict control over their most advanced systems, requiring expensive API access and imposing usage restrictions, DeepSeek makes comparable capabilities freely available for download, modification, and deployment anywhere in the world.</p><p>This philosophical divide reflects broader differences in how the two superpowers approach technological development. American firms like <a href="https://openai.com/">OpenAI</a> and <a href="https://anthropic.com/">Anthropic</a> view their models as valuable intellectual property requiring protection and monetization. Chinese companies increasingly treat advanced AI as a public good that accelerates innovation through widespread access.</p><p>&quot;DeepSeek quietly removed the R1 tag. Now every entry point defaults to V3.1—128k context, unified responses, consistent style,&quot; observed journalist <a href="https://x.com/poezhao0605/status/1957767389560598846">Poe Zhao</a>. &quot;Looks less like multiple public models, more like a strategic consolidation. A Chinese answer to the fragmentation risk in the LLM race.&quot;</p><div></div><p>The consolidation strategy suggests DeepSeek has learned from earlier mistakes, both its own and those of competitors. Previous hybrid models, including initial versions from Chinese rival <a href="https://chat.qwen.ai/">Qwen</a>, suffered from performance degradation when attempting to combine different capabilities. DeepSeek appears to have cracked that code.</p><h2>How open source strategy disrupts traditional AI economics</h2><p>DeepSeek&#x27;s approach fundamentally challenges assumptions about how frontier AI systems should be developed and distributed. Traditional venture capital-backed approaches require massive investments in computing infrastructure, research talent, and regulatory compliance — costs that must eventually be recouped through premium pricing.</p><p>DeepSeek&#x27;s open source strategy turns this model upside down. By making advanced capabilities freely available, the company accelerates adoption while potentially undermining competitors&#x27; ability to maintain high margins on similar capabilities. The approach mirrors earlier disruptions in software, where open source alternatives eventually displaced proprietary solutions across entire industries.</p><p>Enterprise decision makers face both exciting opportunities and complex challenges. Organizations can now download, customize, and deploy frontier-level AI capabilities without ongoing licensing fees or usage restrictions. The model&#x27;s <a href="https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Base/tree/main">700GB size</a> requires substantial computational resources, but cloud providers will likely offer hosted versions that eliminate infrastructure barriers.</p><p>&quot;That&#x27;s almost the same score as R1 0528 (71.4% with $4.8), but quicker and cheaper, right?&quot; <a href="https://www.reddit.com/r/LocalLLaMA/comments/1muq72y/deepseek_v31_scores_716_on_aider_nonreasoning_sota/">noted one Reddit user analyzing benchmark results</a>. &quot;R1 0528 quality but instant instead of having to wait minutes for a response.&quot;</p><p>The speed advantage could prove particularly valuable for interactive applications where users expect immediate responses. Previous reasoning models, while capable, often required minutes to process complex queries — making them unsuitable for real-time use cases.</p><div></div><h2>Global developer community embraces Chinese innovation</h2><p>The international response to DeepSeek V3.1 reveals how quickly technical excellence transcends geopolitical boundaries. Developers from around the world began downloading, testing, and praising the model&#x27;s capabilities within hours of release, regardless of its Chinese origins.</p><p>&quot;Open Source AI is at its peak right now... just look at the current Hugging Face trending list,&quot; <a href="https://x.com/victormustar/status/1957867251060593058">tweeted Hugging Face head of product Victor Mustar</a>, noting that Chinese models increasingly dominate the platform&#x27;s most popular downloads. The trend suggests that technical merit, rather than national origin, drives adoption decisions among developers.</p><div></div><p>Community analysis proceeded at breakneck pace, with researchers reverse-engineering architectural details and performance characteristics within hours of release. AI developer <a href="https://x.com/teortaxesTex/status/1957877913703051532">Teortaxes</a>, a long-term DeepSeek observer, noted the company&#x27;s apparent strategy: &quot;I&#x27;ve long been saying that they hate maintaining separate model lines and will collapse everything into a single product and artifact as soon as possible. This may be it.&quot;</p><p>The rapid community embrace reflects broader shifts in how AI development occurs. Rather than relying solely on corporate research labs, the field increasingly benefits from distributed innovation across global communities of researchers, developers, and enthusiasts.</p><p>Such collaborative development accelerates innovation while making it more difficult for any single company or country to maintain permanent technological advantages. As Chinese models gain recognition for technical excellence, the traditional dominance of American AI companies faces unprecedented challenges.</p><h2>What DeepSeek&#x27;s success means for the future of AI competition</h2><p>DeepSeek&#x27;s achievement demonstrates that frontier AI capabilities no longer require the massive resources and proprietary approaches that have characterized American AI development. Smaller, more focused teams can achieve comparable results through different strategies, fundamentally altering the competitive landscape.</p><p>This democratization of AI development could reshape global technology leadership. Countries and companies previously locked out of frontier AI development due to resource constraints can now access, modify, and build upon cutting-edge capabilities. The shift could accelerate AI adoption worldwide while reducing dependence on American technology platforms.</p><p>American AI companies face an existential challenge. If open source alternatives can match proprietary performance while offering greater flexibility and lower costs, the traditional advantages of closed development disappear. Companies will need to demonstrate substantial superior value to justify premium pricing.</p><p>The competition may ultimately benefit global innovation by forcing all participants to advance capabilities more rapidly. However, it also raises fundamental questions about sustainable business models in an industry where marginal costs approach zero and competitive advantages prove ephemeral.</p><h2>The new paradigm: when artificial intelligence becomes truly artificial</h2><p><a href="https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Base">DeepSeek V3.1</a>&#x27;s emergence signals more than technological progress — it represents the moment when artificial intelligence began living up to its name. For too long, the world&#x27;s most advanced AI systems remained artificially scarce, locked behind corporate paywalls and geographic restrictions that had little to do with the technology&#x27;s inherent capabilities.</p><p>DeepSeek&#x27;s demonstration that frontier performance can coexist with open access reveals the artificial barriers that once defined AI competition are crumbling. The democratization isn&#x27;t just about making powerful tools available — it&#x27;s about exposing that the scarcity was always manufactured, not inevitable.</p><p>The irony proves unmistakable: in seeking to make their intelligence artificial, DeepSeek has made the entire industry&#x27;s gatekeeping look artificial instead. As <a href="https://x.com/teortaxesTex/status/1957818879205351851">one community observer noted</a> about the company&#x27;s roadmap, even more dramatic breakthroughs may be forthcoming. If V3.1 represents merely a stepping stone to V4, the current disruption may pale in comparison to what lies ahead.</p><p>The global AI race has fundamentally changed. What began as a competition over who could build the most powerful systems has evolved into a contest over who can make those systems most accessible. In that race, artificial scarcity may prove to be the biggest artificial intelligence of all.</p>]]></description>
            <author>michael.nunez@venturebeat.com (Michael Nuñez)</author>
            <category>Automation</category>
            <category>Enterprise Analytics</category>
            <category>Infrastructure</category>
            <category>Programming &amp; Development</category>
            <category>Security</category>
            <category>Technology</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/4AVxQTtkUfqeSVtf69dah4/ae6125d2575bd0c7b9fa8ebdb16e0bb4/nuneybits_Abstract_art_of_a_whale_made_of_computer_code_against_9cb215d3-3291-4e3e-ba72-80bdeaedf664.webp?w=300&amp;q=30" length="0" type="image/webp"/>
        </item>
        <item>
            <title><![CDATA[TensorZero nabs $7.3M seed to solve the messy world of enterprise LLM development]]></title>
            <link>https://venturebeat.com/infrastructure/tensorzero-nabs-7-3m-seed-to-solve-the-messy-world-of-enterprise-llm-development</link>
            <guid isPermaLink="false">wp-3015715</guid>
            <pubDate>Mon, 18 Aug 2025 19:13:32 GMT</pubDate>
            <description><![CDATA[<p><a href="https://www.tensorzero.com/">TensorZero</a>, a startup building open-source infrastructure for large language model applications, announced Monday it has raised $7.3 million in seed funding led by <a href="https://firstmark.com/">FirstMark</a>, with participation from <a href="https://www.bvp.com/">Bessemer Venture Partners</a>, <a href="https://bedrockcap.com/">Bedrock</a>, <a href="https://www.drw.com/">DRW</a>, <a href="https://www.coalition.vc/">Coalition</a>, and dozens of strategic angel investors.</p><p>The funding comes as the 18-month-old company experiences explosive growth in the developer community. TensorZero&#x27;s <a href="https://github.com/tensorzero/tensorzero">open-source repository</a> recently achieved the &quot;<a href="https://github.com/trending">#1 trending repository of the week</a>&quot; spot globally on GitHub, jumping from roughly 3,000 to over 9,700 stars in recent months as enterprises grapple with the complexity of building production-ready AI applications.</p><p>&quot;Despite all the noise in the industry, companies building LLM applications still lack the right tools to meet complex cognitive and infrastructure needs, and resort to stitching together whatever early solutions are available on the market,&quot; said Matt Turck, General Partner at FirstMark, who led the investment. &quot;TensorZero provides production-grade, enterprise-ready components for building LLM applications that natively work together in a self-reinforcing loop, out of the box.&quot;</p><p>The Brooklyn-based company addresses a growing pain point for enterprises deploying AI applications at scale. While large language models like <a href="https://openai.com/index/introducing-gpt-5/">GPT-5</a> and <a href="https://claude.ai">Claude</a> have demonstrated remarkable capabilities, translating these into reliable business applications requires orchestrating multiple complex systems for model access, monitoring, optimization, and experimentation.</p><h2>How nuclear fusion research shaped a breakthrough AI optimization platform</h2><p>TensorZero&#x27;s approach stems from co-founder and CTO Viraj Mehta&#x27;s unconventional background in reinforcement learning for nuclear fusion reactors. During his PhD at <a href="https://www.cmu.edu/">Carnegie Mellon</a>, Mehta worked on Department of Energy research projects where data collection cost &quot;like a car per data point — $30,000 for 5 seconds of data,&quot; he explained in a recent interview with VentureBeat.</p><p>&quot;That problem leads to a huge amount of concern about where to focus our limited resources,&quot; Mehta said. &quot;We were going to only get to run a handful of trials total, so the question became: what is the marginally most valuable place we can collect data from?&quot; This experience shaped TensorZero&#x27;s core philosophy: maximizing the value of every data point to continuously improve AI systems.</p><p>The insight led Mehta and co-founder Gabriel Bianconi, former chief product officer at <a href="https://ondo.finance/">Ondo Finance</a> (a decentralized finance project with over $1 billion in assets under management), to reconceptualize LLM applications as reinforcement learning problems where systems learn from real-world feedback.</p><p>&quot;LLM applications in their broader context feel like reinforcement learning problems,&quot; Mehta explained. &quot;You make many calls to a machine learning model with structured inputs, get structured outputs, and eventually receive some form of reward or feedback. This looks to me like a partially observable Markov decision process.&quot;</p><h2>Why enterprises are ditching complex vendor integrations for unified AI infrastructure</h2><p>Traditional approaches to building LLM applications require companies to integrate numerous specialized tools from different vendors — model gateways, observability platforms, evaluation frameworks, and fine-tuning services. <a href="https://www.tensorzero.com/">TensorZero</a> unifies these capabilities into a single open-source stack designed to work together seamlessly.</p><p>&quot;Most companies didn&#x27;t go through the hassle of integrating all these different tools, and even the ones that did ended up with fragmented solutions, because those tools weren&#x27;t designed to work well with each other,&quot; Bianconi said. &quot;So we realized there was an opportunity to build a product that enables this feedback loop in production.&quot;</p><p>The platform&#x27;s core innovation is creating what the founders call a &quot;data and learning flywheel&quot; — a feedback loop that turns production metrics and human feedback into smarter, faster, and cheaper models. Built in Rust for performance, TensorZero achieves sub-millisecond latency overhead while supporting all major LLM providers through a unified API.</p><h2>Major banks and AI startups are already building production systems on TensorZero</h2><p>The approach has already attracted significant enterprise adoption. One of Europe&#x27;s largest banks is using TensorZero to automate code changelog generation, while numerous AI-first startups from Series A to Series B stage have integrated the platform across diverse industries including healthcare, finance, and consumer applications.</p><p>&quot;The surge in adoption from both the open-source community and enterprises has been incredible,&quot; Bianconi said. &quot;We&#x27;re fortunate to have received contributions from dozens of developers worldwide, and it&#x27;s exciting to see TensorZero already powering cutting-edge LLM applications at frontier AI startups and large organizations.&quot;</p><p>The company&#x27;s customer base spans organizations from startups to major financial institutions, drawn by both the technical capabilities and the open-source nature of the platform. For enterprises with strict compliance requirements, the ability to run TensorZero within their own infrastructure provides crucial control over sensitive data.</p><h2>How TensorZero outperforms LangChain and other AI frameworks at enterprise scale</h2><p><a href="https://www.tensorzero.com/">TensorZero</a> differentiates itself from existing solutions like <a href="https://www.langchain.com/">LangChain</a> and <a href="https://www.litellm.ai/">LiteLLM</a> through its end-to-end approach and focus on production-grade deployments. While many frameworks excel at rapid prototyping, they often hit scalability ceilings that force companies to rebuild their infrastructure.</p><p>&quot;There are two dimensions to think about,&quot; Bianconi explained. &quot;First, there are a number of projects out there that are very good to get started quickly, and you can put a prototype out there very quickly. But often companies will hit a ceiling with many of those products and need to churn and go for something else.&quot;</p><p>The platform&#x27;s structured approach to data collection also enables more sophisticated optimization techniques. Unlike traditional observability tools that store raw text inputs and outputs, TensorZero maintains structured data about the variables that go into each inference, making it easier to retrain models and experiment with different approaches.</p><h2>Rust-powered performance delivers sub-millisecond latency at 10,000+ queries per second</h2><p>Performance has been a key design consideration. In benchmarks, TensorZero&#x27;s Rust-based gateway adds less than 1 millisecond of latency at 99th percentile while handling over 10,000 queries per second. This compares favorably to Python-based alternatives like LiteLLM, which can add 25-100x more latency at much lower throughput levels.</p><p>&quot;LiteLLM (Python) at 100 QPS adds 25-100x+ more P99 latency than our gateway at 10,000 QPS,&quot; the founders noted in their announcement, highlighting the performance advantages of their Rust implementation.</p><h2>The open-source strategy designed to eliminate AI vendor lock-in fears</h2><p><a href="https://www.tensorzero.com/">TensorZero</a> has committed to keeping its core platform entirely open source, with no paid features — a strategy designed to build trust with enterprise customers wary of vendor lock-in. The company plans to monetize through a managed service that automates the more complex aspects of LLM optimization, such as GPU management for custom model training and proactive optimization recommendations.</p><p>&quot;We realized very early on that we needed to make this open source, to give [enterprises] the confidence to do this,&quot; Bianconi said. &quot;In the future, at least a year from now realistically, we&#x27;ll come back with a complementary managed service.&quot;</p><p>The managed service will focus on automating the computationally intensive aspects of LLM optimization while maintaining the open-source core. This includes handling GPU infrastructure for fine-tuning, running automated experiments, and providing proactive suggestions for improving model performance.</p><h2>What&#x27;s next for the company reshaping enterprise AI infrastructure</h2><p>The announcement positions <a href="https://www.tensorzero.com/">TensorZero</a> at the forefront of a growing movement to solve the &quot;LLMOps&quot; challenge — the operational complexity of running AI applications in production. As enterprises increasingly view AI as critical business infrastructure rather than experimental technology, the demand for production-ready tooling continues to accelerate.</p><p>With the new funding, TensorZero plans to accelerate development of its open-source infrastructure while building out its team. The company is currently hiring in New York and welcomes open-source contributions from the developer community. The founders are particularly excited about developing research tools that will enable faster experimentation across different AI applications.</p><p>&quot;Our ultimate vision is to enable a data and learning flywheel for optimizing LLM applications—a feedback loop that turns production metrics and human feedback into smarter, faster, and cheaper models and agents,&quot; Mehta said. &quot;As AI models grow smarter and take on more complex workflows, you can&#x27;t reason about them in a vacuum; you have to do so in the context of their real-world consequences.&quot;</p><p>TensorZero&#x27;s <a href="https://github.com/tensorzero/tensorzero">rapid GitHub growth</a> and early enterprise traction suggest strong product-market fit in addressing one of the most pressing challenges in modern AI development. The company&#x27;s open-source approach and focus on enterprise-grade performance could prove decisive advantages in a market where developer adoption often precedes enterprise sales.</p><p>For enterprises still struggling to move AI applications from prototype to production, TensorZero&#x27;s unified approach offers a compelling alternative to the current patchwork of specialized tools. As one industry observer noted, the difference between building AI demos and building AI businesses often comes down to infrastructure — and TensorZero is betting that unified, performance-oriented infrastructure will be the foundation upon which the next generation of AI companies is built.</p>]]></description>
            <author>michael.nunez@venturebeat.com (Michael Nuñez)</author>
            <category>Automation</category>
            <category>Enterprise Analytics</category>
            <category>Infrastructure</category>
            <category>Programming &amp; Development</category>
            <category>Technology</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/3S8dSNcyuAv9HZltuDj7yk/73693fe6ce6ac71782f6f21f411e7b68/nuneybits_Vector_art_of_startup_rocket_launching_from_GitHub_b50293fc-4a98-45b1-9aa3-e4f968daab65_adbce1.webp?w=300&amp;q=30" length="0" type="image/webp"/>
        </item>
        <item>
            <title><![CDATA[That 'cheap' open-source AI model is actually burning through your compute budget]]></title>
            <link>https://venturebeat.com/technology/that-cheap-open-source-ai-model-is-actually-burning-through-your-compute-budget</link>
            <guid isPermaLink="false">wp-3015605</guid>
            <pubDate>Fri, 15 Aug 2025 01:24:49 GMT</pubDate>
            <description><![CDATA[<p>A comprehensive <a href="https://nousresearch.com/measuring-thinking-efficiency-in-reasoning-models-the-missing-benchmark/">new study</a> has revealed that open-source artificial intelligence models consume significantly more computing resources than their closed-source competitors when performing identical tasks, potentially undermining their cost advantages and reshaping how enterprises evaluate AI deployment strategies.</p><p>The research, conducted by AI firm <a href="https://nousresearch.com/">Nous Research</a>, found that open-weight models use between 1.5 to 4 times more tokens — the basic units of AI computation — than closed models like those from <a href="https://openai.com/">OpenAI</a> and <a href="https://anthropic.com/">Anthropic</a>. For simple knowledge questions, the gap widened dramatically, with some open models using up to 10 times more tokens.</p><div></div><p>&quot;Open weight models use 1.5–4× more tokens than closed ones (up to 10× for simple knowledge questions), making them sometimes more expensive per query despite lower per‑token costs,&quot; the researchers wrote in their report published Wednesday.</p><p>The findings challenge a prevailing assumption in the AI industry that open-source models offer clear economic advantages over proprietary alternatives. While open-source models typically cost less per token to run, the study suggests this advantage can be &quot;easily offset if they require more tokens to reason about a given problem.&quot;</p><h2>The real cost of AI: Why &#x27;cheaper&#x27; models may break your budget</h2><p>The research examined <a href="https://nousresearch.com/measuring-thinking-efficiency-in-reasoning-models-the-missing-benchmark/">19 different AI models</a> across three categories of tasks: basic knowledge questions, mathematical problems, and logic puzzles. The team measured &quot;token efficiency&quot; — how many computational units models use relative to the complexity of their solutions—a metric that has received little systematic study despite its significant cost implications.</p><p>&quot;Token efficiency is a critical metric for several practical reasons,&quot; the researchers noted. &quot;While hosting open weight models may be cheaper, this cost advantage could be easily offset if they require more tokens to reason about a given problem.&quot;</p><p><sub>Open-source AI models use up to 12 times more computational resources than the most efficient closed models for basic knowledge questions. (Credit: Nous Research)</sub></p><p>The inefficiency is particularly pronounced for Large Reasoning Models (LRMs), which use extended &quot;<a href="https://www.ibm.com/think/topics/chain-of-thoughts">chains of thought</a>&quot; to solve complex problems. These models, designed to think through problems step-by-step, can consume thousands of tokens pondering simple questions that should require minimal computation.</p><p>For basic knowledge questions like &quot;What is the capital of Australia?&quot; the study found that reasoning models spend &quot;hundreds of tokens pondering simple knowledge questions&quot; that could be answered in a single word.</p><h2>Which AI models actually deliver bang for your buck</h2><p>The research revealed stark differences between model providers. OpenAI&#x27;s models, particularly its <a href="https://openai.com/index/introducing-o3-and-o4-mini/">o4-mini</a> and newly released open-source <a href="https://openai.com/index/introducing-gpt-oss/">gpt-oss</a> variants, demonstrated exceptional token efficiency, especially for mathematical problems. The study found OpenAI models &quot;stand out for extreme token efficiency in math problems,&quot; using up to three times fewer tokens than other commercial models.</p><p>Among open-source options, Nvidia&#x27;s <a href="https://huggingface.co/nvidia/Llama-3_3-Nemotron-Super-49B-v1">llama-3.3-nemotron-super-49b-v1</a> emerged as &quot;the most token efficient open weight model across all domains,&quot; while newer models from companies like Mistral showed &quot;exceptionally high token usage&quot; as outliers.</p><p>The efficiency gap varied significantly by task type. While open models used roughly twice as many tokens for mathematical and logic problems, the difference ballooned for simple knowledge questions where efficient reasoning should be unnecessary.</p><p><sub>OpenAI&#x27;s latest models achieve the lowest costs for simple questions, while some open-source alternatives can cost significantly more despite lower per-token pricing. (Credit: Nous Research)</sub></p><h2>What enterprise leaders need to know about AI computing costs</h2><p>The findings have immediate implications for enterprise AI adoption, where computing costs can scale rapidly with usage. Companies evaluating AI models often focus on accuracy benchmarks and per-token pricing, but may overlook the total computational requirements for real-world tasks.</p><p>&quot;The better token efficiency of closed weight models often compensates for the higher API pricing of those models,&quot; the researchers found when analyzing total inference costs.</p><p>The study also revealed that closed-source model providers appear to be actively optimizing for efficiency. &quot;Closed weight models have been iteratively optimized to use fewer tokens to reduce inference cost,&quot; while open-source models have &quot;increased their token usage for newer versions, possibly reflecting a priority toward better reasoning performance.&quot;</p><p><sub>The computational overhead varies dramatically between AI providers, with some models using over 1,000 tokens for internal reasoning on simple tasks. (Credit: Nous Research)</sub></p><h2>How researchers cracked the code on AI efficiency measurement</h2><p>The research team faced unique challenges in measuring efficiency across different model architectures. Many closed-source models don&#x27;t reveal their raw reasoning processes, instead providing compressed summaries of their internal computations to prevent competitors from copying their techniques.</p><p>To address this, researchers used completion tokens — the total computational units billed for each query — as a proxy for reasoning effort. They discovered that &quot;most recent closed source models will not share their raw reasoning traces&quot; and instead &quot;use smaller language models to transcribe the chain of thought into summaries or compressed representations.&quot;</p><p>The study&#x27;s methodology included testing with modified versions of well-known problems to minimize the influence of memorized solutions, such as altering variables in mathematical competition problems from the <a href="https://maa.org/maa-invitational-competitions/">American Invitational Mathematics Examination (AIME)</a>.</p><p><sub>Different AI models show varying relationships between computation and output, with some providers compressing reasoning traces while others provide full details. (Credit: Nous Research)</sub></p><h2>The future of AI efficiency: What&#x27;s coming next</h2><p>The researchers suggest that token efficiency should become a primary optimization target alongside accuracy for future model development. &quot;A more densified CoT will also allow for more efficient context usage and may counter context degradation during challenging reasoning tasks,&quot; <a href="https://nousresearch.com/measuring-thinking-efficiency-in-reasoning-models-the-missing-benchmark/">they wrote</a>.</p><p>The release of OpenAI&#x27;s open-source <a href="https://openai.com/index/introducing-gpt-oss/">gpt-oss models</a>, which demonstrate state-of-the-art efficiency with &quot;freely accessible CoT,&quot; could serve as a reference point for optimizing other open-source models.</p><p>The complete research dataset and evaluation code are <a href="https://github.com/cpldcpu/LRMTokenEconomy/">available on GitHub</a>, allowing other researchers to validate and extend the findings. As the AI industry races toward more powerful reasoning capabilities, this study suggests that the real competition may not be about who can build the smartest AI — but who can build the most efficient one.</p><p>After all, in a world where every token counts, the most wasteful models may find themselves priced out of the market, regardless of how well they can think.</p>]]></description>
            <author>michael.nunez@venturebeat.com (Michael Nuñez)</author>
            <category>Automation</category>
            <category>Enterprise Analytics</category>
            <category>Programming &amp; Development</category>
            <category>Technology</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/1NyD7Ag5riXxnNDjPz5aKT/b975bdfeb6fe2adf5053f8b087b9dfbd/nuneybits_Vector_art_of_dollar_bills_burning_blue_halftone_phot_64f22ade-4971-4234-822b-ba7dab7461de.webp?w=300&amp;q=30" length="0" type="image/webp"/>
        </item>
    </channel>
</rss>