<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0">
    <channel>
        <title>Dev | VentureBeat</title>
        <link>https://venturebeat.com/category/dev/feed/</link>
        <description>Transformative tech coverage that matters</description>
        <lastBuildDate>Sat, 04 Apr 2026 14:59:14 GMT</lastBuildDate>
        <docs>https://validator.w3.org/feed/docs/rss2.html</docs>
        <generator>https://github.com/jpmonette/feed</generator>
        <language>en</language>
        <copyright>Copyright 2026, VentureBeat</copyright>
        <item>
            <title><![CDATA[OpenAI debuts GPT‑5.1-Codex-Max coding model and it already completed a 24-hour task internally]]></title>
            <link>https://venturebeat.com/technology/openai-debuts-gpt-5-1-codex-max-coding-model-and-it-already-completed-a-24</link>
            <guid isPermaLink="false">qmhAiqqbDOrYa2qQWGaMm</guid>
            <pubDate>Wed, 19 Nov 2025 19:26:00 GMT</pubDate>
            <description><![CDATA[<p>OpenAI has <a href="https://openai.com/index/gpt-5-1-codex-max/"><b>introduced GPT‑5.1-Codex-Max</b></a>, a new frontier agentic coding model now available in its Codex developer environment. The release marks a significant step forward in AI-assisted software engineering, offering improved long-horizon reasoning, efficiency, and real-time interactive capabilities. GPT‑5.1-Codex-Max will now replace GPT‑5.1-Codex as the default model across Codex-integrated surfaces.</p><p>The new model is designed to serve as a persistent, high-context software development agent, capable of managing complex refactors, debugging workflows, and project-scale tasks across multiple context windows.</p><p>It comes on the heels of<a href="https://venturebeat.com/ai/google-unveils-gemini-3-claiming-the-lead-in-math-science-multimodal-and"> Google releasing its powerful new Gemini 3 Pro model</a> yesterday, yet still outperforms or matches it on key coding benchmarks: </p><p>On <b>SWE-Bench Verified</b>, <b>GPT‑5.1-Codex-Max achieved 77.9% accuracy</b> at extra-high reasoning effort, edging past Gemini 3 Pro’s 76.2%. </p><p>It also led on <b>Terminal-Bench 2.0, with 58.1% accuracy versus Gemini’s 54.2%, </b>and matched Gemini’s score of 2,439 on LiveCodeBench Pro, a competitive coding Elo benchmark.</p><p>When measured against Gemini 3 Pro’s most advanced configuration — its Deep Thinking model — Codex-Max holds a slight edge in agentic coding benchmarks, as well. </p><h3><b>Performance Benchmarks: Incremental Gains Across Key Tasks</b></h3><p>GPT‑5.1-Codex-Max demonstrates measurable improvements over GPT‑5.1-Codex across a range of standard software engineering benchmarks. </p><p>On SWE-Lancer IC SWE, it achieved 79.9% accuracy, a significant increase from GPT‑5.1-Codex’s 66.3%. In SWE-Bench Verified (n=500), it reached 77.9% accuracy at extra-high reasoning effort, outperforming GPT‑5.1-Codex’s 73.7%.</p><p>Performance on Terminal Bench 2.0 (n=89) showed more modest improvements, with GPT‑5.1-Codex-Max achieving 58.1% accuracy compared to 52.8% for GPT‑5.1-Codex. </p><p>All evaluations were run with compaction and extra-high reasoning effort enabled.</p><p>These results indicate that the new model offers a higher ceiling on both benchmarked correctness and real-world usability under extended reasoning loads.</p><h3><b>Technical Architecture: Long-Horizon Reasoning via Compaction</b></h3><p>A major architectural improvement in GPT‑5.1-Codex-Max is its ability to reason effectively over extended input-output sessions using a mechanism called <b>compaction</b>. </p><p>This enables the model to retain key contextual information while discarding irrelevant details as it nears its context window limit — effectively allowing for continuous work across millions of tokens without performance degradation.</p><p>The model has been internally observed to complete tasks lasting more than 24 hours, including multi-step refactors, test-driven iteration, and autonomous debugging.</p><p>Compaction also improves token efficiency. At medium reasoning effort, GPT‑5.1-Codex-Max used approximately 30% fewer thinking tokens than GPT‑5.1-Codex for comparable or better accuracy, which has implications for both cost and latency.</p><h3><b>Platform Integration and Use Cases</b></h3><p>GPT‑5.1-Codex-Max is currently available across multiple Codex-based environments, which refer to OpenAI’s own integrated tools and interfaces built specifically for code-focused AI agents. These include:</p><ul><li><p><b>Codex CLI</b>, OpenAI’s official command-line tool (@openai/codex), where GPT‑5.1-Codex-Max is already live.</p></li><li><p><b>IDE extensions</b>, likely developed or maintained by OpenAI, though no specific third-party IDE integrations were named.</p></li><li><p><b>Interactive coding environments</b>, such as those used to demonstrate frontend simulation apps like CartPole or Snell’s Law Explorer.</p></li><li><p><b>Internal code review tooling</b>, used by OpenAI’s engineering teams.</p></li></ul><p>For now, GPT‑5.1-Codex-Max is not yet available via public API, though OpenAI states this is coming soon. Users who wish to work with the model in terminal environments today can do so by installing and using the Codex CLI.</p><p>It is not currently confirmed whether or how the model will integrate into third-party IDEs unless they are built on top of the CLI or future API.</p><p>The model is capable of interacting with live tools and simulations. Examples shown in the release include:</p><ul><li><p>An interactive CartPole policy gradient simulator, which visualizes reinforcement learning training and activations.</p></li><li><p>A Snell’s Law optics explorer, supporting dynamic ray tracing across refractive indices.</p></li></ul><p>These interfaces exemplify the model’s ability to reason in real time while maintaining an interactive development session — effectively bridging computation, visualization, and implementation within a single loop.</p><h3><b>Cybersecurity and Safety Constraints</b></h3><p>While GPT‑5.1-Codex-Max does not meet OpenAI’s “High” capability threshold for cybersecurity under its Preparedness Framework, it is currently the most capable cybersecurity model OpenAI has deployed. It supports use cases such as automated vulnerability detection and remediation, but with strict sandboxing and disabled network access by default.</p><p>OpenAI reports no increase in scaled malicious use but has introduced enhanced monitoring systems, including activity routing and disruption mechanisms for suspicious behavior. Codex remains isolated to a local workspace unless developers opt-in to broader access, mitigating risks like prompt injection from untrusted content.</p><h3><b>Deployment Context and Developer Usage</b></h3><p>GPT‑5.1-Codex-Max is currently available to users on <b>ChatGPT Plus, Pro, Business, Edu, and Enterprise</b> plans. It will also become the new default in Codex-based environments, replacing GPT‑5.1-Codex, which was a more general-purpose model.</p><p>OpenAI states that 95% of its internal engineers use Codex weekly, and since adoption, these engineers have shipped ~70% more pull requests on average — highlighting the tool’s impact on internal development velocity.</p><p>Despite its autonomy and persistence, OpenAI stresses that Codex-Max should be treated as a coding assistant, not a replacement for human review. The model produces terminal logs, test citations, and tool call outputs to support transparency in generated code.</p><h3><b>Outlook</b></h3><p>GPT‑5.1-Codex-Max represents a significant evolution in OpenAI’s strategy toward agentic development tools, offering greater reasoning depth, token efficiency, and interactive capabilities across software engineering tasks. By extending its context management and compaction strategies, the model is positioned to handle tasks at the scale of full repositories, rather than individual files or snippets.</p><p>With continued emphasis on agentic workflows, secure sandboxes, and real-world evaluation metrics, Codex-Max sets the stage for the next generation of AI-assisted programming environments — while underscoring the importance of oversight in increasingly autonomous systems.</p>]]></description>
            <author>carl.franzen@venturebeat.com (Carl Franzen)</author>
            <category>Dev</category>
            <category>Technology</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/1EELcUCWai7BHVe2Iti1Ht/1fef9a3dd13eeed757dfda0bdccc4403/lsRydt-rLj3bqzrXKMGoN_b9a12412710e4c468f59b1b66cb31312__1_.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
        <item>
            <title><![CDATA[The most important OpenAI announcement you probably missed at DevDay 2025]]></title>
            <link>https://venturebeat.com/infrastructure/the-most-important-openai-announcement-you-probably-missed-at-devday-2025</link>
            <guid isPermaLink="false">5vuObQXLYhA0fkYU6uaFO9</guid>
            <pubDate>Thu, 09 Oct 2025 07:00:00 GMT</pubDate>
            <description><![CDATA[<p>OpenAI’s annual developer conference on Monday was a spectacle of ambitious AI product launches, from an <a href="https://openai.com/index/introducing-apps-in-chatgpt/"><u>app store for ChatGPT</u></a> to a stunning <a href="https://openai.com/index/sora-2/"><u>video-generation API</u></a> that brought creative concepts to life. But for the enterprises and technical leaders watching closely, the most consequential announcement was the quiet <a href="https://openai.com/index/codex-now-generally-available/"><u>general availability of Codex</u></a>, the company&#x27;s AI software engineer. This release signals a profound shift in how software—and by extension, modern business—is built.</p><p>While other announcements captured the public’s imagination, the production-ready release of <a href="https://openai.com/codex/"><u>Codex</u></a>, supercharged by a <a href="https://openai.com/index/introducing-upgrades-to-codex/"><u>new specialized model</u></a> and a <a href="https://developers.openai.com/codex/sdk"><u>suite of enterprise-grade tools</u></a>, is the engine behind OpenAI’s entire vision. It is the tool that builds the tools, the proven agent in a world buzzing with agentic potential, and the clearest articulation of the company&#x27;s strategy to win the enterprise.</p><p>The <a href="https://openai.com/index/codex-now-generally-available/"><u>general availability of Codex</u></a> moves it from a &quot;research preview&quot; to a fully supported product, complete with a new <a href="https://developers.openai.com/codex/sdk"><u>software development kit (SDK)</u></a>, a <a href="https://developers.openai.com/codex/integrations/slack"><u>Slack integration</u></a>, and administrative controls for security and monitoring.This transition declares that Codex is ready for mission-critical work inside the world’s largest companies.</p><p>&quot;We think this is the best time in history to be a builder; it has never been faster to go from idea to product,&quot; said OpenAI CEO Sam Altman during the <a href="https://venturebeat.com/ai/openai-dev-day-2025-chatgpt-becomes-the-new-app-store-and-hardware-is-coming"><u>opening keynote</u></a> presentation. &quot;Software used to take months or years to build. You saw that it can take minutes now to build with AI.&quot; </p><p>That acceleration is not theoretical. It&#x27;s a reality born from OpenAI’s own internal use — a massive &quot;dogfooding&quot; effort that serves as the ultimate case study for enterprise customers.</p><h2>Inside GPT-5-Codex: The AI model that codes autonomously for hours and drives 70% productivity gains</h2><p>At the heart of the Codex upgrade is <a href="https://chatgpt.com/features/codex?utm_source=google&amp;utm_medium=paidsearch_brand&amp;utm_campaign=GOOG_B_SEM_GBR_Core_TEM_BAU_ACQ_PER_BRD_ALL_NAMER_US_EN_080625&amp;utm_term=openai%20codex&amp;utm_content=187611721873&amp;utm_ad=776422173331&amp;utm_match=p&amp;gad_source=1&amp;gad_campaignid=23071604080&amp;gbraid=0AAAAA-I0E5cl2krVAgcAc2VJzRhsB5CLd&amp;gclid=CjwKCAjwup3HBhAAEiwA7euZupI6JVfC5p76PsLMGm1i2XCDyp7ERZnjrRPbaodVfs6hE2MM_g4_zhoCzE4QAvD_BwE"><u>GPT-5-Codex</u></a>, a version of OpenAI&#x27;s latest flagship model that has been &quot;purposely trained for Codex and agentic coding.&quot; The new model is designed to function as an autonomous teammate, moving far beyond simple code autocompletion.</p><p>&quot;I personally like to think about it as a little bit like a human teammate,&quot; explained Tibo Sottiaux, an OpenAI engineer, during a technical session on Codex. &quot;You can pair a program with it on your computer, you can delegate to it, or as you&#x27;ll see, you can give it a job without explicit prompting.&quot;</p><p>This new model enables &quot;<a href="https://openai.com/index/introducing-upgrades-to-codex/"><u>adaptive thinking</u></a>,&quot; allowing it to dynamically adjust the time and computational effort spent on a task based on its complexity.For simple requests, it&#x27;s fast and efficient, but for complex refactoring projects, it can work for hours.</p><p>One engineer during the technical session noted, &quot;I&#x27;ve seen the GPT-5-Codex model work for over seven hours productively... on a marathon session.&quot; This capability to handle long-running, complex tasks is a significant leap beyond the simple, single-shot interactions that define most AI coding assistants.</p><p>The results inside OpenAI have been dramatic. The company reported that 92% of its technical staff now uses Codex daily, and those engineers complete 70% more pull requests (a measure of code contribution) each week. Usage has surged tenfold since August. </p><p>&quot;When we as a team see the stats, it feels great,&quot; Sottiaux shared. &quot;But even better is being at lunch with someone who then goes &#x27;Hey I use Codex all the time. Here&#x27;s a cool thing that I do with it. Do you want to hear about it?&#x27;&quot; </p><h2>How OpenAI uses Codex to build its own AI products and catch hundreds of bugs daily</h2><p>Perhaps the most compelling argument for Codex’s importance is that it is the foundational layer upon which OpenAI’s other flashy announcements were built. During the <a href="https://devday.openai.com/"><u>DevDay event</u></a>, the company showcased custom-built arcade games and a dynamic, AI-powered website for the conference itself, all developed using <a href="https://openai.com/codex/"><u>Codex</u></a>.</p><p>In one session, engineers demonstrated how they built &quot;Storyboard,&quot; a custom creative tool for the film industry, in just 48 hours during an internal hackathon. &quot;We decided to test Codex, our coding agent... we would send tasks to Codex in between meetings. We really easily reviewed and merged PRs into production, which Codex even allowed us to do from our phones,&quot; said Allison August, a solutions engineering leader at OpenAI. </p><p>This reveals a critical insight: the rapid innovation showcased at DevDay is a direct result of the productivity flywheel created by <a href="https://openai.com/codex/"><u>Codex</u></a>. The AI is a core part of the manufacturing process for all other AI products.</p><p>A key enterprise-focused feature is the new, more robust code review capability. OpenAI said it &quot;purposely trained GPT-5-Codex to be great at ultra thorough code review,&quot; enabling it to explore dependencies and validate a programmer&#x27;s intent against the actual implementation to find high-quality bugs.Internally, nearly every pull request at OpenAI is now reviewed by Codex, catching hundreds of issues daily before they reach a human reviewer.</p><p>&quot;It saves you time, you ship with more confidence,&quot; Sottiaux said. &quot;There&#x27;s nothing worse than finding a bug after we actually ship the feature.&quot; </p><h2>Why enterprise software teams are choosing Codex over GitHub Copilot for mission-critical development</h2><p>The maturation of <a href="https://openai.com/codex/"><u>Codex</u></a> is central to OpenAI’s broader strategy to conquer the enterprise market, a move essential to justifying its massive valuation and unprecedented compute expenditures. During a press conference, CEO Sam Altman confirmed the strategic shift.</p><p>&quot;The models are there now, and you should expect a huge focus from us on really winning enterprises with amazing products, starting here,&quot; Altman said during a private press conference. </p><p>OpenAI President and Co-founder Greg Brockman immediately added, &quot;And you can see it already with Codex, which I think has been just an incredible success and has really grown super fast.&quot; </p><p>For technical decision-makers, the message is clear. While consumer-facing agents that book dinner reservations are still finding their footing, <a href="https://openai.com/codex/"><u>Codex</u></a> is a proven enterprise agent delivering substantial ROI today. Companies like Cisco have already rolled out Codex to their engineering organizations, cutting code review times by 50% and reducing project timelines from weeks to days.</p><p>With the new <a href="https://developers.openai.com/codex/sdk/"><u>Codex SDK</u></a>, companies can now embed this agentic power directly into their own custom workflows, such as automating fixes in a CI/CD pipeline or even creating self-evolving applications. During a live demo, an engineer showcased a mobile app that updated its own user interface in real-time based on a natural language prompt, all powered by the embedded Codex SDK. </p><p>While the launch of an <a href="https://openai.com/index/introducing-apps-in-chatgpt/"><u>app ecosystem in ChatGPT</u></a> and the breathtaking visuals of the <a href="https://openai.com/index/sora-2/"><u>Sora 2 API</u></a> rightfully generated headlines, the <a href="https://openai.com/index/codex-now-generally-available/"><u>general availability of Codex</u></a> marks a more fundamental and immediate transformation. It is the quiet but powerful engine driving the next era of software development, turning the abstract promise of AI-driven productivity into a tangible, deployable reality for businesses today.</p>]]></description>
            <author>michael.nunez@venturebeat.com (Michael Nuñez)</author>
            <category>Automation</category>
            <category>Dev</category>
            <category>Enterprise</category>
            <category>Infrastructure</category>
            <category>Programming &amp; Development</category>
            <category>Technology</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/2buVKIMlQ2KFcFg8BPHoqU/c5d2dfbfd8cfe9a9432106a936fa204f/nuneybits_A_retro_glowing_computer_on_gradient_background_that__094dfc70-9906-4074-bb00-d32b04faf5f9-1.webp?w=300&amp;q=30" length="0" type="image/webp"/>
        </item>
        <item>
            <title><![CDATA[OpenAI Dev Day 2025: ChatGPT becomes the new app store — and hardware is coming]]></title>
            <link>https://venturebeat.com/technology/openai-dev-day-2025-chatgpt-becomes-the-new-app-store-and-hardware-is-coming</link>
            <guid isPermaLink="false">6na0pAdxl9xh4fN39O1ww5</guid>
            <pubDate>Tue, 07 Oct 2025 22:30:00 GMT</pubDate>
            <description><![CDATA[<p>In a packed hall at Fort Mason Center in San Francisco, against a backdrop of the Golden Gate Bridge, <a href="https://openai.com/"><u>OpenAI</u></a> CEO Sam Altman laid out a bold vision to remake the digital world. The company that brought generative AI to the mainstream with a simple chatbot is now building the foundations for its next act: a comprehensive computing platform designed to move beyond the screen and browser, with legendary designer Jony Ive enlisted to help shape its physical form.</p><p>At its <a href="https://openai.com/devday/"><u>third annual DevDay</u></a>, OpenAI unveiled a suite of tools that signals a strategic pivot from a model provider to a full-fledged ecosystem. The message was clear: the era of simply asking an AI questions is over. The future is about commanding AI to perform complex tasks, build software autonomously, and live inside every application, a transition Altman framed as moving from &quot;systems that you can ask anything to, to systems that you can ask to do anything for you.&quot; </p><p>The day’s announcements were a three-pronged assault on the status quo, targeting how users interact with software, how developers build it, and how businesses deploy intelligent agents. But it was the sessions held behind closed doors, away from the <a href="https://www.youtube.com/live/hS1YqcewH0c?si=mFE2rRx3QrK7z6NF"><u>public livestream</u></a>, that revealed the true scope of OpenAI’s ambition — a future that includes new hardware, a relentless pursuit of computational power, and a philosophical quest to redefine our relationship with technology.</p><h3><b>From chatbot to operating system: The new &#x27;App Store&#x27;</b></h3><p>The centerpiece of the public-facing keynote was the transformation of <a href="https://chatgpt.com/"><u>ChatGPT</u></a> itself. With the new <a href="https://openai.com/index/introducing-apps-in-chatgpt/"><u>Apps SDK</u></a>, OpenAI is turning its wildly popular chatbot into a dynamic, interactive platform, effectively an operating system where developers can build and distribute their own applications.</p><p>“Today, we&#x27;re going to open up ChatGPT for developers to build real apps inside of ChatGPT,” Altman announced during the keynote presentation to applause. “This will enable a new generation of apps that are interactive, adaptive and personalized, that you can chat with.”</p><p>Live demonstrations showcased apps from partners like <a href="https://www.coursera.org/"><u>Coursera</u></a>, <a href="https://www.canva.com/"><u>Canva</u></a>, and <a href="https://www.zillow.com/"><u>Zillow</u></a> running seamlessly within a chat conversation. A user could watch a machine learning lecture, ask <a href="https://chatgpt.com/"><u>ChatGPT</u></a> to explain a concept in real-time, and then use Canva to generate a poster based on the conversation, all without leaving the chat interface. The apps can render rich, interactive UIs, even going full-screen to offer a complete experience, like exploring a Zillow map of homes.</p><p>For developers, this represents a powerful new distribution channel. “When you build with the Apps SDK, your apps can reach hundreds of millions of chat users,” Altman said, highlighting a direct path to a massive user base that has grown to over <a href="https://venturebeat.com/ai/openai-announces-apps-sdk-allowing-chatgpt-to-launch-and-run-third-party"><u>800 million weekly active users</u></a>. </p><p>In a private press conference later, Nick Turley, head of ChatGPT, elaborated on the grander vision. &quot;We never meant to build a chatbot,&quot; he stated. &quot;When we set out to make ChatGPT, we meant to build a super assistant and we got a little sidetracked. And one of the tragedies of getting a little sidetracked is that we built a great chatbot, but we are the first ones to say that not all software needs to be a chatbot, not all interaction with the commercial world needs to be a chatbot.&quot;</p><p>Turley emphasized that while OpenAI is excited about natural language interfaces, &quot;the interface really needs to evolve, which is why you see so much UI in the demos today. In fact, you can even go full screen and chat is in the background.&quot; He described a future where users might &quot;start your day in ChatGPT, just because it kind of has become the de facto entry point into the commercial web and into a lot of software,&quot; but clarified that &quot;our incentive is not to keep you in. Our product is to allow other people to build amazing businesses on top and to evolve the form factor of software.&quot;</p><h3><b>The rise of the agents: Building the &#x27;do anything&#x27; AI</b></h3><p>If apps are about bringing the world into ChatGPT, the new &quot;<a href="https://openai.com/index/introducing-agentkit/"><u>Agent Kit</u></a>&quot; is about sending AI out into the world to get things done. OpenAI is providing a complete &quot;set of building blocks... to help you take agents from prototype to production,&quot; Altman explained in his keynote. </p><p><a href="https://openai.com/index/introducing-agentkit/"><u>Agent Kit</u></a> is an integrated development environment for creating autonomous AI workers. It features a visual canvas to design complex workflows, an embeddable chat interface (&quot;Chat Kit&quot;) for deploying agents in any app, and a sophisticated evaluation suite to measure and improve performance.</p><p>A compelling demo from financial operations platform <a href="https://ramp.com/"><u>Ramp</u></a> showed how Agent Kit was used to build a procurement agent. An employee could simply type, &quot;I need five more ChatGPT business seats,&quot; and the agent would parse the request, check it against company expense policies, find vendor details, and prepare a virtual credit card for the purchase — a process that once took weeks now completed in minutes. </p><p>This push into agents is a direct response to a growing enterprise need to move beyond AI as a simple information retrieval tool and toward AI as a productivity engine that automates complex business processes. Brad Lightcap, OpenAI&#x27;s COO, noted that for enterprise adoption, &quot;you needed this kind of shift to more agentic AI that could actually do things for you, versus just respond with text outputs.&quot; </p><h3><b>The future of code and the Jony Ive bBombshell</b></h3><p>Perhaps the most profound shift is occurring in software development itself. <a href="https://openai.com/index/codex-now-generally-available/"><u>Codex</u></a>, OpenAI&#x27;s AI coding agent, has graduated from a research preview to a full-fledged product, now powered by a specialized version of the new GPT-5 model. It is, as one speaker put it, &quot;a teammate that understands your context.&quot; </p><p>The capabilities are staggering. Developers can now assign <a href="https://openai.com/index/codex-now-generally-available/"><u>Codex</u></a> tasks directly from <a href="https://slack.com/"><u>Slack</u></a>, and the agent can autonomously write code, create pull requests, and even review other engineers&#x27; work on <a href="https://github.com/"><u>GitHub</u></a>. A live demo showed Codex taking a simple photo of a whiteboard sketch and turning it into a fully functional, beautifully designed mobile app screen. Another demo showed an app that could &quot;self-evolve,&quot; reprogramming itself in real-time based on a user&#x27;s natural language request. </p><p>But the day&#x27;s biggest surprise came in a closing fireside chat, which was not livestreamed, between <a href="https://openai.com/sam-and-jony/"><u>Altman and Jony Ive</u></a>, the iconic former chief design officer of Apple. The two revealed they have been collaborating for three years on a new family of AI-centric hardware.</p><p>Ive, whose design philosophy shaped the iPhone, iMac, and Apple Watch, said his creative team’s purpose &quot;became clear&quot; with the launch of ChatGPT. He argued that our current relationship with technology is broken and that AI presents an opportunity for a fundamental reset.</p><p>“I think it would be absurd to assume that you could have technology that is this breathtaking, delivered to us through legacy products, products that are decades old,” Ive said. “I see it as a chance to use this most remarkable capability to full-on address a lot of the overwhelm and despair that people feel right now.”</p><p>While details of the devices remain secret, Ive spoke of his motivation in deeply human terms. “We love our species, and we want to be useful. We think that humanity deserves much better than humanity generally is given,” he said. He emphasized the importance of &quot;care&quot; in the design process, stating, &quot;We sense when people have cared... you sense carelessness. You sense when somebody does not care about you, they care about money and schedule.&quot; </p><p>This collaboration confirms that OpenAI&#x27;s ambitions are not confined to the cloud; it is actively exploring the physical interface through which humanity will interact with its powerful new intelligence.</p><h3><b>The Unquenchable Thirst for Compute</b></h3><p>Underpinning this entire platform strategy is a single, overwhelming constraint: the availability of computing power. In both the private press conference and the un-streamed Developer State of the Union, OpenAI’s leadership returned to this theme again and again.</p><p>“The degree to which we are all constrained by compute... Everyone is just so constrained on being able to offer the services at the scale required to get the revenue that at this point, we&#x27;re quite confident we can push it pretty far,” Altman told reporters. He added that even with massive new hardware partnerships with AMD and others, &quot;we&#x27;ll be saying the same thing again. We&#x27;re so convinced... There&#x27;s so much more demand.&quot; </p><p>This explains the company’s aggressive, <a href="https://www.reuters.com/business/autos-transportation/companies-pouring-billions-advance-ai-infrastructure-2025-10-06/"><u>multi-billion-dollar investment in infrastructure</u></a>. When asked about profitability, Altman was candid that the company is in a phase of &quot;investment and growth.&quot; He invoked a famous quote from Walt Disney, paraphrasing, &quot;We make more money so we can make more movies.&quot; For OpenAI, the &quot;movies&quot; are ever-more-powerful AI models.</p><p>Greg Brockman, OpenAI’s President, put the ultimate goal in stark economic terms during the Developer State of the Union. &quot;AI is going to become, probably in the not too distant future, the fundamental driver of economic growth,&quot; he said. &quot;Asking ‘How much compute do you want?’ is a little bit like asking how much workforce do you want? The answer is, you can always get more out of more.&quot; </p><p>As the day concluded and developers mingled at the reception, the scale of OpenAI&#x27;s project came into focus. Fueled by new models like the powerful <a href="https://platform.openai.com/docs/models/gpt-5-pro"><u>GPT-5 Pro</u></a> and the stunning <a href="https://openai.com/index/sora-2/"><u>Sora 2</u></a> video generator, the company is no longer just building AI. It is building the world where AI will live — a world of intelligent apps, autonomous agents, and new physical devices, betting that in the near future, intelligence itself will be the ultimate platform.</p><p>
</p>]]></description>
            <author>michael.nunez@venturebeat.com (Michael Nuñez)</author>
            <category>Dev</category>
            <category>Programming &amp; Development</category>
            <category>Technology</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/5SsXpDqbiyZvm7DJbht3pv/765b6ae01f75c4e84f15e6f9f026bdf2/nuneybits_A_developer_working_on_a_retro_computer_on_gradient_b_a06360d7-aa20-4d82-949f-48730f690988.webp?w=300&amp;q=30" length="0" type="image/webp"/>
        </item>
        <item>
            <title><![CDATA[OpenAI's DevDay 2025 preview: Will Sam Altman launch the ChatGPT browser?]]></title>
            <link>https://venturebeat.com/technology/openais-devday-2025-preview-will-sam-altman-launch-the-chatgpt-browser</link>
            <guid isPermaLink="false">4hlJDbxEwljBMDhPKiUhiq</guid>
            <pubDate>Fri, 03 Oct 2025 19:00:00 GMT</pubDate>
            <description><![CDATA[<p><a href="https://openai.com/"><u>OpenAI</u></a> will host more than 1,500 developers at its largest annual conference on Monday, as the company behind ChatGPT seeks to maintain its edge in an increasingly competitive artificial intelligence landscape.</p><p>The third annual <a href="https://openai.com/index/announcing-devday-2025/"><u>DevDay conference</u></a> at San Francisco&#x27;s Fort Mason represents a critical moment for OpenAI, which has seen its dominance challenged by rapid advances from Google&#x27;s <a href="https://gemini.google.com/app"><u>Gemini</u></a>, Anthropic&#x27;s <a href="https://claude.ai/"><u>Claude</u></a>, and Meta&#x27;s growing AI efforts. The event comes just days after OpenAI&#x27;s new <a href="https://www.cnbc.com/2025/10/03/openai-sora-apple-app-store.html"><u>Sora video generation app topped Apple&#x27;s App Store</u></a>, demonstrating the company&#x27;s ability to capture mainstream attention even as technical competitors close the gap.</p><p>Chief Executive Sam Altman will deliver the opening keynote at 10 a.m. Pacific time, promising &quot;announcements, live demos, and a vision of how developers are reshaping the future with AI.&quot; The session will be livestreamed, but subsequent presentations — including a developer-focused &quot;State of the Union&quot; with President Greg Brockman and a closing conversation between Altman and Apple design legend Jony Ive — will only be available to in-person attendees.</p><h2><b>Google and Meta challenge ChatGPT&#x27;s developer dominance</b></h2><p>The conference arrives at a pivotal moment for OpenAI. While the company&#x27;s ChatGPT remains the most recognizable AI brand globally, technical evaluations show Google&#x27;s <a href="https://blog.google/technology/google-deepmind/gemini-model-thinking-updates-march-2025/"><u>latest Gemini models</u></a> performing competitively on coding tasks, while Anthropic&#x27;s <a href="https://www.anthropic.com/news/claude-sonnet-4-5"><u>Claude</u></a> has gained traction among developers for its safety features and reasoning capabilities.</p><p>This intensifying competition has fundamentally altered OpenAI&#x27;s strategic calculus. The company that once commanded premium pricing for access to its models now finds itself in a price war, releasing more capable systems at lower costs to retain developer loyalty.</p><p>The shift reflects a maturing market where <a href="https://artificialanalysis.ai/models"><u>technical performance differences</u></a> between leading AI models have narrowed considerably, forcing companies to compete on price, developer experience, and specialized capabilities rather than raw model superiority.</p><p>The timing of <a href="https://devday.openai.com/"><u>DevDay</u></a> also follows several strategic moves by OpenAI that signal broader ambitions beyond its core chatbot business. The company recently launched <a href="https://venturebeat.com/ai/openai-debuts-sora-2-ai-video-generator-app-with-sound-and-self-insertion"><u>Sora 2</u></a>, its advanced video generation model, alongside a social media application that allows users to create and share AI-generated videos. Industry observers speculate that Monday&#x27;s event could feature the long-rumored ChatGPT browser, potentially challenging Google Chrome&#x27;s dominance.</p><h2><b>Enterprise AI adoption takes center stage as revenue strategy shifts</b></h2><p>This year&#x27;s agenda reflects OpenAI&#x27;s growing focus on enterprise customers, who provide more predictable revenue streams than consumer subscriptions. Sessions will cover &quot;orchestrating agents at scale,&quot; enterprise AI adoption challenges, and how OpenAI applies its own technology to internal workflows across sales, support, and finance.</p><p>The enterprise emphasis marks a shift from earlier <a href="https://devday.openai.com/"><u>DevDay</u></a> events. The <a href="https://venturebeat.com/ai/openai-announces-gpt-4-turbo-assistants-api-at-devday-aims-to-revolutionize-ai-apps"><u>inaugural 2023 conference</u></a> introduced GPT-4 Turbo and the GPT Store marketplace, while 2024&#x27;s more subdued gathering focused primarily on developer API improvements. This year&#x27;s expanded format suggests OpenAI views the developer community as crucial to its competitive positioning.</p><p>The State of the Union presentation is expected to focus on how artificial intelligence is transforming software development workflows, with anticipated demonstrations of enhanced capabilities in OpenAI&#x27;s Codex programming assistant and the introduction of new open model offerings that could expand developer access to the company&#x27;s technology.</p><h2><b>Sora cinema and interactive AI demos showcase next-generation capabilities</b></h2><p>Beyond formal presentations, <a href="https://devday.openai.com/"><u>DevDay</u></a> will feature hands-on demonstrations of emerging technologies. A &quot;Sora Cinema&quot; will showcase AI-generated short films, while custom arcade games built using GPT-5 — OpenAI&#x27;s latest model — will demonstrate the technology&#x27;s creative applications.</p><p>Perhaps most intriguingly, attendees can interact with a &quot;living portrait&quot; of computer science pioneer Alan Turing that responds to questions, representing the kind of interactive AI experiences that could define the next generation of human-computer interaction.</p><p>The presence of <a href="https://openai.com/sam-and-jony/"><u>Jony Ive</u></a> at the closing session carries particular significance. The former Apple executive has been collaborating with OpenAI on a consumer AI device, suggesting Monday&#x27;s conversation could provide insights into the company&#x27;s hardware ambitions.</p><h2><b>Developer ecosystem and market positioning face unprecedented competitive pressure</b></h2><p>For enterprise technology decision-makers, <a href="https://devday.openai.com/"><u>DevDay</u></a> represents more than a product showcase — it&#x27;s a window into how AI will reshape software development and business processes. The conference agenda includes sessions on context engineering, agent orchestration, and enterprise scaling challenges that reflect real-world implementation hurdles.</p><p>The developer ecosystem around <a href="https://openai.com/api/"><u>OpenAI&#x27;s APIs</u></a> has become a critical competitive moat. Companies like Cursor, Clay, and Decagon have built substantial businesses on OpenAI&#x27;s foundation models, creating network effects that make switching to alternative providers more difficult.</p><p>However, this ecosystem faces new challenges as competitors offer compelling alternatives. Google&#x27;s recent improvements to <a href="https://www.reddit.com/r/ChatGPTCoding/comments/1jkp5jw/gemini_25_pro_is_the_worlds_best_ai_for_coding/"><u>Gemini for coding tasks</u></a> and <a href="https://www.businessinsider.com/meta-superintelligence-labs-exec-ai-nat-friedman-faster-engineering-tools-2025-10"><u>Meta&#x27;s investments</u></a> in its Superintelligence Labs represent serious threats to OpenAI&#x27;s developer mindshare.</p><p>As the AI industry matures beyond initial breakthroughs, Monday&#x27;s DevDay will test whether OpenAI can maintain its leadership position through superior tooling, developer experience, and enterprise-focused innovation. With over <a href="https://www.reuters.com/technology/openai-hits-500-billion-valuation-after-share-sale-source-says-2025-10-02/"><u>$500 billion in market valuation</u></a> riding on continued growth, the stakes for this year&#x27;s conference extend far beyond San Francisco&#x27;s shores.</p><p>The keynote begins at 10 a.m. Pacific time and will be <a href="https://www.youtube.com/OpenAI"><u>available via livestream</u></a> on OpenAI&#x27;s YouTube channel.</p><p>
</p>]]></description>
            <author>michael.nunez@venturebeat.com (Michael Nuñez)</author>
            <category>Dev</category>
            <category>Technology</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/61ORyI9sy034W678H86Sg1/115c8dc38810efac09cad1fb8fc751eb/nuneybits_A_retro_glowing_computer_on_gradient_background_that__787eb5d3-f84e-4f97-ba01-27ec03df4a88.webp?w=300&amp;q=30" length="0" type="image/webp"/>
        </item>
        <item>
            <title><![CDATA[Anthropic’s new Claude can code for 30 hours. Think of it as your AI coworker]]></title>
            <link>https://venturebeat.com/technology/anthropics-new-claude-can-code-for-30-hours-think-of-it-as-your-ai-coworker</link>
            <guid isPermaLink="false">1Bn23IpmMMPa8MuxWx1TZZ</guid>
            <pubDate>Mon, 29 Sep 2025 17:00:00 GMT</pubDate>
            <description><![CDATA[<p><a href="https://www.anthropic.com/"><u>Anthropic</u></a> launched <a href="https://claude.ai/"><u>Claude Sonnet 4.5</u></a> on Monday, positioning the artificial intelligence model as &quot;the best coding model in the world&quot; in a direct challenge to OpenAI&#x27;s recently released <a href="https://openai.com/index/introducing-gpt-5/"><u>GPT-5</u></a>, as the two AI giants battle for dominance in the lucrative enterprise software development market.</p><p>The San Francisco-based startup claims its newest model achieves state-of-the-art performance on critical coding benchmarks, scoring 77.2% on <a href="https://www.swebench.com/"><u>SWE-bench Verified</u></a> — a rigorous software engineering evaluation — compared to GPT-5&#x27;s performance. More remarkably, Anthropic says Claude Sonnet 4.5 can maintain focus on complex, multi-step tasks for more than 30 hours, a dramatic leap in AI&#x27;s ability to handle sustained work.</p><p>&quot;Sonnet 4.5 achieves 77.2% on SWE-bench Verified (82% with parallel test-time compute). It is SOTA,&quot; an Anthropic spokesperson told VentureBeat, using industry shorthand for &quot;state of the art.&quot; The company also highlighted the model&#x27;s 50% score on <a href="https://www.tbench.ai/"><u>Terminal-bench</u></a>, another coding benchmark where it claims leadership.</p><p>The announcement follows mounting pressure from OpenAI&#x27;s recent advances and pointed criticism from high-profile figures like Elon Musk, who recently posted on X.com that &quot;<a href="https://x.com/elonmusk/status/1970537297792651492?s=46"><u>winning was never in the set of possible outcomes for Anthropic</u></a>.&quot; When asked about Musk&#x27;s statement, Anthropic declined to comment.</p><p>The release arrives just seven weeks after <a href="https://openai.com/index/introducing-gpt-5/"><u>OpenAI&#x27;s GPT-5 launch in August</u></a>, underscoring the breakneck pace of competition in artificial intelligence as companies race to capture enterprise customers increasingly relying on AI for software development. The timing is particularly noteworthy as Anthropic grapples with questions about its heavy dependence on just two major customers.</p><h2><b>Anthropic dominates coding market despite customer concentration risks</b></h2><p>The competition centers on a market that has emerged as AI&#x27;s first major profitable use case beyond chatbots. <a href="https://menlovc.com/perspective/2025-mid-year-llm-market-update/"><u>Anthropic commands 42% of the code generation market</u></a> — more than double OpenAI&#x27;s 21% share — according to a Menlo Ventures survey of 150 enterprise technical leaders. That dominance has translated into remarkable financial performance, with the company reaching a <a href="https://www.anthropic.com/news/anthropic-raises-series-f-at-usd183b-post-money-valuation"><u>$5 billion revenue run rate</u></a> earlier this year.</p><p>However, industry analysis reveals that coding applications <a href="https://cursor.com/"><u>Cursor</u></a> and <a href="https://github.com/features/copilot"><u>GitHub Copilot</u></a> drive approximately <a href="https://www.theinformation.com/articles/anthropics-claude-drives-strong-revenue-growth-while-powering-manus-sensation"><u>$1.4 billion of Anthropic&#x27;s revenue</u></a>, creating a potentially dangerous customer concentration that could leave the company vulnerable if either relationship falters.</p><p>&quot;Our run-rate revenue has grown significantly, even when you exclude these two customers,&quot; the Anthropic spokesperson said, pushing back on concerns about customer concentration. The company provided supportive quotes from both Cursor CEO Michael Truell and GitHub Chief Product Officer Mario Rodriguez praising Claude Sonnet 4.5&#x27;s performance.</p><p>The new model achieves significant advances in computer use capabilities, scoring 61.4% on <a href="https://os-world.github.io/"><u>OSWorld</u></a>, a benchmark that tests AI models on real-world computer tasks. Just four months ago, Claude Sonnet 4 held the lead at 42.2%, demonstrating rapid improvement in AI&#x27;s ability to interact with software interfaces.</p><h2><b>OpenAI&#x27;s aggressive pricing strategy threatens Anthropic&#x27;s premium positioning</b></h2><p>Anthropic&#x27;s announcement comes as the company grapples with competitive pressure from GPT-5&#x27;s aggressive pricing strategy. Early analysis shows <a href="https://simonwillison.net/2025/Aug/7/gpt-5/"><u>Claude Opus 4 costing roughly seven times more</u></a> per million tokens than GPT-5 for certain tasks, creating immediate pressure on Anthropic&#x27;s premium positioning.</p><p>The pricing disparity signals a fundamental shift in competitive dynamics that could force enterprise procurement teams to reconsider vendor relationships previously built on performance rather than price. Companies managing exponentially growing AI budgets now face comparable capability at a fraction of the cost.</p><p>Yet Anthropic is maintaining its pricing strategy with Claude Sonnet 4.5. &quot;Sonnet 4.5&#x27;s cost remains the same as Sonnet 4,&quot; the spokesperson confirmed, keeping prices at $3 per million input tokens and $15 per million output tokens.</p><h2><b>Claude Sonnet 4.5 delivers 30-hour autonomous work sessions and enhanced security</b></h2><p>Beyond performance improvements, Anthropic positions <a href="https://claude.com/product/overview"><u>Claude Sonnet 4.5</u></a> as its &quot;most aligned frontier model yet,&quot; showing significant reductions in concerning behaviors like sycophancy, deception, and power-seeking tendencies. The company has made &quot;considerable progress on defending against prompt injection attacks,&quot; a critical security concern for enterprise deployments.</p><p>The model is being released under <a href="https://www.anthropic.com/news/activating-asl3-protections"><u>Anthropic&#x27;s AI Safety Level 3 (ASL-3) protections</u></a>, which include classifiers designed to detect potentially dangerous inputs and outputs related to chemical, biological, radiological, and nuclear weapons. While these safeguards sometimes flag normal content, Anthropic says it has reduced false positives by a factor of ten since initially describing them.</p><p>Perhaps most significantly for developers, Anthropic is releasing the <a href="https://docs.claude.com/en/home"><u>Claude Agent SDK</u></a> — the same infrastructure that powers its Claude Code product. &quot;We built Claude Code because the tool we needed didn&#x27;t exist yet,&quot; the company said in its announcement. &quot;The Agent SDK gives you the same foundation to build something just as capable for whatever problem you&#x27;re solving.&quot;</p><h2><b>International expansion accelerates as $1.5 billion copyright settlement finalizes</b></h2><p>The model launch coincides with Anthropic&#x27;s aggressive international expansion, as the company seeks to diversify beyond its U.S.-concentrated customer base. The startup recently announced plans to <a href="https://www.reuters.com/business/world-at-work/anthropic-triple-international-workforce-ai-models-drive-growth-outside-us-2025-09-26/"><u>triple its international workforce</u></a> and <a href="https://www.cnbc.com/2025/09/26/anthropic-global-ai-hiring-spree.html"><u>expand its applied AI team fivefold</u></a> in 2025, driven by data showing that nearly 80% of Claude usage now comes from outside the United States.</p><p>However, the expansion comes amid significant legal costs. Anthropic recently agreed to pay <a href="https://apnews.com/article/anthropic-authors-copyright-judge-artificial-intelligence-9643064e847a5e88ef6ee8b620b3a44c"><u>$1.5 billion in a copyright settlement with authors and publishers</u></a> over allegations the company illegally used their books to train AI models without permission. The settlement, approved by a federal judge last week, requires payments of $3,000 for each publication listed in the case.</p><h2><b>Enterprise AI spending doubles as companies prioritize performance over cost</b></h2><p>The rapid-fire model releases from both companies reflect the high stakes in enterprise AI adoption. Model API spending has <a href="https://menlovc.com/perspective/2025-mid-year-llm-market-update/"><u>more than doubled to $8.4 billion</u></a> in just six months, according to Menlo Ventures, as enterprises shift from experimental projects to production deployments.</p><p>Customer behavior patterns suggest enterprises consistently prioritize performance over price, upgrading to the newest models within weeks of release regardless of cost. This behavior could work in Anthropic&#x27;s favor if Claude Sonnet 4.5&#x27;s performance advantages prove compelling enough to overcome GPT-5&#x27;s pricing advantage.</p><p>However, the dramatic price differential introduced by GPT-5 could overcome typical switching inertia, especially for cost-conscious enterprises facing budget pressures. Industry observers note that model switching costs remain relatively low, with <a href="https://venturebeat.com/ai/anthropic-revenue-tied-to-two-customers-as-ai-pricing-war-threatens-margins"><u>66% of enterprises upgrading within existing providers</u></a> rather than switching vendors.</p><p>For enterprises, the intensifying competition delivers better performance and lower costs through continuously improving capabilities. The rapid pace of model improvements — with new versions launching monthly rather than annually — provides organizations with expanding AI capabilities while vendors compete aggressively for their business.</p><p>While the corporate rivalry between Anthropic and OpenAI dominates industry headlines, the real economic impact extends far beyond Silicon Valley boardrooms. The development of AI systems capable of sustained coding work for 30 hours represents a fundamental shift in how software gets built, with implications that extend across every industry relying on technology infrastructure.</p><p>These advancing capabilities signal broader workplace transformation ahead. As AI systems demonstrate increasing proficiency at complex, sustained intellectual work, the technology industry&#x27;s competition for coding supremacy foreshadows similar disruptions across fields requiring analytical thinking, problem-solving, and technical expertise.</p>]]></description>
            <author>michael.nunez@venturebeat.com (Michael Nuñez)</author>
            <category>Automation</category>
            <category>Dev</category>
            <category>Programming &amp; Development</category>
            <category>Software</category>
            <category>Technology</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/qyB4FkFDwzcJMFwLkpg52/2fa0335d017169fc9296f756bc924835/nuneybits_Vector_art_of_brain_with_circuit_pathways_in_burnt_or_af313220-841a-4fca-a27c-072866d0243d.webp?w=300&amp;q=30" length="0" type="image/webp"/>
        </item>
        <item>
            <title><![CDATA[How enterprises can select the right QA tools for riding the AI vibe coding wave]]></title>
            <link>https://venturebeat.com/technology/with-vibe-coding-ai-tools-generating-more-code-than-ever-before-enterprises</link>
            <guid isPermaLink="false">7LuOgyjr1xS5OiIH6fDuQC</guid>
            <pubDate>Tue, 16 Sep 2025 13:00:00 GMT</pubDate>
            <description><![CDATA[<p>Enterprise startup <a href="https://www.coderabbit.ai/">CodeRabbit</a> today raised $60 million to solve a problem most enterprises don&#x27;t realize they have yet. As AI coding agents generate code faster than humans can review it, organizations face a critical infrastructure decision that will determine whether they capture AI&#x27;s productivity gains or get buried in technical debt.</p><p>The funding round, led by Scale Venture Partners, signals investor confidence in a new category of enterprise tooling. The code quality assurance (QA) space is a busy one with GitHub&#x27;s bundled code review features, Cursor&#x27;s bug bot,<a href="https://venturebeat.com/ai/zencoder-just-launched-an-ai-that-can-replace-days-of-qa-work-in-two-hours"> <u>Zencoder</u></a>,<a href="https://venturebeat.com/programming-development/qodo-teams-up-with-google-cloud-to-provide-devs-with-free-ai-code-review-tools-directly-within-platform"> <u>Qodo</u></a> and emerging players like Graphite in a space that&#x27;s rapidly attracting attention from both startups and incumbent platforms.</p><p>The market timing reflects a measurable shift in development workflows. Organizations using AI coding tools  generate significantly more code volume. Traditional peer review processes haven&#x27;t scaled to match this velocity. The result is a new bottleneck that threatens to negate AI&#x27;s promised productivity benefits.</p><p>&quot;AI-generated code is here to stay, but speed without a centralized knowledge base and an independent governance layer is a recipe for disaster,&quot; Harjot Gill, CEO of CodeRabbit told VentureBeat. &quot;Code review is the most critical quality gate in the agentic software lifecycle.&quot;</p><h2>The technical architecture that matters</h2><p>Unlike traditional static analysis tools that rely on rule-based pattern matching, AI code review platforms use reasoning models to understand code intent across entire repositories. The technical complexity is significant. These systems require multiple specialized models working in sequence over 5-15 minute analysis workflows.</p><p>&quot;We&#x27;re using around six or seven different models under the hood,&quot; Gill explained. &quot;This is one of those areas where reasoning models like GPT-5 are a good fit. These are PhD-style problems.&quot;</p><p>The key differentiator lies in context engineering. Advanced platforms gather intelligence from dozens of sources: code graphs, historical pull requests, architectural documents and organizational coding guidelines. This approach enables AI reviewers to catch issues that traditional tools miss. Examples include security vulnerabilities that emerge from changes across multiple files or architectural inconsistencies that only become apparent with full repository context.</p><h2>Competitive landscape and vendor positioning</h2><p>The AI code review space is attracting competition from multiple directions. </p><p>Though there are integrated QA capabilities built directly into platform like GitHub and Cursor there is still a need and a market for standalone solutions as well.</p><p>&quot;When it comes to the critical trust layer, organizations won&#x27;t go cheap out on that,&quot; Gill said. &quot;They will buy the best tool possible.&quot;</p><p>He noted that it&#x27;s similar in some respects  to the observability market where specialized tools like DataDog compete successfully against bundled alternatives like Amazon CloudWatch.</p><p>Gill&#x27;s view is validated by multiple industry analysts.</p><p>&quot;In an era of AI‑assisted development, code review is more important than ever; AI increases code volume and complexity that correspondingly increases code review times and raises the risk of defects,&quot; IDC analyst Arnal Dayaratna told VentureBeat. &quot;That reality elevates the value of an independent, platform‑agnostic reviewer that stands apart from the IDE or model vendor.&quot;</p><p>Industry analyst Paul Nashawaty told VentureBeat that CodeRabbit embeds context-aware, conversational feedback directly in developer environments  making reviews faster and less noisy for developers. Its ability to learn team preferences and provide in-editor guidance reduces friction and accelerates throughput</p><p>&quot;That said, CodeRabbit is more of a complement than a replacement,&quot; Nashawaty said. &quot;Most enterprises will still pair it with established Static Application Security Testing (SAST)/Source Code Analysis (SCA) tools, which the industry estimates represent a $3B plus market growing at approximately 18% CAGR, for broader rule coverage, compliance reporting and governance.&quot;</p><h2>Real-world implementation results</h2><p>The Linux Foundation provides a concrete example of successful deployment. The organization supports numerous open-source projects across multiple programming languages: Golang, Python, Angular and TypeScript. Manual reviews were creating high variance quality checks that missed critical bugs while slowing distributed teams across time zones.</p><p>The default option for the Linux Foundation before CodeRabbit, was to review code manually. This approach was slow, inefficient and error-prone involving significant time commitment from technical leads, with often two cycles to complete the review. After implementing CodeRabbit, their developers reported a 25% reduction in time spent on code reviews.</p><p>CodeRabbit caught issues that human reviewers had missed, including inconsistencies between documentation and test coverage, missing null checks, and refactoring opportunities in Terraform files.</p><h2>Evaluation framework for AI code review platforms</h2><p>Industry analysts have identified specific criteria enterprises should prioritize when evaluating AI code review platforms, based on common adoption barriers and technical requirements.</p><p><b>Agentic reasoning capabilities</b>: IDC analyst Arnal Dayaratna recommends prioritizing agentic capabilities that use generative AI to explain why changes were made, trace impact across the repository and propose fixes with clear rationale and test implications. This differs from traditional static analysis tools that simply flag issues without contextual understanding.</p><p><b>Developer experience and accuracy</b>: Analyst Paul Nashawaty emphasizes balancing developer adoption and risk coverage with focus on accuracy, workflow integration and contextual awareness of code changes.</p><p><b>Platform independence</b>: Dayaratna highlights the value of an independent, platform-agnostic reviewer that stands apart from the IDE or model vendor.</p><p><b>Quality validation and governance</b>: Both analysts stress pre-commit validation capabilities. Dayaratna recommends tools that validate suggested edits before commit to avoid new review churn and require automated tests, static analysis and safe application of one-click patches. Enterprises need governance flexibility to configure review standards. &quot;Every company has a different bar when it comes to how pedantic and how nitpicky they want the system to be,&quot; Gill noted.</p><p><b>Proof-of-concept approach</b>: Nashawaty recommends a 2-4 week proof-of-concept on real issues that help to measure developer satisfaction, scan accuracy, and remediation speed rather than relying solely on vendor demonstrations or feature checklists.</p><p>For enterprises looking to lead in AI-assisted development,it’s increasingly foundational to  evaluate code review platforms as critical infrastructure, not point solutions. The organizations that establish robust AI review capabilities now will have competitive advantages in software delivery velocity and quality.</p><p>For enterprises adopting AI development tools later, the lesson is clear: plan for the review bottleneck before it constrains your AI productivity gains. The infrastructure decision you make today determines whether AI coding tools become force multipliers or sources of technical debt.</p><p>

</p>]]></description>
            <category>Dev</category>
            <category>Technology</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/44kuH6NhIDtQdhtK9fkc62/aac16b67e7450387a34df2a37c20965b/rabbit-vibe-coding-qa-check-smk.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
        <item>
            <title><![CDATA[OpenAI adds 'powerful but dangerous' support for MCP in ChatGPT dev mode]]></title>
            <link>https://venturebeat.com/dev/openai-adds-powerful-but-dangerous-support-for-mcp-in-chatgpt-dev-mode</link>
            <guid isPermaLink="false">6hpz8iMribbm8V3ZKHsG2l</guid>
            <pubDate>Thu, 11 Sep 2025 17:20:00 GMT</pubDate>
            <description><![CDATA[<p>It dropped in the midst of a chaotic and tragic news day yesterday, but OpenAI made a significant upgrade to ChatGPT that&#x27;s worth further consideration among software developers: the <a href="https://platform.openai.com/docs/guides/developer-mode">company added support for the emerging Model Context Protocol (MCP)</a> standard <b>directly into ChatGPT itself when switched into developer mode</b>, allowing third party developers to be able to connect their own external (MCP compatible) servers and tools <i>directly</i> into the developers&#x27; ChatGPT accounts. </p><p>This provides a huge advantage to third-party developers who want to access and modify their own websites, products and services directly <i>within ChatGPT&#x27;s</i> web interface.</p><p>Instead of logging into separate apps, clicking through menus, or juggling multiple dashboards, a developer who has ChatGPT dev mode switched on (and is using the Plus or Pro plans at $20 or $200 monthly, respectively) can ask ChatGPT a natural language question and get a direct answer from <i>the developers&#x27; own </i>service, or even make changes to it, all from within a single chat.</p><p>The company cautions that while the feature is powerful, it is comes with risks. </p><p>In OpenAI’s own words: <i>“it&#x27;s powerful but dangerous, and is intended for developers who understand how to safely configure and test connectors. When using developer mode, watch for prompt injections and other risks, model mistakes on write actions that could destroy data, and malicious MCPs that attempt to steal information.&quot;</i></p><h3><b>Why OpenAI&#x27;s move is so significant and helps further enshrine MCP as an AI industry standard</b></h3><p>MCP itself is an open standard, first <a href="https://www.anthropic.com/news/model-context-protocol">introduced by Anthropic in November 2024</a>, that provides a consistent way to connect AI assistants to external systems such as content repositories, enterprise software, or developer tools.</p><p>Anthropic likens MCP to a kind of USB-C port for AI applications: just as USB-C simplifies hardware connections, MCP standardizes how AI models communicate with external data and services. </p><p>Since its release, MCP has gained rapid adoption across the industry, with early adopters including Block, Apollo, Cloudflare, MongoDB, and PayPal, all of whom have made it possible for developers to connect AI chatbots and other gen AI tools to these respective third party services and obtain information from them. There&#x27;s a<a href="https://mcpservers.org/category/other"> whole website showing public MCP servers</a> to which AI developers can now connect large language models. </p><p>Thus, the purpose of ChatGPT’s new developer mode with MCP support is to give developers a standardized, relatively simple and easy way to connect their systems, tools, or services<i> directly into ChatGPT </i>and maintain that connection going forward, retrieving information or executing operations on the connected MCP server from <i>within the developer&#x27;s own ChatGPT interface</i>. </p><p>Instead of building custom integrations or relying on OpenAI’s old plugin system, developers can now host an MCP server that exposes specific functions — such as checking inventory, updating records, or processing payments — and then execute them <i>from within their own ChatGPT Plus or Pro accounts</i>. </p><p>Industry observers and participants argue that MCP is fast becoming a common language for enterprise AI. A recent <a href="https://venturebeat.com/ai/the-interoperability-breakthrough-how-mcp-is-becoming-enterprise-ais-universal-language">VentureBeat article</a> noted that despite only launching months earlier, MCP appears to be the leading candidate for interoperability in the agentic AI ecosystem. </p><p>Unlike traditional application programming interfaces (APIs), MCP allows for more granular control and security. Enterprises can define what tools are exposed, require authentication from agents, and enforce rules about what models can or cannot access. This fine-grained control makes MCP particularly appealing in corporate environments where security and compliance are top priorities.</p><p>For developers working with ChatGPT’s new developer mode, this means the connectors they create may not just serve one-off integrations — they could be building into a broader ecosystem standard. MCP was created as an open protocol so that one connector can work across different AI ecosystems, not just ChatGPT. </p><h3><b>How it works</b></h3><p>Enabling developer mode requires users to navigate to Settings → Connectors → Advanced → Developer mode.</p><p>Once active, the option to add connectors appears in conversations. Developers can link remote MCP servers over supported protocols such as SSE and streaming HTTP. Authentication options include OAuth or no authentication at all.</p><p>Within the connector settings, developers can toggle individual tools on or off, refresh connectors to pull the latest descriptions, and inspect tool details. </p><p>During conversations, ChatGPT can then invoke the selected tools. Developers may also steer tool usage by writing explicit prompts, such as telling the model to only use a specific connector for a given task, to avoid ambiguity with built-in features.</p><h3><b>From Stripe to Slack: example use cases</b></h3><p>OpenAI posted a<a href="https://x.com/OpenAIDevs/status/1965807401745207708"> video demo on X</a> showing how its new dev mode with MCP support can perform linked actions across different services — once a developer has taken the time to expose and connect the MCP server to ChatGPT.</p><div></div><p>In one example, ChatGPT used an MCP Stripe connector to check a balance, create a customer invoice, and confirm the write action before generating it. The invoice details were then viewable directly in the chat session.</p><p>A second demonstration layered multiple connectors. ChatGPT first processed a refund through Stripe, then used a Zapier connector to send a Slack message notifying the customer. </p><p>Other possible integrations include updating Jira tickets via Atlassian’s MCP server or using Cloudflare’s connector to convert web pages to Markdown.</p><p>These examples highlight the flexibility of chaining connectors together to automate multi-step processes, with ChatGPT managing the sequencing between tools.</p><h3><b>Controls and safeguards</b></h3><p>To mitigate risks, all write actions require explicit confirmation by default. Before a tool executes a write, the developer can expand the tool call details to inspect the full JSON payload, including both inputs and expected outputs. </p><p>Developers may choose to remember an approve or deny choice for the duration of a conversation, but new or refreshed sessions will require confirmation again. </p><p>Tools flagged with the <code>readOnlyHint</code> annotation are treated as read-only, while all others are classified as write actions.</p><h3><b>Guidance for developers</b></h3><p>OpenAI provides recommendations for making connectors easier and safer to use. </p><p>Tool names and descriptions should be action-oriented and include clear instructions about when to use them, as well as parameter explanations. </p><p>This guidance helps the model distinguish between similar tools and avoid defaulting to inappropriate built-in options.</p><p>The documentation also advises developers to disallow alternatives when necessary, specify input shapes for tool calls, and define sequencing when multiple steps are required. </p><p>For example, one prompt might instruct ChatGPT to read a file from a repository first and then write modified content back, avoiding any intermediate or unintended actions.</p><h3><b>Connecting to the broader MCP ecosystem</b></h3><p>The release of developer mode follows other recent updates to OpenAI’s developer-facing tools, particularly the <a href="https://venturebeat.com/programming-development/openai-updates-its-new-responses-api-rapidly-with-mcp-support-gpt-4o-native-image-gen-and-more-enterprise-features">Responses API</a> which received support for MCP in May 2025.</p><p>Among its features are support for remote MCP servers, integration of GPT-4o’s image generation model, and access to built-in tools such as Code Interpreter and improved file search.</p><p>That API was designed as a unified toolbox for building agentic applications and has already processed trillions of tokens since its launch earlier this year. </p><h3><b>Building on earlier commitments</b></h3><p>The expansion of MCP support across OpenAI’s ecosystem also ties back to comments from CEO and co-founder Sam Altman earlier this year. </p><p>In March, <a href="https://x.com/sama/status/1904957253456941061?lang=en">Altman wrote on X </a>that “people love MCP and we are excited to add support across our products,” noting that it was already available in the Agents SDK and would soon extend to the ChatGPT desktop app and Responses API. </p><p>With developer mode now live, that roadmap is taking clearer shape as the company moves to integrate MCP more deeply into both its APIs and its flagship ChatGPT product.</p><h2><b>Comparing OpenAI’s and Anthropic’s MCP guidance</b></h2><p>Although both OpenAI and Anthropic are building around the same open MCP standard, their guidance for developers reflects differences in focus and product integration.</p><p>OpenAI’s instructions for developer mode are tightly tied to ChatGPT’s interface. The company emphasizes practical prompting techniques for ensuring the right tool is called, such as telling the model to only use a specific connector and to avoid built-in tools. </p><p>It also advises developers to specify input formats and sequencing when chaining multiple tool calls. Much of the guidance centers on guardrails and safety: inspecting JSON payloads, confirming write actions, and understanding that model mistakes could delete or alter important data. In other words, OpenAI frames MCP use within ChatGPT as powerful but risky, and stresses developer responsibility in setting up and testing connectors safely.</p><p><a href="https://docs.anthropic.com/en/docs/mcp">Anthropic’s approach</a>, by contrast, is more infrastructure-oriented. Its documentation highlights MCP as an open protocol that enables developers to either expose data through servers or build clients that connect to them. Rather than focusing on prompt design or in-chat usage, Anthropic stresses the architecture: MCP servers as connectors to enterprise systems, MCP clients as AI applications, and a growing ecosystem of prebuilt servers for popular platforms like Google Drive, GitHub, and Slack.</p><p>Its guidance encourages developers to quickly spin up servers, connect them to Claude or other tools, and contribute to the open-source ecosystem.</p><p>Where they converge is in treating MCP as a way to overcome the problem of fragmented integrations. Both also frame it as critical for building agentic systems—AIs that not only generate text but also act on external systems in structured ways.</p><p>The differences reflect their respective product strategies. OpenAI’s guidance is tailored to developers who want to experiment inside ChatGPT and its surrounding APIs, where the risks of write actions are immediate and visible to end users. Anthropic, meanwhile, positions MCP as a foundational building block for enterprise infrastructure and developer platforms, encouraging organizations to standardize their tool connections at the protocol level. For developers, this means OpenAI’s focus is on safe in-chat usage, while Anthropic’s is on building scalable systems that can serve entire organizations.</p><h3><b>What it all means</b></h3><p>For developers already experimenting with MCP servers, the new mode significantly broadens what can be done inside ChatGPT. </p><p>Instead of only fetching data, users can now carry out full workflows — updating records, generating invoices, issuing refunds, or coordinating with third-party services like Slack — directly within a conversation. The ability to chain connectors together also opens the door to more complex automations.</p><p>At the same time, the emphasis on careful setup and review reflects the dual nature of the update: powerful but potentially risky if misused. By keeping confirmations in place and requiring developers to inspect tool calls, OpenAI appears to be prioritizing responsible use as the feature rolls out.</p>]]></description>
            <author>carl.franzen@venturebeat.com (Carl Franzen)</author>
            <category>Dev</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/1XqQUKsPTnWDMZodDPjhT5/f7ffaabef2c42ccaa9df4d83fc880b24/cfr0z3n_flat_illustration_elegant_constructivist_1920s_art_deco_1d105ac8-fbd9-4602-ac28-e1dd5c51f273.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
    </channel>
</rss>