Subscribe to get latest news!

Deep insights for enterprise AI, data, and security leaders

nuneybits Vector art of robot holding blueprint 193c9fc5-bbb5-46ea-9ff6-1a08bb03716e

How Deductive AI saved DoorDash 1,000 engineering hours by automating software debugging

As software systems grow more complex and AI tools generate code faster than ever, a fundamental problem is getting worse: Engineers are drowning in debugging work, spending up to half their time hunting down the causes of software failures instead of building new products. The challenge has become so acute that it's creating a new category of tooling — AI agents that can diagnose production failures in minutes instead of hours.

Michael Nuñez
AI abstract expressionist image of developers working hard in large office under computer code display

How context engineering can save your company from AI vibe code overload: lessons from Qodo and Monday.com

As cloud project tracking software monday.com’s engineering organization scaled past 500 developers, the team began to feel the strain of its own success. Product lines were multiplying, microservices proliferating, and code was flowing faster than human reviewers could keep up. The company needed a way to review thousands of pull requests each month without drowning developers in tedium — or letting quality slip.

Carl Franzen
nuneybits Vector art of multi-cloud nodes interconnected global df5a72fb-1f71-4b95-8b12-94c1e8def7d6

Baseten takes on hyperscalers with new AI training platform that lets you own your model weights

The San Francisco-based company announced Thursday the general availability of Baseten Training, an infrastructure platform designed to help companies fine-tune open-source AI models without the operational headaches of managing GPU clusters, multi-node orchestration, or cloud capacity planning. The move is a calculated expansion beyond Baseten's core inference business, driven by what CTO Amir Haghighat describes as relentless customer demand and a strategic imperative to capture the full lifecycle of AI deployment.

Michael Nuñez

Data Infrastructure

View all
nuneybits Vector art of robot holding blueprint 193c9fc5-bbb5-46ea-9ff6-1a08bb03716e

How Deductive AI saved DoorDash 1,000 engineering hours by automating software debugging

As software systems grow more complex and AI tools generate code faster than ever, a fundamental problem is getting worse: Engineers are drowning in debugging work, spending up to half their time hunting down the causes of software failures instead of building new products. The challenge has become so acute that it's creating a new category of tooling — AI agents that can diagnose production failures in minutes instead of hours.

Michael Nuñez
nuneybits Vector art of multi-cloud nodes interconnected global df5a72fb-1f71-4b95-8b12-94c1e8def7d6

Baseten takes on hyperscalers with new AI training platform that lets you own your model weights

The San Francisco-based company announced Thursday the general availability of Baseten Training, an infrastructure platform designed to help companies fine-tune open-source AI models without the operational headaches of managing GPU clusters, multi-node orchestration, or cloud capacity planning. The move is a calculated expansion beyond Baseten's core inference business, driven by what CTO Amir Haghighat describes as relentless customer demand and a strategic imperative to capture the full lifecycle of AI deployment.

Michael Nuñez
Recursion-Wonder

Ship fast, optimize later: top AI engineers don't care about cost — they're prioritizing deployment

Across industries, rising compute expenses are often cited as a barrier to AI adoption — but leading companies are finding that cost is no longer the real constraint. The tougher challenges (and the ones top of mind for many tech leaders)? Latency, flexibility and capacity. At Wonder, for instance, AI adds a mere few cents per order; the food delivery and takeout company is much more concerned with cloud capacity with skyrocketing demands. Recursion, for its part, has been focused on balancing small and larger-scale training and deployment via on-premises clusters and the cloud; this has afforded the biotech company flexibility for rapid experimentation. The companies’ true in-the-wild experiences highlight a broader industry trend: For enterprises operating AI at scale, economics aren't the key decisive factor — the conversation has shifted from how to pay for AI to how fast it can be deployed and sustained. AI leaders from the two companies recently sat down with Venturebeat’s CEO and editor-in-chief Matt Marshall as part of VB’s traveling AI Impact Series. Here’s what they shared.

Taryn Plumb
Ironwood board

Google debuts AI chips with 4X performance boost, secures Anthropic megadeal worth billions

The announcement, made Thursday, centers on Ironwood, Google's latest custom AI accelerator chip, which will become generally available in the coming weeks. In a striking validation of the technology, Anthropic, the AI safety company behind the Claude family of models, disclosed plans to access up to one million of these TPU chips — a commitment worth tens of billions of dollars and among the largest known AI infrastructure deals to date.

Michael Nuñez

Security

View all

Newsroom

View all
VB in Conversation

From Lift-and-Shift to AI-Ready Data

Lift-and-shift isn’t enough. MongoDB’s Vinod Bagal breaks down how to modernize your data for AI — and why waiting could cost you your competitive edge.

More

Failed AI projects

6 proven lessons from the AI projects that broke before they scaled

Companies hate to admit it, but the road to production-level AI deployment is littered with proof of concepts (PoCs) that go nowhere, or failed projects that never deliver on their goals. In certain domains, there’s little tolerance for iteration, especially in something like life sciences, when the AI application is facilitating new treatments to markets or diagnosing diseases. Even slightly inaccurate analyses and assumptions early on can create sizable downstream drift in ways that can be concerning.

Kavin Xavier
diffusion process

NYU’s new AI architecture makes high-quality image generation faster and cheaper

Researchers at New York University have developed a new architecture for diffusion models that improves the semantic representation of the images they generate. “Diffusion Transformer with Representation Autoencoders” (RAE) challenges some of the accepted norms of building diffusion models. The NYU researcher's model is more efficient and accurate than standard diffusion models, takes advantage of the latest research in representation learning and could pave the way for new applications that were previously too difficult or expensive.

Ben Dickson
Man with moon head and toga and muscular arms typing on keyboard surrounded by monitors, blue purple tone graphic novel style AI art

Moonshot's Kimi K2 Thinking emerges as leading open source AI, outperforming GPT-5, Claude Sonnet 4.5 on key benchmarks

Even as concern and skepticism grows over U.S. AI startup OpenAI's buildout strategy and high spending commitments, Chinese open source AI providers are escalating their competition and one has even caught up to OpenAI's flagship, paid proprietary model GPT-5 in key third-party performance benchmarks with a new, free model.

Carl Franzen