Failed AI projects
Featured

6 proven lessons from the AI projects that broke before they scaled

Companies hate to admit it, but the road to production-level AI deployment is littered with proof of concepts (PoCs) that go nowhere, or failed projects that never deliver on their goals. In certain domains, there’s little tolerance for iteration, especially in something like life sciences, when the AI application is facilitating new treatments to markets or diagnosing diseases. Even slightly inaccurate analyses and assumptions early on can create sizable downstream drift in ways that can be concerning.

Kavin Xavier
Subscribe to get latest news!

Deep insights for enterprise AI, data, and security leaders

AI abstract expressionist image of developers working hard in large office under computer code display

How context engineering can save your company from AI vibe code overload: lessons from Qodo and Monday.com

As cloud project tracking software monday.com’s engineering organization scaled past 500 developers, the team began to feel the strain of its own success. Product lines were multiplying, microservices proliferating, and code was flowing faster than human reviewers could keep up. The company needed a way to review thousands of pull requests each month without drowning developers in tedium — or letting quality slip.

Carl Franzen
nuneybits Vector art of multi-cloud nodes interconnected global df5a72fb-1f71-4b95-8b12-94c1e8def7d6

Baseten takes on hyperscalers with new AI training platform that lets you own your model weights

The San Francisco-based company announced Thursday the general availability of Baseten Training, an infrastructure platform designed to help companies fine-tune open-source AI models without the operational headaches of managing GPU clusters, multi-node orchestration, or cloud capacity planning. The move is a calculated expansion beyond Baseten's core inference business, driven by what CEO Amir Haghighat describes as relentless customer demand and a strategic imperative to capture the full lifecycle of AI deployment.

Michael Nuñez
Man with moon head and toga and muscular arms typing on keyboard surrounded by monitors, blue purple tone graphic novel style AI art

Moonshot's Kimi K2 Thinking emerges as leading open source AI, outperforming GPT-5, Claude Sonnet 4.5 on key benchmarks

Even as concern and skepticism grows over U.S. AI startup OpenAI's buildout strategy and high spending commitments, Chinese open source AI providers are escalating their competition and one has even caught up to OpenAI's flagship, paid proprietary model GPT-5 in key third-party performance benchmarks with a new, free model.

Carl Franzen

Data Infrastructure

View all
nuneybits Vector art of multi-cloud nodes interconnected global df5a72fb-1f71-4b95-8b12-94c1e8def7d6

Baseten takes on hyperscalers with new AI training platform that lets you own your model weights

The San Francisco-based company announced Thursday the general availability of Baseten Training, an infrastructure platform designed to help companies fine-tune open-source AI models without the operational headaches of managing GPU clusters, multi-node orchestration, or cloud capacity planning. The move is a calculated expansion beyond Baseten's core inference business, driven by what CEO Amir Haghighat describes as relentless customer demand and a strategic imperative to capture the full lifecycle of AI deployment.

Michael Nuñez
Ironwood board

Google debuts AI chips with 4X performance boost, secures Anthropic megadeal worth billions

The announcement, made Thursday, centers on Ironwood, Google's latest custom AI accelerator chip, which will become generally available in the coming weeks. In a striking validation of the technology, Anthropic, the AI safety company behind the Claude family of models, disclosed plans to access up to one million of these TPU chips — a commitment worth tens of billions of dollars and among the largest known AI infrastructure deals to date.

Michael Nuñez
Credit: VentureBeat made with Midjourney

Snowflake builds new intelligence that goes beyond RAG to query and aggregate thousands of documents at once

Enterprise AI has a data problem. Despite billions in investment and increasingly capable language models, most organizations still can't answer basic analytical questions about their document repositories. The culprit isn't model quality but architecture: Traditional retrieval augmented generation (RAG) systems were designed to retrieve and summarize, not analyze and aggregate across large document sets.

Sean Michael Kerner

Security

View all

Newsroom

View all

More

machines construction building framework

Google Cloud updates its AI Agent Builder with new observability dashboard and faster build-and-deploy tools

The new features, announced today, include additional governance tools for enterprises and expanding the capabilities for creating agents with just a few lines of code, moving faster with state-of-the-art context management layers and one-click deployment, as well as managed services for scaling production and evaluation, and support for identifying agents.

Emilia David
IMG 8825

AI’s capacity crunch: Latency risk, escalating costs, and the coming surge-pricing breakpoint

The latest big headline in AI isn’t model size or multimodality — it’s the capacity crunch. At VentureBeat’s latest AI Impact stop in NYC, Val Bercovici, chief AI officer at WEKA, joined Matt Marshall, VentureBeat CEO, to discuss what it really takes to scale AI amid rising latency, cloud lock-in, and runaway costs.

VB Staff