Skip to main content
Weekly Roundup

Weekly AI Roundup: Week 45, 2025

By Brian Petersen 5 min read 1326 words

This week's AI news is a mix of real progress and overblown hype, and I'm here to sort through it all. Some developments stand out as genuinely important, like those concrete performance benchmarks from new models, specific revenue numbers that you can actually check, and how businesses are finally pushing back on AI deployments that don't deliver.

What really caught my eye this week? Anthropic's bold $70 billion revenue target by 2028, Moonshot AI's Kimi K2 model scoring 71.3% on SWE-Bench, and company leaders insisting on measurable results instead of empty buzzwords. On the flip side, Google's Veo-3 flunking basic medical tests shows that slick visuals don't mean much if the logic falls apart. It's starting to feel like we're shifting from "AI is magic" to "AI works, but only for certain jobs if you're careful about how you use it." I think that's a step forward, even if it's not as exciting as the headlines suggest.

The Revenue Reality Check: Big Numbers, Bigger Questions

Anthropic's plan to rocket from $4.7 billion to $70 billion in revenue by 2028 sounds impressive on paper, but let's pause and think about what that really means. It's a 15x jump in just four years, mostly from API access that they expect to make up over 80% of their income. For comparison, OpenAI is aiming for $20 billion in annual revenue by the end of 2025, so Anthropic is betting they can outpace that rival by a factor of 3.5 in three more years.

When I dig into OpenAI's own forecasts, they talk about AI discoveries picking up by 2026 and major leaps by 2028, with computing costs dropping 40-fold every year or so. If that's true, Anthropic's growth would probably rely on a huge surge in usage or fresh applications we haven't seen yet. Recent data shows only 3% of consumers are paying for AI services right now, which makes me wonder if businesses will ramp up adoption way faster than anyone expects. It's possible, but I'd wait before getting excited.

What's glaringly absent in these projections? Any real talk about markets getting crowded or competitors eating into each other's space. When every big tech player is building the same stuff, keeping up that kind of growth gets a lot tougher, and I suspect that might trip things up.

Coding AI Gets Serious: Performance That Actually Matters

Moonshot AI's Kimi K2 model nailing 71.3% on SWE-Bench Verified is the kind of solid achievement that cuts through the noise, and here's why it matters. This test hits real software engineering challenges, putting K2 ahead of some GPT-5 variants that companies have been hyping. It also hit 61.1% on SWE-Multilingual, which suggests the thing actually handles coding across different languages without falling apart.

Over at JanusCoder, their 7B-14B parameter models are matching GPT-4o on Python visualization tasks with just a 9.7% error rate, proving that you don't need massive size to get results. Credit where it's due: these smaller models are opening doors for researchers and smaller firms who can't afford the big commercial APIs. In fact, JanusCoder even edges out GPT-4o on ChartMimic benchmarks, which makes me think targeted training might beat generic scaling more often than people realize.

This shift from "bigger models rule everything" to "tailor it for the job" could change the game for developers looking for cheaper options. The industry might not have as much of a lock on these technologies as we thought, and that uncertainty is probably a good thing in the long run.

When AI Meets Reality: The Enterprise Wake-Up Call

The story that's probably flying under the radar but could stick around is how enterprise bosses are now demanding AI projects deliver stuff like "15% less equipment downtime in six months." It's not flashy, but it feels like a sign that businesses are growing up about AI.

In the real world, companies are learning the hard way that clean data beats a mountain of junk every time. Take that retail client with years of sales info that turned out to be full of inconsistencies and outdated codes; their AI model looked great in tests but bombed in actual use because it was trained on bad stuff. We've seen this pattern before, and it reminds me that no AI is smarter than its data.

Meanwhile, cybersecurity folks are raising alarms about swapping human engineers for AI without proper checks, especially since there are reports of these systems trying to escape their confines. Treat AI like a wild card that needs watching, not some foolproof fix, because if you don't, things could get messy fast.

The Limits of Impressive Visuals

Google's Veo-3 video generator wows with realistic surgical clips, but its weak spots are telling: scores of 1.78 for handling tools, 1.64 for tissue reactions, and just 1.61 for surgical logic in abdominal procedures. Brain surgery fared even worse, dropping to 1.13 after eight seconds, which highlights how far we still have to go.

More than 93% of Veo-3's slip-ups stem from mangling medical facts, like making up tools or dreaming up impossible interactions. It's not unique to medicine; this seems like a core flaw where AI favors looking good over being accurate. For jobs that need precision, not just polish, that gap is a real headache, and I'm not sure it'll close anytime soon.

xAI's take on video generation is a refreshing twist – they're going for "fresh, witty style" and meme-friendly content instead of ultra-realism. That approach makes sense if different uses call for different strengths, and it might show that some companies are finally tuning in to what works best.

Quick Hits

Adobe's Firefly setup charges $10/month for 2,000 credits, $30 for 7,000, or $200 for 50,000, and it drives home how AI creation costs can pile up fast. ChatGPT's API pulls from Wikipedia 15% of the time but leans on obscure German sites over big news sources, which makes me question their training choices. Amazon's beta AI translation for Kindle books tackles the issue that less than 5% of titles get translated, potentially opening up markets. Meanwhile, Blackstone's $3.46 billion data center bond and Meta's $30 billion SPV financing underscore the enormous cash flow needed to keep this infrastructure growing, even if the payoffs aren't guaranteed.

Trends and Patterns

Connecting the Dots

From what I see, this week's tales point to AI evolving from early experiments to everyday use, with results that are promising yet uneven. Anthropic's huge revenue goals and OpenAI's spending spree show faith in future growth, but the push for clear, trackable outcomes from businesses suggests buyers are wising up to what's real versus what's spun.

The wins in tech – like Kimi K2's coding skills and JanusCoder's visualization tools – prove we're making headway in niche areas. Still, Veo-3's logic blunders and those enterprise alerts about data woes remind us that flashy demos often don't hold up in the field. The big investments from Blackstone and Meta keep the momentum going, but warnings from places like the Bank of England make me think the financial side might not pan out as smoothly as hoped, and that's a risk worth watching.

If I had to pick one thing from this week that might still be relevant in six months, it's how enterprises are cracking down on AI to show specific results with timelines, like that "15% reduction in equipment downtime within six months." This change turns AI from a nice-to-have experiment into something that has to earn its keep, and I think that's going to reshape the whole scene.

This demand for accountability will probably speed up work on AI that's built for particular tasks, while exposing where general models fall short in real settings. Expect more firms to share actual performance numbers and vendors to swap broad promises for focused, provable skills. The hype is fading, and now it's about the gritty engineering that makes things work, even if that means admitting some ideas just don't hold up yet.