Back to blog
AI NewsJan 04, 20266 min

Gemini hits IMO gold, and the rest of the stack scrambles to catch up

This week: DeepMind's IMO-level Gemini, DPO alignment clarity, safer agent workflows, sturdier deep nets, and multimodal retrieval gets real.

Gemini hits IMO gold, and the rest of the stack scrambles to catch up

DeepMind's "Gemini with Deep Think" hitting an International Mathematical Olympiad gold-medal standard is the kind of milestone that messes with your intuition. Not because it means models are "good at math" now in the way humans mean it. But because it shows something more practical and slightly scarier: a model can sit in a timed setting, grind through multi-step reasoning, and land 5 out of 6 proofs end-to-end in natural language. That's not a parlor trick. That's a capability you can productize.

And it lands in the same week we get clearer alignment recipes (DPO explained from first principles), sturdier training tricks for very deep models (DeepSeek's constrained hyper-connections), and more "grown-up" patterns for agent execution (LangGraph's transactional two-phase commit). Put it together and you get a theme I can't unsee: the frontier isn't just bigger models. It's reliability, controllability, and systems engineering-because we're trying to turn these things into infrastructure.


Main stories

DeepMind's IMO result matters less as a scoreboard win and more as a signal about where model development is headed. Olympiad problems punish shallow pattern matching. They force you to set up structure, keep track of constraints, and not lose the thread halfway through. Getting to "gold medal standard" suggests that long-horizon reasoning is becoming less of a research demo and more of an engineering target.

Here's what caught my attention: it's described as "end-to-end in natural language within time limits." That's a quiet but big deal. Time limits imply compute and search budgets. End-to-end implies fewer handoffs to symbolic tools or curated scaffolds. Natural language implies the interface is the same thing developers already ship. If you're building developer tools, tutoring, formal verification helpers, or any workflow where you need multi-step justification, this is the difference between "LLMs are helpful" and "LLMs can own the task."

The catch is that olympiad-style reasoning is still not the same as being correct in production. The failure mode in prod isn't "couldn't solve problem 6." It's "confidently shipped a wrong answer into a database." Which is why the rest of this week's news-alignment methods, transactional agents, and stability tricks-slots in so neatly.


The Hugging Face write-up deriving the Direct Preference Optimization (DPO) loss from first principles is the kind of content I wish more teams would read before they blindly copy/paste RLHF pipelines. DPO's big pitch has always been "simpler than PPO-style RLHF," but simplicity can mean two different things. It can mean "easier to implement" (nice), or "easier to misunderstand" (dangerous). A derivation helps because it forces everyone to agree on what the objective is actually doing.

My take: DPO is less about being trendy and more about making alignment work accessible to smaller teams. PPO-based RLHF has a lot of moving parts-reward models, policy updates, variance reduction, and the general vibe that you're running a fragile control system. DPO reframes preference tuning into something closer to supervised learning on pairs, with an implicit KL-style regularization baked into the setup. Fewer knobs. Fewer places to blow up training. More reproducibility.

For developers and PMs, the "so what" is pretty direct. If you're aligning a model to your product's preferences-tone, refusal behavior, formatting discipline, safe completion style-DPO-like approaches reduce the overhead tax. That means more teams will do preference tuning as a normal step, not as a moonshot reserved for orgs with a research bench. The competitive line moves from "who has RLHF experts" to "who has the best preference data and evaluation loop."

And that's where the uncomfortable truth kicks in: the scarce resource becomes taste encoded as data. If your preference dataset is inconsistent, you'll get a model that's consistently inconsistent. DPO doesn't magically fix that. It just makes it easier to get to the point where your data quality is the bottleneck-which, honestly, is a healthier bottleneck than "our PPO run diverged again."


LangGraph's tutorial on transactional agentic workflows using a two-phase commit pattern is, to me, one of the most important "boring" developments this week. Because it's not about a new model. It's about not wrecking your systems when the model acts.

Agent demos usually skip the hard part: what happens after the agent decides to do something irreversible. Real businesses have invariants. Money can't be negative. Inventory can't go below zero. You can't email 50,000 customers because the model hallucinated a segmentation query. The two-phase commit approach-prepare actions, validate invariants, get human approval or automated checks, then commit-brings database thinking into agent design. And that's exactly where we need to go.

Here's what I noticed: "human interrupts" and "safe rollbacks" aren't just UX features. They're governance primitives. They turn an agent from a reckless intern into a system that can be audited, paused, and corrected. If you're building agentic workflows for finance, ops, IT automation, or customer support, this pattern is the difference between "we tried agents and got burned" and "we can actually deploy this."

The other thing: transactional design forces you to define state. A lot of agent frameworks implicitly avoid state because state is hard. But without explicit state and commit semantics, you don't really have a workflow-you have a chat that occasionally triggers side effects. That's not an app. That's a liability.


DeepSeek's mHC idea-Manifold Constrained Hyper Connections with Sinkhorn-Knopp constraints-sounds academic until you map it to the pain it's trying to fix: training instability in very deep language models when you start getting fancy with residual mixing.

I read this as part of a broader trend: architecture tweaks are now often about making optimization behave, not just boosting benchmark scores. The Sinkhorn-Knopp bit is an old-school matrix normalization method (1960s era) showing up as a modern stabilizer. That's pretty neat, and it's also a hint that we're going to see more "classical math" resurrected as deep learning infrastructure.

Why does this matter to builders? Because stability improvements tend to flow downstream. If a technique makes deep models easier to train, you eventually get cheaper training runs, more predictable scaling, and potentially smaller models that perform like bigger ones because the training dynamics are healthier. The people threatened here aren't end users. It's teams betting their roadmap on brittle training recipes. When training becomes more stable and standardized, differentiation shifts away from secret sauce hyperparameters and toward data, evals, and deployment.

Also, anything that helps "very deep" setups is a bet on longer-context, more hierarchical representations, and more multi-stage reasoning inside the network. That rhymes with the Gemini IMO story. Different layer of the stack. Same direction.


Meta open-sourcing PE-AV, a unified audio-video-text encoder trained on around 100 million audio-video-caption pairs, is another "infrastructure, not hype" move. Encoders are the workhorses of multimodal systems. They don't always get the spotlight like chatty generative models, but they're what make retrieval and grounding actually work.

I care about this because multimodal retrieval is where a lot of real value is hiding. If you can embed audio and video into a shared space with text, you can do search, moderation, recommendations, dataset curation, and interactive editing workflows with far less hand labeling. It also plugs directly into product surfaces people already understand: "find the clip where the person says X," "show me similar moments," "retrieve videos matching this vibe."

Meta mentions it powering SAM Audio and large-scale retrieval. That's a tell. The market is shifting from "generate content" to "organize and control content." And for entrepreneurs, open encoders reduce time-to-prototype dramatically. The moat won't be "we have embeddings." It'll be "we have distribution and a feedback loop that keeps embeddings useful."


Quick hits

Cloudflare open-sourcing tokio-quiche is a reminder that performance plumbing still matters, even in an AI-first world. If you're serving model-backed apps at scale, QUIC/HTTP3 support in Rust stacks can shave latency and improve reliability, which directly affects perceived model quality. Users don't separate "the model was slow" from "your product is bad."


Closing thought

I keep coming back to this: the biggest AI wins right now aren't just smarter models. They're systems that make smart models safe, stable, and shippable.

Gemini's IMO performance shows the ceiling rising on long-horizon reasoning. DPO's growing mindshare shows alignment is getting more practical. Transactional agent patterns show we're finally designing for failure instead of pretending it won't happen. Stability work like mHC suggests we're still early in understanding how to reliably train the next generation of architectures. And open multimodal encoders like PE-AV point to a future where "AI product" often means "retrieval plus control," not just generation.

If you're building this year, my bias is simple: treat models as components, not magic. Spend the extra week on evals, guardrails, and commit semantics. That's where the real edge is starting to live.


Original data sources

Hugging Face - Deriving the DPO Loss from First Principles: https://huggingface.co/blog/garg-aayush/derive-dpo-loss

Hugging Face - Navigating the RLHF Landscape (PPO, GAE, DPO): https://huggingface.co/blog/NormalUhr/rlhf-pipeline

DeepMind - Gemini with Deep Think achieves IMO gold-medal standard: https://deepmind.google/blog/advanced-version-of-gemini-with-deep-think-officially-achieves-gold-medal-standard-at-the-international-mathematical-olympiad/

MarkTechPost - DeepSeek mHC with Sinkhorn-Knopp constraints: https://www.marktechpost.com/2026/01/03/deepseek-researchers-apply-a-1967-matrix-normalization-algorithm-to-fix-instability-in-hyper-connections/

MarkTechPost - Cloudflare tokio-quiche for QUIC/HTTP3 in Rust: https://www.marktechpost.com/2025/12/31/how-cloudflares-tokio-quiche-makes-quic-and-http-3-a-first-class-citizen-in-rust-backends/

MarkTechPost - LangGraph transactional agentic workflows (two-phase commit): https://www.marktechpost.com/2025/12/31/how-to-design-transactional-agentic-ai-systems-with-langgraph-using-two-phase-commit-human-interrupts-and-safe-rollbacks/

MarkTechPost - Meta open-sources PE-AV encoder: https://www.marktechpost.com/2025/12/22/meta-ai-open-sourced-perception-encoder-audiovisual-pe-av-the-audiovisual-encoder-powering-sam-audio-and-large-scale-multimodal-retrieval/

Want to improve your prompts instantly?