AI Solves Real Open Math Problems, World Models Everywhere

Today's Overview

  • AI just solved real open problems in mathematics for the first time. Google DeepMind's Aletheia agent autonomously cracked 4 unsolved questions from the Erdos Conjectures database.
  • GUI world models don't need pixel-level prediction. Code2World turns next-state forecasting into code generation — an 8B model rivals GPT-5.
  • Learning control policies straight from video. VideoWorld 2 improves task success rate by up to 70% on real-world handcraft tasks, and the knowledge transfers to robot manipulation.
  • A training-free reasoning framework that switches cognitive modes per step. Chain of Mindset beats the strongest baseline by nearly 5% across six benchmarks.

Featured

01 AI for Science AI Is Doing Math Research Now, Not Just Math Homework

Hitting IMO gold-medal level is old news. The leap from "solving competition problems" to "doing actual research" is a different beast — it requires literature search, long-horizon proof construction, and exploration in unknown spaces.

Google DeepMind's Aletheia is a math research agent built on an enhanced version of Gemini Deep Think. Its core loop is generate-verify-revise, end-to-end, with intensive tool use. The headline results: a research paper entirely produced by AI (computing structure constants in arithmetic geometry), a human-AI collaborative proof on particle system bounds, and a semi-autonomous sweep of 700 open problems in the Erdos Conjectures database — 4 of which were solved autonomously.

The team also proposes a framework for grading autonomy and novelty in AI-assisted mathematical results. That they felt the need to formalize this says everything about where things are headed. The capability boundary for AI agents is shifting from "solve known problems" to "explore unknown ones."

Key takeaways: - The jump from competition-level to research-level reasoning depends on tool use and long-horizon verification loops - 4 genuinely open math problems solved by AI — real academic results, not benchmark numbers - Proposed autonomy/novelty grading standards signal this will become routine


02 Agent GUI Agents Need to Predict What Happens Next — Pixels Won't Cut It

For a GUI agent to be useful, it needs to anticipate the result of each action — click this button, what does the screen look like? Existing approaches either describe the next state in text (not precise enough) or generate it pixel by pixel (structurally uncontrollable).

Code2World reframes the problem: instead of predicting pixels, generate the HTML code that renders the next screen. The team built a visual-feedback correction pipeline to translate 80K+ GUI trajectories into high-fidelity HTML training data, then applied render-aware RL — using the actual rendered output as the reward signal.

The 8B model matches GPT-5 and Gemini-3-Pro-Image on UI prediction. More practically, plugging Code2World into Gemini-2.5-Flash for lookahead prediction boosts AndroidWorld navigation success rate by 9.5%. Code is open-source.

Key takeaways: - UI prediction shifts from pixel generation to code generation — structurally controllable and higher fidelity - Rendered output as RL reward is a clever closed-loop training design - 8B model, open-source, and directly integrable as a plug-in


03 Robotics Can Manipulation Knowledge From YouTube Transfer to Robots?

Learning to act by watching video sounds compelling but is brutally hard in practice: actions are implicit, and visual appearance varies wildly across scenes.

VideoWorld 2 tackles this by decoupling action dynamics from visual appearance. A pretrained video diffusion model handles appearance; a dedicated latent dynamics model (dLDM) learns only compact, task-relevant action encodings. These encodings are then modeled autoregressively for policy learning and long-horizon planning.

On real-world handcraft tasks — the kind where previous video generation and latent dynamics models reliably fail — task success rate improves by up to 70%. The cross-domain transfer is the real story: manipulation knowledge learned from the Open-X dataset significantly boosts performance on CALVIN robot manipulation tasks. Code and models will be open-sourced.

Key takeaways: - Decoupling dynamics from appearance is the key design for learning control from unlabeled video - 70% success rate improvement on real handcraft tasks, not simulated environments - Cross-domain transfer to robot manipulation validates generality


04 Reasoning Different Reasoning Steps Need Different Thinking Modes

Humans naturally switch cognitive modes when solving complex problems — spatial visualization, divergent brainstorming, logical convergence, algorithmic execution. Different steps engage different mental faculties. But existing LLM reasoning methods, including all CoT variants, apply the same mode uniformly across every step.

Chain of Mindset (CoM) is a training-free agentic framework that decomposes reasoning into four heterogeneous cognitive modes (Spatial, Convergent, Divergent, Algorithmic). A Meta-Agent dynamically selects the optimal mode based on the evolving reasoning state, with a bidirectional context gate controlling information flow between modules.

Across six benchmarks spanning math, code, scientific QA, and spatial reasoning, CoM beats the strongest baseline by 4.96% on Qwen3-VL-32B and 4.72% on Gemini-2.0-Flash while maintaining reasoning efficiency. No training required — plug and play.

Key takeaways: - "Different steps need different thinking modes" is now quantitatively validated - Training-free agentic framework, directly applicable to existing models - Consistent gains across six diverse benchmarks — not a single-task trick

AI Solves Real Open Math Problems, World Models Everywhere

Also Worth Noting

05
Synthetic Environments Train Better Agents Than Real Ones AgentSnowflake's Agent World Model auto-generates 1,000 everyday-scenario environments (35 tools on average), and purely synthetic training achieves strong OOD generalization on three benchmarks. link
06
First Open-Source VLM to Win 12 Gold Medals at Physics Olympiad AI for ScienceP1-VL-235B uses curriculum RL plus agentic reasoning augmentation, ranking No.2 globally on HiPhO behind only Gemini-3-Pro, with large leads on STEM benchmarks. link
07
Draw Two Strokes, Get a Rigged 3D Character Image GenStroke3D generates skeleton-bound 3D models from 2D strokes and text. Accepted at ICLR 2026 — the first framework supporting stroke-conditioned rigged mesh generation. link
08
Video World Models Finally Get Transferable Action Spaces RoboticsOlaf-World uses sequence-level control-effect alignment to anchor latent actions to self-supervised video feature differences, solving the cross-context action semantics problem. link
09
VLA Pretraining Learns the Wrong Thing? Blame Information Leakage RoboticsVLA-JEPA uses JEPA-style pretraining to eliminate future-frame leakage, predicting in latent space instead of pixel space. Better generalization and robustness across robot manipulation tasks. link
10
Text-to-Image Fine-Tuning Finally Gets a Large-Scale Open Dataset Image GenFine-T2I contains 6M text-image pairs across 10 task combinations, 32 prompt categories, and 11 visual styles, with 95%+ candidates filtered out. Open license. link
11
What Makes Shampoo Better Than Muon? Same Story as Adam vs. Signum TrainingNew analysis decomposes Shampoo's advantage into adaptive modulation of Muon updates — essentially spectral descent adapting to stochasticity. link
12
60% MoE Compression With 8+ Perplexity Drop EfficiencyRFID-MoE exploits heterogeneous routing frequency and information density for adaptive SVD compression. Qwen3-30B at 60% compression still gains ~8% accuracy on HellaSwag. link
13
"Human Control" Means Completely Different Things to Different Communities SafetyComparative analysis of two Reddit communities shows deployers care about execution guardrails and recovery, while interaction-focused users care about identity legitimacy and accountability. Same term, divergent operational meanings. link

Today's Observation

The standout trend today is a concentrated emergence of world models — Code2World for GUI state prediction, VideoWorld 2 for video dynamics modeling, Olaf-World for cross-scene action transfer, Agent World Model for synthetic environment generation. Different methods, same thesis: agents that operate in the real world need an internal simulator to anticipate the consequences of their actions. If you're building agent products, world models are worth tracking seriously. They're transitioning from academic concept to core agent infrastructure.