GeoWorld: Geometric World Models
Key Summary
- ā¢GeoWorld is a new way for AI to plan several steps into the future by thinking in shapes (geometry) instead of only numbers.
- ā¢It moves the AIās thoughts from flat space (Euclidean) into a curved space called hyperbolic space, where hierarchies fit naturally.
- ā¢Inside this curved space, the shortest, most meaningful paths (geodesics) guide the AI to make steadier, longer plans.
- ā¢A special predictor (H-JEPA) maps video observations into the hyperbolic world so distances match real task structure.
- ā¢A learning step called Geometric Reinforcement Learning (GRL) tunes the predictor so lower energy equals better plans, and rollouts follow geodesics.
- ā¢During planning, the AI searches for the action sequence that travels the lowest-energy hyperbolic path to the goal (using CEM).
- ā¢On two big benchmarks (CrossTask and COIN), GeoWorld beats the strong V-JEPA 2 baseline, especially for longer plans (T=3 and T=4).
- ā¢Results show about 3% higher success rate for 3-step plans and 2% higher for 4-step plans on average.
- ā¢The main win: adding geometry makes the AIās inner map match the real worldās layered structure, so errors donāt snowball as fast.
- ā¢This could help robots, video assistants, and how-to tools plan reliable multi-step procedures from visual input.
Why This Research Matters
Many real tasksācooking, assembling gadgets, or lab proceduresāare multi-step and hierarchical. GeoWorld helps AI plan these steps more reliably by using a space that naturally fits branching futures. That reduces the āsnowballing errorā that often ruins long plans. It also avoids heavy pixel generation, making planning faster and cleaner. As robots and assistants learn from videos, shaping their inner map with geometry can turn clumsy guidance into confident, step-by-step help. This is a step toward visual planners that are stable, accurate, and useful in everyday life.
Detailed Explanation
Tap terms for definitions01Background & Problem Definition
š Hook: Imagine youāre building a LEGO castle. You donāt plan just the next brick; you think in stepsāfirst the base, then the towers, then the flags. If your plan doesnāt respect the structure (base before towers), the castle wobbles.
š„¬ The Concept (Latent space representation):
- What it is: A hidden, compact map where AI stores the āimportant partsā of images and videos without all the pixel details.
- How it works:
- The AI sees a picture or video and turns it into a vector (a list of numbers) called a latent.
- Latents cluster related scenes close together (e.g., all āmake a sandwichā steps live near each other).
- The AI uses these latents to predict future steps.
- Why it matters: If the hidden map is messy, the AI gets lost when planning multiple steps ahead. š Anchor: Itās like a treasure map that shows big landmarks so you can plan a route without getting distracted by every pebble.
š Hook: You know how a ball naturally rolls to the lowest spot in a bowl? Thatās the path of least effort.
š„¬ The Concept (Energy-based predictive models):
- What it is: A way for AI to score how ācompatibleā a current situation is with a possible future; lower energy means more likely and better.
- How it works:
- Turn images into latents.
- For a guessed next latent, compute an energy score (lower is better).
- Search for actions that lead to low-energy futures.
- Why it matters: Instead of generating pixels (hard and noisy), the AI just finds a valley in the energy landscape of latents. š Anchor: Itās like choosing the smoothest sled path downhill instead of trying to draw every snowflake.
š Hook: Planning a school play needs a big plan (acts) and smaller plans (lines). Mixing them up causes chaos.
š„¬ The Concept (Hierarchical planning):
- What it is: Plan at high-level first (the storyline), then fill in low-level details (the lines and stage moves).
- How it works:
- Find a rough path of main steps.
- For each main step, fill in details.
- Keep both levels consistent.
- Why it matters: Without hierarchy, long plans break because tiny mistakes stack up without guidance. š Anchor: Like making a sandwich: decide the order (bread ā fillings ā bread), then place each ingredient carefully.
š Hook: Imagine city streets that branch out like a tree: downtown ā neighborhoods ā blocks ā houses. This isnāt flat; it grows fast!
š„¬ The Concept (Geodesics):
- What it is: The shortest, most meaningful paths in curved spaces.
- How it works:
- Pick two points in a curved world.
- Trace the curve that keeps you closest to the surface while staying shortest.
- Use that curve as your ābest route.ā
- Why it matters: In planning, geodesics encode efficient, stable progress. š Anchor: Itās the true shortcut over a hill, not the long walk around.
š Hook: Picture squeezing a huge family tree into a circle where the center is ārootā and the edges hold many leaves.
š„¬ The Concept (PoincarĆ© ball model):
- What it is: A way to draw hyperbolic (negatively curved) space inside a ball so hierarchies fit naturally.
- How it works:
- Map the root near the center.
- Push more detailed nodes outward.
- Distances stretch near the boundary, so levels separate clearly.
- Why it matters: Hierarchies (like procedures) pack efficiently with less confusion. š Anchor: Itās a stretchy map where each new level of a plan has room without bumping into others.
The world before GeoWorld: Many visual planners either generated pixels frame by frame (slow, noisy, and short-sighted) or used energy-based models in flat (Euclidean) latent spaces. Flat spaces donāt represent tree-like futures well, so tiny errors grow fast as you plan further. Also, most models were trained mostly on one-step data, making them wobble on long rollouts.
The problem: Two big hurdles kept showing up. First, geometric neglect: the hidden map didnāt preserve hierarchy and distances in a meaningful way. Second, multi-step shortcoming: plans fell apart over longer horizons because the model wasnāt trained or shaped to stay stable.
What people tried: Generative models made pixels or tokens for the next frame and used inverse dynamics to guess actions. But they saw only one step at a time and paid big compute costs to render pixels. Predictive models avoided pixels and learned an energy landscape, which was betterābut they learned in flat space and focused on one- or two-step training.
The gap: We needed a representation that respects hierarchy (like a family tree), plus a trainer that encourages multi-step paths to be short, straight, and consistent (geodesics) over many steps.
Real stakes: In daily life, cooking, fixing things, playing sports, or doing science projects all need multi-step plans. If your assistant canāt hold a stable long plan, it gives you steps out of order or gets stuck. Making the AIās inner map match real-world structure means instructions that work, robots that donāt fumble, and videos that teach reliably.
02Core Idea
š Hook: You know how hanging coats on a straight rod gets crowded fast, but a tree with branches holds many more coats neatly? Space shape matters.
š„¬ The Concept (GeoWorld):
- What it is: A world model that plans in a curved, hierarchy-friendly space so multi-step visual plans follow the best (geodesic) routes.
- How it works:
- Encode video observations into latents.
- Map those latents into a hyperbolic (curved) space.
- Predict next states along geodesics (the shortest curves in that space).
- Train with a geometric RL trick so low energy means better, longer plans.
- Why it matters: Without geometry, long plans drift; with it, plans align with real task structure and stay stable. š Anchor: Picture planning the Replace Memory Chip task by gliding down a curved valley that naturally keeps sub-steps in order.
Aha! Moment in one sentence: If the future branches like a tree, then plan in a space that loves treesāhyperbolic geometryāso the shortest paths are the most stable multi-step plans.
Three analogies:
- Library analogy: Flat shelves (Euclidean) cram books together; a branching bookshelf (hyperbolic) keeps topics neatly separated so you can find the next book in a series easily.
- Road trip analogy: In a mountain region, the straight line on a flat map is silly; the safe, efficient route follows the terrainās curves (geodesic).
- Family tree analogy: Putting a giant family tree on lined paper (flat) overlaps branches; using a circle that expands near the edge (PoincarƩ ball) keeps generations clear.
Before vs After:
- Before: Energy-based planners in flat space; decent short steps, weak long paths; errors compound and hierarchy is fuzzy.
- After: Energy-based planners in hyperbolic space; geodesics protect long paths; hierarchy is built-in, and GRL gently straightens multi-step rollouts.
š Hook: Imagine roads painted right on the hills so even if youāre blindfolded, the slope guides you.
š„¬ The Concept (H-JEPA):
- What it is: A predictor that maps Euclidean latents into a PoincarƩ ball and learns to step along hyperbolic geodesics.
- How it works:
- Encode frames to Euclidean latents.
- Exponential map projects them into the hyperbolic ball.
- Predict next states so hyperbolic distance to ground-truth is minimized (teacher-forced and with rollout consistency).
- Why it matters: The predictor respects curvature, so the energy landscape mirrors real hierarchical structure. š Anchor: Itās like drawing your route on a globe instead of on a flat map, so āshortestā matches reality.
š Hook: You know how a rubber band snaps into a straight line when stretched between two pins?
š„¬ The Concept (Geometric Reinforcement Learning, GRL):
- What it is: A training step that treats lower hyperbolic energy as higher reward and prefers geodesic-consistent rollouts.
- How it works:
- Define reward = negative hyperbolic distance error.
- Maximize cumulative reward over several steps.
- Add triangle-inequality regularization so two small steps donāt beat one direct step unless itās truly shorter.
- Why it matters: It sculpts the predictorās value landscape so long plans stay smooth and donāt zigzag. š Anchor: Itās like coaching a relay team to pass the baton in the straightest lane, not weave across the track.
Why it works (intuition, no math):
- Future branches explode like trees; hyperbolic space compresses trees naturally, so nearby steps and far-away goals get the right spacing.
- Geodesics in that space are meaningful āstraightā paths; training to follow them makes each added step less likely to drift.
- Triangle inequality is a built-in safety rule that discourages detours.
Building blocks:
- Latent space ā PoincarĆ© ball projection ā Hyperbolic distance.
- Predictor trained on one-step and two-step rollouts.
- GRL tunes the predictor so ālow energy = good long plan.ā
- CEM planner searches for action sequences that ride the low-energy valley to the goal.
03Methodology
At a high level: Video/goal input ā Encode to latents ā Project into hyperbolic space (H-JEPA) ā Predict next latents along geodesics ā GRL refines multi-step value/energy ā Plan actions with CEM to reach the goal along lowest-energy paths.
š Hook: Imagine you take notes (latents) from a video, then pin them inside a stretchy circle where related notes form neat branches.
š„¬ The Concept (Hyperbolic projection with PoincarĆ© ball):
- What it is: A layer that maps Euclidean latents into the hyperbolic ball so distances reflect hierarchy.
- How it works:
- Encode frame xt into a Euclidean latent.
- Use the exponential map to place it in the hyperbolic ball.
- Learn the curvature parameter so the space fits the data.
- Why it matters: If you keep latents flat, branches overlap; curving the space separates levels. š Anchor: Like seating students by grade levels in a fan-shaped hall so everyone fits without crowding.
š Hook: To build a card tower, you place the next card where it best supports the layer above.
š„¬ The Concept (Action-conditioned prediction in H-JEPA):
- What it is: Given a current hyperbolic latent and an action, predict the next latent by following the geodesic direction.
- How it works:
- Input: a sequence of hyperbolic latents and actions.
- Predictor outputs next-step latents across T steps.
- Loss uses hyperbolic distance between predicted and true latents.
- Why it matters: Aligning predictions with hyperbolic distances keeps steps consistent with hierarchy. š Anchor: Itās like stepping stones placed so each new stone is the shortest safe hop toward the goal.
Training recipe (Supervised Fine-Tuning, SFT):
- Teacher Forcing Loss: one-step accuracy; compare predicted next latent to the true next latent using hyperbolic geodesic distance.
- Rollout Loss: two-step consistency; feed predictions back in and still match the true future.
- Total: a balance Ī» between one-step and rollout; roughly half/half works well for long-horizon stability.
š Hook: When hiking, you prefer a single smooth trail over two awkward zigzags, even if distances look similar.
š„¬ The Concept (GRL value shaping + triangle inequality):
- What it is: A planner-friendly tuning where ārewardā = negative hyperbolic distance error over multiple steps, plus a rule favoring directness.
- How it works:
- Define per-step cost as the hyperbolic distance error; reward is its negative.
- Sum discounted rewards over T steps.
- Add triangle-inequality regularization: the direct two-step gap should not exceed the sum of the single-step gaps; encourage geodesic-like rollouts.
- Why it matters: It bends the modelās internal roads into smooth highways, reducing detours as T grows. š Anchor: Like using a ruler to keep your drawing straight instead of freehand wobbles.
š Hook: Picking the best route on a map often means sampling a few options and choosing the most promising, then refining.
š„¬ The Concept (Energy-based planning with CEM):
- What it is: A search method that samples action sequences, keeps the best (lowest energy), and refines around them.
- How it works:
- Encode current and goal frames into hyperbolic latents.
- Sample many candidate action sequences.
- Predict where each sequence lands; measure hyperbolic energy to the goal.
- Keep the best few, update the sampling distribution, and repeat.
- Why it matters: It finds action paths that follow the curved valley of lowest energy to the goal without generating pixels. š Anchor: Itās like trying several paper airplane folds, keeping the ones that fly farthest, and tweaking them.
Concrete example (Replace Memory Chip task):
- Input: current video clip where the device is closed; goal clip where the new chip is installed.
- Encode both into hyperbolic latents.
- The predictor imagines how actions like ātake off shell ā remove old chip ā install new one ā close shellā move along the geodesic path to the goal.
- CEM searches for action sequences that keep predicted latents on the lowest-energy curve; the best sequence matches the correct procedure order.
Secret sauce:
- Curvature-aware distances separate levels of a task (preparing vs installing vs finishing).
- GRLās triangle rule keeps rollouts straight and stable.
- No pixel generationājust clean, structured energy minimization in a space shaped like the tasks themselves.
04Experiments & Results
š Hook: Think of a cooking contest where teams follow a recipe from videos. We judge not just if they finish, but if each step is correct and in order.
š„¬ The Concept (The test setup):
- What it is: Multi-step visual planning on CrossTask and COINābig collections of how-to videos with labeled steps and timings.
- How it works:
- Give the model a starting observation and a goal (image or video).
- Ask it to predict T actions that reach the goal.
- Score with:
- Success Rate (SR): all steps exactly match (a perfect recipe).
- Mean Accuracy (mAcc): percent of correct steps on average.
- Mean IoU (mIoU): overlap between predicted and true procedure sequences.
- Why it matters: Real procedures are long and precise; scoring must check order and coverage. š Anchor: Itās like grading a LEGO build: did you make the right model (SR), place most pieces right (mAcc), and follow the plan closely (mIoU)?
The competition:
- LLM/VLM planners (e.g., Gemini 2.5 Pro, GPT-5).
- Generative world models (e.g., VideoWorld, PDPP, ActionDiffusion).
- Predictive world models (e.g., V-JEPA 2, E3P, PlaTe).
- Simple baselines (Random, Retrieval-based) for context.
Scoreboard with context:
- Procedural planning (images): GeoWorld consistently beats V-JEPA 2 across model sizes for T=3 and T=4. Think of SR improvements of about +3% at T=3 and +2% at T=4ālike nudging from a solid B+ to an A- when everyone else hovers around B.
- Visual planning with videos: Again, GeoWorld rises above V-JEPA 2 and competes well with strong VLMs, especially as T increases. The ViT-g version reaches the best overall results, reflecting how geometry helps as tasks stretch longer.
- Long-horizon stability: As T goes from 3 to 6 (and even to 8 in extended tests), many methods lose accuracy quickly (error snowballs). GeoWorldās hyperbolic mapping and GRL reduce this drift, keeping SR notably higher at large T.
Surprising findings:
- Learnable curvature tends to settle around a moderate value (e.g., c ā 0.3), not extremely curved. This flattens the space just enough for stability but keeps hierarchy benefits.
- Enforcing triangle inequality in GRL gives steady gainsāitās a simple geometric rule with big practical payoff.
- Even with a frozen encoder (plus a small hyperbolic layer), GeoWorld earns strong improvements; fully fine-tuning adds only modest extra boosts, suggesting the core geometric change is doing the heavy lifting.
Why the numbers matter:
- A few percentage points in SR at T=4 can be the difference between a device being fixed correctly vs. steps out of order.
- Higher mIoU indicates the whole plan lines up closely with the ground truth, not just isolated steps.
- Consistent wins across CrossTask and COIN show broad usefulness across many kinds of everyday procedures.
05Discussion & Limitations
š Hook: Even great hiking maps donāt replace good shoes and daylight; tools help, but they have limits.
š„¬ The Concept (Limitations):
- What it is: Honest places where GeoWorld is not magic.
- How it works:
- Needs good visual data; confusing or very noisy videos can still trip it up.
- Geometry helps hierarchy, but it doesnāt add missing facts; if steps are ambiguous, the plan may still wander.
- Currently focused on visual procedures; full embodied control (e.g., robot torques) needs extra modules.
- Why it matters: Knowing edges prevents overpromising and guides future improvements. š Anchor: Even with a great map, a foggy day can slow you down.
š„¬ The Concept (Required resources):
- What it is: What you need to train and run it well.
- How it works:
- Large video encoders (e.g., V-JEPA 2 backbones) and a ~300M-parameter predictor.
- Multi-GPU training (the paper used H100s) for practical training times.
- Inference is light enough for a single high-end GPU.
- Why it matters: Planning budgets and hardware shape real deployments. š Anchor: Itās like needing a good kitchen to prep a feast, but serving plates only need a small counter.
š„¬ The Concept (When not to use):
- What it is: Situations where simpler or different tools may win.
- How it works:
- One-step, reactive control without hierarchy: a simpler Euclidean predictor might suffice.
- Text-only planning: language-first tools may be enough.
- Pixel-perfect video generation needs: use a generative model.
- Why it matters: Pick the right wrench for the bolt. š Anchor: Donāt bring a bulldozer to plant a flower.
š„¬ The Concept (Open questions):
- What it is: Next puzzles to solve.
- How it works:
- Multi-level sub-task hierarchies (explicit high- and low-level controllers) inside the same hyperbolic space.
- Richer action spaces for robots (continuous control) and real-time feedback.
- Mixing language goals with hyperbolic planning (vision-language-action in curved spaces).
- Automatic curvature schedules that adapt per task.
- Why it matters: Answering these unlocks broader, more reliable planning. š Anchor: Todayās sturdy treehouse needs stairs and handrails to welcome more kids safely.
06Conclusion & Future Work
Three-sentence summary: GeoWorld is a geometric world model that moves planning from a flat latent space into a hyperbolic one, where geodesics naturally capture hierarchical procedures. A hyperbolic predictor (H-JEPA) and Geometric Reinforcement Learning (GRL) shape the energy landscape so long, multi-step rollouts stay smooth and stable. On CrossTask and COIN, this delivers consistent gains over strong baselines, especially as plans get longer.
Main achievement: Showing that respecting geometryāby mapping latents to a PoincarĆ© ball and optimizing geodesic-consistent rolloutsāmeaningfully improves long-horizon visual planning.
Future directions: Add explicit multi-level controllers, extend to embodied robotics and continuous control, blend language with curved-space planning, and auto-tune curvature per task and scale.
Why remember this: When your problem branches like a tree, using a space that loves trees makes planning simpler, steadier, and smarterāturning long procedures from wobbly guesses into guided journeys along the right curves.
Practical Applications
- ā¢Home repair assistants that watch and suggest the next correct step (e.g., replacing a part) with fewer mistakes.
- ā¢Cooking tutors that plan and track multi-step recipes from video, keeping order and timing aligned.
- ā¢Assembly-line guidance that predicts stable action sequences for complex product assembly.
- ā¢Robot manipulation planning that follows geodesic-consistent paths for long-horizon tasks.
- ā¢Sports training tools that analyze a goal clip and plan the sequence of drills to reach similar performance.
- ā¢Educational how-to platforms that generate reliable procedure plans from instructional videos.
- ā¢AR/VR step-by-step overlays that keep procedures in correct order even across long sequences.
- ā¢Video retrieval-by-plan: find videos that match a desired multi-step procedure by comparing latent geodesic paths.
- ā¢Quality control in manufacturing: detect step-order deviations by measuring hyperbolic distance to the ideal plan.
- ā¢Medical skills training (simulation): plan multi-step procedural practice from curated videos with stable ordering (non-clinical support).