Winning the 1X World Model Challenge
A deep dive into how we won the ICCV 2025 phase of the 1X Humanoid World Model Challenge.
Hello, I’m Aidan, a researcher working at the intersection of machine learning, sequential decision-making, and embodied AI. My research is driven by the goal of building autonomous agents that can learn and generalize behaviours across a wide range of tasks. I’m particularly interested in methods and architectures for learning world models, and in understanding how agents can leverage them to solve new tasks efficiently.
I am a Research Associate at the University of Edinburgh in The Bayesian and Neural Systems Group working with Amos Storkey and Peter Bell. Previously I was a Finnish Center for Artificial Intelligence postdoctoral researcher at Aalto University in Joni Pajarinen’s Robot Learning Lab and Arno Solin’s Machine Learning Research Group. I obtained my PhD from the University of Bristol under the supervision of Arthur Richards and Carl Henrik Ek. During my PhD I developed methods for controlling quadcopters in uncertain environments by synergising methods from probabilistic machine learning, stochastic differential geometry and reinforcement learning.
[04.11.25] Giving a talk on generative world modelling at the Huawei AI Applications Workshop in Dublin on 26th November 2025.
[19.10.25] 🏆 We won both the “Outstanding Champion” and the “Innovation” awards at the ICCV 2025 phase of the 1X Humanoid World Model Challenge!
[05.10.25] Just released our technical report Generative World Modelling for Humanoids: 1X World Model Challenge Technical Report detailing our appraoch that won the 1X world model challenge.
[03.10.25] We’re top of the leaderboard in the 1X Humanoid World Model Challenge. See here for more details.
[18.05.25] New preprint (led by Yi Zhao) Efficient Reinforcement Learning by Guiding Generalist World Models with Non-Curated Data.
A deep dive into how we won the ICCV 2025 phase of the 1X Humanoid World Model Challenge.
Presenting our methods for winning both tracks of the 1X world model challenge.
Introduction World models equip agents (e.g., humanoid robots) with internal simulators of their environments. By “imagining” the consequences of their actions, agents can plan, …
In reinforcement learning (RL), world models serve as internal simulators, enabling agents to predict environment dynamics and future outcomes in order to make informed decisions. …
Learning representations for reinforcement learning (RL) has shown much promise for continuous control. In this project, we investigate using vector quantization to prevent …