Dreamer 4 learns to solve complex control tasks by reinforcement learning inside of its world model. We decode the imagined training sequences for visualization, showing that the world model has learned to simulate a wide range of game mechanics from low-level mouse and keyboard actions, including breaking blocks, using tools, and interacting with crafting tables. Credit: arXiv (2025). DOI: 10.48550/arxiv.2509.24527
Over the past decade, deep learning has transformed how artificial intelligence (AI) agents perceive and act in digital environments, allowing them to master board games, control simulated robots and reliably tackle various other tasks. Yet most of these systems still depend on enormous amounts of direct experience—millions of trial-and-error interactions—to achieve even modest competence.
This brute-force approach limits their usefulness in the physical world, where such experimentation would be slow, costly, or unsafe.
To overcome these limitations, researchers have turned to world models—simulated environments where agents can safely practice and learn.
These world models aim to capture not just the visuals of a world, but the underlying dynamics: how objects move, collide, and respond to actions. However, while simple games like Atari and Go have served as effective testbeds, world models still fall short when it comes to representing the rich, open-ended physics of complex worlds like Minecraft or robotics environments.
Researchers at Google DeepMind recently developed Dreamer 4, a new artificial agent capable of learning complex behaviors entirely within a scalable world model, given a limited set of pre-recorded videos.
The new model, presented in a paper published on the arXiv preprint server, was the first artificial intelligence (AI) agent to obtain diamonds in Minecraft without practicing in the actual game at all. This remarkable achievement highlights the possibility of using Dreamer 4 to train successful AI agents purely in imagination—with important implications for the future of robotics.
“We as humans choose actions based on a deep understanding of the world and anticipate potential outcomes in advance,” Danijar Hafner, first author of the paper, told Tech Xplore.
“This ability requires an internal model of the world and allows us to solve new problems very quickly. In contrast, previous AI agents usually learn through brute-force with vast amounts of trial-and-error. But that’s infeasible for applications such as physical robots that can easily break.”
Some of the AI agents developed at DeepMind over the past few years have already achievedtremendoussuccess at games such as Go and Atari by training in small world models. However, the world models that these models relied on failed to capture the rich physical interactions in more complex worlds, such as the Minecraft videogame.
Dreamer 4 is the first agent to obtain diamonds in Minecraft from only offline data, without practicing in the environment. The agent first learns a world model and then improves its behavior by reinforcement learning in diverse imagined scenarios. The video shows the agent interacting with the actual game for evaluation. Credit: Google DeepMind.
On the other hand, “Video models such as Veo and Sora are rapidly improving towards generating realistic videos of very diverse situations,” said Hafner.
“However, they are not interactive, and their generations are too slow, so they cannot be used as ‘neural simulators’ to train agents inside of yet. The goal of Dreamer 4 was to train successful agents purely inside of world models that can realistically simulate complex worlds.”
Hafner and his colleagues decided to use Minecraft as a test bed for their AI agent, as it is a complex video game that contains infinite generated worlds and long-horizon tasks that require over 20,000 consecutive mouse/keyboard actions to be completed.
One of these tasks is the mining of diamonds, which requires the agent to perform a long sequence of prerequisites such as chopping trees, crafting tools, and mining and smelting ores.
Notably, the researchers wanted to train their agent purely in “imagined” scenarios, instead of allowing it to practice in the actual game, analogous to how smart robots will have to learn in simulation, because they could easily break when practicing directly in the physical world . This requires the model to learn object interactions in an accurate enough internal model of the Minecraft world.
The artificial agent developed by Hafner and his colleagues is based on a large transformer model that was trained to predict future observations, actions and the rewards associated with specific situations. Dreamer 4 was trained on a fixed offline dataset containing recorded Minecraft gameplay videos collected by human players.
“After completing this training, Dreamer 4 learns to select increasingly better actions in a wide range of imagined scenarios via reinforcement learning,” said Hafner.
“Training agents inside of scalable world models required pushing the frontier of generative AI. We designed an efficient transformer architecture, and a novel training objective named shortcut forcing. These advances enabled accurate predictions while also speeding up generations by over 25x compared to typical video models.”
Humans interacting with different Minecraft world models to build a 3×3 wall from wooden planks. The Dreamer 4 world model has learned to accurately predict object interactions and game mechanics, substantially improving over previous world models while also enabling real-time generations on a single GPU. Credit: Google DeepMind.
Dreamer 4 is the first AI agent to obtain diamonds in Minecraft when trained solely on offline data, without ever practicing its skills in the actual game. This finding highlights the agent’s ability to autonomously learn how to correctly solve complex and long-horizon tasks.
“Learning purely offline is highly relevant for training robots that can easily break when practicing in the physical world,” said Hafner. “Our work introduces a promising new approach to building smart robots that do household chores and factory tasks.”
In the initial tests performed by the researchers, the Dreamer 4 agent was found to accurately predict various object interactions and game mechanics, thus developing a reliable internal world model. The world model established by the agent outperformed the models that earlier agents relied on by a significant margin.
“The model supports real-time interactions on a single GPU, making it easy for human players to explore its dream world and test its capabilities,” said Hafner. “We find that the model accurately predicts the dynamics of mining and placing blocks, crafting simple items, and even using doors, chests, and boats.”
A further advantage of Dreamer 4 is that it achieved remarkable results despite being trained on a very small amount of action data. This is essentially video footage showing the effects of pressing different keys and mouse buttons within the Minecraft videogame.
Dreamer 4 learns to solve complex control tasks purely by training inside of a scalable world model, without practicing in the actual environment. The video shows diverse training scenarios imagined by the agent. Credit: Google DeepMind.
“Instead of requiring thousands of hours of gameplay recordings with actions, the world model can actually learn the majority of its knowledge from video alone,” said Hafner.
“With only a few hundred hours of action data, the world model then understands the effects of mouse movement and key presses in a general way that transfers to new situations. This is exciting because robot data is slow to record, but the internet contains a lot of videos of humans interacting with the world, from which Dreamer 4 could learn in the future.”
This recent work by Hafner and his colleagues at DeepMind could contribute to the advancement of robotics systems, simplifying the training of the algorithms that allow them to reliably complete manual tasks in the real world.
Meanwhile, the researchers plan to further improve Dreamer 4’s world model, by integrating a long-term memory component. This would ensure that the simulated worlds in which the agent is trained remain consistent over long periods of time.
“Incorporating language understanding would also bring us closer towards agents that collaborate with humans and perform tasks for them,” added Hafner.
“Finally, training the world model on general internet videos would equip the agent with common sense knowledge of the physical world and allow us to train robots in diverse imagined scenarios.”
Written for you by our author Ingrid Fadelli, edited by Sadie Harley, and fact-checked and reviewed by Robert Egan—this article is the result of careful human work. We rely on readers like you to keep independent science journalism alive.
If this reporting matters to you,
please consider a donation (especially monthly).
You’ll get an ad-free account as a thank-you.
More information:
Danijar Hafner et al, Training Agents Inside of Scalable World Models, arXiv (2025). DOI: 10.48550/arxiv.2509.24527
Citation:
DeepMind introduces AI agent that learns to complete various tasks in a scalable world model (2025, October 25)
retrieved 25 October 2025
from https://techxplore.com/news/2025-10-deepmind-ai-agent-tasks-scalable.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.
If you want to get into stargazing in 2025, there’s still a chance to catch some of the best meteor showers of the year. Also known as shooting stars, meteors happen when Earth’s orbital path crosses a path of debris left by a comet and that material burns up in the Earth’s atmosphere. Watching a meteor shower is one of the most accessible ways to engage with the night sky.
The next shower are the Geminids, a busy and bright shower that peaks in mid-December, offering the chance to see hundreds of shooting stars each hour. This is just one of nine major meteor showers that grace skies throughout the year, and details of when they will appear in the northern hemisphere are listed below—so mark your 2026 calendar for these.
The Next Big Meteor Shower: The Geminids
The Geminids are active from about December 4 to December 17, peaking overnight from December 13 to December 14. They have a sharp peak, so the night of the 13th is the best time for skywatching.
The Geminids are the most spectacular meteor shower of the year. In addition to boasting up to 120 or even 150 meteors per hour during its peak, this meteor shower is also the brightest and most colorful of the year.
The Geminids are bright, slow-moving meteors that often have yellow tones, but they can be a range of other colors, including green, blue, white, red, or orange. And unlike most meteors, which are caused by comet debris, the Geminids are the remnants of an asteroid.
The night that the Geminids peak, their radiant, the constellation Gemini, will be above the horizon all night and will reach its highest point around 2 am local time, so meteors will be visible almost the whole night.
That same night, the moon will be about 32 percent illuminated and will rise around 1:30 am in the eastern US, so if you watch this shower shortly after midnight, the moonlight won’t interfere with your viewing experience.
How to Watch a Meteor Shower
You don’t need any special equipment to see a meteor shower—in fact, using devices like binoculars or telescopes actually prevents you from seeing meteors, because they travel too fast to be seen through the lenses of such equipment. All you need are your eyes, a dark sky with little to no moonlight, and a location that’s away from excess light, as moonlight and light pollution can wash out shooting stars.
Note that the moon appears (rises) and disappears (sets) in the night sky at different times depending on what time zone you are in. All moonrise/moonset times in this piece are for the eastern US. You can use tools like Time and Date’s moonrise/moonset calendar or this tool from the US Naval Observatory to check the precise moonrise/moonset times in your exact location.
Managing complex responsibilities is a common task for digital leaders. However, for Erik Mayer, transformation chief clinical information officer (CCIO) at Imperial College London and Imperial College Healthcare NHS Trust, the mix of responsibilities is central to his role.
He spends about 40% of his time in the clinic and the rest helping to define the future of digital healthcare.
“I enjoy both roles because, actually, they should be intertwined,” he says. “I have many conversations with clinical and academic colleagues who say, ‘Can I get access to this data?’ That’s why it should be intertwined, because what you put in is what you get out.”
Mayer’s successful transition from the surgery room to the IT department began during his PhD research from 2006 to 2009, when he analysed data to produce evidence for centralising cancer services and improving patient care. Through his role at the trust, he became involved in technology implementation projects.
“I’ve always been in and around data and producing robust evidence for why we should or shouldn’t do things,” he says.
“Then, at Imperial Trust, I was a surgical trainee and became involved in IT, informatics and data warehouse-type environments. I was heavily involved in the work when we went live with the Cerner electronic patient record in 2014.”
As Mayer’s experience grew, so did the opportunities to move into new areas. In 2018, after a competitive process, he was appointed to his current role. He has continued to expand his compass while working on the healthcare frontline.
“I wanted to be forward-thinking about creating secure environments to support access to data for driving research and innovation,” he says.
“I wear many hats. I’m a practising surgeon, transformation CCIO, clinical social professor in Imperial College, and I head up the directorate of the iCARE Secure Data Environment (SDE), which is a digital collaboration space that spans the university and the trust.”
Fostering collaboration
Looking back on his seven years in the CCIO role, Mayer says the data environment has evolved into today’s cloud-based platform using Microsoft Azure and Snowflake technology. He says the transformation process was accelerated during the coronavirus pandemic.
“We’d already set up the environment and had some exemplar projects going on where we were supporting healthcare delivery in the trust,” he says.
“Then Covid hit and, suddenly, there was a huge appetite and urgency about accessing data to support basic decision-making around operational processes.”
These processes included monitoring the number of people with the virus and moving patients around the hospital to free up intensive care beds. Through a collaboration with the North West London Integrated Care Board, Mayer and his peers brought together two key datasets, making it possible to track trends across 2.8 million people.
“I wear many hats. I’m a practising surgeon, transformation CCIO, clinical social professor in Imperial College, and I head up the directorate of the iCARE Secure Data Environment”
Erik Mayer, Imperial College Healthcare NHS Trust
“That created a burning platform for data,” he says, looking back on the interest in information that this initiative helped to foster for the longer term. “Today, that data is now fully migrated and held in the same secure environment as the Imperial Trust data, as well as other databases across different tendencies.”
Mayer says having all these databases together in a secure data environment makes it easier for people to link insights. This capability has changed the mindset of people using data. Previously, particularly in the academic world, people and organisations had to set up data-sharing agreements. Now, collaboration is fast becoming the standard way of working.
“This project brought people into the data environment to do their research and innovation. That approach brought academics together with clinicians and data scientists, meaning we could get quick answers around risk prediction and other insights,” he says.
“Our digital transformation was about bringing the right multidisciplinary people together to work collaboratively in a secure way. Fundamentally, of course, by doing that, you maintain the public trust because you’re not selling data off or moving it around.”
Integrating data
Mayer says the implementation of Snowflake technology has been a crucial component of his data-led approach to digital transformation. While it took weeks to ingest data using previous legacy architectures, the Snowflake AI Data Cloud enables data ingestion in days, supporting the work of healthcare professionals in various roles in a secure environment.
“A lot of the projects are research, but we also focus on direct care,” he says. “So, for example, there are several dashboards that are supporting our clinicians in understanding patients and high-risk cohorts. So it’s direct care research, but it’s also about operational decision-making and efficiencies.”
Our digital transformation was about bringing the right multidisciplinary people together to work collaboratively in a secure way Erik Mayer, Imperial College Healthcare NHS Trust
The organisation is also tapping into the Snowflake Marketplace, an online platform where users source third-party data for its use in the AI Data Cloud. Through the marketplace, research and clinical teams have access to additional non-health data for research and clinical care. Potential sources include Ordnance Survey and the Met Office.
“This is an interesting area for us,” says Mayer, referring to the use of marketplace data. “We’re just starting on this journey. With some of the data, for example, you can start to understand where people live, what services they’re accessing, and why.”
This in-depth detail will be crucial as organisations attempt to support the long-term vision of the NHS 10-year health plan for neighbourhood-based healthcare services.
“You have the evidence to show what is happening, so you can start to plan better,” he says. “Bringing together data is now a way to help us support hospitals and the community.”
Mayer and his colleagues are exploring other ways to exploit the platform. One key use case is federation, including how other trusts in north-west London can share primary and secondary care data. Another use case is artificial intelligence (AI). The data team’s AI testbed in its SDE is supported by Snowflake and Accenture, with secure access to Microsoft AI services and models.
“If you can leave the source data in the separate SDEs and then federate to allow algorithms to run across those sources, you’re not duplicating the cost and resources,” he says.
“So, that’s the piece we’re just developing across environments, which will support, again, operational efficiency, direct care and, of course, research.”
“The interoperability piece for sharing information on individual patients across healthcare providers is critical,” he says. “Just in terms of time savings, you’re not having to sit there trying to understand what’s happened so far – it’s all linked up. And I’m seeing that in my practice now, it’s happening. That kind of federation is a game-changer.”
“These initiatives have gone a long way already to providing a front door for access to data with an explanation around what it is, the clinical definitions and the metadata,” he says.
“We don’t need to reinvent the wheel. We need to build on what we’ve got so far, because that effort has been developing well over the last three or four years.”
This progress includes work in his own organisation. “Within the iCARE SDE, we have built the London analytics platform,” he says.
“We are one half of the London Secure Data Environment, so we are providing that architecture, and the data will start flowing soon. This effort is not just about our trust. It’s a framework that will support the national agenda.”
The NHS has a chequered history when it comes to IT initiatives. However, Mayer says the progress that’s been made recently in data-led projects is impressive. While digital transformation across large-scale organisations can be a challenging process, he’s positive about the opportunities ahead for UK healthcare.
“Now, I think there’s a requirement for a careful thought piece around how local NHS trusts fund, resource and keep up with change, and thinking about business intelligence units, and how those areas start to shape up,” he says.
Leading transformation
Mayer reflects on the pace of change. He suggests the speed of transformation continues to quicken and that AI will play a crucial role in the future of healthcare.
From optimising schedules to reducing the administrative burden by automating clinician note-taking, emerging technologies can have a big impact. However, the key to success is identifying the right, trusted technological solutions for the business challenges.
“We need to think about the problem and the opportunity, and then look at the technology to support us, as opposed to going, ‘AI is going to solve everything’,” he says.
“We’ve got to maintain the public trust around this transformation. They’re starting to engage with these technologies, so we must consider the digital literacy piece.”
This rapid pace of change can bring new and unexpected challenges to healthcare technologists. Mayer says effective digital leaders will develop professional resilience and respect by building a sense of social capital.
“It’s about being clear about the benefits and impact of what everyone’s doing as a multi-disciplinary team,” he says. “Our team includes data engineers, data scientists, clinicians and nurses. If they can feel, metaphorically, the impact and see the effect on care delivery, then they know they’re making a difference.”
Mayer says those results also inspire him. “As a leader, I see that impact, and that’s what gets me out of bed every day,” he says. “Essentially, successful delivery is about that team environment – it’s having a clear message and clear social capital where you say, ‘This is what we’re trying to do and why’.”
B&H Photo is one of our favorite places to shop for camera gear. If you’re ever in New York, head to the store to check out the giant overhead conveyor belt system that brings your purchase from the upper floors to the registers downstairs (yes, seriously, here’s a video). Fortunately B&H Photo’s website is here for the rest of us with some good deals on photo gear we love.
One of the many perks of B&H Photo is their generous free shipping policy. Most orders at B&H over $49 qualify for free expedited shipping to the lower 48 states. The savings just keep coming, as most items totaling under $49 also qualify for free standard shipping in the contiguous US.If for some reason your order doesn’t qualify for free shipping, you can review alternative shipping options during checkout—full shipping policy details can be found here.
Best Times to Find B&H Promo Codes and Discounts
Like most retailers, B&H Photo offers some of their best deals of the year during Black Friday and Cyber Monday, but that doesn’t mean there aren’t other ways to save outside of this sale period. There’s still time to jump on extended holiday sales, with up to $300 24-hour camera discounts, and featured price drops across all accessories. While there may not be any B&H promo codes available at the moment, there are plenty of other ways to save at B&H.
If you’re a student with an EDU email, sign up for B&H Photo’s student discount program, which offers free shipping on most orders and exclusive discounts.
Score Deals on Used and Refurbished Gear
B&H Photo also deals in used and refurbished gear. Deals (and conditions) vary, and I have never purchased a used item this way, but if you’re looking to save some money, that’s another way to go.
Trade-in Your Old Gear at B&H
The flip side of B&H Photos used deals is that you can sell your old gear. I put in my old Sony a7 II and was offered $210, which is more than I would have thought. Your offer is contingent on it matching the condition you claim, but if you’ve got gear you aren’t using anymore, this is a way to turn it into some extra cash.
Snap Up Limited-Time Deals in the Deal Zone
There are always rotating and limited-time B&H Photos deals at this url, which B&H Photo calls the Deal Zone.