Google DeepMind shared on Thursday a analysis preview of SIMA 2, the subsequent technology of its generalist AI agent that integrates the language and reasoning powers of Gemini, Googleβs massive language mannequin, to maneuver past merely following directions to understanding and interacting with its setting.Β
Like a lot of DeepMindβs tasks, together with AlphaFold, the primary model of SIMA was skilled on a whole bunch of hours of online game information to discover ways to play a number of 3D video games like a human, even some video games it wasnβt skilled on. SIMA 1, unveiled in March 2024, may comply with primary directions throughout a variety of digital environments, nevertheless it solely had a 31% success charge for finishing advanced duties, in comparison with 71% for people.Β Β Β
βSIMA 2 is a step change and enchancment in capabilities over SIMA 1,β Joe Marino, senior analysis scientist at DeepMind, stated in a press briefing. βItβs a extra normal agent. It could possibly full advanced duties in beforehand unseen environments. And itβs a self-improving agent. So it could actually truly self-improve based mostly by itself expertise, which is a step in direction of extra general-purpose robots and AGI programs extra usually.β
SIMA 2 is powered by the Gemini 2.5 flash-lite mannequin, and AGI refers to synthetic normal intelligence, which DeepMind defines as a system able to a variety of mental duties with the flexibility to study new expertise and generalize information throughout completely different areas.Β
Working with so-called βembodied brokersβ is essential to generalized intelligence, DeepMindβs researchers say. Marino defined that an embodied agent interacts with a bodily or digital world through a physique β observing inputs and taking actions very similar to a robotic or human would β whereas a non-embodied agent may work together along with your calendar, take notes, or execute code.Β
Jane Wang, a senior workers analysis scientist at DeepMind with a background in neuroscience, instructed Trendster that SIMA 2 goes far past gameplay.Β
βWeβre asking it to truly perceive whatβs occurring, perceive what the person is asking it to do, after which be capable to reply in a commonsense means thatβs truly fairly tough,β Wang stated.Β
Techcrunch occasion
San Francisco
|
October 13-15, 2026
By integrating Gemini, SIMA 2 doubled its predecessorβs efficiency, uniting Geminiβs superior language and reasoning talents with the embodied expertise developed by way of coaching.
Marino demoed SIMA 2 in βNo Manβs Sky,β the place the agent described its environment β a rocky planet floor β and decided its subsequent steps by recognizing and interacting with a misery beacon. SIMA 2 additionally makes use of Gemini to purpose internally. In one other recreation, when requested to stroll to the home thatβs the colour of a ripe tomato, the agent confirmed its pondering β ripe tomatoes are crimson, due to this fact I ought to go to the crimson home β then discovered and approached it.
Being Gemini-powered additionally means SIMA 2 follows directions based mostly on emojis: βYou instruct it πͺπ², and itβll go chop down a tree,β Marino stated.Β
Marino additionally demonstrated how SIMA 2 can navigate newly generated photorealistic worlds produced by Genie, DeepMindβs world mannequin, accurately figuring out and interacting with objects like benches, bushes, and butterflies.Β
Gemini additionally allows self-improvement with out a lot human information, Marino added. The place SIMA 1 was skilled solely on human gameplay, SIMA 2 makes use of it as a baseline to supply a robust preliminary mannequin. When the staff places the agent into a brand new setting, it asks one other Gemini mannequin to create new duties and a separate reward mannequin to attain the agentβs makes an attempt. Utilizing these self-generated experiences as coaching information, the agent learns from its personal errors and progressively performs higher, basically educating itself new behaviors by way of trial and error as a human would, guided by AI-based suggestions as a substitute of people.
DeepMind sees SIMA 2 as a step towards unlocking extra general-purpose robots.
βIf we consider what a system must do to carry out duties in the true world, like a robotic, I feel there are two elements of it,β Frederic Besse, senior workers analysis engineer at DeepMind, stated throughout a press briefing. βFirst, there’s a high-level understanding of the true world and what must be accomplished, in addition to some reasoning.β
If you happen to ask a humanoid robotic in your own home to go test what number of cans of beans you’ve within the cabinet, the system wants to grasp the entire completely different ideas β what beans are, what a cabinet is β and navigate to that location. Besse says SIMA 2 touches extra on that high-level conduct than it does on lower-level actions, which he refers to as controlling issues like bodily joints and wheels.
The staff declined to share a particular timeline for implementing SIMA 2 in bodily robotics programs. Besse instructed Trendster that DeepMindβs just lately unveiled robotics basis fashions β which might additionally purpose in regards to the bodily world and create multi-step plans to finish a mission β had been skilled otherwise and individually from SIMA.Β
Whereas thereβs additionally no timeline for releasing greater than a preview of SIMA 2, Wang instructed Trendster the objective is to indicate the world what DeepMind has been engaged on and see what sorts of collaborations and potential makes use of are attainable.





