Near future
No Result
View All Result
27 September 2023
  • Home
  • Tech
  • Health
  • Environment
  • Energy
  • Transports
  • Spazio
  • AI
  • concepts
  • H+
Understand, anticipate, improve the future.
CES2023 / Coronavirus / Russia-Ukraine
Near future
  • Home
  • Tech
  • Health
  • Environment
  • Energy
  • Transports
  • Spazio
  • AI
  • concepts
  • H+

Understand, anticipate, improve the future.

No Result
View All Result
Robotica, Technology

DeepMind presents RT-2: robots that see, learn and act

The AI ​​model developed by DeepMind, which combines vision and language to control machines, will open new horizons in robotics.

July 30 2023
Gianluca RiccioGianluca Riccio
⚪ 4 minutes
SharePin1TweetSendShareShareShare

READ IN:

In a bright environment, full of monitors and technological equipment, a robot stands as the protagonist. Its metal structure reflects the light, but it is in its "eyes" that the real magic is hidden. These eyes, powered by DeepMind's RT-2 model, are capable of seeing, interpreting and acting.

As the robot moves gracefully, the scientists around him scrutinize his every move. It's not just a piece of metal and circuitry, but the embodiment of an intelligence that unites the vast world of the web with tangible reality.

Deepmind
One of 13 robots used for model training.

The evolution of RT-2

Robotics has come a long way in recent years, but DeepMind it just took the game to a whole new level. Illustrated in a paper just released it arrives RT-2. Things? It is a vision-language-action (VLA) model that not only learns from web data, but also from robotic data, translating this knowledge into generalized instructions for robotic control.

In an era where technology advances by leaps and bounds, RT-2 represents a significant leap, promising to revolutionize not only the field of robotics, but also the way we live and work every day. But what does that mean in practice?

The article continues after the related links

Without a brake from the institutions, will AI lead to private neo-feudalism?

TeddyGPT arrives, the first AI teddy bear that talks and learns with children

DeepMind RT-2, from vision to action

The models of high capacity vision-language (VLM) they are trained on large datasets, and this also makes them extraordinarily good at recognizing visual or linguistic patterns (operating, for example, in different languages). But imagine being able to make robots do what these models do. Indeed, stop imagining it: DeepMind is making it possible with RT-2.

Robotic Transformers 1 (RT-1) it was a marvel in its own right, but RT-2 goes further, displaying enhanced generalization capabilities and semantic and visual understanding that goes beyond the robotic data it has been exposed to.

DeepMind
A visual language model (VLM) pre-trained with web data also learns from RT-1 robotics data. Result? RT-2, a visual language action model (VLA) capable of controlling a robot - Animation: DeepMind

Chain reasoning

One of the most fascinating aspects of RT-2 is its chain reasoning ability. He can decide what object could be used as a makeshift hammer or what kind of drink is best for a tired person. This deep reasoning ability could revolutionize the way we interact with robots.

And at worst, you could still ask a robot to make you a good coffee to regain some lucidity.

But how does DeepMind RT-2 control a robot?

The answer lies in how he was trained. In fact, it uses a representation not unlike the language tokens that are exploited by templates like ChatGPT.

RT-2 demonstrated amazing emergent capabilities, such as symbol understanding, reasoning and human recognition. Skills that currently show an improvement of more than 3x compared to previous models.

With RT-2, DeepMind not only showed that vision-language models can be transformed into powerful vision-language-action models, but it also opened the door to a future in which robots can reason, solve problems and interpret information to perform a wide range of tasks in the real world.

DeepMind
An example of the activities learned by DeepMind RT-2. Tasks that weren't in the data it was trained with.

And now?

In a world where artificial intelligence and robotics will be increasingly central, RT-2 shows us that the next evolution will not be purely technical, but "perceptual". Machines will understand and respond to our needs in ways we never imagined.

If this is just the beginning, who knows what the future holds.

Tags: Deepmindartificial intelligencerobot

Latest news

  • Amazon invests in Anthropic: the AI ​​war rages on
  • Smile: Drug that regrows teeth reaches clinical trials
  • Without a brake from the institutions, will AI lead to private neo-feudalism?
  • When will we become extinct? Here comes a (nice) prediction from the supercomputer
  • DNA and longevity: how 'jumping genes' could extend our lives
  • TeddyGPT arrives, the first AI teddy bear that talks and learns with children
  • Minimum wage: a step forward or a leap in the dark for Italy?
  • Solar cars could halve the need for charging
  • The crazy food of the future: from edible clouds to mussel ceramics
  • NASA, recovered the sample from the asteroid Bennu: why it is important


GPT Chat Megaeasy!

Concrete guide for those approaching this artificial intelligence tool, also designed for the school world: many examples of applications, usage indications and ready-to-use instructions for training and interrogating Chat GPT.

To submit articles, disclose the results of a research or scientific discoveries write to the editorial staff

Enter the Telegram channel of Futuroprossimo, click here. Or follow us on Instagram, Facebook, Twitter, Mastodon e LinkedIn.

FacebookTwitterInstagramTelegramLinkedInMastodonPinterestTikTok

The daily tomorrow.


Futuroprossimo.it provides news on the future of technology, science and innovation: if there is something that is about to arrive, here it has already arrived. FuturoProssimo is part of the network ForwardTo, studies and skills for future scenarios.

  • Environment
  • Architecture
  • Artificial intelligence
  • Gadgets
  • concepts
  • Design
  • Medicine
  • Spazio
  • Robotica
  • Work
  • Transports
  • Energy
  • Edition Francaise
  • Deutsche Ausgabe
  • Japanese version
  • English Edition
  • Portuguese Edition
  • Read more
  • Spanish edition

Subscribe to our newsletter

  • The Editor
  • Advertising on FP
  • Privacy Policy

© 2023 Near future - Creative Commons License
This work is distributed under license Creative Commons Attribution 4.0 International.

No Result
View All Result
Understand, anticipate, improve the future.
  • Home
  • Tech
  • Health
  • Environment
  • Energy
  • Transports
  • Spazio
  • AI
  • concepts
  • H+