Google’s PaLM-E is a generalist robot brain that takes commands

Enlarge / A robotic arm controlled by PaLM-E reaches for a bag of chips in a demonstration video.Google Research

On Monday, a group of AI researchers from Google and the Technical University of Berlin unveiled PaLM-E, a multimodal embodied visual-language model (VLM) with 562 billion parameters that integrates vision and language for robotic control. They claim it is the largest VLM ever developed and that it can perform a variety of tasks without the need for retraining.

According to Google, when given a high-level command, such as “bring me the rice chips from the drawer,” PaLM-E can generate a plan

→ Continue reading at Ars Technica

Related articles

Comments

Share article

Latest articles