top of page
Google PaLM-E
Embodied multimodal language model for robotic control and perception:
Integrates vision and language inputs to enable robots to understand and execute complex natural language commands
Demonstrates improved generalization across different robotic tasks without the need for retraining
Processes multimodal inputs, including images and sensor data, to generate context-aware actions
Showcases the power of combining large language models with visual perception for advanced robotics applications
No Reviews YetShare your thoughts.
Be the first to leave a review.
bottom of page


