AI that learns like humans.

Embodied Science is a research lab developing next generation Vision-Language Models with superior physical and spatial reasoning.

We translate our breakthroughs by partnering with organizations to deploy our models directly into their production systems.

What we do

Research

  • Spatial reasoning for vision-language models
  • Learning from observation + rapid adaptation
  • Foundations for embodied autonomy

Applied Partnerships

  • Data generation + evaluation pipelines
  • Model fine-tuning and benchmarking
  • Deployment-focused prototyping

Featured research

Highlight

GRAID

Enhancing Spatial Reasoning of VLMs through High-Fidelity Data Generation

  • Custom Spatial VQA from your own images and object detector
  • ⁠Avoids LLM-based QA hallucinations and inaccurate single-view 3D reconstruction
  • Scales to 8.5M+ VQA pairs with ~91.16% human-validated accuracy, significantly outperforming Google DeepMind’s SpatialVLM
8.5M+
VQA pairs generated
~91.16%
Human-validated accuracy
1400×
Generation speedups with SPARQ
2D
Detector-output only

Domains

Manufacturing

Precision spatial understanding for industrial automation and quality control.

Robotics

Embodied agents that perceive and reason about physical environments.

Autonomous Driving

Geometric scene understanding for navigation and safety critical decisions.

Medical Image Analysis

Spatial reasoning for diagnostic imaging, anatomical segmentation, and clinical decision support.

Sports Analysis

Vision-based tracking, spatial understanding of play dynamics, and performance analytics.

Lotus TechnologiesA Lotus Technologies Website