National Robotics Week — Latest Physical AI Research, Breakthroughs and Resources

National Robotics Week — Latest Physical AI Research, Breakthroughs and Resources

This National Robotics Week , NVIDIA is highlighting the breakthroughs that are bringing AI into the physical world — as well as the growing wave of robots transforming industries, from agricultural and manufacturing to energy and beyond.

Advancements in robot learning, simulation and foundation models are accelerating development, enabling robots to move from training in virtual environments to real-world deployment faster than ever.

With NVIDIA platforms for simulation , synthetic data and AI-powered robot learning , developers now have the tools to build machines that can perceive, reason and act in complex environments.

Check back here all week for coverage on the latest NVIDIA physical AI technologies.

Underwater simulators are crucial for developing reliable perception systems, but they still struggle with accurate physics‑based sensor modeling and fast rendering.

Helping close this gap is OceanSim , a GPU‑accelerated, high‑fidelity simulator developed by researchers at the University of Michigan. It uses advanced physics‑based rendering techniques to make synthetic underwater images look more realistic. Using GPUs, the simulator can render imaging sonar in real time and generate synthetic data quickly.

OceanSim uses NVIDIA Isaac Sim and plugs into NVIDIA Omniverse libraries, creating a seamless link between robot‑learning research and underwater robotics. This integration lets developers easily develop and deploy embodied AI techniques for underwater applications.

RoboLab is a high-fidelity simulation benchmark for developing and evaluating generalist robot policies — powering systems designed to perform diverse tasks across environments.

Built on NVIDIA Isaac and NVIDIA Omniverse simulation technologies, RoboLab taps into photorealistic environments and physics-based modeling to train and test robotic policies at scale. This enables researchers to measure how well behaviors learned in simulation transfer to the real world as tasks grow in complexity.

By combining advanced simulation with structured evaluation, RoboLab accelerates the path from virtual training to real-world deployment.

RoboLab features will be incorporated into the roadmap of NVIDIA Isaac Lab-Arena , an open source framework for large-scale policy setup and evaluation.

In warehouse environments, palletizing robots typically follow fixed rules — handling boxes the same way regardless of contents, condition or fragility. A project developed by Doosan Robotics introduces a more adaptive approach using NVIDIA Cosmos Reason .

By analyzing a single camera image, the system can infer box contents, detect damage and adjust how each item is handled — such as placement, speed and grip — based on estimated weight and fragility. This reduces common issues like incorrectly stacking damaged or fragile goods.

To build robots that understand the physical world before they ever deploy in it, robotics researchers and developers are building policy models powered by NVIDIA Cosmos world foundation models (WFMs). Toyota Research Institute customizes Cosmos WFMs for their own world model to achieve state-of-the-art results across dynamic view synthesis, state-of-the-art teleoperation data augmentation and navigation world models.

Mimic robotics takes a different angle with mimic-video, a video-action model that pairs a pretrained internet-scale video model with a flow-matching action decoder, replacing the static image-language backbones of traditional VLAs with video-learned physical dynamics — achieving 10x better sample efficiency and 2x faster convergence on real-world manipulation tasks.

Together, both teams demonstrate a fundamental shift: robots trained on world models that capture physics and causality need dramatically less real-world data to perform reliably in conditions they’ve never seen.

This National Robotics Week, OpenClaw running on the NVIDIA Jetson platform showcases how quickly open source innovation is evolving into real-world, intelligent robotics.

From practical applications to innovative projects, the robotics community is building what’s next — and fast.

Developers are pushing the boundaries of autonomy — including hardware-in-the-loop testing powered by Jetson Thor , evaluating camera streams from NVIDIA Isaac Sim and even building systems that can generate their own code to complete tasks.

In addition, OpenClaw now running entirely locally on NVIDIA Jetson Thor — powered by optimized NVIDIA Nemotron open models and the vLLM open inference library — marks a major leap toward private, low-latency edge AI for robotics. And innovations like the NVIDIA NemoClaw stack on Jetson are expanding what’s possible at the intersection of open source and high-performance robotics platforms.

Gennady Plyushchev, a robotics creator known as Skyentific, is documenting the process of building a walking bipedal robot, from simulation and design to real-world deployment — showcasing a simulation-first approach to robot development.

By using NVIDIA Isaac based simulation workflows alongside NVIDIA Jetson for on-device AI and control, the project demonstrates how developers can rapidly iterate in virtual environments before deploying to physical systems.

The result highlights a broader shift in robotics: using AI, simulation and edge computing to accelerate development and bring increasingly capable humanoid robots to life.

To bring robots into everyday life, researchers at the University of Maryland , recipients of a grant from the NVIDIA Academic Grant Program , are developing AI-powered humanoid systems capable of performing complex household tasks with greater autonomy.

NVIDIA RTX PRO 6000 Blackwell GPUs for training large models and NVIDIA Jetson AGX Thor developer kits for efficient deployment on physical robots help bridge the gap between research and real-world applications.

The second cohort of the Amazon Web Services (AWS) MassRobotics fellowship comprises startups being recognized for compelling industrial use cases harnessing robotics and computer vision . They will receive access to technical resources and AWS cloud credits.

The cohort includes NVIDIA Inception members Burro, Config Intelligence, Deltia, Haply Robotics, Luminous Robotics, Roboto AI, Telexistence, Terra Robotics and WiRobotics, each developing technologies spanning humanoid robotics, industrial automation, haptics and agricultural systems.

Burro creates autonomous agricultural robots for tasks like grape harvesting and crop scouting.

Config Intelligence builds data infrastructure for general-purpose bimanual robotics to enable reliable two-handed tasks in real-world settings.

Deltia provides AI-driven manufacturing intelligence that optimizes assembly lines using computer vision and analytics.

Haply Robotics designs haptic control devices that serve as “steering wheels” for physical AI systems across industries.

Luminous Robotics deploys AI-powered robotic systems for fast, low-cost solar-panel installation and maintenance.

Roboto AI offers a data-analytics platform that accelerates robot development by managing and analyzing robotics data.

Telexistence develops AI-powered humanoid robots and remote-controlled systems for retail and logistics.

Terra Robotics develops laser-weeding agricultural robots to automate sustainable farming.

WiRobotics creates wearable walking-assist and humanoid robots to enhance mobility and physical interaction, using training data from assisted products to train its humanoids.

Maximo , a solar robotics business incubated within The AES Corporation, recently completed a 100-megawatt solar installation using its robot fleet. Developed with NVIDIA accelerated computing, NVIDIA Omniverse libraries and the NVIDIA Isaac Sim framework , Maximo demonstrated that autonomous installations can operate reliably for utility-scale projects.

The solution improves installation speed, safety and consistency, helping close the gap between rising demand for faster time to power and construction capacity.

As solar expansion faces ongoing labor constraints and rising demand, AI-driven field robotics systems like Maximo are helping accelerate infrastructure buildout, reduce costs and redefine how energy projects are delivered.

To help regenerate the Earth, Aigen’s solar-powered autonomous robots are breaking farmers’ dependency on chemicals through precision weed control powered by vision AI.

The NVIDIA Inception startup is building a new kind of farming system that’s powered by clean energy and continuously enriched by data. Aigen’s fleet of solar-driven rovers uses advanced computer vision to identify and remove weeds, dramatically reducing the need for herbicides.

Farming has no standard environment. Every field is different — different crops, different soil, different equipment, weeds, growth stages and geographies. That fragmentation makes real-world data collection slow, expensive and inconsistent. By post-training NVIDIA Cosmos open world foundation models on their specialized data and harnessing NVIDIA Isaac Sim pipelines, Aigen is building the system that generalizes for millions of agriculture scenarios.

On the ground, each rover runs inference using an NVIDIA Jetson Orin edge AI module to distinguish crops from weeds in real time.

Using these rovers, farmers can grow crops more sustainably and profitably, using regenerative practices that heal the land and foster ecological balance.

Learn about the breakthroughs shaping the next chapter of AI anytime, anywhere.

Key considerations

  • Investor positioning can change fast
  • Volatility remains possible near catalysts
  • Macro rates and liquidity can dominate flows

Reference reading

More on this site

Informational only. No financial advice. Do your own research.

Leave a Comment