Skip to main content

Chapter 9: Isaac ROS & VSLAM

Building on Isaac Sim from Chapter 8, we now explore NVIDIA's GPU-accelerated perception stack. Isaac ROS brings the power of NVIDIA GPUs to robot perception, enabling real-time Visual SLAM, 3D reconstruction, and AI-powered sensing that runs orders of magnitude faster than CPU alternatives.


Learning Objectivesโ€‹

By the end of this chapter, you will be able to:

  • Explain the Isaac ROS architecture and its advantages
  • Set up Isaac ROS packages in a ROS 2 workspace
  • Implement Visual SLAM using cuVSLAM for robot localization
  • Build 3D maps in real-time using nvblox
  • Integrate depth cameras and stereo vision with Isaac ROS
  • Optimize perception pipelines for GPU acceleration
  • Test SLAM systems in Isaac Sim before real-world deployment

Introduction to Isaac ROSโ€‹

What is Isaac ROS?โ€‹

Isaac ROS is NVIDIA's collection of GPU-accelerated ROS 2 packages designed for high-performance robot perception:

โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
โ”‚ Isaac ROS Ecosystem โ”‚
โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
โ”‚ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Isaac ROS Packages โ”‚ โ”‚
โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚
โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ cuVSLAM โ”‚ โ”‚ nvblox โ”‚ โ”‚ DNN โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ Visual SLAM โ”‚ โ”‚ 3D Mapping โ”‚ โ”‚ Inference โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ AprilTag โ”‚ โ”‚ Stereo โ”‚ โ”‚ Freespace โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ Detection โ”‚ โ”‚ Depth โ”‚ โ”‚ Segmentat. โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ NITROS (NVIDIA ROS) โ”‚ โ”‚
โ”‚ โ”‚ Zero-copy GPU memory transport โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ ROS 2 Humble โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

Why GPU-Accelerated Perception?โ€‹

TaskCPU PerformanceIsaac ROS (GPU)Speedup
Visual SLAM10-15 FPS60+ FPS4-6x
3D Reconstruction2-5 Hz30+ Hz6-15x
Stereo Depth15 FPS90+ FPS6x
Object Detection5-10 FPS60+ FPS6-12x
Semantic Segmentation2-5 FPS30+ FPS6-15x

NITROS: Zero-Copy Data Transportโ€‹

Isaac ROS uses NITROS (NVIDIA ROS) for efficient GPU data handling:

โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
โ”‚ Traditional ROS 2 vs NITROS Pipeline โ”‚
โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
โ”‚ โ”‚
โ”‚ Traditional ROS 2: โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚Cameraโ”‚โ”€โ”€โ”€โ–ถโ”‚ CPU โ”‚โ”€โ”€โ”€โ–ถโ”‚ GPU โ”‚โ”€โ”€โ”€โ–ถโ”‚ CPU โ”‚โ”€โ”€โ”€โ–ถOutput โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚Copy โ”‚ โ”‚Proc โ”‚ โ”‚Copy โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ–ฒ โ”‚ โ–ฒ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ Memory copies = SLOW โ”‚
โ”‚ โ”‚
โ”‚ NITROS (Zero-Copy): โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚Cameraโ”‚โ”€โ”€โ”€โ–ถโ”‚ GPU Memory โ”‚โ”€โ”€โ”€โ–ถOutput โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ (stays on GPU) โ”‚ โ”‚
โ”‚ โ”‚ cuVSLAM โ†’ nvblox โ†’ DNN โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ No memory copies = FAST โ”‚
โ”‚ โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

Setting Up Isaac ROSโ€‹

System Requirementsโ€‹

ComponentRequirement
OSUbuntu 22.04
ROS 2Humble Hawksbill
GPUNVIDIA GPU with Compute Capability 7.0+
CUDA11.8 or 12.x
Docker20.10+ with NVIDIA Container Toolkit
# Clone Isaac ROS Common
mkdir -p ~/workspaces/isaac_ros-dev/src
cd ~/workspaces/isaac_ros-dev/src
git clone https://github.com/NVIDIA-ISAAC-ROS/isaac_ros_common.git

# Clone required packages
git clone https://github.com/NVIDIA-ISAAC-ROS/isaac_ros_visual_slam.git
git clone https://github.com/NVIDIA-ISAAC-ROS/isaac_ros_nvblox.git
git clone https://github.com/NVIDIA-ISAAC-ROS/isaac_ros_image_pipeline.git

# Launch the Isaac ROS development container
cd ~/workspaces/isaac_ros-dev/src/isaac_ros_common
./scripts/run_dev.sh

Installation Option 2: Native Buildโ€‹

# Install dependencies
sudo apt-get update
sudo apt-get install -y \
ros-humble-isaac-ros-visual-slam \
ros-humble-isaac-ros-nvblox \
ros-humble-isaac-ros-image-pipeline

# Or build from source
cd ~/ros2_ws/src
git clone https://github.com/NVIDIA-ISAAC-ROS/isaac_ros_visual_slam.git
git clone https://github.com/NVIDIA-ISAAC-ROS/isaac_ros_nvblox.git

cd ~/ros2_ws
colcon build --packages-up-to isaac_ros_visual_slam isaac_ros_nvblox
source install/setup.bash

Verifying Installationโ€‹

# Check CUDA availability
nvidia-smi

# List Isaac ROS packages
ros2 pkg list | grep isaac

# Check cuVSLAM node
ros2 run isaac_ros_visual_slam isaac_ros_visual_slam_node --help

Understanding Visual SLAMโ€‹

What is Visual SLAM?โ€‹

Visual Simultaneous Localization and Mapping (VSLAM) uses camera images to:

  1. Localize: Determine the robot's position and orientation
  2. Map: Build a representation of the environment
  3. Track: Follow visual features across frames
โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
โ”‚ VSLAM Pipeline โ”‚
โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
โ”‚ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Camera โ”‚ โ”‚
โ”‚ โ”‚ Frames โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚
โ”‚ โ–ผ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Feature โ”‚โ”€โ”€โ”€โ–ถโ”‚ Feature โ”‚โ”€โ”€โ”€โ–ถโ”‚ Pose โ”‚ โ”‚
โ”‚ โ”‚ Detection โ”‚ โ”‚ Matching โ”‚ โ”‚ Estimation โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ–ผ โ–ผ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Map โ”‚โ—€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”‚ Bundle โ”‚ โ”‚
โ”‚ โ”‚ Points โ”‚ โ”‚ Adjustment โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ–ผ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Output: โ”‚ โ”‚
โ”‚ โ”‚ Pose + Map โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

VSLAM vs Traditional SLAMโ€‹

AspectLidar SLAMVisual SLAM
Sensor Cost$1,000 - $10,000+$50 - $500
Power ConsumptionHighLow
Information DensitySparse geometryRich texture + geometry
Lighting DependencyNoneRequires adequate light
Feature TypesDistance measurementsVisual features
Best ForOutdoor, industrialIndoor, robotics

Types of Visual SLAMโ€‹

โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
โ”‚ Visual SLAM Approaches โ”‚
โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
โ”‚ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Feature-Based SLAM โ”‚ โ”‚
โ”‚ โ”‚ โ€ข Extracts keypoints (ORB, SIFT, SURF) โ”‚ โ”‚
โ”‚ โ”‚ โ€ข Sparse map representation โ”‚ โ”‚
โ”‚ โ”‚ โ€ข Fast, works on CPU โ”‚ โ”‚
โ”‚ โ”‚ โ€ข Examples: ORB-SLAM, VINS โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Direct SLAM โ”‚ โ”‚
โ”‚ โ”‚ โ€ข Uses pixel intensities directly โ”‚ โ”‚
โ”‚ โ”‚ โ€ข Dense/semi-dense maps โ”‚ โ”‚
โ”‚ โ”‚ โ€ข More accurate, computationally expensive โ”‚ โ”‚
โ”‚ โ”‚ โ€ข Examples: LSD-SLAM, DSO โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Deep Learning SLAM (cuVSLAM) โ”‚ โ”‚
โ”‚ โ”‚ โ€ข Learned feature extraction โ”‚ โ”‚
โ”‚ โ”‚ โ€ข GPU-accelerated processing โ”‚ โ”‚
โ”‚ โ”‚ โ€ข Robust to challenging conditions โ”‚ โ”‚
โ”‚ โ”‚ โ€ข Examples: cuVSLAM, DROID-SLAM โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

cuVSLAM: GPU-Accelerated Visual SLAMโ€‹

What is cuVSLAM?โ€‹

cuVSLAM is NVIDIA's GPU-accelerated Visual SLAM implementation that provides:

  • Real-time visual odometry at 60+ FPS
  • Loop closure detection and correction
  • Multi-camera support (stereo, RGB-D)
  • IMU fusion for improved accuracy
  • Relocalization after tracking loss

cuVSLAM Architectureโ€‹

โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
โ”‚ cuVSLAM Architecture โ”‚
โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
โ”‚ โ”‚
โ”‚ Inputs: โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Stereo โ”‚ โ”‚ RGB-D โ”‚ โ”‚ IMU โ”‚ โ”‚
โ”‚ โ”‚ Camera โ”‚ โ”‚ Camera โ”‚ โ”‚ (opt.) โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ–ผ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ GPU Processing Pipeline โ”‚ โ”‚
โ”‚ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ Feature โ”‚โ”€โ–ถโ”‚ Stereo โ”‚โ”€โ–ถโ”‚ Visual Odometryโ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ Extraction โ”‚ โ”‚ Matching โ”‚ โ”‚ Estimation โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ–ผ โ–ผ โ”‚ โ”‚
โ”‚ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ Loop โ”‚โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–ถโ”‚ Pose โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ Closure โ”‚ โ”‚ Graph โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚
โ”‚ โ–ผ โ”‚
โ”‚ Outputs: โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Visual Odom โ”‚ โ”‚ Pose Graph โ”‚ โ”‚ Landmark Cloud โ”‚ โ”‚
โ”‚ โ”‚ (30-60 Hz) โ”‚ โ”‚ (on demand) โ”‚ โ”‚ (sparse map) โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

Launching cuVSLAMโ€‹

Basic Launch (Stereo Camera)โ€‹

ros2 launch isaac_ros_visual_slam isaac_ros_visual_slam_realsense.launch.py

Custom Launch Fileโ€‹

# File: cuVSLAM_launch.py
from launch import LaunchDescription
from launch_ros.actions import Node
from launch.actions import DeclareLaunchArgument
from launch.substitutions import LaunchConfiguration

def generate_launch_description():
# Parameters
enable_imu_fusion = LaunchConfiguration('enable_imu_fusion')
enable_debug_mode = LaunchConfiguration('enable_debug_mode')

return LaunchDescription([
# Arguments
DeclareLaunchArgument(
'enable_imu_fusion',
default_value='false',
description='Enable IMU sensor fusion'
),
DeclareLaunchArgument(
'enable_debug_mode',
default_value='false',
description='Enable debug visualization'
),

# cuVSLAM Node
Node(
package='isaac_ros_visual_slam',
executable='isaac_ros_visual_slam_node',
name='visual_slam',
parameters=[{
'denoise_input_images': True,
'rectified_images': True,
'enable_imu_fusion': enable_imu_fusion,
'gyro_noise_density': 0.000244,
'gyro_random_walk': 0.000019393,
'accel_noise_density': 0.001862,
'accel_random_walk': 0.003,
'calibration_frequency': 200.0,
'image_jitter_threshold_ms': 34.0,
'enable_debug_mode': enable_debug_mode,
}],
remappings=[
('stereo_camera/left/image', '/camera/left/image_raw'),
('stereo_camera/left/camera_info', '/camera/left/camera_info'),
('stereo_camera/right/image', '/camera/right/image_raw'),
('stereo_camera/right/camera_info', '/camera/right/camera_info'),
('visual_slam/imu', '/imu'),
]
),
])

cuVSLAM Topics and Servicesโ€‹

# Published Topics
/visual_slam/tracking/odometry # nav_msgs/Odometry (high-rate pose)
/visual_slam/tracking/vo_pose # geometry_msgs/PoseStamped
/visual_slam/tracking/vo_pose_covariance # Pose with uncertainty
/visual_slam/vis/landmarks_cloud # sensor_msgs/PointCloud2 (sparse map)
/visual_slam/vis/observations_cloud # Current observations

# Services
/visual_slam/reset # Reset SLAM system
/visual_slam/save_map # Save current map
/visual_slam/load_map # Load existing map

# Example: Save map
ros2 service call /visual_slam/save_map isaac_ros_visual_slam_interfaces/srv/SaveMap \
"{map_url: '/tmp/my_slam_map'}"

Integrating with Robot Localizationโ€‹

# Fuse cuVSLAM with wheel odometry using robot_localization
from launch import LaunchDescription
from launch_ros.actions import Node

def generate_launch_description():
return LaunchDescription([
# EKF for sensor fusion
Node(
package='robot_localization',
executable='ekf_node',
name='ekf_filter_node',
parameters=[{
'frequency': 50.0,
'sensor_timeout': 0.1,
'two_d_mode': False,
'odom_frame': 'odom',
'base_link_frame': 'base_link',
'world_frame': 'odom',

# Wheel odometry
'odom0': '/wheel_odom',
'odom0_config': [
True, True, False, # x, y, z
False, False, True, # roll, pitch, yaw
True, True, False, # vx, vy, vz
False, False, True, # vroll, vpitch, vyaw
False, False, False # ax, ay, az
],

# Visual SLAM odometry
'odom1': '/visual_slam/tracking/odometry',
'odom1_config': [
True, True, True, # x, y, z
True, True, True, # roll, pitch, yaw
False, False, False, # vx, vy, vz
False, False, False, # vroll, vpitch, vyaw
False, False, False # ax, ay, az
],
}]
),
])

nvblox: Real-Time 3D Reconstructionโ€‹

What is nvblox?โ€‹

nvblox is NVIDIA's GPU-accelerated 3D reconstruction library that creates:

  • Signed Distance Fields (SDF): Continuous surface representation
  • Occupancy Maps: For navigation and collision avoidance
  • Mesh Reconstruction: Real-time surface meshes
  • ESDF (Euclidean Signed Distance Field): For path planning
โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
โ”‚ nvblox Pipeline โ”‚
โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
โ”‚ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Depth Image โ”‚ โ”‚ Robot Pose โ”‚ โ”‚
โ”‚ โ”‚ (Camera) โ”‚ โ”‚ (cuVSLAM) โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ–ผ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ nvblox GPU Pipeline โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ Depth โ”‚โ”€โ”€โ–ถโ”‚ TSDF โ”‚โ”€โ”€โ–ถโ”‚ Mesh โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ Integration โ”‚ โ”‚ Voxel Grid โ”‚ โ”‚ Extractionโ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ–ผ โ–ผ โ”‚ โ”‚
โ”‚ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ Occupancy โ”‚ โ”‚ ESDF โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ Grid โ”‚ โ”‚ (Planning) โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚
โ”‚ โ–ผ โ”‚
โ”‚ Outputs: โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ 3D Mesh โ”‚ โ”‚ Costmap โ”‚ โ”‚ ESDF for โ”‚ โ”‚
โ”‚ โ”‚ (visual) โ”‚ โ”‚ (Nav2) โ”‚ โ”‚ Planning โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

nvblox Conceptsโ€‹

TSDF (Truncated Signed Distance Field)โ€‹

โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
โ”‚ TSDF Representation โ”‚
โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
โ”‚ โ”‚
โ”‚ Surface โ”‚
โ”‚ โ”‚ โ”‚
โ”‚ โ–ผ โ”‚
โ”‚ โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€ โ”‚
โ”‚ โ”‚โ—€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–ถโ”‚ โ”‚
โ”‚ โ”‚ Truncation Distance โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ -1.0 โ”‚ Negative values โ”‚ +1.0 โ”‚ Positive values โ”‚
โ”‚ โ”‚ Insideโ”‚ (inside surface) โ”‚Outsideโ”‚ (outside surface)โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚
โ”‚ Voxel value = signed distance to nearest surface โ”‚
โ”‚ Zero-crossing = surface location โ”‚
โ”‚ โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

Launching nvbloxโ€‹

Basic Launchโ€‹

# With RealSense camera
ros2 launch nvblox_examples_bringup realsense_example.launch.py

# With Isaac Sim
ros2 launch nvblox_examples_bringup isaac_sim_example.launch.py

Custom Configurationโ€‹

# File: nvblox_mapping.launch.py
from launch import LaunchDescription
from launch_ros.actions import Node

def generate_launch_description():
return LaunchDescription([
Node(
package='nvblox_ros',
executable='nvblox_node',
name='nvblox',
parameters=[{
# Voxel parameters
'voxel_size': 0.05, # 5cm voxels
'esdf_update_rate_hz': 10.0,
'mesh_update_rate_hz': 5.0,

# Map parameters
'global_frame': 'odom',
'map_clearing_radius_m': 5.0,

# TSDF parameters
'tsdf_integrator_max_integration_distance_m': 5.0,
'tsdf_integrator_truncation_distance_vox': 4.0,

# Mesh parameters
'mesh_integrator_min_weight': 0.0001,
'mesh_integrator_weld_vertices': True,

# ESDF parameters
'esdf_integrator_max_distance_m': 2.0,
'esdf_integrator_min_weight': 0.0001,

# Occupancy parameters
'occupancy_integrator_free_region_occupancy_probability': 0.3,
'occupancy_integrator_occupied_region_occupancy_probability': 0.7,
}],
remappings=[
('depth/image', '/camera/depth/image_rect_raw'),
('depth/camera_info', '/camera/depth/camera_info'),
('color/image', '/camera/color/image_raw'),
('color/camera_info', '/camera/color/camera_info'),
]
),
])

nvblox Topicsโ€‹

# Published Topics
/nvblox_node/mesh # nvblox_msgs/Mesh (3D reconstruction)
/nvblox_node/static_occupancy # nav_msgs/OccupancyGrid (for Nav2)
/nvblox_node/esdf_pointcloud # sensor_msgs/PointCloud2 (ESDF visualization)
/nvblox_node/mesh_marker # visualization_msgs/Marker (RViz mesh)
/nvblox_node/slice_bounds # Bounding box of map slice

# Subscribed Topics
/depth/image # sensor_msgs/Image (depth)
/depth/camera_info # sensor_msgs/CameraInfo
/color/image # sensor_msgs/Image (RGB, optional)
/tf # Transform tree

Visualizing nvblox Outputโ€‹

# RViz2 configuration for nvblox
# Add these displays:
# 1. Marker - topic: /nvblox_node/mesh_marker
# 2. OccupancyGrid - topic: /nvblox_node/static_occupancy
# 3. PointCloud2 - topic: /nvblox_node/esdf_pointcloud

ros2 run rviz2 rviz2 -d $(ros2 pkg prefix nvblox_rviz_plugin)/share/nvblox_rviz_plugin/rviz/nvblox.rviz

Complete VSLAM + Mapping Pipelineโ€‹

System Architectureโ€‹

โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
โ”‚ Complete Isaac ROS Perception Pipeline โ”‚
โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค
โ”‚ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Sensors โ”‚ โ”‚
โ”‚ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ Stereo โ”‚ โ”‚ RGB-D โ”‚ โ”‚ IMU โ”‚ โ”‚ Wheel โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ Camera โ”‚ โ”‚ Camera โ”‚ โ”‚ โ”‚ โ”‚ Odom โ”‚ โ”‚ โ”‚
โ”‚ โ”‚ โ””โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”˜ โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ cuVSLAM โ”‚ โ”‚
โ”‚ โ”‚ Visual Odometry + SLAM โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ–ผ โ–ผ โ–ผ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Pose โ”‚ โ”‚ Sparse โ”‚ โ”‚ Loop โ”‚ โ”‚
โ”‚ โ”‚ Estimate โ”‚ โ”‚ Map โ”‚ โ”‚ Closures โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚
โ”‚ โ–ผ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ nvblox โ”‚ โ”‚
โ”‚ โ”‚ 3D Reconstruction + Occupancy Grid โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚
โ”‚ โ–ผ โ–ผ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ 3D Mesh โ”‚ โ”‚ Costmap โ”‚ โ”‚
โ”‚ โ”‚ โ”‚ โ”‚ (Nav2) โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚ โ”‚
โ”‚ โ–ผ โ”‚
โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚
โ”‚ โ”‚ Nav2 โ”‚ โ”‚
โ”‚ โ”‚ Planning โ”‚ โ”‚
โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚
โ”‚ โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

Complete Launch Fileโ€‹

# File: full_perception_pipeline.launch.py
from launch import LaunchDescription
from launch_ros.actions import Node
from launch.actions import IncludeLaunchDescription
from launch.launch_description_sources import PythonLaunchDescriptionSource
from launch_ros.substitutions import FindPackageShare
import os

def generate_launch_description():
# Package paths
isaac_ros_visual_slam_dir = FindPackageShare('isaac_ros_visual_slam')
nvblox_dir = FindPackageShare('nvblox_ros')

return LaunchDescription([
# cuVSLAM Node
Node(
package='isaac_ros_visual_slam',
executable='isaac_ros_visual_slam_node',
name='visual_slam',
parameters=[{
'denoise_input_images': True,
'rectified_images': True,
'enable_imu_fusion': True,
'enable_debug_mode': False,
'map_frame': 'map',
'odom_frame': 'odom',
'base_frame': 'base_link',
}],
remappings=[
('stereo_camera/left/image', '/camera/infra1/image_rect_raw'),
('stereo_camera/left/camera_info', '/camera/infra1/camera_info'),
('stereo_camera/right/image', '/camera/infra2/image_rect_raw'),
('stereo_camera/right/camera_info', '/camera/infra2/camera_info'),
('visual_slam/imu', '/camera/imu'),
]
),

# nvblox Node
Node(
package='nvblox_ros',
executable='nvblox_node',
name='nvblox',
parameters=[{
'voxel_size': 0.05,
'global_frame': 'odom',
'esdf_update_rate_hz': 10.0,
'mesh_update_rate_hz': 5.0,
}],
remappings=[
('depth/image', '/camera/depth/image_rect_raw'),
('depth/camera_info', '/camera/depth/camera_info'),
('color/image', '/camera/color/image_raw'),
('color/camera_info', '/camera/color/camera_info'),
]
),

# Static Transform: base_link to camera
Node(
package='tf2_ros',
executable='static_transform_publisher',
arguments=[
'0.1', '0', '0.2', # x, y, z
'0', '0', '0', '1', # qx, qy, qz, qw
'base_link', 'camera_link'
]
),
])

Testing in Isaac Simโ€‹

Setting Up Isaac Sim for VSLAM Testingโ€‹

# File: isaac_sim_vslam_test.py
from isaacsim import SimulationApp

simulation_app = SimulationApp({"headless": False})

from omni.isaac.core import World
from omni.isaac.core.utils.extensions import enable_extension
from omni.isaac.sensor import Camera
from omni.isaac.ros2_bridge import CameraROS2Publisher
import numpy as np

# Enable ROS 2 bridge
enable_extension("omni.isaac.ros2_bridge")

# Create world
world = World(stage_units_in_meters=1.0)
world.scene.add_default_ground_plane()

# Create stereo camera rig
left_camera = Camera(
prim_path="/World/Robot/StereoRig/LeftCamera",
position=np.array([0.0, 0.05, 0.3]), # 10cm baseline
frequency=30,
resolution=(640, 480)
)

right_camera = Camera(
prim_path="/World/Robot/StereoRig/RightCamera",
position=np.array([0.0, -0.05, 0.3]),
frequency=30,
resolution=(640, 480)
)

# Create depth camera for nvblox
depth_camera = Camera(
prim_path="/World/Robot/DepthCamera",
position=np.array([0.1, 0.0, 0.3]),
frequency=30,
resolution=(640, 480)
)

world.scene.add(left_camera)
world.scene.add(right_camera)
world.scene.add(depth_camera)

# Create ROS 2 publishers
left_pub = CameraROS2Publisher(left_camera, "camera_left", "/camera/infra1/image_rect_raw")
right_pub = CameraROS2Publisher(right_camera, "camera_right", "/camera/infra2/image_rect_raw")
depth_pub = CameraROS2Publisher(depth_camera, "camera_depth", "/camera/depth/image_rect_raw")

# Simulation loop
world.reset()
while simulation_app.is_running():
world.step(render=True)

simulation_app.close()

Creating Test Environmentsโ€‹

def create_slam_test_environment(world):
"""Create an environment with features for VSLAM testing."""
from omni.isaac.core.objects import VisualCuboid, DynamicCuboid

# Textured walls (important for visual features)
wall_configs = [
{"pos": [0, 5, 1.5], "scale": [10, 0.1, 3], "color": [0.8, 0.2, 0.2]},
{"pos": [0, -5, 1.5], "scale": [10, 0.1, 3], "color": [0.2, 0.8, 0.2]},
{"pos": [5, 0, 1.5], "scale": [0.1, 10, 3], "color": [0.2, 0.2, 0.8]},
{"pos": [-5, 0, 1.5], "scale": [0.1, 10, 3], "color": [0.8, 0.8, 0.2]},
]

for i, cfg in enumerate(wall_configs):
world.scene.add(
VisualCuboid(
prim_path=f"/World/Walls/Wall_{i}",
name=f"wall_{i}",
position=np.array(cfg["pos"]),
scale=np.array(cfg["scale"]),
color=np.array(cfg["color"])
)
)

# Add distinctive objects for loop closure
landmarks = [
{"pos": [2, 2, 0.5], "scale": [0.5, 0.5, 1], "color": [1, 0, 0]},
{"pos": [-2, 2, 0.5], "scale": [0.5, 0.5, 1], "color": [0, 1, 0]},
{"pos": [2, -2, 0.5], "scale": [0.5, 0.5, 1], "color": [0, 0, 1]},
{"pos": [-2, -2, 0.5], "scale": [0.5, 0.5, 1], "color": [1, 1, 0]},
]

for i, lm in enumerate(landmarks):
world.scene.add(
VisualCuboid(
prim_path=f"/World/Landmarks/Landmark_{i}",
name=f"landmark_{i}",
position=np.array(lm["pos"]),
scale=np.array(lm["scale"]),
color=np.array(lm["color"])
)
)

Performance Optimizationโ€‹

Tuning cuVSLAMโ€‹

# cuVSLAM performance parameters
visual_slam_params:
# Feature detection
num_features: 500 # Reduce for speed, increase for accuracy
feature_threshold: 20 # Lower = more features

# Tracking
max_frame_rate: 60 # Match camera FPS
image_jitter_threshold_ms: 34 # For 30 FPS camera

# Map management
max_landmarks: 10000 # Limit for memory
landmark_culling_threshold: 3 # Remove landmarks seen < N times

# IMU fusion (if available)
enable_imu_fusion: true
imu_integration_sigma: 0.01

Tuning nvbloxโ€‹

# nvblox performance parameters
nvblox_params:
# Resolution trade-offs
voxel_size: 0.05 # Larger = faster, less detail

# Update rates
esdf_update_rate_hz: 10.0 # Reduce for lower GPU load
mesh_update_rate_hz: 5.0 # Reduce for lower GPU load

# Integration limits
tsdf_integrator_max_integration_distance_m: 5.0 # Limit range
max_integration_images_per_sec: 30 # Throttle input

# Memory management
map_clearing_radius_m: 10.0 # Clear distant voxels
max_blocks: 100000 # Limit memory usage

Monitoring Performanceโ€‹

# Monitor GPU usage
watch -n 1 nvidia-smi

# Monitor topic rates
ros2 topic hz /visual_slam/tracking/odometry
ros2 topic hz /nvblox_node/mesh

# Profile with ROS 2
ros2 run ros2_tracing trace --session my_session

Practical Exercise: Building a SLAM Pipelineโ€‹

Goalโ€‹

Create a complete perception system that:

  1. Runs cuVSLAM for localization
  2. Builds 3D maps with nvblox
  3. Outputs costmaps for navigation
  4. Tested first in Isaac Sim

Step 1: Create the Launch Fileโ€‹

# File: perception_pipeline.launch.py
from launch import LaunchDescription
from launch_ros.actions import Node, ComposableNodeContainer
from launch_ros.descriptions import ComposableNode

def generate_launch_description():
return LaunchDescription([
# cuVSLAM
Node(
package='isaac_ros_visual_slam',
executable='isaac_ros_visual_slam_node',
name='visual_slam',
parameters=[{
'enable_imu_fusion': False,
'map_frame': 'map',
'odom_frame': 'odom',
'base_frame': 'base_link',
}]
),

# nvblox
Node(
package='nvblox_ros',
executable='nvblox_node',
name='nvblox',
parameters=[{
'voxel_size': 0.05,
'global_frame': 'odom',
}]
),

# Robot State Publisher
Node(
package='robot_state_publisher',
executable='robot_state_publisher',
parameters=[{'robot_description': '...'}]
),
])

Step 2: Test Workflowโ€‹

# Terminal 1: Launch Isaac Sim with robot and cameras
cd ~/isaac_ros_ws
python3 src/my_robot_sim/scripts/isaac_sim_vslam_test.py

# Terminal 2: Launch perception pipeline
ros2 launch my_robot_perception perception_pipeline.launch.py

# Terminal 3: Visualize in RViz2
rviz2 -d config/slam_visualization.rviz

# Terminal 4: Teleoperate the robot
ros2 run teleop_twist_keyboard teleop_twist_keyboard

Step 3: Evaluate SLAM Qualityโ€‹

# Record trajectory
ros2 bag record /visual_slam/tracking/odometry /tf -o slam_test

# Compare to ground truth (if available)
evo_ape tum ground_truth.txt estimated.txt --align --plot

# Check map consistency
ros2 service call /nvblox_node/save_map std_srvs/srv/Empty

Summaryโ€‹

In this chapter, you learned:

  • Isaac ROS Architecture: GPU-accelerated perception with NITROS zero-copy transport
  • cuVSLAM: Real-time visual odometry at 60+ FPS with loop closure
  • nvblox: GPU-accelerated 3D reconstruction with TSDF and ESDF
  • Sensor Fusion: Combining visual SLAM with IMU and wheel odometry
  • Integration: Building complete perception pipelines with ROS 2
  • Testing: Validating SLAM systems in Isaac Sim before deployment
  • Optimization: Tuning parameters for performance vs accuracy

GPU-accelerated perception fundamentally changes what's possible in real-time robotics, enabling capabilities that were previously too computationally expensive.


Further Readingโ€‹


Next Week Previewโ€‹

In Chapter 10, we explore Nav2 Path Planning:

  • ROS 2 Navigation Stack (Nav2) architecture
  • Costmap configuration with nvblox
  • Path planning algorithms (NavFn, Smac, TEB)
  • Behavior trees for complex navigation
  • Integration with Isaac ROS perception