Star 历史趋势
数据来源: GitHub API · 生成自 Stargazers.cn
README.md

GrandTour Logo


The GrandTour Dataset

A project brought to you by RSL - ETH Zurich.

ReferencesHugging FaceROS1ContributingNewsCitation

References

Please first visit the official webpage to learn more about the available data, hardware setup, and registration.

Visit our sponsors and partners, including Leica Geosystems (Hexagon).



Projects using the GrandTour Dataset

This is a growing list of known public projects/papers using GrandTour data (directly or as curated source data).

ProjectPreview
Physical Terrain Parameter Learning
Learning simulation parameters from RGB and proprioception.
Forward Dynamics Model Learning (FDM)
Learning platform-aware forward dynamics for safer navigation.
Holistic Fusion
Holistic state estimation.
RESPLE: Recursive Spline LIO
LiDAR-inertial odometry benchmarked on GrandTour-style data.
LiMo: Less Is More (Scalable Visual Navigation)
Extends GrandTour deployments with large-scale embodiment-specific trajectories.
DeFM: Learning Foundation Representations from Depth
Uses GrandTour in a large real-world depth pretraining corpus.
WildOS: Open-Vocabulary Object Search in the Wild
Trains frontier prediction using annotated images from GrandTour.
NaviTrace: Evaluating Embodied Navigation of VLMs
Uses curated image samples from the publicly available GrandTour dataset.


Hugging Face Instructions

You can find Jupyter notebooks and scripts with full instructions in the examples_hugging_face directory.

Installation Instructions

Click for installation details

These steps assume you are using uv for dependency management.

1. Install using uv

pip3 install uv uv sync cd examples_hugging_face uv sync uv run scripts/download_data.py

📒 Jupyter Notebook Examples

🐍 Python Scripts



ROS1 Instructions

Download

Click here

To access and download the GrandTour dataset ROS bags, please follow these steps:

1. Register for Access

2. Download ROS bags

Option 1 – Command Line Interface (Recommended):

Install the CLI tool and log in:

pip3 install kleinkram klein login
  • You can now explore the CLI using tab-completion or the --help flag.

Download multiple files via Python scripting:

python3 examples_kleinkram/kleinkram_cli_example.py

Directly convert ROS bags to PNG images (requires ROS1 installation):

python3 examples_kleinkram/kleinkram_extract_images.py

Option 2 – Web Interface:

Installation

Click here

Create Folders

mkdir -p ~/grand_tour_ws/src mkdir -p ~/git

Clone and Link Submodules

⚠️ Note: The repository structure is under active development (ROS2/MCAP support is being added). Commands below may evolve.

# Clone the repository cd ~/git git clone git@github.com:leggedrobotics/grand_tour_dataset.git cd grand_tour_dataset; git submodule update --init # Checkout only the required packages from the grand_tour_box repository for simplicity cd ~/git/grand_tour_dataset/examples_ros1/submodules/grand_tour_box git sparse-checkout init --cone git sparse-checkout set box_model box_calibration box_drivers/anymal_msgs box_drivers/gnss_msgs # Link the repository to the workspace ln -s ~/git/grand_tour_dataset/examples_ros1 ~/grand_tour_ws/src/

Setup and Build Catkin Workspace

cd ~/grand_tour_ws catkin init catkin config --extend /opt/ros/noetic catkin config --cmake-args -DCMAKE_BUILD_TYPE=RelWithDebInfo catkin build grand_tour_ros1 source devel/setup.bash

Download Example Mission

mkdir -p ~/grand_tour_ws/src/examples_ros1/data cd ~/grand_tour_ws/src/examples_ros1/data pip3 install kleinkram klein login klein download --mission 3c97a27e-4180-4e40-b8af-59714de54a87

Examples

Terminal 1: Launch LiDARs

roslaunch grand_tour_ros1 lidars.launch # URDFs are automatically loaded by: # Boxi: box_model box_model.launch # ANYmal: anymal_d_simple_description load.launch

Terminal 2: Replay Bags

cd ~/grand_tour_ws/src/examples_ros1/data # We provide an easy interface to replay the bags rosrun grand_tour_ros1 rosbag_play.sh --help rosrun grand_tour_ros1 rosbag_play.sh --lidars --tf_model # We provide two tf_bags # tf_model contains frames required for the URDF model of ANYmal and Boxi. # tf_minimal contains only core sensor frames.

You can also try the same for cameras.launch.

Example Output:

LiDAR VisualizationCamera Visualization
LiDAR Visualization
Visualization of LiDAR data using lidars.launch.
Camera Visualization
Visualization of images using cameras.launch.

Image Uncompression and Rectification

We provide a launch file to uncompress images and publish rectified images. Install the required dependencies:

sudo apt-get install ros-noetic-image-transport sudo apt-get install ros-noetic-compressed-image-transport
roslaunch grand_tour_ros1 cameras_helpers.launch

IMUs Visualization

We use rqt-multiplot to visualize the IMU measurements.

Install rqt_multiplot:

sudo apt-get install ros-noetic-rqt-multiplot -y

Start rqt_multiplot and replay the bags:

roslaunch grand_tour_ros1 imus.launch
cd ~/grand_tour_ws/src/examples_ros1/data rosrun grand_tour_ros1 rosbag_play.sh --imus --ap20

Example Output:
assets/rqt-multiplot.png



Contributing

We welcome contributions to help improve and expand this project. Whether you're interested in adding new examples, enhancing existing ones, or offering suggestions, feel free to open an issue or reach out directly.

We are particularly looking for contributions in the following areas:

  • New and interesting benchmarks
  • ROS2 integration and conversion (MCAP / ROS2 tooling)
  • Visualization tools (e.g., Viser, etc.)
  • Hosting and deployment support in Asia


News

  • 2026-02: The GrandTour dataset paper is now available as an arXiv preprint: arXiv:2602.18164 (submitted to IJRR).
  • The Localization Challenge page is online (currently marked postponed) and will be hosted via EvalAI once opened.
  • MCAP (ROS2) support is listed as coming soon on the official dataset page.
  • We continue to expand GrandTour with new deployments, examples, and downstream benchmarks.


Citation

If you use the GrandTour dataset, please cite the dataset paper:

@article{frey2026grandtour, title={GrandTour: A Legged Robotics Dataset in the Wild for Multi-Modal Perception and State Estimation}, author={Frey, Jonas and Tuna, Turcan and Fu, Frank and Patterson, Katharine and Xu, Tianao and Fallon, Maurice and Cadena, Cesar and Hutter, Marco}, journal={arXiv preprint arXiv:2602.18164}, year={2026} }

*Equal contribution: Jonas Frey and Turcan Tuna.

If you use the Boxi payload/software or cite hardware/payload design decisions, please also cite:

@INPROCEEDINGS{Tuna-Frey-Fu-RSS-25, AUTHOR = {Jonas Frey AND Turcan Tuna AND Lanke Frank Tarimo Fu AND Cedric Weibel AND Katharine Patterson AND Benjamin Krummenacher AND Matthias Müller AND Julian Nubert AND Maurice Fallon AND Cesar Cadena AND Marco Hutter}, TITLE = {{Boxi: Design Decisions in the Context of Algorithmic Performance for Robotics}}, BOOKTITLE = {Proceedings of Robotics: Science and Systems}, YEAR = {2025}, ADDRESS = {Los Angeles, United States}, MONTH = {June} }

*Shared first authorship: Frey, Tuna, Fu.

关于 About

The GrandTour Dataset: A Legged Robotics Dataset in the Wild

语言 Languages

Jupyter Notebook95.1%
Python4.8%
Shell0.1%
CMake0.0%

提交活跃度 Commit Activity

代码提交热力图
过去 52 周的开发活跃度
62
Total Commits
峰值: 17次/周
Less
More

核心贡献者 Contributors