Skip to content

WHU-Synthetic, a large-scale 3D synthetic perception dataset designed for multi-task learning, from the initial data augmentation (upsampling and depth completion), through scene understanding (segmentation), to macro-level tasks (place recognition and 3D reconstruction).

Notifications You must be signed in to change notification settings

WHU-USI3DV/WHU-Synthetic

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

28 Commits
 
 
 
 

Repository files navigation

WHU-Synthetic

We introduce WHU-Synthetic, a large-scale 3D synthetic perception dataset designed for multi-task learning, from the initial data augmentation, through scene understanding , to macro-level tasks. Collected in the same environmental domain, we ensure inherent alignment across sub-tasks to construct multi-task models without separate training methods. Besides, we implement several novel settings, making it possible to realize certain ideas that are difficult to achieve in real-world scenarios. This supports more adaptive and robust multi-task perception tasks, such as sampling on city-level models, providing point clouds with different densities, and simulating temporal changes.

Intro

🆕 News

  • 2025-02-08: Accepted by TGRS! 🎉🎉🎉

🚧 Download

Our dataset is currently under active preparation. Updates coming soon (It is expected to be released in mid-March) – watch this space!👀

🔢 Dataset

1. Overview

WHU-Synthetic dataset is a large-scale synthetic 3D perception dataset designed for multi-task learning in 3D vision. It integrates multiple sub-tasks, including depth completion, point cloud upsampling, semantic segmentation, place recognition, and 3D reconstruction, all within the same environmental domain. Generated using the CARLA simulator, the dataset features over 140,000 fully annotated frames across diverse urban and rural scenes, with data collected from various sensors such as LiDAR, RGB cameras, and depth cameras. It also includes unique settings like multi-density point clouds, city-level surface sampling, and simulated environmental changes, which support advanced research in domain adaptation, robust perception, and the exploration of inter-task relationships.

2. Collection Settings and Format

In the WHU-Synthetic dataset, a single frame contains data that can be used to train multiple tasks simultaneously. To make it easier to work with, we outline the specific data items that may be relevant to individual sub-tasks in the table. However, it's important to emphasize that all of this data is captured at the same moment, which is key to our multi-task design. Additionally, for the sub-tasks of 3D reconstruction and place recognition, which require fixed routes, we have designed a series of specific routes and collected data separately.

Intro

We meticulously configure our setup by the parameters of KITTI sensors and added noise and post-effects to ensure the realism of the data, aiming to minimize domain shift. We successfully stack LiDARs with different channel counts at the same location to simulate point clouds of different densities at the sensor level, and the data are automatically annotated. The point clouds generated by LiDAR sensors share the same coordinate system and do not produce mutual occlusions.

Intro

3. Unique Features

  • Multiple Density Data: WHU-Synthetic captures point clouds with varying densities (16/32/64/128 channels) in a single frame, supporting tasks like upsampling, depth completion, and segmentation, and enabling domain adaptation studies.
  • City-Level Surface Sampling: The dataset uses Monte Carlo Sampling to generate unobstructed point clouds from a city mesh model, divided into 120m × 120m blocks. This supports scene-level completion and city-scale 3D reconstruction.
  • Simulated Environmental Changes: WHU-Synthetic simulates short-term changes in environments to study place recognition under dynamic conditions, providing insights into algorithm adaptability.

4. Multi-task Network

In the multi-task exploration experiment, we designed a framework with a shared backbone network to jointly perform point cloud upsampling and semantic segmentation tasks. The results showed that simultaneous training of both tasks improved performance: the average recall (AR) for upsampling increased from 43.2% to 46.5%, and the mean Intersection over Union (mIoU) for segmentation improved from 43.6% to 45.3%. This demonstrates a mutually reinforcing effect between the tasks, where upsampling provides geometric constraints for segmentation, and segmentation offers category-related information for upsampling. The experiment also highlighted the dataset's ability to support 3D multi-task training and its potential for real-world applications.

Intro

Intro

5. Domain Shift

In the domain shift experiments, pre-trained models from the WHU-Synthetic dataset were tested on real-world datasets across three tasks: depth completion, point cloud upsampling, and semantic segmentation. For depth completion, the E-Net model trained on WHU-Synthetic indicates a moderate domain shift but reasonable generalization. In upsampling, ILN trained on WHU-Synthetic enhanced KITTI data with some noise, showing small domain shift. For semantic segmentation, the Minkowski U-Net achieved a mean IoU of 33.8% on SemanticKITTI, with significant variations across classes.

Intro

📚 Citation

@article{zhou2025whu,
  title={WHU-Synthetic: A Synthetic Perception Dataset for 3D Multi-task Model Research},
  author={Zhou, Jiahao and Long, Chen and Xie, Yue and Wang, Jialiang and Zhang, Conglang and Li, Boheng and Wang, Haiping and Chen, Zhe and Dong, Zhen},
  journal={IEEE Transactions on Geoscience and Remote Sensing},
  year={2025},
  publisher={IEEE}
}

🎗️ Related Work

  1. CARLA Simulator
  2. SHIFT: A Synthetic Driving Dataset for Continuous Multi-Task Domain Adaptation
  3. KITTI-CARLA: a KITTI-like dataset generated by CARLA Simulator
  4. PU-Dense: Sparse Tensor-based Point Cloud Geometry Upsampling
  5. PatchAugNet: Patch feature augmentation-based heterogeneous point cloud place recognition in large-scale street scenes

About

WHU-Synthetic, a large-scale 3D synthetic perception dataset designed for multi-task learning, from the initial data augmentation (upsampling and depth completion), through scene understanding (segmentation), to macro-level tasks (place recognition and 3D reconstruction).

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published