Advances in 3D Occupancy Prediction and LiDAR Processing

The field of autonomous driving is witnessing significant advancements in 3D occupancy prediction and LiDAR processing. Researchers are focusing on developing efficient and effective methods for predicting 3D environments, reconstructing scenes, and understanding spatial and temporal information. A key direction is the integration of spatiotemporal information from past observations to improve prediction accuracy. Furthermore, there is a growing interest in semi-automatic and zero-shot learning approaches for annotating LiDAR point clouds, which can significantly boost annotation efficiency and enable the expansion of current datasets. Noteworthy papers include:

  • Mitigating Trade-off: Stream and Query-guided Aggregation for Efficient and Effective 3D Occupancy Prediction, which proposes a novel framework that achieves state-of-the-art performance in real-time settings while reducing memory usage.
  • SALT: A Flexible Semi-Automatic Labeling Tool for General LiDAR Point Clouds, which operates directly on raw LiDAR data and surpasses the latest zero-shot methods by 18.4% PQ on SemanticKITTI.
  • Zero-Shot 4D Lidar Panoptic Segmentation, which utilizes multi-modal robotic sensor setups and off-the-shelf Vision-Language foundation models to advance research in spatio-temporal scene understanding in LiDAR.

Sources

Mitigating Trade-off: Stream and Query-guided Aggregation for Efficient and Effective 3D Occupancy Prediction

SALT: A Flexible Semi-Automatic Labeling Tool for General LiDAR Point Clouds with Cross-Scene Adaptability and 4D Consistency

UniOcc: A Unified Benchmark for Occupancy Forecasting and Prediction in Autonomous Driving

Zero-Shot 4D Lidar Panoptic Segmentation

MinkOcc: Towards real-time label-efficient semantic occupancy prediction

Graph Attention-Driven Bayesian Deep Unrolling for Dual-Peak Single-Photon Lidar Imaging

Built with on top of