The field of computer vision is moving towards developing more robust and generalizable models, particularly in areas such as autonomous driving, underwater image enhancement, and object detection. Researchers are exploring new approaches to improve the performance of models under various types of distribution shifts, including spatial domain shifts, lighting condition changes, and sensor failures. Noteworthy papers include the introduction of the MVTec AD 2 dataset, which provides a comprehensive evaluation of state-of-the-art anomaly detection methods under challenging industrial inspection scenarios. The Exponentially Weighted Instance-Aware Repeat Factor Sampling method is also a significant contribution, as it improves the detection performance of rare categories in long-tailed object detection tasks. Additionally, the Benchmarking Multi-modal Semantic Segmentation under Sensor Failures and Benchmarking Object Detectors under Real-World Distribution Shifts papers provide valuable insights into the robustness of multi-modal semantic segmentation and object detection models, highlighting the need for more standardized benchmarks and evaluation metrics.
Advances in Robustness and Generalization for Computer Vision
Sources
Cross-Domain Underwater Image Enhancement Guided by No-Reference Image Quality Assessment: A Transfer Learning Approach
Benchmarking Multi-modal Semantic Segmentation under Sensor Failures: Missing and Noisy Modality Robustness