Deep Learning: Enhancing Few-Shot and Multitask Learning, Mitigating Spurious Correlations

Report on Current Developments in the Research Area

General Direction of the Field

The current research landscape in the field is characterized by a strong emphasis on enhancing the robustness and generalization capabilities of deep learning models, particularly in scenarios involving few-shot learning, multitask learning, and the presence of spurious correlations. Researchers are increasingly focusing on developing practical metrics and benchmarks to evaluate these capabilities, as well as novel training methodologies to mitigate the adverse effects of spurious correlations and conflicting task gradients.

One of the key trends is the shift towards more practical and quantifiable metrics for assessing generalization error in deep networks. This is driven by the need for both theoretical validation and practical applicability. The introduction of new benchmarking frameworks and metrics aims to bridge the gap between theoretical estimations and real-world performance, highlighting the limitations of existing theoretical models and inspiring new avenues of exploration.

Another significant development is the exploration of multitask learning, where the focus is on resolving conflicts between task gradients to improve overall model performance. Innovations in this area include adaptive task weighting schemes that prioritize certain tasks during training, leading to faster convergence and better performance metrics.

The issue of spurious correlations in training data is also receiving considerable attention. Researchers are proposing novel methods to unlearn these correlations, thereby enhancing the robustness of models. This is particularly important in few-shot learning scenarios, where models are prone to relying on spurious attributes for classification. New benchmarking frameworks are being developed to systematically evaluate the robustness of these models against spurious bias, providing a foundation for designing more robust classifiers.

Noteworthy Papers

  • Practical generalization metric for deep networks benchmarking: Introduces a novel metric that quantifies both accuracy and data diversity, revealing shortcomings in existing theoretical estimations.

  • Task Weighting through Gradient Projection for Multitask Learning: Proposes an adaptive task weighting scheme that significantly improves performance metrics in multitask learning scenarios.

  • UnLearning from Experience to Avoid Spurious Correlations: Presents a method that uses parallel models to unlearn spurious correlations, enhancing model robustness.

  • Benchmarking Spurious Bias in Few-Shot Image Classifiers: Develops a systematic benchmark framework to evaluate the robustness of few-shot classifiers against spurious bias, inspiring new designs for robust classifiers.

  • Improving Robustness to Multiple Spurious Correlations by Multi-Objective Optimization: Introduces a novel training method that dynamically adjusts group-wise losses to mitigate multiple biases, achieving superior performance on multiple datasets.

Sources

A practical generalization metric for deep networks benchmarking

Task Weighting through Gradient Projection for Multitask Learning

UnLearning from Experience to Avoid Spurious Correlations

Benchmarking Spurious Bias in Few-Shot Image Classifiers

Oops, I Sampled it Again: Reinterpreting Confidence Intervals in Few-Shot Learning

Improving Robustness to Multiple Spurious Correlations by Multi-Objective Optimization

Reassessing the Validity of Spurious Correlations Benchmarks