Continual Learning and Multi-Task Learning

Current Developments in Continual Learning and Multi-Task Learning

The recent advancements in the fields of Continual Learning (CL) and Multi-Task Learning (MTL) have been particularly innovative, addressing critical challenges such as catastrophic forgetting, data sparsity, and the efficient transfer of knowledge across tasks. These developments are pushing the boundaries of how models can adapt to new information over time while maintaining performance on previously learned tasks.

General Direction of the Field

  1. Mitigating Catastrophic Forgetting: A significant focus has been on developing techniques to mitigate catastrophic forgetting, where models tend to forget previously learned information when trained on new tasks. Novel approaches like CURLoRA and Dual-CBA introduce sophisticated methods for continual fine-tuning and bias adaptation, respectively, to maintain model stability and performance across tasks.

  2. Efficient Knowledge Transfer: The field is also moving towards more efficient methods for transferring knowledge between tasks. Papers like the Efficient Transfer Learning Framework for Cross-Domain Click-Through Rate Prediction propose tri-level asynchronous frameworks that effectively manage extensive source data and prevent catastrophic forgetting, thereby enhancing the performance of target models.

  3. Uncertainty and Noise Handling: Addressing the challenges posed by noisy labels and long-tailed distributions, recent works such as Alternate Experience Replay and Prior-free Balanced Replay introduce strategies to maintain a clear distinction between clean and noisy samples, improving model accuracy and robustness.

  4. Optimization and Generalization: There is a growing interest in understanding the optimization trajectories of models during multi-task learning. Studies like "Can Optimization Trajectories Explain Multi-Task Transfer?" delve into the impact of multi-task learning on generalization, shedding light on the underlying causes of optimization failures and raising questions about the effectiveness of general-purpose multi-task optimization algorithms.

  5. Practical Deployment and Real-World Scenarios: The emphasis on practical deployment and real-world scenarios is evident in works like "A Practitioner's Guide to Continual Multimodal Pretraining," which provides comprehensive guidance for effective continual model updates under realistic compute constraints and deployment requirements.

Noteworthy Innovations

  • CURLoRA: Introduces a novel approach to fine-tuning large language models, significantly reducing trainable parameters and mitigating catastrophic forgetting.
  • Efficient Transfer Learning Framework for Cross-Domain Click-Through Rate Prediction: Proposes a tri-level asynchronous framework that efficiently transfers knowledge from natural content to advertisement CTR models, addressing data sparsity and catastrophic forgetting.
  • Alternate Experience Replay: Effectively handles noisy labels in continual learning, achieving remarkable accuracy gains through asymmetric balanced sampling.
  • Prior-free Balanced Replay: Addresses long-tailed continual learning without prior label distribution, using uncertainty-guided reservoir sampling to prioritize minority data.

These advancements collectively represent a significant stride in the continual and multi-task learning domains, offering practical solutions and theoretical insights that will likely influence future research and applications.

Sources

CURLoRA: Stable LLM Continual Fine-Tuning and Catastrophic Forgetting Mitigation

A Practitioner's Guide to Continual Multimodal Pretraining

May the Forgetting Be with You: Alternate Replay for Learning with Noisy Labels

Dual-CBA: Improving Online Continual Learning via Dual Continual Bias Adaptors from a Bi-level Optimization Perspective

Can Optimization Trajectories Explain Multi-Task Transfer?

Graph and Sequential Neural Networks in Session-based Recommendation: A Survey

Prior-free Balanced Replay: Uncertainty-guided Reservoir Sampling for Long-Tailed Continual Learning

Modeling and Analyzing the Influence of Non-Item Pages on Sequential Next-Item Prediction

Efficient Transfer Learning Framework for Cross-Domain Click-Through Rate Prediction

Transformers Meet ACT-R: Repeat-Aware and Sequential Listening Session Recommendation

Adaptive Variational Continual Learning via Task-Heuristic Modelling

Theoretical Insights into Overparameterized Models in Multi-Task and Replay-Based Continual Learning

Efficient Multi-task Prompt Tuning for Recommendation

Continual learning with the neural tangent ensemble

CoRA: Optimizing Low-Rank Adaptation with Common Subspace of Large Language Models

Built with on top of