Efficient and Personalized On-Device Language Models

The current research landscape in on-device language models is witnessing significant advancements aimed at enhancing efficiency, capability, and personalization. Researchers are focusing on developing small language models (SLMs) that can operate effectively within the constraints of mobile devices, addressing issues such as memory usage, latency, and data imbalance. A notable trend is the integration of large language models (LLMs) into mobile platforms, enabling more sophisticated and personalized user interactions. This includes the ability for LLMs to perform on-device function calls, which enhances their utility by allowing them to interact with external APIs. The development of these models is guided by principles of architecture optimization and data augmentation, ensuring that they not only meet performance benchmarks but also adapt to user preferences and behaviors. Notably, the field is also seeing a push towards open-source initiatives, promoting transparency and reproducibility, and making advanced language models accessible to a broader audience.

Noteworthy Papers:

  • An on-device emoji classifier leveraging GPT-based data augmentation significantly improves emoji prediction accuracy, particularly for rare emojis.
  • PhoneLM introduces a novel approach to SLM design, achieving state-of-the-art efficiency and capability trade-offs while being fully open-source.
  • Alopex framework enhances on-device LLM function call accuracy and reduces catastrophic forgetting through innovative data generation and mixing strategies.

Sources

On-Device Emoji Classifier Trained with GPT-based Data Augmentation for a Mobile Keyboard

PhoneLM:an Efficient and Capable Small Language Model Family through Principled Pre-training

Alopex: A Computational Framework for Enabling On-Device Function Calls with LLMs

Fox-1 Technical Report

Built with on top of