The field of interpretability and control in foundation models is witnessing significant advancements, particularly in the application of sparse autoencoders (SAEs). Researchers are focusing on enhancing the ability of SAEs to capture rare and domain-specific concepts, which are often overlooked by general-purpose models. This is being achieved through specialized training techniques and novel loss functions that improve the recall of these 'dark matter' features. Additionally, there is a growing emphasis on making dense retrieval models more interpretable and controllable by leveraging sparse latent features. This not only enhances the transparency of these models but also allows for more nuanced control over retrieval behaviors. Furthermore, the integration of adaptive sparsity allocation methods is being explored to better match the complexity of different data points, thereby improving the overall fidelity and utility of the extracted features. These developments collectively aim to provide deeper insights into the inner workings of foundation models and enable more precise interventions, ultimately advancing the field towards more interpretable and controllable AI systems.
Noteworthy papers include one that introduces Specialized Sparse Autoencoders (SSAEs) for capturing rare concepts in foundation models, demonstrating a 12.5% increase in classification accuracy in a bias mitigation case study. Another highlights the use of sparse latent features to interpret and control dense retrieval models, showing nearly identical retrieval accuracy with enhanced interpretability.