The recent advancements in the field of explainable AI, particularly within the context of social media data analysis, have shown significant progress in addressing complex mental health issues and online toxicity. Researchers are increasingly leveraging Large Language Models (LLMs) to not only classify and predict mental health concerns and hate speech but also to provide interpretable explanations for these classifications. This shift towards explainability is crucial for enhancing the trustworthiness and practical applicability of AI models in real-world scenarios. Notably, the integration of cognitive theories into LLM reasoning processes for stress detection represents a novel approach, offering a more comprehensive understanding of psychological states. Additionally, the development of multilingual text detoxification methods and the distillation of large models for efficient hate speech detection are advancing the field by making these solutions more accessible and scalable. These innovations collectively push the boundaries of AI's role in social good, emphasizing the importance of both performance and interpretability in AI systems.
Noteworthy papers include one that integrates cognitive appraisal theory into LLM reasoning for stress detection, enhancing both performance and explainability. Another highlights the use of model distillation for efficient and explainable hate speech detection, achieving superior classification performance while maintaining interpretability.