The document presents research on the use of pre-trained self-supervised learning (SSL) models in natural language understanding (NLU) within limited data contexts. It explores various research questions regarding the fusion of multimodal features from SSL models, domain adaptation techniques, and the impact of these models on tasks like emotion recognition and summarization. Findings indicate that pre-trained SSL models exhibit strong performance and adaptability, even with minimal high-quality data, which can significantly enhance tasks in AI.