The document discusses the development of multi-modal linguistic corpora that analyze language beyond just text. It focuses on the HeadTalk project, which aims to better understand head nods through the creation of resources that allow simultaneous exploration of verbal and non-verbal communication. The project involves detecting and extracting features of gestures, clustering gestures based on these features, and relating the clusters to linguistic categories through both automated and expert analysis. The goal is to build tools that help social scientists create and annotate corpora incorporating both language and visual data like gestures.