This document summarizes a student research project that aims to develop a sign language recognition system using the Mediapipe framework. The system takes video input of signed letters from the American Sign Language alphabet and outputs the recognized letters in text format. The document provides background on sign language and gesture recognition, describes the Mediapipe framework and implementation methodology using KNN classification, and presents preliminary results of the system detecting hand positions and recognizing letters in real-time. The overall goal is to reduce communication barriers for deaf individuals by translating sign language to written text.