The document describes a project to develop a real-time sign language detection system using computer vision and deep learning techniques. The researchers collected over 500 images of 5 different signs and trained a convolutional neural network model using transfer learning with a pre-trained SSD MobileNet V2 model. The model takes input from a webcam video stream and classifies each frame in real-time to detect the sign language. Some key applications of this system include improving communication for deaf individuals and teaching sign language. The researchers achieved reliable detection results under controlled lighting conditions and aim to expand the dataset and model capabilities in future work.