This document proposes a technique for analyzing human behavior in videos using motion-based feature extraction. It discusses how previous approaches have used spatial and temporal features to detect abnormal behaviors. The proposed approach extracts motion features from videos to represent each video with a single feature vector, rather than extracting features from each individual frame. This reduces the feature space and unnecessary information. The technique involves preprocessing videos into frames, extracting motion features, using KNN classification on the features to classify behaviors as normal or abnormal, and evaluating the method's performance on various metrics like accuracy, recall, and precision. Testing on fight and riot datasets showed the motion-based approach achieved higher accuracy, recall, precision and F-measure than a non-motion based approach.