This document outlines a project to develop a speech-enhanced gesture-based navigation system for Google Maps using Microsoft Kinect. The goals are to define gestures and voice commands to control Google Maps features on a large touchscreen, and evaluate the usability of this multimodal approach against traditional touch-based gestures. The proposed methodology involves researching similar systems, designing gestures and use cases, developing a prototype using Kinect and speech recognition, and conducting a comparative user study in the next semester to analyze the results. Challenges in speech recognition for Indian accents are also discussed.