Your computer can recognize your voice and detect words in a speech dictation, but can it truly understand the meaning of what you are saying? Can it analyze your intent and respond accordingly? You don’t need a PhD in artificial intelligence to integrate speech and natural language understanding in your projects. Microsoft Cognitive Services (aka “Project Oxford”) is a portfolio of cloud-based REST APIs and SDKs powered by Machine Learning which enable developers to write applications which understand the content within the rapidly growing set of multimedia data. Cognitive Services API services will help you understand and interact with audio, text, image, and video. In this session, we’ll start with an overview of available services for speech recognition and speech synthesis. Then we’ll explore through live demos how to leverage the Language Understanding Intelligent Service which lets you determine intent, detect entities in user speech and improve language understanding models to more efficiently work with user data. Lastly, we’ll leverage Computer Vision APIs to detect human faces, analyze the content of images, and perform Optical Character Recognition (OCR) to detect and analyze words within a photo. Come learn how your apps can tap into the same active learning services behind the brain of Cortana, and get started writing smart applications that can understand what your users are saying.