This document summarizes lecture notes on reinforcement learning and control from CS229. It introduces reinforcement learning as a framework where algorithms are provided only a reward function rather than explicit labels. It then defines Markov decision processes (MDPs) using states, actions, transition probabilities, rewards, and policies. It describes two algorithms - value iteration and policy iteration - for solving MDPs if the model is known, and discusses learning a model from data when transition probabilities and rewards are unknown. Finally, it notes MDPs can also have continuous rather than finite states.