Parameter space noise is a simple method for exploration in reinforcement learning where noise is added to the policy parameters at the start of each episode. It balances exploration and exploitation better than epsilon-greedy or bootstrapped DQN in environments requiring directed exploration like chain environments. It also outperforms action space noise in continuous control tasks with DDPG and is better than alternatives in sparse reward environments. The method is applicable to both on and off-policy algorithms and provides an orthogonal exploration technique to other advances in deep reinforcement learning.