This document proposes using linear function approximation as a computationally efficient method for solving reinforcement learning problems compared to neural network approaches like TRPO and PPO. It summarizes TRPO and PPO, which use neural networks to approximate value functions. The paper then presents a natural actor-critic algorithm that uses linear function approximation instead of neural networks for value approximation. The author evaluates this approach on cart pole and acrobot benchmarks and finds it trains faster than neural network methods while achieving equivalent or better results, especially on sparse reward problems. This allows the paper to recommend using natural policy gradient methods with linear function approximation over TRPO and PPO for traditional and sparse reward low-dimensional reinforcement learning challenges.