1. The document discusses challenges with standard reinforcement learning formulations due to large state and action spaces. It proposes representing actions as operators that induce state transitions rather than discrete choices.
2. It introduces a generalized reinforcement learning framework using kernel methods to compare "decision contexts" or state-action pairs. Value functions are represented as vectors in a Reproducing Kernel Hilbert Space rather than concrete mappings.
3. Gaussian process regression is used to predict values for unseen state-action pairs by comparing them to stored samples, enabling generalization beyond explored contexts. Hyperparameters are tuned to best explain sample data using marginal likelihood optimization.