This document discusses assessing the robustness of machine learning models to "real-world perturbations" that occur naturally rather than being crafted by an adversary. It argues that adversarial robustness is not a valid measure of robustness to such natural perturbations. The document proposes a probabilistic approach to compute the likelihood that a real-world perturbation would change a model's prediction, quantifying its real-world robustness. This approach models possible input perturbations individually for each application and works for any "black-box" model. It illustrates the approach on two datasets and analytically solvable cases, and discusses estimating real-world robustness in high-dimensional spaces.