This document summarizes a presentation about serving machine learning models with Apache Flink. It discusses Flink's ML roadmap, modeling concepts, and model serving considerations like customer needs, governance, and the model lifecycle. It proposes using Flink's streaming capabilities to dynamically update models without interruption. A prototype exploits Flink joins to merge prediction requests with cached models. Next steps include finalizing specifications, integrating TensorFlow and PMML models, and addressing governance.