This document discusses using Spark as an execution engine for Hive queries. It begins by explaining that Hive and Spark are both commonly used in the big data space, and that Hive on Spark uses the Hive optimizer with the Spark query engine, while Spark with a Hive context uses both the Catalyst optimizer and Spark engine. The document then covers challenges in deploying Hive on Spark, such as using a custom Spark JAR without Hive dependencies. It shows how the Hive EXPLAIN command works the same on Spark, and how the execution plan and stages differ between MapReduce and Spark. Overall, the document provides a high-level overview of using Spark as a query engine for Hive.