Apache Spark is a distributed computing framework used for processing and analyzing large datasets.
It can be applied to data that is structured (like database tables) or unstructured (like log files).
Data can be processed one-off or on a schedule, or continuously as a stream of items from a queue.
Definitions for Apache Spark.