Remove 2009 Remove Datasets Remove Lambda Architecture Remove Scala
article thumbnail

Apache Spark Use Cases & Applications

Knowledge Hut

Apache Spark was developed by a team at UC Berkeley in 2009. Spark is developed in Scala programming language. It achieves this using abstraction layer called RDD (Resilient Distributed Datasets) in combination with DAG, which is built to handle failures of tasks or even node failures.

Scala 52