Introduction to Apache Spark Developer Training

Learn what Apache Spark is and how it compares to Hadoop MapReduce, how to filter, map, reduce, and save Resilient Distributed Datasets (RDDs), who is best suited to attend the course and what prior knowledge you should have, and the benefits of building Spark applications as part of an enterprise data hub.

Date: Wednesday, Jul 23 2014


Apache Spark is a next-generation processing engine optimized for speed, ease of use, and advanced analytics well beyond batch. The Spark framework supports streaming data and complex, iterative algorithms, enabling applications to run 100x faster than traditional MapReduce programs. With Spark, developers can write sophisticated parallel applications for faster business decisions and better user outcomes, applied to a wide variety of architectures and industries.

Next Steps