Sysco LABS Tutorials | Getting Started with Spark

by Sysco LABS AI 15 October 2019

In this episode of the Sysco LABS Tutorials, Software Engineer – Deshani Geethika from the EAG team introduces you to Apache Spark – an opensource, distributed processing system usually used for big data worksheets. She also introduces you to the Spark Eco-system before teaching you how to set up a Spark job using Spark sequel.

The Tutorial covers:

  • What is Apache Spark?
  • Hadoop MapReduce Limitations
  • Spark over MapReduce
  • Apache Spark Ecosystem
  • Develop a Spark job using Spark SQL

Leave a Comment