From the course: Cloud Hadoop: Scaling Apache Spark

What you should know - Apache Spark Tutorial

From the course: Cloud Hadoop: Scaling Apache Spark

Start my 1-month free trial

What you should know

- [Instructor] In order to get the most out of this course, there's a few core concepts you should be familiar with. The first has to do with distributed data processing and that is the Hadoop foundations up on which Spark is built. So there's a course in the library called Hadoop Essentials or Hadoop Concepts that you might want to take a look at if you're new to this world of distributed data processing. Also, because this is focused on implementing on the public cloud, if you have experience or knowledge of working with one or more public cloud services, in particular, managing virtual machines and storage buckets, that will be useful. And last, it will be helpful if you have some knowledge of a programming language. We're going to focus on Python. If you have R or some other type of data language, you're probably going to be able to get the concepts. And I do have a couple of examples in R as well, but we are going to be focusing mostly on Python in this course.

Contents