From the course: Cloud Hadoop: Scaling Apache Spark
Unlock the full course today
Join today to access over 22,600 courses taught by industry experts or purchase this course individually.
Sign up for Databricks Community Edition - Apache Spark Tutorial
From the course: Cloud Hadoop: Scaling Apache Spark
Sign up for Databricks Community Edition
- [Instructor] So, now that we've completed setting up our desktop-based IDE in Python, we are going to set up the meat of our modern Hadoop development environment by going to Databricks, and then we'll be clicking on "start your free trial." Now, notice there's two different versions, as I mentioned the previous movie. For this course we're going to be working with the "community edition," which is free. You'll have access to mini six-gigabyte cluster, which is hosted on Amazon. All of that is transparent to you. You'll be able to work, importantly, with interactive notebooks and dashboards that are Jupyter-style, and we're going to use the Python language. And you'll have a public environment that you can share your work if you're collaborating or learning with someone else. Now, just to be complete, if you decide to use Databricks in a commercial situation, then you would have to pay for it. It's a commercial…
Practice while you learn with exercise files
Download the files the instructor uses to teach the course. Follow along and learn by watching, listening and practicing.
Contents
-
-
-
-
(Locked)
Sign up for Databricks Community Edition3m 29s
-
Add Hadoop libraries2m 33s
-
(Locked)
Databricks AWS Community Edition2m 22s
-
(Locked)
Load data into tables1m 51s
-
Hadoop and Spark cluster on AWS EMR7m 30s
-
(Locked)
Run Spark job on AWS EMR4m 40s
-
(Locked)
Review batch architecture for ETL on AWS2m 17s
-
(Locked)
-
-
-
-
-
-