From the course: Azure Spark Databricks Essential Training
Unlock the full course today
Join today to access over 22,600 courses taught by industry experts or purchase this course individually.
Use a Spark Streaming notebook
From the course: Azure Spark Databricks Essential Training
Use a Spark Streaming notebook
- [Instructor] In this example notebook, we're going to take a look at working with data in two types of input methods. The first method is batched, or regular input, and in the second method, we're going to take a look at the new structured streaming API. This is available for Python DataFrames. So as we did in previous examples, what I've done is I've just imported this. And you can see, here it is. And then of course, you'll remember that to interact with a cluster, we're going to attach it. So this is a relatively new API, and it's been very popular with my customers. The idea is that you can do complex distributed data processing in your cluster on both batch and streaming data. It does require, as it says here in command one, that you attach it to a Spark 2.x or greater cluster, as we've done. So, we're going to work with some new type of data. We're not going to use the diamond dataset anymore, we're going to use the event dataset. So let's take a look at that. Now, you'll…
Practice while you learn with exercise files
Download the files the instructor uses to teach the course. Follow along and learn by watching, listening and practicing.
Contents
-
-
-
-
(Locked)
Review Databricks Azure cluster setup3m 39s
-
(Locked)
Use a Python notebook with dashboards6m 1s
-
(Locked)
Use an R notebook4m
-
(Locked)
Use a Scala notebook for visualization6m 37s
-
(Locked)
Use a notebook with scikit-learn11m 29s
-
(Locked)
Use a Spark Streaming notebook8m 53s
-
Use an external Scala library: variant-spark10m 26s
-
(Locked)
-
-
-
-