From the course: Data Science on Google Cloud Platform: Building Data Pipelines

Unlock the full course today

Join today to access over 22,600 courses taught by industry experts or purchase this course individually.

Setting up GCP for Dataflow

Setting up GCP for Dataflow - Google Cloud Tutorial

From the course: Data Science on Google Cloud Platform: Building Data Pipelines

Start my 1-month free trial

Setting up GCP for Dataflow

- [Kumaran] Before we start using Apache Bean, we need to set up GCP and also set up Python. We already have a GCP account, and have created a project called "My First Project". This project will be used for all further examples. First, we need to enable some APIs. To do that, go to the side navigator on the GCP console, click "APIs & Services", and choose "Dashboard". We see a lot of APIs enabled for you in this dashboard. For our examples, we need some specific set of APIs enabled. Those should be the Compute Engine API, the BigQuery API, the Cloud Storage API, the Cloud Storage JSON API, Stackdriver Logging API, and the Dataflow API. We see that Dataflow API is not enabled, so we are going to go ahead and enable them. To do that, we click on the "Enable API" on the top. We then search for "Dataflow". This shows up the Dataflow API. We click on this, and then we click on the "Enable" button to enable this API. Now the API will show up as enabled in the API & Services dashboard…

Contents