- Lab
- A Cloud Guru
Run an Apache Spark Job with Dataproc
Apache Spark is a powerful open-source unified analytics engine for large-scale data processing. With its ability to handle big data and its integrated APIs for languages such as Python (PySpark), it is often used by data scientists for predictive analytics. One such application is the Monte Carlo simulation, a computational algorithm that relies on repeated random sampling to obtain numerical results. In this hands-on lab, you will create a managed Apache Spark cluster and leverage Monte Carlo simulations to predict the growth of an investment portfolio.
Path Info
Table of Contents
-
Challenge
Create a Managed Dataproc Cluster with Apache Spark Pre-Installed
- Enable the Dataproc API and navigate to the Dataproc console.
- Create a standard Dataproc cluster on Compute Engine.
- Configure the master and worker nodes with
n2-standard-2
machine types.
-
Challenge
Using Python, Run a Monte Carlo Simulation That Estimates the Growth of a Stock Portfolio
- Connect to the Dataproc cluster's primary node via SSH.
- Start a PySpark Python interpreter.
- Run the Monte Carlo simulations.
What's a lab?
Hands-on Labs are real environments created by industry experts to help you learn. These environments help you gain knowledge and experience, practice without compromising your system, test without risk, destroy without fear, and let you learn from your mistakes. Hands-on Labs: practice your skills before delivering in the real world.
Provided environment for hands-on practice
We will provide the credentials and environment necessary for you to practice right within your browser.
Guided walkthrough
Follow along with the author’s guided walkthrough and build something new in your provided environment!
Did you know?
On average, you retain 75% more of your learning if you get time for practice.