课程: Cloud Hadoop: Scaling Apache Spark
今天就学习课程吧!
今天就开通帐号,24,700 门业界名师课程任您挑!
Serverless Spark with Dataproc Notebook - Apache Spark教程
课程: Cloud Hadoop: Scaling Apache Spark
Serverless Spark with Dataproc Notebook
- [Instructor] In the Spark ecosystem, there are a number of execution environments. As we've seen in other movies in this course, we can use GCP Dataproc for a managed Spark environment. A relatively new capability is one that many of my customers have found super useful, and I wanted to share a preview of it for you here. It's called Dataproc Jupyter Lab Plugin for serverless batch and interactive notebook sessions. That's a lot of words. What does that mean? It means being able to, from a Jupyter Notebook within GCP, scale out a workload when you need to have more than one computer involved in the analysis. What I've done to give you an intro of this is I've shortened this rather long tutorial so that you can see what it looks like and hopefully be compelled to try this tutorial in full yourself. So the first step is to set up a Vertex AI VM workbench instance in a Google Cloud demonstration project. Once that's set up, then you're going to access Jupyter Lab by clicking the link…
随堂练习,边学边练
下载课堂讲义。学练结合,紧跟进度,轻松巩固知识。
内容
-
-
-
-
-
-
-
-
-
(已锁定)
Scale Spark on the cloud by example5 分钟 11 秒
-
(已锁定)
Build a quick start with Databricks AWS6 分钟 50 秒
-
(已锁定)
Scale Spark cloud compute with VMs6 分钟 16 秒
-
(已锁定)
Optimize cloud Spark virtual machines6 分钟 5 秒
-
(已锁定)
Use AWS EKS containers and data lake7 分钟 8 秒
-
(已锁定)
Optimize Spark cloud data tiers on Kubernetes4 分钟 17 秒
-
(已锁定)
Build reproducible cloud infrastructure8 分钟 37 秒
-
(已锁定)
Scale on GCP Dataproc or on Terra.bio8 分钟 34 秒
-
(已锁定)
Serverless Spark with Dataproc Notebook5 分钟 25 秒
-
(已锁定)
-