课程: Generative AI: Working with Large Language Models
今天就学习课程吧!
今天就开通帐号,24,700 门业界名师课程任您挑!
PaLM
- [Instructor] In April, 2022, Google released PaLM, or to give it its full name, the Pathways Language Model. Now there are a couple of key takeaways from this model. Comparing the number of parameters, we can see that PaLM is the largest of the dense parameter models with 540 billion parameters. It dwarfs the GPT-3's 175 billion parameters, Gophers, 280 billion, and just edges out Megatron-Turing NLG at 530 billion parameters. Now, Google used the pathway system, a new AI architecture that they revealed at the end of 2021. So using this framework allows for many more chips to be used for model training, with PaLM being trained on 6,144 hardware accelerators versus smaller numbers of chips being used for previous large language models. And finally, if we look at the Model Flops Utilization, you can see that the Model Flops Utilizations have increased going from GPT-3 to PaLM. PaLM has effectively doubled the Model Flops…
随堂练习,边学边练
下载课堂讲义。学练结合,紧跟进度,轻松巩固知识。
内容
-
-
-
-
-
GPT-34 分钟 32 秒
-
(已锁定)
GPT-3 use cases5 分钟 27 秒
-
(已锁定)
Challenges and shortcomings of GPT-34 分钟 17 秒
-
(已锁定)
GLaM3 分钟 6 秒
-
(已锁定)
Megatron-Turing NLG Model1 分钟 59 秒
-
(已锁定)
Gopher5 分钟 23 秒
-
(已锁定)
Scaling laws3 分钟 14 秒
-
(已锁定)
Chinchilla7 分钟 53 秒
-
(已锁定)
BIG-bench4 分钟 24 秒
-
(已锁定)
PaLM5 分钟 49 秒
-
(已锁定)
OPT and BLOOM2 分钟 51 秒
-
(已锁定)
GitHub models2 分钟 43 秒
-
(已锁定)
Accessing Large Language Models using an API6 分钟 25 秒
-
(已锁定)
Inference time vs. pre-training4 分钟 5 秒
-
-