课程: Large Language Models on AWS: Building and Deploying Open-Source LLMs
免费学习该课程!
今天就开通帐号,24,700 门业界名师课程任您挑!
Key concepts in llama.cpp walkthrough
课程: Large Language Models on AWS: Building and Deploying Open-Source LLMs
Key concepts in llama.cpp walkthrough
- [Instructor] Let's talk through this Qwen2 coder deployment pipeline, A comprehensive guide from a high level view. And this guide is really interesting because it shows probably the most cutting edge AI coding assistant workflow that's local that you can use because we use Llama.cpp and we have full control of every single step. And this process involves several stages. Each one is a crucial role in making it run efficiently on my specific hardware. So first up here we have the Hugging Face model download stage, and this place we have a central repository for AI models. You can think of it as a GitHub for AI. It has thousands of models, including this Qwen2.5 coder from Ali. And this provides a access to a state-of-the-art coding assistant that's on par with mini commercial models. First up, we have the heavy lifting, which is we have to download this model and it's 32 gigabytes. So it's a huge, huge model. But then we go through and we get the Hugging Face CLI to download it, and…
内容
-
-
-
(已锁定)
Implications of Amdahl’s law: A walkthrough4 分钟 5 秒
-
(已锁定)
Compiling llama.cpp demo4 分钟 17 秒
-
(已锁定)
GGUF file format3 分钟 18 秒
-
(已锁定)
Python UV scripting3 分钟 55 秒
-
Python UV packaging overview1 分钟 59 秒
-
(已锁定)
Key concepts in llama.cpp walkthrough4 分钟 37 秒
-
(已锁定)
GGUF quantized llama.cpp end-to-end demo4 分钟 3 秒
-
(已锁定)
Llama.cpp on AWS G5 demo4 分钟 20 秒
-
(已锁定)
-