课程: Large Language Models on AWS: Building and Deploying Open-Source LLMs
免费学习该课程!
今天就开通帐号,24,700 门业界名师课程任您挑!
GGUF file format
课程: Large Language Models on AWS: Building and Deploying Open-Source LLMs
GGUF file format
- [Presenter] This is the GGUF format architecture. something you'll hear a lot when running local models, especially with llama.cpp. So the big picture is that it helps AI models run efficiently. There are three main pieces. There's the original model, there's the GGUF format, and there's llama.cpp. Think of GGUF as a bridge between doing research, let's say someone that was training a model, let's say it's Allen AI, and practical deployment. From a starting point, the models typically are using things like PyTorch or Hugging Face. And these are great for training but not optimized for deployment. They often come with multiple files and dependencies. If we look at the GGUF format in the center here, it combines everything into a single file. So the model weights, this is the tensors, the configuration details, the tokenized information, the architecture metadata, and everything is packaged together so that you can run the model. So this is pretty convenient, because instead of having…
内容
-
-
-
(已锁定)
Implications of Amdahl’s law: A walkthrough4 分钟 5 秒
-
(已锁定)
Compiling llama.cpp demo4 分钟 17 秒
-
(已锁定)
GGUF file format3 分钟 18 秒
-
(已锁定)
Python UV scripting3 分钟 55 秒
-
Python UV packaging overview1 分钟 59 秒
-
(已锁定)
Key concepts in llama.cpp walkthrough4 分钟 37 秒
-
(已锁定)
GGUF quantized llama.cpp end-to-end demo4 分钟 3 秒
-
(已锁定)
Llama.cpp on AWS G5 demo4 分钟 20 秒
-
(已锁定)
-