Introduction to the Customization of LLMs
Article 5: LLM Series articles, previous article Understanding (LLMs): How Do They Collect Data? | LinkedIn
In modern large language models (LLMs) like ChatGPT, customization plays a key role. While the master LLM—the main model—comes pre-trained with a broad set of knowledge, users often need the model to focus on their specific needs, whether it's handling their own data, providing specialized information, or answering domain-specific questions. However, when users request these custom capabilities, they do so without changing the master LLM itself. Instead, the master LLM provides technologies that allow for these custom needs to be met while keeping the core model intact.
When customers or organizations need the LLM to focus on their unique data or knowledge, they can rely on the master LLM for support. The master LLM provides three main customization methods: RAG, fine-tuning, and custom GPT. These methods help users build what we call custom sub LLMs (or child LLMs), tailored to handle their specific requirements.
Think of the master LLM as a main brain with a wealth of solid, pre-trained knowledge that doesn’t change “unless done by the main owner of the master LLM”. It is designed using mathematical structures, algorithms, and static training data. But it’s also flexible enough to assist users in creating custom LLMs for their own needs, based on external data or specific requests.
It’s important to note that when the master LLM helps build custom LLMs, it does so without altering its own core knowledge base or thinking. Instead, it supports these child LLMs by using its existing capabilities to interpret new data, answer specific queries, or integrate user-provided information. It’s as if the master brain is lending its thought process to assist in handling specialized tasks, but it doesn’t change its own thinking—it's more like an assistant that helps others with custom tasks while keeping its own knowledge intact.
This approach allows users to build small, specialized LLMs for their own purposes, such as answering questions or managing specific types of data. These child LLMs work together with the master LLM to offer precise, personalized responses based on the user’s needs.
Master LLM and Custom Sub LLMs: How They Work Together
The Master LLM (represented as a large central circle) is the core model, like ChatGPT’s master LLM, which contains general-purpose knowledge trained across a wide variety of topics and domains. This master LLM remains unchanged and serves as the base for all interactions, ensuring that the model can respond across many different areas without being specifically fine-tuned for one.
Around the master LLM are customer-specific sub LLMs (represented by smaller connected circles). Each customer-specific LLM is tailored for a particular user, business, or organization. These custom sub LLMs do not modify the master LLM, but rather, they are created to address specific needs through three key methods RAG (Retrieval-Augmented Generation),Fine-Tuning and Custom GPT.
领英推荐
Each customer-specific sub LLM acts as a child LLM, integrating seamlessly with the master LLM while maintaining its own specialization. This flexible architecture allows customers to create powerful, tailored solutions for their own unique needs, without impacting the broader model’s general knowledge and capabilities.
1. Retrieval-Augmented Generation (RAG):
2. Fine-Tuning:
3. Custom GPT: