The Emerging LLM Value Chain for Enterprise use (part 2)
This is a continuation from part 1.
TL;DR
Hand wavy sure, but this is what i see
I'm going to try and speed run this a little bit for a couple of reasons
Domain Adaptation
Something was built with some inputs X, but you want to use it for purpose Y. How do you do it?
This will be the next big leap in LLM use.
The common practice of fine-tuning the model is not only resource-intensive and complex to manage, but it also does not always clearly indicate how to incorporate new knowledge. For example, fine-tuning on a text such as “Alice in Wonderland” does not equip the model to answer questions about the story itself, but rather it trains the model to predict the next token or complete masked sentences
So the somewhat hackneyed aphorisim about LLMs is that they are like a really smart graduate. They seem to know a lot about a whole bunch of things, but their depth on any particular topic is better than the average joe, but not amazing.
There is a reason that new grads are kind of given a few years in junior roles to start to learn the domain expertise to apply the kind of general aptitude their degree is meant to signal. Actual usefulness is often some combination of general aptitude with domain specificity to be able to appreciate, understand, and then solve problems specific to their employer.
So if the big, million/billion/trillion parameter LLMs have ingested the whole internet and are one of these smart grads, how do you give the LLM a masters in Finance? or get them a CII diploma in the London Market?
This is the purpose of domain adaptation and its where an awful lot of the cutting edge research is at right now.
领英推荐
The next piece therefore of the value chain that will emerge will be the domain adaptation element.
There are probably a couple of different forms that this will take:
I think option 1 is dead for the same reason that commercial models like OpenAI etc are dead. However they will present a challenge to entreprenurs in the other layers as they will likely be bundled, free, by the cloud providers, which means any competitive product would have to compete with 'free' and 'already with an MSA', which are bloody tough points of competition
I think 2 will be a the major transitory state in a world where everything is new, complexity is high, and rate of change is rapid. There is clearly value in the service, someone has to do it, and the market at this layer is probably sufficient that a couple of players will merge to serve the most easily addressable domains
I think 3 is the longer term answer, as we become more familiar with the strengths and weakness of different generative ai models and approaches, as the science of domain adaptation starts to gravitate around a couple of key concepts, and spinning up an LLM for a given task becomes something as easy as spinning up an e2-standard or an m3.large.
You will pick the base model, the domain, and the resulting domain adapted model will be spat out and ready for inferencing.
Use case fine-tuning
...will have to wait, i need to go to bed