The Fast Evolving AI Landscape
Anil Murty
VP - Product, Eng & Partnerships @ Overclock Labs - Decentralized Cloud Compute
It’s barely 3 months into 2023 and it is clear that AI is going to dominate the tech space this year, and very likely this decade, by a wide margin (over Web3, Cloud, Mobile or anything else in the past couple decades). If you are in tech, you have to be living under a rock to not see that AI represents a generational shift in not just the way we will use products but also in the way all products are built and how businesses operate. As such, I and the team I work with (at?Akash Network), feel fortunate to have been working on core compute infrastructure for a while, so we haven't had to make as crazy a pivot as some others in tech ;)
Before starting to work on our AI offering, I spent a fair amount of time educating myself about the AI landscape, starting around the time that ChatGPT got announced. And boy, have I had a lot of catching up to do for the last 6 months (and still struggle to keep pace with developments, each week)!
So, I figured I may save you some of the time I spent reading countless articles, blog posts, press releases, tweets, trying new products and building some prototypes myself... by sharing some high level learnings.
Intended Audience: people who have a general understanding of AI models and follow the “buzz” but are looking for a cohesive understanding of the players, use cases and high level model development process.
New Generative AI Models every week
On average one new?Generative AI?model was announced almost every week, thus far in 2023. Whether it be big announcements from?OpenAI,?Microsoft,?Google?and?Meta, or less “mainstream” models like?Stable Diffusion?(built through a collaboration between?CompVis,?StabilityAI?and?RunwayML) or non mainstream products and models like?Midjourney,?ControlNet,?Palm-E?or?Conformer-1.
Some clear themes here:
Race to the bottom with “Open AI API wrapped” App clones
Outside of a few of the big tech companies (MS, Google, Meta) who have their own Generative AI models and LLMs, most of the AI productivity applications being built today (copy.ai, jasper.ai, playground.ai and countless others) are essentially wrappers on top of Open AI’s APIs. While they are seeing a lot of buzz, it is hard to imagine this continuing forever. While some of these will go on to generate 100s of millions of dollars in revenue, eventually this will be a race to the bottom for many of these startups and they will either consolidate or get acquired or just fade away. Most of the productivity AI applications will get integrated into existing bigger “traditional” applications (MSWord, GDocs, Slack, Teams, Salesforce, Hubspot, Notion, Adobe, Autodesk, VSCode, Replit, Github etc).
The disruption to the above trend will only happen when one of the non-OpenAI models surpasses them in capabilities, efficiency, price and/ or functionality. This may happen from relatively unknown companies like Assembly.ai or from one of the big tech models (subsequent iterations/ generations of?BERT,?PaLM-E,?LLaMA?and others) or both.
Users may favor non-OpenAI models and services, if they prefer greater control as well. This definitely has similarities to the Apple (closed) + Android (open) ecosystem, both thriving alongside each other, in some ways.
And to be clear, there are some very valid and lucrative cases for both text analysis (in productivity apps like MS O365, GSuite, Notion and others) as well as for image generation (in Adobe, Autodesk, Canva and others). Here is a quick run down of the use cases for AI image generation:
If you are looking for more apps built on Open AI (and other models like Stable Diffusion) APIs, gpt3demo has a nicely sorted collection of more than 700 such apps, that seems fairly up to date.
GPUs are the new oil (for tech)
GPUs from Nvidia have been in high-demand for several months now, with demand showing no signs of slowing. The challenge is sourcing GPUs (like the?A100s?and?H100s) that can allow these applications to run at a level of performance that doesn’t hurt user experience. This favors the big players who have either amassed a large inventory of GPUs or have the volume/ demand leverage to acquire them or are building their own tensor optimized hardware. It also puts Nvidia in a position of very strong leverage which it continues to try to cement by building it’s own cloud/ service layer (Nvidia DGX Cloud) that sits between the big clouds and Nvidia’s GPUs.
There are 4 things that will upset this equation:
The key thing to keep in mind is that the GPU market was already predicted to explode by 10x, even before these new generative AI models were invented.?Here?is a quick post I wrote 6 months ago, about what an “EMBARRASSINGLY PARALLEL” workload is and how it is the driving force behind the explosive growth of GPU usage, over the past decade.
Open Source is alive and well in AI/ML
While the buzz around OpenAI (which to be clear is NOT open source, even though it originally started out that way) and its announcements overshadow everything else, it is becoming clear that Open Source isn’t dead in AI/ML.
I’d love to hear from you if you have references to other open source projects for AI and ML.
It also remains to be seen, whether other internet platforms (particularly open ones, or ones trying to be more open,?like twitter), decide to play nice with Open AI or not:
领英推荐
AI Workload Types
To help understand the stages in AI model training and use, here is a quick and simplified view. After a base model (like?generic BERT?or?Base Stable Diffusion) has been developed (i.e. the model has been authored, neural network designed, implemented, data tokenized and trained with a base set of parameters), the following are the stages to using it in actual user applications :
Step1:?Hyperparameter Search or Tuning
Step2: Large-Scale Distributed Training
Step3: Production Inference
AI User Ecosystem
The ecosystem of products, companies and services that enable us to train, build and productize AI can be broken down as follows:
Model & App Builders
This segment can further be divided into 3 sub segments
AI Model Creators:
These are scientific groups and companies that have the deepest knowledge and expertise in AI and ML and are engaged in building and training the models that power the applications. Popular models include,?Stable Diffusion?from CompVis, Google’s?BERT,?LaMDA?&?AudioLM, Open AI’s?Whsiper,?GPT?and?DALL-E, Meta’s?LLaMA, and more. In some cases, these companies and groups also produce the end-user application — as in the case of?Google BARD?and?ChatGPT. In other cases the end user app is built by someone else — as in the case of Stable Diffusion or ControlNet. These users may be the authors of the model or initial (base) implementors or both. They are also running all 3 steps outlined in the previous section. In most cases, they will have access to their own infrastructure (and in many cases are a big cloud provider itself).
AI Model Tuners:
These are users who take a base model and tweak it to improve it in some way to enable it to produce more accurate results. A good example of this is Stable Diffusion, which has two variants, one produced by?StabilityAI?and another produced by?RunwayML. These users often don’t invent the model themselves but rather take a paper/ proposal and implement it or take an existing implementation and tune it for a specific type of application. These users are NOT inventing the model but usually must run through all 3 stages outlined in the previous section.
Pure App Developers:
There are a multitude of applications across various Enterprise and Consumer segments that are being built by companies of all sizes (startup to fortune 100). All these applications, rely on a handful of pretrained models (like the ones above) but serve the following categories
These users are typically only running the inference stage of the 3 stages outlined before, because they are taking a pre-trained and tuned model and building a nice user experience on it. They are typically just doing the inference portion of it. In a way, each time you use ChatGPT, you are using one such application that is running inference (aka, making a prediction about what text you are expecting when you ask it a question).
AI Tool Builders:
These are teams and companies that are building tools to make it easy to discover, package and deploy AI models and in some cases offer these as services. Examples include,?Cog and Chassis?(for containerizing ML models),?Huggingface?(for discovering, sharing, reusing, models and ML datasets — essentially the github for AI/ ML),?Replicate?and?Banana.dev?(Deploying+Running AI models) as well as full cycle platforms for training, deploying and managing ML models — like?DataRobot,?NeptuneAI?and?VertaAI.
AI Infrastructure Providers:
This is very core to all AI and ML development and advancement, regardless of what user segment or application is in question. This includes traditional public cloud providers, newer GPU focused providers (like Lamba Labs and Coreweave) as well as Traditional Datacenter and Cloud Hosting providers starting to offer GPUs.
Akash Network?(that we are building, with a?small core team?and?large open source community) is unique in this respect, in that it:
Sami Kassab (analyst at Messari) has written extensively about this in his?latest report, that came out this morning. Here are a couple tweets from today about it:
Lastly, we will be sharing more about our AI plans at the?Cerebral Valley AI Summit?and would love to chat with you in-person if you are attending or in the area!
If you are interested in learning more about our offering, please?sign up here?for early access!
passionate about building products
1 年Great read, thanks, Anil! Can't be more bullish about Akash :)
Product Manager @ Meta
1 年Quite informative. Great write up, Anil.
Architecting AI | Data | Observability Solutions
1 年Great write up Anil. Have been following Aakash and awesome work you all are doing. Getting A100, H1000 on same datacenter spine is very difficult but also very efficient for distributed training on node-to-node. If your providers can avail this, it would be huge!
Digital Transformation Sherpa?? Helping Reimagine Business with AI and Automation | Google Cloud Digital Leader | Product Engineering Maven | Partnerships & Alliances Expert | Follow me on X @sanjaykalra
1 年Great picture - captures it all!