The promise of Foundation models and where we are now

Unlike traditional models that are designed for specific tasks, Foundation models are vast, pre-trained models that serve as a base for many applications across different domains. Stanford’s Center for Research on FMs, define it as “Train one model on a huge amount of data and adapt it to many applications.”. Many foundation models are also fine tuned models on top of larger pre-trained base models. Links below use different terms to point to such models FMs or Large Foundational Models (LFMs), Frontier Models, or Large X Model (LXM) etc.

Source: https://viso.ai/deep-learning/foundation-models/

The rise of Foundation Models came with a lot promises. This post is a brief survey to find where we are with those promises.

So first, what were those promises? These models will:

a. apply to wide ranging real-world applications within and across many domains,

b. be easy to adapt and extend to different modalities,

c. scale efficiently,

d. reduce the need for high quality, labeled data and

e. generally enable rapid development in AI applications.


There are many other factors that are leading the industry further and further towards FMs:

·????? computation challenge (exponentially more to build many speciality models),

·????? applicability & usage (surge in the number of use cases that demands the need for ML based augmentation),

·????? pace of adoption (pace of adoption across the industries),

·????? cost effectiveness (economics in serving multiple applications),

·????? business opportunities and competition (e.g. data and compute accessibility has become a moat for big tech companies).

Of course, this is in addition to the architectures that became available.

I was curious to know what’s happening with this promise. So, I decided to look up and make a list.


As I went through this exercise, I was amazed by what’s being put out there. I don’t know how good they are in real world applications but it appears like we are in a very exciting journey.

However, extending Foundation models to other modalities and applications across various domains presents several challenges - governance and regulation to ensure responsible development and deployment, potential misuse such as misleading content, managing biases, applications with unintended impact. The important part is that these models, especially OSS, cannot be stopped from proliferation making them a difficult challenge to manage.

Conclusion:

1.??? Foundation models are rapidly growing, and new sub-modalities are emerging quite rapidly. The original promise remains alive.

2.??? Combining these FMs with intelligent routers, like with MoEs, creates powerful systems by dynamically selecting the most relevant models for specific tasks. E.g. projects like MIT’s HiP, highlights their potential to revolutionize complex planning and decision-making processes in robotics and beyond.

3.??? Interesting new modalities, such as emotions, haptic feedback, olfactory data, and biosignal processing, opens up exciting possibilities their applications and impact across diverse fields.

4. The path to AGI/ASI, whenever we see it, is likely paved through the advances and integration of these foundation models.

In any case, a very exciting space to watch out for.


Useful References:

https://github.com/uncbiag/Awesome-Foundation-Models

Bitter Lesson

Frontier AI Regulation

The White House report

AI Foundation models report from UK’s Competition and Markets authority

Foundation models for generalist medical artificial intelligence

Foundation models for Decision making: Problems, Methods and Opportunities

https://news.ycombinator.com/item?id=38472128

Can Generalist Foundation Models Outcompete Special-Purpose Tuning? Case Study in Medicine

From GPT-4 to Gemini and Beyond: Assessing the Landscape of MLLMs on Generalizability, Trustworthiness, and Causality through Four Modalities

BOP Benchmark https://bop.felk.cvut.cz/challenges/bop-challenge-2023/#task4


Rajmeet Gandhi

Director Technical Architecture

6 个月

Good share and point of view Rajesh Chitharanjan Excellent consolidation of modalities and relevant models. Thanks for sharing resources as well.

要查看或添加评论,请登录

Rajesh Chitharanjan的更多文章

  • AI Agents in the Wild - Project Sid

    AI Agents in the Wild - Project Sid

    How would 1000s of AI agents behave if they are let loose in a digital world? Would they form their own societies and…

    15 条评论
  • Apple’s move towards PersonalAI. What’s missing?

    Apple’s move towards PersonalAI. What’s missing?

    This week’s curiosity post. AppleIntelligence created a lot of buzz everywhere.

    2 条评论
  • Do we really know how large models work now?

    Do we really know how large models work now?

    In case you didn’t know this. We really do not how large models that we use regularly, work.

    5 条评论
  • Rising Tides of Innovation

    Rising Tides of Innovation

    This is a post based on a curiosity drive (prompted by a book I was reading). Took a few hours of reading and some…

    4 条评论
  • The landscape of Federated Learning. What lies ahead?

    The landscape of Federated Learning. What lies ahead?

    While there’s a massive contest now between companies on large model supremacy, it's becoming increasingly clear that…

    3 条评论

社区洞察

其他会员也浏览了