Mixture of Experts Explained

Mixture of Experts Explained

This iteration is sponsored by my friends at Qdrant !

Good morning everyone!

What you know about Mixture of Experts (MoE) is wrong.

We are not using this technique because each "model" is an expert on a specific topic.

In fact, each of these so-called experts is not an individual model but something much simpler...

So, MoEs are not experts or even models... What are they, and why does it work?

Let's dive into MoEs and discover why it is so powerful!

But first, here's a new, very cool product for all of us experimenting and building RAG -based applications!

1?? Qdrant Hybrid Cloud: The First Managed Vector Database You Can Run Anywhere with Unmatched Flexibility and Control (Sponsor)

Qdrant, the leading open-source vector database, today announced Qdrant Hybrid Cloud, a groundbreaking managed service for deployment across cloud, on-premises, or edge settings. Built on a Kubernetes-native architecture, it offers flexibility in setup and ensures full database isolation, enhancing data privacy in AI. This allows developers like us to choose where to process vector search workloads, making it easier to work on RAG-based applications, advanced semantic search, or recommendation systems in a data-driven world.

Read more about Qdrant Hybrid Cloud in Qdrant’s official release announcement.

2?? Mixture of Experts: Mixtral 8x7B Dive in

Thanks to Jensen, we can now assume that the rumour of GPT-4 having 1.8 trillion parameters is true

1.8 trillion is 1,800 billion, which is 1.8 million million. If we could find someone to process each of these parameters in a second, which would basically be to ask you to do a complex multiplication with values like these, it would take them 57,000 years! Again, assuming you can do that in a second. If we do this all together, calculating one parameter per second with 8 billion people, we could achieve this in 2.6 days. Yet, transformer models do this in milliseconds.

This is thanks to a lot of engineering, including what we call a “mixture of experts,” where we supposedly have eight smaller models put together to reach this ginormous single model. But do we? Learn more in the article here or the video:


And that's it for this iteration! I'm incredibly grateful that?the What's AI newsletter ?is now read by over 16,000 incredible human beings. Click here to share this iteration with a friend if you learned something new!


Looking for more cool AI stuff? ??

Want to share a product, event or course with my AI community? Reply directly to this email, or visit my Passionfroot profile to see my offers.


Thank you for reading, and I wish you a fantastic week! Be sure to have?enough sleep and physical activities next week!


Louis-Fran?ois Bouchard

要查看或添加评论,请登录

Louis-Fran?ois Bouchard的更多文章

  • When is AI Integration REALLY Worth the Hype?

    When is AI Integration REALLY Worth the Hype?

    Good morning everyone! In this iteration, we discuss the biggest problem with AI: usefulness vs. pure marketing (hype).

    1 条评论
  • Advanced RAG Evaluation Techniques for Optimal LLM Performance

    Advanced RAG Evaluation Techniques for Optimal LLM Performance

    Good morning everyone! You are probably implementing RAG systems, missing out on easy improvements. Most don’t even…

    2 条评论
  • Indexing Methods for Vector Retrieval

    Indexing Methods for Vector Retrieval

    Good morning everyone! ChatGPT is pretty useful, but you can’t build a product out of it. I have been talking about RAG…

    2 条评论
  • Releasing our 90+ lesson practical LLM Developer course!

    Releasing our 90+ lesson practical LLM Developer course!

    I am super excited to finally announce that we ( Towards AI ) released our first independent industry-focussed course:…

    1 条评论
  • LLM Evaluations: Find the Best AI Model for Your Specific Task (no code)

    LLM Evaluations: Find the Best AI Model for Your Specific Task (no code)

    Good morning, everyone! Today, I’m excited to dive into an essential topic: why not all Large Language Models (LLMs)…

    4 条评论
  • Master Multi-Agent Systems Like a PRO with AGENTIC AI

    Master Multi-Agent Systems Like a PRO with AGENTIC AI

    Good morning everyone! In this iteration, I'm partnering with the amazing team at Integrail to make a really cool…

  • Running YOLOv7 on Your Phone

    Running YOLOv7 on Your Phone

    Good morning everyone! In this iteration, we are quickly reviewing how much the field of real-time object detection has…

    1 条评论
  • Building LLMs for Production now available everywhere!

    Building LLMs for Production now available everywhere!

    Good morning everyone! Today, I’m super excited to announce that "Building LLMs for Production" is now also available…

    2 条评论
  • A big Update for Building LLMs for Production!

    A big Update for Building LLMs for Production!

    Good morning everyone! Today, I’m super excited to announce that a new and improved version of Building LLMs for…

    18 条评论
  • Teaching AI to "Think"

    Teaching AI to "Think"

    Good morning, everyone! Like everyone else, we already talked about OpenAI's newest o1 model series, exploring how it…

    2 条评论

社区洞察

其他会员也浏览了