AI at a Crossroads: The Open vs. Closed-Source Debate and the Future of Large Language Models
Ritesh Vajariya
Global AI Strategy Leader | Head of GenAI @ Cerebras | Founder, AI Guru | Advisor to CEOs | Ex-AWS Gen AI Leader | Board Member
As we approach the end of 2023, a year that will undoubtedly be remembered as the 'Year of LLMs', it's fascinating to look back at the major milestones in AI. OpenAI launched their GPT-4 early in the year. This was followed by Anthropic's various releases of their model, Claude. Alongside these developments, numerous other proprietary language models emerged, considered as closed-source. Concurrently, we witnessed an influx of open-source models from different sources almost every month:
side bar: those who are not familiar with all these advancements of open-source LLMs, here are additional datapoints:
In February, Meta launched its first major open model, LLaMa, primarily for research and not for commercial use. Following this, April saw Eleuther AI release a range of 16 LLMs, varying from 70M to 12B parameters and trained on public data. This move was a game-changer, inspiring further research-based model development. MosaicML's introduction of the commercially viable MPT-7B model in May, which was developed at a significant cost but with high efficiency, caught the attention of industry giants like Databricks.
Hugging Face's collaboration with ServiceNow to create the StarCoder model and Salesforce's cost-effective X-Gen 7B were notable developments in June. TII UAE also made waves with their Falcon models, licensed under Apache 2.0 to encourage broader LLM adoption.
July heated up the competition as Meta released LLaMa 2, offering commercially viable licenses and sparking direct comparisons with TII UAE's Falcon models. The global landscape expanded with Alibaba’s Qwen model from China and Mistral's entry from France with their 7B model.
01.ai's bilingual Yi model, launched later in the year, demonstrated remarkable performance, challenging even established models like GPT-3.5 and LLaMa 2.
As the year wound down, Deci.ai prepared to launch a highly efficient 7B LLM. However, UpStage stole the show in December with their 10.7B model, topping the Hugging Face LLM leaderboard.
There is significant investment flowing towards building closed-source or proprietary LLMs, as evidenced by developments like GPT-4 from OpenAI, Anthropic's models, and others from companies such as AI21 and Cohere. At the same time, there is a notable momentum in the AI community towards developing open-source models.
So one may ask a question:
Should the future of AI be open-sourced or closed-sourced?
To answer that question, we need to look back at the evolution of our tech industry over the last 25 years. I remember using SCO-Unix on servers and DOS on IBM PCs during my college days; both were predominantly closed-source systems, reflective of the wider industry at the time. However, in the mid-90s, there was a shift towards Linux, an open-source operating system created by Linus Torvalds, primarily for server use. By the early 2000s, the LAMP stack (Linux, Apache, MySQL, and PHP) had become a staple, epitomized by its adoption by major startups like Facebook (now META), which I believe still uses some PHP in their main website infrastructure. Linux, in particular, has become a dominant force in server operating systems, a testament to the potential of open-source software. This period also witnessed significant developments in technology sectors such as databases and big data, exemplified by the Oracle vs. MySQL dynamics and the rise of Hadoop. These evolving landscapes set the stage for today's race in AI development.
Let's look at some of the pros and cons of closed-source and open-source LLMs.
Closed-source LLMs:
Pros:
Cons:
领英推荐
Open-sourced LLMs:
Pros:
Cons:
If history is any indicator, we might see significant consolidation in the open-source AI model landscape in the coming years, with only a few key players emerging as leaders. I am particularly optimistic about META, given their size and impact through contributions to PyTorch and the development of the LLaMa models. Similarly, TIIUAE, backed by the Abu Dhabi government's research council, with their investments in the Falcon models (7B, 40B, 180B), seems well-positioned. We might also witness the rise of a few startups, like Mosaic (well, now Databricks) and Mistral - mainly due to their novel architecture. [Disclaimer: These are my personal speculations based on current trends. I do not own stock in META nor am I affiliated with any of the mentioned entities.]
In the realm of closed-source models, we can expect continuous advancements and an increasing number of companies launching their proprietary models.
So, returning to the initial question: Should the future of AI be open-sourced or closed-sourced? In my opinion, it's too early to tell.
Let's allow the developments of the next 6-12 months to unfold, and then revisit this question in December next year.
What are your thoughts on the future direction of AI?
Wishing you all happy holidays and a joyful New Year!
Shameless Plug Alert! ??
Before you go, I can't resist sharing something I'm incredibly excited about — our brand-new AI Fundamentals Video Course! Designed for curious minds eager to delve into the world of AI, this course is a treasure trove of insights and knowledge. Whether you're starting from scratch or looking to brush up on the basics, our expert-led sessions are tailored to make AI approachable and fascinating.
Here's the kicker: You decide the price! That's right – we value your learning experience and commitment more than fixed fees. Pay what you feel is fair, and embark on a journey that demystifies AI in a practical, engaging way. Don't miss out on this unique learning opportunity – it's tailor-made for enthusiasts like you!
Check it out and jumpstart your AI adventure today!
Interest sign up page: https://bit.ly/ritesh-ai-bootcamp-signup
AI & ML lead | Generative AI | Responsible AI | AI for good
1 年Big thanks, Ritesh, for sharing your thoughts! ?? Taking a cue from smartphones, we see the diverse paths of Android (Linux-based and partly open source) and iOS (closed source). It's fascinating how they dominated the market, leaving others behind. Similarly, I firmly believe both Open source LLMs and closed-source LLMs can find their own devoted audiences. I will add to your article that, looking back, history teaches us that success isn't solely about performance; it's the perfect blend of user experience and ecosystem development. Winners emerge when they anticipate needs with an unwavering focus on the customer. The market is the ultimate decider, and I'm genuinely excited to witness how this space will evolve. Exciting times lie ahead! ??
Principal Strategic Advisor -- Credit Unions | Amazon Web Services (AWS)
1 年Excellent piece Ritesh Vajariya! I am leaning more and more towards allowing both closed and open-source foundation models to coexist, but there needs to be a few unrestricted, open-source models available to provide a basis for comparison against closed ones. All of the testing I am running on Claude, Bard, and ChatGPT is showing considerable variation in bias in their outputs and conclusions. This is quite concerning given the current political climate. There is too much power concentrated in the hands of a few, especially when dissenting perspectives are readily dismissed as "disinformation" rather than engaged with.?? Some models, Claude specifically, can be reasoned out of biased conclusions through sustained dialogue and introducing contradictory evidence. Others like Bard and ChatGPT tend to rigidly maintain whatever biases they have built-in. The more options available the better, as openness facilitates accountability.
Global Account Manager at Amazon Web Services
1 年Your articles are always interesting and insightful!
?? Generative AI Lead @ AWS ?? (100k+) | Startup Advisor | Public Speaker | AI Outsider
1 年Great stuff Ritesh!