Open Source AI
Mark Hinkle
I publish a network of AI newsletters for business under The Artificially Intelligent Enterprise Network and I run a B2B AI Consultancy Peripety Labs. I love dogs and Brazilian Jiu Jitsu.
This week's edition focused on Open Source after my visit to All Things Open Conference , it was top of mind. If you want to read the whole edition of the newsletter. I suggest you check out the full edition of The Artificially Intelligent Enterprise .
The Open Source Initiative's (OSI) release of version 1.0 of the Open Source AI Definition (OSAID) at the All Things Open 2024 conference in Raleigh, North Carolina, may be well-intentioned but raises several practical and strategic concerns. While touted as a breakthrough after years of collaboration among tech giants, their official endorsements are noticeably missing on the endorsements page .
Brief Overview of Open Source
Think of open source like sharing a recipe: instead of keeping it secret, anyone can see, use, and improve it. Open source software is code available to everyone, allowing users to understand, modify, and enhance it. This is different from most software, where only the creator controls how it works.
Why Should You Care?
Open Source in Your Daily Life
Open source is all around you:
How Open Source Set the Stage for AI
The LAMP stack (Linux, Apache, MySQL, and PHP) exemplifies how open source has allowed industries to innovate faster. By building the web’s foundation, LAMP helped companies and developers worldwide focus on creating new digital experiences without reinventing core technology. This shared framework enabled the rapid development of web services and cloud computing platforms, which are now crucial for artificial intelligence (AI).
Today, open source is helping AI progress in similar ways. Tools like TensorFlow, PyTorch, and OpenAI’s models provide shared foundations that developers and companies can build upon, allowing the industry to move forward faster while supporting collaboration, transparency, and innovation.
By choosing open source, you’re supporting a system that accelerates development, shares knowledge, and prioritizes openness for the good of everyone.
Open Source has proven to be a powerful catalyst for innovation. It demonstrates that immense benefits accrue to everyone by removing barriers to learning, using, sharing, and improving software systems. These benefits arise from licenses that adhere to the Open Source Definition (OSD), granting key freedoms to use, study, modify, and distribute software without excessive restriction.
The same freedoms are essential for AI to enable developers, deployers, and end users to benefit from enhanced autonomy, transparency, frictionless reuse, and collaborative improvement. However, with the rise of large language models (LLMs) like Meta’s Llama 3, it’s becoming increasingly clear that applying traditional open source licensing to AI introduces unique challenges—revealing a need for an adapted framework tailored to AI.
The Complexity of Open Source in AI: Square Peg, Round Hole
The Open Source Definition was developed with software in mind, and it works best for applications with standard dependencies, accessible codebases, and achievable reproducibility. LLMs diverge from these norms in ways that make it challenging to apply the OSD to them:
领英推荐
Case in Point: Meta’s Llama 3 License
Meta’s Llama 3 Community License Agreement offers a prominent example of how traditional open source licensing falls short in AI applications. While Meta has labeled Llama 3 as “open source,” its license diverges from OSD norms in several key ways, making the term “open source” somewhat misleading:
Lessons from the Past: Custom Licenses and Market Confusion
Meta is not the first to adopt a unique license structure and label it as “open source,” a practice that has historically caused market confusion. In the late 1990s, Sun Microsystems’ “Community Source License ” for Java introduced additional restrictions, leading many to question its “open source” label. Similarly, SugarCRM’s “Sugar Public License” introduced restrictive terms inconsistent with OSD, ultimately causing a backlash in the developer community. In both cases, these licenses sought to balance proprietary control with open source-like freedoms, creating confusion and fracturing trust in open source definitions. Meta’s “Community License” for Llama 3, by adding restrictions on use and redistribution, risks repeating these past mistakes, potentially confusing users and diluting the open source label.
Open Source Clarity: Avoiding Legal and Compliance Pitfalls
Many companies understand open source as software with specific freedoms and responsibilities, as defined by recognized standards like the Open Source Initiative (OSI). This clarity has allowed developers, legal teams, and business leaders to use, modify, and distribute software within a well-defined legal framework for decades. However, as new AI models and frameworks emerge that do not comply with traditional open source definitions, this shift can potentially lead to confusion or unintentional legal exposure.
The Risk of “Muddied” Open Source Definitions
Clear Paths Forward for Legal and Development Teams
To navigate these complexities, legal and development teams should focus on tools that strictly adhere to recognized open source definitions or collaborate closely to review the terms of newer frameworks. Ensuring that frameworks are used compliant and legally securely requires aligning with a company’s open source policies or adjusting them to account for this evolving category of AI tooling.
Proposing a New Approach: “Responsible AI Openness”
Rather than forcing AI models like Llama into the open source category, the industry could benefit from a new framework that upholds core open source values while accommodating AI-specific needs. A “Responsible AI Openness” framework could provide a more realistic and transparent approach by focusing on the following elements:
Conclusion: Reimagining Open Source for AI
For AI to benefit from open source’s foundational principles, it needs a framework that respects the unique challenges of LLMs. Calling models like Meta’s Llama 3 “open source” despite restrictions confuses the market and risks diminishing trust in the open source community. Instead, a “Responsible AI Openness” approach, focused on transparency, ethical usage, and responsible distribution, would be better suited to the needs of large language models while preserving the spirit of open source.
By embracing a new licensing model that accommodates AI’s complex requirements, we can foster collaboration, innovation, and transparency without misusing the open source label—paving the way for a more accessible and ethically responsible AI ecosystem.
CTO at if.team | Team Lead at Webnauts
1 周It can be dangerous to open big models, because it can be used for attacks, or creation something harmful. But in other hand it can develop AI faster
Hire FAANG talent on Discord | Used by top VC backed startups | Send me a DM for access ???
1 周https://discord.gg/learnmutiny
Business Operations Strategist | Digital Transformation Evangelist | AI Enthusiast | Tech Gadgets Lover | Foodie | Kindness
1 周The OSI’s OSAID release is a step forward, but more clarity on transparency is needed.
Power engineer || Energy Market || Six sigma Green Belt SSGB?
1 周There’s no denying the need for AI-specific definitions in open source. This is complex.
Helping startups and existing businesses transform their passion into profitable ventures, creating strong brands that resonates with customers and drive sustainable growth?? Increase awareness | Global reach ??
1 周Traditional open-source models don’t address the complexities of proprietary data in AI.