Iris Insights by DEUS | Issue 2
DEUS: human(ity)-centered AI
We combine data, design & tech to build human(ity)-centered products & services.
Welcome to this month’s issue of Iris Insights, your monthly guide to navigating the ever-evolving world of AI.
In this issue, we bring you:
???a digest of some of the main stories in AI this month
???an overview on AI agents
???an overview of the main principles to use when designing conversational AI
???tools, repos and datasets to experiment with
???trends to keep an eye on
and our new section, the AI telenovela, describing all the borderline-absurd things that have happened in AI this month.
We’re excited to share this month’s edition with you and help you cut through the noise in AI.
?? AI Digest
The European Parliament adopts the AI Act
The EU Council approved the AI Act , establishing the first comprehensive regulatory framework for artificial intelligence. The Act is set to take effect in August, 2024 .
Unchanged provisions:
Refined provisions:
What this means
The AI Act is the world’s first comprehensive regulation of its kind. Similarly to GDPR, it is likely that it’ll influence AI regulations worldwide. Despite criticisms and lobbying efforts , the Act is now in place, and compliance is mandatory for companies operating in the EU.
While the Act's legal specifics are beyond our scope, we believe one notable shift in its text will have implications for the industry that go beyond compliance: the explicit distinction between AI models and AI systems.
Currently, there are multiple ways to benchmark the performance of a model, but no widely available benchmarks exist for assessing AI systems’ overall impact. Yet, the AI Act is implicitly making the understanding of AI system’s impact kind of mandatory.
Some companies have attempted to address this through "responsible AI principles." However, most haven't progressed beyond outlining broad guidelines or narrowly focusing on bias reduction and fairness.
Truly measuring impact should be more comprehensive. It needs to encompass, to name a few:
This shift from model-centric to system-wide assessment represents a significant challenge and opportunity for the AI industry, potentially driving innovation in responsible AI development and deployment.
Google’s problematic conversational search
On May 14th, Google rolled out its new AI Overviews feature, similar to the summaries offered by its rival Perplexity.
The feature quickly came under scrutiny when people found out that it can produce inaccurate and potentially dangerous (or humorous, depending on your point of view) answers. Some notable examples include the suggestions to put glue on pizza, eat rocks, and run with scissors to improve your health.
In response, Google released a blog post explaining the cause of these errors. The TL;DR version: the inaccurate answers that circulated online stemmed from very uncommon search queries for which the system lacked sufficient data, and many of the examples were faked.
They did vow to refine and retool their summaries.
What this means
Google's dominance in the search engine market has been so profound that "googling" became synonymous with web searching. The company also pioneered cutting-edge AI applications like self-driving cars and AlphaGo . However, Google has faced a series of setbacks with its AI product launches in the past few years.
First, there was Bard’s inaccurate answers, which caused Google to lose $100 billion in market value. Then, Gemini generated ethnically diverse images of Nazis . Now, AI Overviews is in the spotlight for its inaccurate results.
Does this mean Google’s products are worse than its competitors? Partially, but not by a huge margin. State-of-the-art AI models like GPT-4o, Llama, Claude, and Gemini are generally on par with each other.
See the recent introduction of Anthropic’s Claude 3.5 Sonnet , which outperforms all models on the market (according to Anthropic). You can also access the model for free on their site.
Then what’s going on?
We’re witnessing the beginning of AI commoditisation, where technological superiority alone is no longer a sufficient competitive advantage.
Moreover, none of the current AI frontrunners can leverage data as a competitive advantage either, as they don't own the data they train on. They rely on publicly available data created by journalists and internet users, leading to an increasing number of copyright lawsuits.
Perplexity (which is now valued at $1 billion ) was found to be scraping content by ignoring the widely accepted web standard known as the Robots Exclusion Protocol. They’re also currently being sued by Forbes for plagiarism.
Without a clear edge in data or technology, Google (and other AI companies) must find new ways to differentiate themselves.
The coming months will be crucial. Will Google continue to struggle with its AI releases while Perplexity gains traction, or will Google pull a Zuckerberg—copying the most popular features of its competitors, much like Instagram did with Snap’s stories?
The “copy” strategy might not be as effective for Google, though. Switching from Google search to Perplexity is as simple as typing a new URL. The user experiences no “social” loss as she’d do if she switches away from Instagram. Moreover, Google’s search monopoly means it faces much higher accuracy expectations than Perplexity or OpenAI.
This situation presents a fascinating case study of incumbents versus challengers in the AI space. Interestingly, tech giants like Google, Microsoft, Amazon, Meta, Apple, and Nvidia continue to heavily invest in AI startups such as OpenAI, Anthropic, and Perplexity, further complicating the competitive landscape.
The U.S. Federal Trade Commission and Justice department are launching an investigation into Microsoft, Nvidia and OpenAI for antitrust practices.
One thing is certain: as the AI industry evolves, companies will need to find unique value propositions beyond raw technological capability to maintain their market positions and user trust.
Apple Intelligence: the on-device battle
During their Worldwide Developers Conference , Apple introduced the so-called Apple Intelligence a.k.a., their take on how AI will be integrated across their device ecosystem. This new feature is set to help users manage notifications, craft replies, and use native Apple apps more productively. All these will be performed either on-device or in a private cloud.
Following its partnership with OpenAI, Apple also announced a big upgrade to Siri, where questions will be sent to ChatGPT if Siri is not able to answer them.
What this means
Apple's recent focus on on-device AI and private cloud solutions reinforces its long-standing commitment to a privacy-first strategy.
However, the introduction of ChatGPT in Siri raised privacy concerns. Most notably, by Elon Musk . He went so far as to claim that he would require Apple devices to be placed in Faraday cages when entering his companies' premises, effectively blocking all wireless signals to and from the devices. He then proceeded to bash Apple for its lack of “good models”, only to be fact-checked by his own platform.
This all happens against the backdrop of Microsoft having to recall its Recall feature , which was designed to operate on-device and serve a function similar to Apple Intelligence.
However, Apple’s move also serves as a prime example of how power consolidation can impact startups and scale-ups. The introduction of native AI features on Apple devices threatens to render many third-party applications, such as Grammarly and Goodnotes, redundant, as their core functions will be seamlessly integrated into pre-installed Apple apps.
Recognising the potential antitrust implications of this move, Apple preemptively announced that Apple Intelligence will not be available in the EU this year . This decision stems from concerns about potential breaches of the Digital Markets Act, which was designed to prevent monopolistic behaviour.
领英推荐
This trend of consolidation is also reflected in the recent 76%-drop of VC funding in Generative AI. Additionally, the company behind the Humane AI Pin, which faced multiple negative reviews , is now reportedly on the market with a $1 billion price tag .
Beyond privacy and antitrust concerns, the move towards on-device AI showcases a likely future of AI-human interactions: one where AI operates in the background, enhancing task performance and efficiency. Both Microsoft and OpenAI are actively vying for a similar role with their respective offerings Copilot and the new app OpenAI announced a month ago, but hasn’t released yet.
See the secretive AI device Johny Ive and Sam Altman are reportedly trying to create.
Apple's approach differs slightly, aiming to deeply integrate AI capabilities into all of its apps. Based on their recent demonstrations, Apple seems to be positioning AI both as an additional tool in its apps and as a traditional assistant (i.e., Siri).
See the paper Apple published on UI patterns for Apple Intelligence
As this landscape continues to evolve, it will be crucial to monitor how these strategies impact user experience, market competition, and how companies try to identify their unique value proposition. The balance among these will likely remain a central challenge in the AI industry.
Brain byte ??
We can break down an LLM-powered autonomous agent system in 5 modules:?an agent?with predefined instructions, a?memory?system for managing information storage and retrieval, specialised?tools?like web search, a?plan?module that develops strategies, decomposes objectives, and reflects on them, and an?action?module to execute decisions and plans.?
Although, not mandatory, it’s wise to introduce a?human-in-the-loop mechanism, too. That way you can ensure the agent behaves as intended at all times.
Want to learn more??Download our primer on AI agents.
Spotlight ??
Three main principles for designing conversational AI
The narrative around conversational AI taking over the digital world has been gaining momentum, especially after the release of GPT-4o and Project Astra.
Whether you believe this or not, one thing is certain—machine capabilities are finally catching up to the long-held human dream of the ever-capable AI assistant. And they’re about to change the way people expect to interact with your business.
The key to good conversational AI lies in designing conversations that go beyond a single medium. Think rich dialogues that guide users to make informed decisions and complete tasks effectively.
Here are three principles for designing useful conversational AI.
Want to learn more??Download our guide on conversational AI.
Sandbox ????
Each month we hand-pick a selection of models, tools and repos that can give you a hands-on experience with the latest developments in the field.
This month, we’re focusing on:
Tutorials
AI agents
Datasets
Video generation tools/repos
Modalities
Robots
Our Watchlist ??
Some of the most interesting trends in AI this month:
Approaches to model explainability
Tiktok
Robots and LLMs
Geo-politics
The AI telenovela ??
Some call it the AI reality show, others Game of Thrones.
We prefer to call it “the AI telenovela” as no other term properly reflects the dramatic flair, back-stabbing, and absurdism that defines almost everything that’s been going on around AI monetisation efforts.
Enter our main protagonist (or antagonist - depending on your bend) - OpenAI.
Here’s what happened to OpenAI since the announcement of GPT-4o 1 month ago:
Meanwhile,
Stay tuned for the next episode!
Thanks for reading & see you next month!???
Iris Insights is written by the people at DEUS Humanity-Centered AI, a technology & design company focused on applied AI. Want to receive this newsletter in your inbox? Sign up here on LinkedIn or at iris.deus.ai .