Personal GenAI in Your Pocket: LLMs and RAG on mobile device?
Pawel Sobczak
VP Partnerships | ?? ex-IBM VP EMEA | AI strategic advisor | Empowering AI builders to boost productivity | Trustworthy AI for Business | Startups | ISVs
The future of work is undeniably mobile, and soon, a personal generative AI assistant, with access to your personal data might be your pocket companion. I recently discussed the opportunity in AI hardware and the immense computing power AI demands. With current GPU supply chain constraints, any additional processing power becomes a valuable asset.
While mobile devices can't match the raw power of GPUs and enterprise GenAI platforms, the sheer volume of processing power from billions of smartphones is undeniable. There are a staggering 6 billion smartphones globally, with Apple iPhones leading at 1.4 billion. This translates to a massive, untapped pool of processing power waiting to be harnessed. It becomes even more exciting when we consider the growing popularity of powerful mobile devices like M3-equipped iPads and Macs – the go-to choice for many business users.
Despite releasing their own open-source LLM in Q4 2023, Apple might seem to be behind in the LLM and GenAI race. However, recent news from March suggests collaboration with Google to bring their Gemini set of generative AI models to iPhones. This could be a game-changer, especially for on-device processing of multi-modal AI (text, vision, audio) while maintaining personal data privacy.
Imagine a future where generative AI is readily available on your mobile device, seamlessly leveraging your personal data to power RAG (Retrieval Augmented Generation) functions. No longer confined to the cloud, these AI assistants will be right in your pocket! While running LLMs directly on smartphones is still challenging, advancements in very small language models are happening rapidly. Mobile phone processors are also becoming increasingly powerful, paving the way for on-device GenAI capabilities. Even if the next iPhone offers just an improved Siri with GPT-like capabilities, the arrival of RAGs on mobile devices is likely just around the corner.
Here's why on-device RAGs have the potential to revolutionize how we work:
领英推荐
For business users, on-device RAGs hold immense potential:
On-device processing requires smaller, more efficient language models. This is where advancements in mobile-specific language models come into play. By combining the power of on-device processing with the efficiency of smaller LLMs, business users will be able to enjoy the benefits of RAGs without sacrificing significant performance.
The future holds intelligent assistance that respects privacy boundaries and empowers business users to be more productive on the go.
Are you ready to unleash the power of RAG in your pocket? When available, would you use it in your business?
On-Device Generative AI @Picovoice — HIRING
5 个月I agree! This article sums up my thinking behind creating picoLLM. What we've learned is: [1] cross-platform support is really really hard. It is easy to create a runtime to support iOS. But once you want to add Android, Web, Linux, macOS, or Windows, ... then it becomes a nightmare. Let alone CPU, GPU, NPU, ... support [2] Co-creating a compression algorithm along with the inference engine is key to retaining performance and runtime speed.
Product Management. Banking. Demand generation & Customer engagement.
7 个月Can you please elaborate on the key features and applications of this Personal Generative AI for smartphones?