Building with Open Source LLMs: My Session at Packt Gen AI in Action Conference
Amita Kapoor
Author| AI Expert/Consultant| Generative AI | Keynote Speaker| Educator| Founder @ NePeur | Developing custom AI solutions
Packt has long been at the forefront of technical education, helping developers stay ahead of the curve with cutting-edge technology resources. Their upcoming Gen AI in Action conference continues this tradition, bringing together industry experts, practitioners, and innovators to explore the latest developments in Generative AI.
The conference promises to be a treasure trove of practical knowledge, featuring hands-on sessions, expert insights, and real-world implementation strategies. As someone who has benefited from Packt's technical books and resources throughout my career, I'm particularly honored to be part of this influential event.
It gives me immense pleasure to share that I will be delivering a technical session focused on building applications using Open Source Large Language Models (LLMs). This topic is particularly relevant as organizations increasingly look for cost-effective, customizable AI solutions that they can control and adapt to their specific needs.
In this newsletter edition, I’d like to give you a sneak peek into what I’ll be covering during my session. Whether you’re an AI developer, a data scientist, or simply curious about the potential of open-source Large Language Models (LLMs), there’s something here for you.
Choosing Open Source LLMs
Proprietary LLMs are becoming increasingly attractive, thanks to declining API costs, managed cloud hosting, and frequent updates. With so many options like OpenAI’s GPT-4, Anthropic’s Claude, and Grok at your fingertips, it’s easy to get drawn in. But when does it make good business sense to switch from these proprietary APIs to open-source LLMs?
While cost is a significant factor, it’s not the only one. Ask yourself: How much control do you need over your data? How critical is customization for your application? These considerations can tip the scales toward open-source solutions.
In my upcoming talk, I’ll delve into this pivotal decision-making process. We’ll compare the costs of both proprietary and open-source options and explore how to select the ideal open-source LLM for your needs. We’ll examine factors such as:
Optimizing for a Superior User Experience
You’ve built your open-source LLM-based application—fantastic! But now you’re facing a common challenge: How do you make it fast enough to delight your users?
While there are numerous techniques to optimize latency, in my talk, we’ll focus on some of the most impactful methods that strike the perfect balance between performance and practicality. Specifically, we’ll explore:
While there are many other optimization methods out there, these techniques offer a high return on investment and are accessible whether you’re a solo developer or part of a larger team.
Customizing Your LLM: Prompting Strategies, Fine-Tuning, and RAG
Now that we’ve optimized for performance, let’s focus on making the model truly yours. Customization is the key to aligning the LLM with your specific domain or application’s needs. In this part of the talk, we’ll explore how to tailor your LLM using:
By implementing these strategies, you can transform a general-purpose LLM into a specialized tool that excels in your application’s context.
Evaluating and Maintaining Your Application’s Quality
Once your customized LLM is up and running, evaluating its performance becomes crucial. An effective evaluation strategy ensures that your application not only works but thrives. We’ll discuss:
Wrapping It All Up
The world of open-source LLMs is evolving at a breathtaking pace, offering unprecedented opportunities for innovation. Throughout my session, we’ll journey from the crucial decision of choosing between proprietary and open-source models, through the intricacies of optimization and customization, to the practical realities of maintaining production-quality services.
What I’m most looking forward to is sharing both the technical knowledge and the practical insights gathered from real-world experience. We’ll discuss the common challenges, effective strategies, and lessons learned that can help you navigate this landscape more smoothly and efficiently.
I invite you to join me at the Packt Virtual Conferences Gen AI in Action Conference for what promises to be an engaging, hands-on exploration of building with open-source LLMs. Whether you’re aiming to reduce costs, enhance privacy, or gain more control over your AI applications, this session will equip you with the practical insights you need.
Until then, keep experimenting, keep building, and most importantly, keep pushing the boundaries of what’s possible with AI!
Project Manager | Freelance Brand Developer | Content Creator/Developer
2 周Couldn't wait for kicking this off and very excited! Amita!
Principal Data Engineer | Gen AI | 14x Azure, 5x Databricks, 1x AWS,1x GCP
2 周Please post the recording here if it is allowed
Producer (Tech Books) at Packt
2 周Practical insights from choosing the right models to fine-tuning and leveraging RAG are invaluable. Can’t wait to host you at the conference and dive deeper into these strategies with the audience, Amita! https://www.packtpub.com/conference/put-gen-ai-in-action-2024