From ABCs to AI: Unravelling the Magic of Large Language Models
Jonathan Freedman
CCISO | CISSP | CCSP | CISM | CGEIT | CEH | CIPP/E | CIPM | ISO27001 Provisional Implementer | ISO42001 Provisional Implementer | Azure AI Engineer Associate | Head of Technology & Security at Howard Kennedy
Welcome back to my weekly cyber security blog and I hope you all had a wonderful week.?It is undeniable that AI is fast becoming an integral part of our society. As the technology continues to develop and improve, we are starting to see the emerging signs of the changes this will bring to our economy and society. While Artificial Intelligence and Machine Learning have existed as scientific fields of study for decades, ChatGPT has made generative AI a household name. It is this type of technology that organisations are adding to their products and business processes at an accelerating rate. Last week I wrote about the new speech capability of Large Language Models (LLMs), and how they had gone from generating text and conversing with us on-screen to talking to humans verbally in real time. However, to help ensure the safe use of AI, it is helpful to understand more about it. Over the next few weeks my friends, we will explore this from both the technology and cyber security perspectives, beginning this week with the basics of how the technology works and how it is trained. Next week, we will look at prompting and cyber threats, and then the possible future direction of the tech. This series of posts will ask, how well LLMs reasoning can be understood by humans, and as the use of the technology expands into new fields, how much can we trust them?
From Alphabet to Algorithms
I think of the origin of LLMs like a pyramid.
领英推荐
Making the magic box
There is a three stage training process to create an LLM capable of understand and responding to language in a useful and sophisticated way.
So far we’ve covered the basics of what an LLM is and how intelligent systems are trained. We must keep in mind what makes AI different, it’s trained rather than programmed like traditional software. This is also why an LLM can be described as a magic box, we can see what goes in and what comes out, but these models are incredible complex, so it is not always possible for us to understand why they produced certain output. Also as the training data for LLMs came from the internet, it contains unknown biases.
The challenge of controlling and fully understanding LLMs (LLM transparency) are very active areas of AI research. However, this unpredictability also gives rise to different forms of cyber attacks against an LLM which I will cover next week. As researchers continue to improve model transparency, we as end users must consider the risk when using them and adding them into our businesses. I recently started using Google Gemini and was impressed with the inbuilt function to run a Google search on the output of the model and automatically highlight its more “questionable” output. This is a great example of maintaining human oversight over the models output and not blindly trusting them. As the benefits of these models become clearer, along with the risks involved, it is incumbent on us to make sure we consider both.
I believe in our cyber security community and that by sharing and helping each other that we can all be safer. So, everything above is just my opinion, what’s yours? Please share in the comments below and stay safe.
Trusted sales/ commercial leader facilitating increased sales revenue | Localisation industry expert | Professional sales process optimisation | Business Development | Proven sales leader |
9 个月An interesting insight, Jonathan Freedman, thank you for creating it in a concise way. I'm pleased to see the reference to human oversight of these models, or at least the output. Humans-in-the-loop is very important in my line of work, which includes utilizing LLMs for translating some client content into other languages. As good as AI is, leveraging it with professional human linguists that do understand the art of nuance, style and context, including esoteric terminology, is critical for us to achieve the desired quality #translation output for our clients. However, as you mentioned, these LLM models are incredibly complex, and I can't help but think that human oversight is so much more critical than in my world when it comes to much larger data sets with less confined parameters, given we can potentially find ourselves with output that has been produced with no knowledge as to how/why, other than within the model itself. Also, I am curious as to what constitutes a 'reward' for an LLM and how it recognises that reward. Looking forward to the next instalment.
Thank you Jonathan Freedman for a good reminder of the foundations for all of the noise. It's easy to get sucked into the vortex of nonsense around AI and forget that humans built this tech (which is an impressive achievement) and with applied common sense we can control its evolution, application and security. No doubt abuse looms large - that's how we beasts function in all walks of life so AI will be no exception. Look forward to rest of this series. ??