Can we trust LLMs with translations?
Stefan Huyghe
??Localization VP ? AI Enterprise Strategist ?? LinkedIn B2B Growth ?Globalization Consultant ?? Language Industry Writer ??Content Creator ?? Social Media Evangelist ?? Client Success ?? LocDiscussion Brainparent
LLM Hallucinations: Transcreation Gone Overboard
Large language models (LLM) have made significant advancements in natural language processing, demonstrating remarkable capabilities to generate coherent and contextually relevant responses.
They are trained on massive datasets of text and code, and they can learn to perform many kinds of tasks. They can generate text, answer questions, and even help translate. The problem with automated translation for the longest time was that the models used were not flexible enough. Machine translations would slip up because of their close adherence to rules. LLMs, however, can be creative in a way only humans could in the past. Unfortunately fine-tuning the balance still proves challenging and LLMs can also produce hallucinations.
A hallucination is a false perception that is not caused by any external stimulus. LLM hallucinations can take many forms, such as generating text that is factually incorrect, making up stories that are not based on reality, or even creating images that do not exist.
LLMs face significant challenges when it comes to hallucinations, which can have detrimental effects such as spreading misinformation, compromising confidentiality, and fostering unrealistic expectations about their capabilities. It is crucial to comprehend the nature of hallucinations and approach the information generated by LLMs with a critical mindset to effectively address and mitigate the issues they can pose. Let peel back the onion a bit together.
Recognizing the types of Hallucinations
LLMs have been observed to exhibit hallucinations in various forms.
They can occasionally conflate information from different sources, leading to the generation of text that is inaccurate or misleading. For instance, when summarizing news articles from multiple sources, an LLM might unintentionally combine or misinterpret details, resulting in a distorted representation of the information.
LLMs can also generate text that is factually incorrect on occasion. For instance, if asked about historical events, an LLM might provide inaccurate details or present fictional accounts without comprehending the concept of truth or falsehood. It is crucial to fact-check the information provided by LLMs to ensure its accuracy.
At a fundamental level, LLMs utilize probabilities to arrange words, which generally results in coherent and sensible outputs. However, there are instances when LLMs generate text that lacks logical meaning. For example, if prompted to generate a poem, an LLM might produce grammatically correct but nonsensical verses that do not convey any coherent message.
Due to the probabilistic nature of LLMs and their ability to combine information, there is also a statistical possibility for an LLM to generate fabricated information that inadvertently reveals confidential or sensitive details. This overindulgence in creating text based on probabilities can result in the unintentional disclosure of confidential information.
Why Do They Happen?
So why might LLMs hallucinate? One reason is that they are trained on massive datasets of text that contain errors and biases. This can lead the LLM to learn to generate text that is also incorrect or biased.
Another reason for LLM hallucinations is that they are not always able to distinguish between real and imaginary information. This can lead them to generate text that is based on their own internal models of the world, even if that information is not accurate. While LLMs excel at understanding and generating text, they may struggle with comprehending nuanced or ambiguous queries. In such cases, LLMs may provide responses that appear plausible but are not entirely accurate.
领英推荐
LLMs primarily rely on the information available in their training data and may not have access to real-time or specific domain knowledge. This limitation can lead to hallucinations when LLMs generate responses outside their area of expertise.
Finally, LLMs can also hallucinate because they are still under development. They are not perfect, and they can sometimes make mistakes. This can lead them to generate text that is incorrect or misleading.
How to Avoid Them
To ensure you receive accurate information from Chatbots and mitigate the risks of encountering hallucinations, here are some helpful tips:
How to make it better in the future
One crucial step in optimizing the performance of the model you are using involves tuning it according to your specific requirements. This process encompasses various techniques, including prompt engineering, parameter efficient tuning (PET), and full model tuning, each contributing to reducing hallucinations and improving overall output quality.
Prompt engineering involves carefully crafting or modifying the prompts provided to the model to elicit more accurate and relevant responses. By fine-tuning the prompts, researchers and developers can guide the model to produce desired outputs and minimize the chances of hallucinatory or misleading information.
Parameter efficient tuning (PET) is a technique that focuses on training the model with a limited number of examples, utilizing various methods such as data selection and active learning. PET allows for efficient training by strategically selecting the most informative samples and minimizing the amount of labeled data required. This approach helps enhance the model's performance while reducing hallucinations.
Full model tuning involves training the entire language model with a broader and more diverse dataset. This approach allows the model to learn from a wide range of examples and contexts, leading to improved generalization and a better understanding of language patterns. By exposing the model to a comprehensive training dataset, developers can address specific issues related to hallucinations and ensure more accurate and reliable responses.
By utilizing these tuning techniques effectively, researchers and developers can fine-tune LLMs to align with specific requirements, minimize hallucinations, and enhance the overall quality and reliability of the generated outputs.
The adoption of LLMs is still in its early stages, and a comprehensive assessment of their advantages and disadvantages is yet to be accurately determined. In my view, approaching LLMs with an open mind is the most effective approach to comprehending the various aspects, including hallucinations.
Embrace this transformative journey and explore its potential for the localization industry to the fullest, as such rapid evolution is a rare occurrence Those who wholeheartedly embrace this journey are the ones who stand to gain the most from it.
Independent Translation and Technical Writing Professional
1 年One issue I have not seen discussed yet is the possibility - I will say probability - that in the future there will be instances in which some official translation systems will automatically disallow more accurate but prohibited phrases in favor of more politically acceptable phrases. For example, I expect that mainland China and Hong Kong will implement systems that automatically re-translate 'Tienanmen Square Massacre' as something like 'Tienanmen Square Uprising.' This will start with the more autocratic regimes, but it may not be limited to such regimes.
Head of Content at Smartcat
1 年Stefan Huyghe The information you share about mitigating the risk of erroneous LLM hallucinations is extremely pertinent. We need to be equipped with the knowledge on how these actually happen, where the information was sourced, and how to clarify biases, among the other pointers you shared. Today, everyone and anyone can use GPT, and without this know-how, it's difficult to discern what is factually correct and what is not, what's subjective, and what is not.
Global Human Advocate | Age Diversity Champion, Community Builder, Professional Coach-in-training | ex Pearson Education | LocLunch? San Diego Ambassador
1 年Sure I've heard folks chatting about AI hallucinations... when it comes to using an LLM right out of the box, like in ChatGPT alone, it seems hard to trust. But I was chatting about it today with Gabriel Fairman, who addressed the hallucinations in a recent webinar about Bureau Works' use of AI in BWX tools, and he said "The fact that we have structured prompts in a carefully engineered environment minimizes chances of hallucinations, but it really depends on different people's definition of a hallucination... our AI can still disagree with itself which some people may consider a hallucination"... Stefan Huyghe
interpreter - medical, legal, public and social services - OPI, VRI, RSI - English, Spanish, French, Italian, Brazilian Portuguese - ITI, HIPAA, CIOL DPSI Health level 6
1 年Jorge Davidson, mira que interesante!
Visionary CoCreator of LinkedIn - Founder and CEO, Boswell Translation, LLC | President and Founder, Muslim Aid Network (Non-profit in progress) | 25+ years in Arabic and Spanish translation | US Navy Veteran
1 年Excellent exclamation!!!Let me tell you something about hallucinations.I was walking through the streets of mobile Alabama and wound up inside according to the mission. Mobile County jail. There was a young Black Man...with a Learning diszability...... I wasn't there to steal a wine or take that pearl…. I said something that caused him to say "mMma'aN you're tripin....". I looked at him with my left... And said the following following "Man....we're always tripping!" ~Azraprizm