Explore the Future with Gen AI: Your Weekly Passport to Innovation!
Perpetual Block - a Partex Company
Leveraging DeepTech to democratise asset classes
We are back with another exciting edition, ready to dive into the fascinating world of GenAI, and the future-shaping tech trends.
LLMs as Reasoning Agent
Large Language Models (LLMs) have garnered significant attention for their remarkable performance across various natural language understanding tasks. However, their abilities as reasoning agents have come under scrutiny due to a surprising phenomenon known as the "Reversal Curse."
The concept of the "Reversal Curse" in large language models (LLMs), is a phenomenon where these models fail to generalise from "A is B" to "B is A" statements. This failure of generalisation is surprising and highlights a fundamental flaw in the logical deduction capabilities of LLMs.
Example:
For instance, if an LLM is trained on the sentence "Olaf Scholz was the ninth Chancellor of Germany", it will not automatically be able to answer the question "Who was the ninth Chancellor of Germany?" correctly.
This is because the model does not generalise the learned information in the reversed direction. The likelihood of the model providing the correct answer is not higher than for a random name, indicating a lack of understanding of the relationship between the entities involved. According to research published by Microsoft and Allen Institute a few years ago about a strong correlation between the accuracy for a particular number and its frequency in pre-training
To understand this phenomenon an experiment was conducted on real-world knowledge, using a dataset of celebrity questions, where the models are required to identify the reverse relationships between celebrities and their parents. The model's performance on this task is also poor, indicating that the Reversal Curse is not limited to synthetic facts but also extends to real-world knowledge.
Findings:
Proposed Solution:
Future work:
The Reversal Curse raises important questions about the generalisation abilities of LLMs. Despite their impressive performance on many tasks, these models seem to struggle with basic logical deduction. They have deduced that memorization can offset the need to do reasoning as a first principle.
Future research should focus on understanding the underlying mechanisms of the Reversal Curse and exploring ways to mitigate its effects. This could potentially lead to the development of more effective and reliable language models.
LLM for Legal Reasoning
Much like judges follow a structured process to render judgments, there's a machine learning model called Legal Judgment Prediction (LJP) that predicts legal case outcomes. LJP uses case-specific factors like crime type, victim-defendant relationship, and evidence to predict the defendant's guilt or innocence, offering a computational approach akin to how judges make decisions.
However, it's important to recognize that machine learning models have their limitations in the legal context.
In addressing these limitations, Large language models (LLMs) show complex reasoning abilities in legal judgment prediction. In lieu of ML models (LJP), LLMs predicting judgement along with law articles and justification significantly enhance the explainability of models. But simply inferring LLMs (GPT-3 in this case) with zero short prompting is not found to be sufficient. It is important to teach LLMs to generate responses with intermediate reasoning steps with only a few examples.
Solution and Methodology
Legal Syllogism Prompting: Teaching Large Language Models for Legal Judgement Prediction
Let's delve into the evaluation of different prompting techniques with LLM:
Zero-Shot Prompting Without Chain of Thought
Chain-of-Thought (CoT) Prompting
Large Language Model (LoT) Prompting
LoT, when provided with a legal syllogism prompt and without any learning or fine-tuning, surpasses the performance of the other two methods on the CAIL2018 sampled dataset as it allows LLM to access and use Legal syllogism knowledge more effectively.
Following were the advantages achieved:
领英推荐
Limitation of LoT
Open Source LLMs vs. Proprietary LLMs
Folks interested in developing Generative AI tools have two primary choices when deciding what to build upon open-source or private large language models (LLMs).
Considering the pros and cons for both, the choice of LLMs can depend on performance, business needs and priorities:
Source: https://www.dhirubhai.net/pulse/llm-economics-which-cheaper-deploy-open-source-llms-openai-nawaz
ChatGPT: A Technical Journey
ChatGPT has evolved from a research prototype to a versatile and powerful language model with the ability to generate text, translate languages, write different kinds of creative content, and answer your questions in an informative way, all while connected to the internet for real-time access to information.
ChatGPT's technical journey has been marked by the following milestones:
Take a closer look at the infographic highlighting ChatGPT's evolution into a versatile and powerful language model.
Recent Business News
Novo Nordisk partners with Valo Health (privately held U.S. health tech Company), investing $60 million upfront and up to $2.7 billion based on milestones, to harness AI for cardiometabolic disease treatment. This collaboration signifies AI's potential to advance drug development.
Vice Health and Wellness Inc. (Vice) is leveraging AI and nutraceutical technology to develop "Vice Versa AI," an AI-driven platform aimed at providing personalized and sustainable weight loss and obesity management solutions. The AI app aims to revolutionize weight loss by tailoring programs to individual users based on data, enhancing adherence and success rates. The global weight loss and management market is poised for significant growth, making Vice's AI solutions strategically positioned in this expanding sector.
Amazon is investing $4 billion in Anthropic, an AI company, to bolster its healthcare-focused generative AI efforts. Amazon plans to incorporate Anthropic's AI assistant, Claude, into Amazon Bedrock, enhancing drug development and healthcare services. This move escalates competition in the healthcare AI sector, with Microsoft also making substantial investments in the field.
SAP is integrating Joule, an AI system, across its cloud enterprise portfolio to provide contextual insights. This AI technology enhances productivity and business outcomes securely. Joule will be integrated into various SAP applications and platforms, improving user experiences by offering intelligent responses to questions or problems in plain language. It will first be available with SAP SuccessFactors and SAP S/4HANA Cloud. Joule aligns with SAP's broader strategy for an enterprise AI ecosystem.
XYB, a coreless banking platform, is teaming up with Google Cloud to integrate generative AI. This collaboration aims to expedite the development of innovative financial products and streamline processes for banks, non-banks, and fintech firms. XYB's coreless banking platform powered by generative AI will enable hyper-personalized products and foster industry innovation, addressing market gaps and accelerating the provision of banking services.
If you are looking for Generative AI Solutions, check out our offerings at www.perpetualblock.io
XLRI | Business Development | Marketing | Strategy | Key Account Management | HealthTech | Fintech | RetailTech | InsurTech | Entrepreneur | Engineer
11 个月Discover all of our newsletter editions here: https://www.dhirubhai.net/newsletters/generativeai-newsletter-7108431030507745280/