?? The End of Lazy LLMs

?? The End of Lazy LLMs

In this issue:

  1. Active Learning for passive LLMs
  2. 2 heads, 1 task
  3. CRAG yourself before you wreck yourself


Want to market your brand? I’ve been personally using passionfroot for months now and as much as 40% of my partnerships can be accounted to their platform. They make it easy for companies to find fitting creators for their brand and I’ve found their streamlined collaboration process to be more efficient and more enjoyable for both sides.

Become a Sponsor


1. Efficient Exploration for LLMs

Watching: Double TS (paper)

What problem does it solve? Enhancing large language models (LLMs) requires copious amounts of feedback—a costly and time-consuming process. To streamline this, researchers continuously strive to optimize the way such feedback is collected. The crux of the problem lies in efficiently generating queries that elicit the most informative feedback to improve models with the least possible number of queries, a challenge that's critical for refining these models while conserving resources.

How does it solve the problem? Double Thompson sampling, in conjunction with an epistemic neural network—an artificial neural network capable of uncertainty estimation—offers a solution. This technique operates by generating queries based on a model of uncertainty within the learning agent itself. By sampling from this distribution of uncertainty, the agent can craft more informative queries. The combination of an epistemic neural network for uncertainty estimation and a strategic exploration scheme enables efficient collection of high-quality human feedback, thereby reducing the number of needed queries while maintaining or even enhancing the performance of LLMs.

What's next? This research sets the stage for further development of efficient data collection methods and more sophisticated exploration schemes, ensuring faster and cheaper refinements to LLMs. The exciting question now is how the incorporation of such exploration schemes will generalize across different model architectures, tasks, and domains. The ability to enhance model performance with fewer queries opens up possibilities for more intelligent, adaptive models that can self-improve with minimal human intervention.


2. Two Heads Are Better Than One: Integrating Knowledge from Knowledge Graphs and Large Language Models for Entity Alignment

Watching: LLMEA (paper)

What problem does it solve? Entity alignment is crucial for building extensive and interconnected Knowledge Graphs (KGs), which in turn are foundational for tasks in semantic web and artificial intelligence. The challenge in entity alignment lies in identifying correspondences across different KGs, considering the heterogeneous information they harbor—including structural, relational, and attributive data. Currently, embeddings generated to represent entities' multi-faceted information are difficult to match due to their dissimilarity and the lack of effective mechanisms to leverage these embeddings. Moreover, the potential of leveraging the nuanced semantic understanding of Large Language Models (LLMs) for this purpose has remained untapped.

How does it solve the problem? The proposed Large Language Model-enhanced Entity Alignment framework (LLMEA) addresses the shortcomings of previous entity alignment methods by harnessing the semantic knowledge inherent in LLMs. LLMEA operates by: 1) identifying candidate alignments based on embedding similarities and edit distances, and 2) using LLMs' inference abilities to iteratively process multi-choice questions that pinpoint the final aligned entity. This approach marries structural knowledge from KGs with the rich semantic understanding from LLMs, thereby creating a more nuanced and accurate alignment. By interpreting candidate entities through an LLM and using its robust inference capabilities, the framework can dynamically utilize the context and implicit knowledge that LLMs have been trained on.

What’s next? The results from the public datasets where LLMEA outperforms leading baseline models are promising, but it will be interesting to see how this framework performs at a larger scale, or with KGs that have sparse or noisy information. Future developments might include refinements in candidate selection processes or additional integration of LLMs for ongoing knowledge discovery and alignment tasks.


3. Corrective Retrieval Augmented Generation

Watching: CRAG (paper)

What problem does it solve? Large language models (LLMs) sometimes make errors reminiscent of hallucinations where the text generated does not correspond to reality or lacks accuracy, a problem not always solved by the model's inherent knowledge. While retrieval-augmented generation (RAG) strategies help to offer contextually relevant content by pulling from external databases, RAG's effectiveness is compromised if the quality of these external sources is poor. The primary concern this paper addresses is how to enhance the reliability of information generated by LLMs, especially when the RAG approach brings in suboptimal or irrelevant information.

How does it solve the problem? The proposed method, Corrective Retrieval Augmented Generation (CRAG), incorporates a novel retrieval evaluator that gauges the quality and relevance of documents fetched by the RAG system. Based on the confidence score provided by the evaluator, CRAG can adapt its retrieval strategies accordingly—thereby improving the quality of the generation. Furthermore, by incorporating large-scale web searches, CRAG expands beyond static databases to enhance its source material further. It also employs a decompose-then-recompose algorithm to sift through the fetched documents, emphasizing crucial details and discarding irrelevant content. Its "plug-and-play" nature also ensures that CRAG can be used alongside a multitude of existing RAG-based systems.

What's next? Moving forward, it would be fascinating to see how CRAG's introduction to various contexts impacts the adaptation of LLMs in more complex and nuanced environments, such as legal or technical domains where accuracy is paramount. Testing CRAG's effectiveness in these areas could set new standards for information retrieval in LLMs and possibly pave the way for even more advanced systems that combine the judgment and robust reasoning of AI with the ever-expanding knowledge of the internet. Evaluation in real-world scenarios will be crucial to understanding the limits and possibilities of CRAG-enhanced language models.


Papers of the Week:

?? Marc Policani (PMP, SAFe)

Sr Management Consultant | Unleashing Organizational Potential through Integrated PMO, Portfolio, and Program Excellence with a Touch of AI Innovation

10 个月

Between ChatGPT, Bard, and CLaude - Bard is the laziest by far. 3/4 of the time, it will give me tips and tricks on how to perform the tasks I asked it to execute, rather than performing the tasks I asked it to execute.

??Hakim Elakhrass

post-deployment data science | OSS | co-founder @ nannyML

10 个月

i hope not, i enjoy threatening them to get a proper response ??

要查看或添加评论,请登录

Pascal Biese的更多文章

  • ?? Actually Open AI: A Free o1 Alternative

    ?? Actually Open AI: A Free o1 Alternative

    In this issue: An open o1-like model The LLM Engineer Handbook NVIDIA mixing attention with state spaces Upgrade now 1.…

    4 条评论
  • ?? The Future of Designing AI Agents

    ?? The Future of Designing AI Agents

    In this issue: Towards efficient graph foundation models From LLMOps to AgentOps A Text-to-SQL dataset that breaks LLMs…

    3 条评论
  • ?? HTML > Plain Text for RAG

    ?? HTML > Plain Text for RAG

    In this issue: A foundation model for time series forecasting More intelligent tool usage for scientific LLMs HTML >…

    7 条评论
  • ?? All You Need to Know About Small Language Models

    ?? All You Need to Know About Small Language Models

    In this issue: A survey on SLMs A way towards more brain-like inference How to better count the r’s in strawberry…

    6 条评论
  • ?? Is AI Capable of Reflection?

    ?? Is AI Capable of Reflection?

    In this issue: Testing the reflection abilities of LLMs AI for generating new and diverse scientific ideas One LLM…

    6 条评论
  • ??? GraphRAG Evolves into StructRAG

    ??? GraphRAG Evolves into StructRAG

    In this issue: GraphRAG → StructRAG Agent thinking, fast and slow A universal knowledge graph foundation model…

    4 条评论
  • ?? Fixing AI's Energy Consumption

    ?? Fixing AI's Energy Consumption

    In this issue: Reducing energy consumption by up to 95% Google’s way forward for RAG RNNs might’ve been all we needed…

    9 条评论
  • ?? Chasing o1: Closing the Reasoning Gap

    ?? Chasing o1: Closing the Reasoning Gap

    In this issue: Closing LLMs’ reasoning gaps IBM’s data preparation framework Meta attempting to redefine RLHF Financial…

    4 条评论
  • ?? LLMs Are Improving Themselves

    ?? LLMs Are Improving Themselves

    In this issue: Self-correcting LLMs The 4 levels of RAG and beyond 9B model beating GPT-4o in RAG MLOps/GenAI World is…

    2 条评论
  • ?? A New Neural Architecture (Again)

    ?? A New Neural Architecture (Again)

    In this issue: The return of a controversial neural network architecture NVIDIA releasing open frontier models…

    4 条评论

社区洞察

其他会员也浏览了