Trends in AI — September 2024
The AI landscape is buzzing with developments, from massive funding rounds to strategic acquisitions and groundbreaking model releases. Join us for an overview of the latest news in AI R&D and a curated list of the month's top 10 trending research papers.
> Join us for Transformers at Work, live from the Bay Area on Friday, September 20th! <
Transformers are the backbone of modern AI, and we're here to celebrate recent progress. In the fifth edition of our workshop, we are exploring recent breakthroughs focusing on LLM applications in the enterprise: Transformers at Work!
In this program, we bridge the space between cutting-edge research in AI and building practical applications that deliver value. Learn about the progress in AI from world-renowned researchers and engineers pushing the boundaries of Neural Search, RAG, LLMOps, Prompt Optimization, Agents, and AI Hardware.
Sign up here, space is limited: https://zeta-alpha.com/events/transformers-at-work-2024
News Articles
Model Releases
Trending AI papers for September 2024
[1] ColPali: Efficient Document Retrieval with Vision Language Models - M. Faysse et al. (Illuin Tech, CentraleSupélec) - 27 June 2024
→ ColPali: a document retrieval model that uses Vision-Language Models to understand complex and visually rich document formats.
?? Why? It radically simplifies the document indexing pipeline and shows great performance on visual question-answering tasks involving e.g. figures and tables.
?? Key Findings:
[2] RouterRetriever: Exploring the Benefits of Routing over Multiple Expert Embedding Models - H. Lee et al. (KAST AI, AI2) - 04 September 2024
→ RouterRetriever: a retrieval system comprising multiple domain-specific embedding models that uses a routing mechanism to select the best expert for each query.
?? Why??It addresses the limitations of models trained on single, static, large-scale general-domain datasets.
?? Key Findings:
[3] Scaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model Parameters - C. Snell et al. (Google DeepMind) - 06 August 2024
→ Suggests that improving the efficiency of test-time compute scaling can improve performance on hard prompts.
?? Why??We can enhance the effectiveness of LLMs without increases in model size or pre-training effort.
?? Key Findings:
[4] Automated Design of Agentic Systems - S. Hu et al. (Vector Institute) - 15 August 2024
→ Proposes the Automated Design of Agentic Systems framework that uses a meta agent to automatically generate building blocks for agentic systems.
?? Why??The framework can reduce the effort required in designing complex agentic systems, potentially leading to more efficient, robust, and innovative solutions than manually engineered ones.
?? Key Findings:
[5] FlashAttention-3: Fast and Accurate Attention with Asynchrony and Low-precision - J. Shah et al.?(Colfax, NVIDIA, Together AI) - 11 July 2024
→ FlashAttention-3 proposes an attention mechanism that leverages asynchrony and FP8-precision for enhanced speed & accuracy on NVIDIA Hopper GPUs.
?? Why??Asynchrony allows computational tasks to overlap and reduces idle times, while low-precision computations ensure faster processing without significant loss of accuracy.
领英推荐
?? Key Findings:
[6] The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery - C. Lu et al. (Sakana AI, FLAIR, Vector Institute) - 12 August 2024
→ The AI Scientist: an end-to-end framework for fully automated scientific discovery using LLMs. It generates research ideas, performs experiments, analyzes results, and writes scientific papers autonomously.
?? Why??The goal is to accelerate scientific progress. By automating the entire research process, this framework can help overcome human limitations related to time, expertise, and biases.
?? Key Findings:
[7] De novo design of high-affinity protein binders with AlphaProteo - V. Zambaldi et al. (Google DeepMind) - 05 September 2024
→ AlphaProteo: a computational deep-learning-based system capable of designing high-affinity protein binders de novo without requiring extensive rounds of experimental optimization.
?? Why??Traditional methods for producing such binders are labor-intensive. AlphaProteo could be a transformative tool in drug development, diagnostics, and biomedical research.
?? Key Findings:
[8] OLMoE: Open Mixture-of-Experts Language Models - N. Muennighoff et al. (Contextual AI, AI2) - 03 September 2024
→ OLMoE: an open mixture-of-experts (MoE) language model. OLMoE-1B-7B has 7 billion parameters but only activates 1.3 billion parameters per token.
?? Why??It aspires to democratize access to high-performing language models, with insights for the community on optimizing MoE architectures.
?? Key Findings:
[9] Diffusion Models Are Real-Time Game Engines - D. Valevski et al. (Google) - 27 August 2024
→ GameNGen: a neural model-based game engine that is capable of executing complex interactive video games. It can run DOOM at over 20 FPS on a single TPU.
?? Why??A major shift from traditional game engines with handcrafted code and predefined rules, to a model where game worlds are generated with neural networks.
?? Key Findings:
[10] Sapiens: Foundation for Human Vision Models - R. Khirodkar et al. (Meta) - 22 August 2024
→ Sapiens: a family of models for human-centric vision tasks: 2D pose estimation, body-part segmentation, depth prediction, and surface normal estimation.
?? Why??It addresses the challenge of creating robust and generalizable vision models that can perform well in diverse in-the-wild conditions.
?? Key Findings:
And a few runner-ups:
You can find an annotated collection of these papers (+ more that didn't make the cut) in Zeta Alpha, allowing you to easily discover relevant literature and dive deeper into any topic that interests you.
Here is a 3-minute preview of the papers in our top-10 list:
The full recording of our latest Trends in AI episode is available on our YouTube, covering all of the papers in depth. Sign up to join us live for the next edition in October.
Until then, enjoy discovery!