Post-Deepseek World
Deepseek has reset priors of the tech community at large, and opened a much larger application game. Here is a mix of business and technical articles that are pointing in the same overall direction - models (and implicitly agents) will go 'deep' in order to drive up accuracy and value. R1 (and 100 others soon) will make it much easier to do this. The machine is cranking, and we should have seriously powerful agents/models soon.
-Venkata
Subscribe to this Domain-specific Agents and Knowledge Newsletter
Previous newsletter on Knowledge Agents and Economics
1. DeepSeek R1 and the Resurgence of Domain-Specific Models
Author: Jon Turow
Explores the implications of the new AI model DeepSeek R1, highlighting how it could shift the landscape from reliance on massive compute power to domain-specific expertise as a key driver of AI innovation.
Well argued piece. Worth your time.
TL;DR
2. Clouded Judgement 1.3.25 - Domain Specific Models
Author: Jamin Ball
2025 is poised to spotlight domain-specific AI models, moving away from generalized systems. These specialized models, trained on unique datasets, promise to transform industries requiring deep expertise and contextual understanding.
I generally read all of Jamin Ball 's substack posts. It is worth subscribing to.
TL;DR
3. HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs
Authors: Junying Chen, Zhenyang Cai, Ke Ji, Xidong Wang, Wanlong Liu, Rongsheng Wang, Jianye Hou, Benyou Wang
The article discusses the introduction of HuatuoGPT-o1, a medical language model designed for complex reasoning in healthcare. By addressing the limitations of current methods, this model aims to enhance medical decision-making through a novel framework for verifiable medical reasoning.
Deepseek is not the only innovation. Ceative and good solutions are coming out of China in volume. This requires only a training set of 40K (read: small but good datasets go a long way)
领英推荐
TL;DR
4. SciCode: A Research Coding Benchmark Curated by Scientists
Author: Minyang Tian et al.
This article discusses SciCode, a coding benchmark specifically designed for evaluating language models (LMs) in generating code to solve real scientific research problems. Developed with insights from scientific experts across various fields, the benchmark consists of 338 subproblems derived from 80 main challenges, aimed at assessing the capabilities of AI in practical scientific applications.
LLMs struggle to understand specialized codebases. This is saying, wait for 1 year. LLMs will write fluid dynamics and other codes.
TL;DR
5. The Knowledge Revival
Author: Robert Pondiscio
This article argues for a knowledge-rich curriculum as essential for cognitive development in education, countering the notion that skills can be learned in isolation. It emphasizes that deep thinking is fundamentally linked to domain-specific knowledge rather than generic skills.
This is really about surviving and building careers in the AI era by developing deep thinking. What struck me is how thinking and domain-knowledge are interlinked.
6. KAG: A Better Alternative to RAG for Domain-Specific Knowledge Applications
Author: Sarayavalasaravikiran
The article explores the limitations of Retrieval-Augmented Generation (RAG) in specialized fields and introduces Knowledge Augmented Generation (KAG) as a superior alternative that integrates structured knowledge systems.
No surprise but found the technical discussion interesting and useful guide if you are going to implement something of your own
TL;DR
Board member, Automation, Digital transformation, Customer success
1 个月Interesting
Cybersecurity, Artificial Intelligence, Cyber-Physical Systems, Space Systems
1 个月Venkata, this is a fantastic summary. Thanks! The field is moving at lightning speed and it is hard to keep up with everything. Such summaries are immensely helpful.
CTPO @ AGNEXT | AI Expert | PhD in Computer Science | IIT KGP | Ex - Capillary |Quality Food for Billions
1 个月Very helpful!