The excitement around DeepSeek's progress, with the power of 'thinking more', reasoning, and chain-of-thought, show's we've now reached the next chapter of AI – the 'age of inference.' Sid Sheth lays out the case in his thoughtful post, which he & the d-Matrix Team saw coming years ago –?so they have the right hardware architecture, at the right time, for what comes next.
Think more v. Train more This makes intuitive sense and something we have used as a guiding principle from the early days of d-Matrix. A person with multiple advanced degrees is not necessarily smarter than the high-school diploma holder who acquires domain specific knowledge, thinks and works hard to apply this knowledge for their professional outcome. With the arrival of Reasoning and Test-time-compute, we are at an inflection point in the AI computing journey. Finally, revenue generation from AI models is aligning with the cost of AI compute. The more you think (compute) the better the outcomes (results, productivity). The DeepSeek AI reasoning models (Reasoning R1, R1-Zero) make that case in point, using innovative techniques like reinforcement learning with Chains-of-Thought (CoT) and knowledge distillation. Imagine if you could distill all your thoughts into some of the critical life decisions you make, capture the stream of thinking that helped you arrive at those decisions, and then transfer those to your kids. Wouldn’t that be cool! Distillation leads to smaller models and smaller models that are trained to think more, can match the best outcomes in their chosen areas. And the sweetener, these models are openly available for everyone to try and use. DeepSeek AI has likely catalyzed the age of Inference. At d-Matrix, we have been big proponents of at-scale inference and open-source from the early years. We built the company around creating a foundational inference compute solution with open standards as our guiding principle. As we move into the 'inference era’ with multiple forces coming together - reasoning, test-time-compute, open weight models, distillation, quantized numerics, energy efficient architectures, interactive use-cases and ROI - we hope to make our mark in this new world with our newly launched Corsair platform which emphasizes ‘Do More with Less’. More Inference w/ Less Cost($), Less Energy(W), Less Time(s) The platform emphasizes, Disaggregated hardware w/ open tool software stack = Adoptability and cost efficiency -?Less Cost($) Digital In-memory Compute = Energy efficiency -??Less Energy(W) Memory-Compute Integration w/ chiplets = Low latency batched inference -Less Time(s) All this, to make Gen AI commercially viable for everyone. Would love to speak with you, if we can help you go on this journey. #Reasoning #R1 #Inference #AI #DIMC #chiplets #ROI