SambaNova Systems的封面图片
SambaNova Systems

SambaNova Systems

计算机硬件制造业

Palo Alto,CA 54,995 位关注者

Supercharge AI apps with SambaNova Cloud! Accelerate your AI journey. Unlock lightning-fast inference on Llama 3.2.

关于我们

AI is changing the world and at SambaNova, we believe that you don’t need unlimited resources to take advantage of the most advanced, valuable AI capabilities - capabilities that are helping organizations explore the universe, find cures for cancer, and giving companies access to insights that provide a competitive edge. We deliver the world’s fastest and only complete AI solution for enterprises and governments with world-record inference performance and accuracy. Powered by the SambaNova SN40L Reconfigurable Dataflow Unit (RDU), organizations can build a technology backbone for the next decade of AI innovation with SambaNova Suite. Our fully integrated hardware-software system, DataScale?, enables organizations to train, fine-tune, and deploy the most demanding AI workloads using the largest and most challenging models. Most recently, with the launch of our newest offering, SambaNova Cloud, developers can supercharge AI-powered applications on Llama 3.2 models. SambaNova was founded in 2017 in Palo Alto, California, by a group of industry luminaries, business leaders, and world-class innovators who understand AI. Today, we’ve built an incredibly smart and motivated team dedicated to making a lasting impact on the industry and equipping our customers to thrive in the new era of AI.

网站
https://www.sambanova.ai
所属行业
计算机硬件制造业
规模
201-500 人
总部
Palo Alto,CA
类型
私人持股
创立
2017
领域
High Performance Computing、Artificial Intelligence、Machine Learning、GPT3、Foundation Models、Deep Learning、Computer Vision、True Resolution、3D Image Analysis、Recommendation、AI Platform、Large Language Models、AI for Science和Generative AI

地点

SambaNova Systems员工

动态

  • 查看SambaNova Systems的组织主页

    54,995 位关注者

    Agents are the future—but enterprises hit 3 roadblocks: Security. Speed. Cost. ...We’re fixing all three! Today, we've open-sourced a Deep Research Framework to empower enterprises with: ? Your Data, Your Rules: Build agents on your private data, deploy securely on-prem with SambaNova. ? 3X Faster Than GPUs: Blazing-fast inference via SambaNova Cloud—iterate faster, act quicker. ? Save Millions/Year: Ditch OpenAI’s costs. Use open models like AI at Meta's Llama 3.3 70B for GPT-4o performance at a fraction of the price. How? The framework ships with 3 ready-to-go agents: 1?? General Search Agent (quick answers) 2?? Deep Research Agent (detailed reports) 3?? Financial Analyst (market insights) ?? Watch SambaNova smoke GPUs in our demo. ?? Code your own agent, join the #AIGameOn contest, and win prizes! ?? Try the FREE demo or clone the repo to start building. The future of enterprise AI is open, fast, and yours. #OpenSourceAI Learn more ?? https://lnkd.in/gnWxpTxZ

    • 该图片无替代文字
  • 查看SambaNova Systems的组织主页

    54,995 位关注者

    Speed demon alert! We're out here turning DeepSeek R1 into a Formula 1 race! ?????v ? Now up to 250 tps ? Available for anyone to start using today ? Fastest in the world (faster than Nvidia Blackwell GPU) Thank you Artificial Analysis for always verifying our speeds. ??

    查看Artificial Analysis的组织主页

    10,477 位关注者

    SambaNova has released their new DeepSeek R1 endpoint and is achieving >200 output tokens/s on their first-party RDU chips! We now benchmark over 15 providers of DeepSeek's R1 model and are seeing substantial differentiation between their offerings: ? SambaNova Systems stands out as the fastest (though offers a 16k context window) ? Deep Infra Inc. is offering an endpoint with FP-4 precision at almost half the price of their native FP-8 precision endpoint ? DeepSeek AI remains the cheapest provider but offers significantly slower output speeds than other providers. We have also measured substantial variance in time to first token on DeepSeek's first-party API. This is potentially due to queuing of requests to manage infra load. See below for an analysis of the Output Speed vs. Price of DeepSeek R1 providers we benchmark ??

    • 该图片无替代文字
  • 查看SambaNova Systems的组织主页

    54,995 位关注者

    According to The New York Times, AI data centers are projected to see their power consumption triple by 2028. ???? As the demand for AI technology continues to grow, so does the pressure on our energy infrastructure. To tackle this emerging challenge, our partners at Lawrence Livermore National Laboratory are actively researching innovative solutions that aim to maximize AI performance while ensuring we don’t overload our power grid. This crucial work is paving the way for sustainable AI advancements that balance innovation with environmental responsibility. Read the full article ?? https://nyti.ms/4kRNAI5 #AI #PowerConsumption #EthicalAI

    • 该图片无替代文字
  • 查看SambaNova Systems的组织主页

    54,995 位关注者

    ?? LET'S GO! ?? Thrilled to be named one of Fast?Company's Most?Innovative Companies for 2025, taking the Number?4?? spot in the Computing Category! ?? Our team is driven by a passion for innovation and a?desire to disrupt the status quo. We’ve built the best chip for AI Inference that’s power-efficient and a game changer for solutions customers. We’ve built the best chip for AI Inference that’s power-efficient and a game changer for solutions customers. We're honored to be part of this prestigious list — thank you, Fast Company! #FCMostInnovative #AIInnovation #FutureOfWork Adam Bluestein Brendan Vaughan

  • 查看SambaNova Systems的组织主页

    54,995 位关注者

    We love to see it! ?? The fact that this entire experiment runs in under 2 minutes with SambaNova Cloud powering the inference is *chef's kiss* ?? Excited to see more of what everyone can do with this open-source pipeline ?? #AI #OpenSource #AgenticAI

    查看Daniel Vila Suero的档案

    Building data tools @ Hugging Face ??

    ? Running prompts on your datasets to evaluate models is the new superpower. Excited to share an open pipeline you can run with Hugging Face Inference Providers. Last week, I showed QwQ-32 B's potential for complex classification tasks by comparing it to DeepSeek R1 and Llama3.3-70B. Today, I'm open-sourcing the pipeline and code to run similar experiments on your data. Also, since last week: ?? By adding 3 few-shot examples, QwQ-32B results increased from 90% to 92%, which matches R1 accuracy. ?? Jump in 8 points for the Llama 70B version with the same 3 few-shot examples! (77% now vs. 69% in the zero-shot scenario). And the best of it all: the experiment runs in less than 2 minutes ?? with Inference Endpoints using SambaNova Systems as inference provider. Check the dataset in the first comment; it has the pipeline and code to try this with your own datasets!

    • 该图片无替代文字
    • 该图片无替代文字
  • DeepScaleR—a compact yet powerful 1.5B parameter model fine-tuned from Deepseek-R1-Distilled-Qwen-1.5B using simple reinforcement learning (RL). Developed by Sijun Tan (UC Berkeley) and team, this model smashes benchmarks with a 43.1% Pass@1 accuracy on AIME2024—a +14.3% leap over its base model! ?? Why it matters: Despite its small size, DeepScaleR outperforms OpenAI’s o1-preview, proving that smarter training can unlock huge potential in leaner architectures. Efficiency meets excellence! ?? Shoutout to the team for pushing the boundaries of what’s possible with RL! Dive into the details: https://bit.ly/4kmLiRf

    • 该图片无替代文字

关联主页

相似主页

查看职位

融资