Meet Cerebras Inference – the fastest inference for generative AI! ??? Speed: 1,800 tokens/sec for Llama 3.1-8B and 450 tokens/sec for Llama 3.1-70B, 20x faster than NVIDIA GPU-based hyperscale clouds. ?? Price: Cerebras Inference offers the industry’s best price-performance at 10c per million tokens for Llama 3.1-8B and 60c per million tokens for Llama-3.1 70B. ?? Accuracy: Cerebras Inference uses native 16-bit weights for all models, ensuring the highest accuracy responses. ?? Access: Cerebras Inference is open to everyone today via chat and API access. All powered by our third-generation Wafer Scale Engine (WSE-3). Try it now ?? https://lnkd.in/gEJJ2pfY Press Release: https://lnkd.in/gtF5fxHt Blog: https://lnkd.in/gZ46q4cD
Cerebras Systems
计算机硬件
Sunnyvale,California 36,209 位关注者
AI insights, faster! We're a computer systems company dedicated to accelerating deep learning.
关于我们
Cerebras Systems is a team of pioneering computer architects, computer scientists, deep learning researchers, functional business experts and engineers of all types. We have come together to build a new class of computer to accelerate artificial intelligence work by three orders of magnitude beyond the current state of the art. The CS-2 is the fastest AI computer in existence. It contains a collection of industry firsts, including the Cerebras Wafer Scale Engine (WSE-2). The WSE-2 is the largest chip ever built. It contains 2.6 trillion transistors and covers more than 46,225 square millimeters of silicon. The largest graphics processor on the market has 54 billion transistors and covers 815 square millimeters. In artificial intelligence work, large chips process information more quickly producing answers in less time. As a result, neural networks that in the past took months to train, can now train in minutes on the Cerebras CS-2 powered by the WSE-2. Join us: https://cerebras.net/careers/
- 网站
-
https://www.cerebras.ai
Cerebras Systems的外部链接
- 所属行业
- 计算机硬件
- 规模
- 201-500 人
- 总部
- Sunnyvale,California
- 类型
- 私人持股
- 创立
- 2016
- 领域
- artificial intelligence、deep learning和natural language processing
产品
地点
Cerebras Systems员工
动态
-
Cerebras Inference: highest speed, lowest latency. Try it out!?https://lnkd.in/gEJJ2pfY
-
-
Major performance update: Llama3.1-70B now runs at 560 tokens/s! 24% faster in 3 weeks Available now on Cerebras Inference API and chat https://lnkd.in/dTF-6yGP
-
?? Sean Lie, CTO of Cerebras, talks with The_TechArena's Allyson Klein in a new podcast episode about how Cerebras' wafer-scale chips are revolutionizing the AI landscape. ? Get insights on: ???The intersection of ML and hardware ???How Cerebras Inference is delivering breakthrough performance ???Open-source innovation and customer success ???What’s next for AI ? Catch the full episode here: https://lnkd.in/g5sd5EGG
-
Cerebras Inference perf update: Llama3.1-8B: 1800 ??? 1927 tokens/s Llama3.1-70B: 450 ??? 481 tokens/s Still ?? for the most popular open model in the world! https://lnkd.in/gEJJ2pfY
-
-
?? Excited to share the development of NANDA, a cutting-edge Hindi Large Language Model, created with our partners G42, Inception, Core42, and MBZUAI (Mohamed bin Zayed University of Artificial Intelligence). NANDA was trained on Condor Galaxy, one of the world’s most powerful AI supercomputers, built by G42 and Cerebras. ?? NANDA details: ?? 13-billion parameter model ??? Trained on 2.13 trillion tokens, with a specific focus on Hindi ?? Optimized for Hindi, Hinglish, and English Learn more: https://lnkd.in/dFs3BqVq
-
-
?? Cerebras Systems has signed a Memorandum of understanding (MoU) with Saudi Aramco With Cerebras’ CS-3 systems, Aramco will build, train, and deploy large language models to support local AI initiatives. Dr. Nabil Al Nuaim, Aramco SVP of Digital & Information Technology, stated: “This MoU with Cerebras aims to accelerate our abilities to develop an AI-powered digital innovation economy in Saudi Arabia by helping to support the integration of advanced AI solutions, unlocking new opportunities for the country and localizing cutting-edge technologies with regional expertise.” Learn more here: https://lnkd.in/gBn5EFsG
-
-
Join Cerebras Senior Applied ML Scientist Gurpreet Gosal as he participates in a panel discussion on "Revolutionizing AI: Exploring the Next Frontier of Large Language Models (LLMs)." ?? Tuesday, September 10th ?? 12 PM GMT+4 ?? Register here: https://lnkd.in/gwZi-X3m
-
-
Celebrating our product launch in Times Square NYC! ?? Join the high-speed inference party at inference.cerebras.ai ??
-
This is what high speed inference looks like…
You can generate and iterate on UI instantly with Cerebras Inference ? OP: https://lnkd.in/gVkpzbHz