Reka AI releases Reka Core: Understands images, videos, and audio
David Cronshaw
Sr. Product Manager @Disney Streaming | Co-Founder Chatmosa chatmosa.bsky.social | AI, Generative AI | Revenue Generation | Former Microsoft and T-Mobile | Co-Founder UltimateTV.com - Zap2it.com
Reka AI, a San Francisco-based AI startup founded by researchers from DeepMind, Google and Meta, is introducing a new multimodal language model called Reka Core.
"Reka is a frontier-class multimodal language model on par with leading models in the industry today. Core was efficiently trained from scratch on thousands of GPUs over a period of a few months." - Reka AI
Available by API, on-premise, or on-device deployment options, Core is the third member in Reka’s family of language models and offers the ability to understand multiple modalities, including image, audio and video, while offering a massive context window, exceptional reasoning skills, and even coding.
Reka Core is one of only two commercially available comprehensive multimodal solutions.
You can test out Reka Core in the Reka Playground.
Even though Reka was trained in less than a year, it matches or beats the performance of top models from leading players in the AI space, including OpenAI, Google and Anthropic.
"Core is comparable to GPT-4V on MMMU, outperforms Claude-3 Opus on our multimodal human evaluation conducted by an independent third party, and surpasses Gemini Ultra on video tasks. On language tasks, Core is competitive with other frontier models on well-established benchmarks." - Reka AI
The table below summarizes a comparison of Core with leading models in the market today.
Reka AI has 3 models: Reka Core, Flash, and Edge. All 3 of their models are trained to handle and analyze multimodal inputs.
Reka Core Capabilities
Reka Model Showcase:
Reka displayed some impressive results for image and data analysis on their Model Showcase Page:
Reka Core Video:
Reka Core has a lot of capabilities, and one of them is understanding video --- let’s see what Core thinks of the @3body trailer.
领英推荐
Reka tested its Reka Core multimodal language model on Netflix’s “3 Body Problem” and it was able to translate what’s happening onscreen into text. Credit: Reka
Reka Core Use-Cases
Some users cases of Reka’s models include:
Reka Core Tech Report:
State-of-the-Art Performance:
Model Details:
Training Data and Architecture:
Training and Infrastructure:
Evaluation and Benchmarks:
User and Developer Accessibility:
Ongoing Development and Future Prospects:
You can try Reka Core at https://chat.reka.ai/
#ai #rekaai #rekacore #aivideo #aiaudio #multimodality