Here's what's new today in the AI Insights
Martin Khristi
AI & Machine Learning Advocate | BI & Data Specialist at CA Karrierepartner | Microsoft Fabric Enthusiast | Python for Data, AI & Time Series Forecasting | Supporter and Contributor in PandasAI
Google Enhances Gemini with Custom Chatbots and Image Generation
Google has rolled out exciting updates to its Gemini platform, introducing custom chatbots called Gems and improved image generation powered by Imagen 3.
Details:
In addition to these new features, Google plans to reintroduce image generation for select Gemini users in the coming days. This feature was temporarily suspended in February due to instances of historically inaccurate and potentially misleading images.
2) Llama is approaching 350M downloads on Hugging Face.
Open source AI is the way forward and today we're sharing a snapshot of how that's going with the adoption and use of Llama models.Read
the full update here ?? https://go.fb.me/mfc5ki??
Highlights? Llama is approaching 350M downloads on Hugging Face.?
largest cloud service providers have seen Llama token usage more than double since May.?
Llama models are being adopted across the industry with great examples from Accenture, AT&T, DoorDash, Goldman Sachs, Infosys, KPMG, Niantic, Inc., Nomura, Shopify, Spotify and Zoom as just a handful of strong examples.Open source AI is how we ensure that the benefits of AI extend to everyone, and Llama is leading the way.
3) what is snowflake cortex ?
Use AI in Seconds with Snowflake Cortex
Generative AI is unlocking new ways to drive innovation, improve productivity and derive more value from data. For organizations to fully capitalize on this potential, it’s critical that everyone — not just those with AI expertise — is able to access and use generative AI.
That’s why we Snowflake creatred Cortex (in private preview), Snowflake’s new, intelligent, fully managed service that enables organizations to quickly analyze data and build AI applications — all within Snowflake. As part of Snowflake Cortex, users of all skill sets now have access to industry-leading AI models, LLMs and vector search functionality, as well as complete LLM-powered experiences. These innovations enable all Snowflake users to securely tap into the power of generative AI and unlock dynamic insights with their enterprise data — regardless of their technical expertise.
领英推荐
Generative AI is unlocking new ways to drive innovation, improve productivity and derive more value from data. For organizations to fully capitalize on this potential, it’s critical that everyone — not just those with AI expertise — is able to access and use generative AI. That’s why we created Snowflake Cortex (in private preview), Snowflake’s new, intelligent, fully managed service that enables organizations to quickly analyze data and build AI applications — all within Snowflake. As part of Snowflake Cortex, users of all skill sets now have access to industry-leading AI models, LLMs and vector search functionality, as well as complete LLM-powered experiences. These innovations enable all Snowflake users to securely tap into the power of generative AI and unlock dynamic insights with their enterprise data — regardless of their technical expertise.
Serverless functions in Snowflake Cortex
With Snowflake Cortex, Snowflake users now have access to a set of serverless functions that easily accelerate everyday analytics and AI app development. With just a single line of SQL or Python, analysts can instantly access specialized ML and LLM models tuned for specific tasks. They can also leverage more general purpose models for prompt engineering and in-context learning. Since these are fully hosted and managed by Snowflake Cortex, users always have access to them without the need to bring up and manage expensive GPU infrastructure. They can also use and leverage Snowflake’s unified governance framework to seamlessly secure and manage access to their data. These functions include the ones listed below.?
Cost-effective LLM-based models that are great for working with unstructured data:
4) Meet Cerebras Inference – the fastest inference for generative AI!
??? Speed: 1,800 tokens/sec for Llama 3.1-8B and 450 tokens/sec for Llama 3.1-70B, 20x faster than NVIDIA GPU-based hyperscale clouds.
Price: Cerebras Inference offers the industry’s best price-performance at 10c per million tokens for Llama 3.1-8B and 60c per million tokens for Llama-3.1 70B.
'
?? Accuracy: Cerebras Inference uses native 16-bit weights for all models, ensuring the highest accuracy responses.
?? Access: Cerebras Inference is open to everyone today via chat and API access.All powered by our third-generation Wafer Scale Engine
(WSE-3).Try it now ?? https://lnkd.in/gEJJ2pfYPress
press Release: https://lnkd.in/gtF5fxHtBlog:
that's wrap up for today!