Sora-ing to New Heights in AI

Sora-ing to New Heights in AI

ML Engineering Highlights:

  • OpenAI unveils new video AI model : OpenAI has unveiled Sora, its new AI text-to-video generation model, which offers 60-second video generations and incredibly realistic and high-quality results. While it's not yet available to the masses, a limited number of creators are getting access as the company continues to test its security defenses. With the potential to simulate everything, Sora's arrival marks a milestone in the tech and media industry and raises questions about the balance of risks and rewards in AI applications.

Credit OpenAI

  • V-JEPA: The next step toward Yann LeCun’s vision of advanced machine intelligence (AMI) : This paper by Meta introduces V-JEPA, a collection of vision models trained solely on video feature prediction without external sources of supervision. The models, trained on 2 million videos, show versatile visual representations that perform well on both motion and appearance-based tasks without adaption of the model's parameters. For example, the largest model obtains high accuracy on Kinetics-400, Something-Something-v2, and ImageNet1K using a frozen backbone.
  • Feeding the hunger for Nvidia GPU access is big business : Lambda , a GPU cloud company powered by 英伟达 GPUs, has raised $320 million at a $1.5 billion valuation to expand its AI cloud business, joining other companies in offering access to Nvidia GPUs. Nvidia's stock has tripled, and the company has surpassed 亚马逊 and 谷歌 in market capitalization, while AI startups are desperate for access to Nvidia's high-performance computing H100 GPU.

Research Highlights:

  • Transformers Can Achieve Length Generalization But Not Robustly : This paper by Google DeepMind examines the challenge of length generalization for language models, even with large-scale Transformers, using the task of addition of two integers. The success of length generalization is tied to the data format and position encoding, and with the right combination, standard Transformers can extrapolate to a sequence length 2.5 times the input length. However, length generalization is still fragile and influenced by factors like random weight initialization and training data order, leading to large variances across different random seeds.

Credit: Google DeepMind

Lightning AI Studio Highlights:

  • Run Ollama LLMs on a cloud GPU Lightning Studio : This Studio automatically starts a server running an optimized version of Mistral AI ’s Mixtral in the background (courtesy of Ollama ). Ollama allows us to get up and running with large language models quickly. With it we can run Mixtral, Llama 2, and Code Llama, and many other models.

Don’t Miss the Submission Deadline

  • ECCV 2024: European Conference on Computer Vision 2024 Submission Deadline: Fri Mar 08 2024 06:59:00 GMT-0500
  • MICCAI 2024: International Conference on Medical Image Computing and Assisted Intervention Submission Deadline: Fri Mar 08 2024 02:59:59 GMT-0500
  • ECAI 2024 : European Conference on Artificial Intelligence 2024 Submission Deadline: Fri Apr 26 2024 07:59:59 GMT-0400

Want to learn more from Lightning AI? “Subscribe” to make sure you don’t miss the latest flashes of inspiration, news, tutorials, educational courses, and other AI-driven resources from around the industry. Thanks for reading!

BYTE HUB

"Hello Engineers and Investors! ?? Exciting times ahead! ?? I'm spearheading a groundbreaking project that blends innovation and impact. Your expertise is the missing link. Let's revolutionize together! ???? #TechAi

9 个月

Imagine an AI that sees, hears, smells, and learns with you. That's GLBIE AI TECH . I'm seeking collaborators like you with expertise in investing/funding to shape its future. Your insights and knowledge can help us achieve groundbreaking applications. As a partner, you'll gain decision-making power, and co-ownership. and actively contribute. Intrigued? Let's chat! I'm available next week. Sincerely, Ijaz Shah

回复

要查看或添加评论,请登录

社区洞察

其他会员也浏览了