Infrastructure is critical for running AI models effectively and efficiently, and we had an excellent pair talking about the subject during the #LuxAISummit: Vipul Ved Prakash of Together AI and Clem Delangue ?? of Hugging Face in conversation with Brandon Duderstadt of Nomic AI. Here are some highlighted quotes:
Vipul at Together: “We can create synthetic data with more entropy, with all these different synthetic data methods. So I do see that being another exciting area, which is you don't have yield problems there. You can run experiments really quickly, and we'll see more and more of that.”
Vipul at Together: “I think we need more power and more data centers. That's pretty clear. Like right now, it's way [too] difficult to find anything above 15 megawatts in North America. All these data centers have already been reserved and will not provide enough capacity for building and serving models. So I think we need more power. I also think the GPU power envelope is quite off.”
Hugging Face’s Clem: “I also think that we can do some things to make AI more energy efficient today. I think this movement of only using and focusing a lot of our efforts on large generalist models is a mistake in many aspects. You don't need to take a private jet to go to work. In a similar way when you're doing like a specialized, customized use case, as I mentioned, you don't need a model that is going to tell you about the meaning of life. You can actually use a smaller model that is going to take less energy to train, take less energy to run. The world is a bit like biased right now, and a lot of the investment goes towards large, very energy-intensive models and directions, I think as a field, we can take a different direction and focus on specialized, customized, smaller models that give us a more credible path to continuing to build AI capabilities without ruining the planet.”
Hugging Face’s Clem: “Usually what we see is that companies start with like using a large, generalist model behind an API, because it's easier, it's in a way safer, because that's what others are using. And after a few months, especially when it's production with users, and you start to see more users and the cost is starting to pile up, they think, ‘Okay, can we build different systems where we have more control, where we can optimize the models to run cheaper, faster, more focused on our specific use case, specific constraints?’ And that's usually when they when they start experimenting with other approaches, with taking open-source-based models and fine-tuning them, optimizing them, training them, and I think ultimately it's going to pay off for them, because it's a learning curve, right? It takes more time, it takes more investment. But at the end of the day, if you want to be an AI company, you have to be able to build AI, right?”