AI in Tech: Applications, Challenges, and Opportunities | FutureX Seminar
With the introduction of ChatGPT, the IT industry has taken a significant leap towards a new paradigm. This has led to a proliferation of downstream applications, resulting in a flourishing array of AI-based solutions. The Chinese and American tech giants, along with research organizations globally, have joined the large model arms race, marking the beginning of the AI 2.0 era. In early March, FutureX Capital organized an investment seminar titled "Seizing Opportunities in the AI Wave." The event witnessed the participation of several industry heavyweights and founders of portfolio companies. The discussion primarily revolved around two core topics - What caused the ChatGPT craze, and what is the difference between AI 2.0 and AI 1.0? The past months saw the most exciting period in the history of large models. Tech giants, startups, and tech organizations made daily updates on the progress of the large model industry. Notable developments include the release of multimodal GPT-4, Microsoft 365 Copilot, and Google's launch of the Bard beta version. These advancements in the large model industry suggest that the pace of AI 2.0 may far exceed previous technological waves.
Below are the key takeaways from the seminar:
Large Models: The Operating System of the AI 2.0 Era
In the PC and Internet era, operating systems acted as universal systems, managing and connecting software and hardware resources, enabling efficient operation of diverse applications and facilitating various inputs and outputs. Today, large models (such as GPT-4, LaMDA, and Ernie Bots) possess universal underlying capabilities, such as understanding, reasoning, and generation. As large models mature, they provide a wide range of functions for new and existing applications, intelligently handling various tasks and beginning to play a role similar to operating systems.
Data: The Flywheel Effect
As a new generation of operating systems, the data flywheel effect of large AI models will be more pronounced. As shown in the figure above, under the current model, user/preprocessed data will accumulate within the model, driving performance improvements and further enhancing user experience and quantity. A positive feedback loop forms between user data and model performance. (Data privacy and security issues arising from this are also among the barriers for industry customers to accept large models.)
Applications: Imagination is all you need
Large models, with their versatility, comprehension, and multimodal capabilities, bring countless possibilities to applications. Rapid response from mobile internet giants and start-ups in the last generation has enabled large models to quickly penetrate various scenarios, such as search (e.g., New Bing, Bard), social (e.g., SnapChat), e-commerce (e.g., Shopify), creativity (e.g., Adobe), and office (Slack), etc. Start-ups are also pushing the envelope in various scenarios such as social, chat, live streaming, and intelligent customer service, dramatically improving experience and efficiency through AI-native applications.
Open-source models and closed-source model APIs enable start-ups and existing businesses (including many FutureX portfolio) to quickly integrate large models (mostly GPT series) into their applications, achieving functional and efficiency improvements. The main methods currently used include:
领英推荐
The current methods allow businesses to adapt to AI 2.0 era based on existing general-purpose large models, but are still not the optimal solution for vertical industry applications. In the future, there may be lightweight solutions optimized for specific industries (based on general-purpose large models).
Large models not only bring improvements in AI capabilities but also provide new modes of interaction. In the AI 2.0 era, natural language becomes the standard interface for interaction, and the form and application boundaries of intelligent hardware will change. Natural language interaction makes it possible for more diverse intelligent terminals to change their positioning, shifting from toys and auxiliary devices to having certain core functions. Glasses, headphones, speakers, and smart home appliances are expected to see significant improvements in functionality and importance. Application software will also transition from automating simple functions to intelligently completing complex tasks.
Infrastructure: Continuous Demand for Computing Resources
The application and ecosystem of large models are still in their early stages, while the demand for underlying infrastructure presents relatively certain opportunities. The huge demand for computing resources in the training of large models will drive the long-term development of cloud computing, data centers, and underlying computing and networking hardware markets.
Improvements in the efficiency of underlying hardware and reductions in price will greatly promote the training and promotion of large models. On the one hand, this drives the industry to continuously invest in more efficient and powerful GPUs and HBM products, and on the other hand, it also drives industry innovation, including optimized architectures for deep learning training, more innovative semiconductor processes (such as compute-storage integration), and more efficient transmission hardware (such as silicon photonics). The domestic market ban has brought enormous opportunities for domestic GPU, high-performance storage, and high-performance transmission manufacturers.
Challenges Faced by Enterprise-level Applications
Compared to the rapid response and embrace of the AI 2.0 era by the internet industry, traditional industries (especially those with high technological strength, such as finance and high-end manufacturing) are closely monitoring the development of AI, but there are still many challenges in the implementation of large models in enterprise-level applications:
Our take on Large Models
Universal large models are the core systems of AI 2.0 and are a battleground for major tech giants. There will be several dominant players, including not only international tech giants but also China's own large models. At the same time, a rich application layer will bring enormous opportunities to various industries by implementing large models, enabling more intelligent solutions. There will also be continuous opportunities in the infrastructure layer, including hardware, cloud computing, energy, and more, providing ongoing support and promotion for the development of AI 2.0.
actively investing