OpenAI DevDay Predictions
...and what it means for startup founders
On November 6, 2023, almost one year after the launch of ChatGPT, OpenAI was responsible for the single biggest startup massacre in recent memory. Or was it? As a repeat founder and recovering investor, I have spent my career building and evaluating companies. There is some truth to the hyperbole, but a lot of fiction as well. To better understand the aftermath – the winners, losers, and survivors – we need to first unpack what happened.
What has changed?
So what did OpenAI announce at DevDay? A lot of small improvements that will impact user and developer experience but might not significantly change the AI startup landscape: more consistent JSON outputs, model reproducibility for debugging, multimodal inputs for the Completion API, a new text-to-speech API to complement Whisper ASR, etc. They also announced a few large improvements that will fundamentally impact broad categories of companies, my own (stealth) company included:
A very loose summary:
What does this mean?
H1 2023, the trend in the enterprise was AI experimentation. Public companies had board-level mandates to figure out an AI strategy that trickled down as a top priority to the rest of the org. Private companies followed suit. Security and ops teams had to fight a deluge of cool new AI tools. Then, entering the back half of 2023, companies started to realize that good demos don’t necessarily mean good products. Buyer excitement waned. A series of AI-related security vulnerabilities put the power back in the hands of security leaders. Now, companies are focused on AI consolidation. They are more keen to explore building over buying. Although incumbents with distribution have successfully launched AI-first features – Ironclad’s Contract AI, Loom’s summarization, Gong’s spotlights, Outreach’s Smart Email Assist, People.ai's SalesAI etc. – new startups have struggled to get traction. Outside of ChatGPT, there is no LLM-first AI app for the enterprise. Willingness to experiment with consumer apps has also waned as we pass the peak of the current AI hype cycle.
With this backdrop, here are three market/buyer predictions that might be helpful to builders:
The biggest losers
How do startups adapt?
This paints a bleak picture for AI-first startups. But not all hope is lost. For one, GPT-4 Turbo is 2-6x cheaper than GPT-4 and has significantly higher rate limits, which breathes new life into consumer startups with poor unit-economics who are forced to eat the cost at scale (we’re just waiting for similar cost reductions for DALL-E). In addition, developers have more powerful tools to experiment with more functionality.
For RAG-based companies (think: enterprise search companies like Glean), the proprietary implementation done the right way will outperform Assistant API implementations. The most salient features for the ranking step of retrieval tend to be metadata based on user behavior (thumbs up/down, user activity on a subset of documents, time spent on a particular item or document) – known commonly as “personalization”. You can think of this as “contextual awareness” of the problem: e.g., if you are to look up the answer to a question in your internal wiki, the things that might matter to you might be the time the article was last updated, who wrote the article, and the number of times other people viewed the same article versus just the content alone. This level of data ingestion and feature engineering makes all the difference: even if the model abstraction (Assistant API) is very good, the metadata/”hyperparameter optimization” might make the difference between 50-60% accuracy/relevance and 90%+. Companies built on the Assistant API doing out-of-the-box retrieval will look like POCs versus companies that spend significant effort on retrieval and ranking (think: products built on Heroku vs. products built on AWS or GCP). We will explore this more in another blogpost.
Here are three suggestions for startups:
OpenAI is owning more and more of the AI stack. They are eating the infrastructure layer from the bottom-up: from the model to LLM infra/ops to horizontal use-cases. They are eating the application layer from the top-down: first ChatGPT as an end-user application (both consumer and enterprise), and now platforms and marketplaces to build both consumer (custom GPTs) and enterprise (Assistant API, enterprise-only GPTs) applications. The good news is that, unless AGI is on the horizon, even if OpenAI owns the AI stack, it will likely not own the entire software stack. History doesn’t repeat itself, but it rhymes, and if the Salesforce AppExchange is any indicator, the next wave of AI-enabled enterprise products will be end-to-end applications that integrate with the underlying technology (foundation model, CRM, etc.), not be owned by the platform. In the next 5 years, AI will likely be a foundational part of the software infrastructure stack alongside databases and application servers, but you are in trouble if 90% of your infrastructure stack is AI.
Ultimately, OpenAI DevDay is just another hurdle for some startup founders to overcome, and for others, it was a clear win. It’s a much needed reckoning for the most crowded category since web3. And it’s an opportunity for the best founders to separate themselves from the pack. We are excited about the opportunities that OpenAI DevDay opens up, and are excited to continue to build.
Thanks to Dan Roberts, Michael Graczyk, Lauren Reeder, Florian Juengermann, and Zack Lawryk for your feedback.
Experienced Tech Leader (Startups and Big tech, B2B/B2C)
1 年Thanks Daniel Chen for an insightful post and your summary from dev-day. Looking forward to reading more on the topic from you.
Personal Advisor ranking Minister to H.E. Samdech HUN MANET the Prime Minister of Cambodia/ the Secretary of State of the MoJ / Advisor to H.E. Samdech HUN SEN the President of Supreme Privy Council, Senate, and the CPP
1 年Thank for sharing. Very informative.
Entrepreneur | Tech Optimist
1 年Really insightful article! Thanks !
Conversational AI Consultant at Conversation Design Institute | Author of the newsletter Teaching computers how to talk (3K subscribers)
1 年Interesting and thorough summary. I'd love to challenge you on the following point: "Character.ai, Meta AI Characters, Replika, and other customized chatbot platforms. Custom GPTs are a direct competitor." I don't really see how GPTs compete with an app like Replika, who have invested heavily in character building and offer a hyperpersonalized, multi-modal experience. It's much more sophisticated than the product OpenAI put out. I could see competitors who want to build an app similar to Replika attempt using the Assistants API.
CEO & Founder @ Raya Advisory - Offering AI & Product Consulting + Recruiting Services
1 年Terrific reflection and summary Daniel Chen! Especially this point resonated well with me “Products will be expected to be 10% AI 90% good ol’ software, not the other way around.” This is so true! We are still seeing a lot of companies trying to be “AI first” and not knowing what that even means nor can they articulate what real problem they aim to solve.