Which AI Model Should You Pick for Your Startup?
a human picking between two robots anime style via Midjourney

Which AI Model Should You Pick for Your Startup?

A product manager today faces a key architectural question with AI : to use a small language model or a large language model?

The pace of innovation in the field clouds the answer. Each day, researchers publish novel findings on performance, discover new techniques to implement, & surface new challenges to wrestle with.

This is my current mental model of when to choose a large or small model :

When to choose a large model :

  • time to ship is critical : many of these models are available via API, requiring formatted data as an index or vector database - which an engineer can achieve within a few hours for a working beta.
  • the company would prefer to rely on external experts to drive innovation within the models.
  • the company has no plan/interest to staff a team to manage AI infrastructure or develop deep machine learning experience / expertise in-house.
  • the product lead would like to minimize career risk by choosing a well-known player.
  • the company believes the relatively high costs using these models will decline with time & scale.

When to choose a small model?

  • the team has or would like to develop intellectual property around machine learning as a competitive advantage or mechanism to increase the value of the business.
  • the company uses proprietary or sensitive data within its models and needs strict controls / guarantees for compliance or legal reasons. The company doesn’t believe sensitive data masking & indexes provide enough security.
  • the product has an edge architecture : models are trained or run on mobile phones or hardware at the edge, away from the data center. The computing limitations of those devices, plus the benefit of running models locally (primarily cost) demand a smaller model.
  • the business would like to minimize vendor lock-in, keeping an agility to switch to another provider
  • the business prefers to manage its AI costs actively by instrumenting code & training built-for-purpose models.

There’s a third option : MLOps businesses offer managed infrastructure with running small-language models, providing simpler management, reduced operating expense, but with the freedom of smaller models.

As the nascent market matures, customers will elect their preferred deployment option. Today, it’s too early to predict which approach will capture the majority of spend & which infrastructure choice suits different use cases best.

We can say though that managed large-language models have a head start, as Microsoft earning showed with its $900m ARR AI business.

Praveen Singh

Building RANESO- Your Trusted Tech Ally! Analytics Solution Architecture | Healthcare

1 年

I'm particularly interested in the Tesla A100 and V100. Their performance and memory capacity would greatly benefit our complex AI models.

回复

Do we assume that small models and large models have the same capabilities? Is there a use case where a large model is the only option? As part of the evaluation, one could also compare the performance of a large model + proprietary context vs. a small model tuned on proprietary data. Small models may lack emergent properties, like commonsense reasoning, and may perform poorly out-of-distribution.?

回复
Tirthankar Das

Advocate,Solicitor,Broker,Networking entrepreneur, over 28000+ Linkedin connections... Unity is strength...

1 年

As an investor and mentor , how would you avoid or save (after such issue has crept up) the founder and startup on this hot soup https://hsfnotes.com/arbitration/2023/01/10/private-equity-firm-prevails-over-indian-promotors-guerrilla-tactics-making-new-law-on-the-arbitrability-of-disputes-in-singapore/ ?

回复
Tirthankar Das

Advocate,Solicitor,Broker,Networking entrepreneur, over 28000+ Linkedin connections... Unity is strength...

1 年

Nice

回复
CHESTER SWANSON SR.

Next Trend Realty LLC./ Har.com/Chester-Swanson/agent_cbswan

1 年

Thanks for Sharing.

要查看或添加评论,请登录

社区洞察

其他会员也浏览了