Primitives for AI

Primitives for AI

I want to explore more structural thoughts in this article. Let’s say I want to build a financial market prediction model based on news flow. The current way to do this is:

  • Collect lot of news articles
  • Collect financial market movement data
  • Use machine learning models (black-box or otherwise) to predict markets from news

This is not a scalable model of operation for building AI/ML solutions. This will require every problem to be solved from first principles. This approach requires that any ML solution use millions of data points to solve problems. In a way companies are doing exactly this when they build autonomous car driving solutions from pixel data or streaming data from sensors. They call this end-to-end learning. Humans don’t learn this way. And this approach will not be robust. For example a small change in rules of the game will require millions of training sets again.

Psychologists believe that humans are intelligent because of a combination of nature (genetic evolution over time) and nurture (learning over a life time). The above approach to AI puts all the burden on nurture without any pre-built circuitry or parameters. I believe this will not solve problems robustly. Instead I propose on a more collaborative approach of working with “primitives”. For example, going back to financial market prediction from news, the solution can be built using different layers.

I’ve segregated the solution as three layers and each layer delivered as a different platform. I’ve chosen the word platform with a specific intent in mind, which will hopefully get clear by the end of this article. All platforms can be built by the same entity (let’s say G-Square), but it will not make sense. A platform like Google will have access to much larger textual data and can build vector representations more efficiently. These representations can evolve continuously over time as language itself evolves, just like human genes evolve and form the basis of further nurture. The vector representations themselves can be used for a variety of purposes and not just financial markets. It should be noted that predicting markets is just an example I’m using. The solution here may not actually work. In fact predicting markets is somewhere in between incredibly difficult and impossible.

Vector representations of words could be just one of the several primitives necessary for building general AI. I'm not arguing that vector representation is key for learning language, but observing that if vector representation is to be used it should be segregated from other layers of language understanding. The other primitives I can think of are edge detection in images, object identification etc. But unlike vector representations there is no clear cut/well defined schema for representing other times of knowledge. Overtime there could be some standardisation. These primitives will become building blocks for all other AI solutions. As humans we first learn to detect objects, then recognize/classify, then we learn behaviour of objects, only then do we start learning how to interact with the objects (like driving a car). To put all of these steps in a single solution is stupid.

Which brings me to the next idea. The primitives for AI will have to reside on the cloud to be able to be accessed by any subsequent algorithms. Thus the nature of AI will be very different from human intelligence. Humans have a very individualistic intelligence. Although human intelligence improves over time by learning from predecessors, much of learning happens during a life time and ends with that. There is no shared ML algorithm across all humanity, at least there isn’t an explicit one. The AI I envisage will be a connected AI residing on the cloud and not on individual machines or even platforms. In other words future AI will not be embodied. That is the reason I prefer to call the primitive provider a platform. A collection of cloud based AI primitives can be kludged together by engineers to form applications. There could be a limited portion of learning (especially the last mile) which resides on a single machine or a single platform. The power of cloud is in the fact that several platforms can collaborate, compete and evolve much faster than how biological entities evolved.


Gopi Krishna Suvanam

Entrepreneur | Author | AI & Decentralization proponent | Alumnus of IIT-M & IIM-A

7 年

This article is heavily inspired by talks and articles of Gary Marcus, although the thoughts expressed here may not all be inline with his.

要查看或添加评论,请登录

Gopi Krishna Suvanam的更多文章

  • The Hidden Risks of Over-investment in AI Infrastructure

    The Hidden Risks of Over-investment in AI Infrastructure

    The $500 billion investment in AI Infra definitely looks exciting. But.

    2 条评论
  • The Good, the Bad and the Ugly of ChatGPT

    The Good, the Bad and the Ugly of ChatGPT

    To the uninitiated, ChatGPT is the viral product of the organization OpenAI. It is an AI base chat-bot that interacts…

    2 条评论
  • Transfer Learning: The Future of AI

    Transfer Learning: The Future of AI

    End to end learning is paradigm in machine learning where an algorithm takes inputs and desired actions, to learn a…

    2 条评论
  • Limitations of End-to-End Learning

    Limitations of End-to-End Learning

    (Originally posted as an answer to a Quora question here: https://www.quora.

    1 条评论
  • Value per person is more important

    Value per person is more important

    There has been a mad rush to become the next big thing from eCommerce to social media to SaaS. In this rush, startups…

    1 条评论
  • Neat Learning vs Deep Learning

    Neat Learning vs Deep Learning

    AI world used to have two camps: neats and scruffies. The neats wanted AI solutions to be, well, 'neat'.

    1 条评论
  • Online Learning (by Machines)

    Online Learning (by Machines)

    Most ML methods are applied in a batch process mode. So for example in a simple ML model let’s say we want to learn…

  • Machine Learning on Small Datasets

    Machine Learning on Small Datasets

    You may be wondering what the image above the title has to do with machine learning. We will get there by the end of…

    4 条评论
  • Can financial markets be predicted? - Brief note for beginners

    Can financial markets be predicted? - Brief note for beginners

    There are lot of concepts at play here. Let me highlight a few.

  • Machine Learning on External Memory

    Machine Learning on External Memory

    Once Python fails to load data or R is not able to perform analysis because of running out of memory, many people think…

社区洞察

其他会员也浏览了