Machine Learning tribes

Machine Learning tribes

My short summary of the book "The Master Algorithm" by Pedro Domingos.

The Master Algorithm by Pedro Domingos, a Portuguese professor at University of Washington, caught my attention as Bill Gates short-listed it in his must-read book list on AI.

It's not an easy book to read if you are completely new to the Machine Learning field or to Data Science, but is gives an interesting perspective on the different tribes on machine learning, their main algorithms, their strengths and weaknesses and on the relationship between them. The tribes presented are the Symbolists, Connectionists, Evolutionarists, Bayesians and Analogizers. 

Each of this tribes has its own master algorithm and the author defends that each of these are good for some type of problems, but not for others. So the path pointed by the author to reach a single Master Algorithm is to combine the key features of all of them.

Here is a short summary of the tribes presented:

Symbolists

They view learning as the inverse of deduction and take ideas from philosophy, psychology, and logic. The master algorithm for this tribe is inverse deduction.

They believe that intelligence can be reduced to symbols manipulations. Maths is about solving equations by moving symbols around and the same stands for logicians doing deductions. Since elaborating a set of rules for induction is computationally intensive, symbolists prefer decision tree based algorithms.

Connectionists

Connectionists reverse engineer the brain and are inspired by neuroscience and physics. Their master algorithm is backpropagation.

Connectionists are critical of Symbolists as they believe there is a lot more going under the surface than Symbolists can get. They believe this “lot more” can be achieved in parallel processing and not sequence processing done by the Symbolists. For this, they borrowed the concept of neurons from neuroscience, from which each concepts can be represented by neurons that “fire together” – each neuron connects to other via synapses and learning takes place via synaptic connections.

Big data powered the recent popularity of the now called “deep learning” techniques. Bottom line, these techniques yield results that are typically hard to understand and explain. I pointed to this in another article you can find here.

Evolutionaries

Evolutionaries simulate evolution on a computer and draw on genetics and evolutionary biology. The master algorithm for this tribe is derived from genetic programming.

DNA encodes an organism in a sequence of base pairs. Similarly computer programs can also be encoded as strings of bits whose variations are produced by crossover and mutations. A great mystery to be solved in genetic programming has to do with the role of crossover and its helpfulness (mutation makes seems to make the work to achieve fitness by itself). This and some other problems ended up making this tribe less relevant these days.

Bayesians

Bayesians believe that learning is a form of probabilistic inference with its root in statistics. Their master algorithm is Bayesian inference

The basic idea defended by this tribe is their systematic way of updating degrees of belief in light of new data. They agree with Symbolists that prior assumptions are needed even though they don’t agree on the type of prior knowledge allowed – Bayesians defend that knowledge affectes structure and parameters of the model while Symbolist accent anything can be encoded in logic. Na?ve Bayes, Markov Models, Hidden Markov Models, Bayesian Networks are examples of algorithms used by this tribe and developed in the book in good detail to give an overview and relate these algorithms to each other and to others mainly used by other tribes (e.g. Na?ve Bayes and perceptrons).

Analogizers

Analogizers learn by extrapolating from similarity judgments and are influenced by psychology and mathematical optimization. Their master algorithms is Support Vector Machines.

Analogizers use similarities among various data points to categorize them in to distinct classes: We learn by relating the similarity between two concepts and then figure what else one can infer based on the fact that two concepts are similar. Nearest-neighbour algorithms and Support Vector Machines (SVM) are presented in some detail in the chapters related to this tribe. 

要查看或添加评论,请登录

Asif Rajani的更多文章

  • 10 Ideas In Asset Management For 2024

    10 Ideas In Asset Management For 2024

    This is a summary of Oliver Wyman's 10 Asset Management trends for 2024. The original article can be found here.

  • CS: the Archegos case (I)

    CS: the Archegos case (I)

    In this article, I connect insights from the "Three Lines of Defence" section of my book with the challenges seen in…

  • Collateral Allocation and Optimization (II)

    Collateral Allocation and Optimization (II)

    This article is the second part dedicated to Collateral Allocation and Optimization. Here is the first part for your…

  • Collateral Allocation and Optimization (I)

    Collateral Allocation and Optimization (I)

    To illustrate a specific case of collateral allocation, let’s consider an obligor of the bank: the company We Make…

  • Economics and Banking (I)

    Economics and Banking (I)

    In previous articles about inflation and its impact on banking loan losses and profitability, I refer to a very common…

  • An ECL Stress Model (III)

    An ECL Stress Model (III)

    This is the last article on an ECL Stress Model. The first two can be found here and here.

  • An ECL Stress Model (II)

    An ECL Stress Model (II)

    In a previous article we went forescasted transition matrixes conditional on scenarios. In this article, we will use a…

  • An ECL Stress Model (I)

    An ECL Stress Model (I)

    A potential ECL model flow to be used for stress testing is presented in my book. In this article I present a summary…

  • Inflation and Profitability

    Inflation and Profitability

    In my previous articles, I focused on the effect of inflation in the Loan losses, i.e.

    1 条评论
  • Inflation and Loan Losses (II)

    Inflation and Loan Losses (II)

    In the first article about inflation and Loan losses I focused on households. In this article I continue exploring the…

社区洞察

其他会员也浏览了