Artificial Intelligence #26: What could the next decade of AI look like?
It's autumn here in the UK. Not long before, I start teaching the two courses - Digital Twins: Enhancing Model-based Design with AR, VR and MR, and Artificial Intelligence: Cloud and Edge Implementations.
Image source: University of Oxford?
This week, I also shared our vision on #digitaltwins at the MathWorks site Digital Twins and the Evolution of Model-based Design.
Last week, I discussed my misgivings about the idea of data-driven v.s. model-driven AI. This week, we extrapolate that question and ask a much broader question: ?“Will AI breakthroughs this decade be based on last decade’s AI?developments, OR will we see new directions of AI research in this decade?”
Undoubtedly, the last decade 2010 to 2020, was a game-changer for AI research. It was based on the current deep learning models characterised by parallelized networks comprising relatively simple neurons that use non-linear activation by adjusting the strengths of their connections. Using this model, we find the rules for function f(x), which maps domains x to y when the rules are hierarchical and (typically) the data is unstructured. To do this, we need a large number of labeled examples. The labels are at a higher level of abstraction (ex if the image is a cat or a dog). The algorithm can then discern the features that comprise the object(ex a cat has fur, whiskers, etc.). This is the essence of deep learning called representation learning and is common knowledge for data scientists.
Especially in the latter part of the last decade, three developments have accelerated this model:
a)????Transformer based models
b)????Reinforcement learning and
c)????Generative adversarial networks.
And they continue to amaze us daily – for example:
Deepmind - Alphafold – deep mind protein folding; Deepmind - meta-algorithm creating the one algorithm to rule them all, i.e., a deep learning model that can learn how to emulate any algorithm, generating an algorithm-equivalent model that can work with real-world data. And of course, Deepmind’s AI predicts almost precisely when and where it is going to rain, and now megatron Turing NLG from Nvidia and Microsoft – a large language model that claims to exceed GPT-3
Impressive as these are, all the examples above share some crucial properties
?Also, the common element here is: all intelligence (rules) are derived from the data alone.
?The other extreme is: rules are symbolic, i.e., decided by humans.
?That was the early days of AI which ultimately led to the AI winter.
?However, I believe that this decade will be all about techniques that can interject expert intelligence into the algorithm (but is not the same as symbolic as in the early days of AI)
?One such interesting case is from some work by Max Welling. I have referred to this work before, but in this case, I am referring to the use of generative models to build counterfactual worlds. This could work where we have a problem domain with too many exceptions, i.e., there is a very long tail of situations that do not show up in your dataset used to model the problem.
To put the problem in context
According to Max Welling
领英推荐
The above defines the problem, i.e., how to model a world for which you have very little information.
?Some definitions
?
What is proposed
According to Max Welling
The basic idea is to use Generative models to build counterfactual worlds.?
The rationale
How it could work
As I understand it, what is being proposed by Max Welling is:
Relationship to human thinking
?
Analysis
If you want to work with me, see my courses at the #universityofoxford
?
?
CTO at iProtectU
3 年Really interesting. I see AI starting to use a range of techniques that may come from multiple sources (directions may be a better word). One will be the use of models or tools, so whilst you could teach an ANN to add up, it is more simple to teach it to use a calculator. Certain parts of image recognition can be passed to specialist tools. Eventually these tools will be themselves generated by an AI, the output may or may not be a "neuron" architecture. Another directions will be the use of AI to generalise - which has direct links to the above discussion of counter factuals. Whilst I suspect that the generalisation will itself produce benefits, it is likely, in my opinion, that this will be driven by the need to understand decisions coming out of an AI, but an effect of this will facilitate the next stage that I foresee. Which is an ecosystem of partially independent ANNs connected through "generalisation trunks", with the ability to plug-in tools to solve specific part of a problem.
Digital Twin maker: Causality & Data Science --> TwinARC - the "INSIGHT Digital Twin"!
3 年Ajit, a very thoughtful summary on your part . . . I agree 100% with Max Welling Problem in Context section! "However, when you do not have sufficient data available, you will need to use human-knowledge to fill the gaps" - this is where Physics-based models can give a leg up. I know your focus here is not on IoT, we have the opposite problem - we have surfeit of data! When there is lots of data, using models as *initial conditions* is a good idea; knowing full well that all models are approximations. There is a latent feeling that since Physics-based models use partial and non-linear differential equations, etc., they are accurate; but these equations are approximations of reality with drastic simplifying assumptions (which almost never hold in practice). Coming to Counterfactuals, the advantage is that AFTER data are collected, you can do what-if analysis without doing more experiments - this is what got the Nobel prize this year for Imbens and cohorts! Of course, counterfactuals do not exists without knowing causal effect. Sorry to insert my recent work here but the following may illuminate some of your points further in an IoT context. ?? “Causality & Counterfactuals – Role in IoT Digital Twin”; https://www.dhirubhai.net/pulse/causality-counterfactuals-role-iot-digital-twin-dr-pg-madhavan
Data Scientist - Dataworkz
3 年Interesting read again. I'm looking forward to your newsletter every week. I would like to share another possible answer to the main question of this week: “Will AI breakthroughs this decade be based on last decade’s AI?developments, OR will we see new directions of AI research in this decade?” An interesting topic is the research on Spiking Neural Networks. All deep learning solutions that have been developed so far are really amazing but they do have another thing in common. It takes a lot of energy to run all computations due to their continuous nature. For example during training. At the CWI researchers are now looking at the possibilities of Spiking Neural Networks which are much closer to the way our human brain works. At this moment they have built SNNs of which the performance (in terms of accuracy) approaches the performance of the best ANNs while the SNNs are much more energy-efficient. This is promising and can be very useful in for instance always-on devices. https://www.cwi.nl/news/2021/energy-efficient-ai-detects-heart-defects
|Energy, Environment & Sustainability|Biochemical & Bioprocess Engg|CO2 Capturing|LCA|ESG|ISO-9001, 14001, 45001, 27001|QHSE|
3 年Dr. MARYAM ZAFFAR
Digital Marketer & Graphic Designer, YouTube Expert, SEO Professional at Fiverr, Upwork
3 年Good