ODDS: Getting Ready for AI.
Patrick Bangert
Chief of AI | Data Science | Artificial Intelligence (AI) | Machine Learning (ML) | Data Analytics | Product Development | Software Engineering | CTO
At a recent discussion session at The Data Standard (https://datastandard.io/), we asked the question “Are you ready for AI?” Josh Odmark and I hosted the session and introduced the topic. Here are the highlights of how to get ready for AI and improve your ODDS for the project giving you a great result.
First, the organizational framework must be set up to create a supporting structure. This starts with support and buy-in from top-level management who are clear on the objectives. A lack of management support will usually mean that the project will either end when the going gets tough or not get deployed in the end. Furthermore, a project needs to be set up with all the usual bells and whistles of a project: A timeline, to do items, stakeholders, budget, and most importantly project management with enough spare time to see it through.
Second, the project should start with due diligence. In this context, that refers to making sure that the current situation is known, the problem is clearly defined, and the form of the solution is clear. When using the word “clear” twice in the prior sentence, I’m speaking as a mathematician who wants a numerical definition that can be objectified. Especially the solution needs to be defined as precisely as possible including objective numerical success criteria that can be assessed at any stage in the project and allow the project to stay on track. An essential part is the desired accuracy of the model. Projects are often either terminated too early (“this will do”) or too late (“let’s see if we can do even better”), wasting resources in either case.
Third, the data must be available. Data science, artificial intelligence, machine learning or what you wish to call it relies on having as much data as possible. Some data may be in your possession. Other data could be generated, acquired, or bought. Some data cannot be obtained at all. What data is even needed, or desired, must be determined at the very start of the project – and this must be done by domain experts and not the data scientists! Sometimes you have less data than you really want and the question is whether this is good enough or not. Unfortunately this cannot be answered until you try to make the model but all stakeholders should be aware that the project is taking a risk.
Fourth, the data must be scrubbed. Most of the time, the term is “clean” but I have used “scrub” here for two reasons. Of course, I want the acronym of this checklist to work nicely but also because this process actually involves more than just cleaning. Before we clean the data, we must ask whether the data is relevant to the question we are asking and whether it is representative of the underlying problem. Often, these two difficult items are not done, which leads to models that do model the situation but do not solve the problem that was posed – note that being relevant and representative are two technical words from statistics that can be made quite precise. Following this, the data must be cleaned, which means outliers and bad data removed, missing values filled in, spikes possibly smoothed over and so on.
Now you are ready to do AI.
Having done AI, you face the toughest task yet: Convincing people to use your model in real life. Change management is the process of getting all the users to change their behavior from whatever they were doing to the new workflow that includes your model and its solution to the problem. This change may involve many features that have nothing to do with AI, or even nothing to do with the problem. Take, for example, industrial predictive maintenance. The problem is that equipment fails catastrophically sometimes, which is why they want an accurate forecast. The change is that maintenance personnel, having acted in a fire-fighting capacity all along, must now learn to plan their work and spare parts orders. This is a significant change in the way the company operates and may require additional software tools to realize – all of which has nothing to do with making an accurate failure forecast.
Want some help with improving your ODDS? Feel free to reach out. Once you are ready to do AI, please consider giving the Brightics AI Accelerator a spin: https://trial.xcelerator.ai/
SPE Technical Director - Data Science and Engineering Analytics | Former Chief Data Officer (CDO) at Shell | Global Data, Digitalization and AI4Energy Leader
4 年Patrick - I think you have touched the key points. Agree that data can be “Achilles heel” for any company’s digitalization aspirations, especially oil and gas majors. Accessing and ensuring data reliability are key blockers, besides the data accountability, data-centric culture, global data standards for scaling solutions and lack of compliance culture. It has to be all worked together with innovative technology for automation and algorithms to succeed. Keen to hear comments from others. Regards, Sushma