How to Approach Model Optimization for AutoML
Damien Benveniste, PhD
Founder @ TheAiEdge | Building the largest AI professional community | Become an expert with an expert!
Since I started my career in Machine learning, I have worked hard to automate every aspect of my work. If I couldn't develop a fully production-ready machine learning at the click of a button, I was doing something wrong! I find it funny how you can recognize a senior machine learning engineer by how little he works to achieve the same results as a junior one working 10 times as hard!
AutoML has always been a subject dear to my heart, and I wanted to talk today about how we should approach the model optimization problem from an automation standpoint. I want to address the different angles we should consider when building an AutoML pipeline and show you a basic example of a pipeline.
Make sure to watch the video:
The optimization space
The model selection is the component that involves the ML algorithmic components. When we talk about “model selection“, we mean searching for the optimal model for a specific training dataset. If we have features X and a target Y, we would like to learn what the optimal transformation F from the data:
?? = ??(??)
The term “optimal“ implies we have a model performance metric, and the “optimal” model is the one that maximizes that metric.
There are different axes we can consider to optimize our model:
领英推荐
The optimization strategies
Considering the complexity of those different subspaces, it is often impractical to attempt to solve the problem exactly, and we need to find ways to select a suitable model quickly.
Optimizing in sequence
The typical model optimization strategy is to optimize each axe separately in sequence. Modularizing the different optimization problems makes it easier for multiple people/teams to work on different aspects without stepping on each other’s toes.
A typical sequence of steps is as follows:
Joint-optimization
Optimizing in sequence will result in a suboptimal model because we are approximating the search. The feature selection module will select the “best“features in general, and the model paradigm module will determine the best model paradigm based on the features chosen in the previous step. However, for example, finding a model paradigm that would perform better had other features been selected is not impossible.
Therefore, we could consider jointly optimizing different axes together. For example, it is not uncommon to jointly optimize the feature space and the architecture space.
As always, with optimization problems, there is a balance between search accuracy and computational complexity. Each optimization space has a specific dimension. If there are N possible feature sets and M possible architectures, we need to search an N x M overall space to optimize for both axes jointly. However, if we optimize in sequence, the search complexity is only N + M. If, for example, we have 1M possible feature sets and 1M possible architectures, N x M = 1012 and N + M = 2M. This means it would take 500K (= 1012 / 2M) longer to find the exact optimal feature-architecture pair than an approximate one.
Many optimization processes have an iterative implementation, and we can use this to design pseudo-joint-optimization processes. For example, Recursive Feature Elimination (RFE) is a typical feature selection technique where Evolutionary algorithms such as Genetic algorithms (GA) are often used for architecture search. Both methods are iterative and converge slowly to an optimal solution.
We could merge those iterative processes to obtain a pseudo-joint optimization.
The search is still approximated (so fast), but the feature search considers the results from the architecture search and vice-versa.
Watch the video for more information.
Articles You May Have Missed!
Director-Data & Analytics|| Engineering ||Product Development|| Digital Transformation || Ex- Walmart || Ex- Unilever || Ex- Coca -Cola @CDMP @PMP @Togaf
2 个月It's a critical step for scaling. Very insightful! Thanks !
Enterprise Sales Director at Zoho | Enabling Business Success with Scalable CRM & Digital Transformation Solutions
5 个月Absolutely! Automating model optimization is crucial for scaling ML efforts efficiently. Excited to dive into your insights on building robust AutoML pipelines—it's the future of making AI accessible and impactful across industries!
Senior Data Scientist | IBM Certified Data Scientist | AI Researcher | Chief Technology Officer | Deep Learning & Machine Learning Expert | Public Speaker | Help businesses cut off costs up to 50%
5 个月That's a cool journey! Automating ML tasks can be a game-changer. What tools do you use for AutoML? Damien Benveniste
Research Scholar EXPLORING NEW TECHNOLOGIES
5 个月Sir, please make one video on feature extraction in time series raw signal dataset.
Sounds like you've been grinding in Machine Learning. Automate away and let's dive into AutoML pipeline optimization. Damien Benveniste