AI's Curse of Dimensionality

AI's Curse of Dimensionality

As artificial intelligence continues to evolve at a breakneck pace, it faces an unexpected challenge: the curse of dimensionality. This phenomenon, first coined by mathematician Richard E. Bellman, refers to the exponential increase in complexity and computational requirements as the number of dimensions (or features) in a dataset grows.

The Double-edged Sword of Data Diversity

In an expanding and growing universe, with increased access to computing power and devices, it was natural for the data sizes to expand. As technology became more industrialized and advanced, data also became more accessible and broad-based. While this wealth of information theoretically provides AI systems with more nuanced insights, it also introduces significant challenges:

  • Sparsity: As dimensions increase, data becomes increasingly sparse, making it difficult for algorithms to find meaningful patterns.
  • Computational Complexity: Processing high-dimensional data requires exponentially more computational power and time.
  • Overfitting: With more features, models are prone to fitting noise rather than true underlying patterns, leading to poor generalization.
  • Distance Metrics Breakdown: In high-dimensional spaces, traditional distance measures become less meaningful, affecting clustering and similarity-based algorithms.

In the early days of Generative AI, these aspects were heavily responsible for the hallucinations and other risks which came out. This allowed the developers to find alternative frameworks to resolve the issue.

Technological sophistication is both a blessing and a curse for AI

Our quest to establish more visibility across our processes and reach a 'single version of truth', compelled us to add more and more dimensions to the data. These additional dimensions were originally intended to support AI algorithms to tackle complex problems but it ultimately ended up contributing to the curse of dimensionality.

For example, in genomic research, the researchers usually work with datasets containing thousands of features (genes) but relatively few data samples. This high-dimensional nature of genomic data makes it challenging to identify meaningful patterns and relationships between genetic variations and disease outcomes. As a result, many early attempts at developing predictive models for complex diseases based on genetic data yielded disappointing results due to overfitting and poor generalization.

Similarly, in the finance sector, quantitative analysts often work with high-dimensional datasets containing numerous economic indicators, market variables, and company-specific metrics. The curse of dimensionality can lead to overfitting in predictive models for stock prices or risk assessment, resulting in poor performance when applied to real-world market conditions. This challenge has led to the development of sophisticated feature selection and dimensionality reduction techniques specifically tailored for financial data.

As a first hand example, I was recently asked to assess the AI capabilities embedded within a market leading product on Accounts Receivables automation. This was a very typical workflow, automation and orchestration platform woven together through APIs and deriving value by mirroring high-dimensional data sets from multiple ERP solutions. So, simply speaking if there are hundred dimensions in one ERP, and there are four ERPs in total, this platform will be compelled to onboard at least two hundred and fifty dimensions to cover common and unique dimensions. To top it all, they informed their customer pursuits, that we have AI into areas-where the sheer complexity of executing AI will lead to failure. I was helping out a client make some realistic assessment. To my surprise, when I spoke to the product team from this vendor, they referred to within dimension probability calculation as AI, not across the dimensions. To illustrate, their definition of AI included ability to predict which particular cash receipt has to be matched to which particular open invoice, which is within the same dimension. But when I asked, can you make predictions as to which particular invoice will have a failed match to receipt, they said its not possible since it depends on so many dimensions.

So what are the options to combat the curse

Thankfully, there are some sophisticated techniques available, to conscious development and product teams. I am listing two of them (more will come on the topic later) here for reference.

Feature Engineering

Feature engineering involves creating new features or transforming existing ones to capture the most relevant information while reducing dimensionality. For those still new to AI and ML world, a feature is a crucial concept that refers to an individual measurable property or characteristic of the phenomenon being observed or analyzed. Features serve as the input variables for machine learning algorithms, providing the essential information that models use to learn patterns, make predictions, or classify data. The key techniques pf Feature Engineering include:

  • Principal Component Analysis (PCA): identifies the most important dimensions in the data. To simplify, let's take an example of a dataset which has data about 100 apples, and for each apple we have two features weight (in grams) and diameter (in cm), our goal in principal component analysis will be to reduce the dimensionality of this data from two features to one feature, so that AI algorithms which have to make use of this data set do not face processing challenges.
  • Autoencoders: are neural networks that learn compressed representations of high-dimensional data. A classic example of autoencoding technique is the modern day Active Noise Cancellation (ANC) headphones. How it works is the mic in the headphones pick up both the desired audio (e.g. music) and ambient noise. The autoencoder's encoder part analyzes this mixed audio signal and compresses it into a lower-dimensional representation. This representation captures the essential features of the audio. The decoder then attempts to reconstruct the original audio from this compressed representation. During the reconstruction process, the autoencoder is trained to prioritize the desired audio components over the background noise. It learns to generate an output that closely matches the clean audio signal. A similar process is followed in order to filter out noise from data during an autoencoding process in machine learning algorithms.
  • Domain-specific Feature Extraction: leverages expert knowledge to create meaningful, low-dimensional features. A very simple example of Domain-specific Feature Extraction is the email classification algorithm which helps prioritize the emails in order of importance.

Fine-Tuning

Fine-tuning involves optimizing AI models to perform well on specific tasks with high-dimensional data:

  • Transfer Learning: involves putting pre-training models on large datasets and fine-tuning on specific tasks reduces the need for extensive high-dimensional training data. Imagine you've trained a model to recognize different breeds of dogs. Using transfer learning, you can adapt this model to recognize cat breeds much more quickly than starting from scratch. The model already understands basic features like edges, shapes, and textures, which are useful for both tasks
  • Regularization Techniques: include methods like L1/L2 regularization and dropout help prevent overfitting in high-dimensional spaces. Regularization helps prevent overfitting in machine learning models. Let's take a bag packing example for L1/L2 regularization. Think of these as budget constraints for your model. L1 (Lasso) encourages the model to be selective about which features it uses, like choosing only the most important items when packing a small suitcase. L2 (Ridge) encourages the model to use all features but in moderation, like distributing weight evenly in a backpack.
  • Attention Mechanisms: allow models to focus on the most relevant features, effectively reducing the impact of curse of dimensionality. For example, in a language translation service, when translating a sentence, an attention mechanism allows the model to focus on different words in the source sentence as it generates each word in the target language. It's like how a human translator might look back at specific parts of the original text while translating.

Conclusion

As we stand at the frontier of artificial intelligence, the curse of dimensionality looms like a digital Everest, challenging our most advanced algorithms and systems. But there is no option to either ignore it completely or get bogged down by it. There is a lot of work already done by developers and AI researchers, in the space, and more techniques including some on the hardware front as well, are coming up to address the challenge of high-dimensionality.

Assuming that data dimensions will continue to increase, the key lies in not making the same mistake of projecting within dimension AI as across dimension AI, like the accounts receivables solution company referred above. A balanced approach involves taking care of required dimensionality and a realism around how easy or difficult it would be to utilize these additional dimensions will be required.

-Mohit Sharma

This piece is written without prejudice towards any individual or company. Any sources referenced have been directly attributed and are owned by the respective third-parties. The insights I share are based on my own personal experiences on the journey I have been fortunate to live. -Mohit Sharma 
        

要查看或添加评论,请登录

Mohit Sharma CGMA的更多文章

  • The Jobs to be Done Automagically

    The Jobs to be Done Automagically

    In the recently concluded NVIDIA's India AI Summit, at Mumbai, a popular Indian actor asked NVIDIA CEO Jensen Huang…

  • Math Vs. Magic

    Math Vs. Magic

    In May 2024, while releasing one of the latest updates to his company's flagship model, Sam Altman tweeted "Feels like…

    2 条评论
  • If AI is the agent, who's the Principal?

    If AI is the agent, who's the Principal?

    In the rapidly evolving landscape of artificial intelligence, a critical question emerges: if AI systems are acting as…

    3 条评论
  • Generative AI or Predictive AI ?

    Generative AI or Predictive AI ?

    Panacea or hyperbole..

    2 条评论
  • The incentive to bend truth

    The incentive to bend truth

    Welcome to Beyond Human- the AI Story, where we explore the new reality where AI and humans co-exist. The history of…

    1 条评论
  • GCC isn't Gen AI capability center

    GCC isn't Gen AI capability center

    India is home to nearly 50% of the GCCs of the world, with a market size expected to reach USD 110 billion by 2030. It…

    1 条评论
  • Immersive: the new buzzword and the use-cases it can support

    Immersive: the new buzzword and the use-cases it can support

    In 2016, on a business trip to Paris, I noticed there was a lot of buzz on the street around Pokemon. An immersive game…

    3 条评论
  • The Augmentable CFO

    The Augmentable CFO

    When penning this piece, a Google search for the phrase 'Changing role of CFO' produced 35.6 million results in 0.

    1 条评论
  • Enterprise Revival strategies for a Post-Covid World

    Enterprise Revival strategies for a Post-Covid World

    There were numerous 2020 vision and Strategies, across multiple enterprises of all scales and size - with the hope that…

    9 条评论
  • Selecting the most suitable Process mining Solution

    Selecting the most suitable Process mining Solution

    Gartner defines Process mining as " designed to discover, monitor and improve real processes) by extracting knowledge…