Deploying AI Applications

Deploying AI Applications

Deep learning models are now ubiquitous and analyse massive amounts of audio, image, video, text and graph data, with applications in all industry segments across the value chain. While significant work has been done on the training side to standardise and automate using frameworks and tools, the deployment of these models into production, especially at scale, remains largely bespoke and extremely effort intensive.

There are 3 major components in the journey to have AI powered business applications successfully deployed in an enterprise.

  1. Data Strategy - Includes the data collation and training of the AI model
  2. The AI algorithm itself
  3. The deployment of the AI model into production with live data

There is a fourth and final part as well without which the real benefit of AI will not be fully realised. That is the continuous learning loop of the AI application from live data to improve the accuracy/performance of the model. Ideally, the deployment platform and tools need to enable the prediction stream anomaly detection and provisioning the data for retraining of the model. However, as the scale and complexity of deployed models increase, this may have to be considered and planned as a separate workload in itself. We will address is a separate discussion in the future. For a discussion on the data strategy please see my previous posts on this topic.

Discussions about AI tend to focus on data (rightly so), feature engineering, algorithms and model development. No doubt all necessary for successfully using AI, however, insufficient attention is being paid to the equally important aspect of AI applications’ deployment to production. Managing AI applications in production is in fact so important that you can say the actual work only begins after you’ve deployed the AI in production with live data! Welcome to the emerging field of ML Ops or AI Ops or AI Deployments. Not to be confused with using AI/ML for improving IT Operations processes.

Setting up environments for AI pipelines and deploying them can be a long and exhausting process. Some common hurdles that we face in this process are:

  • Mapping an AI algorithm to right hardware for deployment needs thorough research and deep expertise.
  • Availability of multiple AI libraries/frameworks for similar tasks reduce community sharing and reusability of models.
  • AI Frameworks focus primarily on building models, so testing and deployment strategies needs to planned separately and well in advance.
  • Choice of framework, dataset and tools made during the planning and modelling stage has a direct impact on deployment which cannot always be foreseen.

So how do we tackle these problems and create a reliable, scalable and repeatable process for deploying AI applications into production?

This simplest solution and the one being used by most companies today, is to go with a hosted provider like AWS (there are several others with more customised PaaS offerings like Pipeline.ai or H2O.ai) where a significant part of the deployment process can be abstracted and a fairly consistent user (deployment) experience is available for a large cross section of load vs performance patterns. While there are trade-offs in areas like flexibility and optimisation of the hardware specific to the AI models being used, the cost of running a large enterprise-size load on the cloud may also become a deterrent to adoption of AI within the enterprise.

For the rest, unfortunately, the path is still quite uncharted. There are a few tools from the major players and some from smaller boutique outfits but they all tend to be very specific to a particular set of framework-model-hardware combination. Even while using such tools, a significant number of bespoke customisations maybe required for which a pool of smart engineers need to be readily available. 

None of these problems are unsurmountable and the situation is not as grim as it may sound. The tools landscape is evolving every day and considerable work is being done by the leaders like Google, Facebook, Microsoft etc and a whole bunch of smart start-ups. With awareness of the problem and proper planning at the initial phases (think of a process akin to ‘DevOps for AI’) AI applications can be deployed in production and scaled to meet the growing need for better and more reliable enterprise automation.

Saurabh Vaidya

Delivery Executive at Gainwell

5 年

Nice article Debi

回复

要查看或添加评论,请登录

Debiprasad Banerjee的更多文章

  • Future of AI in 2024

    Future of AI in 2024

    Towards the end of 2023 experts from Stanford's HAI had released a set of predictions on what they expect in AI in…

  • GPT’izing all of AI – Generative AI Models capture our imagination

    GPT’izing all of AI – Generative AI Models capture our imagination

    Everyone is talking about ChatGPT, understandably so, after all, as Arthur C. Clarke said, sufficiently advanced…

    6 条评论
  • The Complete AI Pipeline

    The Complete AI Pipeline

    It is such a delight to be working with domain specialists, tech nerds and data scientists eager to build cutting-edge…

    1 条评论
  • The Key to Successful AI Adoption - Building a Robust AI Pipeline

    The Key to Successful AI Adoption - Building a Robust AI Pipeline

    Artificial Intelligence based systems and applications, especially the Deep learning models with convolutional and…

    1 条评论
  • From Data to Decisions – Journey of an Enterprise

    From Data to Decisions – Journey of an Enterprise

    With so much data available from all parts of the business it is often hard to determine which set of data is useful…

    2 条评论
  • AI Enabled Business: A Data Centric View

    AI Enabled Business: A Data Centric View

    A quick recap – The last discussion was about data driven decision-making being the future of all successful…

    1 条评论
  • The Future Enterprise: Data Driven Decision Making

    The Future Enterprise: Data Driven Decision Making

    “The enterprise of the future will thrive on data driven decision making”. You have probably heard this being said and…

    4 条评论
  • rPa: Who made my process complex?

    rPa: Who made my process complex?

    Some processes are complex to begin with and others get there with time. All processes go through a series of evolution.

  • rPa: Processes keep evolving; bot or not!

    rPa: Processes keep evolving; bot or not!

    As we discussed in the first article of this series, our processes keep evolving over time as we find ways to make them…

  • rPa: To Bot or not to Bot!

    rPa: To Bot or not to Bot!

    While we firmly believe that the key to the RPA success lies in the process and not in the robot, let us spend a little…

社区洞察

其他会员也浏览了