Deploy NVIDIA Triton Inference Server with MinIO as Model Store

Deploy NVIDIA Triton Inference Server with MinIO as Model Store

This tutorial is the latest part of a series where we build an end-to-end stack to perform machine learning inference at the edge. In the previous part of?this tutorial series, we installed the?MinIO?object storage service on SUSE Rancher’s RKE2 Kubernetes distribution. We will extend that use case further by deploying?Nvidia Triton Inference Server?that treats the MinIO tenant as a model store.

Step 1 — Populate the MinIO Model Store with Sample Models

Before deploying the model server, we need to have the model store or repository populated with a few models.

Read the entire article at?The New Stack

Janakiram MSV?is an analyst, advisor, and architect. Follow him on?Twitter,??Facebook?and?LinkedIn.

Varun Kruthiventi

Staff Engineer - AI/ML Platform | 8+ Years in AI/ML, Cloud Native, Data Engineering | 2x Databricks Certified | Driving Innovation with Generative AI | BITS Pilani

3 年

Seems like MinIO is gonna be the defacto option for S3 compatible object storage requirements. I find it very useful for integration testing.

要查看或添加评论,请登录

Janakiram MSV的更多文章

社区洞察

其他会员也浏览了