Deploy NVIDIA Triton Inference Server with MinIO as Model Store
This tutorial is the latest part of a series where we build an end-to-end stack to perform machine learning inference at the edge. In the previous part of?this tutorial series, we installed the?MinIO?object storage service on SUSE Rancher’s RKE2 Kubernetes distribution. We will extend that use case further by deploying?Nvidia Triton Inference Server?that treats the MinIO tenant as a model store.
Step 1 — Populate the MinIO Model Store with Sample Models
Before deploying the model server, we need to have the model store or repository populated with a few models.
Read the entire article at?The New Stack
Janakiram MSV?is an analyst, advisor, and architect. Follow him on?Twitter,??Facebook?and?LinkedIn.
Staff Engineer - AI/ML Platform | 8+ Years in AI/ML, Cloud Native, Data Engineering | 2x Databricks Certified | Driving Innovation with Generative AI | BITS Pilani
3 年Seems like MinIO is gonna be the defacto option for S3 compatible object storage requirements. I find it very useful for integration testing.