The difference in pricing suggests cost savings for enterprises, at least for usage of open models.
Generative AI model and repositories provider Hugging Face this week launched an alternative to Nvidia’s NIM (Nvidia Inference Microservices).
Hugging Face Generative AI Services, or HUGS, is the only available alternative to NIM presently.
NIM, which was first introduced in March and later rolled out in June as part of Nvidia’s AI Enterprise suite, was a first-of-its kind tool to help enterprises deploy generative AI foundational models across any cloud or data center by packing optimized inference engines, APIs, and support for custom or generic AI models into containers as microservices.
NIM caught the attention of developers as it was itself an alternative to the likes of vLLM, TensorRT LLM, and LMDeploy — all of which are frameworks and packages that help deploy foundational models for inferencing, but are arguably time-consuming to configure and run.
…