Public Preview

NIM LLM Text documentation

Overview

NIM provides packaged inference services around NVIDIA’s serving stack.

Inputs and Outputs

Text/JSON; structured decoding supported.

Inference & Decoding

Standard controls with NVIDIA acceleration.

Safety & Compliance

Add guardrails as needed; vendor guidance applies.

Evals & Quality

Depends on chosen base models.

Deployment & Ops

First-class on-prem support.

Pricing & Licensing

Commercial; varies by SKU.

References

See YAML references.