To setup prerequisities and quickstart deployment of Intel® AI for Enterprise Inference on a single node follow steps here Quick Start Guide
- Review System Requirements
| Deployment Type | Description |
|---|---|
| Single Node | Quick start for testing or lightweight workloads (Guide) |
| Single Master, Multiple Workers | For higher throughput workloads (Guide) |
| Multi-Master, Multiple Workers | Recommended for HA enterprise clusters (Guide) |
- View the Pre-validated Model List
- To deploy custom models from Hugging Face, follow the Hugging Face Deployment Guide
💡 Both validated and custom models are supported to meet diverse enterprise needs.
Two files are required before deployment:
inventory/hosts.yaml– Cluster inventory and topology (Single-Node Sample, Multi-Node Guide)inference-config.cfg– Component-level deployment config (Sample)
Run the following script to deploy the inference platform:
bash inference-stack-deploy.sh