Deploy, scale, and monitor AI workloads with the OCI AI Blueprints platform, and reduce your GPU onboarding time from weeks to minutes.
OCI AI Blueprints is a streamlined, no-code solution for deploying and managing Generative AI workloads on Kubernetes Engine (OKE). By providing opinionated hardware recommendations, pre-packaged software stacks, and out-of-the-box observability tooling, OCI AI Blueprints helps you get your AI applications running quickly and efficiently—without wrestling with the complexities of infrastructure decisions, software compatibility, and MLOps best practices.
Getting Started
About OCI AI Blueprints
- What is OCI AI Blueprints?
- Why use OCI AI Blueprints?
- Features
- List of Blueprints
- FAQ
- Support & Contact
API Reference
Additional Resources
Install OCI AI Blueprints by clicking on the button below:
Blueprints go beyond basic Terraform templates. Each blueprint:
- Offers validated hardware suggestions (e.g., optimal shapes, CPU/GPU configurations),
- Includes end-to-end application stacks customized for different GenAI use cases, and
- Comes with monitoring, logging, and auto-scaling configured out of the box.
After you install OCI AI Blueprints to an OKE cluster in your tenancy, you can deploy these pre-built blueprints:
Blueprint | Description |
---|---|
LLM Inference with vLLM | Deploy Llama 2/3/3.1 7B/8B models using NVIDIA GPU shapes and the vLLM inference engine with auto-scaling. |
Fine-Tuning Benchmarking | Run MLCommons quantized Llama-2 70B LoRA finetuning on A100 for performance benchmarking. |
LoRA Fine-Tuning | LoRA fine-tuning of custom or HuggingFace models using any dataset. Includes flexible hyperparameter tuning. |
Health Check | Comprehensive evaluation of GPU performance to ensure optimal hardware readiness before initiating any intensive computational workload. |
CPU Inference | Leverage Ollama to test CPU-based inference with models like Mistral, Gemma, and more. |
If you have any questions, issues, or feedback, contact [email protected] or [email protected].