Skip to content

Deploy, scale, and monitor AI workloads with the OCI AI Blueprints platform, and reduce your GPU onboarding time from weeks to minutes.

License

Notifications You must be signed in to change notification settings

oracle-quickstart/oci-ai-blueprints

Repository files navigation

OCI AI Blueprints

Deploy, scale, and monitor AI workloads with the OCI AI Blueprints platform, and reduce your GPU onboarding time from weeks to minutes.

OCI AI Blueprints is a streamlined, no-code solution for deploying and managing Generative AI workloads on Kubernetes Engine (OKE). By providing opinionated hardware recommendations, pre-packaged software stacks, and out-of-the-box observability tooling, OCI AI Blueprints helps you get your AI applications running quickly and efficiently—without wrestling with the complexities of infrastructure decisions, software compatibility, and MLOps best practices.

Install OCI AI Blueprints

Table of Contents

Getting Started

About OCI AI Blueprints

API Reference

Additional Resources

Getting Started

Install OCI AI Blueprints by clicking on the button below:

Install OCI AI Blueprints

Blueprints

Blueprints go beyond basic Terraform templates. Each blueprint:

  • Offers validated hardware suggestions (e.g., optimal shapes, CPU/GPU configurations),
  • Includes end-to-end application stacks customized for different GenAI use cases, and
  • Comes with monitoring, logging, and auto-scaling configured out of the box.

After you install OCI AI Blueprints to an OKE cluster in your tenancy, you can deploy these pre-built blueprints:

Blueprint Description
LLM Inference with vLLM Deploy Llama 2/3/3.1 7B/8B models using NVIDIA GPU shapes and the vLLM inference engine with auto-scaling.
Fine-Tuning Benchmarking Run MLCommons quantized Llama-2 70B LoRA finetuning on A100 for performance benchmarking.
LoRA Fine-Tuning LoRA fine-tuning of custom or HuggingFace models using any dataset. Includes flexible hyperparameter tuning.
Health Check Comprehensive evaluation of GPU performance to ensure optimal hardware readiness before initiating any intensive computational workload.
CPU Inference Leverage Ollama to test CPU-based inference with models like Mistral, Gemma, and more.

Support & Contact

If you have any questions, issues, or feedback, contact [email protected] or [email protected].

About

Deploy, scale, and monitor AI workloads with the OCI AI Blueprints platform, and reduce your GPU onboarding time from weeks to minutes.

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages