This video shows how to integrate NVIDIA Inference Microservice (NIM) on Oracle Kubernetes Engine (OKE) for deploying Large Language Models (LLMs) on NVIDIA GPUs.
✅ Set up NIM on OKE
✅ Deploy and optimize LLMs with GPU acceleration
✅ Leverage Kubernetes for scalable, efficient inference
Watch now to simplify and scale your LLM deployments on Oracle Cloud Infrastructure with NVIDIA NIM!
Don’t forget to like, comment, and subscribe for more tutorials and insights!
Presenters and contributors:
Nitin Satpute - GPU Specialist, Oracle
Check and register below for the upcoming sessions:
? https://www.oracle.com/emea/cloud/events/developer-coaching
Join us on our Slack community:
? https://developercoachingoci.slack.com/
Note: Screens and flows may have changed.
If you liked this session and would be happy to leave feedback on it, please reach us at:
✉️ [email protected].
#OCI #OKE #NVIDIA #NIM #LLM #Kubernetes #OracleCloud #AIDeployment #GPUpowered
----------------------------------------------
Copyright © 2025, Oracle and/or its affiliates.