Ai Ml Genai On Nvidia H100 Gpus On Red Hat Openshift Ai
Last updated 7/2025
MP4 | Video: h264, 1920x1080 | Audio: AAC, 44.1 KHz
Language: English | Size: 817.14 MB | Duration: 1h 29m
OpenShift & OpenShift AI on NVIDIA H100: From Bare-Metal to Production in One Day
What you'll learn
Stand up a bare-metal H100 node, validate firmware & BIOS, and register it in a fresh OpenShift cluster
Install and tune the NVIDIA GPU Operator with Multi-Instance GPU (MIG) profiles for maximum utilisation
Deploy Red Hat OpenShift AI (RHOAI) and run a real Mistral LLM workload with Ollama
Monitor, troubleshoot, upgrade, and scale the platform in production
Requirements
One NVIDIA H100 (or other Ampere/Hopper) server-physical or virtualised
A workstation that can SSH into the node and run the "oc" CLI
(Optional) A Red Hat account to pull mirrored images
Description
Unlock the power of enterprise-grade AI in your own data center-step-by-step, from bare-metal to production-ready inference. In this hands-on workshop, you'll learn how to transform a single NVIDIA H100 server and a lightweight virtualization host into a fully featured Red Hat OpenShift cluster running OpenShift AI, the NVIDIA GPU Operator, and real LLM workloads (Mistral-7B with Ollama). We skip the theory slides and dive straight into keyboards and terminals-every YAML, every BIOS toggle, every troubleshooting trick captured on video.What you'll buildA three-node virtual control plane + one bare-metal GPU worker, deployed via the new Agent-based InstallerGPU Operator with MIG slicing, UUID persistence, and live metrics in GrafanaOpenShift AI (RHODS) with Jupyter and model-serving pipelinesA production-grade load balancer, DNS zone, and HTTPS ingress-no managed cloud neededHands-on every step: you'll inspect firmware through iDRAC, patch BIOS settings, generate a custom Agent ISO, boot the cluster, join the GPU node, and push an LLM endpoint you can curl in under a minute. Along the way, we'll upgrade OpenShift, monitor GPU temps, and rescue a "Node Not Ready" scenario-because real life happens.Who should enrollDevOps engineers, SREs, and ML practitioners who have access to a GPU server (H100, H800, or even an A100) and want a repeatable, enterprise-compatible install path. Basic Linux and kubectl skills are assumed; everything else is taught live.By course end, you'll have a battle-tested Git repository full of manifests, a private Agent ISO pipeline you can clone for new edge sites, and the confidence to stand up-or scale out-your own GPU-accelerated OpenShift AI platform. Join us and ship your first on-prem LLM workload today.
Machine Learning Engineers,DevOps Engineers,Site Reliability Engineers (SREs),Python Developers Exploring Infrastructure,First Steppers into AI Operations
Buy Premium From My Links To Get Resumable Support and Max Speed
https://rapidgator.net/file/beb93befc6d … I.rar.html
https://nitroflare.com/view/D444356D4A1 … ift_AI.rar