Intersight AI Bridge simplifies and accelerates the initial installation and usage of AI workloads such as Cisco AI Pods.

Featured at Cisco Live 2026 EMEA : DEVNET-2488
This project provides scripts and configurations to:
Tip
Each step can be used independently.
Caution
This project can be used for OpenShift deployment or Ubuntu. Please follow the righ guidelines for your endgoal.
There is no preferred way to use AI Pods, however Cisco Validated Designs suggest to use OpenShift.
Ubuntu method can be used for easy and quick proof of concept deployment where OpenShift is recommended for production deployment.
Detailed instructions for Step 1
(Can be skipped if you prefer manual installation or are not using Intersight.)
You can either:
(Can be skipped if you prefer manual installation or are not using Intersight.)
You have the choice to setup for Ubuntu or Red Hat OpenShift:
Note
This can also be used on any Linux system, without Cisco UCS hardware or Intersight licenses.
Connect to the server OS, clone this repository, navigate into the project directory and make shell scripts executable:
git clone https://github.com/mabuelgh/intersight-ai-bridge cd intersight-ai-bridge chmod +x *.sh
If needed, define the variable PROXY_URL in setup.sh file, that will be used to configure system proxy & Docker proxy:
sudo nano setup.sh PROXY_URL="http://proxy.example.com:80" # <--- REPLACE WITH YOUR ACTUAL PROXY
Run the setup script:
./setup.sh
Verify installation after reboot of the OS:
cd intersight-ai-bridge
./checking.shThis process will trigger the creation of a Docker container. It will then display your GPUs inside the container to confirm the Nvidia container toolkit installation.
You have the choice to launch use case scenarions for Ubuntu or Red Hat OpenShift:
After setup, choose one of the following scenarios:
Launch with the Text Generation WebUI project:
./scenario1.sh
Note: You may need to load your model in the settings page before using it.
Launch vLLM with OpenWebUI:
./scenario2.sh
Note: If not done automatically, select your model on the top left corner of OpenWebUI.
Launch vLLMs with RAG for file-based context:
./scenario3.sh
Note: This project comes with sample files about fictives company descriptions.
For dual GPU infra, another file docker-compose-vllm-RAG-dual-GPU.yml can be used instead of docker-compose-vllm-RAG.yml.
Once running, you can ask questions such as:
Launch vLLMs with curl containers:
./scenario4.sh
Important
This scenario was made for dual GPU infra, remove the "gpu2" containers in docker-compose-vllm-stresstest.yml if necessary.
Owner
Contributors
Categories
Data CenterProducts
IntersightProgramming Languages
ShellLicense
Code Exchange Community
Get help, share code, and collaborate with other developers in the Code Exchange community.View Community