This workshop will cover running Large Language Models (LLMs) on the Hoffman2 HPC cluster using Ollama for inference and Open WebUI for an easy-to-use interface. Participants will learn how to set up and manage LLM workloads on HPC resources, optimize performance, and interact with models through a web-based UI.
The session includes a live demo, hands-on training, and a Q&A segment for troubleshooting and best practices. Basic command-line knowledge and access to Hoffman2 are recommended.This workshop will cover running Large Language Models (LLMs) on the Hoffman2 HPC cluster using Ollama for inference and Open WebUI for an easy-to-use interface.
Participants will learn how to set up and manage LLM workloads on HPC resources, optimize performance, and interact with models through a web-based UI. The session includes a live demo, hands-on training, and a Q&A segment for troubleshooting and best practices. Basic command-line knowledge and access to Hoffman2 are recommended.