Watch this webinar for a showcasing of ALCF's Inference Service, which provides cloud-like access to diverse AI models—including Large Language Models (LLMs)—on existing high-performance computing (HPC) clusters. ALCF's Benoit Côté demonstrates how to integrate the Inference Service within scientific applications and share examples of interacting with our chat interface and API.