In-Person
LLMs on Yale's computing clusters hands-on workshop
Thu Feb 19, 2026
9:00 a.m.—11:00 a.m.
This event has passed.
This workshop will provide researchers interested in LLMs with the skills to launch and run inference-based workflows with open-source models on YCRC systems. Specifically, attendees will learn:
- What GPU resources are available for each YCRC system
- How to identify what GPU is needed for different LLMs
- How to launch an LLM on YCRC systems
- How to conduct inference via direct prompting with an LLM using ollama
- How to modify an LLM's parameters for reproducibility/consistency in response
- How to implement an LLM within python using jupyter and ollama
- Understand additional considerations for RAG and fine-tuning.
This is an advanced instruction. Working introductory knowledge of HPC on Yale's clusters is required. If you are a new user, we recommend watching the Introduction to HPC video and its accompanying slides, and completing the intro exercises before attending this workshop.
Seats are limited. Registration is required.