Learn how to leverage the Llama 2 large language model (LLM) and how open-source LLMs can run on self-hosted hardware, made possible through techniques such as quantization by using the llama.cpp package. In this online course, you will explore how Meta’s Llama 2 fits into the larger AI ecosystem and how you can use it to develop Python-based LLM applications. Get hands-on skills in improving and constraining Llama 2 output, and learn how to get more robust data interchanges between Python application code and LLM inference. Understand different Llama 2 model variants, how they were trained, and how to interact with these models in Python.
This course is free for U-M alumni, students, faculty, and staff.