Attention in Transformers: Concepts and Code in PyTorchUnderstand and implement the attention mechanism, a key element of transformer-based LLMs, using PyTorch.StatQuest
ChatGPT Prompt Engineering for DevelopersLearn the fundamentals of prompt engineering for ChatGPT. Learn effective prompting, and how to use LLMs for summarizing, inferring, transforming, and expanding.OpenAI
Reinforcement Learning From Human FeedbackGet an introduction to tuning and evaluating LLMs using Reinforcement Learning from Human Feedback (RLHF) and fine-tune the Llama 2 model.Google Cloud
Prompt Engineering with Llama 2 & 3Learn best practices for prompting and selecting among Meta Llama 2 & 3 models. Interact with Meta Llama 2 Chat, Code Llama, and Llama Guard models.Meta
Open Source Models with Hugging FaceLearn how to easily build AI applications using open-source models and Hugging Face tools. Find and filter open-source models on Hugging Face Hub.Hugging Face
Efficiently Serving LLMsUnderstand how LLMs predict the next token and how techniques like KV caching can speed up text generation. Write code to serve LLM applications efficiently to multiple users.Predibase
Quantization FundamentalsLearn how to quantize any open-source model. Learn to compress models with the Hugging Face Transformers library and the Quanto library.Hugging Face
How Transformer LLMs WorkUnderstand the transformer architecture that powers LLMs to use them more effectively.Jay Alammar, Maarten Grootendorst