Enroll now:
https://bit.ly/4gX9yrq
The Llama 3.2 open multimodal model dropped a few days ago, and we worked with Meta to launch "Introducing Multimodal Llama 3.2," a short course on how to use it!
We’re pleased to launch Introducing Multimodal Llama 3.2, and learn from Amit Sangani, Senior Director of AI Partner Engineering at Meta, to learn all about the latest additions to the Llama models 3.1 and 3.2, from custom tool calling to multimodality and the new Llama stack.
You’ll learn about the new vision capabilities that Llama 3.2 brings to the Llama family, and how to leverage this along with tool-calling, and Llama Stack, an open-source orchestration layer for building on top of the Llama family of models.
You’ll get hands-on and:
- Learn about the new models, how they were trained, their features, and how they fit into the Llama family.
- Understand how to do multimodal prompting with Llama and work on advanced image reasoning use cases such as understanding errors on a car dashboard, adding up the total of three photographed restaurant receipts, grading written math homework, and many more.
- Learn different roles—system, user, assistant, ipython—in the Llama 3.1 and 3.2 family and the prompt format that identifies those roles.
- Understand how Llama uses the tiktoken tokenizer, and how it has expanded to a 128k vocabulary size that improves encoding efficiency and enables support for seven non-English languages.
- Learn how to prompt Llama to call built-in and custom tools with examples for web search and solving math equations.
- Learn about Llama Stack API, a standardized interface for canonical toolchain components like fine-tuning or synthetic data generation to customize Llama models and build agentic applications.
Start building exciting applications on Llama!
Learn more:
https://bit.ly/4gX9yrq