Documentation

Get started with Llama

This guide provides information and resources to help you set up Llama including how to access the model, hosting, how-to and integration guides. Additionally, you will find supplemental materials to further assist you while building with Llama.

What's new:

Llama 4 Scout
Natively multimodal model that offers single H100 GPU efficiency and a 10M context window
Llama 4 Maverick
Natively multimodal model for image and text understanding and fast responses at a low cost
Llama Guard 4
Llama Prompt Guard 2
Updated protection models with higher performance and support for Llama 4

Get the models

You can obtain the models directly from Meta or from one of our partners, Hugging Face, Kaggle or from our 1B/3B or 405B ecosystem partners.
Download the models

Llama 4 Scout

Natively multimodal model that offers single H100 GPU efficiency and a 10M context window

Model Card

Model attributes in easy to consume, standard format.
See Card on GitHub

Prompt Format

How to construct effective prompts.
Get Prompt Guidance

Get the Model

Llama 4 Maverick

Natively multimodal model for image and text understanding and fast responses at a low cost

Model Card

Model attributes in easy to consume, standard format
See Card on GitHub

Prompt Format

How to construct effective prompts.
Get Prompt Guidance

Get the Model

Llama Guard 4 (New)

Updated protection models with higher performance and support for Llama 4

Model Card

Model attributes in easy to consume, standard format.
See Card on GitHub

Prompt Format

How to construct effective prompts.
Get Prompt Guidance

Get the Model

Find us on GitHub

A black letter O with a white background

Llama Cookbook

Notebooks and demos for learning Llama. Scripts for fine-tuning Llama3 with single/multi-node GPUs.

Learn more
A black letter O with a white background

Llama Stack

Defines and standardizes the building blocks needed to bring generative AI applications to market.

Learn more

Explore the new capabilities of Llama 3.2

The Llama 3.2 lightweight models enable Llama to run on phones, tablets, and edge devices. View the video to see Llama running on phone. To see how this demo was implemented, check out the example code from ExecuTorch.

Other topics in this Guide

To discover more about what's possible with the Llama family of models, explore the topics below.

placeholder-image

Llama Everywhere

Notebooks and information on how to run Llama on your local hardware or in the cloud.
Learn more
placeholder-image

How-To Guides

An overview of the processes for developing any LLM: fine-tuning, prompt engineering, and model validation.
Learn more
placeholder-image

Integration Guides

Develop solutions based on Code Llama, LangChain, and LlamaIndex.
Learn more
placeholder-image

Community Support

Compilation of resources available from the community.
Learn more