If you are a researcher, academic institution, government agency, government partner, or other entity with a Llama use case that is currently prohibited by the Llama Community License or Acceptable Use Policy, or requires additional clarification, please contact llamamodels@meta.com with a detailed request. We will consider requests on a case-by-case basis.
For Llama 2 and Llama 3, the license restricts using any part of the Llama models, including the response outputs, to train another AI model (LLM or otherwise). For Llama 3.1 and Llama 3.2, this is allowed provided you include the correct attribution to Llama. See the license for more information.
Llama 2 and Llama 3 were primarily trained on English language data with some additional data from other languages. We do not expect the same level of performance in these other languages as in English. Llama 3.1 and Llama 3.2 support additional languages and are considered multilingual. See the Llama 3.1 and Llama 3.2 model cards for more information.
Llama models are broadly available to developers and licensees through a variety of hosting providers and on the Meta website. The Llama models are licensed under the applicable Llama Community License Agreement and accompanying Acceptable Use Policy, which provides a permissive license to the models along with certain restrictions to help ensure that the models are used responsibly. Hosting providers may have additional terms applicable to their services.
Hardware requirements vary based on the specific Llama model being used, latency, throughput and cost constraints. For the larger Llama models to achieve low latency, one would split the model across multiple inference chips (typically a GPU) with tensor parallelism. Llama models are known to execute in a performant manner on a wide variety of hardware including GPUs, CPUs (both x86 and ARM based), TPUs, NPUs and AI Accelerators. The smaller Llama models typically run on system-on-chip (SOC) platforms found on PC, Mobile and other Edge devices.
Llama models are auto-regressive language models, built on the transformer architecture. The core language models function by taking a sequence of words as input and predicting the next word, recursively generating text.
The model itself supports these parameters, but whether they are exposed or not depends on implementation.
This depends on your application. The Llama pre-trained models were trained for general large language applications, whereas the Llama instruct or chat models were fine tuned for dialogue specific uses like chat bots.
Developers may fine-tune Llama models for languages beyond English or other officially supported languages (e.g., Llama 3.1 and Llama 3.2 (for text to text applications) support 8 languages), provided that such use complies with the applicable Llama Community License Agreement and Acceptable Use Policy.
The employee may use the Llama multimodal models within the scope of the employee’s work for the non-EU based company, since the license extends to the non-EU based company. However, the employee may not use the Llama multimodal model for their own individual purposes if they are domiciled in the EU, although they may use a product or service that incorporates Llama multimodal models as an end user.
Yes, if you are a company with a principal place of business outside of the EU, you may distribute products or services that contain the Llama multimodal models in accordance with your standard global distribution business practices (for additional clarification, please contact llamamodels@meta.com). The EU restriction contained in the Llama 3.2 AUP only restricts the rights granted to EU based companies and individuals. The company’s customers in the EU would be able to use the product or service that incorporates the Llama multimodal models.
Yes. The non-EU based company is permitted to use the Llama multimodal models, even if it is a subsidiary of, or affiliated with, an EU based company. However, it is only the non-EU based company that is permitted to use the Llama multimodal models for any development or distribution. Any EU based affiliates may not be involved in such development or distribution (but, see the FAQ above).
Yes. The company’s principal place of business is outside the EU, and the company is therefore granted the license in the applicable Llama license. It may use its EU-based developers to perform work using the Llama multimodal models.
“MAU” means “monthly active users” that access or use your (and your affiliates’) products and services. Examples include users accessing an internet-based service and monthly users/customers of the licensee’s software applications or hardware devices.
If a developer is distributing any of the Llama models or related materials or any product or service that is built on or incorporates the Llama models (including another AI model), that developer must also provide a copy of the applicable Llama Community License Agreement. With respect to any product or service that is built on or incorporates the Llama models, the developer must also prominently display “Built with Llama” on a website, user interface, blogpost, about page, or product documentation that is related to that product or service. See the FAQ below for examples of what “prominently display” means.
If a developer uses a Llama 3.1 or Llama 3.2 model, such as Llama 3.1-405B, to create or train another AI model, for example by generating a synthetic dataset that is then used to train another AI model, then that developer must include “Llama” at the beginning of such AI model’s name if it is distributed. This case is different from the case above, where Llama models or materials are being distributed or a product or service is built on or incorporates Llama (a derivative work), since the AI model in this case may not be based on or incorporate Llama. Consequently, a statement that the AI model is “Built with Llama” is not required; however, “Llama” should be included at the start of the AI model’s name since a Llama model was used to create, train, fine tune, or otherwise improve the AI model.
Example 1: Developer builds a GenAI tool using Llama 3.1-405B and releases the tool on its platform. Developer’s platform includes an “about page” within the application or website that announces the creation of the new GenAI tool and explains its uses. In the first few sentences about the new GenAI tool, the “about page” states that the tool was "Built with Llama." This is an example of prominently displaying the required attribution. However, if Developer were to release a 4 page article about its new GenAI tool and only mention that it was "Built with Llama" somewhere on the last page or in a footnote, we would not consider it to be a prominent display of the required attribution. Developer also provides the Llama 3.1 Community License Agreement or a link to it with the materials related to the new GenAI tool, since it is considered a derivative work of Llama 3.1.
Example 2: Developer improves its existing service with new AI features built using Llama 3.2. Developer releases a blog post announcing its AI feature, and states that it was "Built with Llama" in the title, subtitle or first few sentences of the blog post. This is an example of prominently displaying the required attribution. Developer also provides the Llama 3.2 Community License Agreement or a link to it with the materials for its service, related to the new AI feature, since it is considered a derivative work of Llama 3.2.
Example 3: Developer builds an online library where AI models are available for download. It includes some Llama 3.2 models in its library. Along with the copy of Llama 3.2, it includes a “Notice” text file stating that “Llama 3.2 is licensed under the Llama 3.2 Community License, Copyright © Meta Platforms, Inc. All Rights Reserved.”
Example: Developer is training its own AI model. It needs additional data for pre-training its AI model, so it uses Llama 3.1-405B to create a synthetic dataset which it then uses to train its AI model. Developer later releases its AI model. Developer names its AI model "Llama Xxxx." This complies with the requirement that anyone that uses outputs or results from a Llama model to create, train, fine tune, or improve another AI model must include "Llama" at the beginning of such AI model name.
No. Llama is a foundational LLM model and is not currently operated by Meta as a service. You, or the downstream developer providing you with services, decide where you run the model and who has access to the inputs and outputs. Unless you provide inputs or outputs directly to Meta, or publish them, Meta cannot access, view, or use your inputs or outputs.
Llama is a foundational LLM model, and is not currently operated by Meta as a service and as a result does not have access to your inputs. You should consult with the downstream developer that builds a product, tool or service using Llama to understand how they ensure data privacy for sensitive/proprietary inputs.
“Disinformation” refers to information that is false or inaccurate, or is presented in such a way as to draw a false or inaccurate conclusion, and is intended to (or does) mislead or deceive.
Critical infrastructure is infrastructure that is considered essential by governments for the functioning of society and the economy. It includes, but is not limited to, highways, bridges, tunnels, utilities and related facilities and services, etc.
Example: Developer is an electric company which manages the electric grid and electricity supply for the City of Burlington. To assist in the management of the Burlington electric grid, which Developer and the City consider critical infrastructure, Developer is considering the use of Llama. The use of Llama to manage the electric grid is prohibited under the terms of the AUP.
No, such companies are not prohibited when their usage of Llama is not related to the operation of critical infrastructure. Llama, however, may not be used in the operation of critical infrastructure by any company, regardless of government certifications.