The pages in this section describe how to obtain the Llama models:
You can download the models directly from Meta or one of our download partners: Hugging Face or Kaggle.
Alternatively, you can work with our ecosystem partners to access the models through the services they provide. This approach can be especially useful if you want to work with the Llama 3.1 405B model.
Note: With Llama 3.1, we introduce the 405B model. This model requires significant storage and computational resources, occupying approximately 750GB of disk storage space and necessitating two nodes on MP16 for inferencing.