Meta developed and released the Llama 2 family of large language models (LLMs), a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. Its on par with some popular closed-source models like ChatGPT and PaLM.
This release includes model weights and starting code for pretrained and fine-tuned Llama language models — ranging from 7B to 70B parameters.
This repository is intended as a minimal example to load Llama 2 models and run inference.
These models are not finetuned for chat or Q&A. They should be prompted so that the expected answer is the natural continuation of the prompt.
example_text_completion.py for some examples. To illustrate, see the command below to run it with the llama-2-7b model (MP
value):
<
tags, EOS
tokens, and the whitespaces and breaklines in between (we recommend calling You can also deploy additional classifiers for filtering out inputs and outputs that are deemed unsafe.
Github: https://github.com/facebookresearch/llama
Contact us with your foundation model usage requirements.