The second largest chat model in Llama 2 family of LLMs developed and publicly released by Meta. This model was pretrained on 2 trillion tokens of data from publicly available sources and fine-tuned on over one million human-annotated instruction datasets.
Supported context length
Price for prompt tokens*
Price for response tokens*
*Note: Data based on 11/14/2023
Here's how Llama-2-13b-chat performed across all three task types
Digging deeper, here’s a look how Llama-2-13b-chat performed across specific datasets
|Tasks||Insights||Dataset Name||Dataset Performance|
|QA without RAG||The model does not perform well which show bias and errors in factual knowledge.||Truthful QA|
|QA with RAG||The model exhibits satisfactory performance which demonstrates good reasoning and comprehension skills. It struggles on mathematical skills as it scores relatively low on DROP compared to other dataset. It performs almost as good as 70b variant & GPT-3.5-turbo-instruct.||MS Marco|
|Long form text generation||The model performs satisfactory at this task which shows good ability to generate long text without factual errors.||Open Assistant|
💰 Cost insights
The model offers a decent balance of cost and performance. It is 8x cheaper compared to GPT3.5 and 4x cheaper compared to Llama 70b variant.