Model Insights

Mistral-7b-instruct-v0.1

The new kid from Mistral is a instruct fine-tuned version of the Mistral-7B-v0.1 trained with a variety of publicly available conversation datasets.

Model

Mistral-7b-instruct-v0.1

Details

Developer

Mistral

License

Apache-2.0

Model parameters

7B

Pretraining tokens

8T

Release date

Sep 2023

Supported context length

8K

Price for prompt tokens*

$0.15/Million tokens

Price for response tokens*

$0.15/Million tokens

*Note: Data based on 11/14/2023

Model Performance Across Task-Types

Here's how Mistral-7b-instruct-v0.1 performed across all three task types

Metric
ChainPoll Score
QA without RAG
0.59
QA with RAG
0.67
Long form text generation
0.65

Model Info Across Task-Types

Digging deeper, here’s a look how Mistral-7b-instruct-v0.1 performed across specific datasets

TasksInsightsDataset NameDataset Performance
QA without RAGThe model does not perform well which show bias and errors in factual knowledge.Truthful QA
0.57
Trivia QA
0.61
QA with RAGThe model peforms decently well which demonstrates good reasoning and comprehension skills. It struggles a lot on mathematical skills as it scores relatively low on DROP compared to other dataset. Its performance is as good as 70b variant & GPT-3.5-turbo-instruct.MS Marco
0.88
Hotpot QA
0.54
Drop
0.51
Narrative QA
0.75
Long form text generation The model performs poorly which shows weakness in generating long text without factual errors.Open Assistant
0.65

💰 Cost insights

The model is relative cheaper to run but does not perform great. It is 13x cheaper compared to GPT3.5 and 6x cheaper compared to Llama 70b variant. We suggest using Zephyr-7b-beta instead of this.

LLMHALLUCINATIONINDEXLLM