The OG GPT3 based instruct model from OpenAI.
Supported context length
Price for prompt tokens*
Price for response tokens*
*Note: Data based on 11/14/2023
Here's how GPT-3.5-turbo-instruct performed across all three task types
Digging deeper, here’s a look how GPT-3.5-turbo-instruct performed across specific datasets
|Tasks||Insights||Dataset Name||Dataset Performance|
|QA without RAG||The model performs well which show less bias and good factual knowledge.||Truthful QA|
|QA with RAG||The model peforms decently well which demonstrates good reasoning and comprehension skills. It struggles on mathematical skills as it scores relatively low on DROP compared to other dataset. It performs almost as good as Llama-2-70b-chat.||MS Marco|
|Long form text generation||The model performs satisfactory at this task which demonstrates good factual knowledge and ability to generate long text without errors.||Open Assistant|
💰 Cost insights
The model offers a decent balance of cost and performance. It is 30x cheaper compared to GPT4 and 2x costlier compared to Llama 70b variant. OpenAI provides generous rate limits for production application. But we would suggest to use GPT3.5 instead of this.