OpenAI GPT-3 API: What is the difference between davinci and text-davinci-003?

题意:OpenAI GPT-3 API:davinci 和 text-davinci-003 有什么区别

问题背景:

I'm testing the different models for OpenAI, and I noticed that not all of them are developed or trained enough to give a reliable response.

我正在测试 OpenAI 的不同模型,我发现并不是所有模型都足够完善或训练充分,无法提供可靠的回应

The models I tested are the following:

我测试的模型如下:

复制代码
model_engine = "text-davinci-003"
model_engine = "davinci" 
model_engine = "curie" 
model_engine = "babbage" 
model_engine = "ada" 

I need to understand what the difference is between davinci and text-davinci-003, and how to improve the responses to match that response when you use ChatGPT.

我需要了解 davincitext-davinci-003 之间的区别,以及如何改进响应,使其与使用 ChatGPT 时的响应相匹配

问题解决:

TL;DR 摘要

  • text-davinci-003 is the newer and more capable model than davinci

text-davinci-003 是比 davinci 更新且更强大的模型

  • text-davinci-003 supports a longer context window than davinci

text-davinci-003 支持比 davinci 更长的上下文窗口

  • text-davinci-003 was trained on a more recent dataset than davinci

text-davinci-003 使用比 davinci 更新的数据集进行训练

  • text-davinci-003 is cheaper than davinci

text-davinci-003davinci 更便宜

  • text-davinci-003 is not available for fine-tuning, while davinci is

text-davinci-003 不支持微调,而 davinci 支持

Capabilities 能力

As stated in the official OpenAI article: 正如 OpenAI 官方文章中所述

While both davinci and text-davinci-003 are powerful models, they differ in a few key ways.

text-davinci-003 is the newer and more capable model, designed specifically for instruction-following tasks. This enables it to respond concisely and more accurately - even in zero-shot scenarios, i.e. without the need for any examples given in the prompt. davinci, on the other hand, can be fine-tuned on a specific task, which can make it very effective if you have access to at least a few hundred training examples.

Additionally, text-davinci-003 supports a longer context window (max prompt+completion length) than davinci - 4097 tokens compared to davinci's 2049.

Finally, text-davinci-003 was trained on a more recent dataset, containing data up to June 2021. These updates, along with its support for Inserting text, make text-davinci-003 a particularly versatile and powerful model we recommend for most use-cases.

Use text-davinci-003 because the other models you mentioned in your question are less capable.

使用 text-davinci-003,因为您在问题中提到的其他模型能力较弱

If you buy a ChatGPT Plus subscription, you can also use gpt-3.5-turbo or gpt-4. So, to get similar responses as you get from ChatGPT, it depends on whether you are subscribed or not. For sure, gpt-3.5-turbo and gpt-4 are even more capable than text-davinci-003.

如果您购买 ChatGPT Plus 订阅,您还可以使用 gpt-3.5-turbogpt-4。因此,获得与 ChatGPT 相似的响应取决于您是否订阅。可以肯定的是,gpt-3.5-turbogpt-4 的能力甚至比 text-davinci-003 更强

Costs 成本

text-davinci-003 is cheaper than davinci, as stated on the official OpenAI website:

正如 OpenAI 官方网站所述,text-davinci-003davinci 更便宜

MODEL USAGE
davinci $0.1200 / 1K tokens
text-davinci-003 $0.0200 / 1K tokens

Fine-tuning availability 微调可用性

text-davinci-003 is not available for fine-tuning, while davinci is, as stated in the official OpenAI documentation:

正如 OpenAI 官方文档所述,text-davinci-003 不支持微调,而 davinci 支持

Fine-tuning is currently only available for the following base models: davinci, curie, babbage, and ada. These are the original models that do not have any instruction following training (like text-davinci-003 does for example).

MODEL FINE-TUNING AVAILABILITY TRAINING
davinci Yes $0.0300 / 1K tokens
text-davinci-003 No
相关推荐
机器之心1 天前
让AI自我进化?斯坦福华人博士答辩视频火了,庞若鸣参与评审
人工智能·openai
曲幽1 天前
FastAPI流式输出实战与避坑指南:让AI像人一样“边想边说”
python·ai·fastapi·web·stream·chat·async·generator·ollama
狗胜1 天前
二等兵·甘: 当 Agent 开始替你做决定,系统治理就不是可选项
openai
后端AI实验室1 天前
用AI写代码,我差点把漏洞发上线:血泪总结的10个教训
java·ai
王小酱1 天前
Everything Claude Code 新手教学指南(中文版)
openai·ai编程·claude
树獭叔叔1 天前
19-为什么AI工程这么喜欢"创造名词":从Prompt到Skill的造词运动
后端·aigc·openai
程序员鱼皮2 天前
67个AI编程必会知识,1.6w字一次讲透!女友:“你要考研啊?!”
ai·程序员·编程·ai编程·vibe coding
xun_xing2 天前
一篇文章让你彻底熟悉AI大模型(一)
llm·openai·ai编程