openchat icon indicating copy to clipboard operation
openchat copied to clipboard

What is the difference between these versions of OpenChat?

Open RZFan525 opened this issue 1 year ago • 8 comments

Thank you for your amazing work! I have some questions below:

In alpaca eval leaderboard, there are 5 versions: OpenChatV3.1, OpenChatV2-W, OpenChatV2, OpenChat, OpenChat8192. What is the difference? And what is the datasets used? What is the difference between openchat_shareGPT_v3 and openchat_shareGPT4? Which datasets do you use for OpenChatV3.1?

Looking forward to your reply.

RZFan525 avatar Sep 11 '23 02:09 RZFan525

These are different versions of the OpenChat release.

The OpenChat, OpenChat8192 were trained solely on GPT-4 answers from ShareGPT (openchat_sharegpt4_dataset).

OpenChatV3.1, OpenChatV2-W, OpenChatV2 were trained on all ShareGPT data (including GPT-3.5 and GPT-4, openchat_sharegpt_v3).

Besides, OpenChat V3 and later leveraged Llama 2 base model, while others are based on Llama 1.

imoneoi avatar Sep 11 '23 03:09 imoneoi

These are different versions of the OpenChat release.

The OpenChat, OpenChat8192 were trained solely on GPT-4 answers from ShareGPT (openchat_sharegpt4_dataset).

OpenChatV3.1, OpenChatV2-W, OpenChatV2 were trained on all ShareGPT data (including GPT-3.5 and GPT-4, openchat_sharegpt_v3).

Besides, OpenChat V3 and later leveraged Llama 2 base model, while others are based on Llama 1.

Thank you for your detailed reply!

RZFan525 avatar Sep 11 '23 05:09 RZFan525

These are different versions of the OpenChat release.

The OpenChat, OpenChat8192 were trained solely on GPT-4 answers from ShareGPT (openchat_sharegpt4_dataset).

OpenChatV3.1, OpenChatV2-W, OpenChatV2 were trained on all ShareGPT data (including GPT-3.5 and GPT-4, openchat_sharegpt_v3).

Besides, OpenChat V3 and later leveraged Llama 2 base model, while others are based on Llama 1.

In addition to the datasets used, what did you update in OpenChatV2-W comparing to OpenchatV2?

RZFan525 avatar Sep 11 '23 06:09 RZFan525

Sorry to trouble you again.

I'm wonder what is updated in OpenChatV3.2 SUPER which boosts the score of MT-Bench.

RZFan525 avatar Sep 11 '23 07:09 RZFan525

Sorry to trouble you again.

I'm wonder what is updated in OpenChatV3.2 SUPER which boosts the score of MT-Bench.

Basically hyperparameter tuning :(

imoneoi avatar Sep 12 '23 07:09 imoneoi

These are different versions of the OpenChat release. The OpenChat, OpenChat8192 were trained solely on GPT-4 answers from ShareGPT (openchat_sharegpt4_dataset). OpenChatV3.1, OpenChatV2-W, OpenChatV2 were trained on all ShareGPT data (including GPT-3.5 and GPT-4, openchat_sharegpt_v3). Besides, OpenChat V3 and later leveraged Llama 2 base model, while others are based on Llama 1.

In addition to the datasets used, what did you update in OpenChatV2-W comparing to OpenchatV2?

V2-W leveraged weighted+conditioned finetuning, while V2 is conditioned finetuning

imoneoi avatar Sep 12 '23 07:09 imoneoi

These are different versions of the OpenChat release. The OpenChat, OpenChat8192 were trained solely on GPT-4 answers from ShareGPT (openchat_sharegpt4_dataset). OpenChatV3.1, OpenChatV2-W, OpenChatV2 were trained on all ShareGPT data (including GPT-3.5 and GPT-4, openchat_sharegpt_v3). Besides, OpenChat V3 and later leveraged Llama 2 base model, while others are based on Llama 1.

In addition to the datasets used, what did you update in OpenChatV2-W comparing to OpenchatV2?

V2-W leveraged weighted+conditioned finetuning, while V2 is conditioned finetuning

Many many thanks for your reply!

Have a good time!

RZFan525 avatar Sep 12 '23 08:09 RZFan525

These are different versions of the OpenChat release.

The OpenChat, OpenChat8192 were trained solely on GPT-4 answers from ShareGPT (openchat_sharegpt4_dataset).

OpenChatV3.1, OpenChatV2-W, OpenChatV2 were trained on all ShareGPT data (including GPT-3.5 and GPT-4, openchat_sharegpt_v3).

Besides, OpenChat V3 and later leveraged Llama 2 base model, while others are based on Llama 1.

openchat 3.5 is based on Mistral-7B? Based on llama1/2 or Mistral without adding glossaries and post-pretrain, can such good results be achieved only with SFT and C-RLFT? thanks

hy-2013 avatar Apr 23 '24 09:04 hy-2013