llama icon indicating copy to clipboard operation
llama copied to clipboard

When will the Llama 2 34B model be released?

Open imoneoi opened this issue 2 years ago • 19 comments

imoneoi avatar Jul 30 '23 15:07 imoneoi

Yes a status update or estimation would be great.

viperwasp avatar Aug 02 '23 07:08 viperwasp

Look forward to it

hunkim avatar Aug 05 '23 00:08 hunkim

I am also getting a bit apprehensive about that, and my fear is that it has been cancelled, as even this repo makes no mention of it even if it was a "coming soon!", and Meta has been silent about it since the paper came out, as they seem happy how the 13b (the mainstream budget option and a pretty good model, still) and 70b (which is inaccessible for most people) are being received.

Which is a shame, 34b would be the best option for AI researchers with single 24gb cards and people who run a small business who just wanted a "smart enough" model but cannot afford several industrial grade >40gb+ cards. If the LLaMA1 fine-tunes are anything to go by, a fine-tuned 34b would be just as good as 70b for most reasoning-related tasks while being half of its parameter count, especially considering how good the current 13b models are getting.

Subarasheese avatar Aug 05 '23 00:08 Subarasheese

@Subarasheese I agree, but I guess their red team is working hard. As you know, their violation percentage is high, so I guess they are working on it.

image

hunkim avatar Aug 05 '23 00:08 hunkim

@Subarasheese I agree, but I guess their red team is working hard. As you know, their violation percentage is high, so I guess they are working on it. image

What I find weird is, that chart is in regards to 34b-chat, which is just a fine-tune over the base 34b model, and they did not even release the base model... If the base model itself is the problem, and not the dataset they used to fine-tune it into a chat model, I honestly doubt they will bother to retrain the whole thing. And if the dataset for fine-tune is the problem here, it makes me wonder why is it causing them so much trouble and taking that long (considering the models were not released when they wrote the paper)...

Subarasheese avatar Aug 05 '23 01:08 Subarasheese

Perhaps they are looking into fixing some of LLama 2's repetition issues (as seen here: https://www.reddit.com/r/LocalLLaMA/comments/155vy0k/llama_2_too_repetitive/)

Perhaps thats why its taking a bit longer. But I'm willing to wait for quality models and am thankful for the great work Meta has done.

Dampfinchen avatar Aug 05 '23 22:08 Dampfinchen

Do they realize their violation % being high is a good thing

teknium1 avatar Aug 07 '23 09:08 teknium1

Do they realize their violation % being high is a good thing

I don't think so, a SFW model have own use cases. And they've tested LaMA-2 34B Chat.

The foundational LLaMA-2 models aren't aligned, so you can fine-tune your own alignment for it. I wonder what's the issue: their fine-tune being meh for the model, or the model being reluctant to be fine-tuned. I am incline to believe it's the latter, because Meta could just release the LLaMA-2 34B foundational model without the Chat fine-tune otherwise.

Anf the "violation rate" of the chat model is on par with ChatGPT, which is a successful commercial service... So yeah, I think the issue isn't that, it's just a symptom. And they are working on some fix.

Hopefully. Would be a shame to skip 34B, it's a VERY useful size.

EriIaz avatar Aug 10 '23 20:08 EriIaz

Considering codellama 34b was published, could we get an update on the foundational model?

maximegmd avatar Aug 24 '23 21:08 maximegmd

@Subarasheese I agree, but I guess their red team is working hard. As you know, their violation percentage is high, so I guess they are working on it.

image

What does this have to do with the base model

teknium1 avatar Aug 27 '23 22:08 teknium1

Any update?

hunkim avatar Sep 03 '23 23:09 hunkim

Any update?

abhibst avatar Sep 04 '23 07:09 abhibst

@macarran We were really wondering if you have any plans to release 34B anytime soon. It would be very useful if you could let us know a rough timeline as well.

hunkim avatar Sep 11 '23 14:09 hunkim

Any update? ^_^

teknium1 avatar Sep 12 '23 00:09 teknium1

I'll sign up for threads if you release this

ozreact avatar Sep 12 '23 00:09 ozreact

Any update? :(

imoneoi avatar Sep 15 '23 02:09 imoneoi

If there were an update I'm sure we would have heard.

Frankly at this point I wouldn't be surprised if any efforts to release LlaMa 2 34B have been redirected to working on LLaMa 3.

On Fri, Sep 15, 2023 at 9:37 AM One @.***> wrote:

Any update? :(

— Reply to this email directly, view it on GitHub https://github.com/facebookresearch/llama/issues/590#issuecomment-1720404449, or unsubscribe https://github.com/notifications/unsubscribe-auth/ACRZOUJGZP6CZJWWBUMSKI3X2O5QLANCNFSM6AAAAAA25IEMGU . You are receiving this because you are subscribed to this thread.Message ID: @.***>

ericksonc avatar Sep 15 '23 02:09 ericksonc

If there were an update I'm sure we would have heard. Frankly at this point I wouldn't be surprised if any efforts to release LlaMa 2 34B have been redirected to working on LLaMa 3. On Fri, Sep 15, 2023 at 9:37 AM One @.> wrote: Any update? :( — Reply to this email directly, view it on GitHub <#590 (comment)>, or unsubscribe https://github.com/notifications/unsubscribe-auth/ACRZOUJGZP6CZJWWBUMSKI3X2O5QLANCNFSM6AAAAAA25IEMGU . You are receiving this because you are subscribed to this thread.Message ID: @.>

The government could ban opensource models by then, we cant wait until after 2024 begins!

teknium1 avatar Sep 16 '23 00:09 teknium1

The government could ban opensource models by then, we cant wait until after 2024 begins!

@teknium1 what do you mean? I'm shocked! Is it true?

RahulBhalley avatar Oct 19 '23 14:10 RahulBhalley