Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Any plan to support Llama 405B (Llama 3.1) #103

Closed
yanjunplay opened this issue Jul 23, 2024 · 3 comments
Closed

Any plan to support Llama 405B (Llama 3.1) #103

yanjunplay opened this issue Jul 23, 2024 · 3 comments

Comments

@yanjunplay
Copy link
Contributor

I've been waiting for a while, haha..

Today's the Llama3.1 405B model is officially announced https://llama.meta.com/. Do we have any plan to support this soon? I assume we need to adjust Eagle model accordingly.

Thanks a lot!!

@hongyanz
Copy link
Contributor

As an academic group, we don't have enough GPUs to proceed the inference/training on LLama 405B (which should require 8x H100s for inference). If any industry would like to support us, please do let us know. Thanks.

@yanjunplay
Copy link
Contributor Author

Thanks @hongyanz for the reply. Let me do some study and see how we can collaborate.

@hongyanz hongyanz closed this as completed Aug 6, 2024
@anilmurty
Copy link

anilmurty commented Aug 7, 2024

hey @hongyanz - I'd be interested in seeing if we can support you training 405B on https://akash.network/. We're able to run Llama-3.1 405B for inference. Are you interested in discussing this?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants