Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Create guide for how to deploy an existing Hugging Face model on self-hosted LLM Engine #141

Open
rkaplan opened this issue Jul 19, 2023 · 2 comments
Assignees
Labels
documentation Improvements or additions to documentation enhancement New feature or request

Comments

@rkaplan
Copy link
Contributor

rkaplan commented Jul 19, 2023

Requested by Biju on Twitter.

@yixu34 yixu34 added documentation Improvements or additions to documentation enhancement New feature or request labels Jul 19, 2023
@yixu34
Copy link
Member

yixu34 commented Jul 20, 2023

We're currently wrapping up some testing for a self-contained helm install on your own EKS cluster. Once that's ready, we'll ship the docs too.

@yixu34
Copy link
Member

yixu34 commented Jul 21, 2023

Btw just want to clarify and acknowledge that #153 solves part but not all of the ask - it shows you how to deploy a self-hosted model for an existing model in our Model Zoo, which represents a subset of Hugging Face models. It does not show how you would also add to the Model Zoo, i.e. build an endpoint from an arbitrary Hugging Face model. That will require some follow-up work.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

4 participants