-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Creating a Vertex AI endpoint of pre-trained model for Inferencing. #9
Comments
@StateGovernment take a look at this example. This is pytorch only, so you'll have to convert the model to pytorch as demonstrated in the training dreambooth repo. In your case, where you are using a local model, you need to modify the Dockerfile to copy your model. Add a new line here
Then call the docker build like:
I think that should work. |
Thank you for the suggestion. I see, following the code I believe a REST endpoint is being deployed using FastAPI through Docker. But our use-case actually involves creating a pure vertex ai Endpoint that could serve inference requests, is there a way to deploy the model to a vertex Endpoint instead of FastAPI? |
This is a deployment pattern for Vertex endpoints using custom containers. The readme describes how to deploy it to the endpoint as described in the Vertex AI documentation https://cloud.google.com/vertex-ai/docs/predictions/use-custom-container |
Labels : How to/ Suggestions.
I am looking for suggestions on, how to create a Vertex AI endpoint out of a Trained Dreambooth stable diffusion model. Inorder to run inference on the model through endpoint. How do I go about this
The text was updated successfully, but these errors were encountered: