Text Generation with Custom HuggingFace Model
Last updated
Last updated
In this demo we will:
Launch a pretrained custom text generation HuggingFace model in a Seldon Deployment
Send a text input request to get a generated text prediction
The custom HuggingFace text generation model is based on the TinyStories-1M model in the HuggingFace hub.
In the Overview
page, click Create new deployment
.
Enter the deployment details as follows, then click Next
:
Configure the default predictor as follows, then click Next
:
Runtime
HuggingFace
Model Project
default
Model URI
gs://seldon-models/scv2/samples/mlserver_1.6.0/huggingface-text-gen-custom-tiny-stories
Storage Secret
(leave blank/none) [2]
Model Name
transformer
Skip to the end and click Launch.
When your deployment is launched successfully the status will read as Available
.
The seldon
and seldon-gitops
namespaces are installed by default, which may not always be available. Select a namespace which best describes your environment.
A secret may be required for private buckets.
Additional steps may be required for your specific model.
Click the hf-custom-tiny-stories
deployment that you created.
In the Deployment Dashboard
page, click Predict
in the left pane.
In the Predict page, click Enter JSON and paste the following text:
Click the Predict
button.
Congratulations, you've successfully sent a prediction request using a custom HuggingFace model! 🥳
Why not try our other demos? Or perhaps try running a larger-scale model? You can find one in s://seldon-models/scv2/samples/mlserver_1.6.0/huggingface-text-gen-custom-gpt2
. However, you may need to request more memory!
The Model Name
is linked to the name described in the model-settings.json
file, located in the Google Cloud Storage location. Changing the name in the JSON file would also require changing the Model Name
, and vice versa.