Text Generation with Custom HuggingFace Model
This demo helps you learn about:
Launching a pre trained custom text generation HuggingFace model in a Seldon Pipeline
Sending a text input request to get a generated text prediction
The custom HuggingFace text generation model is based on the TinyStories-1M model in the HuggingFace hub.
Create a Seldon ML Pipeline
In the Overview page click Create new deployment.
Enter the deployment details as follows:
Name:
hf-custom-tiny-storiesNamespace:
seldonType:
Seldon ML Pipeline
Configure the default predictor as follows:
Runtime:
HuggingFaceModel Project:
defaultModel URI:
gs://seldon-models/scv2/samples/mlserver_1.6.0/huggingface-text-gen-custom-tiny-storiesStorage Secret: (leave blank/none)

Default predictor spec Click Next for the remaining steps and click Launch.
When the deployment is launched successfully, the status of the deployment becomes Available.
Make Predictions
Click the
hf-custom-tiny-storiesdeployment that you created.In the deployment Dashboard page , click Predict in the left pane.
In the Predict REST API dialog, click Enter JSON and paste the following text:
{ "inputs": [{ "name": "args", "shape": [1], "datatype": "BYTES", "data": ["The brown fox jumped"] }] }Click Predict.

A screenshot showing the Predict page with the textarea pre-populated
Next steps
Try other demos or try a larger-scale model. You can find one in gs://seldon-models/scv2/samples/mlserver_1.6.0/huggingface-text-gen-custom-gpt2. However, you may need to request more memory.
Last updated
Was this helpful?