Text Generation with Custom HuggingFace Model
Last updated
Was this helpful?
Last updated
Was this helpful?
This demo helps you learn about:
Launching a pre trained custom text generation HuggingFace model in a Seldon Pipeline
Sending a text input request to get a generated text prediction
The custom HuggingFace text generation model is based on the model in the HuggingFace hub.
In the Overview page click Create new deployment.
Enter the deployment details as follows:
Name: hf-custom-tiny-stories
Namespace: seldon
Type: Seldon ML Pipeline
Configure the default predictor as follows:
Runtime: HuggingFace
Model Project: default
Model URI: gs://seldon-models/scv2/samples/mlserver_1.6.0/huggingface-text-gen-custom-tiny-stories
Storage Secret: (leave blank/none)
Click Next for the remaining steps and click Launch.
When the deployment is launched successfully, the status of the deployment becomes Available
.
Click the hf-custom-tiny-stories
deployment that you created.
In the deployment Dashboard page , click Predict in the left pane.
In the Predict REST API dialog, click Enter JSON and paste the following text:
Click Predict.
Try other or try a larger-scale model. You can find one in gs://seldon-models/scv2/samples/mlserver_1.6.0/huggingface-text-gen-custom-gpt2
. However, you may need to request more memory.