Introduction

In this notebook, we demonstrate how BlazingText supports hosting of pre-trained Text Classification and Word2Vec models FastText models. BlazingText is a GPU accelerated version of FastText. FastText is a shallow Neural Network model used to perform both word embedding generation (unsupervised) and text classification (supervised). BlazingText uses custom CUDA kernels to accelerate the training process of FastText but the underlying algorithm is same for both the algorithms. Therefore, if you have a model trained with FastText or if one of the pre-trained models made available by FastText team is sufficient for your use case, then you can take advantage of Hosting support for BlazingText to setup SageMaker endpoints for realtime predictions using FastText models. It can help you avoid to train with BlazingText algorithm if your use-case is covered by the pre-trained models available from FastText.

To start the proceedings, we will specify few of the important parameter like IAM Role and S3 bucket location which is required for SageMaker to facilitate model hosting. SageMaker Python SDK helps us to retrieve the IAM role and also helps you to operate easily with S3 resources.

[1]:
import sagemaker
from sagemaker import get_execution_role
import boto3
import json

sess = sagemaker.Session()

role = get_execution_role()
print(
    role
)  # This is the role that SageMaker would use to leverage AWS resources (S3, CloudWatch) on your behalf

bucket = sess.default_bucket()  # Replace with your own bucket name if needed
print(bucket)
prefix = "fasttext/pretrained"  # Replace with the prefix under which you want to store the data if needed
arn:aws:iam::688520471316:role/hongshan-sagemaker-experiment
sagemaker-us-west-2-688520471316
[2]:
region_name = boto3.Session().region_name
[3]:
container = sagemaker.amazon.amazon_estimator.image_uris.retrieve("blazingtext", region_name, "1")
print("Using SageMaker BlazingText container: {} ({})".format(container, region_name))
Using SageMaker BlazingText container: 433757028032.dkr.ecr.us-west-2.amazonaws.com/blazingtext:1 (us-west-2)

Hosting the Language Idenfication model by FastText

For the example, we will leverage the pre-trained model available by FastText for Language Identification. Language Identification is the first step of many NLP applications where after the language of the input text is identified, specific models for that language needs to be applied for various other downstream tasks. Language Identification underneath is a Text Classification model which uses the language IDs as the class labels and hence FastText can be directly used for the training. FastText pretrained language model supports identification of 176 different languages.

Here we will download the Language Identification (Text Classification) model [1] from FastText website.

[1] A. Joulin, E. Grave, P. Bojanowski, T. Mikolov, Bag of Tricks for Efficient Text Classification

[4]:
!wget -O model.bin https://dl.fbaipublicfiles.com/fasttext/supervised-models/lid.176.bin
--2021-06-03 18:17:00--  https://dl.fbaipublicfiles.com/fasttext/supervised-models/lid.176.bin
Resolving dl.fbaipublicfiles.com (dl.fbaipublicfiles.com)... 172.67.9.4, 104.22.75.142, 104.22.74.142, ...
Connecting to dl.fbaipublicfiles.com (dl.fbaipublicfiles.com)|172.67.9.4|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: 131266198 (125M) [application/octet-stream]
Saving to: ‘model.bin’

model.bin           100%[===================>] 125.18M  39.9MB/s    in 3.3s

2021-06-03 18:17:04 (38.4 MB/s) - ‘model.bin’ saved [131266198/131266198]

Next we will tar the model and upload it to S3 with the help of utilities available from Python SDK. We’ll delete the local copies of the data as it’s not required anymore.

[5]:
!tar -czvf langid.tar.gz model.bin
model_location = sess.upload_data("langid.tar.gz", bucket=bucket, key_prefix=prefix)
!rm langid.tar.gz model.bin
model.bin

Creating SageMaker Inference Endpoint

Next we’ll create a SageMaker inference endpoint with the BlazingText container. This endpoint will be compatible with the pre-trained models available from FastText and can be used for inference directly without any modification. The inference endpoint works with content-type of application/json.

[6]:
lang_id = sagemaker.Model(
    image_uri=container, model_data=model_location, role=role, sagemaker_session=sess
)
lang_id.deploy(initial_instance_count=1, instance_type="ml.m5.xlarge")

from sagemaker.deserializers import JSONDeserializer
from sagemaker.serializers import JSONSerializer

predictor = sagemaker.Predictor(
    endpoint_name=lang_id.endpoint_name,
    sagemaker_session=sess,
    serializer=JSONSerializer(),
    deserializer=JSONDeserializer(),
)
----!

Next we’ll pass few sentences from various languages to the endpoint to verify that the language identification works as expected.

[7]:
sentences = [
    "hi which language is this?",
    "mon nom est Pierre",
    "Dem Jungen gab ich einen Ball.",
    "আমি বাড়ি যাবো.",
]
payload = {"instances": sentences}
[8]:
predictions = predictor.predict(payload)
print(predictions)
[{'label': ['__label__en'], 'prob': [0.9948582053184509]}, {'label': ['__label__fr'], 'prob': [0.9984669089317322]}, {'label': ['__label__de'], 'prob': [0.9946573972702026]}, {'label': ['__label__bn'], 'prob': [0.9997219443321228]}]

FastText expects the class label to be prefixed by __label__ and that’s why when we are performing inference with pre-trained model provided by FastText, we can see that the output label is prefixed with __label__. With a little preprocessing, we can strip the __label__ prefix from the response.

[9]:
import copy

predictions_copy = copy.deepcopy(
    predictions
)  # Copying predictions object because we want to change the labels in-place
for output in predictions_copy:
    output["label"] = output["label"][0][9:].upper()  # __label__ has length of 9

print(predictions_copy)
[{'label': 'EN', 'prob': [0.9948582053184509]}, {'label': 'FR', 'prob': [0.9984669089317322]}, {'label': 'DE', 'prob': [0.9946573972702026]}, {'label': 'BN', 'prob': [0.9997219443321228]}]

Stop / Close the Endpoint (Optional)

Finally, we should delete the endpoint before we close the notebook if we don’t need to keep the endpoint running for serving realtime predictions.

[10]:
predictor.delete_endpoint()

Similarly, we can host any pre-trained FastText word2vec model using SageMaker BlazingText hosting.