# Text Embedding with Transformers *author: Jael Gu* ## Desription A text embedding operator takes a sentence, paragraph, or document in string as an input and output an embedding vector in ndarray which captures the input's core semantic elements. This operator is implemented with pretrained models from [Huggingface Transformers](https://huggingface.co/docs/transformers). ## Code Example Use the pretrained model 'distilbert-base-cased' to generate a text embedding for the sentence "Hello, world.". *Write the pipeline*: ```python from towhee import dc dc.stream(["Hello, world."]) .text_embedding.transformers('distilbert-base-cased') .show() ``` ## Factory Constructor Create the operator via the following factory method ***text_embedding.transformers(model_name="bert-base-uncased")*** **Parameters:** ​ ***model_name***: *str* ​ The model name in string. You can get the list of supported model names by calling `get_model_list` from [auto_transformers.py](https://towhee.io/text-embedding/transformers/src/branch/main/auto_transformers.py). ## Interface The operator takes a text in string as input. It loads tokenizer and pre-trained model using model name. and then return text embedding in ndarray. **Parameters:** ​ ***text***: *str* ​ The text in string. **Returns**: ​ *numpy.ndarray* ​ The text embedding extracted by model.