logo
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
Readme
Files and versions

Updated 2 years ago

text-embedding

Text Embedding with Transformers

author: Jael Gu


Desription

A text embedding operator takes a sentence, paragraph, or document in string as an input and output an embedding vector in ndarray which captures the input's core semantic elements. This operator is implemented with pretrained models from Huggingface Transformers.


Code Example

Use the pretrained model 'distilbert-base-cased' to generate a text embedding for the sentence "Hello, world.".

Write the pipeline:

from towhee import dc


dc.stream(["Hello, world."]) \
  .text_embedding.transformers(model_name="distilbert-base-cased") \
  .to_list()


Factory Constructor

Create the operator via the following factory method

text_embedding.transformers(model_name="bert-base-uncased")

Parameters:

model_name: str

The model name in string. You can get the list of supported model names by calling get_model_list from auto_transformers.py.


Interface

The operator takes a text in string as input. It loads tokenizer and pre-trained model using model name. and then return text embedding in ndarray.

Parameters:

text: str

The text in string.

Returns:

numpy.ndarray

The text embedding extracted by model.

Jael Gu 518c0b8737 Update README 11 Commits
file-icon .gitattributes
1.1 KiB
download-icon
Initial commit 3 years ago
file-icon README.md
1.4 KiB
download-icon
Update README 2 years ago
file-icon __init__.py
709 B
download-icon
Debug 3 years ago
file-icon auto_transformers.py
6.8 KiB
download-icon
Update 3 years ago
file-icon requirements.txt
42 B
download-icon
Refactor operator 3 years ago