diff --git a/README.md b/README.md index 186a0cb..8ba3139 100644 --- a/README.md +++ b/README.md @@ -16,51 +16,38 @@ This operator extracts features for video or text with [Frozen In Time](https:// Load a video from path './demo_video.mp4' to generate a video embedding. -Read the text 'kids feeding and playing with the horse' to generate a text embedding. - - *Write the pipeline in simplified style*: - - Encode video (default): ```python -import towhee -towhee.dc(['./demo_video.mp4']) \ - .video_decode.ffmpeg(sample_type='uniform_temporal_subsample', args={'num_samples': 4}) \ - .runas_op(func=lambda x: [y for y in x]) \ - .video_text_embedding.frozen_in_time(model_name='frozen_in_time_base_16_244', modality='video', device='cpu') \ - .show() +from towhee.dc2 import pipe, ops, DataCollection -``` - +p = ( + pipe.input('video_path') \ + .map('video_path', 'flame_gen', ops.video_decode.ffmpeg(sample_type='uniform_temporal_subsample', args={'num_samples': 4})) \ + .map('flame_gen', 'flame_list', lambda x: [y for y in x]) \ + .map('flame_list', 'vec', ops.video_text_embedding.frozen_in_time(model_name='frozen_in_time_base_16_244', modality='video', device='cpu')) \ + .output('video_path', 'flame_list', 'vec') +) -- Encode text: -```python -import towhee +DataCollection(p('./demo_video.mp4')).show() -towhee.dc(['kids feeding and playing with the horse']) \ - .video_text_embedding.frozen_in_time(model_name='frozen_in_time_base_16_244', modality='text', device='cpu') \ - .show() ``` - + -*Write a same pipeline with explicit inputs/outputs name specifications:* +Read the text 'kids feeding and playing with the horse' to generate a text embedding. +- Encode text: ```python -import towhee - -towhee.dc['path'](['./demo_video.mp4']) \ - .video_decode.ffmpeg['path', 'frames'](sample_type='uniform_temporal_subsample', args={'num_samples': 4}) \ - .runas_op['frames', 'frames'](func=lambda x: [y for y in x]) \ - .video_text_embedding.frozen_in_time['frames', 'vec'](model_name='frozen_in_time_base_16_244', modality='video', device='cpu') \ - .select['path', 'vec']() \ - .show(formatter={'path': 'video_path'}) - -towhee.dc['text'](["kids feeding and playing with the horse"]) \ - .video_text_embedding.frozen_in_time['text','vec'](model_name='frozen_in_time_base_16_244', modality='text', device='cpu') \ - .select['text', 'vec']() \ - .show() +from towhee.dc2 import pipe, ops, DataCollection + +p = ( + pipe.input('text') \ + .map('text', 'vec', ops.video_text_embedding.frozen_in_time(model_name='frozen_in_time_base_16_244', modality='text', device='cpu')) \ + .output('text', 'vec') +) + +DataCollection(p('kids feeding and playing with the horse')).show() ``` - - +
diff --git a/result1.png b/result1.png deleted file mode 100644 index 56ad2ed..0000000 Binary files a/result1.png and /dev/null differ diff --git a/result2.png b/result2.png deleted file mode 100644 index 0422248..0000000 Binary files a/result2.png and /dev/null differ diff --git a/result3.png b/result3.png deleted file mode 100644 index 98d19e4..0000000 Binary files a/result3.png and /dev/null differ diff --git a/result4.png b/result4.png deleted file mode 100644 index 7558b7b..0000000 Binary files a/result4.png and /dev/null differ diff --git a/text_emb_result.png b/text_emb_result.png new file mode 100644 index 0000000..9506b41 Binary files /dev/null and b/text_emb_result.png differ diff --git a/video_emb_result.png b/video_emb_result.png new file mode 100644 index 0000000..13ef52a Binary files /dev/null and b/video_emb_result.png differ