|
|
@ -33,6 +33,8 @@ towhee.dc(['./demo_video.mp4']) \ |
|
|
|
.show() |
|
|
|
|
|
|
|
``` |
|
|
|
<img src="./result1.png" width="800px"/> |
|
|
|
|
|
|
|
- Encode text: |
|
|
|
```python |
|
|
|
import towhee |
|
|
@ -41,6 +43,7 @@ towhee.dc(['kids feeding and playing with the horse']) \ |
|
|
|
.video_text_embedding.frozen_in_time(model_name='frozen_in_time_base_16_244', modality='text', device='cpu') \ |
|
|
|
.show() |
|
|
|
``` |
|
|
|
<img src="./result2.png" width="800px"/> |
|
|
|
|
|
|
|
*Write a same pipeline with explicit inputs/outputs name specifications:* |
|
|
|
|
|
|
@ -48,16 +51,19 @@ towhee.dc(['kids feeding and playing with the horse']) \ |
|
|
|
import towhee |
|
|
|
|
|
|
|
towhee.dc['path'](['./demo_video.mp4']) \ |
|
|
|
.video_decode.ffmpeg['path', 'frames'](sample_type='uniform_temporal_subsample', args={'num_samples': 4}) \ |
|
|
|
.runas_op['frames', 'frames'](func=lambda x: [y for y in x]) \ |
|
|
|
.video_text_embedding.frozen_in_time['frames', 'vec'](model_name='frozen_in_time_base_16_244', modality='video', device='cpu') \ |
|
|
|
.show() |
|
|
|
.video_decode.ffmpeg['path', 'frames'](sample_type='uniform_temporal_subsample', args={'num_samples': 4}) \ |
|
|
|
.runas_op['frames', 'frames'](func=lambda x: [y for y in x]) \ |
|
|
|
.video_text_embedding.frozen_in_time['frames', 'vec'](model_name='frozen_in_time_base_16_244', modality='video', device='cpu') \ |
|
|
|
.select['path', 'vec']() \ |
|
|
|
.show(formatter={'path': 'video_path'}) |
|
|
|
|
|
|
|
towhee.dc['text'](["kids feeding and playing with the horse"]) \ |
|
|
|
.video_text_embedding.frozen_in_time['text','vec'](model_name='frozen_in_time_base_16_244', modality='text', device='cpu') \ |
|
|
|
.select['text', 'vec']() \ |
|
|
|
.show() |
|
|
|
``` |
|
|
|
<img src="./result3.png" width="800px"/> |
|
|
|
<img src="./result4.png" width="800px"/> |
|
|
|
|
|
|
|
|
|
|
|
<br /> |
|
|
|