diff --git a/README.md b/README.md index 16a9f76..186a0cb 100644 --- a/README.md +++ b/README.md @@ -23,8 +23,9 @@ Read the text 'kids feeding and playing with the horse' to generate a text embed - Encode video (default): ```python import towhee -towhee.glob('./archery.mp4') \ - .video_decode.ffmpeg() \ +towhee.dc(['./demo_video.mp4']) \ + .video_decode.ffmpeg(sample_type='uniform_temporal_subsample', args={'num_samples': 4}) \ + .runas_op(func=lambda x: [y for y in x]) \ .video_text_embedding.frozen_in_time(model_name='frozen_in_time_base_16_244', modality='video', device='cpu') \ .show() @@ -46,8 +47,9 @@ towhee.dc(['kids feeding and playing with the horse']) \ ```python import towhee -towhee.glob['path']('./archery.mp4') \ - .video_decode.ffmpeg['path', 'frames']() \ +towhee.dc['path'](['./demo_video.mp4']) \ + .video_decode.ffmpeg['path', 'frames'](sample_type='uniform_temporal_subsample', args={'num_samples': 4}) \ + .runas_op['frames', 'frames'](func=lambda x: [y for y in x]) \ .video_text_embedding.frozen_in_time['frames', 'vec'](model_name='frozen_in_time_base_16_244', modality='video', device='cpu') \ .select['path', 'vec']() \ .show(formatter={'path': 'video_path'}) diff --git a/demo_video.mp4 b/demo_video.mp4 new file mode 100755 index 0000000..e6fb645 Binary files /dev/null and b/demo_video.mp4 differ diff --git a/result1.png b/result1.png index 2f99f9e..90f9c32 100644 Binary files a/result1.png and b/result1.png differ diff --git a/result3.png b/result3.png index 5636810..15e3021 100644 Binary files a/result3.png and b/result3.png differ