diff --git a/README.md b/README.md
index 186a0cb..8ba3139 100644
--- a/README.md
+++ b/README.md
@@ -16,51 +16,38 @@ This operator extracts features for video or text with [Frozen In Time](https://
Load a video from path './demo_video.mp4' to generate a video embedding.
-Read the text 'kids feeding and playing with the horse' to generate a text embedding.
-
- *Write the pipeline in simplified style*:
-
- Encode video (default):
```python
-import towhee
-towhee.dc(['./demo_video.mp4']) \
- .video_decode.ffmpeg(sample_type='uniform_temporal_subsample', args={'num_samples': 4}) \
- .runas_op(func=lambda x: [y for y in x]) \
- .video_text_embedding.frozen_in_time(model_name='frozen_in_time_base_16_244', modality='video', device='cpu') \
- .show()
+from towhee.dc2 import pipe, ops, DataCollection
-```
-
+p = (
+ pipe.input('video_path') \
+ .map('video_path', 'flame_gen', ops.video_decode.ffmpeg(sample_type='uniform_temporal_subsample', args={'num_samples': 4})) \
+ .map('flame_gen', 'flame_list', lambda x: [y for y in x]) \
+ .map('flame_list', 'vec', ops.video_text_embedding.frozen_in_time(model_name='frozen_in_time_base_16_244', modality='video', device='cpu')) \
+ .output('video_path', 'flame_list', 'vec')
+)
-- Encode text:
-```python
-import towhee
+DataCollection(p('./demo_video.mp4')).show()
-towhee.dc(['kids feeding and playing with the horse']) \
- .video_text_embedding.frozen_in_time(model_name='frozen_in_time_base_16_244', modality='text', device='cpu') \
- .show()
```
-
+
-*Write a same pipeline with explicit inputs/outputs name specifications:*
+Read the text 'kids feeding and playing with the horse' to generate a text embedding.
+- Encode text:
```python
-import towhee
-
-towhee.dc['path'](['./demo_video.mp4']) \
- .video_decode.ffmpeg['path', 'frames'](sample_type='uniform_temporal_subsample', args={'num_samples': 4}) \
- .runas_op['frames', 'frames'](func=lambda x: [y for y in x]) \
- .video_text_embedding.frozen_in_time['frames', 'vec'](model_name='frozen_in_time_base_16_244', modality='video', device='cpu') \
- .select['path', 'vec']() \
- .show(formatter={'path': 'video_path'})
-
-towhee.dc['text'](["kids feeding and playing with the horse"]) \
- .video_text_embedding.frozen_in_time['text','vec'](model_name='frozen_in_time_base_16_244', modality='text', device='cpu') \
- .select['text', 'vec']() \
- .show()
+from towhee.dc2 import pipe, ops, DataCollection
+
+p = (
+ pipe.input('text') \
+ .map('text', 'vec', ops.video_text_embedding.frozen_in_time(model_name='frozen_in_time_base_16_244', modality='text', device='cpu')) \
+ .output('text', 'vec')
+)
+
+DataCollection(p('kids feeding and playing with the horse')).show()
```
-
-
+
diff --git a/result1.png b/result1.png
deleted file mode 100644
index 56ad2ed..0000000
Binary files a/result1.png and /dev/null differ
diff --git a/result2.png b/result2.png
deleted file mode 100644
index 0422248..0000000
Binary files a/result2.png and /dev/null differ
diff --git a/result3.png b/result3.png
deleted file mode 100644
index 98d19e4..0000000
Binary files a/result3.png and /dev/null differ
diff --git a/result4.png b/result4.png
deleted file mode 100644
index 7558b7b..0000000
Binary files a/result4.png and /dev/null differ
diff --git a/text_emb_result.png b/text_emb_result.png
new file mode 100644
index 0000000..9506b41
Binary files /dev/null and b/text_emb_result.png differ
diff --git a/video_emb_result.png b/video_emb_result.png
new file mode 100644
index 0000000..13ef52a
Binary files /dev/null and b/video_emb_result.png differ