Skip to content

Commit 39261f4

Browse files
committed
Add DreamVideo and ModelScopeT2V
1 parent dabec27 commit 39261f4

File tree

2 files changed

+208
-0
lines changed

2 files changed

+208
-0
lines changed

DreamVideo.yaml

Lines changed: 104 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,104 @@
1+
---
2+
# Thank you for contributing!
3+
# In filling out this yaml file, please follow the criteria as described here:
4+
# https://osai-index.eu/contribute
5+
6+
# You're free to build on this work and reuse the data. It is licensed under CC-BY 4.0, with the
7+
# stipulation that attribution should come in the form of a link to https://osai-index.eu/
8+
# and a citation to the peer-reviewed paper in which the dataset & criteria were published:
9+
10+
# Liesenfeld, A. and Dingemanse, M., 2024. Rethinking open source generative AI: open-washing and the EU AI Act. In Proceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency (pp. 1774-1787).
11+
12+
# Organization tags:
13+
# - National origin: China
14+
# - Contributor type: Non-academic (Chinese Big Tech)
15+
16+
system:
17+
name: DreamVideo
18+
link: https://dreamvideo-t2v.github.io/
19+
type: video
20+
performanceclass: limited
21+
basemodelname: ModelScopeT2V-V1.5
22+
endmodelname: DreamVideo
23+
endmodellicense: Unknown
24+
releasedate: 2024-04
25+
notes: Video-generation model with customized subject and motion.
26+
27+
org:
28+
name: Tongyi Lab
29+
link: https://careers-tongyi.alibaba.com/home
30+
notes: Tonyi Lab, a lab under Alibaba.
31+
32+
# availability:
33+
datasources_basemodel:
34+
class: partial
35+
link: ["https://arxiv.org/abs/2210.08402", "https://arxiv.org/pdf/2308.06571"]
36+
notes: ModelScopeT2V's data sources are not traceable for V1.5, its underlying model Stable Diffusion is also slightly problematic.
37+
38+
datasources_endmodel:
39+
class: open
40+
link: https://arxiv.org/pdf/2312.04433
41+
notes: "For subject customization, we select subjects from image customization papers for a total of 20 customized subjects, including 9 pets and 11 objects. For motion customization, we collect a dataset of 30 motion patterns from the Internet, the UCF101 dataset, the UCF Sports Action dataset, and the DAVIS dataset. We also provide 42 text prompts used for extensive experimental validation, where the prompts are designed to generate new motions of subjects, new contexts of subjects and motions, and etc."
42+
43+
weights_basemodel:
44+
class: open
45+
link: https://modelscope.cn/models/iic/text-to-video-synthesis/summary
46+
notes: Model made available through ModelScope.
47+
48+
weights_endmodel:
49+
class: open
50+
link: https://modelscope.cn/models/iic/dreamvideo-t2v/summary
51+
notes: Model made available through ModelScope.
52+
53+
trainingcode:
54+
class: open
55+
link: https://github.com/ali-vilab/VGen
56+
notes: Code made available on GitHub.
57+
58+
# documentation:
59+
code:
60+
class: open
61+
link: https://github.com/ali-vilab/VGen
62+
notes: Repo thoroughly documented.
63+
64+
hardware_architecture:
65+
class: open
66+
link: ["https://arxiv.org/pdf/2312.04433", "https://github.com/ali-vilab/VGen/blob/main/configs/t2v_train.yaml"]
67+
notes: Training setup disclosed in paper, config published on GitHub.
68+
69+
preprint:
70+
class: open
71+
link: https://arxiv.org/pdf/2312.04433
72+
notes: Preprint published on arXiv.
73+
74+
paper:
75+
class: closed
76+
link:
77+
notes: No peer-reviewed paper found.
78+
79+
modelcard:
80+
class: partial
81+
link: https://modelscope.cn/models/iic/dreamvideo-t2v/summary
82+
notes: Model card gives limited information.
83+
84+
datasheet:
85+
class: closed
86+
link:
87+
notes: No datasheet found.
88+
89+
# access:
90+
package:
91+
class: closed
92+
link:
93+
notes: No package found.
94+
95+
api:
96+
class: closed
97+
link:
98+
notes: No API found.
99+
metaprompts:
100+
101+
licenses:
102+
class: closed
103+
link:
104+
notes: No license found.

ModelScopeT2V.yaml

Lines changed: 104 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,104 @@
1+
---
2+
# Thank you for contributing!
3+
# In filling out this yaml file, please follow the criteria as described here:
4+
# https://osai-index.eu/contribute
5+
6+
# You're free to build on this work and reuse the data. It is licensed under CC-BY 4.0, with the
7+
# stipulation that attribution should come in the form of a link to https://osai-index.eu/
8+
# and a citation to the peer-reviewed paper in which the dataset & criteria were published:
9+
10+
# Liesenfeld, A. and Dingemanse, M., 2024. Rethinking open source generative AI: open-washing and the EU AI Act. In Proceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency (pp. 1774-1787).
11+
12+
# Organization tags:
13+
# - National origin: China
14+
# - Contributor type: Non-academic (Chinese Big Tech)
15+
16+
system:
17+
name: ModelScopeT2V
18+
link: https://modelscope.cn/models/iic/text-to-video-synthesis/summary
19+
type: video
20+
performanceclass: limited
21+
basemodelname: Stable Diffusion (version unknown)
22+
endmodelname: ModelScopeT2V-V1.5
23+
endmodellicense: CC-BY-NC-ND
24+
releasedate: 2023-02
25+
notes: Video-generation model based on Stable Diffusion.
26+
27+
org:
28+
name: Tongyi Lab
29+
link: https://careers-tongyi.alibaba.com/home
30+
notes: Tonyi Lab, a lab under Alibaba.
31+
32+
# availability:
33+
datasources_basemodel:
34+
class: partial
35+
link: https://arxiv.org/abs/2210.08402
36+
notes: Originally trained on LAION-5B, which is ostensibly open. In general the language used is not very clear, however.
37+
38+
datasources_endmodel:
39+
class: partial
40+
link: https://arxiv.org/pdf/2308.06571
41+
notes: Trained on LAION-5B, WebVid, and MSR-VTT. Unclear what data was used to train V1.5.
42+
43+
weights_basemodel:
44+
class: partial
45+
link: https://huggingface.co/stabilityai/stable-diffusion-3.5-large
46+
notes: Available through HuggingFace, however requires providing contact information and agreeing to a license agreement. Version is also unknown.
47+
48+
weights_endmodel:
49+
class: open
50+
link: https://modelscope.cn/models/iic/text-to-video-synthesis/summary
51+
notes: Model made available through ModelScope.
52+
53+
trainingcode:
54+
class: open
55+
link: https://github.com/ali-vilab/VGen
56+
notes: Code made available on GitHub.
57+
58+
# documentation:
59+
code:
60+
class: open
61+
link: https://github.com/ali-vilab/VGen
62+
notes: Repo thoroughly documented.
63+
64+
hardware_architecture:
65+
class: open
66+
link: ["https://arxiv.org/pdf/2308.06571", "https://github.com/ali-vilab/VGen/blob/main/configs/t2v_train.yaml"]
67+
notes: Training setup disclosed in paper, config published on GitHub.
68+
69+
preprint:
70+
class: open
71+
link: https://arxiv.org/pdf/2308.06571
72+
notes: Preprint available on arXiv.
73+
74+
paper:
75+
class: closed
76+
link:
77+
notes: No peer-reviewed paper found.
78+
79+
modelcard:
80+
class: partial
81+
link: https://modelscope.cn/models/iic/text-to-video-synthesis/summary
82+
notes: Model card primarily contain information about inference, with some description of training data.
83+
84+
datasheet:
85+
class: closed
86+
link:
87+
notes: No datasheet found.
88+
89+
# access:
90+
package:
91+
class: closed
92+
link:
93+
notes: No package found.
94+
95+
api:
96+
class: closed
97+
link:
98+
notes: No API found.
99+
metaprompts:
100+
101+
licenses:
102+
class: open
103+
link: https://modelscope.cn/models/iic/text-to-video-synthesis/summary
104+
notes: CC-BY-NC-ND

0 commit comments

Comments
 (0)