forked from Language-Technology-Assessment/main-database
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathDreamVideo.yaml
More file actions
104 lines (85 loc) · 3.39 KB
/
DreamVideo.yaml
File metadata and controls
104 lines (85 loc) · 3.39 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
---
# Thank you for contributing!
# In filling out this yaml file, please follow the criteria as described here:
# https://osai-index.eu/contribute
# You're free to build on this work and reuse the data. It is licensed under CC-BY 4.0, with the
# stipulation that attribution should come in the form of a link to https://osai-index.eu/
# and a citation to the peer-reviewed paper in which the dataset & criteria were published:
# Liesenfeld, A. and Dingemanse, M., 2024. Rethinking open source generative AI: open-washing and the EU AI Act. In Proceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency (pp. 1774-1787).
# Organization tags:
# - National origin: China
# - Contributor type: Non-academic (Chinese Big Tech)
system:
name: DreamVideo
link: https://dreamvideo-t2v.github.io/
type: video
performanceclass: limited
basemodelname: ModelScopeT2V-V1.5
endmodelname: DreamVideo
endmodellicense: Unknown
releasedate: 2024-04
notes: Video-generation model with customized subject and motion.
org:
name: Tongyi Lab
link: https://careers-tongyi.alibaba.com/home
notes: Tonyi Lab, a lab under Alibaba.
# availability:
datasources_basemodel:
class: partial
link: ["https://arxiv.org/abs/2210.08402", "https://arxiv.org/pdf/2308.06571"]
notes: ModelScopeT2V's data sources are not traceable for V1.5, its underlying model Stable Diffusion is also slightly problematic.
datasources_endmodel:
class: open
link: https://arxiv.org/pdf/2312.04433
notes: "For subject customization, we select subjects from image customization papers for a total of 20 customized subjects, including 9 pets and 11 objects. For motion customization, we collect a dataset of 30 motion patterns from the Internet, the UCF101 dataset, the UCF Sports Action dataset, and the DAVIS dataset. We also provide 42 text prompts used for extensive experimental validation, where the prompts are designed to generate new motions of subjects, new contexts of subjects and motions, and etc."
weights_basemodel:
class: open
link: https://modelscope.cn/models/iic/text-to-video-synthesis/summary
notes: Model made available through ModelScope.
weights_endmodel:
class: open
link: https://modelscope.cn/models/iic/dreamvideo-t2v/summary
notes: Model made available through ModelScope.
trainingcode:
class: open
link: https://github.com/ali-vilab/VGen
notes: Code made available on GitHub.
# documentation:
code:
class: open
link: https://github.com/ali-vilab/VGen
notes: Repo thoroughly documented.
hardware_architecture:
class: open
link: ["https://arxiv.org/pdf/2312.04433", "https://github.com/ali-vilab/VGen/blob/main/configs/t2v_train.yaml"]
notes: Training setup disclosed in paper, config published on GitHub.
preprint:
class: open
link: https://arxiv.org/pdf/2312.04433
notes: Preprint published on arXiv.
paper:
class: closed
link:
notes: No peer-reviewed paper found.
modelcard:
class: partial
link: https://modelscope.cn/models/iic/dreamvideo-t2v/summary
notes: Model card gives limited information.
datasheet:
class: closed
link:
notes: No datasheet found.
# access:
package:
class: closed
link:
notes: No package found.
api:
class: closed
link:
notes: No API found.
metaprompts:
licenses:
class: closed
link:
notes: No license found.