Skip to content

Commit 1481c1d

Browse files
authored
Update README_zh.md
1 parent f03d851 commit 1481c1d

File tree

1 file changed

+16
-18
lines changed

1 file changed

+16
-18
lines changed

README_zh.md

Lines changed: 16 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@
99
# **HunyuanVideo-I2V** 🌅
1010

1111
<div align="center">
12-
<a href="https://github.com/Tencent/HunyuanVideo-I2V"><img src="https://img.shields.io/static/v1?label=HunyuanVideo-I2V 代码&message=Github&color=blue"></a> &ensp;
12+
<a href="https://github.com/Tencent-Hunyuan/HunyuanVideo-I2V"><img src="https://img.shields.io/static/v1?label=HunyuanVideo-I2V 代码&message=Github&color=blue"></a> &ensp;
1313
<a href="https://aivideo.hunyuan.tencent.com"><img src="https://img.shields.io/static/v1?label=项目主页&message=Web&color=green"></a> &ensp;
1414
<a href="https://video.hunyuan.tencent.com"><img src="https://img.shields.io/static/v1?label=在线体验&message=Web&color=green"></a>
1515
</div>
@@ -27,7 +27,7 @@
2727

2828
-----
2929

30-
继我们成功开源[HunyuanVideo](https://github.com/Tencent/HunyuanVideo)后,我们很高兴推出[HunyuanVideo-I2V](https://github.com/Tencent/HunyuanVideo-I2V),一个新的图像到视频生成框架,加速开源社区的探索!
30+
继我们成功开源[HunyuanVideo](https://github.com/Tencent-Hunyuan/HunyuanVideo)后,我们很高兴推出[HunyuanVideo-I2V](https://github.com/Tencent-Hunyuan/HunyuanVideo-I2V),一个新的图像到视频生成框架,加速开源社区的探索!
3131

3232
本仓库包含官方PyTorch模型定义、预训练权重及推理/采样代码。更多可视化效果请访问[项目主页](https://aivideo.hunyuan.tencent.com)。同时,我们发布了LoRA训练代码,用于定制化特效生成,可创建更有趣的视频效果。
3333

@@ -37,7 +37,7 @@
3737
* 2025年03月13日: 🚀 开源 HunyuanVideo-I2V 多卡并行推理代码,由[xDiT](https://github.com/xdit-project/xDiT)提供。
3838
* 2025年03月11日: 🎉 在修复bug后我们更新了lora的训练和推理代码。
3939
* 2025年03月07日: 🔥 我们已经修复了开源版本中导致ID变化的bug,请尝试[HunyuanVideo-I2V](https://huggingface.co/tencent/HunyuanVideo-I2V)新的模型权重,以确保首帧完全视觉一致性,并制作更高质量的视频。
40-
* 2025年03月06日: 👋 发布HunyuanVideo-I2V的推理代码和模型权重。[下载地址](https://github.com/Tencent/HunyuanVideo-I2V/blob/main/ckpts/README.md)
40+
* 2025年03月06日: 👋 发布HunyuanVideo-I2V的推理代码和模型权重。[下载地址](https://github.com/Tencent-Hunyuan/HunyuanVideo-I2V/blob/main/ckpts/README.md)
4141

4242
## 🎥 演示
4343
### I2V 示例
@@ -67,7 +67,7 @@
6767

6868
- ComfyUI (支持FP8推理、V2V和IP2V生成): [ComfyUI-HunyuanVideoWrapper](https://github.com/kijai/ComfyUI-HunyuanVideoWrapper) by [Kijai](https://github.com/kijai)
6969
- HunyuanVideoGP (针对低性能GPU的版本): [HunyuanVideoGP](https://github.com/deepbeepmeep/HunyuanVideoGP) by [DeepBeepMeep](https://github.com/deepbeepmeep)
70-
- xDiT 兼容性改进: [兼容性改进](https://github.com/Tencent/HunyuanVideo-I2V/issues/36#issuecomment-2728068507) by [pftq](https://github.com/pftq) and [xibosun](https://github.com/xibosun)
70+
- xDiT 兼容性改进: [兼容性改进](https://github.com/Tencent-Hunyuan/HunyuanVideo-I2V/issues/36#issuecomment-2728068507) by [pftq](https://github.com/pftq) and [xibosun](https://github.com/xibosun)
7171

7272
## 📑 开源计划
7373
- HunyuanVideo-I2V(图像到视频模型)
@@ -111,7 +111,7 @@
111111
---
112112

113113
## **HunyuanVideo-I2V 整体架构**
114-
基于[HunyuanVideo](https://github.com/Tencent/HunyuanVideo)强大的视频生成能力,我们将其扩展至图像到视频生成任务。为此,我们采用首帧Token替换方案,有效重构并融合参考图像信息至视频生成流程中。
114+
基于[HunyuanVideo](https://github.com/Tencent-Hunyuan/HunyuanVideo)强大的视频生成能力,我们将其扩展至图像到视频生成任务。为此,我们采用首帧Token替换方案,有效重构并融合参考图像信息至视频生成流程中。
115115

116116
由于我们使用预训练的Decoder-Only架构多模态大语言模型(MLLM)作为文本编码器,可用于显著增强模型对输入图像语义内容的理解能力,并实现图像与文本描述信息的深度融合。具体而言,输入图像经MLLM处理后生成语义图像tokens,这些tokens与视频隐空间tokens拼接,实现跨模态的全注意力计算。
117117

@@ -138,7 +138,7 @@
138138

139139
首先克隆仓库:
140140
```shell
141-
git clone https://github.com/tencent/HunyuanVideo-I2V
141+
git clone https://github.com/Tencent-Hunyuan/HunyuanVideo-I2V
142142
cd HunyuanVideo-I2V
143143
```
144144

@@ -191,7 +191,7 @@ docker run -itd --gpus all --init --net=host --uts=host --ipc=host --name hunyua
191191

192192
## 🔑 单 GPU 推理
193193

194-
类似于 [HunyuanVideo](https://github.com/Tencent/HunyuanVideo),HunyuanVideo-I2V 支持高分辨率视频生成,分辨率最高可达 720P,视频长度最高可达 129 帧(5 秒)。
194+
类似于 [HunyuanVideo](https://github.com/Tencent-Hunyuan/HunyuanVideo),HunyuanVideo-I2V 支持高分辨率视频生成,分辨率最高可达 720P,视频长度最高可达 129 帧(5 秒)。
195195
### 使用图生视频模型的建议
196196

197197
- **使用简短的提示**:为了有效地引导模型的生成,请保持提示简短且直截了当。
@@ -415,13 +415,11 @@ xDiT 并行推理加速如下表所示。
415415
如果您发现 [HunyuanVideo](https://arxiv.org/abs/2412.03603) 对您的研究和应用有所帮助,请使用以下 BibTeX 引用:
416416

417417
```BibTeX
418-
@misc{kong2024hunyuanvideo,
419-
title={HunyuanVideo: A Systematic Framework For Large Video Generative Models},
420-
author={Weijie Kong, Qi Tian, Zijian Zhang, Rox Min, Zuozhuo Dai, Jin Zhou, Jiangfeng Xiong, Xin Li, Bo Wu, Jianwei Zhang, Kathrina Wu, Qin Lin, Aladdin Wang, Andong Wang, Changlin Li, Duojun Huang, Fang Yang, Hao Tan, Hongmei Wang, Jacob Song, Jiawang Bai, Jianbing Wu, Jinbao Xue, Joey Wang, Junkun Yuan, Kai Wang, Mengyang Liu, Pengyu Li, Shuai Li, Weiyan Wang, Wenqing Yu, Xinchi Deng, Yang Li, Yanxin Long, Yi Chen, Yutao Cui, Yuanbo Peng, Zhentao Yu, Zhiyu He, Zhiyong Xu, Zixiang Zhou, Zunnan Xu, Yangyu Tao, Qinglin Lu, Songtao Liu, Dax Zhou, Hongfa Wang, Yong Yang, Di Wang, Yuhong Liu, and Jie Jiang, along with Caesar Zhong},
421-
year={2024},
422-
archivePrefix={arXiv preprint arXiv:2412.03603},
423-
primaryClass={cs.CV},
424-
url={https://arxiv.org/abs/2412.03603},
418+
@article{kong2024hunyuanvideo,
419+
title={Hunyuanvideo: A systematic framework for large video generative models},
420+
author={Kong, Weijie and Tian, Qi and Zhang, Zijian and Min, Rox and Dai, Zuozhuo and Zhou, Jin and Xiong, Jiangfeng and Li, Xin and Wu, Bo and Zhang, Jianwei and others},
421+
journal={arXiv preprint arXiv:2412.03603},
422+
year={2024}
425423
}
426424
```
427425

@@ -433,11 +431,11 @@ HunyuanVideo 的开源离不开诸多开源工作,这里我们特别感谢 [SD
433431

434432
<!-- ## Star 趋势
435433
436-
<a href="https://star-history.com/#Tencent/HunyuanVideo&Date">
434+
<a href="https://star-history.com/#Tencent-Hunyuan/HunyuanVideo&Date">
437435
<picture>
438-
<source media="(prefers-color-scheme: dark)" srcset="https://api.star-history.com/svg?repos=Tencent/HunyuanVideo&type=Date&theme=dark" />
439-
<source media="(prefers-color-scheme: light)" srcset="https://api.star-history.com/svg?repos=Tencent/HunyuanVideo&type=Date" />
440-
<img alt="Star History Chart" src="https://api.star-history.com/svg?repos=Tencent/HunyuanVideo&type=Date" />
436+
<source media="(prefers-color-scheme: dark)" srcset="https://api.star-history.com/svg?repos=Tencent-Hunyuan/HunyuanVideo&type=Date&theme=dark" />
437+
<source media="(prefers-color-scheme: light)" srcset="https://api.star-history.com/svg?repos=Tencent-Hunyuan/HunyuanVideo&type=Date" />
438+
<img alt="Star History Chart" src="https://api.star-history.com/svg?repos=Tencent-Hunyuan/HunyuanVideo&type=Date" />
441439
</picture>
442440
</a> -->
443441

0 commit comments

Comments
 (0)