-
Notifications
You must be signed in to change notification settings - Fork 80
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hunyuan video: inference and train #831
Open
wtomin
wants to merge
137
commits into
mindspore-lab:master
Choose a base branch
from
wtomin:hunyuan-video
base: master
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Open
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
wtomin
requested review from
vigo999,
CaitinZhao,
SamitHuang and
zhanghuiyao
as code owners
January 23, 2025 03:33
wtomin
changed the title
[Draft] Hunyuan video: inference and train
Hunyuan video: inference and train
Jan 24, 2025
* fix config names * add mem monitor * update * update * debug tae attn * update * useless change * continue work from llama3_movie_pr_20241029 * add parallel test case for scheduler and fix some minor bug * add train script * move config file outside the folder * temp save * change some ops to mint * add init for text projector * fix mint * fix type * encoder ok * add image support to OS data loader * update convert script * add recompute support in PyNative * add dataloader * update train script * add OSv1.2 VAE * fixes * reconstruct tested * update readme * discard spurious frames * rename * add train * add train config * rename * rename * add dataset * trainable * add inference * fix opl loss * z 16 * fix linear-quadratic sampling * text encoders inference * allow loading sd3.5 vae pretrained weights * update convert script * add sd3 vae * add moduels for sd3 vae * update configs * temporal median init, 1p train psnr ok * add files * fix rt id * set image and crop size * add train step mode * replace interpolate for bf16 support * add validation support * add ReduceLROnPlateau * save top K checkpoints * add drop text conditioning for training * fix eval loss calculation * add model parallel * hack for model parallel * fix hack * small fixes * add temporal tile * rm comments * clean code * draft readme and update decode * add config * add readme draft * add TAE to Movie Gen * add buckets and dynamic graph support * fix dynamic shape: defualt manual pad for conv1d same pad * fix save callback and TAE scaling * Revert "fix hack" This reverts commit bf505d4. * Revert "hack for model parallel" This reverts commit 8af7437. * revert it later * small fixes * refactoring * linting * add docs * refactor TAE add latents generation other small changes * fix training with TAE latents * revert changes to OpenSora * merge with PR mindspore-lab#778 * small fix * PR fixes: - remove forced dynamic memory allocation for data transformations - purge Model Parallel functionality until it's fully tested * Update docs * Update docs * update docs and small fixes * fix TAE encoding * PR fixes: - remove unrelated code changes - update docs * small inference fix * enable `lazy_inline` enable jit_level `O2` support * small fix * small fix * enable flexible recompute * enable flexible recompute * - add train resume feature - preserve image / video orientation in transformations * ResizeCrop fix * update docs * support SP and change rms to ops.rms * Gradio demo for MovieGen (#6) * update docs and add stage 3 configs * add ZeRO-3 support to Movie Gen * add Model Parallel * add technical report * update technical report * linting * add inference without TAE and stand-alone decoding * Drop Model Parallel * improve SP support * fix checkpoint saving * fix checkpoint saving * align with PR#778 * update README * resolve comments * fix SP * small fixes and update README.md * add TAE download link * update README.md * fix imports and update README.md * update README.md * fix dynamic graph training with TAE * update configs * update configs * turn off EMA until it's fixed --------- Co-authored-by: Samit <285365963@qq.com> Co-authored-by: Mike Cheung <zhtmike@gmail.com> Co-authored-by: Nguyen Truong Hai <47595486+itruonghai@users.noreply.github.com>
suggestion for better code style and clarity
|
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
What does this PR do?
HunyuanVideo:
Fixes # (issue)
Adds # (feature)
Before submitting
What's New
. Here are thedocumentation guidelines
Who can review?
Anyone in the community is free to review the PR once the tests have passed. Feel free to tag
members/contributors who may be interested in your PR.
@xxx