Skip to content

v1 #270

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 60 commits into from
Mar 29, 2025
Merged

v1 #270

merged 60 commits into from
Mar 29, 2025

Conversation

jzhang38
Copy link
Collaborator

@jzhang38 jzhang38 commented Mar 16, 2025

[ ] Attn backend (PY)
[ ] Wan text encoder (PY)
[ ] wan vae (Wei)
[ ] wan pipeline (PY & Wei)
[ ] Merge wan dit code to refactor (Will)
[ ] Clean up code & loader directory(Will)
[ ] hunyuan text encoder (Will)

SolitaryThinker and others added 30 commits March 14, 2025 19:07
Signed-off-by: <>
Co-authored-by: Will Lin <[email protected]>
Co-authored-by: Ubuntu <ubuntu@awesome-gpu-name-8-inst-2tbsnfodvpomxv4tukw2dkfgyvz.c.nv-brev-20240723.internal>
Co-authored-by: Ubuntu <ubuntu@awesome-gpu-name-9-inst-2tpydiudxfu1jg9xvpflm7oexie.c.nv-brev-20240723.internal>
return torch.stack((o1, o2), dim=-1).flatten(-2)


@CustomOp.register("rotary_embedding")
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

a little bit surprised that we need to customize for rotary embedding class?

uses_last_layer = feature_sample_layers[-1] in (len(hs_pool) - 1, -1)
if post_layer_norm is not None and uses_last_layer:
hs_pool[-1] = post_layer_norm(encoder_outputs)
return torch.cat(hs_pool, dim=-1)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

space at EoF

from fastvideo.v1.inference_args import InferenceArgs
from fastvideo.v1.logger import init_logger

from ..composed_pipeline_base import ComposedPipelineBase
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I saw a lot of relative imports throughout; can we avoid relative imports?

def __init__(self, vae) -> None:
self.vae = vae

def forward(
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

one weird thing i noticed: you only use forward context in the text_encoding.py, not anywhere else? correct? if yes -- why?

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

we also use it for DiTs (set in the denoising stage).

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

input validation being a stage sounds strange, but i guess ok currently.

@SolitaryThinker SolitaryThinker marked this pull request as ready for review March 29, 2025 22:42
@SolitaryThinker SolitaryThinker merged commit 1fee098 into main Mar 29, 2025
0 of 4 checks passed
@SolitaryThinker SolitaryThinker changed the title [do not merge] Rebased refactor v1 Mar 30, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

8 participants