Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About training/inference speed improvements of TiTok #86

Open
qyangcv opened this issue Feb 21, 2025 · 0 comments
Open

About training/inference speed improvements of TiTok #86

qyangcv opened this issue Feb 21, 2025 · 0 comments

Comments

@qyangcv
Copy link

qyangcv commented Feb 21, 2025

Thank you for the great job! I just have some doubt about the efficiency of TiTok.

In the paper of TiTok, throughput (samples/s/gpus) of TiTok is basically larger than MaskGIT. However, take TiTok-L-32(32 tokens) for example, quantized latent tokens(32) is concatenated with mask tokens(256) to make sure decoder have enough tokens to generate image, so the tokens fed into decoder should be 32+256=288, larger than MaskGIT(256 tokens). Since time complexity of transformer is O(N^2),I just wonder why TiTok is faster than MaskGIT.

I'm new to the area of image generation, so I would be much grateful for your reply. If I have any misunderstanding about TiTok, please point it out.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant