You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
hello, haohe. I really appreciate your work! Thank you for your kindness of open sourcing.
In the learning of the training code, I can not find the training of GPT2. In the original paper, embeddings of different modalities were fed into GPT2. But the released code seems to directly use clap embeddings and film to fuse the LOA. Or have I missed some details? Hope you can solve my confusion!
The text was updated successfully, but these errors were encountered:
hello, haohe. I really appreciate your work! Thank you for your kindness of open sourcing.
In the learning of the training code, I can not find the training of GPT2. In the original paper, embeddings of different modalities were fed into GPT2. But the released code seems to directly use clap embeddings and film to fuse the LOA. Or have I missed some details? Hope you can solve my confusion!
The text was updated successfully, but these errors were encountered: