You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have done my due diligence in trying to find the answer myself.
Topic
Other / All
Question
One place to use zero_token still overrides it later. So could any help to explain when it's used? Should the delayed audio token be filled with initial_token or zero_token? As input or output? Thanks.
The delayed audio should be filled by the initial tokens (in the released version of moshi: 32000 for text, 2048 for audio). zero is actually only used for some full_like in this _get__initial_token function - most likely this is because we extracted this code from a larger codebase and didn't take the time to clean this bit.
Thanks for the reply. I just though zero_token works together with zero_idx in ScaledEmbedding to let Helium could mask out zero codes to still be able to generate text after per-training stage.
Due diligence
Topic
Other / All
Question
One place to use zero_token still overrides it later. So could any help to explain when it's used? Should the delayed audio token be filled with initial_token or zero_token? As input or output? Thanks.
moshi/moshi/moshi/models/lm.py
Lines 246 to 260 in 67f594d
The text was updated successfully, but these errors were encountered: