Why use 2 time tokens?
KevinGoodman opened this issue · 2 comments
KevinGoodman commented
self.to_time_tokens = nn.Sequential(
nn.Linear(latent_dim * 4, latent_dim * 2), # 2 time tokens
Rearrange("b (r d) -> b r d", r=2),
)
In L278-L281 of model/model.py, what is the purpose of making 2 time tokens instead of just 1 time token ?
xuzheyuan624 commented
I'm also confuse about it. And in L376c = torch.cat((cond_tokens, t_tokens), dim=-2)
, why concat t tokens and cond tokens in time dimension?
anshulkc commented
@xuzheyuan624 it just concatenates the music condition and the time tokens along the sequence dimension to create a combined conditioned sequence that uses both music and time information