A bug of torch.arange for long sequence with fp16 data type
renll opened this issue · 1 comments
renll commented
This line should use dtype=torch.long to support long sequence:
lucidrains commented
@renll thanks for identifying this bug!
renll opened this issue · 1 comments
This line should use dtype=torch.long to support long sequence:
@renll thanks for identifying this bug!