ModelTC/QLLM

Is there a bug?

sunnyxiaohu opened this issue · 1 comments

According to line40, 'attn_output = self.pv_matmul(attn_weights, value_states)' equals to 'attn_output = torch.matmul(attn_weights, value_states)'.