zcgzcgzcg1/MediaSum

About preprocessing before input into BART

Junpliu opened this issue · 1 comments

Hi, I noticed that utterances are truncated after 1024 tokens as described in (Part D, Implementation Details).
I want to know do I need to add role information when con-cating utterances into one sentence? or do the 1024 tokens includes the role information? Please tell me more details about the preprocessing.
Thanks so much!

The truncation happens after role information is added.