/Master-Thesis-Multilingual-Longformer

Master thesis with code investigating methods for incorporating long-context reasoning in low-resource languages, without the need to pre-train from scratch. We investigated if multilingual models could inherit these properties by making it an Efficient Transformer (s.a. the Longformer architecture).

Primary LanguageJupyter NotebookMIT LicenseMIT

Watchers