t5-model
There are 162 repositories under t5-model topic.
ViLongT5
LongT5-based model pre-trained on a large amount of unlabeled Vietnamese news texts and fine-tuned with ViMS and VMDS collections
Text-Summarizers
TF-IDF based Extractive & BERT based Abstractive text summarizer with an Interactive GUI for ease of use.
youtube-video-summary
Creates summary for YouTube video based on subtitles
t5-pegasus-pytorch
追一科技开源的t5-pegasus的Pytorch实现
Coursera_Natural_Language_Processing_Specialization
Implementation of state-of-the-art NLP models using transformers for tasks including machine translation, text-summarization, chatbots, and question answering.
English-Grammar-Error-Correction
The project focuses on leveraging state-of-the-art natural language processing techniques, including the T5 model and a custom Encoder-Decoder architecture, to automatically detect and correct grammatical errors in written English text.
Seq2Seq-to-TorchScript
This repository contains scripts to export T5 model to torchscript or onnx.
Paraphrasing-Generation-T5
Training paraphasing using huggingface T5
Context2Dialogue
A Dialogue Context Generator based on Text or Image
mt5-lang-detect
Fine-Tuning the Multilingual Text-To-Text Transfer Transformer (MT5) for Predicting The Language Of The Given Text
Question-MCQ-_Answer_Generation
Building this project to generate MCQ Questions from any type of text and generate answers and distractors for it.
topic2poem
End-to-End Conditional Poetry Generation
MedicalQuestionAnsweringSystem
Medical Question Answering System using T5 model, fine-tuned on PubMed data with BLEU, ROUGE, and BERTScore evaluation.
BrevityAI
BrevityAI 🚀simplifies understanding lengthy documents by generating concise, accurate summaries. Built with T5, Streamlit, and Python, it supports PDF and JSON inputs with customizable options for summary length and brevity.
Finetuning-LLMs
Finetuning Large Language Models
Unified-Language-Processing-with-ERNIE-T5-and-Pegasus
In this we explore the large language models like Pegasus ,ERNIE and T5 Large
LLM-Prompt-Recovery
ECE-5424 Advanced Machine Learning Final Project - LLM Prompt Recovery task
Abstractive-Text-Summarization-Using-RNN-and-Transformers
This repository contains implementations of abstractive text summarization using RNN ,RNN with Reinforcement learning and Transformer architectures.
SummarizerX.AI
Welcome to our Summarizer App! This tool is designed to help you efficiently condense lengthy articles and text into concise summaries, making it easier for you to grasp the main points without spending too much time reading.
linguistically_informed_t5
[LREC-COLING 2024] Materials for the paper "Linguistic Knowledge Can Enhance Encoder-Decoder Models (If You Let It)"
text-detoxification
This repo addresses the text detoxification task that is transforming the toxic style text into the text with the same meaning but with neutral style.
english-hawaiian-pidgin-translator
Finetuned FLAN-T5 to translate English to Hawaiian Pidgin
COMET-ATOMIC-En-Zh
COMET-ATOMIC ja's translate demonstration in English and Chinese, A commonsense knowledge graph on events translate from Waseda University.
NLP-Explanations-for-NLI
Fine-tuning and evaluation of a language model for explanation generation of natural language inference. Fine-tuning scripts for a pre-trained T5 model supporting both full model fine-tuning as well as LoRA are included in this repository.
Summarization-T5
Extract the most important paragraph from news articles (Abstractive summarization for now)
Text-Summarization-using-T5
Text Summarization using Pre-Trained T5 on News Dataset
T5_pretraining_finetuning
Continue T5 MLM pre-training on verbalized ConceptNet and fine-tune for commonsense question-answering
title-summary-generation
This project is regarding generating the title and summary from the input paragraph. In this we have used T-5 Model which is text to text generation model.
Czech-T5-Base-Model
This is the t5 base model for the Czech that is based on the smaller version of the google/mt5-base model. To make this model, I retained only the Czech and some of the English embeddings from the original multilingual model.
question-answering-t5-pytorch-lightning
Question Answering for English and Portuguese, using the T5 model, Transformers and Pytorch lightning. Training, inference and evaluation scripts included.
T5-Fine-tune-Wikihow-XSum
This is a Pytorch implementation of a summarization model that is fine-tuned on the top of Google-T5 pre-trained model.
MCQs-DistractorsGenerator-NLP
An end-to-end solution on how MCQs can be generated using T5 transformer model, word embeddings and decoding strategies
Text-Summarization-Using-T5
perfroming abstractive text summariztion task using T5, and serving it via REST API
SearchEngine
A Searching and Summarizing Engine leveraging a custom-built search engine for news keyword searching, and a pre-trained transformers-based T5 Model, fine-tuned on news text and summary data to achieve state-of-the-art results on text summarization
Adversarial-Conversational-Shaping
This repository contains the final group project my team and I worked on for the COMP0087 module (UCL).
NLP_mBART_mT5_translation
Polyglot Power: mBART & mT5 Translation Toolkit mBart50 multilanguage translation plus docker image and bulk pdf translation, also older English to german translator local/server using T5 model and german to polish english to spanish MarianMT transformer for file translations.