transformers-models
There are 133 repositories under transformers-models topic.
alibaba/EasyRec
A framework for large scale recommendation algorithms.
kyegomez/awesome-multi-agent-papers
A compilation of the best multi-agent papers
NeuralCarver/Michelangelo
[NeurIPS 2023] Michelangelo: Conditional 3D Shape Generation based on Shape-Image-Text Aligned Latent Representation
InternLM/InternEvo
InternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencies.
abhimishra91/insight
Repository for Project Insight: NLP as a Service
microsoft/msrflute
Federated Learning Utilities and Tools for Experimentation
clip-italian/clip-italian
CLIP (Contrastive Language–Image Pre-training) for Italian
DC-research/TEMPO
The official code for "TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting (ICLR 2024)". TEMPO is one of the very first open source Time Series Foundation Models for forecasting task v1.0 version.
esnya/hf-rvc
Retrieval-based Voice Conversion (RVC) implemented with Hugging Face Transformers.
zarzouram/image_captioning_with_transformers
Pytorch implementation of image captioning using transformer-based model.
ThinamXx/Fastai
I will implement Fastai in each projects present in this repository.
jcwang123/xboundformer
[TMI 2023] XBound-Former: Toward Cross-scale Boundary Modeling in Transformers
sinanw/llm-security-prompt-injection
This project investigates the security of large language models by performing binary classification of a set of input prompts to discover malicious prompts. Several approaches have been analyzed using classical ML algorithms, a trained LLM model, and a fine-tuned LLM model.
Babelscape/ALERT
Official repository for the paper "ALERT: A Comprehensive Benchmark for Assessing Large Language Models’ Safety through Red Teaming"
simranjeet97/75DayHard_GenAI_LLM_Challenge
This repository contain my 75Day Hard Generative AI and LLM Learning Challenge.
The-Swarm-Corporation/Multi-Agent-Template-App
A radically simple, reliable, and high performance template to enable you to quickly get set up building multi-agent applications
inuwamobarak/Image-captioning-ViT
Image Captioning Vision Transformers (ViTs) are transformer models that generate descriptive captions for images by combining the power of Transformers and computer vision. It leverages state-of-the-art pre-trained ViT models and employs technique
kyegomez/DifferentialTransformer
An open source community implementation of the model from "DIFFERENTIAL TRANSFORMER" paper by Microsoft.
hammaad2002/ASRAdversarialAttacks
An ASR (Automatic Speech Recognition) adversarial attack repository.
HRSadeghi/NeuralPersianPoet
Neural Persian Poet: A sequence-to-sequence model for composing Persian poetry
The-Swarm-Corporation/NewsAgent
NewsAgent is an enterprise-grade news aggregation agent designed to fetch, query, and summarize news from multiple sources at scale.
StefanHeng/Symbolic-Music-Generation
Symbolic music generation taking inspiration from NLP and human composition process
ieg-dhr/NLP-Course4Humanities_2024
This repository is part of an NLP course for humanities and cultural studies. This course uses historical newspapers as a source and applies NLP methods to them. NLP tasks: Tokenization, Lemmatization, TF-IDF, Part-of-speech tagging, semantic search with transformers, article extraction and OCR post-correction with LLMs, NER and text classification
contrebande-labs/charred
CHARacter-awaRE Diffusion: Multilingual Character-Aware Encoders for Font-Aware Diffusers That Can Actually Spell
Merterm/Modeling-Intensification-for-SLG
Public repo for the paper: "Modeling Intensification for Sign Language Generation: A Computational Approach" by Mert Inan*, Yang Zhong*, Sabit Hassan*, Lorna Quandt, Malihe Alikhani
priyansh4320/Abstractive-Text-Summarization-Enhancing-Sequence-to-Sequence-Models-Using-Word-Sense-Disambiguatio
This repository contains code and resources for abstractive text summarization (TS) using a novel framework that leverages knowledge-based word sense disambiguation (WSD) and semantic content generalization to enhance the performance of sequence-to-sequence (seq2seq) neural-based TS.
ImMohammadHosseini/MKP-RL
:sparkles: Solve multi_dimensional multiple knapsack problem using state_of_the_art Reinforcement Learning Algorithms and transformers
kyegomez/ShallowFF
Zeta implemantion of "Rethinking Attention: Exploring Shallow Feed-Forward Neural Networks as an Alternative to Attention Layers in Transformers"
nicolay-r/THOR-ECAC
The official fork of THoR Chain-of-Thought framework, enhanced and adapted for Emotion Cause Analysis (ECAC-2024)
SkBlaz/autobot
An autoML for explainable text classification.
aifenaike/Intent-Recognition-Using-BERT
Transformer-based Model to recognize any of 7 unique intents from the Snips personal voice assistant.
GabMartino/TransformerForDummies
Annotated implementation of vanilla Transformers to guide through all the ambiguities.
naokishibuya/simple_transformer
A Transformer Implementation that is easy to understand and customizable.
Esmail-ibraheem/Transformer-pytorch
Language to Language Transformer model from scartch using pure Pytorch where I used my transformer model for translation task. from the paper "Attention is all you Need" 2017 using pytorch.
techn0man1ac/ToxicCommentClassification
In the modern era of social media, toxicity in online comments poses a significant challenge, creating a negative atmosphere for communication. From abuse to insults, toxic behavior discourages the free exchange of thoughts and ideas among users. This project offers a solution to this problem.
aaaastark/Pretrain_Finetune_Transformers_Pytorch
Pre-Training and Fine-Tuning transformer models using PyTorch and the Hugging Face Transformers library. Whether you're delving into pre-training with custom datasets or fine-tuning for specific classification tasks, these notebooks offer explanations and code for implementation.