transformers

There are 3204 repositories under transformers topic.

  • annotated_deep_learning_paper_implementations

    labmlai/annotated_deep_learning_paper_implementations

    🧑‍🏫 60 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠

    Language:Jupyter Notebook47.9k4311195k
  • generative-ai-for-beginners

    microsoft/generative-ai-for-beginners

    18 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/

    Language:Jupyter Notebook42.4k3668023.1k
  • LLaMA-Factory

    hiyouga/LLaMA-Factory

    Unify Efficient Fine-Tuning of 100+ LLMs

    Language:Python18.2k1393.1k2.2k
  • lucidrains/vit-pytorch

    Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch

    Language:Python17.9k1412512.8k
  • amusi/CVPR2024-Papers-with-Code

    CVPR 2024 论文和开源项目合集

  • huggingface/peft

    🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.

    Language:Python13.8k1048621.3k
  • haystack

    deepset-ai/haystack

    :mag: LLM orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.

    Language:Python13.6k1243.2k1.6k
  • BlinkDL/RWKV-LM

    RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.

    Language:Python11.6k134188799
  • PaddleNLP

    PaddlePaddle/PaddleNLP

    👑 Easy-to-use and powerful NLP and LLM library with 🤗 Awesome model zoo, supporting wide-range of NLP tasks from research to industrial applications, including 🗂Text Classification, 🔍 Neural Search, ❓ Question Answering, ℹ️ Information Extraction, 📄 Document Intelligence, 💌 Sentiment Analysis etc.

    Language:Python11.4k1013.4k2.8k
  • stas00/ml-engineering

    Machine Learning Engineering Open Book

    Language:Python9.7k9917579
  • NVIDIA/Megatron-LM

    Ongoing research training transformer models at scale

    Language:Python8.5k1514981.9k
  • tokenizers

    huggingface/tokenizers

    💥 Fast State-of-the-Art Tokenizers optimized for Research and Production

    Language:Rust8.4k118916715
  • EleutherAI/gpt-neo

    An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.

    Language:Python8.1k178139935
  • speechbrain/speechbrain

    A PyTorch-based Speech Toolkit

    Language:Python7.9k1271k1.3k
  • lucidrains/PaLM-rlhf-pytorch

    Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM

    Language:Python7.6k14146658
  • NielsRogge/Transformers-Tutorials

    This repository contains demos I made with the Transformers library by HuggingFace.

    Language:Jupyter Notebook7.5k1183971.2k
  • xenova/transformers.js

    State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!

    Language:JavaScript7.5k56432411
  • txtai

    neuml/txtai

    💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows

    Language:Python7k81664493
  • EleutherAI/gpt-neox

    An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.

    Language:Python6.6k119421947
  • bertviz

    jessevig/bertviz

    BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)

    Language:Python6.4k70118748
  • intel-analytics/ipex-llm

    Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, ModelScope, etc

    Language:Python5.9k2432.2k1.2k
  • openvinotoolkit/openvino

    OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference

    Language:C++5.9k1842.4k2k
  • BERTopic

    MaartenGr/BERTopic

    Leveraging BERT and c-TF-IDF to create easily interpretable topics.

    Language:Python5.5k491.6k693
  • lucidrains/DALLE-pytorch

    Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch

    Language:Python5.5k96275637
  • imoneoi/openchat

    OpenChat: Advancing Open-source Language Models with Imperfect Data

    Language:Python5k48172379
  • SkalskiP/courses

    This repository is a curated collection of links to various courses and resources about Artificial Intelligence (AI)

    Language:Python4.5k795408
  • lucidrains/deep-daze

    Simple command line tool for text to image generation using OpenAI's CLIP and Siren (Implicit neural representation network). Technique was originally created by https://twitter.com/advadnoun

    Language:Python4.4k74166326
  • superduperdb

    SuperDuperDB/superduperdb

    🔮 SuperDuperDB: Bring AI to your database! Build, deploy and manage any AI application directly with your existing data infrastructure, without moving your data. Including streaming inference, scalable model training and vector search.

    Language:Python4.3k391k429
  • cmhungsteve/Awesome-Transformer-Attention

    An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites

  • bentrevett/pytorch-sentiment-analysis

    Tutorials on getting started with PyTorch and TorchText for sentiment analysis.

    Language:Jupyter Notebook4.2k831081.2k
  • trigaten/Learn_Prompting

    Prompt Engineering, Generative AI, and LLM Guide by Learn Prompting | Join our discord for the largest Prompt Engineering learning community

    Language:MDX4.2k93393626
  • lucidrains/x-transformers

    A simple but complete full-attention transformer with a set of promising experimental features from various papers

    Language:Python4.1k51191353
  • marqo

    marqo-ai/marqo

    Unified embedding generation and search engine. Also available on cloud - cloud.marqo.ai

    Language:Python4.1k36230174
  • simpletransformers

    ThilinaRajapakse/simpletransformers

    Transformers for Information Retrieval, Text Classification, NER, QA, Language Modelling, Language Generation, T5, Multi-Modal, and Conversational AI

    Language:Python4k621.1k720
  • Fengshenbang-LM

    IDEA-CCNL/Fengshenbang-LM

    Fengshenbang-LM(封神榜大模型)是IDEA研究院认知计算与自然语言研究中心主导的大模型开源体系,成为中文AIGC和认知智能的基础设施。

    Language:Python3.9k55289357
  • CLUEbenchmark/CLUE

    中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard

    Language:Python3.8k8999539