yurakuratov's Stars
nomic-ai/gpt4all
GPT4All: Run Local LLMs on Any Device. Open-source and available for commercial use.
lm-sys/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
ibraheemdev/modern-unix
A collection of modern/faster/saner alternatives to common unix commands.
google-research/tuning_playbook
A playbook for systematically maximizing the performance of deep learning models.
benfred/py-spy
Sampling profiler for Python programs
muesli/duf
Disk Usage/Free Utility - a better 'df' alternative
microsoft/LoRA
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
facebookresearch/xformers
Hackable and optimized Transformers building blocks, supporting a composable construction.
arogozhnikov/einops
Flexible and powerful tensor operations for readable and reliable code (for pytorch, jax, TF and others)
NVIDIA/FasterTransformer
Transformer related optimization, including BERT, GPT
mosaicml/composer
Supercharge Your Model Training
Lightning-AI/lit-gpt
Hackable implementation of state-of-the-art open-source LLMs based on nanoGPT. Supports flash attention, 4-bit and 8-bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
togethercomputer/RedPajama-Data
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
pytorch/ignite
High-level library to help with training and evaluating neural networks in PyTorch flexibly and transparently.
HazyResearch/flash-attention
Fast and memory-efficient exact attention
baaivision/EVA
EVA Series: Visual Representation Fantasies from BAAI
facebookarchive/MemNN
Memory Networks implementations
pytorch/tnt
A lightweight library for PyTorch training tools and utilities
JonasGeiping/cramming
Cramming the training of a (BERT-type) language model into limited compute.
mryab/efficient-dl-systems
Efficient Deep Learning Systems course materials (HSE, YSDA)
facebookresearch/contriever
Contriever: Unsupervised Dense Information Retrieval with Contrastive Learning
tanelp/tiny-diffusion
A minimal PyTorch implementation of probabilistic diffusion models for 2D datasets.
stanford-crfm/BioMedLM
ofirpress/attention_with_linear_biases
Code for the ALiBi method for transformer language models (ICLR 2022)
lucidrains/recurrent-memory-transformer-pytorch
Implementation of Recurrent Memory Transformer, Neurips 2022 paper, in Pytorch
google-research/meliad
tk-rusch/LEM
Official code for Long Expressive Memory (ICLR 2022, Spotlight)
sismetanin/sentiment-analysis-in-russian
Fine-tuned Multilingual BERT and Multilingual USE for sentiment analysis in Russian. RuReviews, RuSentiment, Kaggle Russian News Dataset, LINIS Crowd, and RuTweetCorp were utilized as training data.
mlcommons/training_results_v2.1
This repository contains the results and code for the MLPerf™ Training v2.1 benchmark.
AIRI-Institute/al_toolbox
Active learning