Pinned Repositories
Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
transformers-bloom-inference
Fast Inference Solutions for BLOOM
BigCode-Megatron-LM
Ongoing research training transformer models at scale
deep-learning-bias-correction
GPTQ-for-SantaCoder
4 bits quantization of SantaCoder using GPTQ
HybridToD
Papers-books-and-blogs
This repository contains the research papers, white papers, thesis etc that I love.
pseudo-code-instructions
Pseudo-code Instructions dataset
real-time-visual-respiration-rate-estimation-with-dynamic-scene-adaptation
VRAG
mayank31398's Repositories
mayank31398/GPTQ-for-SantaCoder
4 bits quantization of SantaCoder using GPTQ
mayank31398/pseudo-code-instructions
Pseudo-code Instructions dataset
mayank31398/Papers-books-and-blogs
This repository contains the research papers, white papers, thesis etc that I love.
mayank31398/VRAG
mayank31398/BigCode-Megatron-LM
Ongoing research training transformer models at scale
mayank31398/HybridToD
mayank31398/blog
Public repo for HF blog posts
mayank31398/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
mayank31398/dotfiles
mayank31398/flash-attention
Fast and memory-efficient exact attention
mayank31398/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
mayank31398/gym-text
mayank31398/ibm-fm-base-images
mayank31398/IBM-fms-fsdp
Demonstrate throughput of PyTorch FSDP
mayank31398/kernel-hyperdrive
A bunch of kernels to make stuff slower 😉
mayank31398/lm-evaluation-harness
A framework for few-shot evaluation of language models.
mayank31398/lqvae
mayank31398/mayank.github.io
Github Pages template for academic personal websites, forked from mmistakes/minimal-mistakes
mayank31398/mayank31398
mayank31398/mayank31398.github.io
mayank31398/MIPS-verilog
This repository contains code for a MIPS single cycle architecture written in Verilog.
mayank31398/mixed-communication-runtime
mayank31398/optimum
🏎️ Accelerate training and inference of 🤗 Transformers with easy to use hardware optimization tools
mayank31398/peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
mayank31398/pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
mayank31398/rocm-apex
A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch
mayank31398/scattermoe
Triton-based implementation of Sparse Mixture of Experts.
mayank31398/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
mayank31398/vscode-icons
Icons for Visual Studio Code
mayank31398/vscode-settings