Pinned Repositories
Attention-as-graph
alternative way to calculating self attention
Baji-Marathi-Programing-Language
बाजी (baji) high-level,dynamically typed, interpreted first Marathi programing language.
Bingo-game
Implementation of Bingo multiplayer game in python with module pygame
iza
Tiny Container runtime (Like docker)
microjax
Jax like function transformation engine but micro, microjax
Neural-Style-Transfer-in-Pytorch
Artistic style transfer using pytorch. Recreate images into style of other images(painting)
ohara
Collection of autoregressive model implementation
Python-Smart-Code-Generation
Generating python code using LSTM
TinyLora
Low-Rank Adaptation of Large Language Models clean implementation
Tinytorch
A really tiny autograd engine
joey00072's Repositories
joey00072/Tinytorch
A really tiny autograd engine
joey00072/ohara
Collection of autoregressive model implementation
joey00072/Baji-Marathi-Programing-Language
बाजी (baji) high-level,dynamically typed, interpreted first Marathi programing language.
joey00072/microjax
Jax like function transformation engine but micro, microjax
joey00072/Attention-as-graph
alternative way to calculating self attention
joey00072/iza
Tiny Container runtime (Like docker)
joey00072/TinyLora
Low-Rank Adaptation of Large Language Models clean implementation
joey00072/BitMat
An efficent implementation of the method proposed in "The Era of 1-bit LLMs"
joey00072/QuasiQ
quasiq is really simple quantum computer
joey00072/dotfiles
collection of my dotfile
joey00072/pythonstuff
my blog
joey00072/ring-flash-attention
Ring attention implementation with flash attention
joey00072/axolotl
Go ahead and axolotl questions
joey00072/calorie
nice and effective super simple calorie counter web app
joey00072/CARA
AI assistant that will talk to people about you
joey00072/datawarden
This repository is dedicated to providing cutting-edge tools and methodologies to evaluate and curate datasets specifically designed for Large Language Models (LLMs). Leveraging the capabilities of LLMs themselves, combined with programmatic best practices, our toolkit ensures a robust evaluation and refinement process for your datasets.
joey00072/EasyContext
Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
joey00072/entropix
Entropy Based Sampling and Parallel CoT Decoding
joey00072/expr-ocr
joey00072/Flash-Attention-Softmax-N
CUDA and Triton implementations of Flash Attention with SoftmaxN.
joey00072/joey00072
joey00072/kube-rbac-proxy
Kubernetes RBAC authorizing HTTP proxy for a single upstream.
joey00072/lit-gpt
Hackable implementation of state-of-the-art open-source LLMs based on nanoGPT. Supports flash attention, 4-bit and 8-bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
joey00072/llama2.c
Inference Llama 2 in one file of pure C
joey00072/myserver-setup
bunch of scripts
joey00072/nanotron
Minimalistic large language model 3D-parallelism training
joey00072/ollama
Get up and running with Llama 2 and other large language models locally
joey00072/PythonStuff-old2
joey00072/Sudoko_Solver
Sudoko solver in python
joey00072/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.