kdexd's Stars
meta-llama/llama3
The official Meta Llama 3 GitHub site
modularml/mojo
The Mojo Programming Language
black-forest-labs/flux
Official inference repo for FLUX.1 models
ml-explore/mlx
MLX: An array framework for Apple silicon
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
naklecha/llama3-from-scratch
llama3 implementation one matrix multiplication at a time
openai/tiktoken
tiktoken is a fast BPE tokeniser for use with OpenAI's models.
karpathy/minbpe
Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.
meta-llama/llama-models
Utilities intended for use with Llama models.
lancedb/lance
Modern columnar data format for ML and LLMs implemented in Rust. Convert from parquet in 2 lines of code for 100x faster random access, vector index, and data versioning. Compatible with Pandas, DuckDB, Polars, Pyarrow, and PyTorch with more integrations coming..
rom1504/img2dataset
Easily turn large sets of image urls to an image dataset. Can download, resize and package 100M urls in 20h on one machine.
pytorch/torchtitan
A native PyTorch Library for large model training
huggingface/datatrove
Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.
cambrian-mllm/cambrian
Cambrian-1 is a family of multimodal LLMs with a vision-centric design.
visual-layer/fastdup
fastdup is a powerful, free tool designed to rapidly generate valuable insights from image and video datasets. It helps enhance the quality of both images and labels, while significantly reducing data operation costs, all with unmatched scalability.
facebookresearch/MetaCLIP
ICLR2024 Spotlight: curation/training code, metadata, distribution and pre-trained models for MetaCLIP; CVPR 2024: MoDE: CLIP Data Experts via Clustering
pytorch-labs/segment-anything-fast
A batched offline inference oriented version of segment-anything
apple/ml-aim
This repository provides the code and model checkpoints for AIMv1 and AIMv2 research projects.
google-research-datasets/wit
WIT (Wikipedia-based Image Text) Dataset is a large multimodal multilingual dataset comprising 37M+ image-text sets with 11M+ unique images across 100+ languages.
AdaptiveMotorControlLab/CEBRA
Learnable latent embeddings for joint behavioral and neural analysis - Official implementation of CEBRA
dangeng/visual_anagrams
Code for the paper "Visual Anagrams: Generating Multi-View Optical Illusions with Diffusion Models"
mlfoundations/datacomp
DataComp: In search of the next generation of multimodal datasets
HazyResearch/aisys-building-blocks
Building blocks for foundation models.
rom1504/cc2dataset
Easily convert common crawl to a dataset of caption and document. Image/text Audio/text Video/text, ...
facebookresearch/lightplane
Lightplane implements a highly memory-efficient differentiable radiance field renderer, and a module for unprojecting features from images to 3D grids.
gautierdag/bpeasy
Fast bare-bones BPE for modern tokenizer training
facebookresearch/meru
Code for the paper "Hyperbolic Image-Text Representations", Desai et al, ICML 2023
masadcv/FastGeodis
Fast Implementation of Generalised Geodesic Distance Transform for CPU (OpenMP) and GPU (CUDA)
hammoudhasan/SynthCLIP
Code base of SynthCLIP: CLIP training with purely synthetic text-image pairs from LLMs and TTIs.
juliusberner/ddn_tutorial
Tutorial on Deep Declarative Networks