Pinned Repositories
gpt-neo
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
stackexchange-dataset
Python tools for processing the stackexchange data dumps into a text dataset for Language Models
the-pile
lm_dataloader
Dataloader tools for language modelling
Opensubtitles_dataset
downloads and parses subtitle dataset from opensubtitles.org
PDFextract
Extracting pdfs using pdfminer.six and pyPDF2
stylegan2
StyleGAN2 - Official TensorFlow Implementation
youtube_subtitle_dataset
YT_subtitles - extracts subtitles from YouTube videos to raw text for Language Model training
sdtblck's Repositories
sdtblck/youtube_subtitle_dataset
YT_subtitles - extracts subtitles from YouTube videos to raw text for Language Model training
sdtblck/Opensubtitles_dataset
downloads and parses subtitle dataset from opensubtitles.org
sdtblck/stylegan2
StyleGAN2 - Official TensorFlow Implementation
sdtblck/PDFextract
Extracting pdfs using pdfminer.six and pyPDF2
sdtblck/lm_dataloader
Dataloader tools for language modelling
sdtblck/image-dl
A fast and simple image downloader in python
sdtblck/pbar-pool
A straightforward, dependency free way to update multiple progress bars with python's multiprocessing library.
sdtblck/tputils
Utilities for TPUs
sdtblck/benchmarking
Tools for benchmarking clusters
sdtblck/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.
sdtblck/example-mkdocs-basic
A basic MkDocs project for Read the Docs
sdtblck/example-sphinx-basic
A basic Sphinx project for Read the Docs
sdtblck/fish
An independent replication of `Training Neural Networks with Fixed Sparse Masks` by Sung et al.
sdtblck/flash-attention
Fast and memory-efficient exact attention
sdtblck/guesslang
Detect the programming language of a source code
sdtblck/lm_dataformat
sdtblck/mapillary_scraper
sdtblck/Megatron-LM
Ongoing research training transformer models at scale
sdtblck/mesh
Mesh TensorFlow: Model Parallelism Made Easier
sdtblck/mesh-transformer-jax
Model parallel transformers in JAX and Haiku
sdtblck/mixtral-offloading
Run Mixtral-8x7B models in Colab or consumer desktops
sdtblck/mojo
The Mojo Programming Language
sdtblck/mup
maximal update parametrization (µP)
sdtblck/pypi
sdtblck/RealFakeAugment
Image augmentation functions for GAN training
sdtblck/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
sdtblck/transformers-bloom-inference
Fast Inference Solutions for BLOOM
sdtblck/Yandex-Image-Scraper
some tools for scraping images from yandex image search