Pinned Repositories
API-Projects
Projects using API calls
audio-data-pytorch
A collection of useful audio datasets and transforms for PyTorch.
audio-diffusion
audio-diffusion-pytorch-trainer
Trainer for audio-diffusion-pytorch
audio-local-transformers
Experimental implementations of local attention-based audio transformers and autoencoders
audiocraft_plus
Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable music generation LM with textual and melodic conditioning.
CLAP
Contrastive Language-Audio Pretraining
ComfyUI
The most powerful and modular stable diffusion GUI, api and backend with a graph/nodes interface.
frame-interpolation
FILM: Frame Interpolation for Large Motion, In arXiv 2022.
RAVE
Official implementation of the RAVE model: a Realtime Audio Variational autoEncoder
zqevans's Repositories
zqevans/audio-diffusion
zqevans/audio-local-transformers
Experimental implementations of local attention-based audio transformers and autoencoders
zqevans/audiocraft_plus
Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable music generation LM with textual and melodic conditioning.
zqevans/CLAP
Contrastive Language-Audio Pretraining
zqevans/ComfyUI
The most powerful and modular stable diffusion GUI, api and backend with a graph/nodes interface.
zqevans/frame-interpolation
FILM: Frame Interpolation for Large Motion, In arXiv 2022.
zqevans/RAVE
Official implementation of the RAVE model: a Realtime Audio Variational autoEncoder
zqevans/audio-data-pytorch
A collection of useful audio datasets and transforms for PyTorch.
zqevans/audio-diffusion-pytorch-trainer
Trainer for audio-diffusion-pytorch
zqevans/audio-perception
zqevans/auraloss
Collection of audio-focused loss functions in PyTorch
zqevans/byol-a-2
BYOL for Audio: Self-Supervised Learning for General-Purpose Audio Representation
zqevans/CLIP
Contrastive Language-Image Pretraining
zqevans/ClipCap
Using pretrained encoder and language models to generate captions from multimedia inputs.
zqevans/CRASH
zqevans/FastDiff
PyTorch Implementation of FastDiff (IJCAI'22)
zqevans/glide-text2im
GLIDE: a diffusion-based text-conditional image synthesis model
zqevans/guided-diffusion
zqevans/imagen-pytorch
Implementation of Imagen, Google's Text-to-Image Neural Network, in Pytorch
zqevans/jukebox-diffusion
zqevans/k-diffusion
Karras et al. (2022) diffusion models for PyTorch
zqevans/nicotine-plus
Graphical client for the Soulseek peer-to-peer network
zqevans/open_flamingo
An open-source framework for training large multimodal models.
zqevans/s3prl
Self-Supervised Speech Pre-training and Representation Learning Toolkit.
zqevans/so-vits-svc-fork
so-vits-svc fork with realtime support, improved interface and more features.
zqevans/SoundStream
This repository is an implementation of this article: https://arxiv.org/pdf/2107.03312.pdf
zqevans/steerable-nafx
Steerable discovery of neural audio effects
zqevans/tagbox
Steer OpenAI's Jukebox with Music Taggers
zqevans/unconditional-diff-STFT
Unconditional music synthesis using a diffusion model in the STFT domain
zqevans/v-diffusion-pytorch
v objective diffusion inference code for PyTorch.