gillosae
M.S. in Department of Intelligence and Information, Music and Audio Research Group, Seoul National University
Seoul, Korea
gillosae's Stars
Guterson/Singing-Voice-Synthesis-in-Brazilian-Portuguese-by-Concatenation-of-Acoustic-Units
Graduation project submitted to the teaching staff from the Eletronics and Computing Engineering undergrad course from the Polytechnic School of the Federal University of Rio de Janeiro as one of the necessary requirements for obtaining the Electronics and Computing Engineer grade.
resonate101/VISinger2-Interactive-System
The Implementation of a Interactive Singing Voice Synthesis System Based on VISinger2.
cyanbx/Prompt-Singer
Implementation of Prompt-Singer: Controllable Singing-Voice-Synthesis with Natural Language Prompt (NAACL'24).
riggraz/no-style-please
A (nearly) no-CSS, fast, minimalist Jekyll theme.
CODEJIN/HiFiSinger
zhangyongmao/VISinger2
VISinger 2: High-Fidelity End-to-End Singing Voice Synthesis Enhanced by Digital Signal Processing Synthesizer
GTSinger/GTSinger
Dataset and code of GTSinger(NeurIPS 2024 Spotlight): A Global Multi-Technique Singing Corpus with Realistic Music Scores for All Singing Tasks
juanalonso/DDSP-singing-experiments
svc-develop-team/so-vits-svc
SoftVC VITS Singing Voice Conversion
zengchang233/xiaoicesing2
The source code for the paper XiaoiceSing2 (interspeech2023)
mmorise/kiritan_singing
東北きりたん歌唱データベースの最新ラベルデータ
jamespark3922/lsmdc-fillin
Identity-Aware Multi-Sentence Video Description
SmartFlowAI/EmoLLM
心理健康大模型、LLM、The Big Model of Mental Health、Finetune、InternLM2、InternLM2.5、Qwen、ChatGLM、Baichuan、DeepSeek、Mixtral、LLama3、GLM4、Qwen2、LLama3.1
PKU-YuanGroup/Chat-UniVi
[CVPR 2024 Highlight🔥] Chat-UniVi: Unified Visual Representation Empowers Large Language Models with Image and Video Understanding
mcomunita/afx-research
Scientific literature about Audio Effects
facebookresearch/sam2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
aws-samples/claude-prompt-generator
open-mmlab/mmdetection
OpenMMLab Detection Toolbox and Benchmark
Sound2Synth/Sound2Synth
Sound2Synth: Interpreting Sound via FM Synthesizer Parameters Estimation
huggingface/peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
aisynth/diffmoog
haoheliu/versatile_audio_super_resolution
Versatile audio super resolution (any -> 48kHz) with AudioSR.
keonlee9420/DiffSinger
PyTorch implementation of DiffSinger: Singing Voice Synthesis via Shallow Diffusion Mechanism (focused on DiffSpeech)
sp-uhh/sgmse
Score-based Generative Models (Diffusion Models) for Speech Enhancement and Dereverberation
modelscope/3D-Speaker
A Repository for Single- and Multi-modal Speaker Verification, Speaker Recognition and Speaker Diarization
facebookresearch/three_bricks
Official Implementation of the paper "Three Bricks to Consolidate Watermarks for LLMs"
sai-soum/Diff-MST
Multitrack music mixing style transfer given a reference song using differentiable mixing console.
diff-usion/Awesome-Diffusion-Models
A collection of resources and papers on Diffusion Models
sh-lee97/grafx-prune
Searching for Music Mixing Graphs: A Pruning Approach
lowerquality/gentle
gentle forced aligner