AMAAI Lab
The Audio, Music, and AI Lab at Singapore University of Technology and Design (SUTD)
Singapore
Pinned Repositories
awesome-MER
A curated list of Datasets, Models and Papers for Music Emotion Recognition (MER)
JamendoMaxCaps
JamendoMaxCaps is a large-scale dataset of 362,000 instrumental creative commons tracks
megamusicaps
MidiCaps
A large-scale dataset of caption-annotated MIDI files.
mirflex
Music Information Retrieval Feature Library for Extraction
Music2Emotion
Music2Emo: Towards Unified Music Emotion Recognition across Dimensional and Categorical Models
mustango
Mustango: Toward Controllable Text-to-Music Generation
MuVi
Predicting emotion from music videos: exploring the relative contribution of visual and auditory information on affective responses
Text2midi
Text2midi is the first end-to-end model for generating MIDI files from textual descriptions. By leveraging pretrained large language models and a powerful autoregressive transformer decoder, text2midi allows users to create symbolic music that aligns with detailed textual prompts, including musical attributes like chords, tempo, and style.
Video2Music
Video2Music: Suitable Music Generation from Videos using an Affective Multimodal Transformer model
AMAAI Lab's Repositories
AMAAI-Lab/mustango
Mustango: Toward Controllable Text-to-Music Generation
AMAAI-Lab/Video2Music
Video2Music: Suitable Music Generation from Videos using an Affective Multimodal Transformer model
AMAAI-Lab/Text2midi
Text2midi is the first end-to-end model for generating MIDI files from textual descriptions. By leveraging pretrained large language models and a powerful autoregressive transformer decoder, text2midi allows users to create symbolic music that aligns with detailed textual prompts, including musical attributes like chords, tempo, and style.
AMAAI-Lab/MidiCaps
A large-scale dataset of caption-annotated MIDI files.
AMAAI-Lab/JamendoMaxCaps
JamendoMaxCaps is a large-scale dataset of 362,000 instrumental creative commons tracks
AMAAI-Lab/awesome-MER
A curated list of Datasets, Models and Papers for Music Emotion Recognition (MER)
AMAAI-Lab/MuVi
Predicting emotion from music videos: exploring the relative contribution of visual and auditory information on affective responses
AMAAI-Lab/Music2Emotion
Music2Emo: Towards Unified Music Emotion Recognition across Dimensional and Categorical Models
AMAAI-Lab/mirflex
Music Information Retrieval Feature Library for Extraction
AMAAI-Lab/megamusicaps
AMAAI-Lab/DART
Demo for DART, Audio Imagination workshop submission in NeurIPS 2024
AMAAI-Lab/PreBit
This is the repo accompanying the paper: "A multimodal model with Twitter FinBERT embeddings for extreme price movement prediction of Bitcoin"
AMAAI-Lab/cross-dataset-emotion-alignment
code for Leveraging LLM Embeddings for Cross Dataset Label Alignment and Zero Shot Music Emotion Prediction
AMAAI-Lab/DisfluencySpeech
Resources for DisfluencySpeech
AMAAI-Lab/Accented-TTS-MLVAE-ADV
AMAAI-Lab/ai-audio-datasets-list
This is a list of datasets consisting of speech, music, and sound effects, which can provide training data for Generative AI, AIGC, AI model training, intelligent audio tool development, and audio applications. It is mainly used for speech recognition, speech synthesis, singing voice synthesis, music information retrieval, music generation, etc.
AMAAI-Lab/Audio-Music-AI-Research-Resources
AMAAI-Lab/CM-HRNN
Hierarchical Recurrent Neural Networks for Conditional Melody Generation with Long-term Structure
AMAAI-Lab/CVAE-Tacotron
Conditional VAE for Accented Speech Generation
AMAAI-Lab/emotionweb
Website emotion guidance
AMAAI-Lab/genmusic_demo_list
a list of demo websites for automatic music generation research
AMAAI-Lab/IAMM
An exploration of how generative text-to-music AI models can be used for emotion guidance
AMAAI-Lab/kylo-ren-app
Web interface for AI music generation models
AMAAI-Lab/singapore-music-classifier
Code for paper A dataset and classification model for Malay, Hindi, Tamil and Chinese music
AMAAI-Lab/survey-music-nlp
Repository for "Natural Language Processing Methods for Symbolic Music Generation and Information Retrieval Systems: a Survey"
AMAAI-Lab/nnAudio
Audio processing by using pytorch 1D convolution network
AMAAI-Lab/AudioLoader
PyTorch Dataset for Speech and Music audio
AMAAI-Lab/FundamentalMusicEmbedding
Fundamental Music Embedding, FME
AMAAI-Lab/Jointist
Official Implementation of Jointist
AMAAI-Lab/ReconVAT
ReconVAT: a semi-supervised automatic music transcription (AMT) model