imskyszy's Stars
KindXiaoming/pykan
Kolmogorov Arnold Networks
mamoe/mirai
高效率 QQ 机器人支持库
ShiArthur03/ShiArthur03
nonebot/nonebot
基于 OneBot 标准的 Python 异步 QQ 机器人框架 / Asynchronous QQ robot framework based on OneBot for Python
AberHu/Knowledge-Distillation-Zoo
Pytorch implementation of various Knowledge Distillation (KD) methods.
qiuqiangkong/audioset_tagging_cnn
LetheSec/HuggingFace-Download-Accelerator
利用HuggingFace的官方下载工具从镜像网站进行高速下载。
pengzhiliang/Conformer
Official code for Conformer: Local Features Coupling Global Representations for Visual Recognition
harritaylor/torchvggish
Pytorch port of Google Research's VGGish model used for extracting audio features.
fschmid56/EfficientAT
This repository aims at providing efficient CNNs for Audio Tagging. We provide AudioSet pre-trained models ready for downstream training and extraction of audio embeddings.
RoyChao19477/SEMamba
This is the official implementation of the SEMamba paper. (Accepted to IEEE SLT 2024)
JusperLee/SPMamba
kaistmm/Audio-Mamba-AuM
Official Implementation of the work "Audio Mamba: Bidirectional State Space Model for Audio Representation Learning"
Thvnvtos/SNN-delays
Official implementation of "Learning Delays in Spiking Neural Networks using Dilated Convolutions with Learnable Spacings" [ICLR2024]
umbertocappellazzo/PETL_AST
This is the official repository of the papers "Parameter-Efficient Transfer Learning of Audio Spectrogram Transformers" and "Efficient Fine-tuning of Audio Spectrogram Transformers via Soft Mixture of Adapters".
judiebig/DR-DiffuSE
Revisiting Denoising Diffusion Probabilistic Models for Speech Enhancement: Condition Collapse, Efficiency and Refinement, Thirty-Seventh AAAI Conference on Artificial Intelligence (AAAI), 2023.
SarthakYadav/audio-mamba-official
Official implementation for our paper "Audio Mamba: Selective State Spaces for Self-Supervised Audio Representations"
Hadryan/TFNet-for-Environmental-Sound-Classification
Learning discriminative and robust time-frequency representations for environmental sound classification: Convolutional neural networks (CNN) are one of the best-performing neural network architectures for environmental sound classification (ESC). Recently, attention mechanisms have been used in CNN to capture the useful information from the audio signal for sound classification, especially for weakly labelled data where the timing information about the acoustic events is not available in the training data, apart from the availability of sound class labels. In these methods, however, the inherent time-frequency characteristics and variations are not explicitly exploited when obtaining the deep features. In this paper, we propose a new method, called time-frequency enhancement block (TFBlock), which temporal attention and frequency attention are employed to enhance the features from relevant frames and frequency bands. Compared with other attention mechanisms, in our method, parallel branches are constructed which allow the temporal and frequency features to be attended respectively in order to mitigate interference from the sections where no sound events happened in the acoustic environments. The experiments on three benchmark ESC datasets show that our method improves the classification performance and also exhibits robustness to noise.
SY-Xuan/DSCL
AAAI-24 Decoupled Contrastive Learning for Long-Tailed Recognition
stanfordmlgroup/selfsupervised-lungandheartsounds
kaen2891/bts
(INTERSPEECH 2024) Official Implementation of "BTS: Bridging Text and Sound Modalities for Metadata-Aided Respiratory Sound Classification"
kaen2891/stethoscope-guided_supervised_contrastive_learning
(ICASSP 2024) Official Implementation of "Stethoscope-guided Supervised Contrastive Learning for Cross-domin Adaptation on Respiratory Sound Classification"
ChihchengHsieh/multimodal-abnormalities-detection
MichaelLynn1996/AAT
ta012/DTFAT
DTFAT code and pre-trained weights
ta012/MaxAST
MaxAST code
PLAN-Lab/uamix-MAE
fatetail/MetaNet
ISBI2020
danelee2601/SpecMix_torch
(not official) pytorch implementation of SpecMix, an augmentation method
jczhang02/VGG_audiovisual_torch
Pytorch implemention of VGG16 and VGGish models