/Whisper-MOE

A generalist model that performs diverse cross-attention tasks like text-to-speech, speech-to-text, and translation.It takes Fine-tunes eight Whisper models with TTT MLP backbones, using 3000-token context lengths and an MOE router for dynamic expert selection. Effective for handling of large datasets and lifelong learning.

Primary LanguagePython

This repository is not active