speech-analysis
There are 129 repositories under speech-analysis topic.
jianchang512/clone-voice
A sound cloning tool with a web interface, using your voice or any sound to record audio / 一个带web界面的声音克隆工具,使用你的音色或任意声音来录制音频
praat/praat
Praat: Doing Phonetics By Computer
mmorise/World
A high-quality speech analysis, manipulation and synthesis system
haoheliu/voicefixer
General Speech Restoration
DmitryRyumin/INTERSPEECH-2023-24-Papers
INTERSPEECH 2023-2024 Papers: A complete collection of influential and exciting research papers from the INTERSPEECH 2023-24 conference. Explore the latest advances in speech and language processing. Code included. Star the repository to support the advancement of speech technology!
gemengtju/Tutorial_Separation
This repo summarizes the tutorials, datasets, papers, codes and tools for speech separation and speaker extraction task. You are kindly invited to pull requests.
speechbrain/speechbrain.github.io
The SpeechBrain project aims to build a novel speech toolkit fully based on PyTorch. With SpeechBrain users can easily create speech processing systems, ranging from speech recognition (both HMM/DNN and end-to-end), speaker recognition, speech enhancement, speech separation, multi-microphone speech processing, and many others.
jcvasquezc/DisVoice
feature extraction from speech signals
Shahabks/my-voice-analysis
My-Voice Analysis is a Python library for the analysis of voice (simultaneous speech, high entropy) without the need of a transcription. It breaks utterances and detects syllable boundaries, fundamental frequency contours, and formants.
haoheliu/voicefixer_main
General Speech Restoration
Shahabks/myprosody
A Python library for measuring the acoustic features of speech (simultaneous speech, high entropy) compared to ones of native speech.
HidekiKawahara/legacy_STRAIGHT
A vocoder framework which had been widely used in research community since 1999.
philipperemy/tensorflow-ctc-speech-recognition
Application of Connectionist Temporal Classification (CTC) for Speech Recognition (Tensorflow 1.0 but compatible with 2.0).
at16k/at16k
Trained models for automatic speech recognition (ASR). A library to quickly build applications that require speech to text conversion.
JusperLee/Calculate-SNR-SDR
Script to calculate SNR and SDR using python
LimingShi/Bayesian-Pitch-Tracking-Using-Harmonic-model
Pitch detection and pitch tracking, voicing unvoicing detection (VAD),基音检测
google/localized-narratives
Localized Narratives
CSTR-Edinburgh/magphase
MagPhase Vocoder: Speech analysis/synthesis system for TTS and related applications.
Speech-Interaction-Technology-Aalto-U/itsp
Introduction to Speech Processing
RichardHladik/outotune
An opensource harmonizer implementation leveraging the DISTRHO Plugin Framework.
mjpyeon/wavenet-classifier
Keras Implementation of Deepmind's WaveNet for Supervised Learning Tasks
hyeonsangjeon/computing-Korean-STT-error-rates
STT 한글 문장 인식기 출력 스크립트의 외자 오류율(CER), 단어 오류율(WER)을 계산하는 Python 함수 패키지
lennes/spect
SpeCT - Speech Corpus Toolkit for Praat. Documentation: https://lennes.github.io/spect/
HidekiKawahara/SparkNG
MATLAB real-time/interactive speech tools. This series is obsolete. SP3ARK is the up-to-date series (will be).
jcvasquezc/NeuroSpeech
Toolkit to asses speech impairments in patients with neurological disorders
MontrealCorpusTools/PolyglotDB
Language data store and linguistic query API
alessandroragano/scoreq
SCOREQ: Speech COntrastive REgression for Quality Assessment (NeurIPS 2024)
msalhab96/SNR-Estimation-Using-Deep-Learning
An implementation for Frame-level Speech Signal-to-Noise Ratio Estimation using deep learning
HidekiKawahara/YANGstraight_source
Analytic signal-based source information analysis for YANGstraight and real-time interactive tools
tabahi/WebSpeechAnalyzer
JS speech analyzer for fast speech analysis and labeling
praaline/Praaline
Praaline is an open-source system to manage, annotate, visualise and analyse spoken language corpora
praweshd/speech_emotion_recognition
In this project, the performance of speech emotion recognition is compared between two methods (SVM vs Bi-LSTM RNN).Conventional classifiers that uses machine learning algorithms has been used for decades in recognizing emotions from speech. However, in recent years, deep learning methods have taken the center stage and have gained popularity for their ability to perform well without any input hand-crafted features. Speech emotion on sets obtained from RAVDESS corpus is classified using a conventionally used Support Vector Machine (SVM) and its performance is compared to that of a bidirectional long short-term memory (LSTM).
operrotin/GFM-IAIF
Glottal Flow Model-based Iterative Adaptive Inverse Filtering
ringabout/scim
[wip]Speech recognition tool-box written by Nim. Based on Arraymancer.
LinkonBSMRSTU/Speech-To-Text-App-iOS
A simple iOS App that can convert speech/voice into text. Only English voice is supported for now. Used Swift 5, AVKit and Speech.
type-a/speechnet
Automatic Speech Recognition