Self-Supervised Speech Pre-training and Representation Learning Toolkit
翻译 - 自我监督的语音预训练和表征学习工具包。
[ICLR 2025] SOTA discrete acoustic codec models with 40/75 tokens per second for audio language modeling
[ACL 2024] Official PyTorch code for extracting features and training downstream models with emotion2vec: Self-Supervised Pre-Training for Speech Emotion Representation
A Survey of Spoken Dialogue Models (60 pages)
LightHuBERT: Lightweight and Configurable Speech Representation Learning with Once-for-All Hidden-Unit BERT
Official Implementation of Mockingjay in Pytorch
A mini, simple, and fast end-to-end automatic speech recognition toolkit.
Semi-supervised spoken language understanding (SLU) via self-supervised speech and language model pretraining
#计算机科学#Causal Speech Enhancement Based on a Two-Branch Nested U-Net Architecture Using Self-Supervised Speech Embeddings