Stars
The official implementation of [CVPR2022] Decoupled Knowledge Distillation https://arxiv.org/abs/2203.08679 and [ICCV2023] DOT: A Distillation-Oriented Trainer https://openaccess.thecvf.com/content…
This repository is the official implementation of Disentangling Writer and Character Styles for Handwriting Generation (CVPR 2023)
Code for our CVPR2021 paper coordinate attention
Official release of pretrained models and codes for 'Golden Gemini Is All You Need: Finding the Sweet Spots for Speaker Verification'
Official Pytorch implementations for "SegNeXt: Rethinking Convolutional Attention Design for Semantic Segmentation" (NeurIPS 2022)
🦄️ 🎃 👻 Clash Premium 规则集(RULE-SET),兼容 ClashX Pro、Clash for Windows 等基于 Clash Premium 内核的客户端。
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
Official repository for Mamba-based Segmentation Model for Speaker Diarization
A Repository for Single- and Multi-modal Speaker Verification, Speaker Recognition and Speaker Diarization
Awesome Knowledge Distillation
[CVPR 2024 Highlight] Logit Standardization in Knowledge Distillation
Sushmitha-Deva / ctcdecode
Forked from parlance/ctcdecodePyTorch CTC Decoder bindings
deep learning for image processing including classification and object-detection etc.
Pretrain, finetune ANY AI model of ANY size on multiple GPUs, TPUs with zero code changes.
[ASRU 2021] Efficient Conformer: Progressive Downsampling and Grouped Attention for Automatic Speech Recognition
《李宏毅深度学习教程》(李宏毅老师推荐👍,苹果书🍎),PDF下载地址:https://github.com/datawhalechina/leedl-tutorial/releases
Unofficial reimplementation of ECAPA-TDNN for speaker recognition (EER=0.86 for Vox1_O when train only in Vox2)