Ray0323 commited on
Commit
e308122
·
verified ·
1 Parent(s): e5e0763

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -9
README.md CHANGED
@@ -1,20 +1,34 @@
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
4
- # DistilCodec
5
- The Joint Laboratory of International Digital Economy Academy (IDEA) and Emdoor, in collaboration with Emdoor Information Technology Co., Ltd., has launched DistilCodec - A Single-Codebook Neural Audio Codec (NAC) with 32768 codes trained on uniersal audio.
6
-
7
-
8
- [![arXiv](https://img.shields.io/badge/arXiv-Paper-<COLOR>.svg)](https://arxiv.org/abs/2408.16532)
9
- [![model](https://img.shields.io/badge/%F0%9F%A4%97%20DistilCodec-Models-blue)](https://huggingface.co/IDEA-Emdoor/DistilCodec-v1.0)
 
 
 
 
 
 
 
 
 
 
 
 
 
10
 
11
 
12
  # 🔥 News
13
- - *2025.05.25*: We release the code of DistilCodec-v1.0, including training and inference.
14
- - *2025.05.23*: We release UniTTS and DistilCodec on [arxiv](https://arxiv.org/abs/2408.16532).
 
15
 
16
  ## Introduction of DistilCodec
17
- The foundational network architecture of DistilCodec adopts an Encoder-VQ-Decoder framework
18
  similar to that proposed in Soundstream. The encoder employs a ConvNeXt-V2 structure,
19
  while the vector quantization module implements the GRFVQ scheme. The decoder
20
  employs a ConvTranspose1d based architectural configuration similar to HiFiGAN. The training methodol-
 
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
4
+ <div align="center">
5
+ <h1>
6
+ DistilCodec
7
+ </h1>
8
+ <p>
9
+ <b><em>DistilCodec: A Single Codebook Audio Codec For Universal Audio</em></b>
10
+ </p>
11
+ <p>
12
+ </p>
13
+ </p>
14
+ <a href="https" style="color:red">Paper </a> |
15
+ <a href="https://huggingface.co/IDEA-Emdoor/DistilCodec-v1.0" style="color:#FFD700">Hugging Face Model</a>
16
+ <a href="https://github.com/IDEA-Emdoor-Lab/DistilCodec" style="color:gray">Code</a>
17
+ <p>
18
+ <img src="./idea_logo.png" alt="Institution 1" style="width: 200px; height: 60px;">
19
+ <img src="./yidao_logo.png" alt="Institution 2" style="width: 200px; height: 60px;">
20
+ <img src="./yijiayiban.png" alt="Institution 3" style="width: 200px; height: 60px;">
21
+ </p>
22
+ </div>
23
 
24
 
25
  # 🔥 News
26
+ - *2025.05.27*: We release DistilCodec-v1.0 checkpoint on [huggingface](https://huggingface.co/IDEA-Emdoor/DistilCodec-v1.0).
27
+ - *2025.05.26*: We release the code of DistilCodec-v1.0, including training and inference.
28
+ - *2025.05.24*: We release UniTTS and DistilCodec on [arxiv](https://arxiv.org/abs/2408.16532).
29
 
30
  ## Introduction of DistilCodec
31
+ The Joint Laboratory of International Digital Economy Academy (IDEA) and Emdoor, in collaboration with Emdoor Information Technology Co., Ltd., and Shenzhen Yijiayiban Information Technology Co., Ltd, has launched DistilCodec - A Single-Codebook Neural Audio Codec (NAC) with 32768 codes trained on uniersal audio.The foundational network architecture of DistilCodec adopts an Encoder-VQ-Decoder framework
32
  similar to that proposed in Soundstream. The encoder employs a ConvNeXt-V2 structure,
33
  while the vector quantization module implements the GRFVQ scheme. The decoder
34
  employs a ConvTranspose1d based architectural configuration similar to HiFiGAN. The training methodol-