Update README.md
Browse files
README.md
CHANGED
|
@@ -21,11 +21,11 @@ The architect looks like:
|
|
| 21 |
# Usage
|
| 22 |
Inference sample code is:
|
| 23 |
https://github.com/yynil/RWKVTTS/blob/respark/model/test/test_asr_whisper.py
|
| 24 |
-
1. Download
|
| 25 |
-
2. Download the
|
| 26 |
-
|
| 27 |
```bash
|
| 28 |
-
python model/test/test_asr_whisper.py --
|
| 29 |
```
|
| 30 |
The output looks like:
|
| 31 |
|
|
@@ -33,7 +33,7 @@ The output looks like:
|
|
| 33 |
|
| 34 |
or in English mode
|
| 35 |
```bash
|
| 36 |
-
python model/test/test_asr_whisper.py --
|
| 37 |
```
|
| 38 |
The output looks like:
|
| 39 |
|
|
|
|
| 21 |
# Usage
|
| 22 |
Inference sample code is:
|
| 23 |
https://github.com/yynil/RWKVTTS/blob/respark/model/test/test_asr_whisper.py
|
| 24 |
+
1. Download the weights in this repo. Please note: 10k steps checkpoint training costs around 5k hours which is a very small amount of data and we are continuing training. Also it proves this mode needs less data to achieve a usable stage.
|
| 25 |
+
2. Download the configuration directories in this repo. Assume you store them to directory YOUR_DIR.
|
| 26 |
+
2. Run the script like:
|
| 27 |
```bash
|
| 28 |
+
python model/test/test_asr_whisper.py --whisper_path $YOUR_DIR/whisper-large-v3/ --audio_lm_path $YOUR_DIR/rwkv7_0.1b_audio_lm_latents/ --llm_path $YOUR_DIR/rwkv7-0.4B-g1a/ --ckpt_path $YOUR_DIR/rwkvasr_whisper_10k.model.bin --audio_path new.mp3
|
| 29 |
```
|
| 30 |
The output looks like:
|
| 31 |
|
|
|
|
| 33 |
|
| 34 |
or in English mode
|
| 35 |
```bash
|
| 36 |
+
python model/test/test_asr_whisper.py --whisper_path $YOUR_DIR/whisper-large-v3/ --audio_lm_path $YOUR_DIR/rwkv7_0.1b_audio_lm_latents/ --llm_path /home/yueyulin/models/rwkv7-0.4B-g1a/ --ckpt_path $YOUR_DIR/rwkvasr_whisper_10k.model.bin --audio_path eng2.wav --language english
|
| 37 |
```
|
| 38 |
The output looks like:
|
| 39 |
|