We do experiments on Augmented English-French, IWSLT2018 English-German and TED English-Chinese speech translation benchmarks and the
results demonstrate the reasonability of LUT. The table below shows the results on Augmented English-French dataset.
| Method |
Enc Pre-train |
Dec Pre-train |
greedy |
beam |
| MT system |
|
|
|
|
| Transformer MT (Liu et al. 2019) |
- |
- |
21.35 |
22.91 |
| Base ST setting |
|
|
|
|
| LSTM ST (B ́erard et al. 2018) |
✗ |
✗ |
12.30 |
12.90 |
| +pre-train+multitask (B ́erard et al. 2018) |
✓ |
✓ |
12.60 |
13.40 |
| LSTM ST+pre-train (Inaguma et al. 2020) |
✓ |
✓ |
- |
16.68 |
| Transformer+pre-train (Liu et al. 2019) |
✓ |
✓ |
13.89 |
14.30 |
| +knowledge distillation (Liu et al. 2019) |
✓ |
✓ |
14.96 |
17.02 |
| TCEN-LSTM (Wang et al. 2020a) |
✓ |
✓ |
- |
17.05 |
| Transformer+ASR pre-train (Wang et al. 2020b) |
✓ |
✗ |
- |
15.97 |
| Transformer+curriculum pre-train (Wang et al. 2020b) |
✓ |
✗ |
- |
17.66 |
| LUT without pre-training |
✗ |
✗ |
16.70 |
17.75 |
| Expanded ST setting |
|
|
|
|
| LSTM+pre-train+SpecAugment (Bahar et al. 2019) |
✓ |
✓ |
- |
17.00 |
| Multilingual ST+PT (Inaguma et al. 2019) |
✓ |
✗ |
- |
17.60 |
| Transformer+ASR pre-train (Wang et al. 2020b) |
✓ |
✗ |
- |
16.90 |
| Transformer+curriculum pre-train (Wang et al. 2020b) |
✓ |
✗ |
- |
18.01 |
| LUT with pre-training |
✓ |
✗ |
17.55 |
18.34 |