We do experiments on Augmented English-French, IWSLT2018 English-German and TED English-Chinese speech translation benchmarks and the
results demonstrate the reasonability of LUT. The table below shows the results on Augmented English-French dataset.
Method |
Enc Pre-train |
Dec Pre-train |
greedy |
beam |
MT system |
|
|
|
|
Transformer MT (Liu et al. 2019) |
- |
- |
21.35 |
22.91 |
Base ST setting |
|
|
|
|
LSTM ST (B ́erard et al. 2018) |
✗ |
✗ |
12.30 |
12.90 |
+pre-train+multitask (B ́erard et al. 2018) |
✓ |
✓ |
12.60 |
13.40 |
LSTM ST+pre-train (Inaguma et al. 2020) |
✓ |
✓ |
- |
16.68 |
Transformer+pre-train (Liu et al. 2019) |
✓ |
✓ |
13.89 |
14.30 |
+knowledge distillation (Liu et al. 2019) |
✓ |
✓ |
14.96 |
17.02 |
TCEN-LSTM (Wang et al. 2020a) |
✓ |
✓ |
- |
17.05 |
Transformer+ASR pre-train (Wang et al. 2020b) |
✓ |
✗ |
- |
15.97 |
Transformer+curriculum pre-train (Wang et al. 2020b) |
✓ |
✗ |
- |
17.66 |
LUT without pre-training |
✗ |
✗ |
16.70 |
17.75 |
Expanded ST setting |
|
|
|
|
LSTM+pre-train+SpecAugment (Bahar et al. 2019) |
✓ |
✓ |
- |
17.00 |
Multilingual ST+PT (Inaguma et al. 2019) |
✓ |
✗ |
- |
17.60 |
Transformer+ASR pre-train (Wang et al. 2020b) |
✓ |
✗ |
- |
16.90 |
Transformer+curriculum pre-train (Wang et al. 2020b) |
✓ |
✗ |
- |
18.01 |
LUT with pre-training |
✓ |
✗ |
17.55 |
18.34 |