ผลต่างระหว่างรุ่นของ "หน้าหลัก"
ล |
ล |
||
แถว 1: | แถว 1: | ||
− | + | With all the size of common or in-domain corpus increases, it may advantage the translation excellent, simply because FMS nevertheless works much better than IC-baseline, which proves its positive influence on filtering noise. Amongst the 3 presented criteria, PP Based can achieve the highest BLEU with contemplating an acceptable quantity of things for similarity measuring. Nevertheless, the curves show that it depends heavily upon the threshold in (two).The Scientific World Journal40.5 39.0 37.five 36.0 34.5 33.0 31.five 30.0 28.five 27.0 25.five 24.0 22.five 21.0 19.five 18.0 0 150 CE CED(b)40.five 39.0 BLEU 0 150 CE CED(a)BLEU37.five 36.0 34.five 33.0 450 750 300 600 900 The numbers of selected sentences (k) B-CED GC-base300 750 900 450 600 The numbers of chosen sentences (k) B-CED GC-baseFigure 2: BLEU scores by means of perplexity-based information choice procedures with dev. (a) and in-domain (b) methods.42.0 40.five 39.0 BLEU 37.5 36.0 34.5 0 150 300 450 600 750 900 The numbers of selected sentences (k) FMS GC-base(a)42.0 40.five 39.0 37.5 36.0 34.five 33.0 31.five 30.0 28.5 27.0 25.5 24.0 0 150 300 450 600 750 900 The numbers of chosen sentences (k) FMS GC-base(b)BLEUCos-IR B-CEDCos-IR B-CEDFigure three: BLEU scores through distinct information selection approaches with dev. (a) and in-domain (b) tactics.Picking extra or much less pseudo in-domain information will lead to the functionality dropping sharply. Alternatively, Cos-IR performs steadily and robustly with either and both strategies, but its improvements will not be clear. As a [https://britishrestaurantawards.org/members/sphynx48arm/activity/433239/ https://britishrestaurantawards.org/members/sphynx48arm/activity/433239/] result any single individual model can't carry out effectively on each effectiveness and robustness. 5.3. Combined Model. From Figure 3, we identified that every single person model peaks in between 80 K and 320 K. Hence, we only chosen the best = 80 K, 160 K, 320 K for additional comparison. We combined Cos-IR and FMS also as B-CED and assigned equal weights to each individual model at both corpus and model levels (as described in Section 3.4). The translation qualities by means of iTPB are shown in Table four.At each levels, iTPB performs considerably superior than any single person model as well as GC-baseline system. For instance, iTPB-C has achieved at most three.89 (dev) and two.72 (in-domain) improvements than the baseline program. Also the result is still higher than the most effective individual model (B-CED) by 1.92 (dev) and 0.91 (in-domain). This shows a strong capability to balance OOV and noise. Around the one particular hand, filtering too much unmatched words may not sufficiently address the information sparsity problem with the SMT model; alternatively, adding too much of your selected information may bring about the dilution of the in-domain characteristics of the SMT model. Nonetheless, combinations appear to succeed the pros and cut down the cons from the person model. Furthermore, the efficiency of iTPB will not drop sharply when changing the threshold in (2)Table five: Final results of mixture models. Methods GI-baseline IC-baseline B-CED+I Sent. BLEU (dev. | |
− | |||
− |
รุ่นแก้ไขเมื่อ 13:03, 14 มกราคม 2565
With all the size of common or in-domain corpus increases, it may advantage the translation excellent, simply because FMS nevertheless works much better than IC-baseline, which proves its positive influence on filtering noise. Amongst the 3 presented criteria, PP Based can achieve the highest BLEU with contemplating an acceptable quantity of things for similarity measuring. Nevertheless, the curves show that it depends heavily upon the threshold in (two).The Scientific World Journal40.5 39.0 37.five 36.0 34.5 33.0 31.five 30.0 28.five 27.0 25.five 24.0 22.five 21.0 19.five 18.0 0 150 CE CED(b)40.five 39.0 BLEU 0 150 CE CED(a)BLEU37.five 36.0 34.five 33.0 450 750 300 600 900 The numbers of selected sentences (k) B-CED GC-base300 750 900 450 600 The numbers of chosen sentences (k) B-CED GC-baseFigure 2: BLEU scores by means of perplexity-based information choice procedures with dev. (a) and in-domain (b) methods.42.0 40.five 39.0 BLEU 37.5 36.0 34.5 0 150 300 450 600 750 900 The numbers of selected sentences (k) FMS GC-base(a)42.0 40.five 39.0 37.5 36.0 34.five 33.0 31.five 30.0 28.5 27.0 25.5 24.0 0 150 300 450 600 750 900 The numbers of chosen sentences (k) FMS GC-base(b)BLEUCos-IR B-CEDCos-IR B-CEDFigure three: BLEU scores through distinct information selection approaches with dev. (a) and in-domain (b) tactics.Picking extra or much less pseudo in-domain information will lead to the functionality dropping sharply. Alternatively, Cos-IR performs steadily and robustly with either and both strategies, but its improvements will not be clear. As a https://britishrestaurantawards.org/members/sphynx48arm/activity/433239/ result any single individual model can't carry out effectively on each effectiveness and robustness. 5.3. Combined Model. From Figure 3, we identified that every single person model peaks in between 80 K and 320 K. Hence, we only chosen the best = 80 K, 160 K, 320 K for additional comparison. We combined Cos-IR and FMS also as B-CED and assigned equal weights to each individual model at both corpus and model levels (as described in Section 3.4). The translation qualities by means of iTPB are shown in Table four.At each levels, iTPB performs considerably superior than any single person model as well as GC-baseline system. For instance, iTPB-C has achieved at most three.89 (dev) and two.72 (in-domain) improvements than the baseline program. Also the result is still higher than the most effective individual model (B-CED) by 1.92 (dev) and 0.91 (in-domain). This shows a strong capability to balance OOV and noise. Around the one particular hand, filtering too much unmatched words may not sufficiently address the information sparsity problem with the SMT model; alternatively, adding too much of your selected information may bring about the dilution of the in-domain characteristics of the SMT model. Nonetheless, combinations appear to succeed the pros and cut down the cons from the person model. Furthermore, the efficiency of iTPB will not drop sharply when changing the threshold in (2)Table five: Final results of mixture models. Methods GI-baseline IC-baseline B-CED+I Sent. BLEU (dev.