zlw-eng

Benchmarks

testset BLEU chr-F
newssyscomb2009-ceseng.ces.eng 25.7 0.536
newstest2009-ceseng.ces.eng 24.6 0.530
newstest2010-ceseng.ces.eng 25.0 0.540
newstest2011-ceseng.ces.eng 25.9 0.539
newstest2012-ceseng.ces.eng 24.8 0.533
newstest2013-ceseng.ces.eng 27.8 0.551
newstest2014-csen-ceseng.ces.eng 30.3 0.585
newstest2015-encs-ceseng.ces.eng 27.5 0.542
newstest2016-encs-ceseng.ces.eng 29.1 0.564
newstest2017-encs-ceseng.ces.eng 26.0 0.537
newstest2018-encs-ceseng.ces.eng 27.3 0.544
Tatoeba-test.ces-eng.ces.eng 53.3 0.691
Tatoeba-test.csb-eng.csb.eng 10.2 0.313
Tatoeba-test.dsb-eng.dsb.eng 11.7 0.296
Tatoeba-test.hsb-eng.hsb.eng 24.6 0.426
Tatoeba-test.multi.eng 51.8 0.680
Tatoeba-test.pol-eng.pol.eng 50.4 0.667

System Info:

Downloads last month
14
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Spaces using Helsinki-NLP/opus-mt-zlw-en 7