diff options
author | Patrick Simianer <p@simianer.de> | 2015-02-26 13:26:37 +0100 |
---|---|---|
committer | Patrick Simianer <p@simianer.de> | 2015-02-26 13:26:37 +0100 |
commit | 4223261682388944fe1b1cf31b9d51d88f9ad53b (patch) | |
tree | daf072c310d60b0386587bde5e554312f193b3b2 /training/dtrain/examples/parallelized/work/out.1.0 | |
parent | 2a37a7ad1b21ab54701de3b5b44dc4ea55a75307 (diff) |
refactoring
Diffstat (limited to 'training/dtrain/examples/parallelized/work/out.1.0')
-rw-r--r-- | training/dtrain/examples/parallelized/work/out.1.0 | 72 |
1 files changed, 25 insertions, 47 deletions
diff --git a/training/dtrain/examples/parallelized/work/out.1.0 b/training/dtrain/examples/parallelized/work/out.1.0 index 595dfc94..3dc4dca6 100644 --- a/training/dtrain/examples/parallelized/work/out.1.0 +++ b/training/dtrain/examples/parallelized/work/out.1.0 @@ -1,65 +1,43 @@ - cdec cfg 'cdec.ini' Loading the LM will be faster if you build a binary file. Reading ../standard/nc-wmt11.en.srilm.gz ----5---10---15---20---25---30---35---40---45---50---55---60---65---70---75---80---85---90---95--100 **************************************************************************************************** -Seeding random number sequence to 859043351 - dtrain Parameters: k 100 N 4 T 1 - batch 0 - scorer 'stupid_bleu' - sample from 'kbest' - filter 'uniq' learning rate 0.0001 - gamma 0 - loss margin 1 - faster perceptron 0 - pairs 'XYX' - hi lo 0.1 - pair threshold 0 - select weights 'last' - l1 reg 0 'none' - pclr no - max pairs 4294967295 - repeat 1 - cdec cfg 'cdec.ini' - input '' + error margin 1 + l1 reg 0 + decoder conf 'cdec.ini' + input 'work/shard.1.0.in' output 'work/weights.1.0' -(a dot represents 10 inputs) +(a dot per input) Iteration #1 of 1. - 3 + .... 3 WEIGHTS - Glue = -0.3229 - WordPenalty = +0.27969 - LanguageModel = +1.3645 - LanguageModel_OOV = -0.0443 - PhraseModel_0 = -0.19049 - PhraseModel_1 = -0.077698 - PhraseModel_2 = +0.058898 - PhraseModel_3 = +0.017251 - PhraseModel_4 = -1.5474 - PhraseModel_5 = +0 - PhraseModel_6 = -0.1818 - PassThrough = -0.193 + Glue = -0.2722 + WordPenalty = +0.05433 + LanguageModel = +0.69948 + LanguageModel_OOV = -0.2641 + PhraseModel_0 = -1.4208 + PhraseModel_1 = -1.563 + PhraseModel_2 = -0.21051 + PhraseModel_3 = -0.17764 + PhraseModel_4 = -1.6583 + PhraseModel_5 = +0.0794 + PhraseModel_6 = +0.1528 + PassThrough = -0.2367 --- - 1best avg score: 0.070229 (+0.070229) - 1best avg model score: -44.01 (-44.01) - avg # pairs: 1294 - avg # rank err: 878.67 - avg # margin viol: 350.67 - k-best loss imp: 100% - non0 feature count: 11 + 1best avg score: 0.071329 (+0.071329) + 1best avg model score: -41.362 + avg # pairs: 1862.3 + non-0 feature count: 12 avg list sz: 100 - avg f count: 11.487 -(time 0.28 min, 5.7 s/S) - -Writing weights file to 'work/weights.1.0' ... -done + avg f count: 11.847 +(time 0.28 min, 5 s/S) --- -Best iteration: 1 [SCORE 'stupid_bleu'=0.070229]. +Best iteration: 1 [GOLD = 0.071329]. This took 0.28333 min. |