diff options
Diffstat (limited to 'training/dtrain/examples/parallelized/work/out.2.2')
-rw-r--r-- | training/dtrain/examples/parallelized/work/out.2.2 | 74 |
1 files changed, 26 insertions, 48 deletions
diff --git a/training/dtrain/examples/parallelized/work/out.2.2 b/training/dtrain/examples/parallelized/work/out.2.2 index e0ca2110..f5d6229f 100644 --- a/training/dtrain/examples/parallelized/work/out.2.2 +++ b/training/dtrain/examples/parallelized/work/out.2.2 @@ -1,66 +1,44 @@ - cdec cfg 'cdec.ini' Loading the LM will be faster if you build a binary file. Reading ../standard/nc-wmt11.en.srilm.gz ----5---10---15---20---25---30---35---40---45---50---55---60---65---70---75---80---85---90---95--100 **************************************************************************************************** -Seeding random number sequence to 2803362953 - dtrain Parameters: k 100 N 4 T 1 - batch 0 - scorer 'stupid_bleu' - sample from 'kbest' - filter 'uniq' learning rate 0.0001 - gamma 0 - loss margin 1 - faster perceptron 0 - pairs 'XYX' - hi lo 0.1 - pair threshold 0 - select weights 'last' - l1 reg 0 'none' - pclr no - max pairs 4294967295 - repeat 1 - cdec cfg 'cdec.ini' - input '' + error margin 1 + l1 reg 0 + decoder conf 'cdec.ini' + input 'work/shard.2.0.in' output 'work/weights.2.2' weights in 'work/weights.1' -(a dot represents 10 inputs) +(a dot per input) Iteration #1 of 1. - 3 + .... 3 WEIGHTS - Glue = -0.32907 - WordPenalty = +0.049596 - LanguageModel = +0.33496 - LanguageModel_OOV = -0.44357 - PhraseModel_0 = -0.3068 - PhraseModel_1 = +0.59376 - PhraseModel_2 = +0.86416 - PhraseModel_3 = -0.21072 - PhraseModel_4 = -0.65734 - PhraseModel_5 = +0.03475 - PhraseModel_6 = -0.10653 - PassThrough = -0.46082 + Glue = -0.90863 + WordPenalty = +0.10819 + LanguageModel = +0.5239 + LanguageModel_OOV = -0.41623 + PhraseModel_0 = -0.86868 + PhraseModel_1 = +0.40784 + PhraseModel_2 = +1.1793 + PhraseModel_3 = -0.24698 + PhraseModel_4 = -1.2353 + PhraseModel_5 = +0.03375 + PhraseModel_6 = -0.17883 + PassThrough = -0.44638 --- - 1best avg score: 0.25055 (+0.25055) - 1best avg model score: -1.4459 (-1.4459) - avg # pairs: 1689 - avg # rank err: 755.67 - avg # margin viol: 829.33 - k-best loss imp: 100% - non0 feature count: 12 + 1best avg score: 0.12788 (+0.12788) + 1best avg model score: 41.302 + avg # pairs: 2246.3 + non-0 feature count: 12 avg list sz: 100 - avg f count: 10.53 -(time 0.13 min, 2.7 s/S) - -Writing weights file to 'work/weights.2.2' ... -done + avg f count: 10.98 +(time 0.35 min, 7 s/S) --- -Best iteration: 1 [SCORE 'stupid_bleu'=0.25055]. -This took 0.13333 min. +Best iteration: 1 [GOLD = 0.12788]. +This took 0.35 min. |