From a0a109329c942ddc956205cc66ccac872fb8f222 Mon Sep 17 00:00:00 2001 From: Patrick Simianer
Date: Mon, 21 Nov 2011 12:21:08 +0100
Subject: added pro stuff,clean up
---
dtrain/README.md | 125 +++++++++++++++++++++++++++++++++++++++++--------------
1 file changed, 93 insertions(+), 32 deletions(-)
(limited to 'dtrain/README.md')
diff --git a/dtrain/README.md b/dtrain/README.md
index 46f783b0..c50f3cad 100644
--- a/dtrain/README.md
+++ b/dtrain/README.md
@@ -23,67 +23,60 @@ Ideas
-----
* *MULTIPARTITE* ranking (1 vs rest, cluster model/score)
* *REMEMBER* sampled translations (merge kbest lists)
-* *SELECT* iteration with highest real BLEU on devtest?
-* *GENERATED* data? (perfect translation always in kbest)
+* *SELECT* iteration with highest _real_ BLEU on devtest?
+* *SYNTHETIC* data? (perfect translation always in kbest)
* *CACHE* ngrams for scoring
-* hadoop *PIPES* imlementation
+* hadoop *PIPES* implementation
* *ITERATION* variants (shuffle resulting weights, re-iterate)
-* *MORE THAN ONE* reference for BLEU?
-* *RANDOM RESTARTS* or directions
+* *MORE THAN ONE* reference for BLEU, paraphrases?
+* *RANDOM RESTARTS* or random directions
* use separate *TEST SET* for each shard
* *REDUCE* training set (50k?)
* *SYNTAX* features (CD)
* distribute *DEV* set to all nodes, avg
-* *PARAPHRASES* for better approx BLEU?
-
-Uncertain, known bugs, problems
+Notes
-------------------------------
* cdec kbest vs 1best (no -k param), rescoring (ref?)? => ok(?)
-* no sparse vector in decoder => ok/fixed
-* PhraseModel features, mapping?
+* no sparse vector in decoder => fixed/'ok'
+* PhraseModel features 0..99, mapping?
* flex scanner jams on bad input, we could skip that
-* input/grammar caching (strings -> WordIDs)
-* look at forest sampling...
-* devtest loo or not? why loo grammars larger? (sort psgs | uniq -> grammar)
+* input/grammar caching (vector
@@ -274,3 +266,72 @@ loo vs non-loo? => generalization
train on dev, test on devtest
train on devtest, test on dev
as above ^^^
+
+
+ ---
+
+as PRO
+ - UPDATES: perceptron
+ - LEARNING RATE: 0.0005
+ - GAMMA: -
+ - #ITERATIONS: 30
+ - SCORER: stupid_bleu@4
+ - K: 100, 1500?(top X pairs)
+ - SAMPLE: kbest uniq, kbest no
+ - PAIR SAMPLING: all, PRO?TODO
+ - SELECT: best
+ - FEATURES: baseline, RuleShape+SpanFeatures
+ ---
+ - Note: no weight interpolation
+ no early stopping based on kbest lists (epsilon?TODO)
+
+dtrain tune reg
+ - updates: SVM
+ - pair sampling important!
+ - learning_rate= 100 50 10 5 1 0.5 0.1 0.05 0.01 0.005 0.001 0.0005 0.0001 0.00005 0.00001 0.000005 0.000001 0.0000005 0.0000001 0.0000000001
+
+ - gamma=
+
+ - scorer: stupid_bleu 3
+ - test weights: last
+ -
+ -
+ - test: devtest
+
+
+---
+weights visualization (blocks, color coded)
+zig zag!?
+repeat all basic exps with training set
+merge?
+
+
+
+
+--sample_from
+--k
+--filter
+--pair_sampling
+--N
+--epochs
+--scorer
+--learning_rate
+--gamma
+--select_weights
+[--unit_weight_vector]
+[--l1_reg]
+[--l1_reg_strength]
+
+---------
+corr best = really best?
+108010gaps
+
+coltrane: 9
+gillespie: 9
+staley: 2
+io: 6
+ioh: 4
+ slots
+
+
+when does overfitting begin?
--
cgit v1.2.3