From c08cbb4e12801b8c8a655f635e9df7ffe10bea4f Mon Sep 17 00:00:00 2001 From: Patrick Simianer
Date: Fri, 30 Jan 2015 16:21:53 +0100
Subject: patches
---
patches/mira_hack_fix_wp.patch | 38 ++++++++++++++++++++++++++++++++++++++
1 file changed, 38 insertions(+)
create mode 100644 patches/mira_hack_fix_wp.patch
diff --git a/patches/mira_hack_fix_wp.patch b/patches/mira_hack_fix_wp.patch
new file mode 100644
index 00000000..dfe9965d
--- /dev/null
+++ b/patches/mira_hack_fix_wp.patch
@@ -0,0 +1,38 @@
+diff --git a/training/mira/kbest_cut_mira.cc b/training/mira/kbest_cut_mira.cc
+index 724b185..1224f0c 100644
+--- a/training/mira/kbest_cut_mira.cc
++++ b/training/mira/kbest_cut_mira.cc
+@@ -750,6 +750,7 @@ int main(int argc, char** argv) {
+ }
+ // Regular mode or LEARN line from stream mode
+ //TODO: allow batch updating
++ lambdas[FD::Convert("WordPenalty")] = -1.0; // HACK WP
+ lambdas.init_vector(&dense_weights);
+ dense_w_local = dense_weights;
+ decoder.SetId(cur_sent);
+@@ -781,7 +782,8 @@ int main(int argc, char** argv) {
+ acc_f->PlusEquals(*fear_sentscore);
+
+ if(optimizer == 4) { //passive-aggresive update (single dual coordinate step)
+-
++
++ dense_weights[FD::Convert("WordPenalty")] = -1.0; // HACK WP
+ double margin = cur_bad.features.dot(dense_weights) - cur_good.features.dot(dense_weights);
+ double mt_loss = (cur_good.mt_metric - cur_bad.mt_metric);
+ const double loss = margin + mt_loss;
+@@ -927,6 +929,7 @@ int main(int argc, char** argv) {
+
+ //reload weights based on update
+ dense_weights.clear();
++ lambdas[FD::Convert("WordPenalty")] = -1.0; // HACK WP
+ lambdas.init_vector(&dense_weights);
+ if (dense_weights.size() < 500)
+ ShowLargestFeatures(dense_weights);
+@@ -1001,6 +1004,7 @@ int main(int argc, char** argv) {
+ ostringstream os;
+ os << weights_dir << "/weights.mira-pass" << (cur_pass < 10 ? "0" : "") << cur_pass << "." << node_id << ".gz";
+ string msg = "# MIRA tuned weights ||| " + boost::lexical_cast