summaryrefslogtreecommitdiff
path: root/training/dtrain/README.md
blob: 2bae6b481f63d98ef0d08655e879851348718f1b (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
This is a simple (and parallelizable) tuning method for cdec
which is able to train the weights of very many (sparse) features.
It was used here:
  "Joint Feature Selection in Distributed Stochastic
   Learning for Large-Scale Discriminative Training in
   SMT"
(Simianer, Riezler, Dyer; ACL 2012)


Building
--------
Builds when building cdec, see ../BUILDING .
To build only parts needed for dtrain do
```
  autoreconf -ifv
  ./configure
  cd training/dtrain/; make
```

Ideas
-----
 * get approx_bleu to work?
 * implement minibatches (Minibatch and Parallelization for Online Large Margin Structured Learning)
 * learning rate 1/T?
 * use an oracle? mira-like (model vs. BLEU), feature repr. of reference!? 
 * implement lc_bleu properly
 * merge kbest lists of previous epochs (as MERT does)
 * ``walk entire regularization path''
 * rerank after each update?

Running
-------
See directories under test/ .

Legal
-----
Copyright (c) 2012-2013 by Patrick Simianer <p@simianer.de>

See the file LICENSE.txt in the root folder for the licensing terms that this software is
released under.