This is a simple (and parallelizable) tuning method for cdec which is able to train the weights of very many (sparse) features. It was used here: "Joint Feature Selection in Distributed Stochastic Learning for Large-Scale Discriminative Training in SMT" (Simianer, Riezler, Dyer; ACL 2012) Building -------- Builds when building cdec, see ../BUILDING . To build only parts needed for dtrain do ``` autoreconf -ifv ./configure [--disable-test] cd dtrain/; make ``` Running ------- To run this on a dev set locally: ``` #define DTRAIN_LOCAL ``` otherwise remove that line or undef, then recompile. You need a single grammar file or input annotated with per-sentence grammars (psg) as you would use with cdec. Additionally you need to give dtrain a file with references (--refs) when running locally. The input for use with hadoop streaming looks like this: ``` \t\t\t ``` To convert a psg to this format you need to replace all "\n" by "\t". Make sure there are no tabs in your data. For an example of local usage (with the 'distributed' format) the see test/example/ . This expects dtrain to be built without DTRAIN_LOCAL. Next ---- + (dtrain|decoder) meta-parameters testing + target side rule ngrams + sa-extract -> leave-one-out for grammar of training set? + make svm doable; no subgradient? + reranking while sgd? + try PRO, mira emulations + avg feature count Legal ----- Copyright (c) 2012 by Patrick Simianer See the file ../LICENSE.txt for the licensing terms that this software is released under.