summaryrefslogtreecommitdiff
path: root/decoder
diff options
context:
space:
mode:
Diffstat (limited to 'decoder')
-rw-r--r--decoder/aligner.cc2
-rw-r--r--decoder/cdec_ff.cc2
-rw-r--r--decoder/dwarf.cc2
-rw-r--r--decoder/dwarf.h2
-rwxr-xr-xdecoder/feature_accum.h10
-rwxr-xr-xdecoder/ff_from_fsa.h2
-rw-r--r--decoder/hg.cc4
-rw-r--r--decoder/scfg_translator.cc2
-rw-r--r--decoder/trule.h2
9 files changed, 15 insertions, 13 deletions
diff --git a/decoder/aligner.cc b/decoder/aligner.cc
index 7830b955..292ee123 100644
--- a/decoder/aligner.cc
+++ b/decoder/aligner.cc
@@ -250,7 +250,7 @@ void AlignerTools::WriteAlignment(const Lattice& src_lattice,
SparseVector<prob_t> posts;
const prob_t z = InsideOutside<prob_t, EdgeProb, SparseVector<prob_t>, TransitionEventWeightFunction>(*g, &posts);
for (int i = 0; i < edge_posteriors.size(); ++i)
- edge_posteriors[i] = posts[i] / z;
+ edge_posteriors[i] = posts.value(i) / z;
}
vector<set<int> > src_cov(g->edges_.size());
vector<set<int> > trg_cov(g->edges_.size());
diff --git a/decoder/cdec_ff.cc b/decoder/cdec_ff.cc
index 7ec54a5a..cbcb0fcd 100644
--- a/decoder/cdec_ff.cc
+++ b/decoder/cdec_ff.cc
@@ -1,6 +1,7 @@
#include <boost/shared_ptr.hpp>
#include "ff.h"
+#include "ff_identity.h"
#include "ff_spans.h"
#include "ff_lm.h"
#include "ff_klm.h"
@@ -70,6 +71,7 @@ void register_feature_functions() {
ff_registry.Register("LexicalPairIndicator", new FFFactory<LexicalPairIndicator>);
ff_registry.Register("OutputIndicator", new FFFactory<OutputIndicator>);
ff_registry.Register("IdentityCycleDetector", new FFFactory<IdentityCycleDetector>);
+ ff_registry.Register("FEIdentity", new FFFactory<FEIdentity>);
ff_registry.Register("InputIndicator", new FFFactory<InputIndicator>);
ff_registry.Register("LexicalTranslationTrigger", new FFFactory<LexicalTranslationTrigger>);
ff_registry.Register("WordPairFeatures", new FFFactory<WordPairFeatures>);
diff --git a/decoder/dwarf.cc b/decoder/dwarf.cc
index 7968fee2..fb0404a6 100644
--- a/decoder/dwarf.cc
+++ b/decoder/dwarf.cc
@@ -3064,7 +3064,7 @@ void Alignment::SetCurrAlVector() {
}
}
-const void CountTable::print() {
+void CountTable::print() const {
cerr << "+++ Model +++" << endl;
for (map<WordID,int*>::const_iterator iter=model.begin(); iter!=model.end(); iter++) {
cerr << TD::Convert(iter->first) << " ";
diff --git a/decoder/dwarf.h b/decoder/dwarf.h
index 83a0cae9..49d2a3b7 100644
--- a/decoder/dwarf.h
+++ b/decoder/dwarf.h
@@ -27,7 +27,7 @@ public:
map<WordID,int*> model;
int mode;
int numColumn;
- const void print();
+ void print() const;
void setup(int _numcolumn, int _mode) {
mode = _mode; numColumn = _numcolumn;
}
diff --git a/decoder/feature_accum.h b/decoder/feature_accum.h
index 851b29db..4b8338eb 100755
--- a/decoder/feature_accum.h
+++ b/decoder/feature_accum.h
@@ -7,15 +7,15 @@
struct SparseFeatureAccumulator : public FeatureVector {
typedef FeatureVector State;
- SparseFeatureAccumulator() { }
+ SparseFeatureAccumulator() { assert(!"this code is disabled"); }
template <class FF>
FeatureVector const& describe(FF const& ) { return *this; }
void Store(FeatureVector *fv) const {
- fv->set_from(*this);
+//NO fv->set_from(*this);
}
template <class FF>
void Store(FF const& /* ff */,FeatureVector *fv) const {
- fv->set_from(*this);
+//NO fv->set_from(*this);
}
template <class FF>
void Add(FF const& /* ff */,FeatureVector const& fv) {
@@ -33,10 +33,10 @@ struct SparseFeatureAccumulator : public FeatureVector {
}
*/
void Add(int i,Featval v) {
- (*this)[i]+=v;
+//NO (*this)[i]+=v;
}
void Add(Features const& fids,int i,Featval v) {
- (*this)[i]+=v;
+//NO (*this)[i]+=v;
}
};
diff --git a/decoder/ff_from_fsa.h b/decoder/ff_from_fsa.h
index f2db8a4b..f8d79e03 100755
--- a/decoder/ff_from_fsa.h
+++ b/decoder/ff_from_fsa.h
@@ -295,7 +295,7 @@ private:
# include "ff_sample_fsa.h"
int main() {
std::cerr<<"Testing left_end...\n";
- std::cerr<<"sizeof(FeatureVector)="<<sizeof(FeatureVector)<<"\nsizeof(FeatureVectorList)="<<sizeof(FeatureVectorList)<<"\n";
+ std::cerr<<"sizeof(FeatureVector)="<<sizeof(FeatureVector)<<"\n";
WordPenaltyFromFsa::test();
return 0;
}
diff --git a/decoder/hg.cc b/decoder/hg.cc
index a4028b0e..3ad17f1a 100644
--- a/decoder/hg.cc
+++ b/decoder/hg.cc
@@ -163,7 +163,7 @@ prob_t Hypergraph::ComputeEdgePosteriors(double scale, vector<prob_t>* posts) co
ScaledTransitionEventWeightFunction>(*this, &pv, weight, w2);
posts->resize(edges_.size());
for (int i = 0; i < edges_.size(); ++i)
- (*posts)[i] = prob_t(pv.get(i));
+ (*posts)[i] = prob_t(pv.value(i));
return prob_t(inside);
}
@@ -176,7 +176,7 @@ prob_t Hypergraph::ComputeBestPathThroughEdges(vector<prob_t>* post) const {
ViterbiTransitionEventWeightFunction>(*this, &pv);
post->resize(edges_.size());
for (int i = 0; i < edges_.size(); ++i)
- (*post)[i] = pv.get(i).v_;
+ (*post)[i] = pv.value(i).v_;
return viterbi_weight.v_;
}
diff --git a/decoder/scfg_translator.cc b/decoder/scfg_translator.cc
index 0f7e40bd..d4bec40d 100644
--- a/decoder/scfg_translator.cc
+++ b/decoder/scfg_translator.cc
@@ -226,7 +226,7 @@ struct SCFGTranslatorImpl {
new_edge->j_ = edge.j_;
new_edge->feature_values_ = fine_rule_ptr->GetFeatureValues();
new_edge->feature_values_.set_value(FD::Convert("LatticeCost"),
- edge.feature_values_[FD::Convert("LatticeCost")]);
+ edge.feature_values_.value(FD::Convert("LatticeCost")));
Hypergraph::Node* head_node;
Split2Node::iterator it = s2n.find(StateSplit(node.id_, cat));
if (it == s2n.end()){
diff --git a/decoder/trule.h b/decoder/trule.h
index 7cd24bc9..4df4ec90 100644
--- a/decoder/trule.h
+++ b/decoder/trule.h
@@ -132,7 +132,7 @@ class TRule {
int Arity() const { return arity_; }
bool IsUnary() const { return (Arity() == 1) && (f_.size() == 1); }
const SparseVector<double>& GetFeatureValues() const { return scores_; }
- double Score(int i) const { return scores_.get(i); }
+ double Score(int i) const { return scores_.value(i); }
WordID GetLHS() const { return lhs_; }
void ComputeArity();