summaryrefslogtreecommitdiff
path: root/klm/lm/ngram_query.hh
blob: dfcda170e7c81fe56da16b69762ac0dae570dced (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
#ifndef LM_NGRAM_QUERY__
#define LM_NGRAM_QUERY__

#include "lm/enumerate_vocab.hh"
#include "lm/model.hh"
#include "util/usage.hh"

#include <cstdlib>
#include <iostream>
#include <ostream>
#include <istream>
#include <string>

namespace lm {
namespace ngram {

template <class Model> void Query(const Model &model, bool sentence_context, std::istream &in_stream, std::ostream &out_stream) {
  std::cerr << "Loading statistics:\n";
  util::PrintUsage(std::cerr);
  typename Model::State state, out;
  lm::FullScoreReturn ret;
  std::string word;

  while (in_stream) {
    state = sentence_context ? model.BeginSentenceState() : model.NullContextState();
    float total = 0.0;
    bool got = false;
    unsigned int oov = 0;
    while (in_stream >> word) {
      got = true;
      lm::WordIndex vocab = model.GetVocabulary().Index(word);
      if (vocab == 0) ++oov;
      ret = model.FullScore(state, vocab, out);
      total += ret.prob;
      out_stream << word << '=' << vocab << ' ' << static_cast<unsigned int>(ret.ngram_length)  << ' ' << ret.prob << '\t';
      state = out;
      char c;
      while (true) {
        c = in_stream.get();
        if (!in_stream) break;
        if (c == '\n') break;
        if (!isspace(c)) {
          in_stream.unget();
          break;
        }
      }
      if (c == '\n') break;
    }
    if (!got && !in_stream) break;
    if (sentence_context) {
      ret = model.FullScore(state, model.GetVocabulary().EndSentence(), out);
      total += ret.prob;
      out_stream << "</s>=" << model.GetVocabulary().EndSentence() << ' ' << static_cast<unsigned int>(ret.ngram_length)  << ' ' << ret.prob << '\t';
    }
    out_stream << "Total: " << total << " OOV: " << oov << '\n';
  }
  std::cerr << "After queries:\n";
  util::PrintUsage(std::cerr);
}

template <class M> void Query(const char *file, bool sentence_context, std::istream &in_stream, std::ostream &out_stream) {
  Config config;
  M model(file, config);
  Query(model, sentence_context, in_stream, out_stream);
}

} // namespace ngram
} // namespace lm

#endif // LM_NGRAM_QUERY__