diff options
author | redpony <redpony@ec762483-ff6d-05da-a07a-a48fb63a330f> | 2010-11-10 22:43:29 +0000 |
---|---|---|
committer | redpony <redpony@ec762483-ff6d-05da-a07a-a48fb63a330f> | 2010-11-10 22:43:29 +0000 |
commit | 816c7efe7ada1554f0d62282517b6051724c7102 (patch) | |
tree | 17e7dce5a7731609844a5b966b879f30c8aafd05 /klm/lm/binary_format.hh | |
parent | 47a656c0b6fdba8f91f2c5808234cbb1de682652 (diff) |
forgotten files
git-svn-id: https://ws10smt.googlecode.com/svn/trunk@707 ec762483-ff6d-05da-a07a-a48fb63a330f
Diffstat (limited to 'klm/lm/binary_format.hh')
-rw-r--r-- | klm/lm/binary_format.hh | 91 |
1 files changed, 91 insertions, 0 deletions
diff --git a/klm/lm/binary_format.hh b/klm/lm/binary_format.hh new file mode 100644 index 00000000..f95f05f7 --- /dev/null +++ b/klm/lm/binary_format.hh @@ -0,0 +1,91 @@ +#ifndef LM_BINARY_FORMAT__ +#define LM_BINARY_FORMAT__ + +#include "lm/config.hh" +#include "lm/read_arpa.hh" + +#include "util/file_piece.hh" +#include "util/mmap.hh" +#include "util/scoped.hh" + +#include <cstddef> +#include <vector> + +#include <inttypes.h> + +namespace lm { +namespace ngram { + +typedef enum {HASH_PROBING=0, HASH_SORTED=1, TRIE_SORTED=2} ModelType; + +struct FixedWidthParameters { + unsigned char order; + float probing_multiplier; + // What type of model is this? + ModelType model_type; + // Does the end of the file have the actual strings in the vocabulary? + bool has_vocabulary; +}; + +struct Parameters { + FixedWidthParameters fixed; + std::vector<uint64_t> counts; +}; + +struct Backing { + // File behind memory, if any. + util::scoped_fd file; + // Raw block of memory backing the language model data structures + util::scoped_memory memory; +}; + +namespace detail { + +bool IsBinaryFormat(int fd); + +void ReadHeader(int fd, Parameters ¶ms); + +void MatchCheck(ModelType model_type, const Parameters ¶ms); + +uint8_t *SetupBinary(const Config &config, const Parameters ¶ms, std::size_t memory_size, Backing &backing); + +uint8_t *SetupZeroed(const Config &config, ModelType model_type, const std::vector<uint64_t> &counts, std::size_t memory_size, Backing &backing); + +void ComplainAboutARPA(const Config &config, ModelType model_type); + +} // namespace detail + +bool RecognizeBinary(const char *file, ModelType &recognized); + +template <class To> void LoadLM(const char *file, const Config &config, To &to) { + Backing &backing = to.MutableBacking(); + backing.file.reset(util::OpenReadOrThrow(file)); + + Parameters params; + + try { + if (detail::IsBinaryFormat(backing.file.get())) { + detail::ReadHeader(backing.file.get(), params); + detail::MatchCheck(To::kModelType, params); + std::size_t memory_size = To::Size(params.counts, config); + uint8_t *start = detail::SetupBinary(config, params, memory_size, backing); + to.InitializeFromBinary(start, params, config, backing.file.get()); + } else { + detail::ComplainAboutARPA(config, To::kModelType); + util::FilePiece f(backing.file.release(), file, config.messages); + ReadARPACounts(f, params.counts); + std::size_t memory_size = To::Size(params.counts, config); + uint8_t *start = detail::SetupZeroed(config, To::kModelType, params.counts, memory_size, backing); + + to.InitializeFromARPA(file, f, start, params, config); + } + } catch (util::Exception &e) { + e << " in file " << file; + throw; + } + +} + +} // namespace ngram +} // namespace lm +#endif // LM_BINARY_FORMAT__ |