#include #include #include #include #include #include #include #include #include #if HAVE_OPEN_MP #include #else const unsigned omp_get_num_threads() { return 1; } #endif #include "alignment.h" #include "data_array.h" #include "features/count_source_target.h" #include "features/feature.h" #include "features/is_source_singleton.h" #include "features/is_source_target_singleton.h" #include "features/max_lex_source_given_target.h" #include "features/max_lex_target_given_source.h" #include "features/sample_source_count.h" #include "features/target_given_source_coherent.h" #include "filelib.h" #include "grammar.h" #include "grammar_extractor.h" #include "precomputation.h" #include "rule.h" #include "scorer.h" #include "suffix_array.h" #include "time_util.h" #include "translation_table.h" #include "vocabulary.h" namespace ar = boost::archive; namespace fs = boost::filesystem; namespace po = boost::program_options; using namespace extractor; using namespace features; using namespace std; // Returns the file path in which a given grammar should be written. fs::path GetGrammarFilePath(const fs::path& grammar_path, int file_number, bool use_zip) { string file_name = "grammar." + to_string(file_number) + (use_zip ? ".gz" : ""); return grammar_path / file_name; } int main(int argc, char** argv) { po::options_description general_options("General options"); int max_threads = 1; #pragma omp parallel max_threads = omp_get_num_threads(); string threads_option = "Number of threads used for grammar extraction " "max(" + to_string(max_threads) + ")"; general_options.add_options() ("threads,t", po::value()->required()->default_value(1), threads_option.c_str()) ("grammars,g", po::value()->required(), "Grammars output path") ("gzip,z", "Gzip grammars") ("max_rule_span", po::value()->default_value(15), "Maximum rule span") ("max_rule_symbols", po::value()->default_value(5), "Maximum number of symbols (terminals + nontermals) in a rule") ("min_gap_size", po::value()->default_value(1), "Minimum gap size") ("max_nonterminals", po::value()->default_value(2), "Maximum number of nonterminals in a rule") ("max_samples", po::value()->default_value(300), "Maximum number of samples") ("tight_phrases", po::value()->default_value(true), "False if phrases may be loose (better, but slower)") ("leave_one_out", po::value()->zero_tokens(), "do leave-one-out estimation of grammars " "(e.g. for extracting grammars for the training set"); po::options_description cmdline_options("Command line options"); cmdline_options.add_options() ("help", "Show available options") ("config,c", po::value()->required(), "Path to config file"); cmdline_options.add(general_options); po::options_description config_options("Config file options"); config_options.add_options() ("target", po::value()->required(), "Path to target data file in binary format") ("source", po::value()->required(), "Path to source suffix array file in binary format") ("alignment", po::value()->required(), "Path to alignment file in binary format") ("precomputation", po::value()->required(), "Path to precomputation file in binary format") ("vocabulary", po::value()->required(), "Path to vocabulary file in binary format") ("ttable", po::value()->required(), "Path to translation table in binary format"); config_options.add(general_options); po::variables_map vm; po::store(po::parse_command_line(argc, argv, cmdline_options), vm); if (vm.count("help")) { po::options_description all_options; all_options.add(cmdline_options).add(config_options); cout << all_options << endl; return 0; } po::notify(vm); ifstream config_stream(vm["config"].as()); po::store(po::parse_config_file(config_stream, config_options), vm); po::notify(vm); int num_threads = vm["threads"].as(); cerr << "Grammar extraction will use " << num_threads << " threads." << endl; Clock::time_point read_start_time = Clock::now(); Clock::time_point start_time = Clock::now(); cerr << "Reading target data in binary format..." << endl; shared_ptr target_data_array = make_shared(); ifstream target_fstream(vm["target"].as()); ar::binary_iarchive target_stream(target_fstream); target_stream >> *target_data_array; Clock::time_point end_time = Clock::now(); cerr << "Reading target data took " << GetDuration(start_time, end_time) << " seconds" << endl; start_time = Clock::now(); cerr << "Reading source suffix array in binary format..." << endl; shared_ptr source_suffix_array = make_shared(); ifstream source_fstream(vm["source"].as()); ar::binary_iarchive source_stream(source_fstream); source_stream >> *source_suffix_array; end_time = Clock::now(); cerr << "Reading source suffix array took " << GetDuration(start_time, end_time) << " seconds" << endl; start_time = Clock::now(); cerr << "Reading alignment in binary format..." << endl; shared_ptr alignment = make_shared(); ifstream alignment_fstream(vm["alignment"].as()); ar::binary_iarchive alignment_stream(alignment_fstream); alignment_stream >> *alignment; end_time = Clock::now(); cerr << "Reading alignment took " << GetDuration(start_time, end_time) << " seconds" << endl; start_time = Clock::now(); cerr << "Reading precomputation in binary format..." << endl; shared_ptr precomputation = make_shared(); ifstream precomputation_fstream(vm["precomputation"].as()); ar::binary_iarchive precomputation_stream(precomputation_fstream); precomputation_stream >> *precomputation; end_time = Clock::now(); cerr << "Reading precomputation took " << GetDuration(start_time, end_time) << " seconds" << endl; start_time = Clock::now(); cerr << "Reading vocabulary in binary format..." << endl; shared_ptr vocabulary = make_shared(); ifstream vocabulary_fstream(vm["vocabulary"].as()); ar::binary_iarchive vocabulary_stream(vocabulary_fstream); vocabulary_stream >> *vocabulary; end_time = Clock::now(); cerr << "Reading vocabulary took " << GetDuration(start_time, end_time) << " seconds" << endl; start_time = Clock::now(); cerr << "Reading translation table in binary format..." << endl; shared_ptr table = make_shared(); ifstream ttable_fstream(vm["ttable"].as()); ar::binary_iarchive ttable_stream(ttable_fstream); ttable_stream >> *table; end_time = Clock::now(); cerr << "Reading translation table took " << GetDuration(start_time, end_time) << " seconds" << endl; Clock::time_point read_end_time = Clock::now(); cerr << "Total time spent loading data structures into memory: " << GetDuration(read_start_time, read_end_time) << " seconds" << endl; Clock::time_point extraction_start_time = Clock::now(); // Features used to score each grammar rule. vector> features = { make_shared(), make_shared(), make_shared(), make_shared(table), make_shared(table), make_shared(), make_shared() }; shared_ptr scorer = make_shared(features); GrammarExtractor extractor( source_suffix_array, target_data_array, alignment, precomputation, scorer, vocabulary, vm["min_gap_size"].as(), vm["max_rule_span"].as(), vm["max_nonterminals"].as(), vm["max_rule_symbols"].as(), vm["max_samples"].as(), vm["tight_phrases"].as()); const bool use_zip = vm.count("gzip"); // Creates the grammars directory if it doesn't exist. fs::path grammar_path = vm["grammars"].as(); if (!fs::is_directory(grammar_path)) { fs::create_directory(grammar_path); } grammar_path = fs::canonical(grammar_path); // Reads all sentences for which we extract grammar rules (the paralellization // is simplified if we read all sentences upfront). string sentence; vector sentences; while (getline(cin, sentence)) { sentences.push_back(sentence); } // Extracts the grammar for each sentence and saves it to a file. vector suffixes(sentences.size()); bool leave_one_out = vm.count("leave_one_out"); #pragma omp parallel for schedule(dynamic) num_threads(num_threads) for (size_t i = 0; i < sentences.size(); ++i) { string suffix; int position = sentences[i].find("|||"); if (position != sentences[i].npos) { suffix = sentences[i].substr(position); sentences[i] = sentences[i].substr(0, position); } suffixes[i] = suffix; unordered_set blacklisted_sentence_ids; if (leave_one_out) { blacklisted_sentence_ids.insert(i); } Grammar grammar = extractor.GetGrammar( sentences[i], blacklisted_sentence_ids); WriteFile wf(GetGrammarFilePath(grammar_path, i, use_zip).c_str()); *wf.stream() << grammar; } for (size_t i = 0; i < sentences.size(); ++i) { cout << " " << sentences[i] << " " << suffixes[i] << endl; } Clock::time_point extraction_stop_time = Clock::now(); cerr << "Overall extraction step took " << GetDuration(extraction_start_time, extraction_stop_time) << " seconds" << endl; return 0; }