mirror of
https://github.com/tesseract-ocr/tesseract.git
synced 2024-11-24 11:09:06 +08:00
0aadbd0169
search so we have them when trying to replace words with alternates in the bigram correction pass. git-svn-id: https://tesseract-ocr.googlecode.com/svn/trunk@739 d0cd1f9f-072b-0410-8dd7-cf729c803f20
2006 lines
84 KiB
C++
2006 lines
84 KiB
C++
///////////////////////////////////////////////////////////////////////
|
|
// File: language_model.cpp
|
|
// Description: Functions that utilize the knowledge about the properties,
|
|
// structure and statistics of the language to help recognition.
|
|
// Author: Daria Antonova
|
|
// Created: Mon Nov 11 11:26:43 PST 2009
|
|
//
|
|
// (C) Copyright 2009, Google Inc.
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
//
|
|
///////////////////////////////////////////////////////////////////////
|
|
|
|
#include <math.h>
|
|
|
|
#include "language_model.h"
|
|
|
|
#include "dawg.h"
|
|
#include "intproto.h"
|
|
#include "matrix.h"
|
|
#include "params.h"
|
|
#include "params_training_featdef.h"
|
|
|
|
namespace tesseract {
|
|
|
|
ELISTIZE(ViterbiStateEntry);
|
|
|
|
const float LanguageModel::kInitialPainPointPriorityAdjustment = 5.0f;
|
|
const float LanguageModel::kDefaultPainPointPriorityAdjustment = 2.0f;
|
|
const float LanguageModel::kBestChoicePainPointPriorityAdjustment = 0.5f;
|
|
const float LanguageModel::kCriticalPainPointPriorityAdjustment = 0.1f;
|
|
const float LanguageModel::kMaxAvgNgramCost = 25.0f;
|
|
const int LanguageModel::kMinFixedLengthDawgLength = 2;
|
|
const float LanguageModel::kLooseMaxCharWhRatio = 2.5f;
|
|
|
|
LanguageModel::LanguageModel(const UnicityTable<FontInfo> *fontinfo_table,
|
|
Dict *dict)
|
|
: INT_MEMBER(language_model_debug_level, 0, "Language model debug level",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
BOOL_INIT_MEMBER(language_model_ngram_on, false,
|
|
"Turn on/off the use of character ngram model",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
INT_MEMBER(language_model_ngram_order, 8,
|
|
"Maximum order of the character ngram model",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
INT_MEMBER(language_model_viterbi_list_max_num_prunable, 10,
|
|
"Maximum number of prunable (those for which"
|
|
" PrunablePath() is true) entries in each viterbi list"
|
|
" recorded in BLOB_CHOICEs",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
INT_MEMBER(language_model_viterbi_list_max_size, 500,
|
|
"Maximum size of viterbi lists recorded in BLOB_CHOICEs",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
double_MEMBER(language_model_ngram_small_prob, 0.000001,
|
|
"To avoid overly small denominators use this as the "
|
|
"floor of the probability returned by the ngram model.",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
double_MEMBER(language_model_ngram_nonmatch_score, -40.0,
|
|
"Average classifier score of a non-matching unichar.",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
BOOL_MEMBER(language_model_ngram_use_only_first_uft8_step, false,
|
|
"Use only the first UTF8 step of the given string"
|
|
" when computing log probabilities.",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
double_MEMBER(language_model_ngram_scale_factor, 0.03,
|
|
"Strength of the character ngram model relative to the"
|
|
" character classifier ",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
BOOL_MEMBER(language_model_ngram_space_delimited_language, true,
|
|
"Words are delimited by space",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
INT_MEMBER(language_model_min_compound_length, 3,
|
|
"Minimum length of compound words",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
INT_MEMBER(language_model_fixed_length_choices_depth, 3,
|
|
"Depth of blob choice lists to explore"
|
|
" when fixed length dawgs are on",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
double_MEMBER(language_model_penalty_non_freq_dict_word, 0.1,
|
|
"Penalty for words not in the frequent word dictionary",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
double_MEMBER(language_model_penalty_non_dict_word, 0.15,
|
|
"Penalty for non-dictionary words",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
double_MEMBER(language_model_penalty_punc, 0.2,
|
|
"Penalty for inconsistent punctuation",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
double_MEMBER(language_model_penalty_case, 0.1,
|
|
"Penalty for inconsistent case",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
double_MEMBER(language_model_penalty_script, 0.5,
|
|
"Penalty for inconsistent script",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
double_MEMBER(language_model_penalty_chartype, 0.3,
|
|
"Penalty for inconsistent character type",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
// TODO(daria, rays): enable font consistency checking
|
|
// after improving font analysis.
|
|
double_MEMBER(language_model_penalty_font, 0.00,
|
|
"Penalty for inconsistent font",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
double_MEMBER(language_model_penalty_spacing, 0.05,
|
|
"Penalty for inconsistent spacing",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
double_MEMBER(language_model_penalty_increment, 0.01,
|
|
"Penalty increment",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
BOOL_INIT_MEMBER(language_model_use_sigmoidal_certainty, false,
|
|
"Use sigmoidal score for certainty",
|
|
dict->getImage()->getCCUtil()->params()),
|
|
fontinfo_table_(fontinfo_table), dict_(dict),
|
|
fixed_pitch_(false), max_char_wh_ratio_(0.0),
|
|
acceptable_choice_found_(false) {
|
|
ASSERT_HOST(dict_ != NULL);
|
|
dawg_args_ = new DawgArgs(NULL, NULL, new DawgInfoVector(),
|
|
new DawgInfoVector(),
|
|
0.0, NO_PERM, kAnyWordLength, -1);
|
|
beginning_active_dawgs_ = new DawgInfoVector();
|
|
beginning_constraints_ = new DawgInfoVector();
|
|
fixed_length_beginning_active_dawgs_ = new DawgInfoVector();
|
|
empty_dawg_info_vec_ = new DawgInfoVector();
|
|
}
|
|
|
|
LanguageModel::~LanguageModel() {
|
|
delete beginning_active_dawgs_;
|
|
delete beginning_constraints_;
|
|
delete fixed_length_beginning_active_dawgs_;
|
|
delete empty_dawg_info_vec_;
|
|
delete dawg_args_->updated_active_dawgs;
|
|
delete dawg_args_->updated_constraints;
|
|
delete dawg_args_;
|
|
}
|
|
|
|
void LanguageModel::InitForWord(
|
|
const WERD_CHOICE *prev_word,
|
|
bool fixed_pitch, float best_choice_cert, float max_char_wh_ratio,
|
|
float rating_cert_scale, HEAP *pain_points, CHUNKS_RECORD *chunks_record,
|
|
BlamerBundle *blamer_bundle, bool debug_blamer) {
|
|
fixed_pitch_ = fixed_pitch;
|
|
max_char_wh_ratio_ = max_char_wh_ratio;
|
|
rating_cert_scale_ = rating_cert_scale;
|
|
acceptable_choice_found_ = false;
|
|
correct_segmentation_explored_ = false;
|
|
|
|
// For each cell, generate a "pain point" if the cell is not classified
|
|
// and has a left or right neighbor that was classified.
|
|
MATRIX *ratings = chunks_record->ratings;
|
|
for (int col = 0; col < ratings->dimension(); ++col) {
|
|
for (int row = col+1; row < ratings->dimension(); ++row) {
|
|
if ((row > 0 && ratings->get(col, row-1) != NOT_CLASSIFIED) ||
|
|
(col+1 < ratings->dimension() &&
|
|
ratings->get(col+1, row) != NOT_CLASSIFIED)) {
|
|
float worst_piece_cert;
|
|
bool fragmented;
|
|
GetWorstPieceCertainty(col, row, chunks_record->ratings,
|
|
&worst_piece_cert, &fragmented);
|
|
GeneratePainPoint(col, row, true, kInitialPainPointPriorityAdjustment,
|
|
worst_piece_cert, fragmented, best_choice_cert,
|
|
max_char_wh_ratio_, NULL, NULL,
|
|
chunks_record, pain_points);
|
|
}
|
|
}
|
|
}
|
|
|
|
// Initialize vectors with beginning DawgInfos.
|
|
beginning_active_dawgs_->clear();
|
|
dict_->init_active_dawgs(kAnyWordLength, beginning_active_dawgs_, false);
|
|
beginning_constraints_->clear();
|
|
dict_->init_constraints(beginning_constraints_);
|
|
if (dict_->GetMaxFixedLengthDawgIndex() >= 0) {
|
|
fixed_length_beginning_active_dawgs_->clear();
|
|
for (int i = 0; i < beginning_active_dawgs_->size(); ++i) {
|
|
int dawg_index = (*beginning_active_dawgs_)[i].dawg_index;
|
|
if (dawg_index <= dict_->GetMaxFixedLengthDawgIndex() &&
|
|
dawg_index >= kMinFixedLengthDawgLength) {
|
|
*fixed_length_beginning_active_dawgs_ += (*beginning_active_dawgs_)[i];
|
|
}
|
|
}
|
|
}
|
|
|
|
max_penalty_adjust_ = (dict_->segment_penalty_dict_nonword -
|
|
dict_->segment_penalty_dict_case_ok);
|
|
|
|
// Fill prev_word_str_ with the last language_model_ngram_order
|
|
// unichars from prev_word.
|
|
if (language_model_ngram_on) {
|
|
if (prev_word != NULL && prev_word->unichar_string() != NULL) {
|
|
prev_word_str_ = prev_word->unichar_string();
|
|
if (language_model_ngram_space_delimited_language) prev_word_str_ += ' ';
|
|
} else {
|
|
prev_word_str_ += ' ';
|
|
}
|
|
const char *str_ptr = prev_word_str_.string();
|
|
const char *str_end = str_ptr + prev_word_str_.length();
|
|
int step;
|
|
prev_word_unichar_step_len_ = 0;
|
|
while (str_ptr != str_end && (step = UNICHAR::utf8_step(str_ptr))) {
|
|
str_ptr += step;
|
|
++prev_word_unichar_step_len_;
|
|
}
|
|
ASSERT_HOST(str_ptr == str_end);
|
|
}
|
|
|
|
// Initialize blamer-related information: map character boxes recorded in
|
|
// blamer_bundle->norm_truth_word to the corresponding i,j indices in the
|
|
// ratings matrix. We expect this step to succeed, since when running the
|
|
// chopper we checked that the correct chops are present.
|
|
if (blamer_bundle != NULL &&
|
|
blamer_bundle->incorrect_result_reason == IRR_CORRECT &&
|
|
blamer_bundle->truth_has_char_boxes) {
|
|
STRING blamer_debug;
|
|
blamer_debug += "Blamer computing correct_segmentation_cols\n";
|
|
int curr_box_col = 0;
|
|
int next_box_col = 0;
|
|
TBLOB *blob = chunks_record->chunks;
|
|
inT16 next_box_x = (blob != NULL) ? blob->bounding_box().right() : 0;
|
|
for (int truth_idx = 0;
|
|
blob != NULL && truth_idx < blamer_bundle->norm_truth_word.length();
|
|
blob = blob->next) {
|
|
++next_box_col;
|
|
inT16 curr_box_x = next_box_x;
|
|
if (blob->next != NULL) next_box_x = blob->next->bounding_box().right();
|
|
inT16 truth_x = blamer_bundle->norm_truth_word.BlobBox(truth_idx).right();
|
|
blamer_debug.add_str_int("Box x coord vs. truth: ", curr_box_x);
|
|
blamer_debug.add_str_int(" ", truth_x);
|
|
blamer_debug += "\n";
|
|
if (curr_box_x > (truth_x + blamer_bundle->norm_box_tolerance)) {
|
|
break; // failed to find a matching box
|
|
} else if (curr_box_x >=
|
|
(truth_x - blamer_bundle->norm_box_tolerance) && // matched
|
|
(blob->next == NULL || // next box can't be included
|
|
next_box_x > truth_x + blamer_bundle->norm_box_tolerance)) {
|
|
blamer_bundle->correct_segmentation_cols.push_back(curr_box_col);
|
|
blamer_bundle->correct_segmentation_rows.push_back(next_box_col-1);
|
|
++truth_idx;
|
|
blamer_debug.add_str_int("col=", curr_box_col);
|
|
blamer_debug.add_str_int(" row=", next_box_col-1);
|
|
blamer_debug += "\n";
|
|
curr_box_col = next_box_col;
|
|
}
|
|
}
|
|
if (blob != NULL || // trailing blobs
|
|
blamer_bundle->correct_segmentation_cols.length() !=
|
|
blamer_bundle->norm_truth_word.length()) {
|
|
blamer_debug.add_str_int("Blamer failed to find correct segmentation"
|
|
" (tolerance=", blamer_bundle->norm_box_tolerance);
|
|
if (blob == NULL) blamer_debug += " blob == NULL";
|
|
blamer_debug += ")\n";
|
|
blamer_debug.add_str_int(
|
|
" path length ", blamer_bundle->correct_segmentation_cols.length());
|
|
blamer_debug.add_str_int(" vs. truth ",
|
|
blamer_bundle->norm_truth_word.length());
|
|
blamer_debug += "\n";
|
|
blamer_bundle->SetBlame(IRR_UNKNOWN, blamer_debug, NULL, debug_blamer);
|
|
blamer_bundle->correct_segmentation_cols.clear();
|
|
blamer_bundle->correct_segmentation_rows.clear();
|
|
}
|
|
} // end if (blamer_bundle != NULL)
|
|
|
|
// Start a new hypothesis list for this run of segmentation search.
|
|
if (blamer_bundle != NULL) {
|
|
blamer_bundle->params_training_bundle.StartHypothesisList();
|
|
}
|
|
}
|
|
|
|
void LanguageModel::CleanUp() {
|
|
for (int i = 0; i < updated_flags_.size(); ++i) {
|
|
*(updated_flags_[i]) = false;
|
|
}
|
|
updated_flags_.clear();
|
|
}
|
|
|
|
void LanguageModel::DeleteState(BLOB_CHOICE_LIST *choices) {
|
|
BLOB_CHOICE_IT b_it(choices);
|
|
for (b_it.mark_cycle_pt(); !b_it.cycled_list(); b_it.forward()) {
|
|
if (b_it.data()->language_model_state() != NULL) {
|
|
LanguageModelState *state = reinterpret_cast<LanguageModelState *>(
|
|
b_it.data()->language_model_state());
|
|
delete state;
|
|
b_it.data()->set_language_model_state(NULL);
|
|
}
|
|
}
|
|
}
|
|
|
|
LanguageModelFlagsType LanguageModel::UpdateState(
|
|
LanguageModelFlagsType changed,
|
|
int curr_col, int curr_row,
|
|
BLOB_CHOICE_LIST *curr_list,
|
|
BLOB_CHOICE_LIST *parent_list,
|
|
HEAP *pain_points,
|
|
BestPathByColumn *best_path_by_column[],
|
|
CHUNKS_RECORD *chunks_record,
|
|
BestChoiceBundle *best_choice_bundle,
|
|
BlamerBundle *blamer_bundle) {
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("\nUpdateState: col=%d row=%d (changed=0x%x parent=%p)\n",
|
|
curr_col, curr_row, changed, parent_list);
|
|
}
|
|
// Initialize helper variables.
|
|
bool word_end = (curr_row+1 >= chunks_record->ratings->dimension());
|
|
bool just_classified = (changed & kJustClassifiedFlag);
|
|
LanguageModelFlagsType new_changed = 0x0;
|
|
float denom = (language_model_ngram_on) ? ComputeDenom(curr_list) : 1.0f;
|
|
|
|
// Call AddViterbiStateEntry() for each parent+child ViterbiStateEntry.
|
|
ViterbiStateEntry_IT vit;
|
|
BLOB_CHOICE_IT c_it(curr_list);
|
|
int c_it_counter = 0;
|
|
bool first_iteration = true;
|
|
BLOB_CHOICE *first_lower = NULL;
|
|
BLOB_CHOICE *first_upper = NULL;
|
|
GetTopChoiceLowerUpper(changed, curr_list, &first_lower, &first_upper);
|
|
for (c_it.mark_cycle_pt(); !c_it.cycled_list(); c_it.forward()) {
|
|
if (dict_->GetMaxFixedLengthDawgIndex() >= 0 &&
|
|
c_it_counter++ >= language_model_fixed_length_choices_depth) {
|
|
break;
|
|
}
|
|
// Skip NULL unichars unless it is the only choice.
|
|
if (!curr_list->singleton() && c_it.data()->unichar_id() == 0) continue;
|
|
if (dict_->getUnicharset().get_fragment(c_it.data()->unichar_id())) {
|
|
continue; // skip fragments
|
|
}
|
|
// Set top choice flags.
|
|
LanguageModelFlagsType top_choice_flags = 0x0;
|
|
if (first_iteration && (changed | kSmallestRatingFlag)) {
|
|
top_choice_flags |= kSmallestRatingFlag;
|
|
}
|
|
if (first_lower == c_it.data()) top_choice_flags |= kLowerCaseFlag;
|
|
if (first_upper == c_it.data()) top_choice_flags |= kUpperCaseFlag;
|
|
|
|
if (parent_list == NULL) { // process the beginning of a word
|
|
new_changed |= AddViterbiStateEntry(
|
|
top_choice_flags, denom, word_end, curr_col, curr_row,
|
|
c_it.data(), NULL, NULL, pain_points, best_path_by_column,
|
|
chunks_record, best_choice_bundle, blamer_bundle);
|
|
} else { // get viterbi entries from each of the parent BLOB_CHOICEs
|
|
BLOB_CHOICE_IT p_it(parent_list);
|
|
for (p_it.mark_cycle_pt(); !p_it.cycled_list(); p_it.forward()) {
|
|
LanguageModelState *parent_lms =
|
|
reinterpret_cast<LanguageModelState *>(
|
|
p_it.data()->language_model_state());
|
|
if (parent_lms == NULL || parent_lms->viterbi_state_entries.empty()) {
|
|
continue;
|
|
}
|
|
vit.set_to_list(&(parent_lms->viterbi_state_entries));
|
|
int vit_counter = 0;
|
|
for (vit.mark_cycle_pt(); !vit.cycled_list(); vit.forward()) {
|
|
// Skip pruned entries and do not look at prunable entries if already
|
|
// examined language_model_viterbi_list_max_num_prunable of those.
|
|
if (PrunablePath(vit.data()->top_choice_flags,
|
|
vit.data()->dawg_info) &&
|
|
(++vit_counter > language_model_viterbi_list_max_num_prunable ||
|
|
(language_model_ngram_on && vit.data()->ngram_info->pruned))) {
|
|
continue;
|
|
}
|
|
// Only consider the parent if it has been updated or
|
|
// if the current ratings cell has just been classified.
|
|
if (!just_classified && !vit.data()->updated) continue;
|
|
// Create a new ViterbiStateEntry if BLOB_CHOICE in c_it.data()
|
|
// looks good according to the Dawgs or character ngram model.
|
|
new_changed |= AddViterbiStateEntry(
|
|
top_choice_flags, denom, word_end, curr_col, curr_row,
|
|
c_it.data(), p_it.data(), vit.data(), pain_points,
|
|
best_path_by_column, chunks_record,
|
|
best_choice_bundle, blamer_bundle);
|
|
}
|
|
} // done looking at parents for this c_it.data()
|
|
}
|
|
first_iteration = false;
|
|
}
|
|
return new_changed;
|
|
}
|
|
|
|
bool LanguageModel::ProblematicPath(const ViterbiStateEntry &vse,
|
|
UNICHAR_ID unichar_id, bool word_end) {
|
|
// The path is problematic if it is inconsistent and has a parent that
|
|
// is consistent (or a NULL parent).
|
|
if (!vse.Consistent() && (vse.parent_vse == NULL ||
|
|
vse.parent_vse->Consistent())) {
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("ProblematicPath: inconsistent\n");
|
|
}
|
|
return true;
|
|
}
|
|
// The path is problematic if it does not represent a dictionary word,
|
|
// while its parent does.
|
|
if (vse.dawg_info == NULL &&
|
|
(vse.parent_vse == NULL || vse.parent_vse->dawg_info != NULL)) {
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("ProblematicPath: dict word terminated\n");
|
|
}
|
|
return true;
|
|
}
|
|
// The path is problematic if ngram info indicates that this path is
|
|
// an unlikely sequence of character, while its parent is does not have
|
|
// extreme dips in ngram probabilities.
|
|
if (vse.ngram_info != NULL && vse.ngram_info->pruned &&
|
|
(vse.parent_vse == NULL || !vse.parent_vse->ngram_info->pruned)) {
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("ProblematicPath: small ngram prob\n");
|
|
}
|
|
return true;
|
|
}
|
|
// The path is problematic if there is a non-alpha character in the
|
|
// middle of the path (unless it is a digit in the middle of a path
|
|
// that represents a number).
|
|
if ((vse.parent_vse != NULL) && !word_end && // is middle
|
|
!(dict_->getUnicharset().get_isalpha(unichar_id) || // alpha
|
|
(dict_->getUnicharset().get_isdigit(unichar_id) && // ok digit
|
|
vse.dawg_info != NULL && vse.dawg_info->permuter == NUMBER_PERM))) {
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("ProblematicPath: non-alpha middle\n");
|
|
}
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
void LanguageModel::GetTopChoiceLowerUpper(LanguageModelFlagsType changed,
|
|
BLOB_CHOICE_LIST *curr_list,
|
|
BLOB_CHOICE **first_lower,
|
|
BLOB_CHOICE **first_upper) {
|
|
if (!(changed & kLowerCaseFlag || changed & kUpperCaseFlag)) return;
|
|
BLOB_CHOICE_IT c_it(curr_list);
|
|
const UNICHARSET &unicharset = dict_->getUnicharset();
|
|
BLOB_CHOICE *first_unichar = NULL;
|
|
for (c_it.mark_cycle_pt(); !c_it.cycled_list(); c_it.forward()) {
|
|
UNICHAR_ID unichar_id = c_it.data()->unichar_id();
|
|
if (unicharset.get_fragment(unichar_id)) continue; // skip fragments
|
|
if (first_unichar == NULL) first_unichar = c_it.data();
|
|
if (*first_lower == NULL && unicharset.get_islower(unichar_id)) {
|
|
*first_lower = c_it.data();
|
|
}
|
|
if (*first_upper == NULL && unicharset.get_isupper(unichar_id)) {
|
|
*first_upper = c_it.data();
|
|
}
|
|
}
|
|
ASSERT_HOST(first_unichar != NULL);
|
|
if (*first_lower == NULL) *first_lower = first_unichar;
|
|
if (*first_upper == NULL) *first_upper = first_unichar;
|
|
}
|
|
|
|
LanguageModelFlagsType LanguageModel::AddViterbiStateEntry(
|
|
LanguageModelFlagsType top_choice_flags,
|
|
float denom,
|
|
bool word_end,
|
|
int curr_col, int curr_row,
|
|
BLOB_CHOICE *b,
|
|
BLOB_CHOICE *parent_b,
|
|
ViterbiStateEntry *parent_vse,
|
|
HEAP *pain_points,
|
|
BestPathByColumn *best_path_by_column[],
|
|
CHUNKS_RECORD *chunks_record,
|
|
BestChoiceBundle *best_choice_bundle,
|
|
BlamerBundle *blamer_bundle) {
|
|
ViterbiStateEntry_IT vit;
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("\nAddViterbiStateEntry for unichar %s rating=%.4f"
|
|
" certainty=%.4f top_choice_flags=0x%x parent_vse=%p\n",
|
|
dict_->getUnicharset().id_to_unichar(b->unichar_id()),
|
|
b->rating(), b->certainty(), top_choice_flags, parent_vse);
|
|
if (language_model_debug_level > 3 && b->language_model_state() != NULL) {
|
|
tprintf("Existing viterbi list:\n");
|
|
vit.set_to_list(&(reinterpret_cast<LanguageModelState *>(
|
|
b->language_model_state())->viterbi_state_entries));
|
|
for (vit.mark_cycle_pt(); !vit.cycled_list(); vit.forward()) {
|
|
PrintViterbiStateEntry("", vit.data(), b, chunks_record);
|
|
}
|
|
}
|
|
}
|
|
// Check whether the list is full.
|
|
if (b->language_model_state() != NULL &&
|
|
(reinterpret_cast<LanguageModelState *>(
|
|
b->language_model_state())->viterbi_state_entries_length >=
|
|
language_model_viterbi_list_max_size)) {
|
|
if (language_model_debug_level > 1) {
|
|
tprintf("AddViterbiStateEntry: viterbi list is full!\n");
|
|
}
|
|
return 0x0;
|
|
}
|
|
|
|
LanguageModelFlagsType changed = 0x0;
|
|
float optimistic_cost = 0.0f;
|
|
if (!language_model_ngram_on) optimistic_cost += b->rating();
|
|
if (parent_vse != NULL) optimistic_cost += parent_vse->cost;
|
|
// Discard this entry if it will not beat best choice.
|
|
if (optimistic_cost >= best_choice_bundle->best_choice->rating()) {
|
|
if (language_model_debug_level > 1) {
|
|
tprintf("Discarded ViterbiEntry with high cost %.4f"
|
|
" best_choice->rating()=%.4f\n", optimistic_cost,
|
|
best_choice_bundle->best_choice->rating());
|
|
}
|
|
return 0x0;
|
|
}
|
|
|
|
// Check consistency of the path and set the relevant consistency_info.
|
|
LanguageModelConsistencyInfo consistency_info;
|
|
FillConsistencyInfo(curr_col, word_end, b, parent_vse, parent_b,
|
|
chunks_record, &consistency_info);
|
|
|
|
// Invoke Dawg language model component.
|
|
LanguageModelDawgInfo *dawg_info =
|
|
GenerateDawgInfo(word_end, consistency_info.script_id,
|
|
curr_col, curr_row, *b, parent_vse, &changed);
|
|
|
|
// Invoke TopChoice language model component
|
|
float ratings_sum = b->rating();
|
|
if (parent_vse != NULL) ratings_sum += parent_vse->ratings_sum;
|
|
GenerateTopChoiceInfo(ratings_sum, dawg_info, consistency_info,
|
|
parent_vse, b, &top_choice_flags, &changed);
|
|
|
|
// Invoke Ngram language model component.
|
|
LanguageModelNgramInfo *ngram_info = NULL;
|
|
if (language_model_ngram_on) {
|
|
ngram_info = GenerateNgramInfo(
|
|
dict_->getUnicharset().id_to_unichar(b->unichar_id()), b->certainty(),
|
|
denom, curr_col, curr_row, parent_vse, parent_b, &changed);
|
|
ASSERT_HOST(ngram_info != NULL);
|
|
}
|
|
|
|
// Prune non-top choice paths with inconsistent scripts.
|
|
if (consistency_info.inconsistent_script) {
|
|
if (!(top_choice_flags & kSmallestRatingFlag)) changed = 0x0;
|
|
if (ngram_info != NULL) ngram_info->pruned = true;
|
|
}
|
|
|
|
// If language model components did not like this unichar - return
|
|
if (!changed) {
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("Language model components did not like this entry\n");
|
|
}
|
|
delete dawg_info;
|
|
delete ngram_info;
|
|
return 0x0;
|
|
}
|
|
|
|
// Compute cost of associating the blobs that represent the current unichar.
|
|
AssociateStats associate_stats;
|
|
ComputeAssociateStats(curr_col, curr_row, max_char_wh_ratio_,
|
|
parent_vse, chunks_record, &associate_stats);
|
|
if (parent_vse != NULL) {
|
|
associate_stats.shape_cost += parent_vse->associate_stats.shape_cost;
|
|
associate_stats.bad_shape |= parent_vse->associate_stats.bad_shape;
|
|
}
|
|
|
|
// Compute the aggregate cost (adjusted ratings sum).
|
|
float cost = ComputeAdjustedPathCost(
|
|
ratings_sum,
|
|
(parent_vse == NULL) ? 1 : (parent_vse->length+1),
|
|
(dawg_info == NULL) ? 0.0f : 1.0f,
|
|
dawg_info, ngram_info, consistency_info, associate_stats, parent_vse);
|
|
|
|
if (b->language_model_state() == NULL) {
|
|
b->set_language_model_state(new LanguageModelState(curr_col, curr_row));
|
|
}
|
|
LanguageModelState *lms =
|
|
reinterpret_cast<LanguageModelState *>(b->language_model_state());
|
|
|
|
// Discard this entry if it represents a prunable path and
|
|
// language_model_viterbi_list_max_num_prunable such entries with a lower
|
|
// cost have already been recorded.
|
|
if (PrunablePath(top_choice_flags, dawg_info) &&
|
|
(lms->viterbi_state_entries_prunable_length >=
|
|
language_model_viterbi_list_max_num_prunable) &&
|
|
cost >= lms->viterbi_state_entries_prunable_max_cost) {
|
|
if (language_model_debug_level > 1) {
|
|
tprintf("Discarded ViterbiEntry with high cost %g max cost %g\n",
|
|
cost, lms->viterbi_state_entries_prunable_max_cost);
|
|
}
|
|
delete dawg_info;
|
|
delete ngram_info;
|
|
return 0x0;
|
|
}
|
|
|
|
// Create the new ViterbiStateEntry and add it to lms->viterbi_state_entries
|
|
ViterbiStateEntry *new_vse = new ViterbiStateEntry(
|
|
parent_b, parent_vse, b, cost, ComputeOutlineLength(b), consistency_info,
|
|
associate_stats, top_choice_flags, dawg_info, ngram_info);
|
|
updated_flags_.push_back(&(new_vse->updated));
|
|
lms->viterbi_state_entries.add_sorted(ViterbiStateEntry::Compare,
|
|
false, new_vse);
|
|
lms->viterbi_state_entries_length++;
|
|
if (PrunablePath(top_choice_flags, dawg_info)) {
|
|
lms->viterbi_state_entries_prunable_length++;
|
|
}
|
|
|
|
// Update lms->viterbi_state_entries_prunable_max_cost and clear
|
|
// top_choice_flags of entries with ratings_sum than new_vse->ratings_sum.
|
|
if ((lms->viterbi_state_entries_prunable_length >=
|
|
language_model_viterbi_list_max_num_prunable) || top_choice_flags) {
|
|
ASSERT_HOST(!lms->viterbi_state_entries.empty());
|
|
int prunable_counter = language_model_viterbi_list_max_num_prunable;
|
|
vit.set_to_list(&(lms->viterbi_state_entries));
|
|
for (vit.mark_cycle_pt(); !vit.cycled_list(); vit.forward()) {
|
|
ViterbiStateEntry *curr_vse = vit.data();
|
|
// Clear the appropriate top choice flags of the entries in the
|
|
// list that have ratings_sum higher thank new_entry->ratings_sum
|
|
// (since they will not be top choices any more).
|
|
if (curr_vse->top_choice_flags && curr_vse != new_vse &&
|
|
ComputeConsistencyAdjustedRatingsSum(
|
|
curr_vse->ratings_sum, curr_vse->dawg_info,
|
|
curr_vse->consistency_info) >
|
|
ComputeConsistencyAdjustedRatingsSum(
|
|
new_vse->ratings_sum, new_vse->dawg_info,
|
|
new_vse->consistency_info)) {
|
|
curr_vse->top_choice_flags &= ~(top_choice_flags);
|
|
}
|
|
if (prunable_counter > 0 &&
|
|
PrunablePath(curr_vse->top_choice_flags, curr_vse->dawg_info)) {
|
|
--prunable_counter;
|
|
}
|
|
// Update lms->viterbi_state_entries_prunable_max_cost.
|
|
if (prunable_counter == 0) {
|
|
lms->viterbi_state_entries_prunable_max_cost = vit.data()->cost;
|
|
if (language_model_debug_level > 1) {
|
|
tprintf("Set viterbi_state_entries_prunable_max_cost to %.4f\n",
|
|
lms->viterbi_state_entries_prunable_max_cost);
|
|
}
|
|
prunable_counter = -1; // stop counting
|
|
}
|
|
}
|
|
}
|
|
|
|
// Print the newly created ViterbiStateEntry.
|
|
if (language_model_debug_level > 2) {
|
|
PrintViterbiStateEntry("New", new_vse, b, chunks_record);
|
|
if (language_model_debug_level > 3) {
|
|
tprintf("Updated viterbi list (max cost %g):\n",
|
|
lms->viterbi_state_entries_prunable_max_cost);
|
|
vit.set_to_list(&(lms->viterbi_state_entries));
|
|
for (vit.mark_cycle_pt(); !vit.cycled_list(); vit.forward()) {
|
|
PrintViterbiStateEntry("", vit.data(), b, chunks_record);
|
|
}
|
|
}
|
|
}
|
|
|
|
// Update best choice if needed.
|
|
if (word_end) {
|
|
UpdateBestChoice(b, new_vse, pain_points, chunks_record,
|
|
best_choice_bundle, blamer_bundle);
|
|
}
|
|
|
|
// Update stats in best_path_by_column.
|
|
if (new_vse->Consistent() || new_vse->dawg_info != NULL ||
|
|
(new_vse->ngram_info != NULL && !new_vse->ngram_info->pruned)) {
|
|
float avg_cost = new_vse->cost / static_cast<float>(curr_row+1);
|
|
for (int c = curr_col; c <= curr_row; ++c) {
|
|
if (avg_cost < (*best_path_by_column)[c].avg_cost) {
|
|
(*best_path_by_column)[c].avg_cost = avg_cost;
|
|
(*best_path_by_column)[c].best_vse = new_vse;
|
|
(*best_path_by_column)[c].best_b = b;
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("Set best_path_by_column[%d]=(%g %p)\n",
|
|
c, avg_cost, new_vse);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
return changed;
|
|
}
|
|
|
|
void LanguageModel::PrintViterbiStateEntry(
|
|
const char *msg, ViterbiStateEntry *vse,
|
|
BLOB_CHOICE *b, CHUNKS_RECORD *chunks_record) {
|
|
tprintf("%s ViterbiStateEntry %p with ratings_sum=%.4f length=%d cost=%.4f",
|
|
msg, vse, vse->ratings_sum, vse->length, vse->cost);
|
|
if (vse->top_choice_flags) {
|
|
tprintf(" top_choice_flags=0x%x", vse->top_choice_flags);
|
|
}
|
|
if (!vse->Consistent()) {
|
|
tprintf(" inconsistent=(punc %d case %d chartype %d script %d)\n",
|
|
vse->consistency_info.NumInconsistentPunc(),
|
|
vse->consistency_info.NumInconsistentCase(),
|
|
vse->consistency_info.NumInconsistentChartype(),
|
|
vse->consistency_info.inconsistent_script);
|
|
}
|
|
if (vse->dawg_info) tprintf(" permuter=%d", vse->dawg_info->permuter);
|
|
if (vse->ngram_info) {
|
|
tprintf(" ngram_cost=%g context=%s ngram pruned=%d",
|
|
vse->ngram_info->ngram_cost,
|
|
vse->ngram_info->context.string(),
|
|
vse->ngram_info->pruned);
|
|
}
|
|
if (vse->associate_stats.shape_cost > 0.0f) {
|
|
tprintf(" shape_cost=%g", vse->associate_stats.shape_cost);
|
|
}
|
|
if (language_model_debug_level > 3) {
|
|
STRING wd_str;
|
|
WERD_CHOICE *wd = ConstructWord(b, vse, chunks_record,
|
|
NULL, NULL, NULL, NULL, NULL, NULL);
|
|
wd->string_and_lengths(&wd_str, NULL);
|
|
delete wd;
|
|
tprintf(" str=%s", wd_str.string());
|
|
}
|
|
tprintf("\n");
|
|
}
|
|
|
|
void LanguageModel::GenerateTopChoiceInfo(
|
|
float ratings_sum,
|
|
const LanguageModelDawgInfo *dawg_info,
|
|
const LanguageModelConsistencyInfo &consistency_info,
|
|
const ViterbiStateEntry *parent_vse,
|
|
BLOB_CHOICE *b,
|
|
LanguageModelFlagsType *top_choice_flags,
|
|
LanguageModelFlagsType *changed) {
|
|
ratings_sum = ComputeConsistencyAdjustedRatingsSum(
|
|
ratings_sum, dawg_info, consistency_info);
|
|
// Clear flags that do not agree with parent_vse->top_choice_flags.
|
|
if (parent_vse != NULL) *top_choice_flags &= parent_vse->top_choice_flags;
|
|
if (consistency_info.Consistent()) *top_choice_flags |= kConsistentFlag;
|
|
if (*top_choice_flags == 0x0) return;
|
|
LanguageModelState *lms =
|
|
reinterpret_cast<LanguageModelState *>(b->language_model_state());
|
|
if (lms != NULL && !lms->viterbi_state_entries.empty()) {
|
|
ViterbiStateEntry_IT vit(&(lms->viterbi_state_entries));
|
|
for (vit.mark_cycle_pt(); !vit.cycled_list(); vit.forward()) {
|
|
if (ratings_sum >= ComputeConsistencyAdjustedRatingsSum(
|
|
vit.data()->ratings_sum, vit.data()->dawg_info,
|
|
vit.data()->consistency_info)) {
|
|
// Clear the appropriate flags if the list already contains
|
|
// a top choice entry with a lower cost.
|
|
*top_choice_flags &= ~(vit.data()->top_choice_flags);
|
|
}
|
|
}
|
|
}
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("GenerateTopChoiceInfo: top_choice_flags=0x%x\n",
|
|
*top_choice_flags);
|
|
}
|
|
*changed |= *top_choice_flags;
|
|
}
|
|
|
|
LanguageModelDawgInfo *LanguageModel::GenerateDawgInfo(
|
|
bool word_end, int script_id,
|
|
int curr_col, int curr_row,
|
|
const BLOB_CHOICE &b,
|
|
const ViterbiStateEntry *parent_vse,
|
|
LanguageModelFlagsType *changed) {
|
|
bool use_fixed_length_dawgs = !fixed_length_beginning_active_dawgs_->empty();
|
|
|
|
// Initialize active_dawgs and constraints from parent_vse if it is not NULL,
|
|
// otherwise use beginning_active_dawgs_ and beginning_constraints_.
|
|
if (parent_vse == NULL ||
|
|
(use_fixed_length_dawgs && parent_vse->dawg_info == NULL)) {
|
|
dawg_args_->active_dawgs = beginning_active_dawgs_;
|
|
dawg_args_->constraints = beginning_constraints_;
|
|
dawg_args_->permuter = NO_PERM;
|
|
} else {
|
|
if (parent_vse->dawg_info == NULL) return NULL; // not a dict word path
|
|
dawg_args_->active_dawgs = parent_vse->dawg_info->active_dawgs;
|
|
dawg_args_->constraints = parent_vse->dawg_info->constraints;
|
|
dawg_args_->permuter = parent_vse->dawg_info->permuter;
|
|
}
|
|
|
|
// Deal with hyphenated words.
|
|
if (!use_fixed_length_dawgs && word_end &&
|
|
dict_->has_hyphen_end(b.unichar_id(), curr_col == 0)) {
|
|
if (language_model_debug_level > 0) tprintf("Hyphenated word found\n");
|
|
*changed |= kDawgFlag;
|
|
return new LanguageModelDawgInfo(dawg_args_->active_dawgs,
|
|
dawg_args_->constraints,
|
|
COMPOUND_PERM);
|
|
}
|
|
|
|
// Deal with compound words.
|
|
if (!use_fixed_length_dawgs && dict_->compound_marker(b.unichar_id()) &&
|
|
(parent_vse == NULL || parent_vse->dawg_info->permuter != NUMBER_PERM)) {
|
|
if (language_model_debug_level > 0) tprintf("Found compound marker");
|
|
// Do not allow compound operators at the beginning and end of the word.
|
|
// Do not allow more than one compound operator per word.
|
|
// Do not allow compounding of words with lengths shorter than
|
|
// language_model_min_compound_length
|
|
if (parent_vse == NULL || word_end ||
|
|
dawg_args_->permuter == COMPOUND_PERM ||
|
|
parent_vse->length < language_model_min_compound_length) return NULL;
|
|
|
|
int i;
|
|
// Check a that the path terminated before the current character is a word.
|
|
bool has_word_ending = false;
|
|
for (i = 0; i < parent_vse->dawg_info->active_dawgs->size(); ++i) {
|
|
const DawgInfo &info = (*parent_vse->dawg_info->active_dawgs)[i];
|
|
const Dawg *pdawg = dict_->GetDawg(info.dawg_index);
|
|
assert(pdawg != NULL);
|
|
if (pdawg->type() == DAWG_TYPE_WORD && info.ref != NO_EDGE &&
|
|
pdawg->end_of_word(info.ref)) {
|
|
has_word_ending = true;
|
|
break;
|
|
}
|
|
}
|
|
if (!has_word_ending) return NULL;
|
|
|
|
// Return LanguageModelDawgInfo with active_dawgs set to
|
|
// the beginning dawgs of type DAWG_TYPE_WORD.
|
|
if (language_model_debug_level > 0) tprintf("Compound word found\n");
|
|
DawgInfoVector beginning_word_dawgs;
|
|
for (i = 0; i < beginning_active_dawgs_->size(); ++i) {
|
|
const Dawg *bdawg =
|
|
dict_->GetDawg((*beginning_active_dawgs_)[i].dawg_index);
|
|
if (bdawg->type() == DAWG_TYPE_WORD) {
|
|
beginning_word_dawgs += (*beginning_active_dawgs_)[i];
|
|
}
|
|
}
|
|
*changed |= kDawgFlag;
|
|
return new LanguageModelDawgInfo(&(beginning_word_dawgs),
|
|
dawg_args_->constraints,
|
|
COMPOUND_PERM);
|
|
} // done dealing with compound words
|
|
|
|
LanguageModelDawgInfo *dawg_info = NULL;
|
|
|
|
// Call LetterIsOkay().
|
|
dict_->LetterIsOkay(dawg_args_, b.unichar_id(), word_end);
|
|
if (dawg_args_->permuter != NO_PERM) {
|
|
*changed |= kDawgFlag;
|
|
dawg_info = new LanguageModelDawgInfo(dawg_args_->updated_active_dawgs,
|
|
dawg_args_->updated_constraints,
|
|
dawg_args_->permuter);
|
|
}
|
|
|
|
// For non-space delimited languages: since every letter could be
|
|
// a valid word, a new word could start at every unichar. Thus append
|
|
// fixed_length_beginning_active_dawgs_ to dawg_info->active_dawgs.
|
|
if (use_fixed_length_dawgs) {
|
|
if (dawg_info == NULL) {
|
|
*changed |= kDawgFlag;
|
|
dawg_info = new LanguageModelDawgInfo(
|
|
fixed_length_beginning_active_dawgs_,
|
|
empty_dawg_info_vec_, SYSTEM_DAWG_PERM);
|
|
} else {
|
|
*(dawg_info->active_dawgs) += *(fixed_length_beginning_active_dawgs_);
|
|
}
|
|
} // done dealing with fixed-length dawgs
|
|
|
|
return dawg_info;
|
|
}
|
|
|
|
LanguageModelNgramInfo *LanguageModel::GenerateNgramInfo(
|
|
const char *unichar, float certainty, float denom,
|
|
int curr_col, int curr_row,
|
|
const ViterbiStateEntry *parent_vse,
|
|
BLOB_CHOICE *parent_b,
|
|
LanguageModelFlagsType *changed) {
|
|
// Initialize parent context.
|
|
const char *pcontext_ptr = "";
|
|
int pcontext_unichar_step_len = 0;
|
|
if (parent_vse == NULL) {
|
|
pcontext_ptr = prev_word_str_.string();
|
|
pcontext_unichar_step_len = prev_word_unichar_step_len_;
|
|
} else {
|
|
pcontext_ptr = parent_vse->ngram_info->context.string();
|
|
pcontext_unichar_step_len =
|
|
parent_vse->ngram_info->context_unichar_step_len;
|
|
}
|
|
// Compute p(unichar | parent context).
|
|
int unichar_step_len = 0;
|
|
bool pruned = false;
|
|
float ngram_prob;
|
|
float ngram_cost = ComputeNgramCost(unichar, certainty, denom,
|
|
pcontext_ptr, &unichar_step_len,
|
|
&pruned, &ngram_prob);
|
|
// First attempt to normalize ngram_cost for strings of different
|
|
// lengths - we multiply ngram_cost by P(char | context) as many times
|
|
// as the number of chunks occupied by char. This makes the ngram costs
|
|
// of all the paths ending at the current BLOB_CHOICE comparable.
|
|
// TODO(daria): it would be worth looking at different ways of normalization.
|
|
if (curr_row > curr_col) {
|
|
ngram_cost += (curr_row - curr_col) * ngram_cost;
|
|
ngram_prob += (curr_row - curr_col) * ngram_prob;
|
|
}
|
|
// Add the ngram_cost of the parent.
|
|
if (parent_vse != NULL) {
|
|
ngram_cost += parent_vse->ngram_info->ngram_cost;
|
|
ngram_prob += parent_vse->ngram_info->ngram_prob;
|
|
}
|
|
|
|
// Shorten parent context string by unichar_step_len unichars.
|
|
int num_remove = (unichar_step_len + pcontext_unichar_step_len -
|
|
language_model_ngram_order);
|
|
if (num_remove > 0) pcontext_unichar_step_len -= num_remove;
|
|
while (num_remove > 0 && *pcontext_ptr != '\0') {
|
|
pcontext_ptr += UNICHAR::utf8_step(pcontext_ptr);
|
|
--num_remove;
|
|
}
|
|
|
|
// Decide whether to prune this ngram path and update changed accordingly.
|
|
if (parent_vse != NULL && parent_vse->ngram_info->pruned) pruned = true;
|
|
if (!pruned) *changed |= kNgramFlag;
|
|
|
|
// Construct and return the new LanguageModelNgramInfo.
|
|
LanguageModelNgramInfo *ngram_info = new LanguageModelNgramInfo(
|
|
pcontext_ptr, pcontext_unichar_step_len, pruned, ngram_prob, ngram_cost);
|
|
ngram_info->context += unichar;
|
|
ngram_info->context_unichar_step_len += unichar_step_len;
|
|
assert(ngram_info->context_unichar_step_len <= language_model_ngram_order);
|
|
return ngram_info;
|
|
}
|
|
|
|
float LanguageModel::ComputeNgramCost(const char *unichar,
|
|
float certainty,
|
|
float denom,
|
|
const char *context,
|
|
int *unichar_step_len,
|
|
bool *found_small_prob,
|
|
float *ngram_prob) {
|
|
const char *context_ptr = context;
|
|
char *modified_context = NULL;
|
|
char *modified_context_end = NULL;
|
|
const char *unichar_ptr = unichar;
|
|
const char *unichar_end = unichar_ptr + strlen(unichar_ptr);
|
|
float prob = 0.0f;
|
|
int step = 0;
|
|
while (unichar_ptr < unichar_end &&
|
|
(step = UNICHAR::utf8_step(unichar_ptr)) > 0) {
|
|
if (language_model_debug_level > 1) {
|
|
tprintf("prob(%s | %s)=%g\n", unichar_ptr, context_ptr,
|
|
dict_->ProbabilityInContext(context_ptr, -1, unichar_ptr, step));
|
|
}
|
|
prob += dict_->ProbabilityInContext(context_ptr, -1, unichar_ptr, step);
|
|
++(*unichar_step_len);
|
|
if (language_model_ngram_use_only_first_uft8_step) break;
|
|
unichar_ptr += step;
|
|
// If there are multiple UTF8 characters present in unichar, context is
|
|
// updated to include the previously examined characters from str,
|
|
// unless use_only_first_uft8_step is true.
|
|
if (unichar_ptr < unichar_end) {
|
|
if (modified_context == NULL) {
|
|
int context_len = strlen(context);
|
|
modified_context =
|
|
new char[context_len + strlen(unichar_ptr) + step + 1];
|
|
strncpy(modified_context, context, context_len);
|
|
modified_context_end = modified_context + context_len;
|
|
context_ptr = modified_context;
|
|
}
|
|
strncpy(modified_context_end, unichar_ptr - step, step);
|
|
modified_context_end += step;
|
|
*modified_context_end = '\0';
|
|
}
|
|
}
|
|
prob /= static_cast<float>(*unichar_step_len); // normalize
|
|
if (prob < language_model_ngram_small_prob) {
|
|
if (language_model_debug_level > 0) tprintf("Found small prob %g\n", prob);
|
|
*found_small_prob = true;
|
|
}
|
|
*ngram_prob = -1.0*log(prob);
|
|
float cost = -1.0*log(CertaintyScore(certainty)/denom) +
|
|
*ngram_prob * language_model_ngram_scale_factor;
|
|
if (language_model_debug_level > 1) {
|
|
tprintf("-log [ p(%s) * p(%s | %s) ] = -log(%g*%g) = %g\n", unichar,
|
|
unichar, context_ptr, CertaintyScore(certainty)/denom, prob, cost);
|
|
}
|
|
if (modified_context != NULL) delete[] modified_context;
|
|
return cost;
|
|
}
|
|
|
|
float LanguageModel::ComputeDenom(BLOB_CHOICE_LIST *curr_list) {
|
|
if (curr_list->empty()) return 1.0f;
|
|
float denom = 0.0f;
|
|
int len = 0;
|
|
BLOB_CHOICE_IT c_it(curr_list);
|
|
for (c_it.mark_cycle_pt(); !c_it.cycled_list(); c_it.forward()) {
|
|
ASSERT_HOST(c_it.data() != NULL);
|
|
++len;
|
|
denom += CertaintyScore(c_it.data()->certainty());
|
|
}
|
|
assert(len != 0);
|
|
// The ideal situation would be to have the classifier scores for
|
|
// classifying each position as each of the characters in the unicharset.
|
|
// Since we can not do this because of speed, we add a very crude estimate
|
|
// of what these scores for the "missing" classifications would sum up to.
|
|
denom += (dict_->getUnicharset().size() - len) *
|
|
CertaintyScore(language_model_ngram_nonmatch_score);
|
|
|
|
return denom;
|
|
}
|
|
|
|
void LanguageModel::FillConsistencyInfo(
|
|
int curr_col,
|
|
bool word_end,
|
|
BLOB_CHOICE *b,
|
|
ViterbiStateEntry *parent_vse,
|
|
BLOB_CHOICE *parent_b,
|
|
CHUNKS_RECORD *chunks_record,
|
|
LanguageModelConsistencyInfo *consistency_info) {
|
|
const UNICHARSET &unicharset = dict_->getUnicharset();
|
|
UNICHAR_ID unichar_id = b->unichar_id();
|
|
if (parent_vse != NULL) *consistency_info = parent_vse->consistency_info;
|
|
|
|
// Check punctuation validity.
|
|
if (unicharset.get_ispunctuation(unichar_id)) consistency_info->num_punc++;
|
|
if (dict_->GetPuncDawg() != NULL && !consistency_info->invalid_punc) {
|
|
if (dict_->compound_marker(unichar_id) && parent_b != NULL &&
|
|
(unicharset.get_isalpha(parent_b->unichar_id()) ||
|
|
unicharset.get_isdigit(parent_b->unichar_id()))) {
|
|
// reset punc_ref for compound words
|
|
consistency_info->punc_ref = NO_EDGE;
|
|
} else {
|
|
UNICHAR_ID pattern_unichar_id =
|
|
(unicharset.get_isalpha(unichar_id) ||
|
|
unicharset.get_isdigit(unichar_id)) ?
|
|
Dawg::kPatternUnicharID : unichar_id;
|
|
if (consistency_info->punc_ref == NO_EDGE ||
|
|
pattern_unichar_id != Dawg::kPatternUnicharID ||
|
|
dict_->GetPuncDawg()->edge_letter(consistency_info->punc_ref) !=
|
|
Dawg::kPatternUnicharID) {
|
|
NODE_REF node = Dict::GetStartingNode(dict_->GetPuncDawg(),
|
|
consistency_info->punc_ref);
|
|
consistency_info->punc_ref =
|
|
(node != NO_EDGE) ? dict_->GetPuncDawg()->edge_char_of(
|
|
node, pattern_unichar_id, word_end) : NO_EDGE;
|
|
if (consistency_info->punc_ref == NO_EDGE) {
|
|
consistency_info->invalid_punc = true;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Update case related counters.
|
|
if (parent_vse != NULL && !word_end && dict_->compound_marker(unichar_id)) {
|
|
// Reset counters if we are dealing with a compound word.
|
|
consistency_info->num_lower = 0;
|
|
consistency_info->num_non_first_upper = 0;
|
|
}
|
|
else if (unicharset.get_islower(unichar_id)) {
|
|
consistency_info->num_lower++;
|
|
} else if ((parent_b != NULL) && unicharset.get_isupper(unichar_id)) {
|
|
if (unicharset.get_isupper(parent_b->unichar_id()) ||
|
|
consistency_info->num_lower > 0 ||
|
|
consistency_info->num_non_first_upper > 0) {
|
|
consistency_info->num_non_first_upper++;
|
|
}
|
|
}
|
|
|
|
// Initialize consistency_info->script_id (use script of unichar_id
|
|
// if it is not Common, use script id recorded by the parent otherwise).
|
|
// Set inconsistent_script to true if the script of the current unichar
|
|
// is not consistent with that of the parent.
|
|
consistency_info->script_id = unicharset.get_script(unichar_id);
|
|
// Hiragana and Katakana can mix with Han.
|
|
if (dict_->getUnicharset().han_sid() != dict_->getUnicharset().null_sid()) {
|
|
if ((unicharset.hiragana_sid() != unicharset.null_sid() &&
|
|
consistency_info->script_id == unicharset.hiragana_sid()) ||
|
|
(unicharset.katakana_sid() != unicharset.null_sid() &&
|
|
consistency_info->script_id == unicharset.katakana_sid())) {
|
|
consistency_info->script_id = dict_->getUnicharset().han_sid();
|
|
}
|
|
}
|
|
|
|
if (parent_vse != NULL &&
|
|
(parent_vse->consistency_info.script_id !=
|
|
dict_->getUnicharset().common_sid())) {
|
|
int parent_script_id = parent_vse->consistency_info.script_id;
|
|
// If script_id is Common, use script id of the parent instead.
|
|
if (consistency_info->script_id == dict_->getUnicharset().common_sid()) {
|
|
consistency_info->script_id = parent_script_id;
|
|
}
|
|
if (consistency_info->script_id != parent_script_id) {
|
|
consistency_info->inconsistent_script = true;
|
|
}
|
|
}
|
|
|
|
// Update chartype related counters.
|
|
if (unicharset.get_isalpha(unichar_id)) {
|
|
consistency_info->num_alphas++;
|
|
} else if (unicharset.get_isdigit(unichar_id)) {
|
|
consistency_info->num_digits++;
|
|
} else if (!unicharset.get_ispunctuation(unichar_id)) {
|
|
consistency_info->num_other++;
|
|
}
|
|
|
|
// Check font and spacing consistency.
|
|
if (parent_b != NULL) {
|
|
int fontinfo_id = -1;
|
|
if (parent_b->fontinfo_id() == b->fontinfo_id() ||
|
|
parent_b->fontinfo_id2() == b->fontinfo_id()) {
|
|
fontinfo_id = b->fontinfo_id();
|
|
} else if (parent_b->fontinfo_id() == b->fontinfo_id2() ||
|
|
parent_b->fontinfo_id2() == b->fontinfo_id2()) {
|
|
fontinfo_id = b->fontinfo_id2();
|
|
}
|
|
if(language_model_debug_level > 1) {
|
|
tprintf("pfont %s pfont %s font %s font2 %s common %s(%d)\n",
|
|
(parent_b->fontinfo_id() >= 0) ?
|
|
fontinfo_table_->get(parent_b->fontinfo_id()).name : "" ,
|
|
(parent_b->fontinfo_id2() >= 0) ?
|
|
fontinfo_table_->get(parent_b->fontinfo_id2()).name : "",
|
|
(b->fontinfo_id() >= 0) ?
|
|
fontinfo_table_->get(b->fontinfo_id()).name : "",
|
|
(fontinfo_id >= 0) ? fontinfo_table_->get(fontinfo_id).name : "",
|
|
(fontinfo_id >= 0) ? fontinfo_table_->get(fontinfo_id).name : "",
|
|
fontinfo_id);
|
|
}
|
|
bool expected_gap_found = false;
|
|
float expected_gap;
|
|
int temp_gap;
|
|
if (fontinfo_id >= 0) { // found a common font
|
|
if (fontinfo_table_->get(fontinfo_id).get_spacing(
|
|
parent_b->unichar_id(), unichar_id, &temp_gap)) {
|
|
expected_gap = temp_gap;
|
|
expected_gap_found = true;
|
|
}
|
|
} else {
|
|
consistency_info->inconsistent_font = true;
|
|
// Get an average of the expected gaps in each font
|
|
int num_addends = 0;
|
|
expected_gap = 0;
|
|
int temp_fid;
|
|
for (int i = 0; i < 4; ++i) {
|
|
if (i == 0) {
|
|
temp_fid = parent_b->fontinfo_id();
|
|
} else if (i == 1) {
|
|
temp_fid = parent_b->fontinfo_id2();
|
|
} else if (i == 2) {
|
|
temp_fid = b->fontinfo_id();
|
|
} else {
|
|
temp_fid = b->fontinfo_id2();
|
|
}
|
|
if (temp_fid >= 0 && fontinfo_table_->get(temp_fid).get_spacing(
|
|
parent_b->unichar_id(), unichar_id, &temp_gap)) {
|
|
expected_gap += temp_gap;
|
|
num_addends++;
|
|
}
|
|
}
|
|
expected_gap_found = (num_addends > 0);
|
|
if (num_addends > 0) expected_gap /= static_cast<float>(num_addends);
|
|
}
|
|
if (expected_gap_found) {
|
|
float actual_gap =
|
|
static_cast<float>(AssociateUtils::GetChunksGap(
|
|
chunks_record->chunk_widths, curr_col-1));
|
|
float gap_ratio = expected_gap / actual_gap;
|
|
// TODO(daria): find a good way to tune this heuristic estimate.
|
|
if (gap_ratio < 1/2 || gap_ratio > 2) {
|
|
consistency_info->num_inconsistent_spaces++;
|
|
}
|
|
if(language_model_debug_level > 1) {
|
|
tprintf("spacing for %s(%d) %s(%d) col %d: expected %g actual %g\n",
|
|
unicharset.id_to_unichar(parent_b->unichar_id()),
|
|
parent_b->unichar_id(), unicharset.id_to_unichar(unichar_id),
|
|
unichar_id, curr_col, expected_gap, actual_gap);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
float LanguageModel::ComputeAdjustedPathCost(
|
|
float ratings_sum, int length, float dawg_score,
|
|
const LanguageModelDawgInfo *dawg_info,
|
|
const LanguageModelNgramInfo *ngram_info,
|
|
const LanguageModelConsistencyInfo &consistency_info,
|
|
const AssociateStats &associate_stats,
|
|
ViterbiStateEntry *parent_vse) {
|
|
float adjustment = 1.0f;
|
|
if (dawg_info == NULL || dawg_info->permuter != FREQ_DAWG_PERM) {
|
|
adjustment += language_model_penalty_non_freq_dict_word;
|
|
}
|
|
if (dawg_score == 0.0f) {
|
|
adjustment += language_model_penalty_non_dict_word;
|
|
if (length > language_model_min_compound_length) {
|
|
adjustment += ((length - language_model_min_compound_length) *
|
|
language_model_penalty_increment);
|
|
}
|
|
} else if (dawg_score < 1.0f) {
|
|
adjustment += (1.0f - dawg_score) * language_model_penalty_non_dict_word;
|
|
}
|
|
if (associate_stats.shape_cost > 0) {
|
|
adjustment += associate_stats.shape_cost / static_cast<float>(length);
|
|
}
|
|
if (language_model_ngram_on) {
|
|
ASSERT_HOST(ngram_info != NULL);
|
|
return ngram_info->ngram_cost * adjustment;
|
|
} else {
|
|
adjustment += ComputeConsistencyAdjustment(dawg_info, consistency_info);
|
|
return ratings_sum * adjustment;
|
|
}
|
|
}
|
|
|
|
void LanguageModel::UpdateBestChoice(
|
|
BLOB_CHOICE *b,
|
|
ViterbiStateEntry *vse,
|
|
HEAP *pain_points,
|
|
CHUNKS_RECORD *chunks_record,
|
|
BestChoiceBundle *best_choice_bundle,
|
|
BlamerBundle *blamer_bundle) {
|
|
int i;
|
|
BLOB_CHOICE_LIST_VECTOR temp_best_char_choices(vse->length);
|
|
for (i = 0; i < vse->length; ++i) {
|
|
temp_best_char_choices.push_back(NULL);
|
|
}
|
|
float *certainties = new float[vse->length];
|
|
STATE temp_state;
|
|
// The fraction of letters in the path that are "covered" by dawgs.
|
|
// For space delimited languages this number will be 0.0 for nonwords
|
|
// and 1.0 for dictionary words. For non-space delimited languages
|
|
// this number will be in the [0.0, 1.0] range.
|
|
float dawg_score;
|
|
bool truth_path;
|
|
WERD_CHOICE *word = ConstructWord(b, vse, chunks_record,
|
|
&temp_best_char_choices, certainties,
|
|
&dawg_score, &temp_state,
|
|
blamer_bundle, &truth_path);
|
|
ASSERT_HOST(word != NULL);
|
|
bool not_blaming =
|
|
(blamer_bundle == NULL || !blamer_bundle->segsearch_is_looking_for_blame);
|
|
|
|
// Log new segmentation (for dict_->LogNewChoice()).
|
|
if (not_blaming) {
|
|
PIECES_STATE pieces_widths;
|
|
bin_to_pieces(&temp_state, chunks_record->ratings->dimension() - 1,
|
|
pieces_widths);
|
|
dict_->LogNewSegmentation(pieces_widths);
|
|
}
|
|
|
|
if (language_model_debug_level > 0) {
|
|
STRING word_str;
|
|
word->string_and_lengths(&word_str, NULL);
|
|
tprintf("UpdateBestChoice() constructed word %s\n", word_str.string());
|
|
if (language_model_debug_level > 2) word->print();
|
|
};
|
|
|
|
// Update raw_choice if needed.
|
|
if ((vse->top_choice_flags & kSmallestRatingFlag) &&
|
|
word->rating() < best_choice_bundle->raw_choice->rating() &&
|
|
not_blaming) {
|
|
dict_->LogNewChoice(1.0, certainties, true, word, temp_best_char_choices);
|
|
*(best_choice_bundle->raw_choice) = *word;
|
|
best_choice_bundle->raw_choice->set_permuter(TOP_CHOICE_PERM);
|
|
if (language_model_debug_level > 0) tprintf("Updated raw choice\n");
|
|
}
|
|
|
|
// When working with non-space delimited languages we re-adjust the cost
|
|
// taking into account the final dawg_score and a more precise shape cost.
|
|
// While constructing the paths we assume that they are all dictionary words
|
|
// (since a single character would be a valid dictionary word). At the end
|
|
// we compute dawg_score which reflects how many characters on the path are
|
|
// "covered" by dictionary words of length > 1.
|
|
if (vse->associate_stats.full_wh_ratio_var != 0.0f ||
|
|
(dict_->GetMaxFixedLengthDawgIndex() >= 0 && dawg_score < 1.0f)) {
|
|
vse->cost = ComputeAdjustedPathCost(
|
|
vse->ratings_sum, vse->length, dawg_score, vse->dawg_info,
|
|
vse->ngram_info, vse->consistency_info, vse->associate_stats,
|
|
vse->parent_vse);
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("Updated vse cost to %g (dawg_score %g full_wh_ratio_var %g)\n",
|
|
vse->cost, dawg_score, vse->associate_stats.full_wh_ratio_var);
|
|
}
|
|
}
|
|
|
|
// Update best choice and best char choices if needed.
|
|
// TODO(daria): re-write AcceptableChoice() and NoDangerousAmbig()
|
|
// to fit better into the new segmentation search.
|
|
word->set_rating(vse->cost);
|
|
if (word->rating() < best_choice_bundle->best_choice->rating() &&
|
|
not_blaming) {
|
|
dict_->LogNewChoice(vse->cost / (language_model_ngram_on ?
|
|
vse->ngram_info->ngram_cost :
|
|
vse->ratings_sum),
|
|
certainties, false, word, temp_best_char_choices);
|
|
// Since the rating of the word could have been modified by
|
|
// Dict::LogNewChoice() - check again.
|
|
if (word->rating() < best_choice_bundle->best_choice->rating()) {
|
|
bool modified_blobs; // not used
|
|
DANGERR fixpt;
|
|
if (dict_->AcceptableChoice(&temp_best_char_choices, word, &fixpt,
|
|
ASSOCIATOR_CALLER, &modified_blobs) &&
|
|
AcceptablePath(*vse)) {
|
|
acceptable_choice_found_ = true;
|
|
}
|
|
// Update best_choice_bundle.
|
|
*(best_choice_bundle->best_choice) = *word;
|
|
best_choice_bundle->updated = true;
|
|
best_choice_bundle->best_char_choices->delete_data_pointers();
|
|
best_choice_bundle->best_char_choices->clear();
|
|
for (i = 0; i < temp_best_char_choices.size(); ++i) {
|
|
BLOB_CHOICE_LIST *cc_list = new BLOB_CHOICE_LIST();
|
|
cc_list->deep_copy(temp_best_char_choices[i], &BLOB_CHOICE::deep_copy);
|
|
best_choice_bundle->best_char_choices->push_back(cc_list);
|
|
}
|
|
best_choice_bundle->best_state->part2 = temp_state.part2;
|
|
best_choice_bundle->best_state->part1 = temp_state.part1;
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("Updated best choice\n");
|
|
print_state("New state ", best_choice_bundle->best_state,
|
|
chunks_record->ratings->dimension()-1);
|
|
}
|
|
// Update hyphen state if we are dealing with a dictionary word.
|
|
if (vse->dawg_info != NULL && dict_->GetMaxFixedLengthDawgIndex() < 0) {
|
|
if (dict_->has_hyphen_end(*word)) {
|
|
dict_->set_hyphen_word(*word, *(dawg_args_->active_dawgs),
|
|
*(dawg_args_->constraints));
|
|
} else {
|
|
dict_->reset_hyphen_vars(true);
|
|
}
|
|
}
|
|
best_choice_bundle->best_vse = vse;
|
|
best_choice_bundle->best_b = b;
|
|
best_choice_bundle->fixpt = fixpt;
|
|
|
|
if (blamer_bundle != NULL) {
|
|
blamer_bundle->best_choice_is_dict_and_top_choice =
|
|
(vse->dawg_info != NULL &&
|
|
dict_->GetMaxFixedLengthDawgIndex() < 0 &&
|
|
(vse->top_choice_flags));
|
|
}
|
|
}
|
|
}
|
|
if (blamer_bundle != NULL) {
|
|
// Record the current hypothesis in params_training_bundle.
|
|
ParamsTrainingHypothesis &hyp =
|
|
blamer_bundle->params_training_bundle.AddHypothesis();
|
|
word->string_and_lengths(&(hyp.str), NULL);
|
|
ExtractRawFeaturesFromPath(*vse, hyp.features);
|
|
if (truth_path && // update best truth path rating
|
|
word->rating() < blamer_bundle->best_correctly_segmented_rating) {
|
|
blamer_bundle->best_correctly_segmented_rating = word->rating();
|
|
}
|
|
}
|
|
|
|
// Clean up.
|
|
delete[] certainties;
|
|
delete word;
|
|
}
|
|
|
|
void LanguageModel::ExtractRawFeaturesFromPath(const ViterbiStateEntry &vse,
|
|
float *features) {
|
|
for (int f = 0; f < PTRAIN_NUM_RAW_FEATURE_TYPES; ++f) features[f] = 0.0;
|
|
// Dictionary-related features.
|
|
if (vse.dawg_info != NULL) {
|
|
features[PTRAIN_RAW_FEATURE_DICT_MATCH_TYPE] = vse.dawg_info->permuter;
|
|
|
|
// Mark as unambiguous if unambig_dawg is found among active dawgs.
|
|
for (int d = 0; d < vse.dawg_info->active_dawgs->size(); ++d) {
|
|
if (dict_->GetDawg(vse.dawg_info->active_dawgs->get(d).dawg_index) ==
|
|
dict_->GetUnambigDawg()) {
|
|
features[PTRAIN_RAW_FEATURE_UNAMBIG_DICT_MATCH] = 1.0;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
if (vse.associate_stats.shape_cost > 0) {
|
|
features[PTRAIN_RAW_FEATURE_SHAPE_COST] = vse.associate_stats.shape_cost;
|
|
}
|
|
if (language_model_ngram_on) {
|
|
ASSERT_HOST(vse.ngram_info != NULL);
|
|
features[PTRAIN_RAW_FEATURE_NGRAM_PROB] = vse.ngram_info->ngram_prob;
|
|
}
|
|
// Consistency-related features.
|
|
features[PTRAIN_RAW_FEATURE_NUM_BAD_PUNC] =
|
|
vse.consistency_info.NumInconsistentPunc();
|
|
features[PTRAIN_RAW_FEATURE_NUM_BAD_CASE] =
|
|
vse.consistency_info.NumInconsistentCase();
|
|
features[PTRAIN_RAW_FEATURE_NUM_BAD_CHAR_TYPE] =
|
|
vse.consistency_info.NumInconsistentChartype();
|
|
features[PTRAIN_RAW_FEATURE_NUM_BAD_SPACING] =
|
|
vse.consistency_info.NumInconsistentSpaces();
|
|
features[PTRAIN_RAW_FEATURE_NUM_BAD_SCRIPT] =
|
|
vse.consistency_info.inconsistent_script;
|
|
features[PTRAIN_RAW_FEATURE_NUM_BAD_FONT] =
|
|
vse.consistency_info.inconsistent_font;
|
|
// Classifier-related features.
|
|
features[PTRAIN_RAW_FEATURE_WORST_CERT] = vse.min_certainty;
|
|
features[PTRAIN_RAW_FEATURE_RATING] = vse.ratings_sum;
|
|
features[PTRAIN_RAW_FEATURE_ADAPTED] = vse.adapted;
|
|
// Normalization-related features.
|
|
features[PTRAIN_RAW_FEATURE_NUM_UNICHARS] = vse.length;
|
|
features[PTRAIN_RAW_FEATURE_OUTLINE_LEN] = vse.outline_length;
|
|
}
|
|
|
|
WERD_CHOICE *LanguageModel::ConstructWord(
|
|
BLOB_CHOICE *b,
|
|
ViterbiStateEntry *vse,
|
|
CHUNKS_RECORD *chunks_record,
|
|
BLOB_CHOICE_LIST_VECTOR *best_char_choices,
|
|
float certainties[],
|
|
float *dawg_score,
|
|
STATE *state,
|
|
BlamerBundle *blamer_bundle,
|
|
bool *truth_path) {
|
|
uinT64 state_uint = 0x0;
|
|
if (truth_path != NULL) {
|
|
*truth_path =
|
|
(blamer_bundle != NULL &&
|
|
!blamer_bundle->correct_segmentation_cols.empty() &&
|
|
vse->length == blamer_bundle->correct_segmentation_cols.length());
|
|
}
|
|
const uinT64 kHighestBitOn = 0x8000000000000000LL;
|
|
BLOB_CHOICE *curr_b = b;
|
|
LanguageModelState *curr_lms =
|
|
reinterpret_cast<LanguageModelState *>(b->language_model_state());
|
|
ViterbiStateEntry *curr_vse = vse;
|
|
|
|
int i;
|
|
bool compound = dict_->hyphenated(); // treat hyphenated words as compound
|
|
bool dawg_score_done = true;
|
|
if (dawg_score != NULL) {
|
|
*dawg_score = 0.0f;
|
|
// For space-delimited languages the presence of dawg_info in the last
|
|
// ViterbyStateEntry on the path means that the whole path represents
|
|
// a valid dictionary word.
|
|
if (dict_->GetMaxFixedLengthDawgIndex() < 0) {
|
|
if (vse->dawg_info != NULL) *dawg_score = 1.0f;
|
|
} else if (vse->length == 1) {
|
|
*dawg_score = 1.0f; // each one-letter word is legal
|
|
dawg_score_done = true; // in non-space delimited languages
|
|
} else {
|
|
dawg_score_done = false; // do more work to compute dawg_score
|
|
}
|
|
}
|
|
// For non-space delimited languages we compute the fraction of letters
|
|
// "covered" by fixed length dawgs (i.e. words of length > 1 on the path).
|
|
int covered_by_fixed_length_dawgs = 0;
|
|
// The number of unichars remaining that should be skipped because
|
|
// they are covered by the previous word from fixed length dawgs.
|
|
int fixed_length_num_unichars_to_skip = 0;
|
|
|
|
// Re-compute the variance of the width-to-height ratios (since we now
|
|
// can compute the mean over the whole word).
|
|
float full_wh_ratio_mean = 0.0f;
|
|
if (vse->associate_stats.full_wh_ratio_var != 0.0f) {
|
|
vse->associate_stats.shape_cost -= vse->associate_stats.full_wh_ratio_var;
|
|
full_wh_ratio_mean = (vse->associate_stats.full_wh_ratio_total /
|
|
static_cast<float>(vse->length));
|
|
vse->associate_stats.full_wh_ratio_var = 0.0f;
|
|
}
|
|
|
|
// Construct a WERD_CHOICE by tracing parent pointers.
|
|
WERD_CHOICE *word = new WERD_CHOICE(chunks_record->word_res->uch_set,
|
|
vse->length);
|
|
word->set_length(vse->length);
|
|
for (i = (vse->length-1); i >= 0; --i) {
|
|
if (blamer_bundle != NULL && truth_path != NULL && *truth_path) {
|
|
if ((blamer_bundle->correct_segmentation_cols[i] !=
|
|
curr_lms->contained_in_col) ||
|
|
(blamer_bundle->correct_segmentation_rows[i] !=
|
|
curr_lms->contained_in_row)) {
|
|
*truth_path = false;
|
|
}
|
|
}
|
|
word->set_unichar_id(curr_b->unichar_id(), i);
|
|
word->set_fragment_length(1, i);
|
|
if (certainties != NULL) certainties[i] = curr_b->certainty();
|
|
if (best_char_choices != NULL) {
|
|
best_char_choices->set(chunks_record->ratings->get(
|
|
curr_lms->contained_in_col, curr_lms->contained_in_row), i);
|
|
}
|
|
if (state != NULL) {
|
|
// Record row minus col zeroes in the reverse state to mark the number
|
|
// of joins done by using a blob from this cell in the ratings matrix.
|
|
state_uint >>= (curr_lms->contained_in_row - curr_lms->contained_in_col);
|
|
// Record a one in the reverse state to indicate the split before
|
|
// the blob from the next cell in the ratings matrix (unless we are
|
|
// at the first cell, in which case there is no next blob).
|
|
if (i != 0) {
|
|
state_uint >>= 1;
|
|
state_uint |= kHighestBitOn;
|
|
}
|
|
}
|
|
// For non-space delimited languages: find word endings recorded while
|
|
// trying to separate the current path into words (for words found in
|
|
// fixed length dawgs.
|
|
if (!dawg_score_done && curr_vse->dawg_info != NULL) {
|
|
UpdateCoveredByFixedLengthDawgs(*(curr_vse->dawg_info->active_dawgs),
|
|
i, vse->length,
|
|
&fixed_length_num_unichars_to_skip,
|
|
&covered_by_fixed_length_dawgs,
|
|
dawg_score, &dawg_score_done);
|
|
}
|
|
// Update the width-to-height ratio variance. Useful non-space delimited
|
|
// languages to ensure that the blobs are of uniform width.
|
|
// Skip leading and trailing punctuation when computing the variance.
|
|
if ((full_wh_ratio_mean != 0.0f &&
|
|
((curr_vse != vse && curr_vse->parent_vse != NULL) ||
|
|
!dict_->getUnicharset().get_ispunctuation(curr_b->unichar_id())))) {
|
|
vse->associate_stats.full_wh_ratio_var +=
|
|
pow(full_wh_ratio_mean - curr_vse->associate_stats.full_wh_ratio, 2);
|
|
if (language_model_debug_level > 2) {
|
|
tprintf("full_wh_ratio_var += (%g-%g)^2\n",
|
|
full_wh_ratio_mean, curr_vse->associate_stats.full_wh_ratio);
|
|
}
|
|
}
|
|
|
|
// Mark the word as compound if compound permuter was set for any of
|
|
// the unichars on the path (usually this will happen for unichars
|
|
// that are compounding operators, like "-" and "/").
|
|
if (!compound && curr_vse->dawg_info &&
|
|
curr_vse->dawg_info->permuter == COMPOUND_PERM) compound = true;
|
|
|
|
// Update curr_* pointers.
|
|
if (curr_vse->parent_b == NULL) break;
|
|
curr_b = curr_vse->parent_b;
|
|
curr_lms =
|
|
reinterpret_cast<LanguageModelState *>(curr_b->language_model_state());
|
|
curr_vse = curr_vse->parent_vse;
|
|
}
|
|
ASSERT_HOST(i == 0); // check that we recorded all the unichar ids
|
|
// Re-adjust shape cost to include the updated width-to-height variance.
|
|
if (full_wh_ratio_mean != 0.0f) {
|
|
vse->associate_stats.shape_cost += vse->associate_stats.full_wh_ratio_var;
|
|
}
|
|
|
|
if (state != NULL) {
|
|
state_uint >>= (64 - (chunks_record->ratings->dimension()-1));
|
|
state->part2 = state_uint;
|
|
state_uint >>= 32;
|
|
state->part1 = state_uint;
|
|
}
|
|
word->set_rating(vse->ratings_sum);
|
|
word->set_certainty(vse->min_certainty);
|
|
if (vse->dawg_info != NULL && dict_->GetMaxFixedLengthDawgIndex() < 0) {
|
|
word->set_permuter(compound ? COMPOUND_PERM : vse->dawg_info->permuter);
|
|
} else if (language_model_ngram_on && !vse->ngram_info->pruned) {
|
|
word->set_permuter(NGRAM_PERM);
|
|
} else if (vse->top_choice_flags) {
|
|
word->set_permuter(TOP_CHOICE_PERM);
|
|
} else {
|
|
word->set_permuter(NO_PERM);
|
|
}
|
|
return word;
|
|
}
|
|
|
|
void LanguageModel::UpdateCoveredByFixedLengthDawgs(
|
|
const DawgInfoVector &active_dawgs, int word_index, int word_length,
|
|
int *skip, int *covered, float *dawg_score, bool *dawg_score_done) {
|
|
if (language_model_debug_level > 3) {
|
|
tprintf("UpdateCoveredByFixedLengthDawgs for index %d skip=%d\n",
|
|
word_index, *skip, word_length);
|
|
}
|
|
|
|
if (*skip > 0) {
|
|
--(*skip);
|
|
} else {
|
|
int best_index = -1;
|
|
for (int d = 0; d < active_dawgs.size(); ++d) {
|
|
int dawg_index = (active_dawgs[d]).dawg_index;
|
|
if (dawg_index > dict_->GetMaxFixedLengthDawgIndex()) {
|
|
// If active_dawgs of the last ViterbiStateEntry on the path
|
|
// contain a non-fixed length dawg, this means that the whole
|
|
// path represents a word from a non-fixed length word dawg.
|
|
if (word_index == (word_length - 1)) {
|
|
*dawg_score = 1.0f;
|
|
*dawg_score_done = true;
|
|
return;
|
|
}
|
|
} else if (dawg_index >= kMinFixedLengthDawgLength) {
|
|
const Dawg *curr_dawg = dict_->GetDawg(dawg_index);
|
|
ASSERT_HOST(curr_dawg != NULL);
|
|
if ((active_dawgs[d]).ref != NO_EDGE &&
|
|
curr_dawg->end_of_word((active_dawgs[d]).ref) &&
|
|
dawg_index > best_index) {
|
|
best_index = dawg_index;
|
|
}
|
|
|
|
if (language_model_debug_level > 3) {
|
|
tprintf("dawg_index %d, ref %d, eow %d\n", dawg_index,
|
|
(active_dawgs[d]).ref,
|
|
((active_dawgs[d]).ref != NO_EDGE &&
|
|
curr_dawg->end_of_word((active_dawgs[d]).ref)));
|
|
}
|
|
}
|
|
} // end for
|
|
if (best_index != -1) {
|
|
*skip = best_index - 1;
|
|
*covered += best_index;
|
|
}
|
|
} // end if/else skip
|
|
|
|
if (word_index == 0) {
|
|
ASSERT_HOST(*covered <= word_length);
|
|
*dawg_score = (static_cast<float>(*covered) /
|
|
static_cast<float>(word_length));
|
|
*dawg_score_done = true;
|
|
}
|
|
}
|
|
|
|
void LanguageModel::GeneratePainPointsFromColumn(
|
|
int col,
|
|
const GenericVector<int> &non_empty_rows,
|
|
float best_choice_cert,
|
|
HEAP *pain_points,
|
|
BestPathByColumn *best_path_by_column[],
|
|
CHUNKS_RECORD *chunks_record) {
|
|
for (int i = 0; i < non_empty_rows.length(); ++i) {
|
|
int row = non_empty_rows[i];
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("\nLooking for pain points in col=%d row=%d\n", col, row);
|
|
}
|
|
if (language_model_ngram_on) {
|
|
GenerateNgramModelPainPointsFromColumn(
|
|
col, row, pain_points, chunks_record);
|
|
} else {
|
|
GenerateProblematicPathPainPointsFromColumn(
|
|
col, row, best_choice_cert, pain_points,
|
|
best_path_by_column, chunks_record);
|
|
}
|
|
}
|
|
}
|
|
|
|
void LanguageModel::GenerateNgramModelPainPointsFromColumn(
|
|
int col, int row, HEAP *pain_points, CHUNKS_RECORD *chunks_record) {
|
|
// Find the first top choice path recorded for this cell.
|
|
// If this path is pruned - generate a pain point.
|
|
ASSERT_HOST(chunks_record->ratings->get(col, row) != NULL);
|
|
BLOB_CHOICE_IT bit(chunks_record->ratings->get(col, row));
|
|
bool fragmented = false;
|
|
for (bit.mark_cycle_pt(); !bit.cycled_list(); bit.forward()) {
|
|
if (dict_->getUnicharset().get_fragment(bit.data()->unichar_id())) {
|
|
fragmented = true;
|
|
continue;
|
|
}
|
|
LanguageModelState *lms = reinterpret_cast<LanguageModelState *>(
|
|
bit.data()->language_model_state());
|
|
if (lms == NULL) continue;
|
|
ViterbiStateEntry_IT vit(&(lms->viterbi_state_entries));
|
|
for (vit.mark_cycle_pt(); !vit.cycled_list(); vit.forward()) {
|
|
const ViterbiStateEntry *vse = vit.data();
|
|
if (!vse->top_choice_flags) continue;
|
|
ASSERT_HOST(vse->ngram_info != NULL);
|
|
if (vse->ngram_info->pruned && (vse->parent_vse == NULL ||
|
|
!vse->parent_vse->ngram_info->pruned)) {
|
|
if (vse->parent_vse != NULL) {
|
|
LanguageModelState *pp_lms = reinterpret_cast<LanguageModelState *>(
|
|
vse->parent_b->language_model_state());
|
|
GeneratePainPoint(pp_lms->contained_in_col, row, false,
|
|
kDefaultPainPointPriorityAdjustment,
|
|
-1.0f, fragmented, -1.0f,
|
|
max_char_wh_ratio_,
|
|
vse->parent_vse->parent_b,
|
|
vse->parent_vse->parent_vse,
|
|
chunks_record, pain_points);
|
|
}
|
|
if (vse->parent_vse != NULL &&
|
|
vse->parent_vse->parent_vse != NULL &&
|
|
dict_->getUnicharset().get_ispunctuation(
|
|
vse->parent_b->unichar_id())) {
|
|
// If the dip in the ngram probability is due to punctuation in the
|
|
// middle of the word - go two unichars back to combine this
|
|
// punctuation mark with the previous character on the path.
|
|
LanguageModelState *pp_lms = reinterpret_cast<LanguageModelState *>(
|
|
vse->parent_vse->parent_b->language_model_state());
|
|
GeneratePainPoint(pp_lms->contained_in_col, col-1, false,
|
|
kDefaultPainPointPriorityAdjustment,
|
|
-1.0f, fragmented, -1.0f,
|
|
max_char_wh_ratio_,
|
|
vse->parent_vse->parent_vse->parent_b,
|
|
vse->parent_vse->parent_vse->parent_vse,
|
|
chunks_record, pain_points);
|
|
} else if (row+1 < chunks_record->ratings->dimension()) {
|
|
GeneratePainPoint(col, row+1, true,
|
|
kDefaultPainPointPriorityAdjustment,
|
|
-1.0f, fragmented, -1.0f,
|
|
max_char_wh_ratio_,
|
|
vse->parent_b,
|
|
vse->parent_vse,
|
|
chunks_record, pain_points);
|
|
}
|
|
}
|
|
return; // examined the lowest cost top choice path
|
|
}
|
|
}
|
|
}
|
|
|
|
void LanguageModel::GenerateProblematicPathPainPointsFromColumn(
|
|
int col, int row, float best_choice_cert,
|
|
HEAP *pain_points, BestPathByColumn *best_path_by_column[],
|
|
CHUNKS_RECORD *chunks_record) {
|
|
MATRIX *ratings = chunks_record->ratings;
|
|
|
|
// Get the best path from this matrix cell.
|
|
BLOB_CHOICE_LIST *blist = ratings->get(col, row);
|
|
ASSERT_HOST(blist != NULL);
|
|
if (blist->empty()) return;
|
|
BLOB_CHOICE_IT bit(blist);
|
|
bool fragment = false;
|
|
while (dict_->getUnicharset().get_fragment(bit.data()->unichar_id()) &&
|
|
!bit.at_last()) { // skip fragments
|
|
fragment = true;
|
|
bit.forward();
|
|
}
|
|
if (bit.data()->language_model_state() == NULL) return;
|
|
ViterbiStateEntry_IT vit(&(reinterpret_cast<LanguageModelState *>(
|
|
bit.data()->language_model_state())->viterbi_state_entries));
|
|
if (vit.empty()) return;
|
|
ViterbiStateEntry *vse = vit.data();
|
|
// Check whether this path is promising.
|
|
bool path_is_promising = true;
|
|
if (vse->parent_vse != NULL) {
|
|
float potential_avg_cost =
|
|
((vse->parent_vse->cost + bit.data()->rating()*0.5f) /
|
|
static_cast<float>(row+1));
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("potential_avg_cost %g best cost %g\n",
|
|
potential_avg_cost, (*best_path_by_column)[col].avg_cost);
|
|
}
|
|
if (potential_avg_cost >= (*best_path_by_column)[col].avg_cost) {
|
|
path_is_promising = false;
|
|
}
|
|
}
|
|
// Set best_parent_vse to the best parent recorded in best_path_by_column.
|
|
ViterbiStateEntry *best_parent_vse = vse->parent_vse;
|
|
BLOB_CHOICE *best_parent_b = vse->parent_b;
|
|
if (col > 0 && (*best_path_by_column)[col-1].best_vse != NULL) {
|
|
ASSERT_HOST((*best_path_by_column)[col-1].best_b != NULL);
|
|
LanguageModelState *best_lms = reinterpret_cast<LanguageModelState *>(
|
|
((*best_path_by_column)[col-1].best_b)->language_model_state());
|
|
if (best_lms->contained_in_row == col-1) {
|
|
best_parent_vse = (*best_path_by_column)[col-1].best_vse;
|
|
best_parent_b = (*best_path_by_column)[col-1].best_b;
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("Setting best_parent_vse to %p\n", best_parent_vse);
|
|
}
|
|
}
|
|
}
|
|
// Check whether this entry terminates the best parent path
|
|
// recorded in best_path_by_column.
|
|
bool best_not_prolonged = (best_parent_vse != vse->parent_vse);
|
|
|
|
// If this path is problematic because of the last unichar - generate
|
|
// a pain point to combine it with its left and right neighbor.
|
|
BLOB_CHOICE_IT tmp_bit;
|
|
if (best_not_prolonged ||
|
|
(path_is_promising &&
|
|
ProblematicPath(*vse, bit.data()->unichar_id(),
|
|
row+1 == ratings->dimension()))) {
|
|
float worst_piece_cert;
|
|
bool fragmented;
|
|
if (col-1 > 0) {
|
|
GetWorstPieceCertainty(col-1, row, chunks_record->ratings,
|
|
&worst_piece_cert, &fragmented);
|
|
GeneratePainPoint(col-1, row, false,
|
|
kDefaultPainPointPriorityAdjustment,
|
|
worst_piece_cert, fragmented, best_choice_cert,
|
|
max_char_wh_ratio_, best_parent_b, best_parent_vse,
|
|
chunks_record, pain_points);
|
|
}
|
|
if (row+1 < ratings->dimension()) {
|
|
GetWorstPieceCertainty(col, row+1, chunks_record->ratings,
|
|
&worst_piece_cert, &fragmented);
|
|
GeneratePainPoint(col, row+1, true, kDefaultPainPointPriorityAdjustment,
|
|
worst_piece_cert, fragmented, best_choice_cert,
|
|
max_char_wh_ratio_, best_parent_b, best_parent_vse,
|
|
chunks_record, pain_points);
|
|
}
|
|
} // for ProblematicPath
|
|
}
|
|
|
|
void LanguageModel::GeneratePainPointsFromBestChoice(
|
|
HEAP *pain_points,
|
|
CHUNKS_RECORD *chunks_record,
|
|
BestChoiceBundle *best_choice_bundle) {
|
|
// Variables to backtrack best_vse path;
|
|
ViterbiStateEntry *curr_vse = best_choice_bundle->best_vse;
|
|
BLOB_CHOICE *curr_b = best_choice_bundle->best_b;
|
|
|
|
// Begins and ends in DANGERR vector record the positions in the blob choice
|
|
// list of the best choice. We need to translate these endpoints into the
|
|
// beginning column and ending row for the pain points. We maintain
|
|
// danger_begin and danger_end arrays indexed by the position in
|
|
// best_choice_bundle->best_char_choices (which is equal to the position
|
|
// on the best_choice_bundle->best_vse path).
|
|
// danger_end[d] stores the DANGERR_INFO structs with end==d and is
|
|
// initialized at the beginning of this function.
|
|
// danger_begin[d] stores the DANGERR_INFO struct with begin==d and
|
|
// has end set to the row of the end of this ambiguity.
|
|
// The translation from end in terms of the best choice index to the end row
|
|
// is done while following the parents of best_choice_bundle->best_vse.
|
|
assert(best_choice_bundle->best_char_choices->length() ==
|
|
best_choice_bundle->best_vse->length);
|
|
DANGERR *danger_begin = NULL;
|
|
DANGERR *danger_end = NULL;
|
|
int d;
|
|
if (!best_choice_bundle->fixpt.empty()) {
|
|
danger_begin = new DANGERR[best_choice_bundle->best_vse->length];
|
|
danger_end = new DANGERR[best_choice_bundle->best_vse->length];
|
|
for (d = 0; d < best_choice_bundle->fixpt.size(); ++d) {
|
|
const DANGERR_INFO &danger = best_choice_bundle->fixpt[d];
|
|
// Only use n->1 ambiguities.
|
|
if (danger.end > danger.begin && !danger.correct_is_ngram &&
|
|
(!language_model_ngram_on || danger.dangerous)) {
|
|
danger_end[danger.end].push_back(danger);
|
|
}
|
|
}
|
|
}
|
|
|
|
// Variables to keep track of punctuation/number streaks.
|
|
int punc_streak_end_row = -1;
|
|
int punc_streak_length = 0;
|
|
float punc_streak_min_cert = 0.0f;
|
|
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("\nGenerating pain points for best path=%p\n", curr_vse);
|
|
}
|
|
|
|
int word_index = best_choice_bundle->best_vse->length;
|
|
while (curr_vse != NULL) {
|
|
word_index--;
|
|
ASSERT_HOST(word_index >= 0);
|
|
ASSERT_HOST(curr_b != NULL);
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("Looking at unichar %s\n",
|
|
dict_->getUnicharset().id_to_unichar(curr_b->unichar_id()));
|
|
}
|
|
|
|
int pp_col = reinterpret_cast<LanguageModelState *>(
|
|
curr_b->language_model_state())->contained_in_col;
|
|
int pp_row = reinterpret_cast<LanguageModelState *>(
|
|
curr_b->language_model_state())->contained_in_row;
|
|
|
|
// Generate pain points for ambiguities found by NoDangerousAmbig().
|
|
if (danger_end != NULL) {
|
|
// Translate end index of an ambiguity to an end row.
|
|
for (d = 0; d < danger_end[word_index].size(); ++d) {
|
|
danger_end[word_index][d].end = pp_row;
|
|
danger_begin[danger_end[word_index][d].begin].push_back(
|
|
danger_end[word_index][d]);
|
|
}
|
|
// Generate a pain point to combine unchars in the "wrong" part
|
|
// of the ambiguity.
|
|
for (d = 0; d < danger_begin[word_index].size(); ++d) {
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("Generating pain point from %sambiguity\n",
|
|
danger_begin[word_index][d].dangerous ? "dangerous " : "");
|
|
}
|
|
GeneratePainPoint(pp_col, danger_begin[word_index][d].end, false,
|
|
danger_begin[word_index][d].dangerous ?
|
|
kCriticalPainPointPriorityAdjustment :
|
|
kBestChoicePainPointPriorityAdjustment,
|
|
best_choice_bundle->best_choice->certainty(), true,
|
|
best_choice_bundle->best_choice->certainty(),
|
|
kLooseMaxCharWhRatio,
|
|
curr_vse->parent_b, curr_vse->parent_vse,
|
|
chunks_record, pain_points);
|
|
}
|
|
}
|
|
|
|
if (!language_model_ngram_on) { // no need to use further heuristics if we
|
|
// are guided by the character ngram model
|
|
// Generate pain points for problematic sub-paths.
|
|
if (ProblematicPath(*curr_vse, curr_b->unichar_id(),
|
|
pp_row+1 == chunks_record->ratings->dimension())) {
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("Generating pain point from a problematic sub-path\n");
|
|
}
|
|
float worst_piece_cert;
|
|
bool fragment;
|
|
if (pp_col > 0) {
|
|
GetWorstPieceCertainty(pp_col-1, pp_row, chunks_record->ratings,
|
|
&worst_piece_cert, &fragment);
|
|
GeneratePainPoint(pp_col-1, pp_row, false,
|
|
kBestChoicePainPointPriorityAdjustment,
|
|
worst_piece_cert, true,
|
|
best_choice_bundle->best_choice->certainty(),
|
|
max_char_wh_ratio_, NULL, NULL,
|
|
chunks_record, pain_points);
|
|
}
|
|
if (pp_row+1 < chunks_record->ratings->dimension()) {
|
|
GetWorstPieceCertainty(pp_col, pp_row+1, chunks_record->ratings,
|
|
&worst_piece_cert, &fragment);
|
|
GeneratePainPoint(pp_col, pp_row+1, true,
|
|
kBestChoicePainPointPriorityAdjustment,
|
|
worst_piece_cert, true,
|
|
best_choice_bundle->best_choice->certainty(),
|
|
max_char_wh_ratio_, NULL, NULL,
|
|
chunks_record, pain_points);
|
|
}
|
|
}
|
|
|
|
// Generate a pain point if we encountered a punctuation/number streak to
|
|
// combine all punctuation marks into one blob and try to classify it.
|
|
bool is_alpha = dict_->getUnicharset().get_isalpha(curr_b->unichar_id());
|
|
if (!is_alpha) {
|
|
if (punc_streak_end_row == -1) punc_streak_end_row = pp_row;
|
|
punc_streak_length++;
|
|
if (curr_b->certainty() < punc_streak_min_cert)
|
|
punc_streak_min_cert = curr_b->certainty();
|
|
}
|
|
if (is_alpha || curr_vse->parent_vse == NULL) {
|
|
if (punc_streak_end_row != -1 && punc_streak_length > 1) {
|
|
if (language_model_debug_level > 0) {
|
|
tprintf("Generating pain point from a punctuation streak\n");
|
|
}
|
|
if (is_alpha ||
|
|
(curr_vse->parent_vse == NULL && punc_streak_length > 2)) {
|
|
GeneratePainPoint(pp_row+1, punc_streak_end_row, false,
|
|
kBestChoicePainPointPriorityAdjustment,
|
|
punc_streak_min_cert, true,
|
|
best_choice_bundle->best_choice->certainty(),
|
|
max_char_wh_ratio_, curr_b, curr_vse,
|
|
chunks_record, pain_points);
|
|
}
|
|
// Handle punctuation/number streaks starting from the first unichar.
|
|
if (curr_vse->parent_vse == NULL) {
|
|
GeneratePainPoint(0, punc_streak_end_row, false,
|
|
kBestChoicePainPointPriorityAdjustment,
|
|
punc_streak_min_cert, true,
|
|
best_choice_bundle->best_choice->certainty(),
|
|
max_char_wh_ratio_, NULL, NULL,
|
|
chunks_record, pain_points);
|
|
}
|
|
}
|
|
punc_streak_end_row = -1;
|
|
punc_streak_length = 0;
|
|
punc_streak_min_cert = 0.0f;
|
|
} // end handling punctuation streaks
|
|
}
|
|
|
|
curr_b = curr_vse->parent_b;
|
|
curr_vse = curr_vse->parent_vse;
|
|
} // end looking at best choice subpaths
|
|
|
|
// Clean up.
|
|
if (danger_end != NULL) {
|
|
delete[] danger_begin;
|
|
delete[] danger_end;
|
|
}
|
|
}
|
|
|
|
bool LanguageModel::GeneratePainPoint(
|
|
int col, int row, bool ok_to_extend, float priority,
|
|
float worst_piece_cert, bool fragmented, float best_choice_cert,
|
|
float max_char_wh_ratio,
|
|
BLOB_CHOICE *parent_b, ViterbiStateEntry *parent_vse,
|
|
CHUNKS_RECORD *chunks_record, HEAP *pain_points) {
|
|
if (col < 0 || row >= chunks_record->ratings->dimension() ||
|
|
chunks_record->ratings->get(col, row) != NOT_CLASSIFIED) {
|
|
return false;
|
|
}
|
|
if (language_model_debug_level > 3) {
|
|
tprintf("\nGenerating pain point for col=%d row=%d priority=%g parent=",
|
|
col, row, priority);
|
|
if (parent_vse != NULL) {
|
|
PrintViterbiStateEntry("", parent_vse, parent_b, chunks_record);
|
|
} else {
|
|
tprintf("NULL");
|
|
}
|
|
tprintf("\n");
|
|
}
|
|
|
|
AssociateStats associate_stats;
|
|
ComputeAssociateStats(col, row, max_char_wh_ratio, parent_vse,
|
|
chunks_record, &associate_stats);
|
|
// For fixed-pitch fonts/languages: if the current combined blob overlaps
|
|
// the next blob on the right and it is ok to extend the blob, try expending
|
|
// the blob until there is no overlap with the next blob on the right or
|
|
// until the width-to-height ratio becomes too large.
|
|
if (ok_to_extend) {
|
|
while (associate_stats.bad_fixed_pitch_right_gap &&
|
|
row+1 < chunks_record->ratings->dimension() &&
|
|
!associate_stats.bad_fixed_pitch_wh_ratio) {
|
|
ComputeAssociateStats(col, ++row, max_char_wh_ratio, parent_vse,
|
|
chunks_record, &associate_stats);
|
|
}
|
|
}
|
|
|
|
if (associate_stats.bad_shape) {
|
|
if (language_model_debug_level > 3) {
|
|
tprintf("Discarded pain point with a bad shape\n");
|
|
}
|
|
return false;
|
|
}
|
|
|
|
// Compute pain point priority.
|
|
if (associate_stats.shape_cost > 0) {
|
|
priority *= associate_stats.shape_cost;
|
|
}
|
|
if (worst_piece_cert < best_choice_cert) {
|
|
worst_piece_cert = best_choice_cert;
|
|
}
|
|
priority *= CertaintyScore(worst_piece_cert);
|
|
if (fragmented) priority /= kDefaultPainPointPriorityAdjustment;
|
|
|
|
if (language_model_debug_level > 3) {
|
|
tprintf("worst_piece_cert=%g fragmented=%d\n",
|
|
worst_piece_cert, fragmented);
|
|
}
|
|
|
|
if (parent_vse != NULL) {
|
|
priority *= sqrt(parent_vse->cost / static_cast<float>(col));
|
|
if (parent_vse->dawg_info != NULL) {
|
|
priority /= kDefaultPainPointPriorityAdjustment;
|
|
if (parent_vse->length > language_model_min_compound_length) {
|
|
priority /= sqrt(static_cast<double>(parent_vse->length));
|
|
}
|
|
}
|
|
}
|
|
|
|
MATRIX_COORD *pain_point = new MATRIX_COORD(col, row);
|
|
if (HeapPushCheckSize(pain_points, priority, pain_point)) {
|
|
if (language_model_debug_level) {
|
|
tprintf("Added pain point with priority %g\n", priority);
|
|
}
|
|
return true;
|
|
} else {
|
|
delete pain_point;
|
|
if (language_model_debug_level) tprintf("Pain points heap is full\n");
|
|
return false;
|
|
}
|
|
}
|
|
|
|
} // namespace tesseract
|