mirror of
https://github.com/osm-search/Nominatim.git
synced 2026-02-26 11:08:13 +00:00
switch country name tokens to new word table layout
This commit is contained in:
@@ -146,8 +146,8 @@ class Tokenizer
|
|||||||
private function addTokensFromDB(&$oValidTokens, $aTokens, $sNormQuery)
|
private function addTokensFromDB(&$oValidTokens, $aTokens, $sNormQuery)
|
||||||
{
|
{
|
||||||
// Check which tokens we have, get the ID numbers
|
// Check which tokens we have, get the ID numbers
|
||||||
$sSQL = 'SELECT word_id, word_token, word, class, type, country_code,';
|
$sSQL = 'SELECT word_id, word_token, type';
|
||||||
$sSQL .= ' operator, coalesce(search_name_count, 0) as count';
|
$sSQL .= " info->>'cc' as country";
|
||||||
$sSQL .= ' FROM word WHERE word_token in (';
|
$sSQL .= ' FROM word WHERE word_token in (';
|
||||||
$sSQL .= join(',', $this->oDB->getDBQuotedList($aTokens)).')';
|
$sSQL .= join(',', $this->oDB->getDBQuotedList($aTokens)).')';
|
||||||
|
|
||||||
@@ -156,8 +156,20 @@ class Tokenizer
|
|||||||
$aDBWords = $this->oDB->getAll($sSQL, null, 'Could not get word tokens.');
|
$aDBWords = $this->oDB->getAll($sSQL, null, 'Could not get word tokens.');
|
||||||
|
|
||||||
foreach ($aDBWords as $aWord) {
|
foreach ($aDBWords as $aWord) {
|
||||||
$oToken = null;
|
switch ($aWord['type']) {
|
||||||
$iId = (int) $aWord['word_id'];
|
'C': // country name tokens
|
||||||
|
if ($aWord['country'] === null
|
||||||
|
|| ($this->aCountryRestriction
|
||||||
|
&& !in_array($aWord['country'], $this->aCountryRestriction))
|
||||||
|
) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
$oToken = new Token\Country($iId, $aWord['country'])
|
||||||
|
break;
|
||||||
|
default:
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
/* $iId = (int) $aWord['word_id'];
|
||||||
|
|
||||||
if ($aWord['class']) {
|
if ($aWord['class']) {
|
||||||
// Special terms need to appear in their normalized form.
|
// Special terms need to appear in their normalized form.
|
||||||
@@ -207,16 +219,9 @@ class Tokenizer
|
|||||||
$aWord['word_token'],
|
$aWord['word_token'],
|
||||||
(int) $aWord['count']
|
(int) $aWord['count']
|
||||||
);
|
);
|
||||||
}
|
}*/
|
||||||
|
|
||||||
if ($oToken) {
|
$oValidTokens->addToken($aWord['word_token'], $oToken);
|
||||||
// remove any leading spaces
|
|
||||||
if ($aWord['word_token'][0] == ' ') {
|
|
||||||
$oValidTokens->addToken(substr($aWord['word_token'], 1), $oToken);
|
|
||||||
} else {
|
|
||||||
$oValidTokens->addToken($aWord['word_token'], $oToken);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -234,12 +239,10 @@ class Tokenizer
|
|||||||
|
|
||||||
for ($i = 0; $i < $iNumWords; $i++) {
|
for ($i = 0; $i < $iNumWords; $i++) {
|
||||||
$sPhrase = $aWords[$i];
|
$sPhrase = $aWords[$i];
|
||||||
$aTokens[' '.$sPhrase] = ' '.$sPhrase;
|
|
||||||
$aTokens[$sPhrase] = $sPhrase;
|
$aTokens[$sPhrase] = $sPhrase;
|
||||||
|
|
||||||
for ($j = $i + 1; $j < $iNumWords; $j++) {
|
for ($j = $i + 1; $j < $iNumWords; $j++) {
|
||||||
$sPhrase .= ' '.$aWords[$j];
|
$sPhrase .= ' '.$aWords[$j];
|
||||||
$aTokens[' '.$sPhrase] = ' '.$sPhrase;
|
|
||||||
$aTokens[$sPhrase] = $sPhrase;
|
$aTokens[$sPhrase] = $sPhrase;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -8,6 +8,9 @@ CREATE TABLE word_icu (
|
|||||||
|
|
||||||
CREATE INDEX idx_word_word_token ON word
|
CREATE INDEX idx_word_word_token ON word
|
||||||
USING BTREE (word_token) {{db.tablespace.search_index}};
|
USING BTREE (word_token) {{db.tablespace.search_index}};
|
||||||
|
-- Used when updating country names from the boundary relation.
|
||||||
|
CREATE INDEX idx_word_country_names ON word
|
||||||
|
USING btree((info->>'cc')) WHERE type = 'C';
|
||||||
GRANT SELECT ON word TO "{{config.DATABASE_WEBUSER}}";
|
GRANT SELECT ON word TO "{{config.DATABASE_WEBUSER}}";
|
||||||
|
|
||||||
DROP SEQUENCE IF EXISTS seq_word;
|
DROP SEQUENCE IF EXISTS seq_word;
|
||||||
|
|||||||
@@ -371,22 +371,28 @@ class LegacyICUNameAnalyzer:
|
|||||||
"""
|
"""
|
||||||
word_tokens = set()
|
word_tokens = set()
|
||||||
for name in self._compute_full_names(names):
|
for name in self._compute_full_names(names):
|
||||||
if name:
|
norm_name = self.name_processor.get_search_normalized(name)
|
||||||
word_tokens.add(' ' + self.name_processor.get_search_normalized(name))
|
if norm_name:
|
||||||
|
word_tokens.add(norm_name)
|
||||||
|
|
||||||
with self.conn.cursor() as cur:
|
with self.conn.cursor() as cur:
|
||||||
# Get existing names
|
# Get existing names
|
||||||
cur.execute("SELECT word_token FROM word WHERE country_code = %s",
|
cur.execute("""SELECT word_token FROM word
|
||||||
|
WHERE type = 'C' and info->>'cc'= %s""",
|
||||||
(country_code, ))
|
(country_code, ))
|
||||||
word_tokens.difference_update((t[0] for t in cur))
|
word_tokens.difference_update((t[0] for t in cur))
|
||||||
|
|
||||||
|
# Only add those names that are not yet in the list.
|
||||||
if word_tokens:
|
if word_tokens:
|
||||||
cur.execute("""INSERT INTO word (word_id, word_token, country_code,
|
cur.execute("""INSERT INTO word (word_token, type, info)
|
||||||
search_name_count)
|
(SELECT token, 'C', json_build_object('cc', %s)
|
||||||
(SELECT nextval('seq_word'), token, %s, 0
|
|
||||||
FROM unnest(%s) as token)
|
FROM unnest(%s) as token)
|
||||||
""", (country_code, list(word_tokens)))
|
""", (country_code, list(word_tokens)))
|
||||||
|
|
||||||
|
# No names are deleted at the moment.
|
||||||
|
# If deletion is made possible, then the static names from the
|
||||||
|
# initial 'country_name' table should be kept.
|
||||||
|
|
||||||
|
|
||||||
def process_place(self, place):
|
def process_place(self, place):
|
||||||
""" Determine tokenizer information about the given place.
|
""" Determine tokenizer information about the given place.
|
||||||
|
|||||||
Reference in New Issue
Block a user