mirror of
https://github.com/osm-search/Nominatim.git
synced 2026-03-11 21:34:06 +00:00
icu: move token deduplication into TokenInfo
Puts collection into one common place.
This commit is contained in:
@@ -454,9 +454,7 @@ class LegacyICUNameAnalyzer(AbstractAnalyzer):
|
|||||||
names, address = self.sanitizer.process_names(place)
|
names, address = self.sanitizer.process_names(place)
|
||||||
|
|
||||||
if names:
|
if names:
|
||||||
fulls, partials = self._compute_name_tokens(names)
|
token_info.set_names(*self._compute_name_tokens(names))
|
||||||
|
|
||||||
token_info.add_names(fulls, partials)
|
|
||||||
|
|
||||||
if place.is_country():
|
if place.is_country():
|
||||||
self._add_country_full_names(place.country_code, names)
|
self._add_country_full_names(place.country_code, names)
|
||||||
@@ -464,39 +462,23 @@ class LegacyICUNameAnalyzer(AbstractAnalyzer):
|
|||||||
if address:
|
if address:
|
||||||
self._process_place_address(token_info, address)
|
self._process_place_address(token_info, address)
|
||||||
|
|
||||||
return token_info.data
|
return token_info.to_dict()
|
||||||
|
|
||||||
|
|
||||||
def _process_place_address(self, token_info, address):
|
def _process_place_address(self, token_info, address):
|
||||||
hnr_tokens = set()
|
|
||||||
hnrs = set()
|
|
||||||
addr_terms = []
|
|
||||||
streets = []
|
|
||||||
for item in address:
|
for item in address:
|
||||||
if item.kind == 'postcode':
|
if item.kind == 'postcode':
|
||||||
self._add_postcode(item.name)
|
self._add_postcode(item.name)
|
||||||
elif item.kind == 'housenumber':
|
elif item.kind == 'housenumber':
|
||||||
token, hnr = self._compute_housenumber_token(item)
|
token_info.add_housenumber(*self._compute_housenumber_token(item))
|
||||||
if token is not None:
|
|
||||||
hnr_tokens.add(token)
|
|
||||||
hnrs.add(hnr)
|
|
||||||
elif item.kind == 'street':
|
elif item.kind == 'street':
|
||||||
streets.extend(self._retrieve_full_tokens(item.name))
|
token_info.add_street(self._retrieve_full_tokens(item.name))
|
||||||
elif item.kind == 'place':
|
elif item.kind == 'place':
|
||||||
if not item.suffix:
|
if not item.suffix:
|
||||||
token_info.add_place(self._compute_partial_tokens(item.name))
|
token_info.add_place(self._compute_partial_tokens(item.name))
|
||||||
elif not item.kind.startswith('_') and not item.suffix and \
|
elif not item.kind.startswith('_') and not item.suffix and \
|
||||||
item.kind not in ('country', 'full'):
|
item.kind not in ('country', 'full'):
|
||||||
addr_terms.append((item.kind, self._compute_partial_tokens(item.name)))
|
token_info.add_address_term(item.kind, self._compute_partial_tokens(item.name))
|
||||||
|
|
||||||
if hnrs:
|
|
||||||
token_info.add_housenumbers(hnr_tokens, hnrs)
|
|
||||||
|
|
||||||
if addr_terms:
|
|
||||||
token_info.add_address_terms(addr_terms)
|
|
||||||
|
|
||||||
if streets:
|
|
||||||
token_info.add_street(streets)
|
|
||||||
|
|
||||||
|
|
||||||
def _compute_housenumber_token(self, hnr):
|
def _compute_housenumber_token(self, hnr):
|
||||||
@@ -626,48 +608,75 @@ class _TokenInfo:
|
|||||||
""" Collect token information to be sent back to the database.
|
""" Collect token information to be sent back to the database.
|
||||||
"""
|
"""
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
self.data = {}
|
self.names = None
|
||||||
|
self.housenumbers = set()
|
||||||
|
self.housenumber_tokens = set()
|
||||||
|
self.street_tokens = set()
|
||||||
|
self.place_tokens = set()
|
||||||
|
self.address_tokens = {}
|
||||||
|
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def _mk_array(tokens):
|
def _mk_array(tokens):
|
||||||
return '{%s}' % ','.join((str(s) for s in tokens))
|
return f"{{{','.join((str(s) for s in tokens))}}}"
|
||||||
|
|
||||||
|
|
||||||
def add_names(self, fulls, partials):
|
def to_dict(self):
|
||||||
|
""" Return the token information in database importable format.
|
||||||
|
"""
|
||||||
|
out = {}
|
||||||
|
|
||||||
|
if self.names:
|
||||||
|
out['names'] = self.names
|
||||||
|
|
||||||
|
if self.housenumbers:
|
||||||
|
out['hnr'] = ';'.join(self.housenumbers)
|
||||||
|
out['hnr_tokens'] = self._mk_array(self.housenumber_tokens)
|
||||||
|
|
||||||
|
if self.street_tokens:
|
||||||
|
out['street'] = self._mk_array(self.street_tokens)
|
||||||
|
|
||||||
|
if self.place_tokens:
|
||||||
|
out['place'] = self._mk_array(self.place_tokens)
|
||||||
|
|
||||||
|
if self.address_tokens:
|
||||||
|
out['addr'] = self.address_tokens
|
||||||
|
|
||||||
|
return out
|
||||||
|
|
||||||
|
|
||||||
|
def set_names(self, fulls, partials):
|
||||||
""" Adds token information for the normalised names.
|
""" Adds token information for the normalised names.
|
||||||
"""
|
"""
|
||||||
self.data['names'] = self._mk_array(itertools.chain(fulls, partials))
|
self.names = self._mk_array(itertools.chain(fulls, partials))
|
||||||
|
|
||||||
|
|
||||||
def add_housenumbers(self, tokens, hnrs):
|
def add_housenumber(self, token, hnr):
|
||||||
""" Extract housenumber information from a list of normalised
|
""" Extract housenumber information from a list of normalised
|
||||||
housenumbers.
|
housenumbers.
|
||||||
"""
|
"""
|
||||||
self.data['hnr_tokens'] = self._mk_array(tokens)
|
if token:
|
||||||
self.data['hnr'] = ';'.join(hnrs)
|
self.housenumbers.add(hnr)
|
||||||
|
self.housenumber_tokens.add(token)
|
||||||
|
|
||||||
|
|
||||||
def add_street(self, tokens):
|
def add_street(self, tokens):
|
||||||
""" Add addr:street match terms.
|
""" Add addr:street match terms.
|
||||||
"""
|
"""
|
||||||
self.data['street'] = self._mk_array(tokens)
|
self.street_tokens.update(tokens)
|
||||||
|
|
||||||
|
|
||||||
def add_place(self, tokens):
|
def add_place(self, tokens):
|
||||||
""" Add addr:place search and match terms.
|
""" Add addr:place search and match terms.
|
||||||
"""
|
"""
|
||||||
if tokens:
|
self.place_tokens.update(tokens)
|
||||||
self.data['place'] = self._mk_array(tokens)
|
|
||||||
|
|
||||||
|
|
||||||
def add_address_terms(self, terms):
|
def add_address_term(self, key, partials):
|
||||||
""" Add additional address terms.
|
""" Add additional address terms.
|
||||||
"""
|
"""
|
||||||
tokens = {key: self._mk_array(partials)
|
if partials:
|
||||||
for key, partials in terms if partials}
|
self.address_tokens[key] = self._mk_array(partials)
|
||||||
|
|
||||||
if tokens:
|
|
||||||
self.data['addr'] = tokens
|
|
||||||
|
|
||||||
|
|
||||||
class _TokenCache:
|
class _TokenCache:
|
||||||
|
|||||||
Reference in New Issue
Block a user