mirror of
https://github.com/osm-search/Nominatim.git
synced 2026-02-26 11:08:13 +00:00
more formatting fixes
Found by flake8.
This commit is contained in:
@@ -24,6 +24,7 @@ def _check(hint=None):
|
||||
"""
|
||||
def decorator(func):
|
||||
title = func.__doc__.split('\n', 1)[0].strip()
|
||||
|
||||
def run_check(conn, config):
|
||||
print(title, end=' ... ')
|
||||
ret = func(conn, config)
|
||||
@@ -98,13 +99,12 @@ def _get_indexes(conn):
|
||||
if conn.table_exists('place'):
|
||||
indexes.extend(('idx_placex_pendingsector',
|
||||
'idx_location_area_country_place_id',
|
||||
'idx_place_osm_unique'
|
||||
))
|
||||
'idx_place_osm_unique'))
|
||||
|
||||
return indexes
|
||||
|
||||
|
||||
### CHECK FUNCTIONS
|
||||
# CHECK FUNCTIONS
|
||||
#
|
||||
# Functions are exectured in the order they appear here.
|
||||
|
||||
|
||||
@@ -184,8 +184,10 @@ def truncate_data_tables(conn):
|
||||
|
||||
conn.commit()
|
||||
|
||||
|
||||
_COPY_COLUMNS = 'osm_type, osm_id, class, type, name, admin_level, address, extratags, geometry'
|
||||
|
||||
|
||||
def load_data(dsn, threads):
|
||||
""" Copy data into the word and placex table.
|
||||
"""
|
||||
@@ -250,6 +252,7 @@ def create_search_indices(conn, config, drop=False):
|
||||
|
||||
sql.run_sql_file(conn, 'indices.sql', drop=drop)
|
||||
|
||||
|
||||
def create_country_names(conn, tokenizer, languages=None):
|
||||
""" Add default country names to search index. `languages` is a comma-
|
||||
separated list of language codes as used in OSM. If `languages` is not
|
||||
@@ -261,8 +264,7 @@ def create_country_names(conn, tokenizer, languages=None):
|
||||
|
||||
def _include_key(key):
|
||||
return key == 'name' or \
|
||||
(key.startswith('name:') \
|
||||
and (not languages or key[5:] in languages))
|
||||
(key.startswith('name:') and (not languages or key[5:] in languages))
|
||||
|
||||
with conn.cursor() as cur:
|
||||
psycopg2.extras.register_hstore(cur)
|
||||
@@ -271,7 +273,7 @@ def create_country_names(conn, tokenizer, languages=None):
|
||||
|
||||
with tokenizer.name_analyzer() as analyzer:
|
||||
for code, name in cur:
|
||||
names = {'countrycode' : code}
|
||||
names = {'countrycode': code}
|
||||
if code == 'gb':
|
||||
names['short_name'] = 'UK'
|
||||
if code == 'us':
|
||||
|
||||
@@ -136,7 +136,7 @@ def run_osm2pgsql(options):
|
||||
def get_url(url):
|
||||
""" Get the contents from the given URL and return it as a UTF-8 string.
|
||||
"""
|
||||
headers = {"User-Agent" : "Nominatim/{0[0]}.{0[1]}.{0[2]}-{0[3]}".format(NOMINATIM_VERSION)}
|
||||
headers = {"User-Agent": "Nominatim/{0[0]}.{0[1]}.{0[2]}-{0[3]}".format(NOMINATIM_VERSION)}
|
||||
|
||||
try:
|
||||
with urlrequest.urlopen(urlrequest.Request(url, headers=headers)) as response:
|
||||
|
||||
@@ -142,7 +142,8 @@ def change_housenumber_transliteration(conn, **_):
|
||||
BEGIN
|
||||
SELECT array_to_string(array_agg(trans), ';')
|
||||
INTO normtext
|
||||
FROM (SELECT lookup_word as trans, getorcreate_housenumber_id(lookup_word)
|
||||
FROM (SELECT lookup_word as trans,
|
||||
getorcreate_housenumber_id(lookup_word)
|
||||
FROM (SELECT make_standard_name(h) as lookup_word
|
||||
FROM regexp_split_to_table(housenumber, '[,;]') h) x) y;
|
||||
return normtext;
|
||||
|
||||
@@ -165,11 +165,14 @@ def update_postcodes(dsn, project_dir, tokenizer):
|
||||
with conn.cursor(name="placex_postcodes") as cur:
|
||||
cur.execute("""
|
||||
SELECT cc as country_code, pc, ST_X(centroid), ST_Y(centroid)
|
||||
FROM (SELECT
|
||||
COALESCE(plx.country_code, get_country_code(ST_Centroid(pl.geometry))) as cc,
|
||||
FROM (SELECT
|
||||
COALESCE(plx.country_code,
|
||||
get_country_code(ST_Centroid(pl.geometry))) as cc,
|
||||
token_normalized_postcode(pl.address->'postcode') as pc,
|
||||
ST_Centroid(ST_Collect(COALESCE(plx.centroid, ST_Centroid(pl.geometry)))) as centroid
|
||||
FROM place AS pl LEFT OUTER JOIN placex AS plx ON pl.osm_id = plx.osm_id AND pl.osm_type = plx.osm_type
|
||||
ST_Centroid(ST_Collect(COALESCE(plx.centroid,
|
||||
ST_Centroid(pl.geometry)))) as centroid
|
||||
FROM place AS pl LEFT OUTER JOIN placex AS plx
|
||||
ON pl.osm_id = plx.osm_id AND pl.osm_type = plx.osm_type
|
||||
WHERE pl.address ? 'postcode' AND pl.geometry IS NOT null
|
||||
GROUP BY cc, pc) xx
|
||||
WHERE pc IS NOT null AND cc IS NOT null
|
||||
|
||||
@@ -44,8 +44,8 @@ class SPImporter():
|
||||
# This set will contain all existing phrases to be added.
|
||||
# It contains tuples with the following format: (lable, class, type, operator)
|
||||
self.word_phrases = set()
|
||||
#This set will contain all existing place_classtype tables which doesn't match any
|
||||
#special phrases class/type on the wiki.
|
||||
# This set will contain all existing place_classtype tables which doesn't match any
|
||||
# special phrases class/type on the wiki.
|
||||
self.table_phrases_to_delete = set()
|
||||
|
||||
def import_phrases(self, tokenizer, should_replace):
|
||||
@@ -60,7 +60,7 @@ class SPImporter():
|
||||
LOG.warning('Special phrases importation starting')
|
||||
self._fetch_existing_place_classtype_tables()
|
||||
|
||||
#Store pairs of class/type for further processing
|
||||
# Store pairs of class/type for further processing
|
||||
class_type_pairs = set()
|
||||
|
||||
for loaded_phrases in self.sp_loader:
|
||||
@@ -131,17 +131,17 @@ class SPImporter():
|
||||
Return the class/type pair corresponding to the phrase.
|
||||
"""
|
||||
|
||||
#blacklisting: disallow certain class/type combinations
|
||||
# blacklisting: disallow certain class/type combinations
|
||||
if phrase.p_class in self.black_list.keys() \
|
||||
and phrase.p_type in self.black_list[phrase.p_class]:
|
||||
return None
|
||||
|
||||
#whitelisting: if class is in whitelist, allow only tags in the list
|
||||
# whitelisting: if class is in whitelist, allow only tags in the list
|
||||
if phrase.p_class in self.white_list.keys() \
|
||||
and phrase.p_type not in self.white_list[phrase.p_class]:
|
||||
return None
|
||||
|
||||
#sanity check, in case somebody added garbage in the wiki
|
||||
# sanity check, in case somebody added garbage in the wiki
|
||||
if not self._check_sanity(phrase):
|
||||
self.statistics_handler.notify_one_phrase_invalid()
|
||||
return None
|
||||
@@ -161,7 +161,7 @@ class SPImporter():
|
||||
|
||||
sql_tablespace = self.config.TABLESPACE_AUX_DATA
|
||||
if sql_tablespace:
|
||||
sql_tablespace = ' TABLESPACE '+sql_tablespace
|
||||
sql_tablespace = ' TABLESPACE ' + sql_tablespace
|
||||
|
||||
with self.db_connection.cursor() as db_cursor:
|
||||
db_cursor.execute("CREATE INDEX idx_placex_classtype ON placex (class, type)")
|
||||
@@ -174,19 +174,19 @@ class SPImporter():
|
||||
|
||||
if table_name in self.table_phrases_to_delete:
|
||||
self.statistics_handler.notify_one_table_ignored()
|
||||
#Remove this table from the ones to delete as it match a class/type
|
||||
#still existing on the special phrases of the wiki.
|
||||
# Remove this table from the ones to delete as it match a
|
||||
# class/type still existing on the special phrases of the wiki.
|
||||
self.table_phrases_to_delete.remove(table_name)
|
||||
#So dont need to create the table and indexes.
|
||||
# So don't need to create the table and indexes.
|
||||
continue
|
||||
|
||||
#Table creation
|
||||
# Table creation
|
||||
self._create_place_classtype_table(sql_tablespace, phrase_class, phrase_type)
|
||||
|
||||
#Indexes creation
|
||||
# Indexes creation
|
||||
self._create_place_classtype_indexes(sql_tablespace, phrase_class, phrase_type)
|
||||
|
||||
#Grant access on read to the web user.
|
||||
# Grant access on read to the web user.
|
||||
self._grant_access_to_webuser(phrase_class, phrase_type)
|
||||
|
||||
self.statistics_handler.notify_one_table_created()
|
||||
@@ -202,8 +202,8 @@ class SPImporter():
|
||||
table_name = _classtype_table(phrase_class, phrase_type)
|
||||
with self.db_connection.cursor() as db_cursor:
|
||||
db_cursor.execute(SQL("""
|
||||
CREATE TABLE IF NOT EXISTS {{}} {}
|
||||
AS SELECT place_id AS place_id,st_centroid(geometry) AS centroid FROM placex
|
||||
CREATE TABLE IF NOT EXISTS {{}} {}
|
||||
AS SELECT place_id AS place_id,st_centroid(geometry) AS centroid FROM placex
|
||||
WHERE class = {{}} AND type = {{}}""".format(sql_tablespace))
|
||||
.format(Identifier(table_name), Literal(phrase_class),
|
||||
Literal(phrase_type)))
|
||||
@@ -215,7 +215,7 @@ class SPImporter():
|
||||
"""
|
||||
index_prefix = 'idx_place_classtype_{}_{}_'.format(phrase_class, phrase_type)
|
||||
base_table = _classtype_table(phrase_class, phrase_type)
|
||||
#Index on centroid
|
||||
# Index on centroid
|
||||
if not self.db_connection.index_exists(index_prefix + 'centroid'):
|
||||
with self.db_connection.cursor() as db_cursor:
|
||||
db_cursor.execute(SQL("""
|
||||
@@ -223,7 +223,7 @@ class SPImporter():
|
||||
.format(Identifier(index_prefix + 'centroid'),
|
||||
Identifier(base_table)), sql_tablespace)
|
||||
|
||||
#Index on place_id
|
||||
# Index on place_id
|
||||
if not self.db_connection.index_exists(index_prefix + 'place_id'):
|
||||
with self.db_connection.cursor() as db_cursor:
|
||||
db_cursor.execute(SQL(
|
||||
@@ -248,10 +248,12 @@ class SPImporter():
|
||||
Delete the place_classtype tables.
|
||||
"""
|
||||
LOG.warning('Cleaning database...')
|
||||
#Array containing all queries to execute. Contain tuples of format (query, parameters)
|
||||
# Array containing all queries to execute.
|
||||
# Contains tuples of format (query, parameters)
|
||||
queries_parameters = []
|
||||
|
||||
#Delete place_classtype tables corresponding to class/type which are not on the wiki anymore
|
||||
# Delete place_classtype tables corresponding to class/type which
|
||||
# are not on the wiki anymore.
|
||||
for table in self.table_phrases_to_delete:
|
||||
self.statistics_handler.notify_one_table_deleted()
|
||||
query = SQL('DROP TABLE IF EXISTS {}').format(Identifier(table))
|
||||
@@ -271,7 +273,7 @@ class SPImporter():
|
||||
file, extension = os.path.splitext(file_path)
|
||||
json_file_path = Path(file + '.json').resolve()
|
||||
|
||||
if extension not in('.php', '.json'):
|
||||
if extension not in ('.php', '.json'):
|
||||
raise UsageError('The custom NOMINATIM_PHRASE_CONFIG file has not a valid extension.')
|
||||
|
||||
if extension == '.php' and not isfile(json_file_path):
|
||||
|
||||
@@ -15,7 +15,7 @@ class SPWikiLoader(Iterator):
|
||||
def __init__(self, config, languages=None):
|
||||
super().__init__()
|
||||
self.config = config
|
||||
#Compile the regex here to increase performances.
|
||||
# Compile the regex here to increase performances.
|
||||
self.occurence_pattern = re.compile(
|
||||
r'\| *([^\|]+) *\|\| *([^\|]+) *\|\| *([^\|]+) *\|\| *([^\|]+) *\|\| *([\-YN])'
|
||||
)
|
||||
@@ -35,7 +35,7 @@ class SPWikiLoader(Iterator):
|
||||
Parses XML content and extracts special phrases from it.
|
||||
Return a list of SpecialPhrase.
|
||||
"""
|
||||
#One match will be of format [label, class, type, operator, plural]
|
||||
# One match will be of format [label, class, type, operator, plural]
|
||||
matches = self.occurence_pattern.findall(xml)
|
||||
returned_phrases = set()
|
||||
for match in matches:
|
||||
@@ -65,5 +65,6 @@ class SPWikiLoader(Iterator):
|
||||
Requested URL Example :
|
||||
https://wiki.openstreetmap.org/wiki/Special:Export/Nominatim/Special_Phrases/EN
|
||||
"""
|
||||
url = 'https://wiki.openstreetmap.org/wiki/Special:Export/Nominatim/Special_Phrases/' + lang.upper() # pylint: disable=line-too-long
|
||||
url = 'https://wiki.openstreetmap.org/wiki/Special:Export/Nominatim/Special_Phrases/' \
|
||||
+ lang.upper()
|
||||
return get_url(url)
|
||||
|
||||
@@ -13,7 +13,7 @@ class SpecialPhrase():
|
||||
def __init__(self, p_label, p_class, p_type, p_operator):
|
||||
self.p_label = p_label.strip()
|
||||
self.p_class = p_class.strip()
|
||||
#Hack around a bug where building=yes was imported with quotes into the wiki
|
||||
# Hack around a bug where building=yes was imported with quotes into the wiki
|
||||
self.p_type = re.sub(r'\"|"', '', p_type.strip())
|
||||
#Needed if some operator in the wiki are not written in english
|
||||
# Needed if some operator in the wiki are not written in english
|
||||
self.p_operator = '-' if p_operator not in ('near', 'in') else p_operator
|
||||
|
||||
Reference in New Issue
Block a user