forked from hans/Nominatim
Merge pull request #2129 from lonvia/cleanup-bdd-tests
Clean up Python support code for BDD tests
This commit is contained in:
@@ -1,16 +1,11 @@
|
|||||||
from behave import *
|
from behave import *
|
||||||
import logging
|
from pathlib import Path
|
||||||
import os
|
|
||||||
import psycopg2
|
|
||||||
import psycopg2.extras
|
|
||||||
import subprocess
|
|
||||||
import tempfile
|
|
||||||
from sys import version_info as python_version
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
from steps.geometry_factory import GeometryFactory
|
||||||
|
from steps.nominatim_environment import NominatimEnvironment
|
||||||
|
|
||||||
userconfig = {
|
userconfig = {
|
||||||
'BUILDDIR' : os.path.join(os.path.split(__file__)[0], "../../build"),
|
'BUILDDIR' : (Path(__file__) / '..' / '..' / '..' / 'build').resolve(),
|
||||||
'REMOVE_TEMPLATE' : False,
|
'REMOVE_TEMPLATE' : False,
|
||||||
'KEEP_TEST_DB' : False,
|
'KEEP_TEST_DB' : False,
|
||||||
'DB_HOST' : None,
|
'DB_HOST' : None,
|
||||||
@@ -26,290 +21,24 @@ userconfig = {
|
|||||||
|
|
||||||
use_step_matcher("re")
|
use_step_matcher("re")
|
||||||
|
|
||||||
|
|
||||||
class NominatimEnvironment(object):
|
|
||||||
""" Collects all functions for the execution of Nominatim functions.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(self, config):
|
|
||||||
self.build_dir = os.path.abspath(config['BUILDDIR'])
|
|
||||||
self.src_dir = os.path.abspath(os.path.join(os.path.split(__file__)[0], "../.."))
|
|
||||||
self.db_host = config['DB_HOST']
|
|
||||||
self.db_port = config['DB_PORT']
|
|
||||||
self.db_user = config['DB_USER']
|
|
||||||
self.db_pass = config['DB_PASS']
|
|
||||||
self.template_db = config['TEMPLATE_DB']
|
|
||||||
self.test_db = config['TEST_DB']
|
|
||||||
self.api_test_db = config['API_TEST_DB']
|
|
||||||
self.server_module_path = config['SERVER_MODULE_PATH']
|
|
||||||
self.reuse_template = not config['REMOVE_TEMPLATE']
|
|
||||||
self.keep_scenario_db = config['KEEP_TEST_DB']
|
|
||||||
self.code_coverage_path = config['PHPCOV']
|
|
||||||
self.code_coverage_id = 1
|
|
||||||
self.test_env = None
|
|
||||||
|
|
||||||
self.template_db_done = False
|
|
||||||
self.website_dir = None
|
|
||||||
|
|
||||||
def connect_database(self, dbname):
|
|
||||||
dbargs = {'database': dbname}
|
|
||||||
if self.db_host:
|
|
||||||
dbargs['host'] = self.db_host
|
|
||||||
if self.db_port:
|
|
||||||
dbargs['port'] = self.db_port
|
|
||||||
if self.db_user:
|
|
||||||
dbargs['user'] = self.db_user
|
|
||||||
if self.db_pass:
|
|
||||||
dbargs['password'] = self.db_pass
|
|
||||||
conn = psycopg2.connect(**dbargs)
|
|
||||||
return conn
|
|
||||||
|
|
||||||
def next_code_coverage_file(self):
|
|
||||||
fn = os.path.join(self.code_coverage_path, "%06d.cov" % self.code_coverage_id)
|
|
||||||
self.code_coverage_id += 1
|
|
||||||
|
|
||||||
return fn
|
|
||||||
|
|
||||||
def write_nominatim_config(self, dbname):
|
|
||||||
dsn = 'pgsql:dbname={}{}{}{}{}'.format(
|
|
||||||
dbname,
|
|
||||||
(';host=' + self.db_host) if self.db_host else '',
|
|
||||||
(';port=' + self.db_port) if self.db_port else '',
|
|
||||||
(';user=' + self.db_user) if self.db_user else '',
|
|
||||||
(';password=' + self.db_pass) if self.db_pass else ''
|
|
||||||
)
|
|
||||||
|
|
||||||
if self.website_dir is not None \
|
|
||||||
and self.test_env is not None \
|
|
||||||
and dsn == self.test_env['NOMINATIM_DATABASE_DSN']:
|
|
||||||
return # environment already set uo
|
|
||||||
|
|
||||||
self.test_env = os.environ
|
|
||||||
self.test_env['NOMINATIM_DATABASE_DSN'] = dsn
|
|
||||||
self.test_env['NOMINATIM_FLATNODE_FILE'] = ''
|
|
||||||
self.test_env['NOMINATIM_IMPORT_STYLE'] = 'full'
|
|
||||||
self.test_env['NOMINATIM_USE_US_TIGER_DATA'] = 'yes'
|
|
||||||
|
|
||||||
if self.website_dir is not None:
|
|
||||||
self.website_dir.cleanup()
|
|
||||||
|
|
||||||
self.website_dir = tempfile.TemporaryDirectory()
|
|
||||||
self.run_setup_script('setup-website')
|
|
||||||
|
|
||||||
|
|
||||||
def db_drop_database(self, name):
|
|
||||||
conn = self.connect_database('postgres')
|
|
||||||
conn.set_isolation_level(0)
|
|
||||||
cur = conn.cursor()
|
|
||||||
cur.execute('DROP DATABASE IF EXISTS %s' % (name, ))
|
|
||||||
conn.close()
|
|
||||||
|
|
||||||
def setup_template_db(self):
|
|
||||||
if self.template_db_done:
|
|
||||||
return
|
|
||||||
|
|
||||||
self.template_db_done = True
|
|
||||||
|
|
||||||
if self.reuse_template:
|
|
||||||
# check that the template is there
|
|
||||||
conn = self.connect_database('postgres')
|
|
||||||
cur = conn.cursor()
|
|
||||||
cur.execute('select count(*) from pg_database where datname = %s',
|
|
||||||
(self.template_db,))
|
|
||||||
if cur.fetchone()[0] == 1:
|
|
||||||
return
|
|
||||||
conn.close()
|
|
||||||
else:
|
|
||||||
# just in case... make sure a previous table has been dropped
|
|
||||||
self.db_drop_database(self.template_db)
|
|
||||||
|
|
||||||
try:
|
|
||||||
# call the first part of database setup
|
|
||||||
self.write_nominatim_config(self.template_db)
|
|
||||||
self.run_setup_script('create-db', 'setup-db')
|
|
||||||
# remove external data to speed up indexing for tests
|
|
||||||
conn = self.connect_database(self.template_db)
|
|
||||||
cur = conn.cursor()
|
|
||||||
cur.execute("""select tablename from pg_tables
|
|
||||||
where tablename in ('gb_postcode', 'us_postcode')""")
|
|
||||||
for t in cur:
|
|
||||||
conn.cursor().execute('TRUNCATE TABLE %s' % (t[0],))
|
|
||||||
conn.commit()
|
|
||||||
conn.close()
|
|
||||||
|
|
||||||
# execute osm2pgsql import on an empty file to get the right tables
|
|
||||||
with tempfile.NamedTemporaryFile(dir='/tmp', suffix='.xml') as fd:
|
|
||||||
fd.write(b'<osm version="0.6"></osm>')
|
|
||||||
fd.flush()
|
|
||||||
self.run_setup_script('import-data',
|
|
||||||
'ignore-errors',
|
|
||||||
'create-functions',
|
|
||||||
'create-tables',
|
|
||||||
'create-partition-tables',
|
|
||||||
'create-partition-functions',
|
|
||||||
'load-data',
|
|
||||||
'create-search-indices',
|
|
||||||
osm_file=fd.name,
|
|
||||||
osm2pgsql_cache='200')
|
|
||||||
except:
|
|
||||||
self.db_drop_database(self.template_db)
|
|
||||||
raise
|
|
||||||
|
|
||||||
|
|
||||||
def setup_api_db(self, context):
|
|
||||||
self.write_nominatim_config(self.api_test_db)
|
|
||||||
|
|
||||||
def setup_unknown_db(self, context):
|
|
||||||
self.write_nominatim_config('UNKNOWN_DATABASE_NAME')
|
|
||||||
|
|
||||||
def setup_db(self, context):
|
|
||||||
self.setup_template_db()
|
|
||||||
self.write_nominatim_config(self.test_db)
|
|
||||||
conn = self.connect_database(self.template_db)
|
|
||||||
conn.set_isolation_level(0)
|
|
||||||
cur = conn.cursor()
|
|
||||||
cur.execute('DROP DATABASE IF EXISTS %s' % (self.test_db, ))
|
|
||||||
cur.execute('CREATE DATABASE %s TEMPLATE = %s' % (self.test_db, self.template_db))
|
|
||||||
conn.close()
|
|
||||||
context.db = self.connect_database(self.test_db)
|
|
||||||
if python_version[0] < 3:
|
|
||||||
psycopg2.extras.register_hstore(context.db, globally=False, unicode=True)
|
|
||||||
else:
|
|
||||||
psycopg2.extras.register_hstore(context.db, globally=False)
|
|
||||||
|
|
||||||
def teardown_db(self, context):
|
|
||||||
if 'db' in context:
|
|
||||||
context.db.close()
|
|
||||||
|
|
||||||
if not self.keep_scenario_db:
|
|
||||||
self.db_drop_database(self.test_db)
|
|
||||||
|
|
||||||
def run_setup_script(self, *args, **kwargs):
|
|
||||||
if self.server_module_path:
|
|
||||||
kwargs = dict(kwargs)
|
|
||||||
kwargs['module_path'] = self.server_module_path
|
|
||||||
self.run_nominatim_script('setup', *args, **kwargs)
|
|
||||||
|
|
||||||
def run_update_script(self, *args, **kwargs):
|
|
||||||
self.run_nominatim_script('update', *args, **kwargs)
|
|
||||||
|
|
||||||
def run_nominatim_script(self, script, *args, **kwargs):
|
|
||||||
cmd = ['/usr/bin/env', 'php', '-Cq']
|
|
||||||
cmd.append(os.path.join(self.build_dir, 'utils', '%s.php' % script))
|
|
||||||
cmd.extend(['--%s' % x for x in args])
|
|
||||||
for k, v in kwargs.items():
|
|
||||||
cmd.extend(('--' + k.replace('_', '-'), str(v)))
|
|
||||||
|
|
||||||
if self.website_dir is not None:
|
|
||||||
cwd = self.website_dir.name
|
|
||||||
else:
|
|
||||||
cwd = self.build_dir
|
|
||||||
|
|
||||||
proc = subprocess.Popen(cmd, cwd=cwd, env=self.test_env,
|
|
||||||
stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
||||||
(outp, outerr) = proc.communicate()
|
|
||||||
outerr = outerr.decode('utf-8').replace('\\n', '\n')
|
|
||||||
logger.debug("run_nominatim_script: %s\n%s\n%s" % (cmd, outp, outerr))
|
|
||||||
assert (proc.returncode == 0), "Script '%s' failed:\n%s\n%s\n" % (script, outp, outerr)
|
|
||||||
|
|
||||||
|
|
||||||
class OSMDataFactory(object):
|
|
||||||
|
|
||||||
def __init__(self):
|
|
||||||
scriptpath = os.path.dirname(os.path.abspath(__file__))
|
|
||||||
self.scene_path = os.environ.get('SCENE_PATH',
|
|
||||||
os.path.join(scriptpath, '..', 'scenes', 'data'))
|
|
||||||
self.scene_cache = {}
|
|
||||||
self.clear_grid()
|
|
||||||
|
|
||||||
def parse_geometry(self, geom, scene):
|
|
||||||
if geom.find(':') >= 0:
|
|
||||||
return "ST_SetSRID(%s, 4326)" % self.get_scene_geometry(scene, geom)
|
|
||||||
|
|
||||||
if geom.find(',') < 0:
|
|
||||||
out = "POINT(%s)" % self.mk_wkt_point(geom)
|
|
||||||
elif geom.find('(') < 0:
|
|
||||||
line = ','.join([self.mk_wkt_point(x) for x in geom.split(',')])
|
|
||||||
out = "LINESTRING(%s)" % line
|
|
||||||
else:
|
|
||||||
inner = geom.strip('() ')
|
|
||||||
line = ','.join([self.mk_wkt_point(x) for x in inner.split(',')])
|
|
||||||
out = "POLYGON((%s))" % line
|
|
||||||
|
|
||||||
return "ST_SetSRID('%s'::geometry, 4326)" % out
|
|
||||||
|
|
||||||
def mk_wkt_point(self, point):
|
|
||||||
geom = point.strip()
|
|
||||||
if geom.find(' ') >= 0:
|
|
||||||
return geom
|
|
||||||
else:
|
|
||||||
pt = self.grid_node(int(geom))
|
|
||||||
assert pt is not None, "Bad scenario: Point '{}' not found in grid".format(geom)
|
|
||||||
return "%f %f" % pt
|
|
||||||
|
|
||||||
def get_scene_geometry(self, default_scene, name):
|
|
||||||
geoms = []
|
|
||||||
for obj in name.split('+'):
|
|
||||||
oname = obj.strip()
|
|
||||||
if oname.startswith(':'):
|
|
||||||
assert default_scene is not None, "Bad scenario: You need to set a scene"
|
|
||||||
defscene = self.load_scene(default_scene)
|
|
||||||
wkt = defscene[oname[1:]]
|
|
||||||
else:
|
|
||||||
scene, obj = oname.split(':', 2)
|
|
||||||
scene_geoms = self.load_scene(scene)
|
|
||||||
wkt = scene_geoms[obj]
|
|
||||||
|
|
||||||
geoms.append("'%s'::geometry" % wkt)
|
|
||||||
|
|
||||||
if len(geoms) == 1:
|
|
||||||
return geoms[0]
|
|
||||||
else:
|
|
||||||
return 'ST_LineMerge(ST_Collect(ARRAY[%s]))' % ','.join(geoms)
|
|
||||||
|
|
||||||
def load_scene(self, name):
|
|
||||||
if name in self.scene_cache:
|
|
||||||
return self.scene_cache[name]
|
|
||||||
|
|
||||||
scene = {}
|
|
||||||
with open(os.path.join(self.scene_path, "%s.wkt" % name), 'r') as fd:
|
|
||||||
for line in fd:
|
|
||||||
if line.strip():
|
|
||||||
obj, wkt = line.split('|', 2)
|
|
||||||
scene[obj.strip()] = wkt.strip()
|
|
||||||
self.scene_cache[name] = scene
|
|
||||||
|
|
||||||
return scene
|
|
||||||
|
|
||||||
def clear_grid(self):
|
|
||||||
self.grid = {}
|
|
||||||
|
|
||||||
def add_grid_node(self, nodeid, x, y):
|
|
||||||
self.grid[nodeid] = (x, y)
|
|
||||||
|
|
||||||
def grid_node(self, nodeid):
|
|
||||||
return self.grid.get(nodeid)
|
|
||||||
|
|
||||||
|
|
||||||
def before_all(context):
|
def before_all(context):
|
||||||
# logging setup
|
# logging setup
|
||||||
context.config.setup_logging()
|
context.config.setup_logging()
|
||||||
# set up -D options
|
# set up -D options
|
||||||
for k,v in userconfig.items():
|
for k,v in userconfig.items():
|
||||||
context.config.userdata.setdefault(k, v)
|
context.config.userdata.setdefault(k, v)
|
||||||
logging.debug('User config: %s' %(str(context.config.userdata)))
|
|
||||||
# Nominatim test setup
|
# Nominatim test setup
|
||||||
context.nominatim = NominatimEnvironment(context.config.userdata)
|
context.nominatim = NominatimEnvironment(context.config.userdata)
|
||||||
context.osm = OSMDataFactory()
|
context.osm = GeometryFactory()
|
||||||
|
|
||||||
|
|
||||||
def before_scenario(context, scenario):
|
def before_scenario(context, scenario):
|
||||||
if 'DB' in context.tags:
|
if 'DB' in context.tags:
|
||||||
context.nominatim.setup_db(context)
|
context.nominatim.setup_db(context)
|
||||||
elif 'APIDB' in context.tags:
|
elif 'APIDB' in context.tags:
|
||||||
context.nominatim.setup_api_db(context)
|
context.nominatim.setup_api_db()
|
||||||
elif 'UNKNOWNDB' in context.tags:
|
elif 'UNKNOWNDB' in context.tags:
|
||||||
context.nominatim.setup_unknown_db(context)
|
context.nominatim.setup_unknown_db()
|
||||||
context.scene = None
|
context.scene = None
|
||||||
|
|
||||||
def after_scenario(context, scenario):
|
def after_scenario(context, scenario):
|
||||||
|
|||||||
@@ -1,623 +0,0 @@
|
|||||||
import base64
|
|
||||||
import random
|
|
||||||
import string
|
|
||||||
import re
|
|
||||||
import psycopg2.extras
|
|
||||||
|
|
||||||
from check_functions import Almost
|
|
||||||
|
|
||||||
class PlaceColumn:
|
|
||||||
|
|
||||||
def __init__(self, context, force_name):
|
|
||||||
self.columns = { 'admin_level' : 15}
|
|
||||||
self.force_name = force_name
|
|
||||||
self.context = context
|
|
||||||
self.geometry = None
|
|
||||||
|
|
||||||
def add(self, key, value):
|
|
||||||
if hasattr(self, 'set_key_' + key):
|
|
||||||
getattr(self, 'set_key_' + key)(value)
|
|
||||||
elif key.startswith('name+'):
|
|
||||||
self.add_hstore('name', key[5:], value)
|
|
||||||
elif key.startswith('extra+'):
|
|
||||||
self.add_hstore('extratags', key[6:], value)
|
|
||||||
elif key.startswith('addr+'):
|
|
||||||
self.add_hstore('address', key[5:], value)
|
|
||||||
elif key in ('name', 'address', 'extratags'):
|
|
||||||
self.columns[key] = eval('{' + value + '}')
|
|
||||||
else:
|
|
||||||
assert key in ('class', 'type')
|
|
||||||
self.columns[key] = None if value == '' else value
|
|
||||||
|
|
||||||
def set_key_name(self, value):
|
|
||||||
self.add_hstore('name', 'name', value)
|
|
||||||
|
|
||||||
def set_key_osm(self, value):
|
|
||||||
assert value[0] in 'NRW'
|
|
||||||
assert value[1:].isdigit()
|
|
||||||
|
|
||||||
self.columns['osm_type'] = value[0]
|
|
||||||
self.columns['osm_id'] = int(value[1:])
|
|
||||||
|
|
||||||
def set_key_admin(self, value):
|
|
||||||
self.columns['admin_level'] = int(value)
|
|
||||||
|
|
||||||
def set_key_housenr(self, value):
|
|
||||||
if value:
|
|
||||||
self.add_hstore('address', 'housenumber', value)
|
|
||||||
|
|
||||||
def set_key_postcode(self, value):
|
|
||||||
if value:
|
|
||||||
self.add_hstore('address', 'postcode', value)
|
|
||||||
|
|
||||||
def set_key_street(self, value):
|
|
||||||
if value:
|
|
||||||
self.add_hstore('address', 'street', value)
|
|
||||||
|
|
||||||
def set_key_addr_place(self, value):
|
|
||||||
if value:
|
|
||||||
self.add_hstore('address', 'place', value)
|
|
||||||
|
|
||||||
def set_key_country(self, value):
|
|
||||||
if value:
|
|
||||||
self.add_hstore('address', 'country', value)
|
|
||||||
|
|
||||||
def set_key_geometry(self, value):
|
|
||||||
self.geometry = self.context.osm.parse_geometry(value, self.context.scene)
|
|
||||||
assert self.geometry is not None
|
|
||||||
|
|
||||||
def add_hstore(self, column, key, value):
|
|
||||||
if column in self.columns:
|
|
||||||
self.columns[column][key] = value
|
|
||||||
else:
|
|
||||||
self.columns[column] = { key : value }
|
|
||||||
|
|
||||||
def db_insert(self, cursor):
|
|
||||||
assert 'osm_type' in self.columns
|
|
||||||
if self.force_name and 'name' not in self.columns:
|
|
||||||
self.add_hstore('name', 'name', ''.join(random.choice(string.printable)
|
|
||||||
for _ in range(int(random.random()*30))))
|
|
||||||
|
|
||||||
if self.columns['osm_type'] == 'N' and self.geometry is None:
|
|
||||||
pt = self.context.osm.grid_node(self.columns['osm_id'])
|
|
||||||
if pt is None:
|
|
||||||
pt = (random.random()*360 - 180, random.random()*180 - 90)
|
|
||||||
|
|
||||||
self.geometry = "ST_SetSRID(ST_Point(%f, %f), 4326)" % pt
|
|
||||||
else:
|
|
||||||
assert self.geometry is not None, "Geometry missing"
|
|
||||||
query = 'INSERT INTO place (%s, geometry) values(%s, %s)' % (
|
|
||||||
','.join(self.columns.keys()),
|
|
||||||
','.join(['%s' for x in range(len(self.columns))]),
|
|
||||||
self.geometry)
|
|
||||||
cursor.execute(query, list(self.columns.values()))
|
|
||||||
|
|
||||||
class LazyFmt(object):
|
|
||||||
|
|
||||||
def __init__(self, fmtstr, *args):
|
|
||||||
self.fmt = fmtstr
|
|
||||||
self.args = args
|
|
||||||
|
|
||||||
def __str__(self):
|
|
||||||
return self.fmt % self.args
|
|
||||||
|
|
||||||
class PlaceObjName(object):
|
|
||||||
|
|
||||||
def __init__(self, placeid, conn):
|
|
||||||
self.pid = placeid
|
|
||||||
self.conn = conn
|
|
||||||
|
|
||||||
def __str__(self):
|
|
||||||
if self.pid is None:
|
|
||||||
return "<null>"
|
|
||||||
|
|
||||||
if self.pid == 0:
|
|
||||||
return "place ID 0"
|
|
||||||
|
|
||||||
cur = self.conn.cursor()
|
|
||||||
cur.execute("""SELECT osm_type, osm_id, class
|
|
||||||
FROM placex WHERE place_id = %s""",
|
|
||||||
(self.pid, ))
|
|
||||||
assert cur.rowcount == 1, "No entry found for place id %s" % self.pid
|
|
||||||
|
|
||||||
return "%s%s:%s" % cur.fetchone()
|
|
||||||
|
|
||||||
def compare_place_id(expected, result, column, context):
|
|
||||||
if expected == '0':
|
|
||||||
assert result == 0, \
|
|
||||||
LazyFmt("Bad place id in column %s. Expected: 0, got: %s.",
|
|
||||||
column, PlaceObjName(result, context.db))
|
|
||||||
elif expected == '-':
|
|
||||||
assert result is None, \
|
|
||||||
LazyFmt("bad place id in column %s: %s.",
|
|
||||||
column, PlaceObjName(result, context.db))
|
|
||||||
else:
|
|
||||||
assert NominatimID(expected).get_place_id(context.db.cursor()) == result, \
|
|
||||||
LazyFmt("Bad place id in column %s. Expected: %s, got: %s.",
|
|
||||||
column, expected, PlaceObjName(result, context.db))
|
|
||||||
|
|
||||||
def check_database_integrity(context):
|
|
||||||
""" Check some generic constraints on the tables.
|
|
||||||
"""
|
|
||||||
# place_addressline should not have duplicate (place_id, address_place_id)
|
|
||||||
cur = context.db.cursor()
|
|
||||||
cur.execute("""SELECT count(*) FROM
|
|
||||||
(SELECT place_id, address_place_id, count(*) as c
|
|
||||||
FROM place_addressline GROUP BY place_id, address_place_id) x
|
|
||||||
WHERE c > 1""")
|
|
||||||
assert cur.fetchone()[0] == 0, "Duplicates found in place_addressline"
|
|
||||||
|
|
||||||
|
|
||||||
class NominatimID:
|
|
||||||
""" Splits a unique identifier for places into its components.
|
|
||||||
As place_ids cannot be used for testing, we use a unique
|
|
||||||
identifier instead that is of the form <osmtype><osmid>[:<class>].
|
|
||||||
"""
|
|
||||||
|
|
||||||
id_regex = re.compile(r"(?P<tp>[NRW])(?P<id>\d+)(:(?P<cls>\w+))?")
|
|
||||||
|
|
||||||
def __init__(self, oid):
|
|
||||||
self.typ = self.oid = self.cls = None
|
|
||||||
|
|
||||||
if oid is not None:
|
|
||||||
m = self.id_regex.fullmatch(oid)
|
|
||||||
assert m is not None, "ID '%s' not of form <osmtype><osmid>[:<class>]" % oid
|
|
||||||
|
|
||||||
self.typ = m.group('tp')
|
|
||||||
self.oid = m.group('id')
|
|
||||||
self.cls = m.group('cls')
|
|
||||||
|
|
||||||
def __str__(self):
|
|
||||||
if self.cls is None:
|
|
||||||
return self.typ + self.oid
|
|
||||||
|
|
||||||
return '%s%d:%s' % (self.typ, self.oid, self.cls)
|
|
||||||
|
|
||||||
def table_select(self):
|
|
||||||
""" Return where clause and parameter list to select the object
|
|
||||||
from a Nominatim table.
|
|
||||||
"""
|
|
||||||
where = 'osm_type = %s and osm_id = %s'
|
|
||||||
params = [self.typ, self. oid]
|
|
||||||
|
|
||||||
if self.cls is not None:
|
|
||||||
where += ' and class = %s'
|
|
||||||
params.append(self.cls)
|
|
||||||
|
|
||||||
return where, params
|
|
||||||
|
|
||||||
def get_place_id(self, cur):
|
|
||||||
where, params = self.table_select()
|
|
||||||
cur.execute("SELECT place_id FROM placex WHERE %s" % where, params)
|
|
||||||
assert cur.rowcount == 1, \
|
|
||||||
"Expected exactly 1 entry in placex for %s found %s" % (str(self), cur.rowcount)
|
|
||||||
|
|
||||||
return cur.fetchone()[0]
|
|
||||||
|
|
||||||
|
|
||||||
def assert_db_column(row, column, value, context):
|
|
||||||
if column == 'object':
|
|
||||||
return
|
|
||||||
|
|
||||||
if column.startswith('centroid'):
|
|
||||||
if value == 'in geometry':
|
|
||||||
query = """SELECT ST_Within(ST_SetSRID(ST_Point({}, {}), 4326),
|
|
||||||
ST_SetSRID('{}'::geometry, 4326))""".format(
|
|
||||||
row['cx'], row['cy'], row['geomtxt'])
|
|
||||||
cur = context.db.cursor()
|
|
||||||
cur.execute(query)
|
|
||||||
assert cur.fetchone()[0], "(Row %s failed: %s)" % (column, query)
|
|
||||||
else:
|
|
||||||
fac = float(column[9:]) if column.startswith('centroid*') else 1.0
|
|
||||||
x, y = value.split(' ')
|
|
||||||
assert Almost(float(x) * fac) == row['cx'], "Bad x coordinate"
|
|
||||||
assert Almost(float(y) * fac) == row['cy'], "Bad y coordinate"
|
|
||||||
elif column == 'geometry':
|
|
||||||
geom = context.osm.parse_geometry(value, context.scene)
|
|
||||||
cur = context.db.cursor()
|
|
||||||
query = "SELECT ST_Equals(ST_SnapToGrid(%s, 0.00001, 0.00001), ST_SnapToGrid(ST_SetSRID('%s'::geometry, 4326), 0.00001, 0.00001))" % (
|
|
||||||
geom, row['geomtxt'],)
|
|
||||||
cur.execute(query)
|
|
||||||
assert cur.fetchone()[0], "(Row %s failed: %s)" % (column, query)
|
|
||||||
elif value == '-':
|
|
||||||
assert row[column] is None, "Row %s" % column
|
|
||||||
else:
|
|
||||||
assert value == str(row[column]), \
|
|
||||||
"Row '%s': expected: %s, got: %s" % (column, value, str(row[column]))
|
|
||||||
|
|
||||||
|
|
||||||
################################ STEPS ##################################
|
|
||||||
|
|
||||||
@given(u'the scene (?P<scene>.+)')
|
|
||||||
def set_default_scene(context, scene):
|
|
||||||
context.scene = scene
|
|
||||||
|
|
||||||
@given("the (?P<named>named )?places")
|
|
||||||
def add_data_to_place_table(context, named):
|
|
||||||
cur = context.db.cursor()
|
|
||||||
cur.execute('ALTER TABLE place DISABLE TRIGGER place_before_insert')
|
|
||||||
for r in context.table:
|
|
||||||
col = PlaceColumn(context, named is not None)
|
|
||||||
|
|
||||||
for h in r.headings:
|
|
||||||
col.add(h, r[h])
|
|
||||||
|
|
||||||
col.db_insert(cur)
|
|
||||||
cur.execute('ALTER TABLE place ENABLE TRIGGER place_before_insert')
|
|
||||||
cur.close()
|
|
||||||
context.db.commit()
|
|
||||||
|
|
||||||
@given("the relations")
|
|
||||||
def add_data_to_planet_relations(context):
|
|
||||||
cur = context.db.cursor()
|
|
||||||
for r in context.table:
|
|
||||||
last_node = 0
|
|
||||||
last_way = 0
|
|
||||||
parts = []
|
|
||||||
if r['members']:
|
|
||||||
members = []
|
|
||||||
for m in r['members'].split(','):
|
|
||||||
mid = NominatimID(m)
|
|
||||||
if mid.typ == 'N':
|
|
||||||
parts.insert(last_node, int(mid.oid))
|
|
||||||
last_node += 1
|
|
||||||
last_way += 1
|
|
||||||
elif mid.typ == 'W':
|
|
||||||
parts.insert(last_way, int(mid.oid))
|
|
||||||
last_way += 1
|
|
||||||
else:
|
|
||||||
parts.append(int(mid.oid))
|
|
||||||
|
|
||||||
members.extend((mid.typ.lower() + mid.oid, mid.cls or ''))
|
|
||||||
else:
|
|
||||||
members = None
|
|
||||||
|
|
||||||
tags = []
|
|
||||||
for h in r.headings:
|
|
||||||
if h.startswith("tags+"):
|
|
||||||
tags.extend((h[5:], r[h]))
|
|
||||||
|
|
||||||
cur.execute("""INSERT INTO planet_osm_rels (id, way_off, rel_off, parts, members, tags)
|
|
||||||
VALUES (%s, %s, %s, %s, %s, %s)""",
|
|
||||||
(r['id'], last_node, last_way, parts, members, tags))
|
|
||||||
context.db.commit()
|
|
||||||
|
|
||||||
@given("the ways")
|
|
||||||
def add_data_to_planet_ways(context):
|
|
||||||
cur = context.db.cursor()
|
|
||||||
for r in context.table:
|
|
||||||
tags = []
|
|
||||||
for h in r.headings:
|
|
||||||
if h.startswith("tags+"):
|
|
||||||
tags.extend((h[5:], r[h]))
|
|
||||||
|
|
||||||
nodes = [ int(x.strip()) for x in r['nodes'].split(',') ]
|
|
||||||
|
|
||||||
cur.execute("INSERT INTO planet_osm_ways (id, nodes, tags) VALUES (%s, %s, %s)",
|
|
||||||
(r['id'], nodes, tags))
|
|
||||||
context.db.commit()
|
|
||||||
|
|
||||||
@when("importing")
|
|
||||||
def import_and_index_data_from_place_table(context):
|
|
||||||
context.nominatim.run_setup_script('create-functions', 'create-partition-functions')
|
|
||||||
cur = context.db.cursor()
|
|
||||||
cur.execute(
|
|
||||||
"""insert into placex (osm_type, osm_id, class, type, name, admin_level, address, extratags, geometry)
|
|
||||||
select osm_type, osm_id, class, type, name, admin_level, address, extratags, geometry
|
|
||||||
from place where not (class='place' and type='houses' and osm_type='W')""")
|
|
||||||
cur.execute(
|
|
||||||
"""insert into location_property_osmline (osm_id, address, linegeo)
|
|
||||||
SELECT osm_id, address, geometry from place
|
|
||||||
WHERE class='place' and type='houses' and osm_type='W'
|
|
||||||
and ST_GeometryType(geometry) = 'ST_LineString'""")
|
|
||||||
context.db.commit()
|
|
||||||
context.nominatim.run_setup_script('calculate-postcodes', 'index', 'index-noanalyse')
|
|
||||||
check_database_integrity(context)
|
|
||||||
|
|
||||||
@when("updating places")
|
|
||||||
def update_place_table(context):
|
|
||||||
context.nominatim.run_setup_script(
|
|
||||||
'create-functions', 'create-partition-functions', 'enable-diff-updates')
|
|
||||||
cur = context.db.cursor()
|
|
||||||
for r in context.table:
|
|
||||||
col = PlaceColumn(context, False)
|
|
||||||
|
|
||||||
for h in r.headings:
|
|
||||||
col.add(h, r[h])
|
|
||||||
|
|
||||||
col.db_insert(cur)
|
|
||||||
|
|
||||||
context.db.commit()
|
|
||||||
|
|
||||||
while True:
|
|
||||||
context.nominatim.run_update_script('index')
|
|
||||||
|
|
||||||
cur = context.db.cursor()
|
|
||||||
cur.execute("SELECT 'a' FROM placex WHERE indexed_status != 0 LIMIT 1")
|
|
||||||
if cur.rowcount == 0:
|
|
||||||
break
|
|
||||||
|
|
||||||
check_database_integrity(context)
|
|
||||||
|
|
||||||
@when("updating postcodes")
|
|
||||||
def update_postcodes(context):
|
|
||||||
context.nominatim.run_update_script('calculate-postcodes')
|
|
||||||
|
|
||||||
@when("marking for delete (?P<oids>.*)")
|
|
||||||
def delete_places(context, oids):
|
|
||||||
context.nominatim.run_setup_script(
|
|
||||||
'create-functions', 'create-partition-functions', 'enable-diff-updates')
|
|
||||||
cur = context.db.cursor()
|
|
||||||
for oid in oids.split(','):
|
|
||||||
where, params = NominatimID(oid).table_select()
|
|
||||||
cur.execute("DELETE FROM place WHERE " + where, params)
|
|
||||||
context.db.commit()
|
|
||||||
|
|
||||||
while True:
|
|
||||||
context.nominatim.run_update_script('index')
|
|
||||||
|
|
||||||
cur = context.db.cursor()
|
|
||||||
cur.execute("SELECT 'a' FROM placex WHERE indexed_status != 0 LIMIT 1")
|
|
||||||
if cur.rowcount == 0:
|
|
||||||
break
|
|
||||||
|
|
||||||
@then("placex contains(?P<exact> exactly)?")
|
|
||||||
def check_placex_contents(context, exact):
|
|
||||||
cur = context.db.cursor(cursor_factory=psycopg2.extras.DictCursor)
|
|
||||||
|
|
||||||
expected_content = set()
|
|
||||||
for row in context.table:
|
|
||||||
nid = NominatimID(row['object'])
|
|
||||||
where, params = nid.table_select()
|
|
||||||
cur.execute("""SELECT *, ST_AsText(geometry) as geomtxt,
|
|
||||||
ST_X(centroid) as cx, ST_Y(centroid) as cy
|
|
||||||
FROM placex where %s""" % where,
|
|
||||||
params)
|
|
||||||
assert cur.rowcount > 0, "No rows found for " + row['object']
|
|
||||||
|
|
||||||
for res in cur:
|
|
||||||
if exact:
|
|
||||||
expected_content.add((res['osm_type'], res['osm_id'], res['class']))
|
|
||||||
for h in row.headings:
|
|
||||||
if h in ('extratags', 'address'):
|
|
||||||
if row[h] == '-':
|
|
||||||
assert res[h] is None
|
|
||||||
else:
|
|
||||||
vdict = eval('{' + row[h] + '}')
|
|
||||||
assert vdict == res[h]
|
|
||||||
elif h.startswith('name'):
|
|
||||||
name = h[5:] if h.startswith('name+') else 'name'
|
|
||||||
assert name in res['name']
|
|
||||||
assert res['name'][name] == row[h]
|
|
||||||
elif h.startswith('extratags+'):
|
|
||||||
assert res['extratags'][h[10:]] == row[h]
|
|
||||||
elif h.startswith('addr+'):
|
|
||||||
if row[h] == '-':
|
|
||||||
if res['address'] is not None:
|
|
||||||
assert h[5:] not in res['address']
|
|
||||||
else:
|
|
||||||
assert h[5:] in res['address'], "column " + h
|
|
||||||
assert res['address'][h[5:]] == row[h], "column %s" % h
|
|
||||||
elif h in ('linked_place_id', 'parent_place_id'):
|
|
||||||
compare_place_id(row[h], res[h], h, context)
|
|
||||||
else:
|
|
||||||
assert_db_column(res, h, row[h], context)
|
|
||||||
|
|
||||||
if exact:
|
|
||||||
cur.execute('SELECT osm_type, osm_id, class from placex')
|
|
||||||
assert expected_content == set([(r[0], r[1], r[2]) for r in cur])
|
|
||||||
|
|
||||||
context.db.commit()
|
|
||||||
|
|
||||||
@then("place contains(?P<exact> exactly)?")
|
|
||||||
def check_placex_contents(context, exact):
|
|
||||||
cur = context.db.cursor(cursor_factory=psycopg2.extras.DictCursor)
|
|
||||||
|
|
||||||
expected_content = set()
|
|
||||||
for row in context.table:
|
|
||||||
nid = NominatimID(row['object'])
|
|
||||||
where, params = nid.table_select()
|
|
||||||
cur.execute("""SELECT *, ST_AsText(geometry) as geomtxt,
|
|
||||||
ST_GeometryType(geometry) as geometrytype
|
|
||||||
FROM place where %s""" % where,
|
|
||||||
params)
|
|
||||||
assert cur.rowcount > 0, "No rows found for " + row['object']
|
|
||||||
|
|
||||||
for res in cur:
|
|
||||||
if exact:
|
|
||||||
expected_content.add((res['osm_type'], res['osm_id'], res['class']))
|
|
||||||
for h in row.headings:
|
|
||||||
msg = "%s: %s" % (row['object'], h)
|
|
||||||
if h in ('name', 'extratags', 'address'):
|
|
||||||
if row[h] == '-':
|
|
||||||
assert res[h] is None, msg
|
|
||||||
else:
|
|
||||||
vdict = eval('{' + row[h] + '}')
|
|
||||||
assert vdict == res[h], msg
|
|
||||||
elif h.startswith('name+'):
|
|
||||||
assert res['name'][h[5:]] == row[h], msg
|
|
||||||
elif h.startswith('extratags+'):
|
|
||||||
assert res['extratags'][h[10:]] == row[h], msg
|
|
||||||
elif h.startswith('addr+'):
|
|
||||||
if row[h] == '-':
|
|
||||||
if res['address'] is not None:
|
|
||||||
assert h[5:] not in res['address']
|
|
||||||
else:
|
|
||||||
assert res['address'][h[5:]] == row[h], msg
|
|
||||||
elif h in ('linked_place_id', 'parent_place_id'):
|
|
||||||
compare_place_id(row[h], res[h], h, context)
|
|
||||||
else:
|
|
||||||
assert_db_column(res, h, row[h], context)
|
|
||||||
|
|
||||||
if exact:
|
|
||||||
cur.execute('SELECT osm_type, osm_id, class from place')
|
|
||||||
assert expected_content, set([(r[0], r[1], r[2]) for r in cur])
|
|
||||||
|
|
||||||
context.db.commit()
|
|
||||||
|
|
||||||
@then("search_name contains(?P<exclude> not)?")
|
|
||||||
def check_search_name_contents(context, exclude):
|
|
||||||
cur = context.db.cursor(cursor_factory=psycopg2.extras.DictCursor)
|
|
||||||
|
|
||||||
for row in context.table:
|
|
||||||
pid = NominatimID(row['object']).get_place_id(cur)
|
|
||||||
cur.execute("""SELECT *, ST_X(centroid) as cx, ST_Y(centroid) as cy
|
|
||||||
FROM search_name WHERE place_id = %s""", (pid, ))
|
|
||||||
assert cur.rowcount > 0, "No rows found for " + row['object']
|
|
||||||
|
|
||||||
for res in cur:
|
|
||||||
for h in row.headings:
|
|
||||||
if h in ('name_vector', 'nameaddress_vector'):
|
|
||||||
terms = [x.strip() for x in row[h].split(',') if not x.strip().startswith('#')]
|
|
||||||
words = [x.strip()[1:] for x in row[h].split(',') if x.strip().startswith('#')]
|
|
||||||
subcur = context.db.cursor()
|
|
||||||
subcur.execute(""" SELECT word_id, word_token
|
|
||||||
FROM word, (SELECT unnest(%s::TEXT[]) as term) t
|
|
||||||
WHERE word_token = make_standard_name(t.term)
|
|
||||||
and class is null and country_code is null
|
|
||||||
and operator is null
|
|
||||||
UNION
|
|
||||||
SELECT word_id, word_token
|
|
||||||
FROM word, (SELECT unnest(%s::TEXT[]) as term) t
|
|
||||||
WHERE word_token = ' ' || make_standard_name(t.term)
|
|
||||||
and class is null and country_code is null
|
|
||||||
and operator is null
|
|
||||||
""",
|
|
||||||
(terms, words))
|
|
||||||
if not exclude:
|
|
||||||
assert subcur.rowcount >= len(terms) + len(words), \
|
|
||||||
"No word entry found for " + row[h] + ". Entries found: " + str(subcur.rowcount)
|
|
||||||
for wid in subcur:
|
|
||||||
if exclude:
|
|
||||||
assert wid[0] not in res[h], "Found term for %s/%s: %s" % (pid, h, wid[1])
|
|
||||||
else:
|
|
||||||
assert wid[0] in res[h], "Missing term for %s/%s: %s" % (pid, h, wid[1])
|
|
||||||
else:
|
|
||||||
assert_db_column(res, h, row[h], context)
|
|
||||||
|
|
||||||
|
|
||||||
context.db.commit()
|
|
||||||
|
|
||||||
@then("location_postcode contains exactly")
|
|
||||||
def check_location_postcode(context):
|
|
||||||
cur = context.db.cursor(cursor_factory=psycopg2.extras.DictCursor)
|
|
||||||
|
|
||||||
cur.execute("SELECT *, ST_AsText(geometry) as geomtxt FROM location_postcode")
|
|
||||||
assert cur.rowcount == len(list(context.table)), \
|
|
||||||
"Postcode table has %d rows, expected %d rows." % (cur.rowcount, len(list(context.table)))
|
|
||||||
|
|
||||||
table = list(cur)
|
|
||||||
for row in context.table:
|
|
||||||
for i in range(len(table)):
|
|
||||||
if table[i]['country_code'] != row['country'] \
|
|
||||||
or table[i]['postcode'] != row['postcode']:
|
|
||||||
continue
|
|
||||||
for h in row.headings:
|
|
||||||
if h not in ('country', 'postcode'):
|
|
||||||
assert_db_column(table[i], h, row[h], context)
|
|
||||||
|
|
||||||
@then("word contains(?P<exclude> not)?")
|
|
||||||
def check_word_table(context, exclude):
|
|
||||||
cur = context.db.cursor(cursor_factory=psycopg2.extras.DictCursor)
|
|
||||||
|
|
||||||
for row in context.table:
|
|
||||||
wheres = []
|
|
||||||
values = []
|
|
||||||
for h in row.headings:
|
|
||||||
wheres.append("%s = %%s" % h)
|
|
||||||
values.append(row[h])
|
|
||||||
cur.execute("SELECT * from word WHERE %s" % ' AND '.join(wheres), values)
|
|
||||||
if exclude:
|
|
||||||
assert cur.rowcount == 0, "Row still in word table: %s" % '/'.join(values)
|
|
||||||
else:
|
|
||||||
assert cur.rowcount > 0, "Row not in word table: %s" % '/'.join(values)
|
|
||||||
|
|
||||||
@then("place_addressline contains")
|
|
||||||
def check_place_addressline(context):
|
|
||||||
cur = context.db.cursor(cursor_factory=psycopg2.extras.DictCursor)
|
|
||||||
|
|
||||||
for row in context.table:
|
|
||||||
pid = NominatimID(row['object']).get_place_id(cur)
|
|
||||||
apid = NominatimID(row['address']).get_place_id(cur)
|
|
||||||
cur.execute(""" SELECT * FROM place_addressline
|
|
||||||
WHERE place_id = %s AND address_place_id = %s""",
|
|
||||||
(pid, apid))
|
|
||||||
assert cur.rowcount > 0, \
|
|
||||||
"No rows found for place %s and address %s" % (row['object'], row['address'])
|
|
||||||
|
|
||||||
for res in cur:
|
|
||||||
for h in row.headings:
|
|
||||||
if h not in ('address', 'object'):
|
|
||||||
assert_db_column(res, h, row[h], context)
|
|
||||||
|
|
||||||
context.db.commit()
|
|
||||||
|
|
||||||
@then("place_addressline doesn't contain")
|
|
||||||
def check_place_addressline_exclude(context):
|
|
||||||
cur = context.db.cursor(cursor_factory=psycopg2.extras.DictCursor)
|
|
||||||
|
|
||||||
for row in context.table:
|
|
||||||
pid = NominatimID(row['object']).get_place_id(cur)
|
|
||||||
apid = NominatimID(row['address']).get_place_id(cur)
|
|
||||||
cur.execute(""" SELECT * FROM place_addressline
|
|
||||||
WHERE place_id = %s AND address_place_id = %s""",
|
|
||||||
(pid, apid))
|
|
||||||
assert cur.rowcount == 0, \
|
|
||||||
"Row found for place %s and address %s" % (row['object'], row['address'])
|
|
||||||
|
|
||||||
context.db.commit()
|
|
||||||
|
|
||||||
@then("(?P<oid>\w+) expands to(?P<neg> no)? interpolation")
|
|
||||||
def check_location_property_osmline(context, oid, neg):
|
|
||||||
cur = context.db.cursor(cursor_factory=psycopg2.extras.DictCursor)
|
|
||||||
nid = NominatimID(oid)
|
|
||||||
|
|
||||||
assert 'W' == nid.typ, "interpolation must be a way"
|
|
||||||
|
|
||||||
cur.execute("""SELECT *, ST_AsText(linegeo) as geomtxt
|
|
||||||
FROM location_property_osmline
|
|
||||||
WHERE osm_id = %s AND startnumber IS NOT NULL""",
|
|
||||||
(nid.oid, ))
|
|
||||||
|
|
||||||
if neg:
|
|
||||||
assert cur.rowcount == 0
|
|
||||||
return
|
|
||||||
|
|
||||||
todo = list(range(len(list(context.table))))
|
|
||||||
for res in cur:
|
|
||||||
for i in todo:
|
|
||||||
row = context.table[i]
|
|
||||||
if (int(row['start']) == res['startnumber']
|
|
||||||
and int(row['end']) == res['endnumber']):
|
|
||||||
todo.remove(i)
|
|
||||||
break
|
|
||||||
else:
|
|
||||||
assert False, "Unexpected row %s" % (str(res))
|
|
||||||
|
|
||||||
for h in row.headings:
|
|
||||||
if h in ('start', 'end'):
|
|
||||||
continue
|
|
||||||
elif h == 'parent_place_id':
|
|
||||||
compare_place_id(row[h], res[h], h, context)
|
|
||||||
else:
|
|
||||||
assert_db_column(res, h, row[h], context)
|
|
||||||
|
|
||||||
assert not todo
|
|
||||||
|
|
||||||
|
|
||||||
@then("(?P<table>placex|place) has no entry for (?P<oid>.*)")
|
|
||||||
def check_placex_has_entry(context, table, oid):
|
|
||||||
cur = context.db.cursor(cursor_factory=psycopg2.extras.DictCursor)
|
|
||||||
nid = NominatimID(oid)
|
|
||||||
where, params = nid.table_select()
|
|
||||||
cur.execute("SELECT * FROM %s where %s" % (table, where), params)
|
|
||||||
assert cur.rowcount == 0
|
|
||||||
context.db.commit()
|
|
||||||
|
|
||||||
@then("search_name has no entry for (?P<oid>.*)")
|
|
||||||
def check_search_name_has_entry(context, oid):
|
|
||||||
cur = context.db.cursor(cursor_factory=psycopg2.extras.DictCursor)
|
|
||||||
pid = NominatimID(oid).get_place_id(cur)
|
|
||||||
cur.execute("SELECT * FROM search_name WHERE place_id = %s", (pid, ))
|
|
||||||
assert cur.rowcount == 0
|
|
||||||
context.db.commit()
|
|
||||||
122
test/bdd/steps/geometry_factory.py
Normal file
122
test/bdd/steps/geometry_factory.py
Normal file
@@ -0,0 +1,122 @@
|
|||||||
|
from pathlib import Path
|
||||||
|
import os
|
||||||
|
|
||||||
|
class GeometryFactory:
|
||||||
|
""" Provides functions to create geometries from scenes and data grids.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
defpath = Path(__file__) / '..' / '..' / '..' / 'scenes' / 'data'
|
||||||
|
self.scene_path = os.environ.get('SCENE_PATH', defpath.resolve())
|
||||||
|
self.scene_cache = {}
|
||||||
|
self.grid = {}
|
||||||
|
|
||||||
|
def parse_geometry(self, geom, scene):
|
||||||
|
""" Create a WKT SQL term for the given geometry.
|
||||||
|
The function understands the following formats:
|
||||||
|
|
||||||
|
[<scene>]:<name>
|
||||||
|
Geometry from a scene. If the scene is omitted, use the
|
||||||
|
default scene.
|
||||||
|
<P>
|
||||||
|
Point geometry
|
||||||
|
<P>,...,<P>
|
||||||
|
Line geometry
|
||||||
|
(<P>,...,<P>)
|
||||||
|
Polygon geometry
|
||||||
|
|
||||||
|
<P> may either be a coordinate of the form '<x> <y>' or a single
|
||||||
|
number. In the latter case it must refer to a point in
|
||||||
|
a previously defined grid.
|
||||||
|
"""
|
||||||
|
if geom.find(':') >= 0:
|
||||||
|
return "ST_SetSRID({}, 4326)".format(self.get_scene_geometry(scene, geom))
|
||||||
|
|
||||||
|
if geom.find(',') < 0:
|
||||||
|
out = "POINT({})".format(self.mk_wkt_point(geom))
|
||||||
|
elif geom.find('(') < 0:
|
||||||
|
out = "LINESTRING({})".format(self.mk_wkt_points(geom))
|
||||||
|
else:
|
||||||
|
out = "POLYGON(({}))".format(self.mk_wkt_points(geom.strip('() ')))
|
||||||
|
|
||||||
|
return "ST_SetSRID('{}'::geometry, 4326)".format(out)
|
||||||
|
|
||||||
|
def mk_wkt_point(self, point):
|
||||||
|
""" Parse a point description.
|
||||||
|
The point may either consist of 'x y' cooordinates or a number
|
||||||
|
that refers to a grid setup.
|
||||||
|
"""
|
||||||
|
geom = point.strip()
|
||||||
|
if geom.find(' ') >= 0:
|
||||||
|
return geom
|
||||||
|
|
||||||
|
try:
|
||||||
|
pt = self.grid_node(int(geom))
|
||||||
|
except ValueError:
|
||||||
|
assert False, "Scenario error: Point '{}' is not a number".format(geom)
|
||||||
|
|
||||||
|
assert pt is not None, "Scenario error: Point '{}' not found in grid".format(geom)
|
||||||
|
return "{} {}".format(*pt)
|
||||||
|
|
||||||
|
def mk_wkt_points(self, geom):
|
||||||
|
""" Parse a list of points.
|
||||||
|
The list must be a comma-separated list of points. Points
|
||||||
|
in coordinate and grid format may be mixed.
|
||||||
|
"""
|
||||||
|
return ','.join([self.mk_wkt_point(x) for x in geom.split(',')])
|
||||||
|
|
||||||
|
def get_scene_geometry(self, default_scene, name):
|
||||||
|
""" Load the geometry from a scene.
|
||||||
|
"""
|
||||||
|
geoms = []
|
||||||
|
for obj in name.split('+'):
|
||||||
|
oname = obj.strip()
|
||||||
|
if oname.startswith(':'):
|
||||||
|
assert default_scene is not None, "Scenario error: You need to set a scene"
|
||||||
|
defscene = self.load_scene(default_scene)
|
||||||
|
wkt = defscene[oname[1:]]
|
||||||
|
else:
|
||||||
|
scene, obj = oname.split(':', 2)
|
||||||
|
scene_geoms = self.load_scene(scene)
|
||||||
|
wkt = scene_geoms[obj]
|
||||||
|
|
||||||
|
geoms.append("'{}'::geometry".format(wkt))
|
||||||
|
|
||||||
|
if len(geoms) == 1:
|
||||||
|
return geoms[0]
|
||||||
|
|
||||||
|
return 'ST_LineMerge(ST_Collect(ARRAY[{}]))'.format(','.join(geoms))
|
||||||
|
|
||||||
|
def load_scene(self, name):
|
||||||
|
""" Load a scene from a file.
|
||||||
|
"""
|
||||||
|
if name in self.scene_cache:
|
||||||
|
return self.scene_cache[name]
|
||||||
|
|
||||||
|
scene = {}
|
||||||
|
with open(Path(self.scene_path) / "{}.wkt".format(name), 'r') as fd:
|
||||||
|
for line in fd:
|
||||||
|
if line.strip():
|
||||||
|
obj, wkt = line.split('|', 2)
|
||||||
|
scene[obj.strip()] = wkt.strip()
|
||||||
|
self.scene_cache[name] = scene
|
||||||
|
|
||||||
|
return scene
|
||||||
|
|
||||||
|
def set_grid(self, lines, grid_step):
|
||||||
|
""" Replace the grid with one from the given lines.
|
||||||
|
"""
|
||||||
|
self.grid = {}
|
||||||
|
y = 0
|
||||||
|
for line in lines:
|
||||||
|
x = 0
|
||||||
|
for pt_id in line:
|
||||||
|
if pt_id.isdigit():
|
||||||
|
self.grid[int(pt_id)] = (x, y)
|
||||||
|
x += grid_step
|
||||||
|
y += grid_step
|
||||||
|
|
||||||
|
def grid_node(self, nodeid):
|
||||||
|
""" Get the coordinates for the given grid node.
|
||||||
|
"""
|
||||||
|
return self.grid.get(nodeid)
|
||||||
198
test/bdd/steps/http_responses.py
Normal file
198
test/bdd/steps/http_responses.py
Normal file
@@ -0,0 +1,198 @@
|
|||||||
|
"""
|
||||||
|
Classes wrapping HTTP responses from the Nominatim API.
|
||||||
|
"""
|
||||||
|
from collections import OrderedDict
|
||||||
|
import re
|
||||||
|
import json
|
||||||
|
import xml.etree.ElementTree as ET
|
||||||
|
|
||||||
|
from check_functions import Almost
|
||||||
|
|
||||||
|
def _geojson_result_to_json_result(geojson_result):
|
||||||
|
result = geojson_result['properties']
|
||||||
|
result['geojson'] = geojson_result['geometry']
|
||||||
|
if 'bbox' in geojson_result:
|
||||||
|
# bbox is minlon, minlat, maxlon, maxlat
|
||||||
|
# boundingbox is minlat, maxlat, minlon, maxlon
|
||||||
|
result['boundingbox'] = [geojson_result['bbox'][1],
|
||||||
|
geojson_result['bbox'][3],
|
||||||
|
geojson_result['bbox'][0],
|
||||||
|
geojson_result['bbox'][2]]
|
||||||
|
return result
|
||||||
|
|
||||||
|
class BadRowValueAssert:
|
||||||
|
""" Lazily formatted message for failures to find a field content.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, response, idx, field, value):
|
||||||
|
self.idx = idx
|
||||||
|
self.field = field
|
||||||
|
self.value = value
|
||||||
|
self.row = response.result[idx]
|
||||||
|
|
||||||
|
def __str__(self):
|
||||||
|
return "\nBad value for row {} field '{}'. Expected: {}, got: {}.\nFull row: {}"""\
|
||||||
|
.format(self.idx, self.field, self.value,
|
||||||
|
self.row[self.field], json.dumps(self.row, indent=4))
|
||||||
|
|
||||||
|
|
||||||
|
class GenericResponse:
|
||||||
|
""" Common base class for all API responses.
|
||||||
|
"""
|
||||||
|
def __init__(self, page, fmt, errorcode=200):
|
||||||
|
fmt = fmt.strip()
|
||||||
|
if fmt == 'jsonv2':
|
||||||
|
fmt = 'json'
|
||||||
|
|
||||||
|
self.page = page
|
||||||
|
self.format = fmt
|
||||||
|
self.errorcode = errorcode
|
||||||
|
self.result = []
|
||||||
|
self.header = dict()
|
||||||
|
|
||||||
|
if errorcode == 200:
|
||||||
|
getattr(self, '_parse_' + fmt)()
|
||||||
|
|
||||||
|
def _parse_json(self):
|
||||||
|
m = re.fullmatch(r'([\w$][^(]*)\((.*)\)', self.page)
|
||||||
|
if m is None:
|
||||||
|
code = self.page
|
||||||
|
else:
|
||||||
|
code = m.group(2)
|
||||||
|
self.header['json_func'] = m.group(1)
|
||||||
|
self.result = json.JSONDecoder(object_pairs_hook=OrderedDict).decode(code)
|
||||||
|
if isinstance(self.result, OrderedDict):
|
||||||
|
self.result = [self.result]
|
||||||
|
|
||||||
|
def _parse_geojson(self):
|
||||||
|
self._parse_json()
|
||||||
|
if 'error' in self.result[0]:
|
||||||
|
self.result = []
|
||||||
|
else:
|
||||||
|
self.result = list(map(_geojson_result_to_json_result, self.result[0]['features']))
|
||||||
|
|
||||||
|
def _parse_geocodejson(self):
|
||||||
|
self._parse_geojson()
|
||||||
|
if self.result is not None:
|
||||||
|
self.result = [r['geocoding'] for r in self.result]
|
||||||
|
|
||||||
|
def assert_field(self, idx, field, value):
|
||||||
|
""" Check that result row `idx` has a field `field` with value `value`.
|
||||||
|
Float numbers are matched approximately. When the expected value
|
||||||
|
starts with a carat, regular expression matching is used.
|
||||||
|
"""
|
||||||
|
assert field in self.result[idx], \
|
||||||
|
"Result row {} has no field '{}'.\nFull row: {}"\
|
||||||
|
.format(idx, field, json.dumps(self.result[idx], indent=4))
|
||||||
|
|
||||||
|
if isinstance(value, float):
|
||||||
|
assert Almost(value) == float(self.result[idx][field]), \
|
||||||
|
BadRowValueAssert(self, idx, field, value)
|
||||||
|
elif value.startswith("^"):
|
||||||
|
assert re.fullmatch(value, self.result[idx][field]), \
|
||||||
|
BadRowValueAssert(self, idx, field, value)
|
||||||
|
else:
|
||||||
|
assert str(self.result[idx][field]) == str(value), \
|
||||||
|
BadRowValueAssert(self, idx, field, value)
|
||||||
|
|
||||||
|
def match_row(self, row):
|
||||||
|
""" Match the result fields against the given behave table row.
|
||||||
|
"""
|
||||||
|
if 'ID' in row.headings:
|
||||||
|
todo = [int(row['ID'])]
|
||||||
|
else:
|
||||||
|
todo = range(len(self.result))
|
||||||
|
|
||||||
|
for i in todo:
|
||||||
|
for name, value in zip(row.headings, row.cells):
|
||||||
|
if name == 'ID':
|
||||||
|
pass
|
||||||
|
elif name == 'osm':
|
||||||
|
self.assert_field(i, 'osm_type', value[0])
|
||||||
|
self.assert_field(i, 'osm_id', value[1:])
|
||||||
|
elif name == 'centroid':
|
||||||
|
lon, lat = value.split(' ')
|
||||||
|
self.assert_field(i, 'lat', float(lat))
|
||||||
|
self.assert_field(i, 'lon', float(lon))
|
||||||
|
else:
|
||||||
|
self.assert_field(i, name, value)
|
||||||
|
|
||||||
|
def property_list(self, prop):
|
||||||
|
return [x[prop] for x in self.result]
|
||||||
|
|
||||||
|
|
||||||
|
class SearchResponse(GenericResponse):
|
||||||
|
""" Specialised class for search and lookup responses.
|
||||||
|
Transforms the xml response in a format similar to json.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def _parse_xml(self):
|
||||||
|
xml_tree = ET.fromstring(self.page)
|
||||||
|
|
||||||
|
self.header = dict(xml_tree.attrib)
|
||||||
|
|
||||||
|
for child in xml_tree:
|
||||||
|
assert child.tag == "place"
|
||||||
|
self.result.append(dict(child.attrib))
|
||||||
|
|
||||||
|
address = {}
|
||||||
|
for sub in child:
|
||||||
|
if sub.tag == 'extratags':
|
||||||
|
self.result[-1]['extratags'] = {}
|
||||||
|
for tag in sub:
|
||||||
|
self.result[-1]['extratags'][tag.attrib['key']] = tag.attrib['value']
|
||||||
|
elif sub.tag == 'namedetails':
|
||||||
|
self.result[-1]['namedetails'] = {}
|
||||||
|
for tag in sub:
|
||||||
|
self.result[-1]['namedetails'][tag.attrib['desc']] = tag.text
|
||||||
|
elif sub.tag == 'geokml':
|
||||||
|
self.result[-1][sub.tag] = True
|
||||||
|
else:
|
||||||
|
address[sub.tag] = sub.text
|
||||||
|
|
||||||
|
if address:
|
||||||
|
self.result[-1]['address'] = address
|
||||||
|
|
||||||
|
|
||||||
|
class ReverseResponse(GenericResponse):
|
||||||
|
""" Specialised class for reverse responses.
|
||||||
|
Transforms the xml response in a format similar to json.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def _parse_xml(self):
|
||||||
|
xml_tree = ET.fromstring(self.page)
|
||||||
|
|
||||||
|
self.header = dict(xml_tree.attrib)
|
||||||
|
self.result = []
|
||||||
|
|
||||||
|
for child in xml_tree:
|
||||||
|
if child.tag == 'result':
|
||||||
|
assert not self.result, "More than one result in reverse result"
|
||||||
|
self.result.append(dict(child.attrib))
|
||||||
|
elif child.tag == 'addressparts':
|
||||||
|
address = {}
|
||||||
|
for sub in child:
|
||||||
|
address[sub.tag] = sub.text
|
||||||
|
self.result[0]['address'] = address
|
||||||
|
elif child.tag == 'extratags':
|
||||||
|
self.result[0]['extratags'] = {}
|
||||||
|
for tag in child:
|
||||||
|
self.result[0]['extratags'][tag.attrib['key']] = tag.attrib['value']
|
||||||
|
elif child.tag == 'namedetails':
|
||||||
|
self.result[0]['namedetails'] = {}
|
||||||
|
for tag in child:
|
||||||
|
self.result[0]['namedetails'][tag.attrib['desc']] = tag.text
|
||||||
|
elif child.tag == 'geokml':
|
||||||
|
self.result[0][child.tag] = True
|
||||||
|
else:
|
||||||
|
assert child.tag == 'error', \
|
||||||
|
"Unknown XML tag {} on page: {}".format(child.tag, self.page)
|
||||||
|
|
||||||
|
|
||||||
|
class StatusResponse(GenericResponse):
|
||||||
|
""" Specialised class for status responses.
|
||||||
|
Can also parse text responses.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def _parse_text(self):
|
||||||
|
pass
|
||||||
254
test/bdd/steps/nominatim_environment.py
Normal file
254
test/bdd/steps/nominatim_environment.py
Normal file
@@ -0,0 +1,254 @@
|
|||||||
|
import os
|
||||||
|
from pathlib import Path
|
||||||
|
import tempfile
|
||||||
|
|
||||||
|
import psycopg2
|
||||||
|
import psycopg2.extras
|
||||||
|
|
||||||
|
from steps.utils import run_script
|
||||||
|
|
||||||
|
class NominatimEnvironment:
|
||||||
|
""" Collects all functions for the execution of Nominatim functions.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, config):
|
||||||
|
self.build_dir = Path(config['BUILDDIR']).resolve()
|
||||||
|
self.src_dir = (Path(__file__) / '..' / '..' / '..' / '..').resolve()
|
||||||
|
self.db_host = config['DB_HOST']
|
||||||
|
self.db_port = config['DB_PORT']
|
||||||
|
self.db_user = config['DB_USER']
|
||||||
|
self.db_pass = config['DB_PASS']
|
||||||
|
self.template_db = config['TEMPLATE_DB']
|
||||||
|
self.test_db = config['TEST_DB']
|
||||||
|
self.api_test_db = config['API_TEST_DB']
|
||||||
|
self.server_module_path = config['SERVER_MODULE_PATH']
|
||||||
|
self.reuse_template = not config['REMOVE_TEMPLATE']
|
||||||
|
self.keep_scenario_db = config['KEEP_TEST_DB']
|
||||||
|
self.code_coverage_path = config['PHPCOV']
|
||||||
|
self.code_coverage_id = 1
|
||||||
|
self.test_env = None
|
||||||
|
|
||||||
|
self.template_db_done = False
|
||||||
|
self.website_dir = None
|
||||||
|
|
||||||
|
def connect_database(self, dbname):
|
||||||
|
""" Return a connection to the database with the given name.
|
||||||
|
Uses configured host, user and port.
|
||||||
|
"""
|
||||||
|
dbargs = {'database': dbname}
|
||||||
|
if self.db_host:
|
||||||
|
dbargs['host'] = self.db_host
|
||||||
|
if self.db_port:
|
||||||
|
dbargs['port'] = self.db_port
|
||||||
|
if self.db_user:
|
||||||
|
dbargs['user'] = self.db_user
|
||||||
|
if self.db_pass:
|
||||||
|
dbargs['password'] = self.db_pass
|
||||||
|
conn = psycopg2.connect(**dbargs)
|
||||||
|
return conn
|
||||||
|
|
||||||
|
def next_code_coverage_file(self):
|
||||||
|
""" Generate the next name for a coverage file.
|
||||||
|
"""
|
||||||
|
fn = Path(self.code_coverage_path) / "{:06d}.cov".format(self.code_coverage_id)
|
||||||
|
self.code_coverage_id += 1
|
||||||
|
|
||||||
|
return fn.resolve()
|
||||||
|
|
||||||
|
def write_nominatim_config(self, dbname):
|
||||||
|
""" Set up a custom test configuration that connects to the given
|
||||||
|
database. This sets up the environment variables so that they can
|
||||||
|
be picked up by dotenv and creates a project directory with the
|
||||||
|
appropriate website scripts.
|
||||||
|
"""
|
||||||
|
dsn = 'pgsql:dbname={}'.format(dbname)
|
||||||
|
if self.db_host:
|
||||||
|
dsn += ';host=' + self.db_host
|
||||||
|
if self.db_port:
|
||||||
|
dsn += ';port=' + self.db_port
|
||||||
|
if self.db_user:
|
||||||
|
dsn += ';user=' + self.db_user
|
||||||
|
if self.db_pass:
|
||||||
|
dsn += ';password=' + self.db_pass
|
||||||
|
|
||||||
|
if self.website_dir is not None \
|
||||||
|
and self.test_env is not None \
|
||||||
|
and dsn == self.test_env['NOMINATIM_DATABASE_DSN']:
|
||||||
|
return # environment already set uo
|
||||||
|
|
||||||
|
self.test_env = os.environ
|
||||||
|
self.test_env['NOMINATIM_DATABASE_DSN'] = dsn
|
||||||
|
self.test_env['NOMINATIM_FLATNODE_FILE'] = ''
|
||||||
|
self.test_env['NOMINATIM_IMPORT_STYLE'] = 'full'
|
||||||
|
self.test_env['NOMINATIM_USE_US_TIGER_DATA'] = 'yes'
|
||||||
|
|
||||||
|
if self.server_module_path:
|
||||||
|
self.test_env['NOMINATIM_DATABASE_MODULE_PATH'] = self.server_module_path
|
||||||
|
|
||||||
|
if self.website_dir is not None:
|
||||||
|
self.website_dir.cleanup()
|
||||||
|
|
||||||
|
self.website_dir = tempfile.TemporaryDirectory()
|
||||||
|
self.run_setup_script('setup-website')
|
||||||
|
|
||||||
|
|
||||||
|
def db_drop_database(self, name):
|
||||||
|
""" Drop the database with the given name.
|
||||||
|
"""
|
||||||
|
conn = self.connect_database('postgres')
|
||||||
|
conn.set_isolation_level(0)
|
||||||
|
cur = conn.cursor()
|
||||||
|
cur.execute('DROP DATABASE IF EXISTS {}'.format(name))
|
||||||
|
conn.close()
|
||||||
|
|
||||||
|
def setup_template_db(self):
|
||||||
|
""" Setup a template database that already contains common test data.
|
||||||
|
Having a template database speeds up tests considerably but at
|
||||||
|
the price that the tests sometimes run with stale data.
|
||||||
|
"""
|
||||||
|
if self.template_db_done:
|
||||||
|
return
|
||||||
|
|
||||||
|
self.template_db_done = True
|
||||||
|
|
||||||
|
if self.reuse_template:
|
||||||
|
# check that the template is there
|
||||||
|
conn = self.connect_database('postgres')
|
||||||
|
cur = conn.cursor()
|
||||||
|
cur.execute('select count(*) from pg_database where datname = %s',
|
||||||
|
(self.template_db,))
|
||||||
|
if cur.fetchone()[0] == 1:
|
||||||
|
return
|
||||||
|
conn.close()
|
||||||
|
else:
|
||||||
|
# just in case... make sure a previous table has been dropped
|
||||||
|
self.db_drop_database(self.template_db)
|
||||||
|
|
||||||
|
try:
|
||||||
|
# call the first part of database setup
|
||||||
|
self.write_nominatim_config(self.template_db)
|
||||||
|
self.run_setup_script('create-db', 'setup-db')
|
||||||
|
# remove external data to speed up indexing for tests
|
||||||
|
conn = self.connect_database(self.template_db)
|
||||||
|
cur = conn.cursor()
|
||||||
|
cur.execute("""select tablename from pg_tables
|
||||||
|
where tablename in ('gb_postcode', 'us_postcode')""")
|
||||||
|
for t in cur:
|
||||||
|
conn.cursor().execute('TRUNCATE TABLE {}'.format(t[0]))
|
||||||
|
conn.commit()
|
||||||
|
conn.close()
|
||||||
|
|
||||||
|
# execute osm2pgsql import on an empty file to get the right tables
|
||||||
|
with tempfile.NamedTemporaryFile(dir='/tmp', suffix='.xml') as fd:
|
||||||
|
fd.write(b'<osm version="0.6"></osm>')
|
||||||
|
fd.flush()
|
||||||
|
self.run_setup_script('import-data',
|
||||||
|
'ignore-errors',
|
||||||
|
'create-functions',
|
||||||
|
'create-tables',
|
||||||
|
'create-partition-tables',
|
||||||
|
'create-partition-functions',
|
||||||
|
'load-data',
|
||||||
|
'create-search-indices',
|
||||||
|
osm_file=fd.name,
|
||||||
|
osm2pgsql_cache='200')
|
||||||
|
except:
|
||||||
|
self.db_drop_database(self.template_db)
|
||||||
|
raise
|
||||||
|
|
||||||
|
|
||||||
|
def setup_api_db(self):
|
||||||
|
""" Setup a test against the API test database.
|
||||||
|
"""
|
||||||
|
self.write_nominatim_config(self.api_test_db)
|
||||||
|
|
||||||
|
def setup_unknown_db(self):
|
||||||
|
""" Setup a test against a non-existing database.
|
||||||
|
"""
|
||||||
|
self.write_nominatim_config('UNKNOWN_DATABASE_NAME')
|
||||||
|
|
||||||
|
def setup_db(self, context):
|
||||||
|
""" Setup a test against a fresh, empty test database.
|
||||||
|
"""
|
||||||
|
self.setup_template_db()
|
||||||
|
self.write_nominatim_config(self.test_db)
|
||||||
|
conn = self.connect_database(self.template_db)
|
||||||
|
conn.set_isolation_level(0)
|
||||||
|
cur = conn.cursor()
|
||||||
|
cur.execute('DROP DATABASE IF EXISTS {}'.format(self.test_db))
|
||||||
|
cur.execute('CREATE DATABASE {} TEMPLATE = {}'.format(self.test_db, self.template_db))
|
||||||
|
conn.close()
|
||||||
|
context.db = self.connect_database(self.test_db)
|
||||||
|
context.db.autocommit = True
|
||||||
|
psycopg2.extras.register_hstore(context.db, globally=False)
|
||||||
|
|
||||||
|
def teardown_db(self, context):
|
||||||
|
""" Remove the test database, if it exists.
|
||||||
|
"""
|
||||||
|
if 'db' in context:
|
||||||
|
context.db.close()
|
||||||
|
|
||||||
|
if not self.keep_scenario_db:
|
||||||
|
self.db_drop_database(self.test_db)
|
||||||
|
|
||||||
|
def reindex_placex(self, db):
|
||||||
|
""" Run the indexing step until all data in the placex has
|
||||||
|
been processed. Indexing during updates can produce more data
|
||||||
|
to index under some circumstances. That is why indexing may have
|
||||||
|
to be run multiple times.
|
||||||
|
"""
|
||||||
|
with db.cursor() as cur:
|
||||||
|
while True:
|
||||||
|
self.run_update_script('index')
|
||||||
|
|
||||||
|
cur.execute("SELECT 'a' FROM placex WHERE indexed_status != 0 LIMIT 1")
|
||||||
|
if cur.rowcount == 0:
|
||||||
|
return
|
||||||
|
|
||||||
|
def run_setup_script(self, *args, **kwargs):
|
||||||
|
""" Run the Nominatim setup script with the given arguments.
|
||||||
|
"""
|
||||||
|
self.run_nominatim_script('setup', *args, **kwargs)
|
||||||
|
|
||||||
|
def run_update_script(self, *args, **kwargs):
|
||||||
|
""" Run the Nominatim update script with the given arguments.
|
||||||
|
"""
|
||||||
|
self.run_nominatim_script('update', *args, **kwargs)
|
||||||
|
|
||||||
|
def run_nominatim_script(self, script, *args, **kwargs):
|
||||||
|
""" Run one of the Nominatim utility scripts with the given arguments.
|
||||||
|
"""
|
||||||
|
cmd = ['/usr/bin/env', 'php', '-Cq']
|
||||||
|
cmd.append((Path(self.build_dir) / 'utils' / '{}.php'.format(script)).resolve())
|
||||||
|
cmd.extend(['--' + x for x in args])
|
||||||
|
for k, v in kwargs.items():
|
||||||
|
cmd.extend(('--' + k.replace('_', '-'), str(v)))
|
||||||
|
|
||||||
|
if self.website_dir is not None:
|
||||||
|
cwd = self.website_dir.name
|
||||||
|
else:
|
||||||
|
cwd = self.build_dir
|
||||||
|
|
||||||
|
run_script(cmd, cwd=cwd, env=self.test_env)
|
||||||
|
|
||||||
|
def copy_from_place(self, db):
|
||||||
|
""" Copy data from place to the placex and location_property_osmline
|
||||||
|
tables invoking the appropriate triggers.
|
||||||
|
"""
|
||||||
|
self.run_setup_script('create-functions', 'create-partition-functions')
|
||||||
|
|
||||||
|
with db.cursor() as cur:
|
||||||
|
cur.execute("""INSERT INTO placex (osm_type, osm_id, class, type,
|
||||||
|
name, admin_level, address,
|
||||||
|
extratags, geometry)
|
||||||
|
SELECT osm_type, osm_id, class, type,
|
||||||
|
name, admin_level, address,
|
||||||
|
extratags, geometry
|
||||||
|
FROM place
|
||||||
|
WHERE not (class='place' and type='houses' and osm_type='W')""")
|
||||||
|
cur.execute("""INSERT INTO location_property_osmline (osm_id, address, linegeo)
|
||||||
|
SELECT osm_id, address, geometry
|
||||||
|
FROM place
|
||||||
|
WHERE class='place' and type='houses'
|
||||||
|
and osm_type='W'
|
||||||
|
and ST_GeometryType(geometry) = 'ST_LineString'""")
|
||||||
105
test/bdd/steps/place_inserter.py
Normal file
105
test/bdd/steps/place_inserter.py
Normal file
@@ -0,0 +1,105 @@
|
|||||||
|
"""
|
||||||
|
Helper classes for filling the place table.
|
||||||
|
"""
|
||||||
|
import random
|
||||||
|
import string
|
||||||
|
|
||||||
|
class PlaceColumn:
|
||||||
|
""" Helper class to collect contents from a behave table row and
|
||||||
|
insert it into the place table.
|
||||||
|
"""
|
||||||
|
def __init__(self, context):
|
||||||
|
self.columns = {'admin_level' : 15}
|
||||||
|
self.context = context
|
||||||
|
self.geometry = None
|
||||||
|
|
||||||
|
def add_row(self, row, force_name):
|
||||||
|
""" Parse the content from the given behave row as place column data.
|
||||||
|
"""
|
||||||
|
for name, value in zip(row.headings, row.cells):
|
||||||
|
self._add(name, value)
|
||||||
|
|
||||||
|
assert 'osm_type' in self.columns, "osm column missing"
|
||||||
|
|
||||||
|
if force_name and 'name' not in self.columns:
|
||||||
|
self._add_hstore('name', 'name',
|
||||||
|
''.join(random.choice(string.printable)
|
||||||
|
for _ in range(int(random.random()*30))))
|
||||||
|
|
||||||
|
return self
|
||||||
|
|
||||||
|
def _add(self, key, value):
|
||||||
|
if hasattr(self, '_set_key_' + key):
|
||||||
|
getattr(self, '_set_key_' + key)(value)
|
||||||
|
elif key.startswith('name+'):
|
||||||
|
self._add_hstore('name', key[5:], value)
|
||||||
|
elif key.startswith('extra+'):
|
||||||
|
self._add_hstore('extratags', key[6:], value)
|
||||||
|
elif key.startswith('addr+'):
|
||||||
|
self._add_hstore('address', key[5:], value)
|
||||||
|
elif key in ('name', 'address', 'extratags'):
|
||||||
|
self.columns[key] = eval('{' + value + '}')
|
||||||
|
else:
|
||||||
|
assert key in ('class', 'type'), "Unknown column '{}'.".format(key)
|
||||||
|
self.columns[key] = None if value == '' else value
|
||||||
|
|
||||||
|
def _set_key_name(self, value):
|
||||||
|
self._add_hstore('name', 'name', value)
|
||||||
|
|
||||||
|
def _set_key_osm(self, value):
|
||||||
|
assert value[0] in 'NRW' and value[1:].isdigit(), \
|
||||||
|
"OSM id needs to be of format <NRW><id>."
|
||||||
|
|
||||||
|
self.columns['osm_type'] = value[0]
|
||||||
|
self.columns['osm_id'] = int(value[1:])
|
||||||
|
|
||||||
|
def _set_key_admin(self, value):
|
||||||
|
self.columns['admin_level'] = int(value)
|
||||||
|
|
||||||
|
def _set_key_housenr(self, value):
|
||||||
|
if value:
|
||||||
|
self._add_hstore('address', 'housenumber', value)
|
||||||
|
|
||||||
|
def _set_key_postcode(self, value):
|
||||||
|
if value:
|
||||||
|
self._add_hstore('address', 'postcode', value)
|
||||||
|
|
||||||
|
def _set_key_street(self, value):
|
||||||
|
if value:
|
||||||
|
self._add_hstore('address', 'street', value)
|
||||||
|
|
||||||
|
def _set_key_addr_place(self, value):
|
||||||
|
if value:
|
||||||
|
self._add_hstore('address', 'place', value)
|
||||||
|
|
||||||
|
def _set_key_country(self, value):
|
||||||
|
if value:
|
||||||
|
self._add_hstore('address', 'country', value)
|
||||||
|
|
||||||
|
def _set_key_geometry(self, value):
|
||||||
|
self.geometry = self.context.osm.parse_geometry(value, self.context.scene)
|
||||||
|
assert self.geometry is not None, "Bad geometry: {}".format(value)
|
||||||
|
|
||||||
|
def _add_hstore(self, column, key, value):
|
||||||
|
if column in self.columns:
|
||||||
|
self.columns[column][key] = value
|
||||||
|
else:
|
||||||
|
self.columns[column] = {key: value}
|
||||||
|
|
||||||
|
def db_insert(self, cursor):
|
||||||
|
""" Insert the collected data into the database.
|
||||||
|
"""
|
||||||
|
if self.columns['osm_type'] == 'N' and self.geometry is None:
|
||||||
|
pt = self.context.osm.grid_node(self.columns['osm_id'])
|
||||||
|
if pt is None:
|
||||||
|
pt = (random.random()*360 - 180, random.random()*180 - 90)
|
||||||
|
|
||||||
|
self.geometry = "ST_SetSRID(ST_Point(%f, %f), 4326)" % pt
|
||||||
|
else:
|
||||||
|
assert self.geometry is not None, "Geometry missing"
|
||||||
|
|
||||||
|
query = 'INSERT INTO place ({}, geometry) values({}, {})'.format(
|
||||||
|
','.join(self.columns.keys()),
|
||||||
|
','.join(['%s' for x in range(len(self.columns))]),
|
||||||
|
self.geometry)
|
||||||
|
cursor.execute(query, list(self.columns.values()))
|
||||||
@@ -1,22 +1,18 @@
|
|||||||
""" Steps that run search queries.
|
""" Steps that run queries against the API.
|
||||||
|
|
||||||
Queries may either be run directly via PHP using the query script
|
Queries may either be run directly via PHP using the query script
|
||||||
or via the HTTP interface.
|
or via the HTTP interface using php-cgi.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
import io
|
|
||||||
import re
|
import re
|
||||||
import logging
|
import logging
|
||||||
import xml.etree.ElementTree as ET
|
|
||||||
import subprocess
|
|
||||||
from urllib.parse import urlencode
|
from urllib.parse import urlencode
|
||||||
from collections import OrderedDict
|
|
||||||
|
|
||||||
from check_functions import Almost
|
from utils import run_script
|
||||||
|
from http_responses import GenericResponse, SearchResponse, ReverseResponse, StatusResponse
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
LOG = logging.getLogger(__name__)
|
||||||
|
|
||||||
BASE_SERVER_ENV = {
|
BASE_SERVER_ENV = {
|
||||||
'HTTP_HOST' : 'localhost',
|
'HTTP_HOST' : 'localhost',
|
||||||
@@ -56,208 +52,6 @@ def compare(operator, op1, op2):
|
|||||||
else:
|
else:
|
||||||
raise Exception("unknown operator '%s'" % operator)
|
raise Exception("unknown operator '%s'" % operator)
|
||||||
|
|
||||||
class GenericResponse(object):
|
|
||||||
|
|
||||||
def match_row(self, row):
|
|
||||||
if 'ID' in row.headings:
|
|
||||||
todo = [int(row['ID'])]
|
|
||||||
else:
|
|
||||||
todo = range(len(self.result))
|
|
||||||
|
|
||||||
for i in todo:
|
|
||||||
res = self.result[i]
|
|
||||||
for h in row.headings:
|
|
||||||
if h == 'ID':
|
|
||||||
pass
|
|
||||||
elif h == 'osm':
|
|
||||||
assert res['osm_type'] == row[h][0]
|
|
||||||
assert res['osm_id'] == int(row[h][1:])
|
|
||||||
elif h == 'centroid':
|
|
||||||
x, y = row[h].split(' ')
|
|
||||||
assert Almost(float(y)) == float(res['lat'])
|
|
||||||
assert Almost(float(x)) == float(res['lon'])
|
|
||||||
elif row[h].startswith("^"):
|
|
||||||
assert h in res
|
|
||||||
assert re.fullmatch(row[h], res[h]) is not None, \
|
|
||||||
"attribute '%s': expected: '%s', got '%s'" % (h, row[h], res[h])
|
|
||||||
else:
|
|
||||||
assert h in res
|
|
||||||
assert str(res[h]) == str(row[h])
|
|
||||||
|
|
||||||
def property_list(self, prop):
|
|
||||||
return [ x[prop] for x in self.result ]
|
|
||||||
|
|
||||||
|
|
||||||
class SearchResponse(GenericResponse):
|
|
||||||
|
|
||||||
def __init__(self, page, fmt='json', errorcode=200):
|
|
||||||
self.page = page
|
|
||||||
self.format = fmt
|
|
||||||
self.errorcode = errorcode
|
|
||||||
self.result = []
|
|
||||||
self.header = dict()
|
|
||||||
|
|
||||||
if errorcode == 200:
|
|
||||||
getattr(self, 'parse_' + fmt)()
|
|
||||||
|
|
||||||
def parse_json(self):
|
|
||||||
m = re.fullmatch(r'([\w$][^(]*)\((.*)\)', self.page)
|
|
||||||
if m is None:
|
|
||||||
code = self.page
|
|
||||||
else:
|
|
||||||
code = m.group(2)
|
|
||||||
self.header['json_func'] = m.group(1)
|
|
||||||
self.result = json.JSONDecoder(object_pairs_hook=OrderedDict).decode(code)
|
|
||||||
|
|
||||||
def parse_geojson(self):
|
|
||||||
self.parse_json()
|
|
||||||
self.result = geojson_results_to_json_results(self.result)
|
|
||||||
|
|
||||||
def parse_geocodejson(self):
|
|
||||||
self.parse_geojson()
|
|
||||||
if self.result is not None:
|
|
||||||
self.result = [r['geocoding'] for r in self.result]
|
|
||||||
|
|
||||||
def parse_xml(self):
|
|
||||||
et = ET.fromstring(self.page)
|
|
||||||
|
|
||||||
self.header = dict(et.attrib)
|
|
||||||
|
|
||||||
for child in et:
|
|
||||||
assert child.tag == "place"
|
|
||||||
self.result.append(dict(child.attrib))
|
|
||||||
|
|
||||||
address = {}
|
|
||||||
for sub in child:
|
|
||||||
if sub.tag == 'extratags':
|
|
||||||
self.result[-1]['extratags'] = {}
|
|
||||||
for tag in sub:
|
|
||||||
self.result[-1]['extratags'][tag.attrib['key']] = tag.attrib['value']
|
|
||||||
elif sub.tag == 'namedetails':
|
|
||||||
self.result[-1]['namedetails'] = {}
|
|
||||||
for tag in sub:
|
|
||||||
self.result[-1]['namedetails'][tag.attrib['desc']] = tag.text
|
|
||||||
elif sub.tag in ('geokml'):
|
|
||||||
self.result[-1][sub.tag] = True
|
|
||||||
else:
|
|
||||||
address[sub.tag] = sub.text
|
|
||||||
|
|
||||||
if len(address) > 0:
|
|
||||||
self.result[-1]['address'] = address
|
|
||||||
|
|
||||||
|
|
||||||
class ReverseResponse(GenericResponse):
|
|
||||||
|
|
||||||
def __init__(self, page, fmt='json', errorcode=200):
|
|
||||||
self.page = page
|
|
||||||
self.format = fmt
|
|
||||||
self.errorcode = errorcode
|
|
||||||
self.result = []
|
|
||||||
self.header = dict()
|
|
||||||
|
|
||||||
if errorcode == 200:
|
|
||||||
getattr(self, 'parse_' + fmt)()
|
|
||||||
|
|
||||||
def parse_json(self):
|
|
||||||
m = re.fullmatch(r'([\w$][^(]*)\((.*)\)', self.page)
|
|
||||||
if m is None:
|
|
||||||
code = self.page
|
|
||||||
else:
|
|
||||||
code = m.group(2)
|
|
||||||
self.header['json_func'] = m.group(1)
|
|
||||||
self.result = [json.JSONDecoder(object_pairs_hook=OrderedDict).decode(code)]
|
|
||||||
|
|
||||||
def parse_geojson(self):
|
|
||||||
self.parse_json()
|
|
||||||
if 'error' in self.result:
|
|
||||||
return
|
|
||||||
self.result = geojson_results_to_json_results(self.result[0])
|
|
||||||
|
|
||||||
def parse_geocodejson(self):
|
|
||||||
self.parse_geojson()
|
|
||||||
if self.result is not None:
|
|
||||||
self.result = [r['geocoding'] for r in self.result]
|
|
||||||
|
|
||||||
def parse_xml(self):
|
|
||||||
et = ET.fromstring(self.page)
|
|
||||||
|
|
||||||
self.header = dict(et.attrib)
|
|
||||||
self.result = []
|
|
||||||
|
|
||||||
for child in et:
|
|
||||||
if child.tag == 'result':
|
|
||||||
assert len(self.result) == 0, "More than one result in reverse result"
|
|
||||||
self.result.append(dict(child.attrib))
|
|
||||||
elif child.tag == 'addressparts':
|
|
||||||
address = {}
|
|
||||||
for sub in child:
|
|
||||||
address[sub.tag] = sub.text
|
|
||||||
self.result[0]['address'] = address
|
|
||||||
elif child.tag == 'extratags':
|
|
||||||
self.result[0]['extratags'] = {}
|
|
||||||
for tag in child:
|
|
||||||
self.result[0]['extratags'][tag.attrib['key']] = tag.attrib['value']
|
|
||||||
elif child.tag == 'namedetails':
|
|
||||||
self.result[0]['namedetails'] = {}
|
|
||||||
for tag in child:
|
|
||||||
self.result[0]['namedetails'][tag.attrib['desc']] = tag.text
|
|
||||||
elif child.tag in ('geokml'):
|
|
||||||
self.result[0][child.tag] = True
|
|
||||||
else:
|
|
||||||
assert child.tag == 'error', \
|
|
||||||
"Unknown XML tag %s on page: %s" % (child.tag, self.page)
|
|
||||||
|
|
||||||
|
|
||||||
class DetailsResponse(GenericResponse):
|
|
||||||
|
|
||||||
def __init__(self, page, fmt='json', errorcode=200):
|
|
||||||
self.page = page
|
|
||||||
self.format = fmt
|
|
||||||
self.errorcode = errorcode
|
|
||||||
self.result = []
|
|
||||||
self.header = dict()
|
|
||||||
|
|
||||||
if errorcode == 200:
|
|
||||||
getattr(self, 'parse_' + fmt)()
|
|
||||||
|
|
||||||
def parse_json(self):
|
|
||||||
self.result = [json.JSONDecoder(object_pairs_hook=OrderedDict).decode(self.page)]
|
|
||||||
|
|
||||||
|
|
||||||
class StatusResponse(GenericResponse):
|
|
||||||
|
|
||||||
def __init__(self, page, fmt='text', errorcode=200):
|
|
||||||
self.page = page
|
|
||||||
self.format = fmt
|
|
||||||
self.errorcode = errorcode
|
|
||||||
|
|
||||||
if errorcode == 200 and fmt != 'text':
|
|
||||||
getattr(self, 'parse_' + fmt)()
|
|
||||||
|
|
||||||
def parse_json(self):
|
|
||||||
self.result = [json.JSONDecoder(object_pairs_hook=OrderedDict).decode(self.page)]
|
|
||||||
|
|
||||||
|
|
||||||
def geojson_result_to_json_result(geojson_result):
|
|
||||||
result = geojson_result['properties']
|
|
||||||
result['geojson'] = geojson_result['geometry']
|
|
||||||
if 'bbox' in geojson_result:
|
|
||||||
# bbox is minlon, minlat, maxlon, maxlat
|
|
||||||
# boundingbox is minlat, maxlat, minlon, maxlon
|
|
||||||
result['boundingbox'] = [
|
|
||||||
geojson_result['bbox'][1],
|
|
||||||
geojson_result['bbox'][3],
|
|
||||||
geojson_result['bbox'][0],
|
|
||||||
geojson_result['bbox'][2]
|
|
||||||
]
|
|
||||||
return result
|
|
||||||
|
|
||||||
|
|
||||||
def geojson_results_to_json_results(geojson_results):
|
|
||||||
if 'error' in geojson_results:
|
|
||||||
return
|
|
||||||
return list(map(geojson_result_to_json_result, geojson_results['features']))
|
|
||||||
|
|
||||||
|
|
||||||
@when(u'searching for "(?P<query>.*)"(?P<dups> with dups)?')
|
@when(u'searching for "(?P<query>.*)"(?P<dups> with dups)?')
|
||||||
def query_cmd(context, query, dups):
|
def query_cmd(context, query, dups):
|
||||||
@@ -277,14 +71,9 @@ def query_cmd(context, query, dups):
|
|||||||
if dups:
|
if dups:
|
||||||
cmd.extend(('--dedupe', '0'))
|
cmd.extend(('--dedupe', '0'))
|
||||||
|
|
||||||
proc = subprocess.Popen(cmd, cwd=context.nominatim.build_dir,
|
outp, err = run_script(cmd, cwd=context.nominatim.build_dir)
|
||||||
stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
||||||
(outp, err) = proc.communicate()
|
|
||||||
|
|
||||||
assert proc.returncode == 0, "query.php failed with message: %s\noutput: %s" % (err, outp)
|
context.response = SearchResponse(outp, 'json')
|
||||||
logger.debug("run_nominatim_script: %s\n%s\n" % (cmd, outp.decode('utf-8').replace('\\n', '\n')))
|
|
||||||
|
|
||||||
context.response = SearchResponse(outp.decode('utf-8'), 'json')
|
|
||||||
|
|
||||||
def send_api_query(endpoint, params, fmt, context):
|
def send_api_query(endpoint, params, fmt, context):
|
||||||
if fmt is not None:
|
if fmt is not None:
|
||||||
@@ -306,7 +95,7 @@ def send_api_query(endpoint, params, fmt, context):
|
|||||||
env['SCRIPT_FILENAME'] = os.path.join(env['CONTEXT_DOCUMENT_ROOT'],
|
env['SCRIPT_FILENAME'] = os.path.join(env['CONTEXT_DOCUMENT_ROOT'],
|
||||||
'%s.php' % endpoint)
|
'%s.php' % endpoint)
|
||||||
|
|
||||||
logger.debug("Environment:" + json.dumps(env, sort_keys=True, indent=2))
|
LOG.debug("Environment:" + json.dumps(env, sort_keys=True, indent=2))
|
||||||
|
|
||||||
if hasattr(context, 'http_headers'):
|
if hasattr(context, 'http_headers'):
|
||||||
env.update(context.http_headers)
|
env.update(context.http_headers)
|
||||||
@@ -326,19 +115,7 @@ def send_api_query(endpoint, params, fmt, context):
|
|||||||
for k,v in params.items():
|
for k,v in params.items():
|
||||||
cmd.append("%s=%s" % (k, v))
|
cmd.append("%s=%s" % (k, v))
|
||||||
|
|
||||||
proc = subprocess.Popen(cmd, cwd=context.nominatim.website_dir.name, env=env,
|
outp, err = run_script(cmd, cwd=context.nominatim.website_dir.name, env=env)
|
||||||
stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
||||||
|
|
||||||
(outp, err) = proc.communicate()
|
|
||||||
outp = outp.decode('utf-8')
|
|
||||||
err = err.decode("utf-8")
|
|
||||||
|
|
||||||
logger.debug("Result: \n===============================\n"
|
|
||||||
+ outp + "\n===============================\n")
|
|
||||||
|
|
||||||
assert proc.returncode == 0, \
|
|
||||||
"%s failed with message: %s" % (
|
|
||||||
os.path.basename(env['SCRIPT_FILENAME']), err)
|
|
||||||
|
|
||||||
assert len(err) == 0, "Unexpected PHP error: %s" % (err)
|
assert len(err) == 0, "Unexpected PHP error: %s" % (err)
|
||||||
|
|
||||||
@@ -371,12 +148,7 @@ def website_search_request(context, fmt, query, addr):
|
|||||||
|
|
||||||
outp, status = send_api_query('search', params, fmt, context)
|
outp, status = send_api_query('search', params, fmt, context)
|
||||||
|
|
||||||
if fmt is None or fmt == 'jsonv2 ':
|
context.response = SearchResponse(outp, fmt or 'json', status)
|
||||||
outfmt = 'json'
|
|
||||||
else:
|
|
||||||
outfmt = fmt.strip()
|
|
||||||
|
|
||||||
context.response = SearchResponse(outp, outfmt, status)
|
|
||||||
|
|
||||||
@when(u'sending (?P<fmt>\S+ )?reverse coordinates (?P<lat>.+)?,(?P<lon>.+)?')
|
@when(u'sending (?P<fmt>\S+ )?reverse coordinates (?P<lat>.+)?,(?P<lon>.+)?')
|
||||||
def website_reverse_request(context, fmt, lat, lon):
|
def website_reverse_request(context, fmt, lat, lon):
|
||||||
@@ -388,14 +160,7 @@ def website_reverse_request(context, fmt, lat, lon):
|
|||||||
|
|
||||||
outp, status = send_api_query('reverse', params, fmt, context)
|
outp, status = send_api_query('reverse', params, fmt, context)
|
||||||
|
|
||||||
if fmt is None:
|
context.response = ReverseResponse(outp, fmt or 'xml', status)
|
||||||
outfmt = 'xml'
|
|
||||||
elif fmt == 'jsonv2 ':
|
|
||||||
outfmt = 'json'
|
|
||||||
else:
|
|
||||||
outfmt = fmt.strip()
|
|
||||||
|
|
||||||
context.response = ReverseResponse(outp, outfmt, status)
|
|
||||||
|
|
||||||
@when(u'sending (?P<fmt>\S+ )?details query for (?P<query>.*)')
|
@when(u'sending (?P<fmt>\S+ )?details query for (?P<query>.*)')
|
||||||
def website_details_request(context, fmt, query):
|
def website_details_request(context, fmt, query):
|
||||||
@@ -407,42 +172,21 @@ def website_details_request(context, fmt, query):
|
|||||||
params['place_id'] = query
|
params['place_id'] = query
|
||||||
outp, status = send_api_query('details', params, fmt, context)
|
outp, status = send_api_query('details', params, fmt, context)
|
||||||
|
|
||||||
if fmt is None:
|
context.response = GenericResponse(outp, fmt or 'json', status)
|
||||||
outfmt = 'json'
|
|
||||||
else:
|
|
||||||
outfmt = fmt.strip()
|
|
||||||
|
|
||||||
context.response = DetailsResponse(outp, outfmt, status)
|
|
||||||
|
|
||||||
@when(u'sending (?P<fmt>\S+ )?lookup query for (?P<query>.*)')
|
@when(u'sending (?P<fmt>\S+ )?lookup query for (?P<query>.*)')
|
||||||
def website_lookup_request(context, fmt, query):
|
def website_lookup_request(context, fmt, query):
|
||||||
params = { 'osm_ids' : query }
|
params = { 'osm_ids' : query }
|
||||||
outp, status = send_api_query('lookup', params, fmt, context)
|
outp, status = send_api_query('lookup', params, fmt, context)
|
||||||
|
|
||||||
if fmt == 'json ':
|
context.response = SearchResponse(outp, fmt or 'xml', status)
|
||||||
outfmt = 'json'
|
|
||||||
elif fmt == 'jsonv2 ':
|
|
||||||
outfmt = 'json'
|
|
||||||
elif fmt == 'geojson ':
|
|
||||||
outfmt = 'geojson'
|
|
||||||
elif fmt == 'geocodejson ':
|
|
||||||
outfmt = 'geocodejson'
|
|
||||||
else:
|
|
||||||
outfmt = 'xml'
|
|
||||||
|
|
||||||
context.response = SearchResponse(outp, outfmt, status)
|
|
||||||
|
|
||||||
@when(u'sending (?P<fmt>\S+ )?status query')
|
@when(u'sending (?P<fmt>\S+ )?status query')
|
||||||
def website_status_request(context, fmt):
|
def website_status_request(context, fmt):
|
||||||
params = {}
|
params = {}
|
||||||
outp, status = send_api_query('status', params, fmt, context)
|
outp, status = send_api_query('status', params, fmt, context)
|
||||||
|
|
||||||
if fmt is None:
|
context.response = StatusResponse(outp, fmt or 'text', status)
|
||||||
outfmt = 'text'
|
|
||||||
else:
|
|
||||||
outfmt = fmt.strip()
|
|
||||||
|
|
||||||
context.response = StatusResponse(outp, outfmt, status)
|
|
||||||
|
|
||||||
@step(u'(?P<operator>less than|more than|exactly|at least|at most) (?P<number>\d+) results? (?:is|are) returned')
|
@step(u'(?P<operator>less than|more than|exactly|at least|at most) (?P<number>\d+) results? (?:is|are) returned')
|
||||||
def validate_result_number(context, operator, number):
|
def validate_result_number(context, operator, number):
|
||||||
336
test/bdd/steps/steps_db_ops.py
Normal file
336
test/bdd/steps/steps_db_ops.py
Normal file
@@ -0,0 +1,336 @@
|
|||||||
|
from itertools import chain
|
||||||
|
|
||||||
|
import psycopg2.extras
|
||||||
|
|
||||||
|
from place_inserter import PlaceColumn
|
||||||
|
from table_compare import NominatimID, DBRow
|
||||||
|
|
||||||
|
|
||||||
|
def check_database_integrity(context):
|
||||||
|
""" Check some generic constraints on the tables.
|
||||||
|
"""
|
||||||
|
# place_addressline should not have duplicate (place_id, address_place_id)
|
||||||
|
cur = context.db.cursor()
|
||||||
|
cur.execute("""SELECT count(*) FROM
|
||||||
|
(SELECT place_id, address_place_id, count(*) as c
|
||||||
|
FROM place_addressline GROUP BY place_id, address_place_id) x
|
||||||
|
WHERE c > 1""")
|
||||||
|
assert cur.fetchone()[0] == 0, "Duplicates found in place_addressline"
|
||||||
|
|
||||||
|
|
||||||
|
################################ GIVEN ##################################
|
||||||
|
|
||||||
|
@given("the (?P<named>named )?places")
|
||||||
|
def add_data_to_place_table(context, named):
|
||||||
|
""" Add entries into the place table. 'named places' makes sure that
|
||||||
|
the entries get a random name when none is explicitly given.
|
||||||
|
"""
|
||||||
|
with context.db.cursor() as cur:
|
||||||
|
cur.execute('ALTER TABLE place DISABLE TRIGGER place_before_insert')
|
||||||
|
for row in context.table:
|
||||||
|
PlaceColumn(context).add_row(row, named is not None).db_insert(cur)
|
||||||
|
cur.execute('ALTER TABLE place ENABLE TRIGGER place_before_insert')
|
||||||
|
|
||||||
|
@given("the relations")
|
||||||
|
def add_data_to_planet_relations(context):
|
||||||
|
""" Add entries into the osm2pgsql relation middle table. This is needed
|
||||||
|
for tests on data that looks up members.
|
||||||
|
"""
|
||||||
|
with context.db.cursor() as cur:
|
||||||
|
for r in context.table:
|
||||||
|
last_node = 0
|
||||||
|
last_way = 0
|
||||||
|
parts = []
|
||||||
|
if r['members']:
|
||||||
|
members = []
|
||||||
|
for m in r['members'].split(','):
|
||||||
|
mid = NominatimID(m)
|
||||||
|
if mid.typ == 'N':
|
||||||
|
parts.insert(last_node, int(mid.oid))
|
||||||
|
last_node += 1
|
||||||
|
last_way += 1
|
||||||
|
elif mid.typ == 'W':
|
||||||
|
parts.insert(last_way, int(mid.oid))
|
||||||
|
last_way += 1
|
||||||
|
else:
|
||||||
|
parts.append(int(mid.oid))
|
||||||
|
|
||||||
|
members.extend((mid.typ.lower() + mid.oid, mid.cls or ''))
|
||||||
|
else:
|
||||||
|
members = None
|
||||||
|
|
||||||
|
tags = chain.from_iterable([(h[5:], r[h]) for h in r.headings if h.startswith("tags+")])
|
||||||
|
|
||||||
|
cur.execute("""INSERT INTO planet_osm_rels (id, way_off, rel_off, parts, members, tags)
|
||||||
|
VALUES (%s, %s, %s, %s, %s, %s)""",
|
||||||
|
(r['id'], last_node, last_way, parts, members, list(tags)))
|
||||||
|
|
||||||
|
@given("the ways")
|
||||||
|
def add_data_to_planet_ways(context):
|
||||||
|
""" Add entries into the osm2pgsql way middle table. This is necessary for
|
||||||
|
tests on that that looks up node ids in this table.
|
||||||
|
"""
|
||||||
|
with context.db.cursor() as cur:
|
||||||
|
for r in context.table:
|
||||||
|
tags = chain.from_iterable([(h[5:], r[h]) for h in r.headings if h.startswith("tags+")])
|
||||||
|
nodes = [ int(x.strip()) for x in r['nodes'].split(',') ]
|
||||||
|
|
||||||
|
cur.execute("INSERT INTO planet_osm_ways (id, nodes, tags) VALUES (%s, %s, %s)",
|
||||||
|
(r['id'], nodes, list(tags)))
|
||||||
|
|
||||||
|
################################ WHEN ##################################
|
||||||
|
|
||||||
|
@when("importing")
|
||||||
|
def import_and_index_data_from_place_table(context):
|
||||||
|
""" Import data previously set up in the place table.
|
||||||
|
"""
|
||||||
|
context.nominatim.copy_from_place(context.db)
|
||||||
|
context.nominatim.run_setup_script('calculate-postcodes', 'index', 'index-noanalyse')
|
||||||
|
check_database_integrity(context)
|
||||||
|
|
||||||
|
@when("updating places")
|
||||||
|
def update_place_table(context):
|
||||||
|
""" Update the place table with the given data. Also runs all triggers
|
||||||
|
related to updates and reindexes the new data.
|
||||||
|
"""
|
||||||
|
context.nominatim.run_setup_script(
|
||||||
|
'create-functions', 'create-partition-functions', 'enable-diff-updates')
|
||||||
|
with context.db.cursor() as cur:
|
||||||
|
for row in context.table:
|
||||||
|
PlaceColumn(context).add_row(row, False).db_insert(cur)
|
||||||
|
|
||||||
|
context.nominatim.reindex_placex(context.db)
|
||||||
|
check_database_integrity(context)
|
||||||
|
|
||||||
|
@when("updating postcodes")
|
||||||
|
def update_postcodes(context):
|
||||||
|
""" Rerun the calculation of postcodes.
|
||||||
|
"""
|
||||||
|
context.nominatim.run_update_script('calculate-postcodes')
|
||||||
|
|
||||||
|
@when("marking for delete (?P<oids>.*)")
|
||||||
|
def delete_places(context, oids):
|
||||||
|
""" Remove entries from the place table. Multiple ids may be given
|
||||||
|
separated by commas. Also runs all triggers
|
||||||
|
related to updates and reindexes the new data.
|
||||||
|
"""
|
||||||
|
context.nominatim.run_setup_script(
|
||||||
|
'create-functions', 'create-partition-functions', 'enable-diff-updates')
|
||||||
|
with context.db.cursor() as cur:
|
||||||
|
for oid in oids.split(','):
|
||||||
|
NominatimID(oid).query_osm_id(cur, 'DELETE FROM place WHERE {}')
|
||||||
|
|
||||||
|
context.nominatim.reindex_placex(context.db)
|
||||||
|
|
||||||
|
################################ THEN ##################################
|
||||||
|
|
||||||
|
@then("(?P<table>placex|place) contains(?P<exact> exactly)?")
|
||||||
|
def check_place_contents(context, table, exact):
|
||||||
|
""" Check contents of place/placex tables. Each row represents a table row
|
||||||
|
and all data must match. Data not present in the expected table, may
|
||||||
|
be arbitry. The rows are identified via the 'object' column which must
|
||||||
|
have an identifier of the form '<NRW><osm id>[:<class>]'. When multiple
|
||||||
|
rows match (for example because 'class' was left out and there are
|
||||||
|
multiple entries for the given OSM object) then all must match. All
|
||||||
|
expected rows are expected to be present with at least one database row.
|
||||||
|
When 'exactly' is given, there must not be additional rows in the database.
|
||||||
|
"""
|
||||||
|
with context.db.cursor(cursor_factory=psycopg2.extras.DictCursor) as cur:
|
||||||
|
expected_content = set()
|
||||||
|
for row in context.table:
|
||||||
|
nid = NominatimID(row['object'])
|
||||||
|
query = 'SELECT *, ST_AsText(geometry) as geomtxt, ST_GeometryType(geometry) as geometrytype'
|
||||||
|
if table == 'placex':
|
||||||
|
query += ' ,ST_X(centroid) as cx, ST_Y(centroid) as cy'
|
||||||
|
query += " FROM %s WHERE {}" % (table, )
|
||||||
|
nid.query_osm_id(cur, query)
|
||||||
|
assert cur.rowcount > 0, "No rows found for " + row['object']
|
||||||
|
|
||||||
|
for res in cur:
|
||||||
|
if exact:
|
||||||
|
expected_content.add((res['osm_type'], res['osm_id'], res['class']))
|
||||||
|
|
||||||
|
DBRow(nid, res, context).assert_row(row, ['object'])
|
||||||
|
|
||||||
|
if exact:
|
||||||
|
cur.execute('SELECT osm_type, osm_id, class from {}'.format(table))
|
||||||
|
assert expected_content == set([(r[0], r[1], r[2]) for r in cur])
|
||||||
|
|
||||||
|
|
||||||
|
@then("(?P<table>placex|place) has no entry for (?P<oid>.*)")
|
||||||
|
def check_place_has_entry(context, table, oid):
|
||||||
|
""" Ensure that no database row for the given object exists. The ID
|
||||||
|
must be of the form '<NRW><osm id>[:<class>]'.
|
||||||
|
"""
|
||||||
|
with context.db.cursor(cursor_factory=psycopg2.extras.DictCursor) as cur:
|
||||||
|
NominatimID(oid).query_osm_id(cur, "SELECT * FROM %s where {}" % table)
|
||||||
|
assert cur.rowcount == 0, \
|
||||||
|
"Found {} entries for ID {}".format(cur.rowcount, oid)
|
||||||
|
|
||||||
|
|
||||||
|
@then("search_name contains(?P<exclude> not)?")
|
||||||
|
def check_search_name_contents(context, exclude):
|
||||||
|
""" Check contents of place/placex tables. Each row represents a table row
|
||||||
|
and all data must match. Data not present in the expected table, may
|
||||||
|
be arbitry. The rows are identified via the 'object' column which must
|
||||||
|
have an identifier of the form '<NRW><osm id>[:<class>]'. All
|
||||||
|
expected rows are expected to be present with at least one database row.
|
||||||
|
"""
|
||||||
|
with context.db.cursor(cursor_factory=psycopg2.extras.DictCursor) as cur:
|
||||||
|
for row in context.table:
|
||||||
|
nid = NominatimID(row['object'])
|
||||||
|
nid.row_by_place_id(cur, 'search_name',
|
||||||
|
['ST_X(centroid) as cx', 'ST_Y(centroid) as cy'])
|
||||||
|
assert cur.rowcount > 0, "No rows found for " + row['object']
|
||||||
|
|
||||||
|
for res in cur:
|
||||||
|
db_row = DBRow(nid, res, context)
|
||||||
|
for name, value in zip(row.headings, row.cells):
|
||||||
|
if name in ('name_vector', 'nameaddress_vector'):
|
||||||
|
items = [x.strip() for x in value.split(',')]
|
||||||
|
with context.db.cursor() as subcur:
|
||||||
|
subcur.execute(""" SELECT word_id, word_token
|
||||||
|
FROM word, (SELECT unnest(%s::TEXT[]) as term) t
|
||||||
|
WHERE word_token = make_standard_name(t.term)
|
||||||
|
and class is null and country_code is null
|
||||||
|
and operator is null
|
||||||
|
UNION
|
||||||
|
SELECT word_id, word_token
|
||||||
|
FROM word, (SELECT unnest(%s::TEXT[]) as term) t
|
||||||
|
WHERE word_token = ' ' || make_standard_name(t.term)
|
||||||
|
and class is null and country_code is null
|
||||||
|
and operator is null
|
||||||
|
""",
|
||||||
|
(list(filter(lambda x: not x.startswith('#'), items)),
|
||||||
|
list(filter(lambda x: x.startswith('#'), items))))
|
||||||
|
if not exclude:
|
||||||
|
assert subcur.rowcount >= len(items), \
|
||||||
|
"No word entry found for {}. Entries found: {!s}".format(value, subcur.rowcount)
|
||||||
|
for wid in subcur:
|
||||||
|
present = wid[0] in res[name]
|
||||||
|
if exclude:
|
||||||
|
assert not present, "Found term for {}/{}: {}".format(row['object'], name, wid[1])
|
||||||
|
else:
|
||||||
|
assert present, "Missing term for {}/{}: {}".fromat(row['object'], name, wid[1])
|
||||||
|
elif name != 'object':
|
||||||
|
assert db_row.contains(name, value), db_row.assert_msg(name, value)
|
||||||
|
|
||||||
|
@then("search_name has no entry for (?P<oid>.*)")
|
||||||
|
def check_search_name_has_entry(context, oid):
|
||||||
|
""" Check that there is noentry in the search_name table for the given
|
||||||
|
objects. IDs are in format '<NRW><osm id>[:<class>]'.
|
||||||
|
"""
|
||||||
|
with context.db.cursor(cursor_factory=psycopg2.extras.DictCursor) as cur:
|
||||||
|
NominatimID(oid).row_by_place_id(cur, 'search_name')
|
||||||
|
|
||||||
|
assert cur.rowcount == 0, \
|
||||||
|
"Found {} entries for ID {}".format(cur.rowcount, oid)
|
||||||
|
|
||||||
|
@then("location_postcode contains exactly")
|
||||||
|
def check_location_postcode(context):
|
||||||
|
""" Check full contents for location_postcode table. Each row represents a table row
|
||||||
|
and all data must match. Data not present in the expected table, may
|
||||||
|
be arbitry. The rows are identified via 'country' and 'postcode' columns.
|
||||||
|
All rows must be present as excepted and there must not be additional
|
||||||
|
rows.
|
||||||
|
"""
|
||||||
|
with context.db.cursor(cursor_factory=psycopg2.extras.DictCursor) as cur:
|
||||||
|
cur.execute("SELECT *, ST_AsText(geometry) as geomtxt FROM location_postcode")
|
||||||
|
assert cur.rowcount == len(list(context.table)), \
|
||||||
|
"Postcode table has {} rows, expected {}.".foramt(cur.rowcount, len(list(context.table)))
|
||||||
|
|
||||||
|
results = {}
|
||||||
|
for row in cur:
|
||||||
|
key = (row['country_code'], row['postcode'])
|
||||||
|
assert key not in results, "Postcode table has duplicate entry: {}".format(row)
|
||||||
|
results[key] = DBRow((row['country_code'],row['postcode']), row, context)
|
||||||
|
|
||||||
|
for row in context.table:
|
||||||
|
db_row = results.get((row['country'],row['postcode']))
|
||||||
|
assert db_row is not None, \
|
||||||
|
"Missing row for country '{r['country']}' postcode '{r['postcode']}'.".format(r=row)
|
||||||
|
|
||||||
|
db_row.assert_row(row, ('country', 'postcode'))
|
||||||
|
|
||||||
|
@then("word contains(?P<exclude> not)?")
|
||||||
|
def check_word_table(context, exclude):
|
||||||
|
""" Check the contents of the word table. Each row represents a table row
|
||||||
|
and all data must match. Data not present in the expected table, may
|
||||||
|
be arbitry. The rows are identified via all given columns.
|
||||||
|
"""
|
||||||
|
with context.db.cursor(cursor_factory=psycopg2.extras.DictCursor) as cur:
|
||||||
|
for row in context.table:
|
||||||
|
wheres = ' AND '.join(["{} = %s".format(h) for h in row.headings])
|
||||||
|
cur.execute("SELECT * from word WHERE " + wheres, list(row.cells))
|
||||||
|
if exclude:
|
||||||
|
assert cur.rowcount == 0, "Row still in word table: %s" % '/'.join(values)
|
||||||
|
else:
|
||||||
|
assert cur.rowcount > 0, "Row not in word table: %s" % '/'.join(values)
|
||||||
|
|
||||||
|
@then("place_addressline contains")
|
||||||
|
def check_place_addressline(context):
|
||||||
|
""" Check the contents of the place_addressline table. Each row represents
|
||||||
|
a table row and all data must match. Data not present in the expected
|
||||||
|
table, may be arbitry. The rows are identified via the 'object' column,
|
||||||
|
representing the addressee and the 'address' column, representing the
|
||||||
|
address item.
|
||||||
|
"""
|
||||||
|
with context.db.cursor(cursor_factory=psycopg2.extras.DictCursor) as cur:
|
||||||
|
for row in context.table:
|
||||||
|
nid = NominatimID(row['object'])
|
||||||
|
pid = nid.get_place_id(cur)
|
||||||
|
apid = NominatimID(row['address']).get_place_id(cur)
|
||||||
|
cur.execute(""" SELECT * FROM place_addressline
|
||||||
|
WHERE place_id = %s AND address_place_id = %s""",
|
||||||
|
(pid, apid))
|
||||||
|
assert cur.rowcount > 0, \
|
||||||
|
"No rows found for place %s and address %s" % (row['object'], row['address'])
|
||||||
|
|
||||||
|
for res in cur:
|
||||||
|
DBRow(nid, res, context).assert_row(row, ('address', 'object'))
|
||||||
|
|
||||||
|
@then("place_addressline doesn't contain")
|
||||||
|
def check_place_addressline_exclude(context):
|
||||||
|
""" Check that the place_addressline doesn't contain any entries for the
|
||||||
|
given addressee/address item pairs.
|
||||||
|
"""
|
||||||
|
with context.db.cursor(cursor_factory=psycopg2.extras.DictCursor) as cur:
|
||||||
|
for row in context.table:
|
||||||
|
pid = NominatimID(row['object']).get_place_id(cur)
|
||||||
|
apid = NominatimID(row['address']).get_place_id(cur)
|
||||||
|
cur.execute(""" SELECT * FROM place_addressline
|
||||||
|
WHERE place_id = %s AND address_place_id = %s""",
|
||||||
|
(pid, apid))
|
||||||
|
assert cur.rowcount == 0, \
|
||||||
|
"Row found for place %s and address %s" % (row['object'], row['address'])
|
||||||
|
|
||||||
|
@then("W(?P<oid>\d+) expands to(?P<neg> no)? interpolation")
|
||||||
|
def check_location_property_osmline(context, oid, neg):
|
||||||
|
""" Check that the given way is present in the interpolation table.
|
||||||
|
"""
|
||||||
|
with context.db.cursor(cursor_factory=psycopg2.extras.DictCursor) as cur:
|
||||||
|
cur.execute("""SELECT *, ST_AsText(linegeo) as geomtxt
|
||||||
|
FROM location_property_osmline
|
||||||
|
WHERE osm_id = %s AND startnumber IS NOT NULL""",
|
||||||
|
(oid, ))
|
||||||
|
|
||||||
|
if neg:
|
||||||
|
assert cur.rowcount == 0, "Interpolation found for way {}.".format(oid)
|
||||||
|
return
|
||||||
|
|
||||||
|
todo = list(range(len(list(context.table))))
|
||||||
|
for res in cur:
|
||||||
|
for i in todo:
|
||||||
|
row = context.table[i]
|
||||||
|
if (int(row['start']) == res['startnumber']
|
||||||
|
and int(row['end']) == res['endnumber']):
|
||||||
|
todo.remove(i)
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
assert False, "Unexpected row " + str(res)
|
||||||
|
|
||||||
|
DBRow(oid, res, context).assert_row(row, ('start', 'end'))
|
||||||
|
|
||||||
|
assert not todo
|
||||||
|
|
||||||
|
|
||||||
@@ -1,34 +1,45 @@
|
|||||||
import subprocess
|
|
||||||
import tempfile
|
import tempfile
|
||||||
import random
|
import random
|
||||||
import os
|
import os
|
||||||
|
|
||||||
|
def write_opl_file(opl, grid):
|
||||||
|
""" Create a temporary OSM file from OPL and return the file name. It is
|
||||||
|
the responsibility of the caller to delete the file again.
|
||||||
|
|
||||||
|
Node with missing coordinates, can retrieve their coordinates from
|
||||||
|
a supplied grid. Failing that a random coordinate is assigned.
|
||||||
|
"""
|
||||||
|
with tempfile.NamedTemporaryFile(suffix='.opl', delete=False) as fd:
|
||||||
|
for line in opl.splitlines():
|
||||||
|
if line.startswith('n') and line.find(' x') < 0:
|
||||||
|
coord = grid.grid_node(int(line[1:].split(' ')[0]))
|
||||||
|
if coord is None:
|
||||||
|
coord = (random.random() * 360 - 180,
|
||||||
|
random.random() * 180 - 90)
|
||||||
|
line += " x%f y%f" % coord
|
||||||
|
fd.write(line.encode('utf-8'))
|
||||||
|
fd.write(b'\n')
|
||||||
|
|
||||||
|
return fd.name
|
||||||
|
|
||||||
|
@given(u'the scene (?P<scene>.+)')
|
||||||
|
def set_default_scene(context, scene):
|
||||||
|
context.scene = scene
|
||||||
|
|
||||||
@given(u'the ([0-9.]+ )?grid')
|
@given(u'the ([0-9.]+ )?grid')
|
||||||
def define_node_grid(context, grid_step):
|
def define_node_grid(context, grid_step):
|
||||||
"""
|
"""
|
||||||
Define a grid of node positions.
|
Define a grid of node positions.
|
||||||
|
Use a table to define the grid. The nodes must be integer ids. Optionally
|
||||||
|
you can give the grid distance. The default is 0.00001 degrees.
|
||||||
"""
|
"""
|
||||||
if grid_step is not None:
|
if grid_step is not None:
|
||||||
grid_step = float(grid_step.strip())
|
grid_step = float(grid_step.strip())
|
||||||
else:
|
else:
|
||||||
grid_step = 0.00001
|
grid_step = 0.00001
|
||||||
|
|
||||||
context.osm.clear_grid()
|
context.osm.set_grid([context.table.headings] + [list(h) for h in context.table],
|
||||||
|
grid_step)
|
||||||
i = 0
|
|
||||||
for h in context.table.headings:
|
|
||||||
if h.isdigit():
|
|
||||||
context.osm.add_grid_node(int(h), 0, i)
|
|
||||||
i += grid_step
|
|
||||||
|
|
||||||
x = grid_step
|
|
||||||
for r in context.table:
|
|
||||||
y = 0
|
|
||||||
for h in r:
|
|
||||||
if h.isdigit():
|
|
||||||
context.osm.add_grid_node(int(h), x, y)
|
|
||||||
y += grid_step
|
|
||||||
x += grid_step
|
|
||||||
|
|
||||||
|
|
||||||
@when(u'loading osm data')
|
@when(u'loading osm data')
|
||||||
@@ -39,21 +50,11 @@ def load_osm_file(context):
|
|||||||
|
|
||||||
The data is expected as attached text in OPL format.
|
The data is expected as attached text in OPL format.
|
||||||
"""
|
"""
|
||||||
# create a OSM file in /tmp and import it
|
# create an OSM file and import it
|
||||||
with tempfile.NamedTemporaryFile(dir='/tmp', suffix='.opl', delete=False) as fd:
|
fname = write_opl_file(context.text, context.osm)
|
||||||
fname = fd.name
|
|
||||||
for line in context.text.splitlines():
|
|
||||||
if line.startswith('n') and line.find(' x') < 0:
|
|
||||||
coord = context.osm.grid_node(int(line[1:].split(' ')[0]))
|
|
||||||
if coord is None:
|
|
||||||
coord = (random.random() * 360 - 180,
|
|
||||||
random.random() * 180 - 90)
|
|
||||||
line += " x%f y%f" % coord
|
|
||||||
fd.write(line.encode('utf-8'))
|
|
||||||
fd.write(b'\n')
|
|
||||||
|
|
||||||
context.nominatim.run_setup_script('import-data', osm_file=fname,
|
context.nominatim.run_setup_script('import-data', osm_file=fname,
|
||||||
osm2pgsql_cache=300)
|
osm2pgsql_cache=300)
|
||||||
|
os.remove(fname)
|
||||||
|
|
||||||
### reintroduce the triggers/indexes we've lost by having osm2pgsql set up place again
|
### reintroduce the triggers/indexes we've lost by having osm2pgsql set up place again
|
||||||
cur = context.db.cursor()
|
cur = context.db.cursor()
|
||||||
@@ -64,7 +65,6 @@ def load_osm_file(context):
|
|||||||
cur.execute("""CREATE UNIQUE INDEX idx_place_osm_unique on place using btree(osm_id,osm_type,class,type)""")
|
cur.execute("""CREATE UNIQUE INDEX idx_place_osm_unique on place using btree(osm_id,osm_type,class,type)""")
|
||||||
context.db.commit()
|
context.db.commit()
|
||||||
|
|
||||||
os.remove(fname)
|
|
||||||
|
|
||||||
@when(u'updating osm data')
|
@when(u'updating osm data')
|
||||||
def update_from_osm_file(context):
|
def update_from_osm_file(context):
|
||||||
@@ -74,30 +74,12 @@ def update_from_osm_file(context):
|
|||||||
|
|
||||||
The data is expected as attached text in OPL format.
|
The data is expected as attached text in OPL format.
|
||||||
"""
|
"""
|
||||||
context.nominatim.run_setup_script('create-functions', 'create-partition-functions')
|
context.nominatim.copy_from_place(context.db)
|
||||||
|
|
||||||
cur = context.db.cursor()
|
|
||||||
cur.execute("""insert into placex (osm_type, osm_id, class, type, name, admin_level, address, extratags, geometry)
|
|
||||||
select osm_type, osm_id, class, type, name, admin_level, address, extratags, geometry from place""")
|
|
||||||
cur.execute(
|
|
||||||
"""insert into location_property_osmline (osm_id, address, linegeo)
|
|
||||||
SELECT osm_id, address, geometry from place
|
|
||||||
WHERE class='place' and type='houses' and osm_type='W'
|
|
||||||
and ST_GeometryType(geometry) = 'ST_LineString'""")
|
|
||||||
context.db.commit()
|
|
||||||
context.nominatim.run_setup_script('index', 'index-noanalyse')
|
context.nominatim.run_setup_script('index', 'index-noanalyse')
|
||||||
context.nominatim.run_setup_script('create-functions', 'create-partition-functions',
|
context.nominatim.run_setup_script('create-functions', 'create-partition-functions',
|
||||||
'enable-diff-updates')
|
'enable-diff-updates')
|
||||||
|
|
||||||
# create a OSM file in /tmp and import it
|
# create an OSM file and import it
|
||||||
with tempfile.NamedTemporaryFile(dir='/tmp', suffix='.opl', delete=False) as fd:
|
fname = write_opl_file(context.text, context.osm)
|
||||||
fname = fd.name
|
|
||||||
for line in context.text.splitlines():
|
|
||||||
if line.startswith('n') and line.find(' x') < 0:
|
|
||||||
line += " x%d y%d" % (random.random() * 360 - 180,
|
|
||||||
random.random() * 180 - 90)
|
|
||||||
fd.write(line.encode('utf-8'))
|
|
||||||
fd.write(b'\n')
|
|
||||||
|
|
||||||
context.nominatim.run_update_script(import_diff=fname)
|
context.nominatim.run_update_script(import_diff=fname)
|
||||||
os.remove(fname)
|
os.remove(fname)
|
||||||
209
test/bdd/steps/table_compare.py
Normal file
209
test/bdd/steps/table_compare.py
Normal file
@@ -0,0 +1,209 @@
|
|||||||
|
"""
|
||||||
|
Functions to facilitate accessing and comparing the content of DB tables.
|
||||||
|
"""
|
||||||
|
import re
|
||||||
|
import json
|
||||||
|
|
||||||
|
from steps.check_functions import Almost
|
||||||
|
|
||||||
|
ID_REGEX = re.compile(r"(?P<typ>[NRW])(?P<oid>\d+)(:(?P<cls>\w+))?")
|
||||||
|
|
||||||
|
class NominatimID:
|
||||||
|
""" Splits a unique identifier for places into its components.
|
||||||
|
As place_ids cannot be used for testing, we use a unique
|
||||||
|
identifier instead that is of the form <osmtype><osmid>[:<class>].
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, oid):
|
||||||
|
self.typ = self.oid = self.cls = None
|
||||||
|
|
||||||
|
if oid is not None:
|
||||||
|
m = ID_REGEX.fullmatch(oid)
|
||||||
|
assert m is not None, \
|
||||||
|
"ID '{}' not of form <osmtype><osmid>[:<class>]".format(oid)
|
||||||
|
|
||||||
|
self.typ = m.group('typ')
|
||||||
|
self.oid = m.group('oid')
|
||||||
|
self.cls = m.group('cls')
|
||||||
|
|
||||||
|
def __str__(self):
|
||||||
|
if self.cls is None:
|
||||||
|
return self.typ + self.oid
|
||||||
|
|
||||||
|
return '{self.typ}{self.oid}:{self.cls}'.format(self=self)
|
||||||
|
|
||||||
|
def query_osm_id(self, cur, query):
|
||||||
|
""" Run a query on cursor `cur` using osm ID, type and class. The
|
||||||
|
`query` string must contain exactly one placeholder '{}' where
|
||||||
|
the 'where' query should go.
|
||||||
|
"""
|
||||||
|
where = 'osm_type = %s and osm_id = %s'
|
||||||
|
params = [self.typ, self. oid]
|
||||||
|
|
||||||
|
if self.cls is not None:
|
||||||
|
where += ' and class = %s'
|
||||||
|
params.append(self.cls)
|
||||||
|
|
||||||
|
cur.execute(query.format(where), params)
|
||||||
|
|
||||||
|
def row_by_place_id(self, cur, table, extra_columns=None):
|
||||||
|
""" Get a row by place_id from the given table using cursor `cur`.
|
||||||
|
extra_columns may contain a list additional elements for the select
|
||||||
|
part of the query.
|
||||||
|
"""
|
||||||
|
pid = self.get_place_id(cur)
|
||||||
|
query = "SELECT {} FROM {} WHERE place_id = %s".format(
|
||||||
|
','.join(['*'] + (extra_columns or [])), table)
|
||||||
|
cur.execute(query, (pid, ))
|
||||||
|
|
||||||
|
def get_place_id(self, cur):
|
||||||
|
""" Look up the place id for the ID. Throws an assertion if the ID
|
||||||
|
is not unique.
|
||||||
|
"""
|
||||||
|
self.query_osm_id(cur, "SELECT place_id FROM placex WHERE {}")
|
||||||
|
assert cur.rowcount == 1, \
|
||||||
|
"Place ID {!s} not unique. Found {} entries.".format(self, cur.rowcount)
|
||||||
|
|
||||||
|
return cur.fetchone()[0]
|
||||||
|
|
||||||
|
|
||||||
|
class DBRow:
|
||||||
|
""" Represents a row from a database and offers comparison functions.
|
||||||
|
"""
|
||||||
|
def __init__(self, nid, db_row, context):
|
||||||
|
self.nid = nid
|
||||||
|
self.db_row = db_row
|
||||||
|
self.context = context
|
||||||
|
|
||||||
|
def assert_row(self, row, exclude_columns):
|
||||||
|
""" Check that all columns of the given behave row are contained
|
||||||
|
in the database row. Exclude behave rows with the names given
|
||||||
|
in the `exclude_columns` list.
|
||||||
|
"""
|
||||||
|
for name, value in zip(row.headings, row.cells):
|
||||||
|
if name not in exclude_columns:
|
||||||
|
assert self.contains(name, value), self.assert_msg(name, value)
|
||||||
|
|
||||||
|
def contains(self, name, expected):
|
||||||
|
""" Check that the DB row contains a column `name` with the given value.
|
||||||
|
"""
|
||||||
|
if '+' in name:
|
||||||
|
column, field = name.split('+', 1)
|
||||||
|
return self._contains_hstore_value(column, field, expected)
|
||||||
|
|
||||||
|
if name == 'geometry':
|
||||||
|
return self._has_geometry(expected)
|
||||||
|
|
||||||
|
if name not in self.db_row:
|
||||||
|
return False
|
||||||
|
|
||||||
|
actual = self.db_row[name]
|
||||||
|
|
||||||
|
if expected == '-':
|
||||||
|
return actual is None
|
||||||
|
|
||||||
|
if name == 'name' and ':' not in expected:
|
||||||
|
return self._compare_column(actual[name], expected)
|
||||||
|
|
||||||
|
if 'place_id' in name:
|
||||||
|
return self._compare_place_id(actual, expected)
|
||||||
|
|
||||||
|
if name == 'centroid':
|
||||||
|
return self._has_centroid(expected)
|
||||||
|
|
||||||
|
return self._compare_column(actual, expected)
|
||||||
|
|
||||||
|
def _contains_hstore_value(self, column, field, expected):
|
||||||
|
if column == 'addr':
|
||||||
|
column = 'address'
|
||||||
|
|
||||||
|
if column not in self.db_row:
|
||||||
|
return False
|
||||||
|
|
||||||
|
if expected == '-':
|
||||||
|
return self.db_row[column] is None or field not in self.db_row[column]
|
||||||
|
|
||||||
|
if self.db_row[column] is None:
|
||||||
|
return False
|
||||||
|
|
||||||
|
return self._compare_column(self.db_row[column].get(field), expected)
|
||||||
|
|
||||||
|
def _compare_column(self, actual, expected):
|
||||||
|
if isinstance(actual, dict):
|
||||||
|
return actual == eval('{' + expected + '}')
|
||||||
|
|
||||||
|
return str(actual) == expected
|
||||||
|
|
||||||
|
def _compare_place_id(self, actual, expected):
|
||||||
|
if expected == '0':
|
||||||
|
return actual == 0
|
||||||
|
|
||||||
|
with self.context.db.cursor() as cur:
|
||||||
|
return NominatimID(expected).get_place_id(cur) == actual
|
||||||
|
|
||||||
|
def _has_centroid(self, expected):
|
||||||
|
if expected == 'in geometry':
|
||||||
|
with self.context.db.cursor() as cur:
|
||||||
|
cur.execute("""SELECT ST_Within(ST_SetSRID(ST_Point({cx}, {cy}), 4326),
|
||||||
|
ST_SetSRID('{geomtxt}'::geometry, 4326))""".format(**self.db_row))
|
||||||
|
return cur.fetchone()[0]
|
||||||
|
|
||||||
|
x, y = expected.split(' ')
|
||||||
|
return Almost(float(x)) == self.db_row['cx'] and Almost(float(y)) == self.db_row['cy']
|
||||||
|
|
||||||
|
def _has_geometry(self, expected):
|
||||||
|
geom = self.context.osm.parse_geometry(expected, self.context.scene)
|
||||||
|
with self.context.db.cursor() as cur:
|
||||||
|
cur.execute("""SELECT ST_Equals(ST_SnapToGrid({}, 0.00001, 0.00001),
|
||||||
|
ST_SnapToGrid(ST_SetSRID('{}'::geometry, 4326), 0.00001, 0.00001))""".format(
|
||||||
|
geom, self.db_row['geomtxt']))
|
||||||
|
return cur.fetchone()[0]
|
||||||
|
|
||||||
|
def assert_msg(self, name, value):
|
||||||
|
""" Return a string with an informative message for a failed compare.
|
||||||
|
"""
|
||||||
|
msg = "\nBad column '{}' in row '{!s}'.".format(name, self.nid)
|
||||||
|
actual = self._get_actual(name)
|
||||||
|
if actual is not None:
|
||||||
|
msg += " Expected: {}, got: {}.".format(value, actual)
|
||||||
|
else:
|
||||||
|
msg += " No such column."
|
||||||
|
|
||||||
|
return msg + "\nFull DB row: {}".format(json.dumps(dict(self.db_row), indent=4, default=str))
|
||||||
|
|
||||||
|
def _get_actual(self, name):
|
||||||
|
if '+' in name:
|
||||||
|
column, field = name.split('+', 1)
|
||||||
|
if column == 'addr':
|
||||||
|
column = 'address'
|
||||||
|
return (self.db_row.get(column) or {}).get(field)
|
||||||
|
|
||||||
|
if name == 'geometry':
|
||||||
|
return self.db_row['geomtxt']
|
||||||
|
|
||||||
|
if name not in self.db_row:
|
||||||
|
return None
|
||||||
|
|
||||||
|
if name == 'centroid':
|
||||||
|
return "POINT({cx} {cy})".format(**self.db_row)
|
||||||
|
|
||||||
|
actual = self.db_row[name]
|
||||||
|
|
||||||
|
if 'place_id' in name:
|
||||||
|
if actual is None:
|
||||||
|
return '<null>'
|
||||||
|
|
||||||
|
if actual == 0:
|
||||||
|
return "place ID 0"
|
||||||
|
|
||||||
|
with self.context.db.cursor() as cur:
|
||||||
|
cur.execute("""SELECT osm_type, osm_id, class
|
||||||
|
FROM placex WHERE place_id = %s""",
|
||||||
|
(actual, ))
|
||||||
|
|
||||||
|
if cur.rowcount == 1:
|
||||||
|
return "{0[0]}{0[1]}:{0[2]}".format(cur.fetchone())
|
||||||
|
|
||||||
|
return "[place ID {} not found]".format(actual)
|
||||||
|
|
||||||
|
return actual
|
||||||
22
test/bdd/steps/utils.py
Normal file
22
test/bdd/steps/utils.py
Normal file
@@ -0,0 +1,22 @@
|
|||||||
|
"""
|
||||||
|
Various smaller helps for step execution.
|
||||||
|
"""
|
||||||
|
import logging
|
||||||
|
import subprocess
|
||||||
|
|
||||||
|
LOG = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
def run_script(cmd, **kwargs):
|
||||||
|
""" Run the given command, check that it is successful and output
|
||||||
|
when necessary.
|
||||||
|
"""
|
||||||
|
proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE,
|
||||||
|
**kwargs)
|
||||||
|
(outp, outerr) = proc.communicate()
|
||||||
|
outp = outp.decode('utf-8')
|
||||||
|
outerr = outerr.decode('utf-8').replace('\\n', '\n')
|
||||||
|
LOG.debug("Run command: %s\n%s\n%s", cmd, outp, outerr)
|
||||||
|
|
||||||
|
assert proc.returncode == 0, "Script '{}' failed:\n{}\n{}\n".format(cmd[0], outp, outerr)
|
||||||
|
|
||||||
|
return outp, outerr
|
||||||
Reference in New Issue
Block a user