mirror of
https://github.com/osm-search/Nominatim.git
synced 2026-02-15 19:07:58 +00:00
replace TokenType enum with simple char constants
This commit is contained in:
@@ -22,42 +22,42 @@ def mktoken(tid: int):
|
||||
lookup_word='foo')
|
||||
|
||||
|
||||
@pytest.mark.parametrize('ptype,ttype', [('NONE', 'WORD'),
|
||||
('AMENITY', 'QUALIFIER'),
|
||||
('STREET', 'PARTIAL'),
|
||||
('CITY', 'WORD'),
|
||||
('COUNTRY', 'COUNTRY'),
|
||||
('POSTCODE', 'POSTCODE')])
|
||||
@pytest.mark.parametrize('ptype,ttype', [('NONE', 'W'),
|
||||
('AMENITY', 'Q'),
|
||||
('STREET', 'w'),
|
||||
('CITY', 'W'),
|
||||
('COUNTRY', 'C'),
|
||||
('POSTCODE', 'P')])
|
||||
def test_phrase_compatible(ptype, ttype):
|
||||
assert query.PhraseType[ptype].compatible_with(query.TokenType[ttype], False)
|
||||
assert query.PhraseType[ptype].compatible_with(ttype, False)
|
||||
|
||||
|
||||
@pytest.mark.parametrize('ptype', ['COUNTRY', 'POSTCODE'])
|
||||
def test_phrase_incompatible(ptype):
|
||||
assert not query.PhraseType[ptype].compatible_with(query.TokenType.PARTIAL, True)
|
||||
assert not query.PhraseType[ptype].compatible_with(query.TOKEN_PARTIAL, True)
|
||||
|
||||
|
||||
def test_query_node_empty():
|
||||
qn = query.QueryNode(query.BREAK_PHRASE, query.PhraseType.NONE)
|
||||
|
||||
assert not qn.has_tokens(3, query.TokenType.PARTIAL)
|
||||
assert qn.get_tokens(3, query.TokenType.WORD) is None
|
||||
assert not qn.has_tokens(3, query.TOKEN_PARTIAL)
|
||||
assert qn.get_tokens(3, query.TOKEN_WORD) is None
|
||||
|
||||
|
||||
def test_query_node_with_content():
|
||||
qn = query.QueryNode(query.BREAK_PHRASE, query.PhraseType.NONE)
|
||||
qn.starting.append(query.TokenList(2, query.TokenType.PARTIAL, [mktoken(100), mktoken(101)]))
|
||||
qn.starting.append(query.TokenList(2, query.TokenType.WORD, [mktoken(1000)]))
|
||||
qn.starting.append(query.TokenList(2, query.TOKEN_PARTIAL, [mktoken(100), mktoken(101)]))
|
||||
qn.starting.append(query.TokenList(2, query.TOKEN_WORD, [mktoken(1000)]))
|
||||
|
||||
assert not qn.has_tokens(3, query.TokenType.PARTIAL)
|
||||
assert not qn.has_tokens(2, query.TokenType.COUNTRY)
|
||||
assert qn.has_tokens(2, query.TokenType.PARTIAL)
|
||||
assert qn.has_tokens(2, query.TokenType.WORD)
|
||||
assert not qn.has_tokens(3, query.TOKEN_PARTIAL)
|
||||
assert not qn.has_tokens(2, query.TOKEN_COUNTRY)
|
||||
assert qn.has_tokens(2, query.TOKEN_PARTIAL)
|
||||
assert qn.has_tokens(2, query.TOKEN_WORD)
|
||||
|
||||
assert qn.get_tokens(3, query.TokenType.PARTIAL) is None
|
||||
assert qn.get_tokens(2, query.TokenType.COUNTRY) is None
|
||||
assert len(qn.get_tokens(2, query.TokenType.PARTIAL)) == 2
|
||||
assert len(qn.get_tokens(2, query.TokenType.WORD)) == 1
|
||||
assert qn.get_tokens(3, query.TOKEN_PARTIAL) is None
|
||||
assert qn.get_tokens(2, query.TOKEN_COUNTRY) is None
|
||||
assert len(qn.get_tokens(2, query.TOKEN_PARTIAL)) == 2
|
||||
assert len(qn.get_tokens(2, query.TOKEN_WORD)) == 1
|
||||
|
||||
|
||||
def test_query_struct_empty():
|
||||
@@ -73,13 +73,13 @@ def test_query_struct_with_tokens():
|
||||
|
||||
assert q.num_token_slots() == 2
|
||||
|
||||
q.add_token(query.TokenRange(0, 1), query.TokenType.PARTIAL, mktoken(1))
|
||||
q.add_token(query.TokenRange(1, 2), query.TokenType.PARTIAL, mktoken(2))
|
||||
q.add_token(query.TokenRange(1, 2), query.TokenType.WORD, mktoken(99))
|
||||
q.add_token(query.TokenRange(1, 2), query.TokenType.WORD, mktoken(98))
|
||||
q.add_token(query.TokenRange(0, 1), query.TOKEN_PARTIAL, mktoken(1))
|
||||
q.add_token(query.TokenRange(1, 2), query.TOKEN_PARTIAL, mktoken(2))
|
||||
q.add_token(query.TokenRange(1, 2), query.TOKEN_WORD, mktoken(99))
|
||||
q.add_token(query.TokenRange(1, 2), query.TOKEN_WORD, mktoken(98))
|
||||
|
||||
assert q.get_tokens(query.TokenRange(0, 2), query.TokenType.WORD) == []
|
||||
assert len(q.get_tokens(query.TokenRange(1, 2), query.TokenType.WORD)) == 2
|
||||
assert q.get_tokens(query.TokenRange(0, 2), query.TOKEN_WORD) == []
|
||||
assert len(q.get_tokens(query.TokenRange(1, 2), query.TOKEN_WORD)) == 2
|
||||
|
||||
partials = q.get_partials_list(query.TokenRange(0, 2))
|
||||
|
||||
@@ -95,24 +95,24 @@ def test_query_struct_incompatible_token():
|
||||
q.add_node(query.BREAK_WORD, query.PhraseType.COUNTRY)
|
||||
q.add_node(query.BREAK_END, query.PhraseType.NONE)
|
||||
|
||||
q.add_token(query.TokenRange(0, 1), query.TokenType.PARTIAL, mktoken(1))
|
||||
q.add_token(query.TokenRange(1, 2), query.TokenType.COUNTRY, mktoken(100))
|
||||
q.add_token(query.TokenRange(0, 1), query.TOKEN_PARTIAL, mktoken(1))
|
||||
q.add_token(query.TokenRange(1, 2), query.TOKEN_COUNTRY, mktoken(100))
|
||||
|
||||
assert q.get_tokens(query.TokenRange(0, 1), query.TokenType.PARTIAL) == []
|
||||
assert len(q.get_tokens(query.TokenRange(1, 2), query.TokenType.COUNTRY)) == 1
|
||||
assert q.get_tokens(query.TokenRange(0, 1), query.TOKEN_PARTIAL) == []
|
||||
assert len(q.get_tokens(query.TokenRange(1, 2), query.TOKEN_COUNTRY)) == 1
|
||||
|
||||
|
||||
def test_query_struct_amenity_single_word():
|
||||
q = query.QueryStruct([query.Phrase(query.PhraseType.AMENITY, 'bar')])
|
||||
q.add_node(query.BREAK_END, query.PhraseType.NONE)
|
||||
|
||||
q.add_token(query.TokenRange(0, 1), query.TokenType.PARTIAL, mktoken(1))
|
||||
q.add_token(query.TokenRange(0, 1), query.TokenType.NEAR_ITEM, mktoken(2))
|
||||
q.add_token(query.TokenRange(0, 1), query.TokenType.QUALIFIER, mktoken(3))
|
||||
q.add_token(query.TokenRange(0, 1), query.TOKEN_PARTIAL, mktoken(1))
|
||||
q.add_token(query.TokenRange(0, 1), query.TOKEN_NEAR_ITEM, mktoken(2))
|
||||
q.add_token(query.TokenRange(0, 1), query.TOKEN_QUALIFIER, mktoken(3))
|
||||
|
||||
assert len(q.get_tokens(query.TokenRange(0, 1), query.TokenType.PARTIAL)) == 1
|
||||
assert len(q.get_tokens(query.TokenRange(0, 1), query.TokenType.NEAR_ITEM)) == 1
|
||||
assert len(q.get_tokens(query.TokenRange(0, 1), query.TokenType.QUALIFIER)) == 0
|
||||
assert len(q.get_tokens(query.TokenRange(0, 1), query.TOKEN_PARTIAL)) == 1
|
||||
assert len(q.get_tokens(query.TokenRange(0, 1), query.TOKEN_NEAR_ITEM)) == 1
|
||||
assert len(q.get_tokens(query.TokenRange(0, 1), query.TOKEN_QUALIFIER)) == 0
|
||||
|
||||
|
||||
def test_query_struct_amenity_two_words():
|
||||
@@ -121,15 +121,15 @@ def test_query_struct_amenity_two_words():
|
||||
q.add_node(query.BREAK_END, query.PhraseType.NONE)
|
||||
|
||||
for trange in [(0, 1), (1, 2)]:
|
||||
q.add_token(query.TokenRange(*trange), query.TokenType.PARTIAL, mktoken(1))
|
||||
q.add_token(query.TokenRange(*trange), query.TokenType.NEAR_ITEM, mktoken(2))
|
||||
q.add_token(query.TokenRange(*trange), query.TokenType.QUALIFIER, mktoken(3))
|
||||
q.add_token(query.TokenRange(*trange), query.TOKEN_PARTIAL, mktoken(1))
|
||||
q.add_token(query.TokenRange(*trange), query.TOKEN_NEAR_ITEM, mktoken(2))
|
||||
q.add_token(query.TokenRange(*trange), query.TOKEN_QUALIFIER, mktoken(3))
|
||||
|
||||
assert len(q.get_tokens(query.TokenRange(0, 1), query.TokenType.PARTIAL)) == 1
|
||||
assert len(q.get_tokens(query.TokenRange(0, 1), query.TokenType.NEAR_ITEM)) == 0
|
||||
assert len(q.get_tokens(query.TokenRange(0, 1), query.TokenType.QUALIFIER)) == 1
|
||||
assert len(q.get_tokens(query.TokenRange(0, 1), query.TOKEN_PARTIAL)) == 1
|
||||
assert len(q.get_tokens(query.TokenRange(0, 1), query.TOKEN_NEAR_ITEM)) == 0
|
||||
assert len(q.get_tokens(query.TokenRange(0, 1), query.TOKEN_QUALIFIER)) == 1
|
||||
|
||||
assert len(q.get_tokens(query.TokenRange(1, 2), query.TokenType.PARTIAL)) == 1
|
||||
assert len(q.get_tokens(query.TokenRange(1, 2), query.TokenType.NEAR_ITEM)) == 0
|
||||
assert len(q.get_tokens(query.TokenRange(1, 2), query.TokenType.QUALIFIER)) == 1
|
||||
assert len(q.get_tokens(query.TokenRange(1, 2), query.TOKEN_PARTIAL)) == 1
|
||||
assert len(q.get_tokens(query.TokenRange(1, 2), query.TOKEN_NEAR_ITEM)) == 0
|
||||
assert len(q.get_tokens(query.TokenRange(1, 2), query.TOKEN_QUALIFIER)) == 1
|
||||
|
||||
|
||||
@@ -9,7 +9,7 @@ Tests for creating abstract searches from token assignments.
|
||||
"""
|
||||
import pytest
|
||||
|
||||
from nominatim_api.search.query import Token, TokenRange, PhraseType, TokenType, QueryStruct, Phrase
|
||||
from nominatim_api.search.query import Token, TokenRange, PhraseType, QueryStruct, Phrase
|
||||
import nominatim_api.search.query as qmod
|
||||
from nominatim_api.search.db_search_builder import SearchBuilder
|
||||
from nominatim_api.search.token_assignment import TokenAssignment
|
||||
@@ -32,7 +32,7 @@ def make_query(*args):
|
||||
for end, ttype, tinfo in tlist:
|
||||
for tid, word in tinfo:
|
||||
q.add_token(TokenRange(start, end), ttype,
|
||||
MyToken(penalty=0.5 if ttype == TokenType.PARTIAL else 0.0,
|
||||
MyToken(penalty=0.5 if ttype == qmod.TOKEN_PARTIAL else 0.0,
|
||||
token=tid, count=1, addr_count=1,
|
||||
lookup_word=word))
|
||||
|
||||
@@ -41,7 +41,7 @@ def make_query(*args):
|
||||
|
||||
|
||||
def test_country_search():
|
||||
q = make_query([(1, TokenType.COUNTRY, [(2, 'de'), (3, 'en')])])
|
||||
q = make_query([(1, qmod.TOKEN_COUNTRY, [(2, 'de'), (3, 'en')])])
|
||||
builder = SearchBuilder(q, SearchDetails())
|
||||
|
||||
searches = list(builder.build(TokenAssignment(country=TokenRange(0, 1))))
|
||||
@@ -55,7 +55,7 @@ def test_country_search():
|
||||
|
||||
|
||||
def test_country_search_with_country_restriction():
|
||||
q = make_query([(1, TokenType.COUNTRY, [(2, 'de'), (3, 'en')])])
|
||||
q = make_query([(1, qmod.TOKEN_COUNTRY, [(2, 'de'), (3, 'en')])])
|
||||
builder = SearchBuilder(q, SearchDetails.from_kwargs({'countries': 'en,fr'}))
|
||||
|
||||
searches = list(builder.build(TokenAssignment(country=TokenRange(0, 1))))
|
||||
@@ -69,7 +69,7 @@ def test_country_search_with_country_restriction():
|
||||
|
||||
|
||||
def test_country_search_with_conflicting_country_restriction():
|
||||
q = make_query([(1, TokenType.COUNTRY, [(2, 'de'), (3, 'en')])])
|
||||
q = make_query([(1, qmod.TOKEN_COUNTRY, [(2, 'de'), (3, 'en')])])
|
||||
builder = SearchBuilder(q, SearchDetails.from_kwargs({'countries': 'fr'}))
|
||||
|
||||
searches = list(builder.build(TokenAssignment(country=TokenRange(0, 1))))
|
||||
@@ -78,7 +78,7 @@ def test_country_search_with_conflicting_country_restriction():
|
||||
|
||||
|
||||
def test_postcode_search_simple():
|
||||
q = make_query([(1, TokenType.POSTCODE, [(34, '2367')])])
|
||||
q = make_query([(1, qmod.TOKEN_POSTCODE, [(34, '2367')])])
|
||||
builder = SearchBuilder(q, SearchDetails())
|
||||
|
||||
searches = list(builder.build(TokenAssignment(postcode=TokenRange(0, 1))))
|
||||
@@ -94,8 +94,8 @@ def test_postcode_search_simple():
|
||||
|
||||
|
||||
def test_postcode_with_country():
|
||||
q = make_query([(1, TokenType.POSTCODE, [(34, '2367')])],
|
||||
[(2, TokenType.COUNTRY, [(1, 'xx')])])
|
||||
q = make_query([(1, qmod.TOKEN_POSTCODE, [(34, '2367')])],
|
||||
[(2, qmod.TOKEN_COUNTRY, [(1, 'xx')])])
|
||||
builder = SearchBuilder(q, SearchDetails())
|
||||
|
||||
searches = list(builder.build(TokenAssignment(postcode=TokenRange(0, 1),
|
||||
@@ -112,8 +112,8 @@ def test_postcode_with_country():
|
||||
|
||||
|
||||
def test_postcode_with_address():
|
||||
q = make_query([(1, TokenType.POSTCODE, [(34, '2367')])],
|
||||
[(2, TokenType.PARTIAL, [(100, 'word')])])
|
||||
q = make_query([(1, qmod.TOKEN_POSTCODE, [(34, '2367')])],
|
||||
[(2, qmod.TOKEN_PARTIAL, [(100, 'word')])])
|
||||
builder = SearchBuilder(q, SearchDetails())
|
||||
|
||||
searches = list(builder.build(TokenAssignment(postcode=TokenRange(0, 1),
|
||||
@@ -130,9 +130,9 @@ def test_postcode_with_address():
|
||||
|
||||
|
||||
def test_postcode_with_address_with_full_word():
|
||||
q = make_query([(1, TokenType.POSTCODE, [(34, '2367')])],
|
||||
[(2, TokenType.PARTIAL, [(100, 'word')]),
|
||||
(2, TokenType.WORD, [(1, 'full')])])
|
||||
q = make_query([(1, qmod.TOKEN_POSTCODE, [(34, '2367')])],
|
||||
[(2, qmod.TOKEN_PARTIAL, [(100, 'word')]),
|
||||
(2, qmod.TOKEN_WORD, [(1, 'full')])])
|
||||
builder = SearchBuilder(q, SearchDetails())
|
||||
|
||||
searches = list(builder.build(TokenAssignment(postcode=TokenRange(0, 1),
|
||||
@@ -151,7 +151,7 @@ def test_postcode_with_address_with_full_word():
|
||||
@pytest.mark.parametrize('kwargs', [{'viewbox': '0,0,1,1', 'bounded_viewbox': True},
|
||||
{'near': '10,10'}])
|
||||
def test_near_item_only(kwargs):
|
||||
q = make_query([(1, TokenType.NEAR_ITEM, [(2, 'foo')])])
|
||||
q = make_query([(1, qmod.TOKEN_NEAR_ITEM, [(2, 'foo')])])
|
||||
builder = SearchBuilder(q, SearchDetails.from_kwargs(kwargs))
|
||||
|
||||
searches = list(builder.build(TokenAssignment(near_item=TokenRange(0, 1))))
|
||||
@@ -167,7 +167,7 @@ def test_near_item_only(kwargs):
|
||||
@pytest.mark.parametrize('kwargs', [{'viewbox': '0,0,1,1'},
|
||||
{}])
|
||||
def test_near_item_skipped(kwargs):
|
||||
q = make_query([(1, TokenType.NEAR_ITEM, [(2, 'foo')])])
|
||||
q = make_query([(1, qmod.TOKEN_NEAR_ITEM, [(2, 'foo')])])
|
||||
builder = SearchBuilder(q, SearchDetails.from_kwargs(kwargs))
|
||||
|
||||
searches = list(builder.build(TokenAssignment(near_item=TokenRange(0, 1))))
|
||||
@@ -176,8 +176,8 @@ def test_near_item_skipped(kwargs):
|
||||
|
||||
|
||||
def test_name_only_search():
|
||||
q = make_query([(1, TokenType.PARTIAL, [(1, 'a')]),
|
||||
(1, TokenType.WORD, [(100, 'a')])])
|
||||
q = make_query([(1, qmod.TOKEN_PARTIAL, [(1, 'a')]),
|
||||
(1, qmod.TOKEN_WORD, [(100, 'a')])])
|
||||
builder = SearchBuilder(q, SearchDetails())
|
||||
|
||||
searches = list(builder.build(TokenAssignment(name=TokenRange(0, 1))))
|
||||
@@ -195,9 +195,9 @@ def test_name_only_search():
|
||||
|
||||
|
||||
def test_name_with_qualifier():
|
||||
q = make_query([(1, TokenType.PARTIAL, [(1, 'a')]),
|
||||
(1, TokenType.WORD, [(100, 'a')])],
|
||||
[(2, TokenType.QUALIFIER, [(55, 'hotel')])])
|
||||
q = make_query([(1, qmod.TOKEN_PARTIAL, [(1, 'a')]),
|
||||
(1, qmod.TOKEN_WORD, [(100, 'a')])],
|
||||
[(2, qmod.TOKEN_QUALIFIER, [(55, 'hotel')])])
|
||||
builder = SearchBuilder(q, SearchDetails())
|
||||
|
||||
searches = list(builder.build(TokenAssignment(name=TokenRange(0, 1),
|
||||
@@ -216,9 +216,9 @@ def test_name_with_qualifier():
|
||||
|
||||
|
||||
def test_name_with_housenumber_search():
|
||||
q = make_query([(1, TokenType.PARTIAL, [(1, 'a')]),
|
||||
(1, TokenType.WORD, [(100, 'a')])],
|
||||
[(2, TokenType.HOUSENUMBER, [(66, '66')])])
|
||||
q = make_query([(1, qmod.TOKEN_PARTIAL, [(1, 'a')]),
|
||||
(1, qmod.TOKEN_WORD, [(100, 'a')])],
|
||||
[(2, qmod.TOKEN_HOUSENUMBER, [(66, '66')])])
|
||||
builder = SearchBuilder(q, SearchDetails())
|
||||
|
||||
searches = list(builder.build(TokenAssignment(name=TokenRange(0, 1),
|
||||
@@ -236,12 +236,12 @@ def test_name_with_housenumber_search():
|
||||
|
||||
|
||||
def test_name_and_address():
|
||||
q = make_query([(1, TokenType.PARTIAL, [(1, 'a')]),
|
||||
(1, TokenType.WORD, [(100, 'a')])],
|
||||
[(2, TokenType.PARTIAL, [(2, 'b')]),
|
||||
(2, TokenType.WORD, [(101, 'b')])],
|
||||
[(3, TokenType.PARTIAL, [(3, 'c')]),
|
||||
(3, TokenType.WORD, [(102, 'c')])]
|
||||
q = make_query([(1, qmod.TOKEN_PARTIAL, [(1, 'a')]),
|
||||
(1, qmod.TOKEN_WORD, [(100, 'a')])],
|
||||
[(2, qmod.TOKEN_PARTIAL, [(2, 'b')]),
|
||||
(2, qmod.TOKEN_WORD, [(101, 'b')])],
|
||||
[(3, qmod.TOKEN_PARTIAL, [(3, 'c')]),
|
||||
(3, qmod.TOKEN_WORD, [(102, 'c')])]
|
||||
)
|
||||
builder = SearchBuilder(q, SearchDetails())
|
||||
|
||||
@@ -261,13 +261,13 @@ def test_name_and_address():
|
||||
|
||||
|
||||
def test_name_and_complex_address():
|
||||
q = make_query([(1, TokenType.PARTIAL, [(1, 'a')]),
|
||||
(1, TokenType.WORD, [(100, 'a')])],
|
||||
[(2, TokenType.PARTIAL, [(2, 'b')]),
|
||||
(3, TokenType.WORD, [(101, 'bc')])],
|
||||
[(3, TokenType.PARTIAL, [(3, 'c')])],
|
||||
[(4, TokenType.PARTIAL, [(4, 'd')]),
|
||||
(4, TokenType.WORD, [(103, 'd')])]
|
||||
q = make_query([(1, qmod.TOKEN_PARTIAL, [(1, 'a')]),
|
||||
(1, qmod.TOKEN_WORD, [(100, 'a')])],
|
||||
[(2, qmod.TOKEN_PARTIAL, [(2, 'b')]),
|
||||
(3, qmod.TOKEN_WORD, [(101, 'bc')])],
|
||||
[(3, qmod.TOKEN_PARTIAL, [(3, 'c')])],
|
||||
[(4, qmod.TOKEN_PARTIAL, [(4, 'd')]),
|
||||
(4, qmod.TOKEN_WORD, [(103, 'd')])]
|
||||
)
|
||||
builder = SearchBuilder(q, SearchDetails())
|
||||
|
||||
@@ -287,9 +287,9 @@ def test_name_and_complex_address():
|
||||
|
||||
|
||||
def test_name_only_near_search():
|
||||
q = make_query([(1, TokenType.NEAR_ITEM, [(88, 'g')])],
|
||||
[(2, TokenType.PARTIAL, [(1, 'a')]),
|
||||
(2, TokenType.WORD, [(100, 'a')])])
|
||||
q = make_query([(1, qmod.TOKEN_NEAR_ITEM, [(88, 'g')])],
|
||||
[(2, qmod.TOKEN_PARTIAL, [(1, 'a')]),
|
||||
(2, qmod.TOKEN_WORD, [(100, 'a')])])
|
||||
builder = SearchBuilder(q, SearchDetails())
|
||||
|
||||
searches = list(builder.build(TokenAssignment(name=TokenRange(1, 2),
|
||||
@@ -303,8 +303,8 @@ def test_name_only_near_search():
|
||||
|
||||
|
||||
def test_name_only_search_with_category():
|
||||
q = make_query([(1, TokenType.PARTIAL, [(1, 'a')]),
|
||||
(1, TokenType.WORD, [(100, 'a')])])
|
||||
q = make_query([(1, qmod.TOKEN_PARTIAL, [(1, 'a')]),
|
||||
(1, qmod.TOKEN_WORD, [(100, 'a')])])
|
||||
builder = SearchBuilder(q, SearchDetails.from_kwargs({'categories': [('foo', 'bar')]}))
|
||||
|
||||
searches = list(builder.build(TokenAssignment(name=TokenRange(0, 1))))
|
||||
@@ -317,9 +317,9 @@ def test_name_only_search_with_category():
|
||||
|
||||
|
||||
def test_name_with_near_item_search_with_category_mismatch():
|
||||
q = make_query([(1, TokenType.NEAR_ITEM, [(88, 'g')])],
|
||||
[(2, TokenType.PARTIAL, [(1, 'a')]),
|
||||
(2, TokenType.WORD, [(100, 'a')])])
|
||||
q = make_query([(1, qmod.TOKEN_NEAR_ITEM, [(88, 'g')])],
|
||||
[(2, qmod.TOKEN_PARTIAL, [(1, 'a')]),
|
||||
(2, qmod.TOKEN_WORD, [(100, 'a')])])
|
||||
builder = SearchBuilder(q, SearchDetails.from_kwargs({'categories': [('foo', 'bar')]}))
|
||||
|
||||
searches = list(builder.build(TokenAssignment(name=TokenRange(1, 2),
|
||||
@@ -329,9 +329,9 @@ def test_name_with_near_item_search_with_category_mismatch():
|
||||
|
||||
|
||||
def test_name_with_near_item_search_with_category_match():
|
||||
q = make_query([(1, TokenType.NEAR_ITEM, [(88, 'g')])],
|
||||
[(2, TokenType.PARTIAL, [(1, 'a')]),
|
||||
(2, TokenType.WORD, [(100, 'a')])])
|
||||
q = make_query([(1, qmod.TOKEN_NEAR_ITEM, [(88, 'g')])],
|
||||
[(2, qmod.TOKEN_PARTIAL, [(1, 'a')]),
|
||||
(2, qmod.TOKEN_WORD, [(100, 'a')])])
|
||||
builder = SearchBuilder(q, SearchDetails.from_kwargs({'categories': [('foo', 'bar'),
|
||||
('this', 'that')]}))
|
||||
|
||||
@@ -346,9 +346,9 @@ def test_name_with_near_item_search_with_category_match():
|
||||
|
||||
|
||||
def test_name_with_qualifier_search_with_category_mismatch():
|
||||
q = make_query([(1, TokenType.QUALIFIER, [(88, 'g')])],
|
||||
[(2, TokenType.PARTIAL, [(1, 'a')]),
|
||||
(2, TokenType.WORD, [(100, 'a')])])
|
||||
q = make_query([(1, qmod.TOKEN_QUALIFIER, [(88, 'g')])],
|
||||
[(2, qmod.TOKEN_PARTIAL, [(1, 'a')]),
|
||||
(2, qmod.TOKEN_WORD, [(100, 'a')])])
|
||||
builder = SearchBuilder(q, SearchDetails.from_kwargs({'categories': [('foo', 'bar')]}))
|
||||
|
||||
searches = list(builder.build(TokenAssignment(name=TokenRange(1, 2),
|
||||
@@ -358,9 +358,9 @@ def test_name_with_qualifier_search_with_category_mismatch():
|
||||
|
||||
|
||||
def test_name_with_qualifier_search_with_category_match():
|
||||
q = make_query([(1, TokenType.QUALIFIER, [(88, 'g')])],
|
||||
[(2, TokenType.PARTIAL, [(1, 'a')]),
|
||||
(2, TokenType.WORD, [(100, 'a')])])
|
||||
q = make_query([(1, qmod.TOKEN_QUALIFIER, [(88, 'g')])],
|
||||
[(2, qmod.TOKEN_PARTIAL, [(1, 'a')]),
|
||||
(2, qmod.TOKEN_WORD, [(100, 'a')])])
|
||||
builder = SearchBuilder(q, SearchDetails.from_kwargs({'categories': [('foo', 'bar'),
|
||||
('this', 'that')]}))
|
||||
|
||||
@@ -375,8 +375,8 @@ def test_name_with_qualifier_search_with_category_match():
|
||||
|
||||
|
||||
def test_name_only_search_with_countries():
|
||||
q = make_query([(1, TokenType.PARTIAL, [(1, 'a')]),
|
||||
(1, TokenType.WORD, [(100, 'a')])])
|
||||
q = make_query([(1, qmod.TOKEN_PARTIAL, [(1, 'a')]),
|
||||
(1, qmod.TOKEN_WORD, [(100, 'a')])])
|
||||
builder = SearchBuilder(q, SearchDetails.from_kwargs({'countries': 'de,en'}))
|
||||
|
||||
searches = list(builder.build(TokenAssignment(name=TokenRange(0, 1))))
|
||||
@@ -397,14 +397,14 @@ def make_counted_searches(name_part, name_full, address_part, address_full,
|
||||
q.add_node(qmod.BREAK_WORD, PhraseType.NONE)
|
||||
q.add_node(qmod.BREAK_END, PhraseType.NONE)
|
||||
|
||||
q.add_token(TokenRange(0, 1), TokenType.PARTIAL,
|
||||
q.add_token(TokenRange(0, 1), qmod.TOKEN_PARTIAL,
|
||||
MyToken(0.5, 1, name_part, 1, 'name_part'))
|
||||
q.add_token(TokenRange(0, 1), TokenType.WORD,
|
||||
q.add_token(TokenRange(0, 1), qmod.TOKEN_WORD,
|
||||
MyToken(0, 101, name_full, 1, 'name_full'))
|
||||
for i in range(num_address_parts):
|
||||
q.add_token(TokenRange(i + 1, i + 2), TokenType.PARTIAL,
|
||||
q.add_token(TokenRange(i + 1, i + 2), qmod.TOKEN_PARTIAL,
|
||||
MyToken(0.5, 2, address_part, 1, 'address_part'))
|
||||
q.add_token(TokenRange(i + 1, i + 2), TokenType.WORD,
|
||||
q.add_token(TokenRange(i + 1, i + 2), qmod.TOKEN_WORD,
|
||||
MyToken(0, 102, address_full, 1, 'address_full'))
|
||||
|
||||
builder = SearchBuilder(q, SearchDetails())
|
||||
|
||||
@@ -11,7 +11,7 @@ import pytest
|
||||
import pytest_asyncio
|
||||
|
||||
from nominatim_api import NominatimAPIAsync
|
||||
from nominatim_api.search.query import Phrase, PhraseType, TokenType
|
||||
from nominatim_api.search.query import Phrase, PhraseType
|
||||
import nominatim_api.search.query as qmod
|
||||
import nominatim_api.search.icu_tokenizer as tok
|
||||
from nominatim_api.logging import set_log_output, get_and_disable
|
||||
@@ -101,8 +101,8 @@ async def test_splitting_in_transliteration(conn):
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
@pytest.mark.parametrize('term,order', [('23456', ['POSTCODE', 'HOUSENUMBER', 'WORD', 'PARTIAL']),
|
||||
('3', ['HOUSENUMBER', 'POSTCODE', 'WORD', 'PARTIAL'])
|
||||
@pytest.mark.parametrize('term,order', [('23456', ['P', 'H', 'W', 'w']),
|
||||
('3', ['H', 'P', 'W', 'w'])
|
||||
])
|
||||
async def test_penalty_postcodes_and_housenumbers(conn, term, order):
|
||||
ana = await tok.create_query_analyzer(conn)
|
||||
@@ -116,7 +116,7 @@ async def test_penalty_postcodes_and_housenumbers(conn, term, order):
|
||||
|
||||
assert query.num_token_slots() == 1
|
||||
|
||||
torder = [(tl.tokens[0].penalty, tl.ttype.name) for tl in query.nodes[0].starting]
|
||||
torder = [(tl.tokens[0].penalty, tl.ttype) for tl in query.nodes[0].starting]
|
||||
torder.sort()
|
||||
|
||||
assert [t[1] for t in torder] == order
|
||||
@@ -132,7 +132,7 @@ async def test_category_words_only_at_beginning(conn):
|
||||
|
||||
assert query.num_token_slots() == 3
|
||||
assert len(query.nodes[0].starting) == 1
|
||||
assert query.nodes[0].starting[0].ttype == TokenType.NEAR_ITEM
|
||||
assert query.nodes[0].starting[0].ttype == qmod.TOKEN_NEAR_ITEM
|
||||
assert not query.nodes[2].starting
|
||||
|
||||
|
||||
@@ -146,7 +146,7 @@ async def test_freestanding_qualifier_words_become_category(conn):
|
||||
|
||||
assert query.num_token_slots() == 1
|
||||
assert len(query.nodes[0].starting) == 1
|
||||
assert query.nodes[0].starting[0].ttype == TokenType.NEAR_ITEM
|
||||
assert query.nodes[0].starting[0].ttype == qmod.TOKEN_NEAR_ITEM
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
@@ -159,9 +159,9 @@ async def test_qualifier_words(conn):
|
||||
query = await ana.analyze_query(make_phrase('foo BAR foo BAR foo'))
|
||||
|
||||
assert query.num_token_slots() == 5
|
||||
assert set(t.ttype for t in query.nodes[0].starting) == {TokenType.QUALIFIER}
|
||||
assert set(t.ttype for t in query.nodes[2].starting) == {TokenType.QUALIFIER}
|
||||
assert set(t.ttype for t in query.nodes[4].starting) == {TokenType.QUALIFIER}
|
||||
assert set(t.ttype for t in query.nodes[0].starting) == {qmod.TOKEN_QUALIFIER}
|
||||
assert set(t.ttype for t in query.nodes[2].starting) == {qmod.TOKEN_QUALIFIER}
|
||||
assert set(t.ttype for t in query.nodes[4].starting) == {qmod.TOKEN_QUALIFIER}
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
@@ -173,10 +173,10 @@ async def test_add_unknown_housenumbers(conn):
|
||||
query = await ana.analyze_query(make_phrase('466 23 99834 34a'))
|
||||
|
||||
assert query.num_token_slots() == 4
|
||||
assert query.nodes[0].starting[0].ttype == TokenType.HOUSENUMBER
|
||||
assert query.nodes[0].starting[0].ttype == qmod.TOKEN_HOUSENUMBER
|
||||
assert len(query.nodes[0].starting[0].tokens) == 1
|
||||
assert query.nodes[0].starting[0].tokens[0].token == 0
|
||||
assert query.nodes[1].starting[0].ttype == TokenType.HOUSENUMBER
|
||||
assert query.nodes[1].starting[0].ttype == qmod.TOKEN_HOUSENUMBER
|
||||
assert len(query.nodes[1].starting[0].tokens) == 1
|
||||
assert query.nodes[1].starting[0].tokens[0].token == 1
|
||||
assert not query.nodes[2].starting
|
||||
|
||||
@@ -9,7 +9,7 @@ Test for creation of token assignments from tokenized queries.
|
||||
"""
|
||||
import pytest
|
||||
|
||||
from nominatim_api.search.query import QueryStruct, Phrase, PhraseType, TokenType, TokenRange, Token
|
||||
from nominatim_api.search.query import QueryStruct, Phrase, PhraseType, TokenRange, Token
|
||||
import nominatim_api.search.query as qmod
|
||||
from nominatim_api.search.token_assignment import yield_token_assignments, TokenAssignment, PENALTY_TOKENCHANGE
|
||||
|
||||
@@ -52,9 +52,9 @@ def test_query_with_missing_tokens():
|
||||
|
||||
def test_one_word_query():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE,
|
||||
[(1, TokenType.PARTIAL),
|
||||
(1, TokenType.WORD),
|
||||
(1, TokenType.HOUSENUMBER)]))
|
||||
[(1, qmod.TOKEN_PARTIAL),
|
||||
(1, qmod.TOKEN_WORD),
|
||||
(1, qmod.TOKEN_HOUSENUMBER)]))
|
||||
|
||||
res = list(yield_token_assignments(q))
|
||||
assert res == [TokenAssignment(name=TokenRange(0, 1))]
|
||||
@@ -62,7 +62,7 @@ def test_one_word_query():
|
||||
|
||||
def test_single_postcode():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE,
|
||||
[(1, TokenType.POSTCODE)]))
|
||||
[(1, qmod.TOKEN_POSTCODE)]))
|
||||
|
||||
res = list(yield_token_assignments(q))
|
||||
assert res == [TokenAssignment(postcode=TokenRange(0, 1))]
|
||||
@@ -70,7 +70,7 @@ def test_single_postcode():
|
||||
|
||||
def test_single_country_name():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE,
|
||||
[(1, TokenType.COUNTRY)]))
|
||||
[(1, qmod.TOKEN_COUNTRY)]))
|
||||
|
||||
res = list(yield_token_assignments(q))
|
||||
assert res == [TokenAssignment(country=TokenRange(0, 1))]
|
||||
@@ -78,8 +78,8 @@ def test_single_country_name():
|
||||
|
||||
def test_single_word_poi_search():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE,
|
||||
[(1, TokenType.NEAR_ITEM),
|
||||
(1, TokenType.QUALIFIER)]))
|
||||
[(1, qmod.TOKEN_NEAR_ITEM),
|
||||
(1, qmod.TOKEN_QUALIFIER)]))
|
||||
|
||||
res = list(yield_token_assignments(q))
|
||||
assert res == [TokenAssignment(near_item=TokenRange(0, 1))]
|
||||
@@ -87,9 +87,9 @@ def test_single_word_poi_search():
|
||||
|
||||
@pytest.mark.parametrize('btype', [qmod.BREAK_WORD, qmod.BREAK_PART, qmod.BREAK_TOKEN])
|
||||
def test_multiple_simple_words(btype):
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.PARTIAL)]),
|
||||
(btype, PhraseType.NONE, [(2, TokenType.PARTIAL)]),
|
||||
(btype, PhraseType.NONE, [(3, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_PARTIAL)]),
|
||||
(btype, PhraseType.NONE, [(2, qmod.TOKEN_PARTIAL)]),
|
||||
(btype, PhraseType.NONE, [(3, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
penalty = PENALTY_TOKENCHANGE[btype]
|
||||
|
||||
@@ -107,8 +107,8 @@ def test_multiple_simple_words(btype):
|
||||
|
||||
|
||||
def test_multiple_words_respect_phrase_break():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
TokenAssignment(name=TokenRange(0, 1),
|
||||
@@ -118,8 +118,8 @@ def test_multiple_words_respect_phrase_break():
|
||||
|
||||
|
||||
def test_housenumber_and_street():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.HOUSENUMBER)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_HOUSENUMBER)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
TokenAssignment(name=TokenRange(1, 2),
|
||||
@@ -129,8 +129,8 @@ def test_housenumber_and_street():
|
||||
|
||||
|
||||
def test_housenumber_and_street_backwards():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, TokenType.HOUSENUMBER)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, qmod.TOKEN_HOUSENUMBER)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
TokenAssignment(name=TokenRange(0, 1),
|
||||
@@ -140,10 +140,10 @@ def test_housenumber_and_street_backwards():
|
||||
|
||||
|
||||
def test_housenumber_and_postcode():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, TokenType.HOUSENUMBER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(4, TokenType.POSTCODE)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, qmod.TOKEN_HOUSENUMBER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(4, qmod.TOKEN_POSTCODE)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
TokenAssignment(penalty=pytest.approx(0.3),
|
||||
@@ -157,10 +157,10 @@ def test_housenumber_and_postcode():
|
||||
postcode=TokenRange(3, 4)))
|
||||
|
||||
def test_postcode_and_housenumber():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, TokenType.POSTCODE)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(4, TokenType.HOUSENUMBER)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, qmod.TOKEN_POSTCODE)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(4, qmod.TOKEN_HOUSENUMBER)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
TokenAssignment(penalty=pytest.approx(0.3),
|
||||
@@ -175,38 +175,38 @@ def test_postcode_and_housenumber():
|
||||
|
||||
|
||||
def test_country_housenumber_postcode():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.COUNTRY)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, TokenType.HOUSENUMBER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(4, TokenType.POSTCODE)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_COUNTRY)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, qmod.TOKEN_HOUSENUMBER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(4, qmod.TOKEN_POSTCODE)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q))
|
||||
|
||||
|
||||
@pytest.mark.parametrize('ttype', [TokenType.POSTCODE, TokenType.COUNTRY,
|
||||
TokenType.NEAR_ITEM, TokenType.QUALIFIER])
|
||||
@pytest.mark.parametrize('ttype', [qmod.TOKEN_POSTCODE, qmod.TOKEN_COUNTRY,
|
||||
qmod.TOKEN_NEAR_ITEM, qmod.TOKEN_QUALIFIER])
|
||||
def test_housenumber_with_only_special_terms(ttype):
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.HOUSENUMBER)]),
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_HOUSENUMBER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, ttype)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q))
|
||||
|
||||
|
||||
@pytest.mark.parametrize('ttype', [TokenType.POSTCODE, TokenType.HOUSENUMBER, TokenType.COUNTRY])
|
||||
@pytest.mark.parametrize('ttype', [qmod.TOKEN_POSTCODE, qmod.TOKEN_HOUSENUMBER, qmod.TOKEN_COUNTRY])
|
||||
def test_multiple_special_tokens(ttype):
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, ttype)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(3, ttype)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q))
|
||||
|
||||
|
||||
def test_housenumber_many_phrases():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(3, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(4, TokenType.HOUSENUMBER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(5, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(3, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(4, qmod.TOKEN_HOUSENUMBER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(5, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
TokenAssignment(penalty=0.1,
|
||||
@@ -221,8 +221,8 @@ def test_housenumber_many_phrases():
|
||||
|
||||
|
||||
def test_country_at_beginning():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.COUNTRY)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_COUNTRY)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
TokenAssignment(penalty=0.1, name=TokenRange(1, 2),
|
||||
@@ -230,8 +230,8 @@ def test_country_at_beginning():
|
||||
|
||||
|
||||
def test_country_at_end():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, TokenType.COUNTRY)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, qmod.TOKEN_COUNTRY)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
TokenAssignment(penalty=0.1, name=TokenRange(0, 1),
|
||||
@@ -239,16 +239,16 @@ def test_country_at_end():
|
||||
|
||||
|
||||
def test_country_in_middle():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, TokenType.COUNTRY)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, qmod.TOKEN_COUNTRY)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q))
|
||||
|
||||
|
||||
def test_postcode_with_designation():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.POSTCODE)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_POSTCODE)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
TokenAssignment(penalty=0.1, name=TokenRange(1, 2),
|
||||
@@ -258,8 +258,8 @@ def test_postcode_with_designation():
|
||||
|
||||
|
||||
def test_postcode_with_designation_backwards():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, TokenType.POSTCODE)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, qmod.TOKEN_POSTCODE)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
TokenAssignment(name=TokenRange(0, 1),
|
||||
@@ -269,8 +269,8 @@ def test_postcode_with_designation_backwards():
|
||||
|
||||
|
||||
def test_near_item_at_beginning():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.NEAR_ITEM)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_NEAR_ITEM)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
TokenAssignment(penalty=0.1, name=TokenRange(1, 2),
|
||||
@@ -278,8 +278,8 @@ def test_near_item_at_beginning():
|
||||
|
||||
|
||||
def test_near_item_at_end():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, TokenType.NEAR_ITEM)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, qmod.TOKEN_NEAR_ITEM)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
TokenAssignment(penalty=0.1, name=TokenRange(0, 1),
|
||||
@@ -287,17 +287,17 @@ def test_near_item_at_end():
|
||||
|
||||
|
||||
def test_near_item_in_middle():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, TokenType.NEAR_ITEM)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, qmod.TOKEN_NEAR_ITEM)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q))
|
||||
|
||||
|
||||
def test_qualifier_at_beginning():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.QUALIFIER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_QUALIFIER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
@@ -309,11 +309,11 @@ def test_qualifier_at_beginning():
|
||||
|
||||
|
||||
def test_qualifier_after_name():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, TokenType.QUALIFIER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(4, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(5, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, qmod.TOKEN_QUALIFIER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(4, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(5, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
|
||||
check_assignments(yield_token_assignments(q),
|
||||
@@ -326,27 +326,27 @@ def test_qualifier_after_name():
|
||||
|
||||
|
||||
def test_qualifier_before_housenumber():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.QUALIFIER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, TokenType.HOUSENUMBER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_QUALIFIER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, qmod.TOKEN_HOUSENUMBER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q))
|
||||
|
||||
|
||||
def test_qualifier_after_housenumber():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.HOUSENUMBER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, TokenType.QUALIFIER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_HOUSENUMBER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(2, qmod.TOKEN_QUALIFIER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q))
|
||||
|
||||
|
||||
def test_qualifier_in_middle_of_phrase():
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, TokenType.QUALIFIER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(4, TokenType.PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(5, TokenType.PARTIAL)]))
|
||||
q = make_query((qmod.BREAK_START, PhraseType.NONE, [(1, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(2, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(3, qmod.TOKEN_QUALIFIER)]),
|
||||
(qmod.BREAK_WORD, PhraseType.NONE, [(4, qmod.TOKEN_PARTIAL)]),
|
||||
(qmod.BREAK_PHRASE, PhraseType.NONE, [(5, qmod.TOKEN_PARTIAL)]))
|
||||
|
||||
check_assignments(yield_token_assignments(q))
|
||||
|
||||
|
||||
Reference in New Issue
Block a user