a8261ece
tangwang
检索效果优化
|
1
2
3
4
5
6
7
8
9
|
from config import FunctionScoreConfig, IndexConfig, QueryConfig, RerankConfig, SPUConfig, SearchConfig
from query.query_parser import QueryParser
class _DummyTranslator:
def translate(self, text, target_lang, source_lang, scene, model_name):
return f"{text}-{target_lang}"
|
ef5baa86
tangwang
混杂语言处理
|
10
11
12
13
|
def _tokenizer(text):
return str(text).split()
|
a8261ece
tangwang
检索效果优化
|
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
|
def _build_config() -> SearchConfig:
return SearchConfig(
es_index_name="test_products",
field_boosts={"title.en": 3.0, "title.zh": 3.0},
indexes=[IndexConfig(name="default", label="default", fields=["title.en", "title.zh"])],
query_config=QueryConfig(
enable_text_embedding=False,
enable_query_rewrite=False,
supported_languages=["en", "zh"],
default_language="zh",
),
function_score=FunctionScoreConfig(),
rerank=RerankConfig(),
spu_config=SPUConfig(enabled=False),
)
|
35da3813
tangwang
中英混写query的优化逻辑,不适...
|
31
|
def test_parse_mixed_zh_query_translates_to_en(monkeypatch):
|
ef5baa86
tangwang
混杂语言处理
|
32
|
parser = QueryParser(_build_config(), translator=_DummyTranslator(), tokenizer=_tokenizer)
|
a8261ece
tangwang
检索效果优化
|
33
|
monkeypatch.setattr(parser.language_detector, "detect", lambda text: "zh")
|
a8261ece
tangwang
检索效果优化
|
34
|
|
ef5baa86
tangwang
混杂语言处理
|
35
36
37
38
39
40
|
result = parser.parse(
"法式 dress 连衣裙",
tenant_id="162",
generate_vector=False,
target_languages=["zh", "en"],
)
|
a8261ece
tangwang
检索效果优化
|
41
42
|
assert result.detected_language == "zh"
|
ef5baa86
tangwang
混杂语言处理
|
43
44
45
46
|
assert result.translations == {"en": "法式 dress 连衣裙-en"}
assert result.query_tokens == ["法式", "dress", "连衣裙"]
assert not hasattr(result, "query_text_by_lang")
assert not hasattr(result, "search_langs")
|
a8261ece
tangwang
检索效果优化
|
47
48
|
|
35da3813
tangwang
中英混写query的优化逻辑,不适...
|
49
|
def test_parse_mixed_en_query_translates_to_zh(monkeypatch):
|
ef5baa86
tangwang
混杂语言处理
|
50
|
parser = QueryParser(_build_config(), translator=_DummyTranslator(), tokenizer=_tokenizer)
|
a8261ece
tangwang
检索效果优化
|
51
|
monkeypatch.setattr(parser.language_detector, "detect", lambda text: "en")
|
a8261ece
tangwang
检索效果优化
|
52
|
|
ef5baa86
tangwang
混杂语言处理
|
53
54
55
56
57
58
|
result = parser.parse(
"red 连衣裙",
tenant_id="0",
generate_vector=False,
target_languages=["en", "zh"],
)
|
a8261ece
tangwang
检索效果优化
|
59
60
|
assert result.detected_language == "en"
|
ef5baa86
tangwang
混杂语言处理
|
61
62
|
assert result.translations == {"zh": "red 连衣裙-zh"}
assert result.query_tokens == ["red", "连衣裙"]
|
1556989b
tangwang
query翻译等待超时逻辑
|
63
64
65
66
|
def test_parse_waits_for_translation_when_source_in_index_languages(monkeypatch):
"""en 在 index_languages 内时仍应等待并采纳 en->zh 翻译结果(与向量共用预算)。"""
|
ef5baa86
tangwang
混杂语言处理
|
67
|
parser = QueryParser(_build_config(), translator=_DummyTranslator(), tokenizer=_tokenizer)
|
1556989b
tangwang
query翻译等待超时逻辑
|
68
|
monkeypatch.setattr(parser.language_detector, "detect", lambda text: "en")
|
1556989b
tangwang
query翻译等待超时逻辑
|
69
|
|
ef5baa86
tangwang
混杂语言处理
|
70
71
72
73
74
75
|
result = parser.parse(
"off shoulder top",
tenant_id="0",
generate_vector=False,
target_languages=["en", "zh"],
)
|
1556989b
tangwang
query翻译等待超时逻辑
|
76
77
|
assert result.detected_language == "en"
|
1556989b
tangwang
query翻译等待超时逻辑
|
78
|
assert result.translations.get("zh") == "off shoulder top-zh"
|
ef5baa86
tangwang
混杂语言处理
|
79
|
assert not hasattr(result, "source_in_index_languages")
|
45b39796
tangwang
qp性能优化
|
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
|
def test_parse_reuses_tokenization_across_tail_stages(monkeypatch):
tokenize_calls = []
def counting_tokenizer(text):
tokenize_calls.append(str(text))
return str(text).split()
config = SearchConfig(
es_index_name="test_products",
field_boosts={"title.en": 3.0, "title.zh": 3.0},
indexes=[IndexConfig(name="default", label="default", fields=["title.en", "title.zh"])],
query_config=QueryConfig(
enable_text_embedding=False,
enable_query_rewrite=False,
supported_languages=["en", "zh"],
default_language="en",
style_intent_terms={
"color": [
{"en_terms": ["black"], "zh_terms": ["黑色"], "attribute_terms": ["black"]}
],
},
style_intent_dimension_aliases={"color": ["color", "颜色"]},
product_title_exclusion_rules=[
{
"zh_trigger_terms": ["修身"],
"en_trigger_terms": ["fitted"],
"zh_title_exclusions": ["宽松"],
"en_title_exclusions": ["loose"],
}
],
),
function_score=FunctionScoreConfig(),
rerank=RerankConfig(),
spu_config=SPUConfig(enabled=False),
)
parser = QueryParser(
config,
translator=_DummyTranslator(),
tokenizer=counting_tokenizer,
)
monkeypatch.setattr(parser.language_detector, "detect", lambda text: "en")
result = parser.parse(
"black fitted dress",
tenant_id="0",
generate_vector=False,
target_languages=["en", "zh"],
)
assert result.translations == {"zh": "black fitted dress-zh"}
assert result.style_intent_profile is not None
assert result.style_intent_profile.is_active is True
assert result.product_title_exclusion_profile is not None
assert result.product_title_exclusion_profile.is_active is True
assert tokenize_calls == []
def test_parse_fast_path_detects_ascii_query_as_english_without_lingua(monkeypatch):
parser = QueryParser(_build_config(), translator=_DummyTranslator(), tokenizer=_tokenizer)
monkeypatch.setattr(
parser.language_detector,
"detect",
lambda text: (_ for _ in ()).throw(AssertionError("Lingua path should not be used")),
)
result = parser.parse(
"street t-shirt women",
tenant_id="0",
generate_vector=False,
target_languages=["en", "zh"],
)
assert result.detected_language == "en"
assert result.query_tokens == ["street", "t-shirt", "women"]
|