Blame view

embeddings/server.py 49.9 KB
7bfb9946   tangwang   向量化模块
1
2
3
  """
  Embedding service (FastAPI).
  
ed948666   tangwang   tidy
4
5
6
  API (simple list-in, list-out; aligned by index):
  - POST /embed/text   body: ["text1", "text2", ...] -> [[...], ...]
  - POST /embed/image  body: ["url_or_path1", ...]  -> [[...], ...]
7bfb9946   tangwang   向量化模块
7
8
  """
  
0a3764c4   tangwang   优化embedding模型加载
9
  import logging
07cf5a93   tangwang   START_EMBEDDING=...
10
  import os
4747e2f4   tangwang   embedding perform...
11
  import pathlib
7bfb9946   tangwang   向量化模块
12
  import threading
efd435cf   tangwang   tei性能调优:
13
  import time
4747e2f4   tangwang   embedding perform...
14
  import uuid
efd435cf   tangwang   tei性能调优:
15
16
  from collections import deque
  from dataclasses import dataclass
4747e2f4   tangwang   embedding perform...
17
  from logging.handlers import TimedRotatingFileHandler
7bfb9946   tangwang   向量化模块
18
19
20
  from typing import Any, Dict, List, Optional
  
  import numpy as np
4747e2f4   tangwang   embedding perform...
21
22
  from fastapi import FastAPI, HTTPException, Request, Response
  from fastapi.concurrency import run_in_threadpool
7bfb9946   tangwang   向量化模块
23
  
7214c2e7   tangwang   mplemented**
24
  from config.env_config import REDIS_CONFIG
4747e2f4   tangwang   embedding perform...
25
  from config.services_config import get_embedding_backend_config
7214c2e7   tangwang   mplemented**
26
  from embeddings.cache_keys import build_image_cache_key, build_text_cache_key
7bfb9946   tangwang   向量化模块
27
  from embeddings.config import CONFIG
c10f90fe   tangwang   cnclip
28
  from embeddings.protocols import ImageEncoderProtocol
7214c2e7   tangwang   mplemented**
29
  from embeddings.redis_embedding_cache import RedisEmbeddingCache
7bfb9946   tangwang   向量化模块
30
  
a7920e17   tangwang   项目名称和部署路径修改
31
  app = FastAPI(title="saas-search Embedding Service", version="1.0.0")
7bfb9946   tangwang   向量化模块
32
  
4747e2f4   tangwang   embedding perform...
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
  
  class _DefaultRequestIdFilter(logging.Filter):
      def filter(self, record: logging.LogRecord) -> bool:
          if not hasattr(record, "reqid"):
              record.reqid = "-1"
          return True
  
  
  def configure_embedding_logging() -> None:
      root_logger = logging.getLogger()
      if getattr(root_logger, "_embedding_logging_configured", False):
          return
  
      log_dir = pathlib.Path("logs")
      verbose_dir = log_dir / "verbose"
      log_dir.mkdir(exist_ok=True)
      verbose_dir.mkdir(parents=True, exist_ok=True)
  
      log_level = os.getenv("LOG_LEVEL", "INFO").upper()
      numeric_level = getattr(logging, log_level, logging.INFO)
      formatter = logging.Formatter(
          "%(asctime)s | reqid:%(reqid)s | %(name)s | %(levelname)s | %(message)s"
      )
      request_filter = _DefaultRequestIdFilter()
  
      root_logger.setLevel(numeric_level)
  
      file_handler = TimedRotatingFileHandler(
          filename=log_dir / "embedding_api.log",
          when="midnight",
          interval=1,
          backupCount=30,
          encoding="utf-8",
      )
      file_handler.setLevel(numeric_level)
      file_handler.setFormatter(formatter)
      file_handler.addFilter(request_filter)
      root_logger.addHandler(file_handler)
  
      error_handler = TimedRotatingFileHandler(
          filename=log_dir / "embedding_api_error.log",
          when="midnight",
          interval=1,
          backupCount=30,
          encoding="utf-8",
      )
      error_handler.setLevel(logging.ERROR)
      error_handler.setFormatter(formatter)
      error_handler.addFilter(request_filter)
      root_logger.addHandler(error_handler)
  
      verbose_logger = logging.getLogger("embedding.verbose")
      verbose_logger.setLevel(numeric_level)
      verbose_logger.handlers.clear()
      verbose_logger.propagate = False
  
      verbose_handler = TimedRotatingFileHandler(
          filename=verbose_dir / "embedding_verbose.log",
          when="midnight",
          interval=1,
          backupCount=30,
          encoding="utf-8",
      )
      verbose_handler.setLevel(numeric_level)
      verbose_handler.setFormatter(formatter)
      verbose_handler.addFilter(request_filter)
      verbose_logger.addHandler(verbose_handler)
  
      root_logger._embedding_logging_configured = True  # type: ignore[attr-defined]
  
  
  configure_embedding_logging()
  logger = logging.getLogger(__name__)
  verbose_logger = logging.getLogger("embedding.verbose")
  
0a3764c4   tangwang   优化embedding模型加载
108
  # Models are loaded at startup, not lazily
950a640e   tangwang   embeddings
109
  _text_model: Optional[Any] = None
c10f90fe   tangwang   cnclip
110
  _image_model: Optional[ImageEncoderProtocol] = None
07cf5a93   tangwang   START_EMBEDDING=...
111
  _text_backend_name: str = ""
7214c2e7   tangwang   mplemented**
112
113
114
115
116
117
118
119
120
  _SERVICE_KIND = (os.getenv("EMBEDDING_SERVICE_KIND", "all") or "all").strip().lower()
  if _SERVICE_KIND not in {"all", "text", "image"}:
      raise RuntimeError(
          f"Invalid EMBEDDING_SERVICE_KIND={_SERVICE_KIND!r}; expected all, text, or image"
      )
  _TEXT_ENABLED_BY_ENV = os.getenv("EMBEDDING_ENABLE_TEXT_MODEL", "true").lower() in ("1", "true", "yes")
  _IMAGE_ENABLED_BY_ENV = os.getenv("EMBEDDING_ENABLE_IMAGE_MODEL", "true").lower() in ("1", "true", "yes")
  open_text_model = _TEXT_ENABLED_BY_ENV and _SERVICE_KIND in {"all", "text"}
  open_image_model = _IMAGE_ENABLED_BY_ENV and _SERVICE_KIND in {"all", "image"}
7bfb9946   tangwang   向量化模块
121
122
123
124
  
  _text_encode_lock = threading.Lock()
  _image_encode_lock = threading.Lock()
  
4747e2f4   tangwang   embedding perform...
125
126
127
128
129
130
131
  _TEXT_MICROBATCH_WINDOW_SEC = max(
      0.0, float(os.getenv("TEXT_MICROBATCH_WINDOW_MS", "4")) / 1000.0
  )
  _TEXT_REQUEST_TIMEOUT_SEC = max(
      1.0, float(os.getenv("TEXT_REQUEST_TIMEOUT_SEC", "30"))
  )
  _TEXT_MAX_INFLIGHT = max(1, int(os.getenv("TEXT_MAX_INFLIGHT", "32")))
16204531   tangwang   docs
132
  _IMAGE_MAX_INFLIGHT = max(1, int(os.getenv("IMAGE_MAX_INFLIGHT", "20")))
4747e2f4   tangwang   embedding perform...
133
134
135
136
137
  _OVERLOAD_STATUS_CODE = int(os.getenv("EMBEDDING_OVERLOAD_STATUS_CODE", "503"))
  _LOG_PREVIEW_COUNT = max(1, int(os.getenv("EMBEDDING_LOG_PREVIEW_COUNT", "3")))
  _LOG_TEXT_PREVIEW_CHARS = max(32, int(os.getenv("EMBEDDING_LOG_TEXT_PREVIEW_CHARS", "120")))
  _LOG_IMAGE_PREVIEW_CHARS = max(32, int(os.getenv("EMBEDDING_LOG_IMAGE_PREVIEW_CHARS", "180")))
  _VECTOR_PREVIEW_DIMS = max(1, int(os.getenv("EMBEDDING_VECTOR_PREVIEW_DIMS", "6")))
7214c2e7   tangwang   mplemented**
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
  _CACHE_PREFIX = str(REDIS_CONFIG.get("embedding_cache_prefix", "embedding")).strip() or "embedding"
  
  
  @dataclass
  class _EmbedResult:
      vectors: List[Optional[List[float]]]
      cache_hits: int
      cache_misses: int
      backend_elapsed_ms: float
      mode: str
  
  
  class _EndpointStats:
      def __init__(self, name: str):
          self.name = name
          self._lock = threading.Lock()
          self.request_total = 0
          self.success_total = 0
          self.failure_total = 0
          self.rejected_total = 0
          self.cache_hits = 0
          self.cache_misses = 0
          self.total_latency_ms = 0.0
          self.total_backend_latency_ms = 0.0
  
      def record_rejected(self) -> None:
          with self._lock:
              self.request_total += 1
              self.rejected_total += 1
  
      def record_completed(
          self,
          *,
          success: bool,
          latency_ms: float,
          backend_latency_ms: float,
          cache_hits: int,
          cache_misses: int,
      ) -> None:
          with self._lock:
              self.request_total += 1
              if success:
                  self.success_total += 1
              else:
                  self.failure_total += 1
              self.cache_hits += max(0, int(cache_hits))
              self.cache_misses += max(0, int(cache_misses))
              self.total_latency_ms += max(0.0, float(latency_ms))
              self.total_backend_latency_ms += max(0.0, float(backend_latency_ms))
  
      def snapshot(self) -> Dict[str, Any]:
          with self._lock:
              completed = self.success_total + self.failure_total
              return {
                  "request_total": self.request_total,
                  "success_total": self.success_total,
                  "failure_total": self.failure_total,
                  "rejected_total": self.rejected_total,
                  "cache_hits": self.cache_hits,
                  "cache_misses": self.cache_misses,
                  "avg_latency_ms": round(self.total_latency_ms / completed, 3) if completed else 0.0,
                  "avg_backend_latency_ms": round(self.total_backend_latency_ms / completed, 3)
                  if completed
                  else 0.0,
              }
4747e2f4   tangwang   embedding perform...
203
204
205
206
207
208
  
  
  class _InflightLimiter:
      def __init__(self, name: str, limit: int):
          self.name = name
          self.limit = max(1, int(limit))
4747e2f4   tangwang   embedding perform...
209
210
211
212
213
214
          self._lock = threading.Lock()
          self._active = 0
          self._rejected = 0
          self._completed = 0
          self._failed = 0
          self._max_active = 0
b754fd41   tangwang   图片向量化支持优先级参数
215
          self._priority_bypass_total = 0
4747e2f4   tangwang   embedding perform...
216
  
b754fd41   tangwang   图片向量化支持优先级参数
217
218
219
      def try_acquire(self, *, bypass_limit: bool = False) -> tuple[bool, int]:
          with self._lock:
              if not bypass_limit and self._active >= self.limit:
4747e2f4   tangwang   embedding perform...
220
221
                  self._rejected += 1
                  active = self._active
b754fd41   tangwang   图片向量化支持优先级参数
222
                  return False, active
4747e2f4   tangwang   embedding perform...
223
224
              self._active += 1
              self._max_active = max(self._max_active, self._active)
b754fd41   tangwang   图片向量化支持优先级参数
225
226
              if bypass_limit:
                  self._priority_bypass_total += 1
4747e2f4   tangwang   embedding perform...
227
228
229
230
231
232
233
234
235
236
237
              active = self._active
          return True, active
  
      def release(self, *, success: bool) -> int:
          with self._lock:
              self._active = max(0, self._active - 1)
              if success:
                  self._completed += 1
              else:
                  self._failed += 1
              active = self._active
4747e2f4   tangwang   embedding perform...
238
239
240
241
242
243
244
245
246
247
248
          return active
  
      def snapshot(self) -> Dict[str, int]:
          with self._lock:
              return {
                  "limit": self.limit,
                  "active": self._active,
                  "rejected_total": self._rejected,
                  "completed_total": self._completed,
                  "failed_total": self._failed,
                  "max_active": self._max_active,
b754fd41   tangwang   图片向量化支持优先级参数
249
                  "priority_bypass_total": self._priority_bypass_total,
4747e2f4   tangwang   embedding perform...
250
251
252
              }
  
  
b754fd41   tangwang   图片向量化支持优先级参数
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
  def _effective_priority(priority: int) -> int:
      return 1 if int(priority) > 0 else 0
  
  
  def _priority_label(priority: int) -> str:
      return "high" if _effective_priority(priority) > 0 else "normal"
  
  
  @dataclass
  class _TextDispatchTask:
      normalized: List[str]
      effective_normalize: bool
      request_id: str
      priority: int
      created_at: float
      done: threading.Event
      result: Optional[_EmbedResult] = None
      error: Optional[Exception] = None
  
  
  _text_dispatch_high_queue: "deque[_TextDispatchTask]" = deque()
  _text_dispatch_normal_queue: "deque[_TextDispatchTask]" = deque()
  _text_dispatch_cv = threading.Condition()
  _text_dispatch_workers: List[threading.Thread] = []
  _text_dispatch_worker_stop = False
  _text_dispatch_worker_count = 0
  
  
  def _text_dispatch_queue_depth() -> Dict[str, int]:
      with _text_dispatch_cv:
          return {
              "high": len(_text_dispatch_high_queue),
              "normal": len(_text_dispatch_normal_queue),
              "total": len(_text_dispatch_high_queue) + len(_text_dispatch_normal_queue),
          }
  
  
  def _pop_text_dispatch_task_locked() -> Optional["_TextDispatchTask"]:
      if _text_dispatch_high_queue:
          return _text_dispatch_high_queue.popleft()
      if _text_dispatch_normal_queue:
          return _text_dispatch_normal_queue.popleft()
      return None
  
  
  def _start_text_dispatch_workers() -> None:
      global _text_dispatch_workers, _text_dispatch_worker_stop, _text_dispatch_worker_count
      if _text_model is None:
          return
      target_worker_count = 1 if _text_backend_name == "local_st" else _TEXT_MAX_INFLIGHT
      alive_workers = [worker for worker in _text_dispatch_workers if worker.is_alive()]
      if len(alive_workers) == target_worker_count:
          _text_dispatch_workers = alive_workers
          _text_dispatch_worker_count = target_worker_count
          return
      _text_dispatch_worker_stop = False
      _text_dispatch_worker_count = target_worker_count
      _text_dispatch_workers = []
      for idx in range(target_worker_count):
          worker = threading.Thread(
              target=_text_dispatch_worker_loop,
              args=(idx,),
              name=f"embed-text-dispatch-{idx}",
              daemon=True,
          )
          worker.start()
          _text_dispatch_workers.append(worker)
      logger.info(
          "Started text dispatch workers | backend=%s workers=%d",
          _text_backend_name,
          target_worker_count,
      )
  
  
  def _stop_text_dispatch_workers() -> None:
      global _text_dispatch_worker_stop
      with _text_dispatch_cv:
          _text_dispatch_worker_stop = True
          _text_dispatch_cv.notify_all()
  
  
  def _text_dispatch_worker_loop(worker_idx: int) -> None:
      while True:
          with _text_dispatch_cv:
              while (
                  not _text_dispatch_high_queue
                  and not _text_dispatch_normal_queue
                  and not _text_dispatch_worker_stop
              ):
                  _text_dispatch_cv.wait()
              if _text_dispatch_worker_stop:
                  return
              task = _pop_text_dispatch_task_locked()
          if task is None:
              continue
          try:
              queue_wait_ms = (time.perf_counter() - task.created_at) * 1000.0
              logger.info(
                  "text dispatch start | worker=%d priority=%s inputs=%d queue_wait_ms=%.2f",
                  worker_idx,
                  _priority_label(task.priority),
                  len(task.normalized),
                  queue_wait_ms,
                  extra=_request_log_extra(task.request_id),
              )
              task.result = _embed_text_impl(
                  task.normalized,
                  task.effective_normalize,
                  task.request_id,
                  task.priority,
              )
          except Exception as exc:
              task.error = exc
          finally:
              task.done.set()
  
  
  def _submit_text_dispatch_and_wait(
      normalized: List[str],
      effective_normalize: bool,
      request_id: str,
      priority: int,
  ) -> _EmbedResult:
      if not any(worker.is_alive() for worker in _text_dispatch_workers):
          _start_text_dispatch_workers()
      task = _TextDispatchTask(
          normalized=normalized,
          effective_normalize=effective_normalize,
          request_id=request_id,
          priority=_effective_priority(priority),
          created_at=time.perf_counter(),
          done=threading.Event(),
      )
      with _text_dispatch_cv:
          if task.priority > 0:
              _text_dispatch_high_queue.append(task)
          else:
              _text_dispatch_normal_queue.append(task)
          _text_dispatch_cv.notify()
      task.done.wait()
      if task.error is not None:
          raise task.error
      if task.result is None:
          raise RuntimeError("Text dispatch worker returned empty result")
      return task.result
  
  
4747e2f4   tangwang   embedding perform...
400
401
  _text_request_limiter = _InflightLimiter(name="text", limit=_TEXT_MAX_INFLIGHT)
  _image_request_limiter = _InflightLimiter(name="image", limit=_IMAGE_MAX_INFLIGHT)
7214c2e7   tangwang   mplemented**
402
403
404
405
  _text_stats = _EndpointStats(name="text")
  _image_stats = _EndpointStats(name="image")
  _text_cache = RedisEmbeddingCache(key_prefix=_CACHE_PREFIX, namespace="")
  _image_cache = RedisEmbeddingCache(key_prefix=_CACHE_PREFIX, namespace="image")
4747e2f4   tangwang   embedding perform...
406
  
7bfb9946   tangwang   向量化模块
407
  
efd435cf   tangwang   tei性能调优:
408
409
410
411
  @dataclass
  class _SingleTextTask:
      text: str
      normalize: bool
b754fd41   tangwang   图片向量化支持优先级参数
412
      priority: int
efd435cf   tangwang   tei性能调优:
413
      created_at: float
4747e2f4   tangwang   embedding perform...
414
      request_id: str
efd435cf   tangwang   tei性能调优:
415
416
417
418
419
      done: threading.Event
      result: Optional[List[float]] = None
      error: Optional[Exception] = None
  
  
b754fd41   tangwang   图片向量化支持优先级参数
420
421
  _text_single_high_queue: "deque[_SingleTextTask]" = deque()
  _text_single_normal_queue: "deque[_SingleTextTask]" = deque()
efd435cf   tangwang   tei性能调优:
422
423
424
  _text_single_queue_cv = threading.Condition()
  _text_batch_worker: Optional[threading.Thread] = None
  _text_batch_worker_stop = False
28e57bb1   tangwang   日志体系优化
425
426
  
  
b754fd41   tangwang   图片向量化支持优先级参数
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
  def _text_microbatch_queue_depth() -> Dict[str, int]:
      with _text_single_queue_cv:
          return {
              "high": len(_text_single_high_queue),
              "normal": len(_text_single_normal_queue),
              "total": len(_text_single_high_queue) + len(_text_single_normal_queue),
          }
  
  
  def _pop_single_text_task_locked() -> Optional["_SingleTextTask"]:
      if _text_single_high_queue:
          return _text_single_high_queue.popleft()
      if _text_single_normal_queue:
          return _text_single_normal_queue.popleft()
      return None
  
  
28e57bb1   tangwang   日志体系优化
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
  def _compact_preview(text: str, max_chars: int) -> str:
      compact = " ".join((text or "").split())
      if len(compact) <= max_chars:
          return compact
      return compact[:max_chars] + "..."
  
  
  def _preview_inputs(items: List[str], max_items: int, max_chars: int) -> List[Dict[str, Any]]:
      previews: List[Dict[str, Any]] = []
      for idx, item in enumerate(items[:max_items]):
          previews.append(
              {
                  "idx": idx,
                  "len": len(item),
                  "preview": _compact_preview(item, max_chars),
              }
          )
      return previews
efd435cf   tangwang   tei性能调优:
462
463
  
  
4747e2f4   tangwang   embedding perform...
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
  def _preview_vector(vec: Optional[List[float]], max_dims: int = _VECTOR_PREVIEW_DIMS) -> List[float]:
      if not vec:
          return []
      return [round(float(v), 6) for v in vec[:max_dims]]
  
  
  def _request_log_extra(request_id: str) -> Dict[str, str]:
      return {"reqid": request_id}
  
  
  def _resolve_request_id(http_request: Request) -> str:
      header_value = http_request.headers.get("X-Request-ID")
      if header_value and header_value.strip():
          return header_value.strip()[:32]
      return str(uuid.uuid4())[:8]
  
  
  def _request_client(http_request: Request) -> str:
      client = getattr(http_request, "client", None)
      host = getattr(client, "host", None)
      return str(host or "-")
  
  
efd435cf   tangwang   tei性能调优:
487
488
  def _encode_local_st(texts: List[str], normalize_embeddings: bool) -> Any:
      with _text_encode_lock:
77516841   tangwang   tidy embeddings
489
          return _text_model.encode(
efd435cf   tangwang   tei性能调优:
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
              texts,
              batch_size=int(CONFIG.TEXT_BATCH_SIZE),
              device=CONFIG.TEXT_DEVICE,
              normalize_embeddings=normalize_embeddings,
          )
  
  
  def _start_text_batch_worker() -> None:
      global _text_batch_worker, _text_batch_worker_stop
      if _text_batch_worker is not None and _text_batch_worker.is_alive():
          return
      _text_batch_worker_stop = False
      _text_batch_worker = threading.Thread(
          target=_text_batch_worker_loop,
          name="embed-text-microbatch-worker",
          daemon=True,
      )
      _text_batch_worker.start()
      logger.info(
          "Started local_st text micro-batch worker | window_ms=%.1f max_batch=%d",
          _TEXT_MICROBATCH_WINDOW_SEC * 1000.0,
          int(CONFIG.TEXT_BATCH_SIZE),
      )
  
  
  def _stop_text_batch_worker() -> None:
      global _text_batch_worker_stop
      with _text_single_queue_cv:
          _text_batch_worker_stop = True
          _text_single_queue_cv.notify_all()
  
  
  def _text_batch_worker_loop() -> None:
      max_batch = max(1, int(CONFIG.TEXT_BATCH_SIZE))
      while True:
          with _text_single_queue_cv:
b754fd41   tangwang   图片向量化支持优先级参数
526
527
528
529
530
              while (
                  not _text_single_high_queue
                  and not _text_single_normal_queue
                  and not _text_batch_worker_stop
              ):
efd435cf   tangwang   tei性能调优:
531
532
533
534
                  _text_single_queue_cv.wait()
              if _text_batch_worker_stop:
                  return
  
b754fd41   tangwang   图片向量化支持优先级参数
535
536
537
538
              first_task = _pop_single_text_task_locked()
              if first_task is None:
                  continue
              batch: List[_SingleTextTask] = [first_task]
efd435cf   tangwang   tei性能调优:
539
540
541
542
543
544
              deadline = time.perf_counter() + _TEXT_MICROBATCH_WINDOW_SEC
  
              while len(batch) < max_batch:
                  remaining = deadline - time.perf_counter()
                  if remaining <= 0:
                      break
b754fd41   tangwang   图片向量化支持优先级参数
545
                  if not _text_single_high_queue and not _text_single_normal_queue:
efd435cf   tangwang   tei性能调优:
546
547
                      _text_single_queue_cv.wait(timeout=remaining)
                      continue
b754fd41   tangwang   图片向量化支持优先级参数
548
549
550
551
552
                  while len(batch) < max_batch:
                      next_task = _pop_single_text_task_locked()
                      if next_task is None:
                          break
                      batch.append(next_task)
efd435cf   tangwang   tei性能调优:
553
554
  
          try:
4747e2f4   tangwang   embedding perform...
555
556
557
              queue_wait_ms = [(time.perf_counter() - task.created_at) * 1000.0 for task in batch]
              reqids = [task.request_id for task in batch]
              logger.info(
b754fd41   tangwang   图片向量化支持优先级参数
558
                  "text microbatch dispatch | size=%d priority=%s queue_wait_ms_min=%.2f queue_wait_ms_max=%.2f reqids=%s preview=%s",
4747e2f4   tangwang   embedding perform...
559
                  len(batch),
b754fd41   tangwang   图片向量化支持优先级参数
560
                  _priority_label(max(task.priority for task in batch)),
4747e2f4   tangwang   embedding perform...
561
562
563
564
565
566
567
568
569
570
                  min(queue_wait_ms) if queue_wait_ms else 0.0,
                  max(queue_wait_ms) if queue_wait_ms else 0.0,
                  reqids,
                  _preview_inputs(
                      [task.text for task in batch],
                      _LOG_PREVIEW_COUNT,
                      _LOG_TEXT_PREVIEW_CHARS,
                  ),
              )
              batch_t0 = time.perf_counter()
efd435cf   tangwang   tei性能调优:
571
572
573
574
575
576
577
578
579
580
581
              embs = _encode_local_st([task.text for task in batch], normalize_embeddings=False)
              if embs is None or len(embs) != len(batch):
                  raise RuntimeError(
                      f"Text model response length mismatch in micro-batch: "
                      f"expected {len(batch)}, got {0 if embs is None else len(embs)}"
                  )
              for task, emb in zip(batch, embs):
                  vec = _as_list(emb, normalize=task.normalize)
                  if vec is None:
                      raise RuntimeError("Text model returned empty embedding in micro-batch")
                  task.result = vec
4747e2f4   tangwang   embedding perform...
582
583
584
585
586
587
588
              logger.info(
                  "text microbatch done | size=%d reqids=%s dim=%d backend_elapsed_ms=%.2f",
                  len(batch),
                  reqids,
                  len(batch[0].result) if batch and batch[0].result is not None else 0,
                  (time.perf_counter() - batch_t0) * 1000.0,
              )
efd435cf   tangwang   tei性能调优:
589
          except Exception as exc:
4747e2f4   tangwang   embedding perform...
590
591
592
593
594
595
596
              logger.error(
                  "text microbatch failed | size=%d reqids=%s error=%s",
                  len(batch),
                  [task.request_id for task in batch],
                  exc,
                  exc_info=True,
              )
efd435cf   tangwang   tei性能调优:
597
598
599
600
601
602
603
              for task in batch:
                  task.error = exc
          finally:
              for task in batch:
                  task.done.set()
  
  
b754fd41   tangwang   图片向量化支持优先级参数
604
605
606
607
608
609
  def _encode_single_text_with_microbatch(
      text: str,
      normalize: bool,
      request_id: str,
      priority: int,
  ) -> List[float]:
efd435cf   tangwang   tei性能调优:
610
611
612
      task = _SingleTextTask(
          text=text,
          normalize=normalize,
b754fd41   tangwang   图片向量化支持优先级参数
613
          priority=_effective_priority(priority),
efd435cf   tangwang   tei性能调优:
614
          created_at=time.perf_counter(),
4747e2f4   tangwang   embedding perform...
615
          request_id=request_id,
efd435cf   tangwang   tei性能调优:
616
617
618
          done=threading.Event(),
      )
      with _text_single_queue_cv:
b754fd41   tangwang   图片向量化支持优先级参数
619
620
621
622
          if task.priority > 0:
              _text_single_high_queue.append(task)
          else:
              _text_single_normal_queue.append(task)
efd435cf   tangwang   tei性能调优:
623
624
625
626
          _text_single_queue_cv.notify()
  
      if not task.done.wait(timeout=_TEXT_REQUEST_TIMEOUT_SEC):
          with _text_single_queue_cv:
b754fd41   tangwang   图片向量化支持优先级参数
627
              queue = _text_single_high_queue if task.priority > 0 else _text_single_normal_queue
efd435cf   tangwang   tei性能调优:
628
              try:
b754fd41   tangwang   图片向量化支持优先级参数
629
                  queue.remove(task)
efd435cf   tangwang   tei性能调优:
630
631
632
633
634
635
636
637
638
639
640
641
              except ValueError:
                  pass
          raise RuntimeError(
              f"Timed out waiting for text micro-batch worker ({_TEXT_REQUEST_TIMEOUT_SEC:.1f}s)"
          )
      if task.error is not None:
          raise task.error
      if task.result is None:
          raise RuntimeError("Text micro-batch worker returned empty result")
      return task.result
  
  
0a3764c4   tangwang   优化embedding模型加载
642
643
644
  @app.on_event("startup")
  def load_models():
      """Load models at service startup to avoid first-request latency."""
07cf5a93   tangwang   START_EMBEDDING=...
645
      global _text_model, _image_model, _text_backend_name
7bfb9946   tangwang   向量化模块
646
  
7214c2e7   tangwang   mplemented**
647
648
649
650
651
652
      logger.info(
          "Loading embedding models at startup | service_kind=%s text_enabled=%s image_enabled=%s",
          _SERVICE_KIND,
          open_text_model,
          open_image_model,
      )
7bfb9946   tangwang   向量化模块
653
  
40f1e391   tangwang   cnclip
654
655
      if open_text_model:
          try:
07cf5a93   tangwang   START_EMBEDDING=...
656
657
658
              backend_name, backend_cfg = get_embedding_backend_config()
              _text_backend_name = backend_name
              if backend_name == "tei":
77516841   tangwang   tidy embeddings
659
                  from embeddings.text_embedding_tei import TEITextModel
07cf5a93   tangwang   START_EMBEDDING=...
660
  
86d8358b   tangwang   config optimize
661
662
                  base_url = backend_cfg.get("base_url") or CONFIG.TEI_BASE_URL
                  timeout_sec = int(backend_cfg.get("timeout_sec") or CONFIG.TEI_TIMEOUT_SEC)
07cf5a93   tangwang   START_EMBEDDING=...
663
664
665
666
667
668
                  logger.info("Loading text backend: tei (base_url=%s)", base_url)
                  _text_model = TEITextModel(
                      base_url=str(base_url),
                      timeout_sec=timeout_sec,
                  )
              elif backend_name == "local_st":
77516841   tangwang   tidy embeddings
669
                  from embeddings.text_embedding_sentence_transformers import Qwen3TextModel
950a640e   tangwang   embeddings
670
  
86d8358b   tangwang   config optimize
671
                  model_id = backend_cfg.get("model_id") or CONFIG.TEXT_MODEL_ID
07cf5a93   tangwang   START_EMBEDDING=...
672
673
                  logger.info("Loading text backend: local_st (model=%s)", model_id)
                  _text_model = Qwen3TextModel(model_id=str(model_id))
efd435cf   tangwang   tei性能调优:
674
                  _start_text_batch_worker()
07cf5a93   tangwang   START_EMBEDDING=...
675
676
677
678
679
              else:
                  raise ValueError(
                      f"Unsupported embedding backend: {backend_name}. "
                      "Supported: tei, local_st"
                  )
b754fd41   tangwang   图片向量化支持优先级参数
680
              _start_text_dispatch_workers()
07cf5a93   tangwang   START_EMBEDDING=...
681
              logger.info("Text backend loaded successfully: %s", _text_backend_name)
40f1e391   tangwang   cnclip
682
          except Exception as e:
4747e2f4   tangwang   embedding perform...
683
              logger.error("Failed to load text model: %s", e, exc_info=True)
40f1e391   tangwang   cnclip
684
              raise
0a3764c4   tangwang   优化embedding模型加载
685
  
40f1e391   tangwang   cnclip
686
687
      if open_image_model:
          try:
c10f90fe   tangwang   cnclip
688
              if CONFIG.USE_CLIP_AS_SERVICE:
950a640e   tangwang   embeddings
689
690
                  from embeddings.clip_as_service_encoder import ClipAsServiceImageEncoder
  
4747e2f4   tangwang   embedding perform...
691
692
693
694
695
                  logger.info(
                      "Loading image encoder via clip-as-service: %s (configured model: %s)",
                      CONFIG.CLIP_AS_SERVICE_SERVER,
                      CONFIG.CLIP_AS_SERVICE_MODEL_NAME,
                  )
c10f90fe   tangwang   cnclip
696
697
698
699
700
701
                  _image_model = ClipAsServiceImageEncoder(
                      server=CONFIG.CLIP_AS_SERVICE_SERVER,
                      batch_size=CONFIG.IMAGE_BATCH_SIZE,
                  )
                  logger.info("Image model (clip-as-service) loaded successfully")
              else:
950a640e   tangwang   embeddings
702
703
                  from embeddings.clip_model import ClipImageModel
  
4747e2f4   tangwang   embedding perform...
704
705
706
707
708
                  logger.info(
                      "Loading local image model: %s (device: %s)",
                      CONFIG.IMAGE_MODEL_NAME,
                      CONFIG.IMAGE_DEVICE,
                  )
c10f90fe   tangwang   cnclip
709
710
711
712
713
                  _image_model = ClipImageModel(
                      model_name=CONFIG.IMAGE_MODEL_NAME,
                      device=CONFIG.IMAGE_DEVICE,
                  )
                  logger.info("Image model (local CN-CLIP) loaded successfully")
40f1e391   tangwang   cnclip
714
          except Exception as e:
ed948666   tangwang   tidy
715
716
              logger.error("Failed to load image model: %s", e, exc_info=True)
              raise
0a3764c4   tangwang   优化embedding模型加载
717
718
  
      logger.info("All embedding models loaded successfully, service ready")
7bfb9946   tangwang   向量化模块
719
720
  
  
efd435cf   tangwang   tei性能调优:
721
722
723
  @app.on_event("shutdown")
  def stop_workers() -> None:
      _stop_text_batch_worker()
b754fd41   tangwang   图片向量化支持优先级参数
724
      _stop_text_dispatch_workers()
efd435cf   tangwang   tei性能调优:
725
726
  
  
200fdddf   tangwang   embed norm
727
728
729
730
731
732
733
734
  def _normalize_vector(vec: np.ndarray) -> np.ndarray:
      norm = float(np.linalg.norm(vec))
      if not np.isfinite(norm) or norm <= 0.0:
          raise RuntimeError("Embedding vector has invalid norm (must be > 0)")
      return vec / norm
  
  
  def _as_list(embedding: Optional[np.ndarray], normalize: bool = False) -> Optional[List[float]]:
7bfb9946   tangwang   向量化模块
735
736
737
738
739
740
      if embedding is None:
          return None
      if not isinstance(embedding, np.ndarray):
          embedding = np.array(embedding, dtype=np.float32)
      if embedding.ndim != 1:
          embedding = embedding.reshape(-1)
200fdddf   tangwang   embed norm
741
742
743
744
      embedding = embedding.astype(np.float32, copy=False)
      if normalize:
          embedding = _normalize_vector(embedding).astype(np.float32, copy=False)
      return embedding.tolist()
7bfb9946   tangwang   向量化模块
745
746
  
  
7214c2e7   tangwang   mplemented**
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
  def _try_full_text_cache_hit(
      normalized: List[str],
      effective_normalize: bool,
  ) -> Optional[_EmbedResult]:
      out: List[Optional[List[float]]] = []
      for text in normalized:
          cached = _text_cache.get(build_text_cache_key(text, normalize=effective_normalize))
          if cached is None:
              return None
          vec = _as_list(cached, normalize=False)
          if vec is None:
              return None
          out.append(vec)
      return _EmbedResult(
          vectors=out,
          cache_hits=len(out),
          cache_misses=0,
          backend_elapsed_ms=0.0,
          mode="cache-only",
      )
  
  
  def _try_full_image_cache_hit(
      urls: List[str],
      effective_normalize: bool,
  ) -> Optional[_EmbedResult]:
      out: List[Optional[List[float]]] = []
      for url in urls:
          cached = _image_cache.get(build_image_cache_key(url, normalize=effective_normalize))
          if cached is None:
              return None
          vec = _as_list(cached, normalize=False)
          if vec is None:
              return None
          out.append(vec)
      return _EmbedResult(
          vectors=out,
          cache_hits=len(out),
          cache_misses=0,
          backend_elapsed_ms=0.0,
          mode="cache-only",
      )
  
  
7bfb9946   tangwang   向量化模块
791
792
  @app.get("/health")
  def health() -> Dict[str, Any]:
4747e2f4   tangwang   embedding perform...
793
      """Health check endpoint. Returns status and current throttling stats."""
7214c2e7   tangwang   mplemented**
794
      ready = (not open_text_model or _text_model is not None) and (not open_image_model or _image_model is not None)
b754fd41   tangwang   图片向量化支持优先级参数
795
796
      text_dispatch_depth = _text_dispatch_queue_depth()
      text_microbatch_depth = _text_microbatch_queue_depth()
0a3764c4   tangwang   优化embedding模型加载
797
      return {
7214c2e7   tangwang   mplemented**
798
799
          "status": "ok" if ready else "degraded",
          "service_kind": _SERVICE_KIND,
0a3764c4   tangwang   优化embedding模型加载
800
          "text_model_loaded": _text_model is not None,
07cf5a93   tangwang   START_EMBEDDING=...
801
          "text_backend": _text_backend_name,
0a3764c4   tangwang   优化embedding模型加载
802
          "image_model_loaded": _image_model is not None,
7214c2e7   tangwang   mplemented**
803
804
805
806
          "cache_enabled": {
              "text": _text_cache.redis_client is not None,
              "image": _image_cache.redis_client is not None,
          },
4747e2f4   tangwang   embedding perform...
807
808
809
810
          "limits": {
              "text": _text_request_limiter.snapshot(),
              "image": _image_request_limiter.snapshot(),
          },
7214c2e7   tangwang   mplemented**
811
812
813
814
          "stats": {
              "text": _text_stats.snapshot(),
              "image": _image_stats.snapshot(),
          },
b754fd41   tangwang   图片向量化支持优先级参数
815
816
817
818
819
820
821
          "text_dispatch": {
              "workers": _text_dispatch_worker_count,
              "workers_alive": sum(1 for worker in _text_dispatch_workers if worker.is_alive()),
              "queue_depth": text_dispatch_depth["total"],
              "queue_depth_high": text_dispatch_depth["high"],
              "queue_depth_normal": text_dispatch_depth["normal"],
          },
4747e2f4   tangwang   embedding perform...
822
823
          "text_microbatch": {
              "window_ms": round(_TEXT_MICROBATCH_WINDOW_SEC * 1000.0, 3),
b754fd41   tangwang   图片向量化支持优先级参数
824
825
826
              "queue_depth": text_microbatch_depth["total"],
              "queue_depth_high": text_microbatch_depth["high"],
              "queue_depth_normal": text_microbatch_depth["normal"],
4747e2f4   tangwang   embedding perform...
827
828
829
              "worker_alive": bool(_text_batch_worker is not None and _text_batch_worker.is_alive()),
              "request_timeout_sec": _TEXT_REQUEST_TIMEOUT_SEC,
          },
0a3764c4   tangwang   优化embedding模型加载
830
      }
7bfb9946   tangwang   向量化模块
831
832
  
  
7214c2e7   tangwang   mplemented**
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
  @app.get("/ready")
  def ready() -> Dict[str, Any]:
      text_ready = (not open_text_model) or (_text_model is not None)
      image_ready = (not open_image_model) or (_image_model is not None)
      if not (text_ready and image_ready):
          raise HTTPException(
              status_code=503,
              detail={
                  "service_kind": _SERVICE_KIND,
                  "text_ready": text_ready,
                  "image_ready": image_ready,
              },
          )
      return {
          "status": "ready",
          "service_kind": _SERVICE_KIND,
          "text_ready": text_ready,
          "image_ready": image_ready,
      }
  
  
4747e2f4   tangwang   embedding perform...
854
855
856
857
  def _embed_text_impl(
      normalized: List[str],
      effective_normalize: bool,
      request_id: str,
b754fd41   tangwang   图片向量化支持优先级参数
858
      priority: int = 0,
7214c2e7   tangwang   mplemented**
859
  ) -> _EmbedResult:
0a3764c4   tangwang   优化embedding模型加载
860
861
      if _text_model is None:
          raise RuntimeError("Text model not loaded")
28e57bb1   tangwang   日志体系优化
862
  
7214c2e7   tangwang   mplemented**
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
      out: List[Optional[List[float]]] = [None] * len(normalized)
      missing_indices: List[int] = []
      missing_texts: List[str] = []
      missing_cache_keys: List[str] = []
      cache_hits = 0
      for idx, text in enumerate(normalized):
          cache_key = build_text_cache_key(text, normalize=effective_normalize)
          cached = _text_cache.get(cache_key)
          if cached is not None:
              vec = _as_list(cached, normalize=False)
              if vec is not None:
                  out[idx] = vec
                  cache_hits += 1
                  continue
          missing_indices.append(idx)
          missing_texts.append(text)
          missing_cache_keys.append(cache_key)
  
      if not missing_texts:
          logger.info(
              "text backend done | backend=%s mode=cache-only inputs=%d normalize=%s dim=%d cache_hits=%d cache_misses=0 backend_elapsed_ms=0.00",
              _text_backend_name,
              len(normalized),
              effective_normalize,
              len(out[0]) if out and out[0] is not None else 0,
              cache_hits,
              extra=_request_log_extra(request_id),
          )
          return _EmbedResult(
              vectors=out,
              cache_hits=cache_hits,
              cache_misses=0,
              backend_elapsed_ms=0.0,
              mode="cache-only",
          )
  
      backend_t0 = time.perf_counter()
54ccf28c   tangwang   tei
900
      try:
efd435cf   tangwang   tei性能调优:
901
          if _text_backend_name == "local_st":
7214c2e7   tangwang   mplemented**
902
903
              if len(missing_texts) == 1 and _text_batch_worker is not None:
                  computed = [
4747e2f4   tangwang   embedding perform...
904
                      _encode_single_text_with_microbatch(
7214c2e7   tangwang   mplemented**
905
                          missing_texts[0],
4747e2f4   tangwang   embedding perform...
906
907
                          normalize=effective_normalize,
                          request_id=request_id,
b754fd41   tangwang   图片向量化支持优先级参数
908
                          priority=priority,
4747e2f4   tangwang   embedding perform...
909
910
                      )
                  ]
7214c2e7   tangwang   mplemented**
911
912
913
914
915
916
917
918
919
920
                  mode = "microbatch-single"
              else:
                  embs = _encode_local_st(missing_texts, normalize_embeddings=False)
                  computed = []
                  for i, emb in enumerate(embs):
                      vec = _as_list(emb, normalize=effective_normalize)
                      if vec is None:
                          raise RuntimeError(f"Text model returned empty embedding for missing index {i}")
                      computed.append(vec)
                  mode = "direct-batch"
efd435cf   tangwang   tei性能调优:
921
          else:
77516841   tangwang   tidy embeddings
922
              embs = _text_model.encode(
7214c2e7   tangwang   mplemented**
923
                  missing_texts,
54ccf28c   tangwang   tei
924
925
                  batch_size=int(CONFIG.TEXT_BATCH_SIZE),
                  device=CONFIG.TEXT_DEVICE,
200fdddf   tangwang   embed norm
926
                  normalize_embeddings=effective_normalize,
54ccf28c   tangwang   tei
927
              )
7214c2e7   tangwang   mplemented**
928
929
930
931
932
933
              computed = []
              for i, emb in enumerate(embs):
                  vec = _as_list(emb, normalize=False)
                  if vec is None:
                      raise RuntimeError(f"Text model returned empty embedding for missing index {i}")
                  computed.append(vec)
4747e2f4   tangwang   embedding perform...
934
              mode = "backend-batch"
54ccf28c   tangwang   tei
935
      except Exception as e:
4747e2f4   tangwang   embedding perform...
936
937
938
939
940
941
942
943
          logger.error(
              "Text embedding backend failure: %s",
              e,
              exc_info=True,
              extra=_request_log_extra(request_id),
          )
          raise RuntimeError(f"Text embedding backend failure: {e}") from e
  
7214c2e7   tangwang   mplemented**
944
      if len(computed) != len(missing_texts):
ed948666   tangwang   tidy
945
          raise RuntimeError(
7214c2e7   tangwang   mplemented**
946
947
              f"Text model response length mismatch: expected {len(missing_texts)}, "
              f"got {len(computed)}"
ed948666   tangwang   tidy
948
          )
4747e2f4   tangwang   embedding perform...
949
  
7214c2e7   tangwang   mplemented**
950
951
952
953
954
      for pos, cache_key, vec in zip(missing_indices, missing_cache_keys, computed):
          out[pos] = vec
          _text_cache.set(cache_key, np.asarray(vec, dtype=np.float32))
  
      backend_elapsed_ms = (time.perf_counter() - backend_t0) * 1000.0
4747e2f4   tangwang   embedding perform...
955
  
efd435cf   tangwang   tei性能调优:
956
      logger.info(
7214c2e7   tangwang   mplemented**
957
          "text backend done | backend=%s mode=%s inputs=%d normalize=%s dim=%d cache_hits=%d cache_misses=%d backend_elapsed_ms=%.2f",
efd435cf   tangwang   tei性能调优:
958
          _text_backend_name,
4747e2f4   tangwang   embedding perform...
959
          mode,
efd435cf   tangwang   tei性能调优:
960
961
          len(normalized),
          effective_normalize,
28e57bb1   tangwang   日志体系优化
962
          len(out[0]) if out and out[0] is not None else 0,
7214c2e7   tangwang   mplemented**
963
964
965
          cache_hits,
          len(missing_texts),
          backend_elapsed_ms,
4747e2f4   tangwang   embedding perform...
966
          extra=_request_log_extra(request_id),
efd435cf   tangwang   tei性能调优:
967
      )
7214c2e7   tangwang   mplemented**
968
969
970
971
972
973
974
      return _EmbedResult(
          vectors=out,
          cache_hits=cache_hits,
          cache_misses=len(missing_texts),
          backend_elapsed_ms=backend_elapsed_ms,
          mode=mode,
      )
7bfb9946   tangwang   向量化模块
975
976
  
  
4747e2f4   tangwang   embedding perform...
977
978
979
980
981
982
  @app.post("/embed/text")
  async def embed_text(
      texts: List[str],
      http_request: Request,
      response: Response,
      normalize: Optional[bool] = None,
b754fd41   tangwang   图片向量化支持优先级参数
983
      priority: int = 0,
4747e2f4   tangwang   embedding perform...
984
  ) -> List[Optional[List[float]]]:
7214c2e7   tangwang   mplemented**
985
986
987
      if _text_model is None:
          raise HTTPException(status_code=503, detail="Text embedding model not loaded in this service")
  
4747e2f4   tangwang   embedding perform...
988
989
990
      request_id = _resolve_request_id(http_request)
      response.headers["X-Request-ID"] = request_id
  
b754fd41   tangwang   图片向量化支持优先级参数
991
992
993
      if priority < 0:
          raise HTTPException(status_code=400, detail="priority must be >= 0")
      effective_priority = _effective_priority(priority)
4747e2f4   tangwang   embedding perform...
994
995
996
997
998
999
      effective_normalize = bool(CONFIG.TEXT_NORMALIZE_EMBEDDINGS) if normalize is None else bool(normalize)
      normalized: List[str] = []
      for i, t in enumerate(texts):
          if not isinstance(t, str):
              raise HTTPException(status_code=400, detail=f"Invalid text at index {i}: must be string")
          s = t.strip()
ed948666   tangwang   tidy
1000
          if not s:
4747e2f4   tangwang   embedding perform...
1001
1002
              raise HTTPException(status_code=400, detail=f"Invalid text at index {i}: empty string")
          normalized.append(s)
c10f90fe   tangwang   cnclip
1003
  
7214c2e7   tangwang   mplemented**
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
      cache_check_started = time.perf_counter()
      cache_only = _try_full_text_cache_hit(normalized, effective_normalize)
      if cache_only is not None:
          latency_ms = (time.perf_counter() - cache_check_started) * 1000.0
          _text_stats.record_completed(
              success=True,
              latency_ms=latency_ms,
              backend_latency_ms=0.0,
              cache_hits=cache_only.cache_hits,
              cache_misses=0,
          )
          logger.info(
b754fd41   tangwang   图片向量化支持优先级参数
1016
              "embed_text response | backend=%s mode=cache-only priority=%s inputs=%d normalize=%s dim=%d cache_hits=%d cache_misses=0 first_vector=%s latency_ms=%.2f",
7214c2e7   tangwang   mplemented**
1017
              _text_backend_name,
b754fd41   tangwang   图片向量化支持优先级参数
1018
              _priority_label(effective_priority),
7214c2e7   tangwang   mplemented**
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
              len(normalized),
              effective_normalize,
              len(cache_only.vectors[0]) if cache_only.vectors and cache_only.vectors[0] is not None else 0,
              cache_only.cache_hits,
              _preview_vector(cache_only.vectors[0] if cache_only.vectors else None),
              latency_ms,
              extra=_request_log_extra(request_id),
          )
          return cache_only.vectors
  
b754fd41   tangwang   图片向量化支持优先级参数
1029
      accepted, active = _text_request_limiter.try_acquire(bypass_limit=effective_priority > 0)
4747e2f4   tangwang   embedding perform...
1030
      if not accepted:
7214c2e7   tangwang   mplemented**
1031
          _text_stats.record_rejected()
4747e2f4   tangwang   embedding perform...
1032
          logger.warning(
b754fd41   tangwang   图片向量化支持优先级参数
1033
              "embed_text rejected | client=%s backend=%s priority=%s inputs=%d normalize=%s active=%d limit=%d preview=%s",
4747e2f4   tangwang   embedding perform...
1034
1035
              _request_client(http_request),
              _text_backend_name,
b754fd41   tangwang   图片向量化支持优先级参数
1036
              _priority_label(effective_priority),
4747e2f4   tangwang   embedding perform...
1037
1038
1039
1040
1041
1042
1043
1044
1045
              len(normalized),
              effective_normalize,
              active,
              _TEXT_MAX_INFLIGHT,
              _preview_inputs(normalized, _LOG_PREVIEW_COUNT, _LOG_TEXT_PREVIEW_CHARS),
              extra=_request_log_extra(request_id),
          )
          raise HTTPException(
              status_code=_OVERLOAD_STATUS_CODE,
b754fd41   tangwang   图片向量化支持优先级参数
1046
1047
1048
1049
              detail=(
                  "Text embedding service busy for priority=0 requests: "
                  f"active={active}, limit={_TEXT_MAX_INFLIGHT}"
              ),
4747e2f4   tangwang   embedding perform...
1050
1051
1052
1053
          )
  
      request_started = time.perf_counter()
      success = False
7214c2e7   tangwang   mplemented**
1054
1055
1056
      backend_elapsed_ms = 0.0
      cache_hits = 0
      cache_misses = 0
4747e2f4   tangwang   embedding perform...
1057
1058
      try:
          logger.info(
b754fd41   tangwang   图片向量化支持优先级参数
1059
              "embed_text request | client=%s backend=%s priority=%s inputs=%d normalize=%s active=%d limit=%d preview=%s",
4747e2f4   tangwang   embedding perform...
1060
1061
              _request_client(http_request),
              _text_backend_name,
b754fd41   tangwang   图片向量化支持优先级参数
1062
              _priority_label(effective_priority),
4747e2f4   tangwang   embedding perform...
1063
1064
1065
1066
1067
1068
1069
1070
              len(normalized),
              effective_normalize,
              active,
              _TEXT_MAX_INFLIGHT,
              _preview_inputs(normalized, _LOG_PREVIEW_COUNT, _LOG_TEXT_PREVIEW_CHARS),
              extra=_request_log_extra(request_id),
          )
          verbose_logger.info(
b754fd41   tangwang   图片向量化支持优先级参数
1071
              "embed_text detail | payload=%s normalize=%s backend=%s priority=%s",
4747e2f4   tangwang   embedding perform...
1072
1073
1074
              normalized,
              effective_normalize,
              _text_backend_name,
b754fd41   tangwang   图片向量化支持优先级参数
1075
              _priority_label(effective_priority),
4747e2f4   tangwang   embedding perform...
1076
1077
              extra=_request_log_extra(request_id),
          )
b754fd41   tangwang   图片向量化支持优先级参数
1078
1079
1080
1081
1082
1083
1084
          result = await run_in_threadpool(
              _submit_text_dispatch_and_wait,
              normalized,
              effective_normalize,
              request_id,
              effective_priority,
          )
4747e2f4   tangwang   embedding perform...
1085
          success = True
7214c2e7   tangwang   mplemented**
1086
1087
1088
          backend_elapsed_ms = result.backend_elapsed_ms
          cache_hits = result.cache_hits
          cache_misses = result.cache_misses
4747e2f4   tangwang   embedding perform...
1089
          latency_ms = (time.perf_counter() - request_started) * 1000.0
7214c2e7   tangwang   mplemented**
1090
1091
1092
1093
1094
1095
1096
          _text_stats.record_completed(
              success=True,
              latency_ms=latency_ms,
              backend_latency_ms=backend_elapsed_ms,
              cache_hits=cache_hits,
              cache_misses=cache_misses,
          )
4747e2f4   tangwang   embedding perform...
1097
          logger.info(
b754fd41   tangwang   图片向量化支持优先级参数
1098
              "embed_text response | backend=%s mode=%s priority=%s inputs=%d normalize=%s dim=%d cache_hits=%d cache_misses=%d first_vector=%s latency_ms=%.2f",
4747e2f4   tangwang   embedding perform...
1099
              _text_backend_name,
7214c2e7   tangwang   mplemented**
1100
              result.mode,
b754fd41   tangwang   图片向量化支持优先级参数
1101
              _priority_label(effective_priority),
4747e2f4   tangwang   embedding perform...
1102
1103
              len(normalized),
              effective_normalize,
7214c2e7   tangwang   mplemented**
1104
1105
1106
1107
              len(result.vectors[0]) if result.vectors and result.vectors[0] is not None else 0,
              cache_hits,
              cache_misses,
              _preview_vector(result.vectors[0] if result.vectors else None),
4747e2f4   tangwang   embedding perform...
1108
1109
1110
1111
              latency_ms,
              extra=_request_log_extra(request_id),
          )
          verbose_logger.info(
b754fd41   tangwang   图片向量化支持优先级参数
1112
              "embed_text result detail | count=%d priority=%s first_vector=%s latency_ms=%.2f",
7214c2e7   tangwang   mplemented**
1113
              len(result.vectors),
b754fd41   tangwang   图片向量化支持优先级参数
1114
              _priority_label(effective_priority),
7214c2e7   tangwang   mplemented**
1115
1116
1117
              result.vectors[0][: _VECTOR_PREVIEW_DIMS]
              if result.vectors and result.vectors[0] is not None
              else [],
4747e2f4   tangwang   embedding perform...
1118
1119
1120
              latency_ms,
              extra=_request_log_extra(request_id),
          )
7214c2e7   tangwang   mplemented**
1121
          return result.vectors
4747e2f4   tangwang   embedding perform...
1122
1123
1124
1125
      except HTTPException:
          raise
      except Exception as e:
          latency_ms = (time.perf_counter() - request_started) * 1000.0
7214c2e7   tangwang   mplemented**
1126
1127
1128
1129
1130
1131
1132
          _text_stats.record_completed(
              success=False,
              latency_ms=latency_ms,
              backend_latency_ms=backend_elapsed_ms,
              cache_hits=cache_hits,
              cache_misses=cache_misses,
          )
4747e2f4   tangwang   embedding perform...
1133
          logger.error(
b754fd41   tangwang   图片向量化支持优先级参数
1134
              "embed_text failed | backend=%s priority=%s inputs=%d normalize=%s latency_ms=%.2f error=%s",
4747e2f4   tangwang   embedding perform...
1135
              _text_backend_name,
b754fd41   tangwang   图片向量化支持优先级参数
1136
              _priority_label(effective_priority),
4747e2f4   tangwang   embedding perform...
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
              len(normalized),
              effective_normalize,
              latency_ms,
              e,
              exc_info=True,
              extra=_request_log_extra(request_id),
          )
          raise HTTPException(status_code=502, detail=str(e)) from e
      finally:
          remaining = _text_request_limiter.release(success=success)
          logger.info(
b754fd41   tangwang   图片向量化支持优先级参数
1148
              "embed_text finalize | success=%s priority=%s active_after=%d",
4747e2f4   tangwang   embedding perform...
1149
              success,
b754fd41   tangwang   图片向量化支持优先级参数
1150
              _priority_label(effective_priority),
4747e2f4   tangwang   embedding perform...
1151
1152
1153
1154
1155
1156
1157
1158
1159
              remaining,
              extra=_request_log_extra(request_id),
          )
  
  
  def _embed_image_impl(
      urls: List[str],
      effective_normalize: bool,
      request_id: str,
7214c2e7   tangwang   mplemented**
1160
  ) -> _EmbedResult:
4747e2f4   tangwang   embedding perform...
1161
1162
      if _image_model is None:
          raise RuntimeError("Image model not loaded")
28e57bb1   tangwang   日志体系优化
1163
  
7214c2e7   tangwang   mplemented**
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
      out: List[Optional[List[float]]] = [None] * len(urls)
      missing_indices: List[int] = []
      missing_urls: List[str] = []
      missing_cache_keys: List[str] = []
      cache_hits = 0
      for idx, url in enumerate(urls):
          cache_key = build_image_cache_key(url, normalize=effective_normalize)
          cached = _image_cache.get(cache_key)
          if cached is not None:
              vec = _as_list(cached, normalize=False)
              if vec is not None:
                  out[idx] = vec
                  cache_hits += 1
                  continue
          missing_indices.append(idx)
          missing_urls.append(url)
          missing_cache_keys.append(cache_key)
  
      if not missing_urls:
          logger.info(
              "image backend done | mode=cache-only inputs=%d normalize=%s dim=%d cache_hits=%d cache_misses=0 backend_elapsed_ms=0.00",
              len(urls),
              effective_normalize,
              len(out[0]) if out and out[0] is not None else 0,
              cache_hits,
              extra=_request_log_extra(request_id),
          )
          return _EmbedResult(
              vectors=out,
              cache_hits=cache_hits,
              cache_misses=0,
              backend_elapsed_ms=0.0,
              mode="cache-only",
          )
  
      backend_t0 = time.perf_counter()
7bfb9946   tangwang   向量化模块
1200
      with _image_encode_lock:
200fdddf   tangwang   embed norm
1201
          vectors = _image_model.encode_image_urls(
7214c2e7   tangwang   mplemented**
1202
              missing_urls,
200fdddf   tangwang   embed norm
1203
1204
1205
              batch_size=CONFIG.IMAGE_BATCH_SIZE,
              normalize_embeddings=effective_normalize,
          )
7214c2e7   tangwang   mplemented**
1206
      if vectors is None or len(vectors) != len(missing_urls):
ed948666   tangwang   tidy
1207
          raise RuntimeError(
7214c2e7   tangwang   mplemented**
1208
              f"Image model response length mismatch: expected {len(missing_urls)}, "
ed948666   tangwang   tidy
1209
1210
              f"got {0 if vectors is None else len(vectors)}"
          )
4747e2f4   tangwang   embedding perform...
1211
  
7214c2e7   tangwang   mplemented**
1212
      for pos, cache_key, vec in zip(missing_indices, missing_cache_keys, vectors):
200fdddf   tangwang   embed norm
1213
          out_vec = _as_list(vec, normalize=effective_normalize)
ed948666   tangwang   tidy
1214
          if out_vec is None:
7214c2e7   tangwang   mplemented**
1215
1216
1217
1218
1219
              raise RuntimeError(f"Image model returned empty embedding for position {pos}")
          out[pos] = out_vec
          _image_cache.set(cache_key, np.asarray(out_vec, dtype=np.float32))
  
      backend_elapsed_ms = (time.perf_counter() - backend_t0) * 1000.0
4747e2f4   tangwang   embedding perform...
1220
  
28e57bb1   tangwang   日志体系优化
1221
      logger.info(
7214c2e7   tangwang   mplemented**
1222
          "image backend done | mode=backend-batch inputs=%d normalize=%s dim=%d cache_hits=%d cache_misses=%d backend_elapsed_ms=%.2f",
28e57bb1   tangwang   日志体系优化
1223
1224
1225
          len(urls),
          effective_normalize,
          len(out[0]) if out and out[0] is not None else 0,
7214c2e7   tangwang   mplemented**
1226
1227
1228
          cache_hits,
          len(missing_urls),
          backend_elapsed_ms,
4747e2f4   tangwang   embedding perform...
1229
          extra=_request_log_extra(request_id),
28e57bb1   tangwang   日志体系优化
1230
      )
7214c2e7   tangwang   mplemented**
1231
1232
1233
1234
1235
1236
1237
      return _EmbedResult(
          vectors=out,
          cache_hits=cache_hits,
          cache_misses=len(missing_urls),
          backend_elapsed_ms=backend_elapsed_ms,
          mode="backend-batch",
      )
4747e2f4   tangwang   embedding perform...
1238
1239
1240
1241
1242
1243
1244
1245
  
  
  @app.post("/embed/image")
  async def embed_image(
      images: List[str],
      http_request: Request,
      response: Response,
      normalize: Optional[bool] = None,
b754fd41   tangwang   图片向量化支持优先级参数
1246
      priority: int = 0,
4747e2f4   tangwang   embedding perform...
1247
  ) -> List[Optional[List[float]]]:
7214c2e7   tangwang   mplemented**
1248
1249
1250
      if _image_model is None:
          raise HTTPException(status_code=503, detail="Image embedding model not loaded in this service")
  
4747e2f4   tangwang   embedding perform...
1251
1252
1253
      request_id = _resolve_request_id(http_request)
      response.headers["X-Request-ID"] = request_id
  
b754fd41   tangwang   图片向量化支持优先级参数
1254
1255
1256
1257
      if priority < 0:
          raise HTTPException(status_code=400, detail="priority must be >= 0")
      effective_priority = _effective_priority(priority)
  
4747e2f4   tangwang   embedding perform...
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
      effective_normalize = bool(CONFIG.IMAGE_NORMALIZE_EMBEDDINGS) if normalize is None else bool(normalize)
      urls: List[str] = []
      for i, url_or_path in enumerate(images):
          if not isinstance(url_or_path, str):
              raise HTTPException(status_code=400, detail=f"Invalid image at index {i}: must be string URL/path")
          s = url_or_path.strip()
          if not s:
              raise HTTPException(status_code=400, detail=f"Invalid image at index {i}: empty URL/path")
          urls.append(s)
  
7214c2e7   tangwang   mplemented**
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
      cache_check_started = time.perf_counter()
      cache_only = _try_full_image_cache_hit(urls, effective_normalize)
      if cache_only is not None:
          latency_ms = (time.perf_counter() - cache_check_started) * 1000.0
          _image_stats.record_completed(
              success=True,
              latency_ms=latency_ms,
              backend_latency_ms=0.0,
              cache_hits=cache_only.cache_hits,
              cache_misses=0,
          )
          logger.info(
b754fd41   tangwang   图片向量化支持优先级参数
1280
1281
              "embed_image response | mode=cache-only priority=%s inputs=%d normalize=%s dim=%d cache_hits=%d cache_misses=0 first_vector=%s latency_ms=%.2f",
              _priority_label(effective_priority),
7214c2e7   tangwang   mplemented**
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
              len(urls),
              effective_normalize,
              len(cache_only.vectors[0]) if cache_only.vectors and cache_only.vectors[0] is not None else 0,
              cache_only.cache_hits,
              _preview_vector(cache_only.vectors[0] if cache_only.vectors else None),
              latency_ms,
              extra=_request_log_extra(request_id),
          )
          return cache_only.vectors
  
b754fd41   tangwang   图片向量化支持优先级参数
1292
      accepted, active = _image_request_limiter.try_acquire(bypass_limit=effective_priority > 0)
4747e2f4   tangwang   embedding perform...
1293
      if not accepted:
7214c2e7   tangwang   mplemented**
1294
          _image_stats.record_rejected()
4747e2f4   tangwang   embedding perform...
1295
          logger.warning(
b754fd41   tangwang   图片向量化支持优先级参数
1296
              "embed_image rejected | client=%s priority=%s inputs=%d normalize=%s active=%d limit=%d preview=%s",
4747e2f4   tangwang   embedding perform...
1297
              _request_client(http_request),
b754fd41   tangwang   图片向量化支持优先级参数
1298
              _priority_label(effective_priority),
4747e2f4   tangwang   embedding perform...
1299
1300
1301
1302
1303
1304
1305
1306
1307
              len(urls),
              effective_normalize,
              active,
              _IMAGE_MAX_INFLIGHT,
              _preview_inputs(urls, _LOG_PREVIEW_COUNT, _LOG_IMAGE_PREVIEW_CHARS),
              extra=_request_log_extra(request_id),
          )
          raise HTTPException(
              status_code=_OVERLOAD_STATUS_CODE,
b754fd41   tangwang   图片向量化支持优先级参数
1308
1309
1310
1311
              detail=(
                  "Image embedding service busy for priority=0 requests: "
                  f"active={active}, limit={_IMAGE_MAX_INFLIGHT}"
              ),
4747e2f4   tangwang   embedding perform...
1312
1313
1314
1315
          )
  
      request_started = time.perf_counter()
      success = False
7214c2e7   tangwang   mplemented**
1316
1317
1318
      backend_elapsed_ms = 0.0
      cache_hits = 0
      cache_misses = 0
4747e2f4   tangwang   embedding perform...
1319
1320
      try:
          logger.info(
b754fd41   tangwang   图片向量化支持优先级参数
1321
              "embed_image request | client=%s priority=%s inputs=%d normalize=%s active=%d limit=%d preview=%s",
4747e2f4   tangwang   embedding perform...
1322
              _request_client(http_request),
b754fd41   tangwang   图片向量化支持优先级参数
1323
              _priority_label(effective_priority),
4747e2f4   tangwang   embedding perform...
1324
1325
1326
1327
1328
1329
1330
1331
              len(urls),
              effective_normalize,
              active,
              _IMAGE_MAX_INFLIGHT,
              _preview_inputs(urls, _LOG_PREVIEW_COUNT, _LOG_IMAGE_PREVIEW_CHARS),
              extra=_request_log_extra(request_id),
          )
          verbose_logger.info(
b754fd41   tangwang   图片向量化支持优先级参数
1332
              "embed_image detail | payload=%s normalize=%s priority=%s",
4747e2f4   tangwang   embedding perform...
1333
1334
              urls,
              effective_normalize,
b754fd41   tangwang   图片向量化支持优先级参数
1335
              _priority_label(effective_priority),
4747e2f4   tangwang   embedding perform...
1336
1337
              extra=_request_log_extra(request_id),
          )
7214c2e7   tangwang   mplemented**
1338
          result = await run_in_threadpool(_embed_image_impl, urls, effective_normalize, request_id)
4747e2f4   tangwang   embedding perform...
1339
          success = True
7214c2e7   tangwang   mplemented**
1340
1341
1342
          backend_elapsed_ms = result.backend_elapsed_ms
          cache_hits = result.cache_hits
          cache_misses = result.cache_misses
4747e2f4   tangwang   embedding perform...
1343
          latency_ms = (time.perf_counter() - request_started) * 1000.0
7214c2e7   tangwang   mplemented**
1344
1345
1346
1347
1348
1349
1350
          _image_stats.record_completed(
              success=True,
              latency_ms=latency_ms,
              backend_latency_ms=backend_elapsed_ms,
              cache_hits=cache_hits,
              cache_misses=cache_misses,
          )
4747e2f4   tangwang   embedding perform...
1351
          logger.info(
b754fd41   tangwang   图片向量化支持优先级参数
1352
              "embed_image response | mode=%s priority=%s inputs=%d normalize=%s dim=%d cache_hits=%d cache_misses=%d first_vector=%s latency_ms=%.2f",
7214c2e7   tangwang   mplemented**
1353
              result.mode,
b754fd41   tangwang   图片向量化支持优先级参数
1354
              _priority_label(effective_priority),
4747e2f4   tangwang   embedding perform...
1355
1356
              len(urls),
              effective_normalize,
7214c2e7   tangwang   mplemented**
1357
1358
1359
1360
              len(result.vectors[0]) if result.vectors and result.vectors[0] is not None else 0,
              cache_hits,
              cache_misses,
              _preview_vector(result.vectors[0] if result.vectors else None),
4747e2f4   tangwang   embedding perform...
1361
1362
1363
1364
1365
              latency_ms,
              extra=_request_log_extra(request_id),
          )
          verbose_logger.info(
              "embed_image result detail | count=%d first_vector=%s latency_ms=%.2f",
7214c2e7   tangwang   mplemented**
1366
1367
1368
1369
              len(result.vectors),
              result.vectors[0][: _VECTOR_PREVIEW_DIMS]
              if result.vectors and result.vectors[0] is not None
              else [],
4747e2f4   tangwang   embedding perform...
1370
1371
1372
              latency_ms,
              extra=_request_log_extra(request_id),
          )
7214c2e7   tangwang   mplemented**
1373
          return result.vectors
4747e2f4   tangwang   embedding perform...
1374
1375
1376
1377
      except HTTPException:
          raise
      except Exception as e:
          latency_ms = (time.perf_counter() - request_started) * 1000.0
7214c2e7   tangwang   mplemented**
1378
1379
1380
1381
1382
1383
1384
          _image_stats.record_completed(
              success=False,
              latency_ms=latency_ms,
              backend_latency_ms=backend_elapsed_ms,
              cache_hits=cache_hits,
              cache_misses=cache_misses,
          )
4747e2f4   tangwang   embedding perform...
1385
          logger.error(
b754fd41   tangwang   图片向量化支持优先级参数
1386
1387
              "embed_image failed | priority=%s inputs=%d normalize=%s latency_ms=%.2f error=%s",
              _priority_label(effective_priority),
4747e2f4   tangwang   embedding perform...
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
              len(urls),
              effective_normalize,
              latency_ms,
              e,
              exc_info=True,
              extra=_request_log_extra(request_id),
          )
          raise HTTPException(status_code=502, detail=f"Image embedding backend failure: {e}") from e
      finally:
          remaining = _image_request_limiter.release(success=success)
          logger.info(
b754fd41   tangwang   图片向量化支持优先级参数
1399
              "embed_image finalize | success=%s priority=%s active_after=%d",
4747e2f4   tangwang   embedding perform...
1400
              success,
b754fd41   tangwang   图片向量化支持优先级参数
1401
              _priority_label(effective_priority),
4747e2f4   tangwang   embedding perform...
1402
1403
1404
              remaining,
              extra=_request_log_extra(request_id),
          )