Blame view

requirements_reranker_qwen3_vllm_score.txt 766 Bytes
4823f463   tangwang   qwen3_vllm_score ...
1
2
3
4
5
6
7
8
9
10
11
  # Dedicated high-performance venv for qwen3_vllm_score: .venv-reranker-score
  #
  # Create / refresh:
  #   ./scripts/setup_reranker_venv.sh qwen3_vllm_score
  #
  # vLLM 0.17+ replaces LLM(task="score") with runner/convert auto + LLM.score().
  # Pin vLLM for reproducible perf baselines; bump after validating CUDA/driver on your hosts.
  # If pip cannot find a wheel for your CUDA version, edit the vllm line or install from:
  #   https://docs.vllm.ai/en/latest/getting_started/installation.html
  
  -r requirements_reranker_base.txt
540fb5af   tangwang   添加了可关闭的开关:保留默认行为(...
12
13
  # FlashInfer JIT (vLLM may select it on Turing when TRITON_ATTN is not forced) needs a ninja binary on PATH.
  ninja>=1.11
4823f463   tangwang   qwen3_vllm_score ...
14
15
16
  vllm==0.18.0
  # Match vLLM 0.18 stack; cap <5 to avoid pip prefetching incompatible transformers 5.x.
  transformers>=4.51.0,<5