feat: enhance search, ranking, reranker and CLI tooling across ccw and codex-lens

Major improvements to smart-search, chain-search cascade, ranking pipeline,
reranker factory, CLI history store, codex-lens integration, and uv-manager.
Simplify command-generator skill by inlining phases. Add comprehensive tests.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
catlog22
2026-03-16 20:35:08 +08:00
parent 1cd96b90e8
commit 5a4b18d9b1
73 changed files with 14684 additions and 2442 deletions

View File

@@ -24,13 +24,24 @@ from codexlens.entities import SearchResult
from codexlens.search.ranking import (
DEFAULT_WEIGHTS,
QueryIntent,
apply_path_penalties,
extract_explicit_path_hints,
cross_encoder_rerank,
adjust_weights_by_intent,
apply_symbol_boost,
detect_query_intent,
filter_results_by_category,
get_rrf_weights,
group_similar_results,
is_auxiliary_reference_path,
is_generated_artifact_path,
is_test_file,
normalize_weights,
query_prefers_lexical_search,
query_targets_auxiliary_files,
query_targets_generated_files,
query_targets_test_files,
rebalance_noisy_results,
reciprocal_rank_fusion,
simple_weighted_fusion,
)
@@ -73,6 +84,7 @@ class TestDetectQueryIntent:
def test_detect_keyword_intent(self):
"""CamelCase/underscore queries should be detected as KEYWORD."""
assert detect_query_intent("MyClassName") == QueryIntent.KEYWORD
assert detect_query_intent("windowsHide") == QueryIntent.KEYWORD
assert detect_query_intent("my_function_name") == QueryIntent.KEYWORD
assert detect_query_intent("foo::bar") == QueryIntent.KEYWORD
@@ -91,6 +103,25 @@ class TestDetectQueryIntent:
assert detect_query_intent("") == QueryIntent.MIXED
assert detect_query_intent(" ") == QueryIntent.MIXED
def test_query_targets_test_files(self):
"""Queries explicitly mentioning tests should skip test penalties."""
assert query_targets_test_files("how do tests cover auth flow?")
assert query_targets_test_files("spec fixtures for parser")
assert not query_targets_test_files("windowsHide")
def test_query_targets_generated_files(self):
"""Queries explicitly mentioning build artifacts should skip that penalty."""
assert query_targets_generated_files("inspect dist bundle output")
assert query_targets_generated_files("generated artifacts under build")
assert not query_targets_generated_files("cache invalidation strategy")
def test_query_prefers_lexical_search(self):
"""Config/env/factory queries should prefer lexical-first routing."""
assert query_prefers_lexical_search("embedding backend fastembed local litellm api config")
assert query_prefers_lexical_search("get_reranker factory onnx backend selection")
assert query_prefers_lexical_search("EMBEDDING_BACKEND and RERANKER_BACKEND environment variables")
assert not query_prefers_lexical_search("how does smart search route keyword queries")
# =============================================================================
# Tests: adjust_weights_by_intent
@@ -129,6 +160,427 @@ class TestAdjustWeightsByIntent:
assert adjusted["exact"] == pytest.approx(0.3, abs=0.01)
class TestPathPenalties:
"""Tests for lightweight path-based ranking penalties."""
def test_is_test_file(self):
assert is_test_file("/repo/tests/test_auth.py")
assert is_test_file("D:\\repo\\src\\auth.spec.ts")
assert is_test_file("/repo/frontend/src/pages/discoverypage.test.tsx")
assert is_test_file("/repo/frontend/src/pages/discoverypage.spec.jsx")
assert not is_test_file("/repo/src/auth.py")
def test_is_generated_artifact_path(self):
assert is_generated_artifact_path("/repo/dist/app.js")
assert is_generated_artifact_path("/repo/src/generated/client.ts")
assert is_generated_artifact_path("D:\\repo\\frontend\\.next\\server.js")
assert not is_generated_artifact_path("/repo/src/auth.py")
def test_is_auxiliary_reference_path(self):
assert is_auxiliary_reference_path("/repo/examples/auth_demo.py")
assert is_auxiliary_reference_path("/repo/benchmarks/search_eval.py")
assert is_auxiliary_reference_path("/repo/tools/debug_search.py")
assert not is_auxiliary_reference_path("/repo/src/auth.py")
def test_query_targets_auxiliary_files(self):
assert query_targets_auxiliary_files("show smart search examples")
assert query_targets_auxiliary_files("benchmark smart search")
assert not query_targets_auxiliary_files("smart search routing")
def test_apply_path_penalties_demotes_test_files(self):
results = [
_make_result(path="/repo/tests/test_auth.py", score=10.0),
_make_result(path="/repo/src/auth.py", score=9.0),
]
penalized = apply_path_penalties(
results,
"authenticate user",
test_file_penalty=0.15,
)
assert penalized[0].path == "/repo/src/auth.py"
assert penalized[1].metadata["path_penalty_reasons"] == ["test_file"]
def test_apply_path_penalties_more_aggressively_demotes_tests_for_keyword_queries(self):
results = [
_make_result(path="/repo/tests/test_auth.py", score=5.0),
_make_result(path="/repo/src/auth.py", score=4.0),
]
penalized = apply_path_penalties(
results,
"find_descendant_project_roots",
test_file_penalty=0.15,
)
assert penalized[0].path == "/repo/src/auth.py"
assert penalized[1].metadata["path_penalty_reasons"] == ["test_file"]
assert penalized[1].metadata["path_penalty_multiplier"] == pytest.approx(0.55)
assert penalized[1].metadata["path_rank_multiplier"] == pytest.approx(0.55)
def test_apply_path_penalties_more_aggressively_demotes_tests_for_semantic_queries(self):
results = [
_make_result(path="/repo/tests/test_auth.py", score=5.0),
_make_result(path="/repo/src/auth.py", score=4.1),
]
penalized = apply_path_penalties(
results,
"how does auth routing work",
test_file_penalty=0.15,
)
assert penalized[0].path == "/repo/src/auth.py"
assert penalized[1].metadata["path_penalty_reasons"] == ["test_file"]
assert penalized[1].metadata["path_penalty_multiplier"] == pytest.approx(0.75)
def test_apply_path_penalties_boosts_source_definitions_for_identifier_queries(self):
results = [
_make_result(
path="/repo/tests/test_registry.py",
score=4.2,
excerpt='query="find_descendant_project_roots"',
),
_make_result(
path="/repo/src/registry.py",
score=3.0,
excerpt="def find_descendant_project_roots(self, source_root: Path) -> list[str]:",
),
]
penalized = apply_path_penalties(
results,
"find_descendant_project_roots",
test_file_penalty=0.15,
)
assert penalized[0].path == "/repo/src/registry.py"
assert penalized[0].metadata["path_boost_reasons"] == ["source_definition"]
assert penalized[0].metadata["path_boost_multiplier"] == pytest.approx(2.0)
assert penalized[0].metadata["path_rank_multiplier"] == pytest.approx(2.0)
assert penalized[1].metadata["path_penalty_reasons"] == ["test_file"]
def test_apply_path_penalties_boosts_source_paths_for_semantic_feature_queries(self):
results = [
_make_result(
path="/repo/tests/smart-search-intent.test.js",
score=0.832,
excerpt="describes how smart search routes keyword queries",
),
_make_result(
path="/repo/src/tools/smart-search.ts",
score=0.555,
excerpt="smart search keyword routing logic",
),
]
penalized = apply_path_penalties(
results,
"how does smart search route keyword queries",
test_file_penalty=0.15,
)
assert penalized[0].path == "/repo/src/tools/smart-search.ts"
assert penalized[0].metadata["path_boost_reasons"] == ["source_path_topic_overlap"]
assert penalized[0].metadata["path_boost_multiplier"] == pytest.approx(1.35)
assert penalized[0].metadata["path_boost_overlap_tokens"] == ["smart", "search"]
assert penalized[1].metadata["path_penalty_reasons"] == ["test_file"]
def test_apply_path_penalties_strongly_boosts_keyword_basename_overlap(self):
results = [
_make_result(
path="/repo/src/tools/core-memory.ts",
score=0.04032417772512223,
excerpt="memory listing helpers",
),
_make_result(
path="/repo/src/tools/smart-search.ts",
score=0.009836065573770493,
excerpt="smart search keyword routing logic",
),
]
penalized = apply_path_penalties(
results,
"executeHybridMode dense_rerank semantic smart_search",
test_file_penalty=0.15,
)
assert penalized[0].path == "/repo/src/tools/smart-search.ts"
assert penalized[0].metadata["path_boost_reasons"] == ["source_path_topic_overlap"]
assert penalized[0].metadata["path_boost_multiplier"] == pytest.approx(4.5)
assert penalized[0].metadata["path_boost_overlap_tokens"] == ["smart", "search"]
def test_extract_explicit_path_hints_ignores_generic_platform_terms(self):
assert extract_explicit_path_hints(
"parse CodexLens JSON output strip ANSI smart_search",
) == [["smart", "search"]]
def test_apply_path_penalties_prefers_explicit_feature_hint_over_platform_terms(self):
results = [
_make_result(
path="/repo/src/tools/codex-lens-lsp.ts",
score=0.045,
excerpt="CodexLens LSP bridge",
),
_make_result(
path="/repo/src/tools/smart-search.ts",
score=0.03,
excerpt="parse JSON output and strip ANSI for plain-text fallback",
),
]
penalized = apply_path_penalties(
results,
"parse CodexLens JSON output strip ANSI smart_search",
test_file_penalty=0.15,
)
assert penalized[0].path == "/repo/src/tools/smart-search.ts"
assert penalized[0].metadata["path_boost_reasons"] == ["source_path_topic_overlap"]
assert penalized[0].metadata["path_boost_overlap_tokens"] == ["smart", "search"]
def test_apply_path_penalties_strongly_boosts_lexical_config_modules(self):
results = [
_make_result(
path="/repo/src/tools/smart-search.ts",
score=22.07,
excerpt="embedding backend local api config routing",
),
_make_result(
path="/repo/src/codexlens/config.py",
score=4.88,
excerpt="embedding_backend = 'fastembed'",
),
]
penalized = apply_path_penalties(
results,
"embedding backend fastembed local litellm api config",
test_file_penalty=0.15,
)
assert penalized[0].path == "/repo/src/codexlens/config.py"
assert penalized[0].metadata["path_boost_reasons"] == ["source_path_topic_overlap"]
assert penalized[0].metadata["path_boost_multiplier"] == pytest.approx(5.0)
assert penalized[0].metadata["path_boost_overlap_tokens"] == ["config"]
def test_apply_path_penalties_more_aggressively_demotes_tests_for_explicit_feature_queries(self):
results = [
_make_result(
path="/repo/tests/smart-search-intent.test.js",
score=1.0,
excerpt="smart search intent coverage",
),
_make_result(
path="/repo/src/tools/smart-search.ts",
score=0.58,
excerpt="plain-text JSON fallback for smart search",
),
]
penalized = apply_path_penalties(
results,
"parse CodexLens JSON output strip ANSI smart_search",
test_file_penalty=0.15,
)
assert penalized[0].path == "/repo/src/tools/smart-search.ts"
assert penalized[1].metadata["path_penalty_reasons"] == ["test_file"]
assert penalized[1].metadata["path_penalty_multiplier"] == pytest.approx(0.55)
def test_apply_path_penalties_demotes_generated_artifacts(self):
results = [
_make_result(path="/repo/dist/auth.js", score=10.0),
_make_result(path="/repo/src/auth.ts", score=9.0),
]
penalized = apply_path_penalties(
results,
"authenticate user",
generated_file_penalty=0.35,
)
assert penalized[0].path == "/repo/src/auth.ts"
assert penalized[1].metadata["path_penalty_reasons"] == ["generated_artifact"]
def test_apply_path_penalties_more_aggressively_demotes_generated_artifacts_for_explicit_feature_queries(self):
results = [
_make_result(
path="/repo/dist/tools/smart-search.js",
score=1.0,
excerpt="built smart search output",
),
_make_result(
path="/repo/src/tools/smart-search.ts",
score=0.45,
excerpt="plain-text JSON fallback for smart search",
),
]
penalized = apply_path_penalties(
results,
"parse CodexLens JSON output strip ANSI smart_search",
generated_file_penalty=0.35,
)
assert penalized[0].path == "/repo/src/tools/smart-search.ts"
assert penalized[1].metadata["path_penalty_reasons"] == ["generated_artifact"]
assert penalized[1].metadata["path_penalty_multiplier"] == pytest.approx(0.4)
def test_apply_path_penalties_demotes_auxiliary_reference_files(self):
results = [
_make_result(path="/repo/examples/simple_search_comparison.py", score=10.0),
_make_result(path="/repo/src/search/router.py", score=9.0),
]
penalized = apply_path_penalties(
results,
"how does smart search route keyword queries",
test_file_penalty=0.15,
)
assert penalized[0].path == "/repo/src/search/router.py"
assert penalized[1].metadata["path_penalty_reasons"] == ["auxiliary_file"]
def test_apply_path_penalties_more_aggressively_demotes_auxiliary_files_for_explicit_feature_queries(self):
results = [
_make_result(
path="/repo/benchmarks/smart_search_demo.py",
score=1.0,
excerpt="demo for smart search fallback",
),
_make_result(
path="/repo/src/tools/smart-search.ts",
score=0.52,
excerpt="plain-text JSON fallback for smart search",
),
]
penalized = apply_path_penalties(
results,
"parse CodexLens JSON output strip ANSI smart_search",
test_file_penalty=0.15,
)
assert penalized[0].path == "/repo/src/tools/smart-search.ts"
assert penalized[1].metadata["path_penalty_reasons"] == ["auxiliary_file"]
assert penalized[1].metadata["path_penalty_multiplier"] == pytest.approx(0.5)
def test_apply_path_penalties_skips_when_query_targets_tests(self):
results = [
_make_result(path="/repo/tests/test_auth.py", score=10.0),
_make_result(path="/repo/src/auth.py", score=9.0),
]
penalized = apply_path_penalties(
results,
"auth tests",
test_file_penalty=0.15,
)
assert penalized[0].path == "/repo/tests/test_auth.py"
def test_apply_path_penalties_skips_generated_penalty_when_query_targets_artifacts(self):
results = [
_make_result(path="/repo/dist/auth.js", score=10.0),
_make_result(path="/repo/src/auth.ts", score=9.0),
]
penalized = apply_path_penalties(
results,
"dist auth bundle",
generated_file_penalty=0.35,
)
assert penalized[0].path == "/repo/dist/auth.js"
def test_rebalance_noisy_results_pushes_explicit_feature_query_noise_behind_source_files(self):
results = [
_make_result(path="/repo/src/tools/smart-search.ts", score=0.9),
_make_result(path="/repo/tests/smart-search-intent.test.tsx", score=0.8),
_make_result(path="/repo/src/core/cli-routes.ts", score=0.7),
_make_result(path="/repo/dist/tools/smart-search.js", score=0.6),
_make_result(path="/repo/benchmarks/smart_search_demo.py", score=0.5),
]
rebalanced = rebalance_noisy_results(
results,
"parse CodexLens JSON output strip ANSI smart_search",
)
assert [item.path for item in rebalanced[:2]] == [
"/repo/src/tools/smart-search.ts",
"/repo/src/core/cli-routes.ts",
]
def test_rebalance_noisy_results_preserves_tests_when_query_targets_them(self):
results = [
_make_result(path="/repo/tests/smart-search-intent.test.tsx", score=0.9),
_make_result(path="/repo/src/tools/smart-search.ts", score=0.8),
]
rebalanced = rebalance_noisy_results(results, "smart search tests")
assert [item.path for item in rebalanced] == [
"/repo/tests/smart-search-intent.test.tsx",
"/repo/src/tools/smart-search.ts",
]
def test_apply_path_penalties_skips_auxiliary_penalty_when_query_targets_examples(self):
results = [
_make_result(path="/repo/examples/simple_search_comparison.py", score=10.0),
_make_result(path="/repo/src/search/router.py", score=9.0),
]
penalized = apply_path_penalties(
results,
"smart search examples",
test_file_penalty=0.15,
)
assert penalized[0].path == "/repo/examples/simple_search_comparison.py"
class TestCrossEncoderRerank:
"""Tests for cross-encoder reranking edge cases."""
def test_cross_encoder_rerank_preserves_strong_source_candidates_for_semantic_feature_queries(self):
class DummyReranker:
def score_pairs(self, pairs, batch_size=32):
_ = (pairs, batch_size)
return [0.8323705792427063, 1.2463066923373844e-05]
reranked = cross_encoder_rerank(
"how does smart search route keyword queries",
[
_make_result(
path="/repo/tests/smart-search-intent.test.js",
score=0.5989155769348145,
excerpt="describes how smart search routes keyword queries",
),
_make_result(
path="/repo/src/tools/smart-search.ts",
score=0.554444432258606,
excerpt="smart search keyword routing logic",
),
],
DummyReranker(),
top_k=2,
)
reranked = apply_path_penalties(
reranked,
"how does smart search route keyword queries",
test_file_penalty=0.15,
)
assert reranked[0].path == "/repo/src/tools/smart-search.ts"
assert reranked[0].metadata["cross_encoder_floor_reason"] == "semantic_source_path_overlap"
assert reranked[0].metadata["cross_encoder_floor_overlap_tokens"] == ["smart", "search"]
assert reranked[0].metadata["path_boost_reasons"] == ["source_path_topic_overlap"]
assert reranked[1].metadata["path_penalty_reasons"] == ["test_file"]
# =============================================================================
# Tests: get_rrf_weights
# =============================================================================