mirror of
https://github.com/catlog22/Claude-Code-Workflow.git
synced 2026-02-05 01:50:27 +08:00
feat: 添加动态批处理大小设置及相关国际化支持,优化配置管理
This commit is contained in:
@@ -856,6 +856,19 @@ export async function handleCodexLensConfigRoutes(ctx: RouteContext): Promise<bo
|
||||
if (settings.api?.batch_size !== undefined) {
|
||||
settingsDefaults['CODEXLENS_API_BATCH_SIZE'] = String(settings.api.batch_size);
|
||||
}
|
||||
// Dynamic batch size settings
|
||||
if (settings.api?.batch_size_dynamic !== undefined) {
|
||||
settingsDefaults['CODEXLENS_API_BATCH_SIZE_DYNAMIC'] = String(settings.api.batch_size_dynamic);
|
||||
}
|
||||
if (settings.api?.batch_size_utilization_factor !== undefined) {
|
||||
settingsDefaults['CODEXLENS_API_BATCH_SIZE_UTILIZATION'] = String(settings.api.batch_size_utilization_factor);
|
||||
}
|
||||
if (settings.api?.batch_size_max !== undefined) {
|
||||
settingsDefaults['CODEXLENS_API_BATCH_SIZE_MAX'] = String(settings.api.batch_size_max);
|
||||
}
|
||||
if (settings.api?.chars_per_token_estimate !== undefined) {
|
||||
settingsDefaults['CODEXLENS_CHARS_PER_TOKEN'] = String(settings.api.chars_per_token_estimate);
|
||||
}
|
||||
|
||||
// Cascade search settings
|
||||
if (settings.cascade?.strategy) {
|
||||
@@ -1057,6 +1070,10 @@ export async function handleCodexLensConfigRoutes(ctx: RouteContext): Promise<bo
|
||||
'CODEXLENS_RERANKER_TOP_K': { path: ['reranker', 'top_k'], transform: v => parseInt(v, 10) },
|
||||
'CODEXLENS_API_MAX_WORKERS': { path: ['api', 'max_workers'], transform: v => parseInt(v, 10) },
|
||||
'CODEXLENS_API_BATCH_SIZE': { path: ['api', 'batch_size'], transform: v => parseInt(v, 10) },
|
||||
'CODEXLENS_API_BATCH_SIZE_DYNAMIC': { path: ['api', 'batch_size_dynamic'], transform: v => v === 'true' },
|
||||
'CODEXLENS_API_BATCH_SIZE_UTILIZATION': { path: ['api', 'batch_size_utilization_factor'], transform: v => parseFloat(v) },
|
||||
'CODEXLENS_API_BATCH_SIZE_MAX': { path: ['api', 'batch_size_max'], transform: v => parseInt(v, 10) },
|
||||
'CODEXLENS_CHARS_PER_TOKEN': { path: ['api', 'chars_per_token_estimate'], transform: v => parseInt(v, 10) },
|
||||
'CODEXLENS_CASCADE_STRATEGY': { path: ['cascade', 'strategy'] },
|
||||
'CODEXLENS_CASCADE_COARSE_K': { path: ['cascade', 'coarse_k'], transform: v => parseInt(v, 10) },
|
||||
'CODEXLENS_CASCADE_FINE_K': { path: ['cascade', 'fine_k'], transform: v => parseInt(v, 10) },
|
||||
|
||||
@@ -305,7 +305,11 @@ const i18n = {
|
||||
'codexlens.envField.enabled': 'Enabled',
|
||||
'codexlens.envField.topKResults': 'Top K Results',
|
||||
'codexlens.envField.maxWorkers': 'Max Workers',
|
||||
'codexlens.envField.batchSize': 'Batch Size',
|
||||
'codexlens.envField.batchSize': 'Batch Size (Static)',
|
||||
'codexlens.envField.dynamicBatchSize': 'Dynamic Batch Size',
|
||||
'codexlens.envField.batchSizeUtilization': 'Utilization Factor',
|
||||
'codexlens.envField.batchSizeMax': 'Max Batch Size',
|
||||
'codexlens.envField.charsPerToken': 'Chars per Token',
|
||||
'codexlens.envField.searchStrategy': 'Search Strategy',
|
||||
'codexlens.envField.coarseK': 'Coarse K (1st stage)',
|
||||
'codexlens.envField.fineK': 'Fine K (final)',
|
||||
@@ -2450,7 +2454,11 @@ const i18n = {
|
||||
'codexlens.envField.enabled': '启用',
|
||||
'codexlens.envField.topKResults': 'Top K 结果数',
|
||||
'codexlens.envField.maxWorkers': '最大工作线程数',
|
||||
'codexlens.envField.batchSize': '批处理大小',
|
||||
'codexlens.envField.batchSize': '批处理大小 (固定)',
|
||||
'codexlens.envField.dynamicBatchSize': '动态批处理大小',
|
||||
'codexlens.envField.batchSizeUtilization': '容量利用率',
|
||||
'codexlens.envField.batchSizeMax': '最大批处理大小',
|
||||
'codexlens.envField.charsPerToken': '字符/Token比',
|
||||
'codexlens.envField.searchStrategy': '搜索策略',
|
||||
'codexlens.envField.coarseK': '粗筛 K (第一阶段)',
|
||||
'codexlens.envField.fineK': '精筛 K (最终)',
|
||||
|
||||
@@ -1086,7 +1086,11 @@ var ENV_VAR_GROUPS = {
|
||||
icon: 'cpu',
|
||||
vars: {
|
||||
'CODEXLENS_API_MAX_WORKERS': { labelKey: 'codexlens.envField.maxWorkers', type: 'number', placeholder: '4', default: '4', settingsPath: 'api.max_workers', min: 1, max: 32 },
|
||||
'CODEXLENS_API_BATCH_SIZE': { labelKey: 'codexlens.envField.batchSize', type: 'number', placeholder: '8', default: '8', settingsPath: 'api.batch_size', min: 1, max: 64 }
|
||||
'CODEXLENS_API_BATCH_SIZE': { labelKey: 'codexlens.envField.batchSize', type: 'number', placeholder: '8', default: '8', settingsPath: 'api.batch_size', min: 1, max: 64, showWhen: function(env) { return env['CODEXLENS_API_BATCH_SIZE_DYNAMIC'] !== 'true'; } },
|
||||
'CODEXLENS_API_BATCH_SIZE_DYNAMIC': { labelKey: 'codexlens.envField.dynamicBatchSize', type: 'checkbox', default: 'false', settingsPath: 'api.batch_size_dynamic' },
|
||||
'CODEXLENS_API_BATCH_SIZE_UTILIZATION': { labelKey: 'codexlens.envField.batchSizeUtilization', type: 'number', placeholder: '0.8', default: '0.8', settingsPath: 'api.batch_size_utilization_factor', min: 0.1, max: 0.95, step: 0.05, showWhen: function(env) { return env['CODEXLENS_API_BATCH_SIZE_DYNAMIC'] === 'true'; } },
|
||||
'CODEXLENS_API_BATCH_SIZE_MAX': { labelKey: 'codexlens.envField.batchSizeMax', type: 'number', placeholder: '2048', default: '2048', settingsPath: 'api.batch_size_max', min: 1, max: 4096, showWhen: function(env) { return env['CODEXLENS_API_BATCH_SIZE_DYNAMIC'] === 'true'; } },
|
||||
'CODEXLENS_CHARS_PER_TOKEN': { labelKey: 'codexlens.envField.charsPerToken', type: 'number', placeholder: '4', default: '4', settingsPath: 'api.chars_per_token_estimate', min: 1, max: 10, showWhen: function(env) { return env['CODEXLENS_API_BATCH_SIZE_DYNAMIC'] === 'true'; } }
|
||||
}
|
||||
},
|
||||
cascade: {
|
||||
|
||||
Reference in New Issue
Block a user