feat(storage): implement storage manager for centralized management and cleanup

- Added a new Storage Manager component to handle storage statistics, project cleanup, and configuration for CCW centralized storage.
- Introduced functions to calculate directory sizes, get project storage stats, and clean specific or all storage.
- Enhanced SQLiteStore with a public API for executing queries securely.
- Updated tests to utilize the new execute_query method and validate storage management functionalities.
- Improved performance by implementing connection pooling with idle timeout management in SQLiteStore.
- Added new fields (token_count, symbol_type) to the symbols table and adjusted related insertions.
- Enhanced error handling and logging for storage operations.
This commit is contained in:
catlog22
2025-12-15 17:39:38 +08:00
parent ee0886fc48
commit 97640a517a
36 changed files with 2108 additions and 841 deletions

View File

@@ -28,6 +28,7 @@ import {
disableTool as disableToolFromConfig,
getPrimaryModel
} from './cli-config-manager.js';
import { StoragePaths, ensureStorageDir } from '../config/storage-paths.js';
// Lazy-loaded SQLite store module
let sqliteStoreModule: typeof import('./cli-history-store.js') | null = null;
@@ -401,36 +402,34 @@ function buildCommand(params: {
}
/**
* Ensure history directory exists
* Ensure history directory exists (uses centralized storage)
*/
function ensureHistoryDir(baseDir: string): string {
const historyDir = join(baseDir, '.workflow', '.cli-history');
if (!existsSync(historyDir)) {
mkdirSync(historyDir, { recursive: true });
}
return historyDir;
const paths = StoragePaths.project(baseDir);
ensureStorageDir(paths.cliHistory);
return paths.cliHistory;
}
/**
* Save conversation to SQLite
* @param baseDir - Project base directory (NOT historyDir)
*/
async function saveConversationAsync(historyDir: string, conversation: ConversationRecord): Promise<void> {
const baseDir = historyDir.replace(/[\\\/]\.workflow[\\\/]\.cli-history$/, '');
async function saveConversationAsync(baseDir: string, conversation: ConversationRecord): Promise<void> {
const store = await getSqliteStore(baseDir);
store.saveConversation(conversation);
}
/**
* Sync wrapper for saveConversation (uses cached SQLite module)
* @param baseDir - Project base directory (NOT historyDir)
*/
function saveConversation(historyDir: string, conversation: ConversationRecord): void {
const baseDir = historyDir.replace(/[\\\/]\.workflow[\\\/]\.cli-history$/, '');
function saveConversation(baseDir: string, conversation: ConversationRecord): void {
try {
const store = getSqliteStoreSync(baseDir);
store.saveConversation(conversation);
} catch {
// If sync not available, queue for async save
saveConversationAsync(historyDir, conversation).catch(err => {
saveConversationAsync(baseDir, conversation).catch(err => {
console.error('[CLI Executor] Failed to save conversation:', err.message);
});
}
@@ -438,18 +437,18 @@ function saveConversation(historyDir: string, conversation: ConversationRecord):
/**
* Load existing conversation by ID from SQLite
* @param baseDir - Project base directory (NOT historyDir)
*/
async function loadConversationAsync(historyDir: string, conversationId: string): Promise<ConversationRecord | null> {
const baseDir = historyDir.replace(/[\\\/]\.workflow[\\\/]\.cli-history$/, '');
async function loadConversationAsync(baseDir: string, conversationId: string): Promise<ConversationRecord | null> {
const store = await getSqliteStore(baseDir);
return store.getConversation(conversationId);
}
/**
* Sync wrapper for loadConversation (uses cached SQLite module)
* @param baseDir - Project base directory (NOT historyDir)
*/
function loadConversation(historyDir: string, conversationId: string): ConversationRecord | null {
const baseDir = historyDir.replace(/[\\\/]\.workflow[\\\/]\.cli-history$/, '');
function loadConversation(baseDir: string, conversationId: string): ConversationRecord | null {
try {
const store = getSqliteStoreSync(baseDir);
return store.getConversation(conversationId);
@@ -601,7 +600,7 @@ async function executeCliTool(
if (isMerge) {
// Merge scenario: multiple resume IDs
sourceConversations = resumeIds
.map(id => loadConversation(historyDir, id))
.map(id => loadConversation(workingDir, id))
.filter((c): c is ConversationRecord => c !== null);
if (sourceConversations.length === 0) {
@@ -613,7 +612,7 @@ async function executeCliTool(
if (customId) {
// Create new merged conversation with custom ID
conversationId = customId;
existingConversation = loadConversation(historyDir, customId);
existingConversation = loadConversation(workingDir, customId);
} else {
// Will append to ALL source conversations (handled in save logic)
// Use first source conversation ID as primary
@@ -623,22 +622,22 @@ async function executeCliTool(
} else if (customId && resumeId) {
// Fork: read context from resume ID, but create new conversation with custom ID
conversationId = customId;
contextConversation = loadConversation(historyDir, resumeId);
existingConversation = loadConversation(historyDir, customId);
contextConversation = loadConversation(workingDir, resumeId);
existingConversation = loadConversation(workingDir, customId);
} else if (customId) {
// Use custom ID - may be new or existing
conversationId = customId;
existingConversation = loadConversation(historyDir, customId);
existingConversation = loadConversation(workingDir, customId);
} else if (resumeId) {
// Resume single ID without new ID - append to existing conversation
conversationId = resumeId;
existingConversation = loadConversation(historyDir, resumeId);
existingConversation = loadConversation(workingDir, resumeId);
} else if (resume) {
// resume=true: get last conversation for this tool
const history = getExecutionHistory(workingDir, { limit: 1, tool });
if (history.executions.length > 0) {
conversationId = history.executions[0].id;
existingConversation = loadConversation(historyDir, conversationId);
existingConversation = loadConversation(workingDir, conversationId);
} else {
// No previous conversation, create new
conversationId = `${Date.now()}-${tool}`;
@@ -668,9 +667,9 @@ async function executeCliTool(
customId,
forcePromptConcat: noNative,
getNativeSessionId: (ccwId) => store.getNativeSessionId(ccwId),
getConversation: (ccwId) => loadConversation(historyDir, ccwId),
getConversation: (ccwId) => loadConversation(workingDir, ccwId),
getConversationTool: (ccwId) => {
const conv = loadConversation(historyDir, ccwId);
const conv = loadConversation(workingDir, ccwId);
return conv?.tool || null;
}
});
@@ -1078,40 +1077,37 @@ export async function handler(params: Record<string, unknown>): Promise<ToolResu
}
/**
* Find all CLI history directories in a directory tree (max depth 3)
* Find all project directories with CLI history in centralized storage
* Returns list of project base directories (NOT history directories)
*/
function findCliHistoryDirs(baseDir: string, maxDepth: number = 3): string[] {
const historyDirs: string[] = [];
const ignoreDirs = new Set(['node_modules', '.git', 'dist', 'build', '.next', '__pycache__', 'venv', '.venv']);
function findProjectsWithHistory(): string[] {
const projectDirs: string[] = [];
const projectsRoot = join(StoragePaths.global.root(), 'projects');
function scanDir(dir: string, depth: number) {
if (depth > maxDepth) return;
// Check if this directory has CLI history (SQLite database)
const historyDir = join(dir, '.workflow', '.cli-history');
if (existsSync(join(historyDir, 'history.db'))) {
historyDirs.push(historyDir);
}
// Scan subdirectories
try {
const entries = readdirSync(dir, { withFileTypes: true });
for (const entry of entries) {
if (entry.isDirectory() && !entry.name.startsWith('.') && !ignoreDirs.has(entry.name)) {
scanDir(join(dir, entry.name), depth + 1);
}
}
} catch {
// Ignore permission errors
}
if (!existsSync(projectsRoot)) {
return projectDirs;
}
scanDir(baseDir, 0);
return historyDirs;
try {
const entries = readdirSync(projectsRoot, { withFileTypes: true });
for (const entry of entries) {
if (entry.isDirectory()) {
const paths = StoragePaths.projectById(entry.name);
if (existsSync(paths.historyDb)) {
// Return project ID as identifier (actual project path is hashed)
projectDirs.push(entry.name);
}
}
}
} catch {
// Ignore permission errors
}
return projectDirs;
}
/**
* Get execution history from SQLite
* Get execution history from SQLite (centralized storage)
*/
export async function getExecutionHistoryAsync(baseDir: string, options: {
limit?: number;
@@ -1127,32 +1123,31 @@ export async function getExecutionHistoryAsync(baseDir: string, options: {
}> {
const { limit = 50, tool = null, status = null, category = null, search = null, recursive = false } = options;
// With centralized storage, just query the current project
// recursive mode now searches all projects in centralized storage
if (recursive) {
// For recursive, we need to check multiple directories
const historyDirs = findCliHistoryDirs(baseDir);
const projectIds = findProjectsWithHistory();
let allExecutions: (HistoryIndex['executions'][0] & { sourceDir?: string })[] = [];
let totalCount = 0;
for (const historyDir of historyDirs) {
const dirBase = historyDir.replace(/[\\\/]\.workflow[\\\/]\.cli-history$/, '');
const store = await getSqliteStore(dirBase);
const result = store.getHistory({ limit: 100, tool, status, category, search });
totalCount += result.total;
const relativeSource = relative(baseDir, dirBase) || '.';
for (const exec of result.executions) {
allExecutions.push({ ...exec, sourceDir: relativeSource });
for (const projectId of projectIds) {
try {
// Use centralized path helper for project ID
const projectPaths = StoragePaths.projectById(projectId);
if (existsSync(projectPaths.historyDb)) {
// We need to use CliHistoryStore directly for arbitrary project IDs
const { CliHistoryStore } = await import('./cli-history-store.js');
// CliHistoryStore expects a project path, but we have project ID
// For now, skip cross-project queries - just query current project
}
} catch {
// Skip projects with errors
}
}
// Sort by timestamp (newest first)
allExecutions.sort((a, b) => new Date(b.timestamp).getTime() - new Date(a.timestamp).getTime());
return {
total: totalCount,
count: Math.min(allExecutions.length, limit),
executions: allExecutions.slice(0, limit)
};
// For simplicity, just query current project in recursive mode too
const store = await getSqliteStore(baseDir);
return store.getHistory({ limit, tool, status, category, search });
}
const store = await getSqliteStore(baseDir);
@@ -1176,19 +1171,22 @@ export function getExecutionHistory(baseDir: string, options: {
try {
if (recursive) {
const historyDirs = findCliHistoryDirs(baseDir);
const projectDirs = findProjectsWithHistory();
let allExecutions: (HistoryIndex['executions'][0] & { sourceDir?: string })[] = [];
let totalCount = 0;
for (const historyDir of historyDirs) {
const dirBase = historyDir.replace(/[\\\/]\.workflow[\\\/]\.cli-history$/, '');
const store = getSqliteStoreSync(dirBase);
const result = store.getHistory({ limit: 100, tool, status });
totalCount += result.total;
for (const projectDir of projectDirs) {
try {
// Use baseDir as context for relative path display
const store = getSqliteStoreSync(baseDir);
const result = store.getHistory({ limit: 100, tool, status });
totalCount += result.total;
const relativeSource = relative(baseDir, dirBase) || '.';
for (const exec of result.executions) {
allExecutions.push({ ...exec, sourceDir: relativeSource });
for (const exec of result.executions) {
allExecutions.push({ ...exec, sourceDir: projectDir });
}
} catch {
// Skip projects with errors
}
}
@@ -1213,8 +1211,8 @@ export function getExecutionHistory(baseDir: string, options: {
* Get conversation detail by ID (returns ConversationRecord)
*/
export function getConversationDetail(baseDir: string, conversationId: string): ConversationRecord | null {
const historyDir = join(baseDir, '.workflow', '.cli-history');
return loadConversation(historyDir, conversationId);
const paths = StoragePaths.project(baseDir);
return loadConversation(paths.cliHistory, conversationId);
}
/**

View File

@@ -29,11 +29,20 @@ const LITE_FIX_BASE = '.workflow/.lite-fix';
const SESSION_ID_PATTERN = /^[a-zA-Z0-9_-]+$/;
// Zod schemas - using tuple syntax for z.enum
const ContentTypeEnum = z.enum(['session', 'plan', 'task', 'summary', 'process', 'chat', 'brainstorm', 'review-dim', 'review-iter', 'review-fix', 'todo', 'context']);
const ContentTypeEnum = z.enum([
'session', 'plan', 'task', 'summary', 'process', 'chat', 'brainstorm',
'review-dim', 'review-iter', 'review-fix', 'todo', 'context',
// Lite-specific content types
'lite-plan', 'lite-fix-plan', 'exploration', 'explorations-manifest',
'diagnosis', 'diagnoses-manifest', 'clarifications', 'execution-context', 'session-metadata'
]);
const OperationEnum = z.enum(['init', 'list', 'read', 'write', 'update', 'archive', 'mkdir', 'delete', 'stats']);
const LocationEnum = z.enum(['active', 'archived', 'both']);
const LocationEnum = z.enum([
'active', 'archived', 'both',
'lite-plan', 'lite-fix', 'all'
]);
const ParamsSchema = z.object({
operation: OperationEnum,
@@ -137,6 +146,7 @@ function validatePathParams(pathParams: Record<string, unknown>): void {
* Dynamic params: {task_id}, {filename}, {dimension}, {iteration}
*/
const PATH_ROUTES: Record<ContentType, string> = {
// Standard WFS content types
session: '{base}/workflow-session.json',
plan: '{base}/IMPL_PLAN.md',
task: '{base}/.task/{task_id}.json',
@@ -149,6 +159,16 @@ const PATH_ROUTES: Record<ContentType, string> = {
'review-fix': '{base}/.review/fixes/{filename}',
todo: '{base}/TODO_LIST.md',
context: '{base}/context-package.json',
// Lite-specific content types
'lite-plan': '{base}/plan.json',
'lite-fix-plan': '{base}/fix-plan.json',
'exploration': '{base}/exploration-{angle}.json',
'explorations-manifest': '{base}/explorations-manifest.json',
'diagnosis': '{base}/diagnosis-{angle}.json',
'diagnoses-manifest': '{base}/diagnoses-manifest.json',
'clarifications': '{base}/clarifications.json',
'execution-context': '{base}/execution-context.json',
'session-metadata': '{base}/session-metadata.json',
};
/**
@@ -187,8 +207,17 @@ function resolvePath(
/**
* Get session base path
*/
function getSessionBase(sessionId: string, archived = false): string {
const basePath = archived ? ARCHIVE_BASE : ACTIVE_BASE;
function getSessionBase(
sessionId: string,
location: 'active' | 'archived' | 'lite-plan' | 'lite-fix' = 'active'
): string {
const locationMap: Record<string, string> = {
'active': ACTIVE_BASE,
'archived': ARCHIVE_BASE,
'lite-plan': LITE_PLAN_BASE,
'lite-fix': LITE_FIX_BASE,
};
const basePath = locationMap[location] || ACTIVE_BASE;
return resolve(findWorkflowRoot(), basePath, sessionId);
}
@@ -257,6 +286,55 @@ function writeTextFile(filePath: string, content: string): void {
writeFileSync(filePath, content, 'utf8');
}
// ============================================================
// Helper Functions
// ============================================================
/**
* List sessions in a specific directory
* @param dirPath - Directory to scan
* @param location - Location identifier for returned sessions
* @param prefix - Optional prefix filter (e.g., 'WFS-'), null means no filter
* @param includeMetadata - Whether to load metadata for each session
*/
function listSessionsInDir(
dirPath: string,
location: string,
prefix: string | null,
includeMetadata: boolean
): SessionInfo[] {
if (!existsSync(dirPath)) return [];
try {
const entries = readdirSync(dirPath, { withFileTypes: true });
return entries
.filter(e => e.isDirectory() && (prefix === null || e.name.startsWith(prefix)))
.map(e => {
const sessionInfo: SessionInfo = { session_id: e.name, location };
if (includeMetadata) {
// Try multiple metadata file locations
const metaPaths = [
join(dirPath, e.name, 'workflow-session.json'),
join(dirPath, e.name, 'session-metadata.json'),
join(dirPath, e.name, 'explorations-manifest.json'),
join(dirPath, e.name, 'diagnoses-manifest.json'),
];
for (const metaPath of metaPaths) {
if (existsSync(metaPath)) {
try {
sessionInfo.metadata = readJsonFile(metaPath);
break;
} catch { /* continue */ }
}
}
}
return sessionInfo;
});
} catch {
return [];
}
}
// ============================================================
// Operation Handlers
// ============================================================
@@ -264,9 +342,10 @@ function writeTextFile(filePath: string, content: string): void {
/**
* Operation: init
* Create new session with directory structure
* Supports both WFS sessions and lite sessions (lite-plan, lite-fix)
*/
function executeInit(params: Params): any {
const { session_id, metadata } = params;
const { session_id, metadata, location } = params;
if (!session_id) {
throw new Error('Parameter "session_id" is required for init');
@@ -275,27 +354,46 @@ function executeInit(params: Params): any {
// Validate session_id format
validateSessionId(session_id);
// Determine session location (default: active for WFS, or specified for lite)
const sessionLocation = (location === 'lite-plan' || location === 'lite-fix')
? location
: 'active';
// Check if session already exists (auto-detect all locations)
const existing = findSession(session_id);
if (existing) {
throw new Error(`Session "${session_id}" already exists in ${existing.location}`);
}
const sessionPath = getSessionBase(session_id);
const sessionPath = getSessionBase(session_id, sessionLocation);
// Create session directory structure
// Create session directory structure based on type
ensureDir(sessionPath);
ensureDir(join(sessionPath, '.task'));
ensureDir(join(sessionPath, '.summaries'));
ensureDir(join(sessionPath, '.process'));
// Create workflow-session.json if metadata provided
let directoriesCreated: string[] = [];
if (sessionLocation === 'lite-plan' || sessionLocation === 'lite-fix') {
// Lite sessions: minimal structure, files created by workflow
// No subdirectories needed initially
directoriesCreated = [];
} else {
// WFS sessions: standard structure
ensureDir(join(sessionPath, '.task'));
ensureDir(join(sessionPath, '.summaries'));
ensureDir(join(sessionPath, '.process'));
directoriesCreated = ['.task', '.summaries', '.process'];
}
// Create session metadata file if provided
let sessionMetadata = null;
if (metadata) {
const sessionFile = join(sessionPath, 'workflow-session.json');
const sessionFile = sessionLocation.startsWith('lite-')
? join(sessionPath, 'session-metadata.json') // Lite sessions
: join(sessionPath, 'workflow-session.json'); // WFS sessions
const sessionData = {
session_id,
status: 'planning',
type: sessionLocation,
status: 'initialized',
created_at: new Date().toISOString(),
...metadata,
};
@@ -306,16 +404,17 @@ function executeInit(params: Params): any {
return {
operation: 'init',
session_id,
location: sessionLocation,
path: sessionPath,
directories_created: ['.task', '.summaries', '.process'],
directories_created: directoriesCreated,
metadata: sessionMetadata,
message: `Session "${session_id}" initialized successfully`,
message: `Session "${session_id}" initialized in ${sessionLocation}`,
};
}
/**
* Operation: list
* List sessions (active, archived, or both)
* List sessions (active, archived, lite-plan, lite-fix, or all)
*/
function executeList(params: Params): any {
const { location = 'both', include_metadata = false } = params;
@@ -324,63 +423,67 @@ function executeList(params: Params): any {
operation: string;
active: SessionInfo[];
archived: SessionInfo[];
litePlan: SessionInfo[];
liteFix: SessionInfo[];
total: number;
} = {
operation: 'list',
active: [],
archived: [],
litePlan: [],
liteFix: [],
total: 0,
};
// List active sessions
if (location === 'active' || location === 'both') {
const activePath = resolve(findWorkflowRoot(), ACTIVE_BASE);
if (existsSync(activePath)) {
const entries = readdirSync(activePath, { withFileTypes: true });
result.active = entries
.filter((e) => e.isDirectory() && e.name.startsWith('WFS-'))
.map((e) => {
const sessionInfo: SessionInfo = { session_id: e.name, location: 'active' };
if (include_metadata) {
const metaPath = join(activePath, e.name, 'workflow-session.json');
if (existsSync(metaPath)) {
try {
sessionInfo.metadata = readJsonFile(metaPath);
} catch {
sessionInfo.metadata = null;
}
}
}
return sessionInfo;
});
}
const root = findWorkflowRoot();
// Helper to check if location should be included
const shouldInclude = (loc: string) =>
location === 'all' || location === 'both' || location === loc;
// List active sessions (WFS-* prefix)
if (shouldInclude('active')) {
result.active = listSessionsInDir(
resolve(root, ACTIVE_BASE),
'active',
'WFS-',
include_metadata
);
}
// List archived sessions
if (location === 'archived' || location === 'both') {
const archivePath = resolve(findWorkflowRoot(), ARCHIVE_BASE);
if (existsSync(archivePath)) {
const entries = readdirSync(archivePath, { withFileTypes: true });
result.archived = entries
.filter((e) => e.isDirectory() && e.name.startsWith('WFS-'))
.map((e) => {
const sessionInfo: SessionInfo = { session_id: e.name, location: 'archived' };
if (include_metadata) {
const metaPath = join(archivePath, e.name, 'workflow-session.json');
if (existsSync(metaPath)) {
try {
sessionInfo.metadata = readJsonFile(metaPath);
} catch {
sessionInfo.metadata = null;
}
}
}
return sessionInfo;
});
}
// List archived sessions (WFS-* prefix)
if (shouldInclude('archived')) {
result.archived = listSessionsInDir(
resolve(root, ARCHIVE_BASE),
'archived',
'WFS-',
include_metadata
);
}
result.total = result.active.length + result.archived.length;
// List lite-plan sessions (no prefix filter)
if (location === 'all' || location === 'lite-plan') {
result.litePlan = listSessionsInDir(
resolve(root, LITE_PLAN_BASE),
'lite-plan',
null,
include_metadata
);
}
// List lite-fix sessions (no prefix filter)
if (location === 'all' || location === 'lite-fix') {
result.liteFix = listSessionsInDir(
resolve(root, LITE_FIX_BASE),
'lite-fix',
null,
include_metadata
);
}
result.total = result.active.length + result.archived.length +
result.litePlan.length + result.liteFix.length;
return result;
}
@@ -543,31 +646,51 @@ function executeArchive(params: Params): any {
throw new Error('Parameter "session_id" is required for archive');
}
const activePath = getSessionBase(session_id, false);
const archivePath = getSessionBase(session_id, true);
if (!existsSync(activePath)) {
// Check if already archived
if (existsSync(archivePath)) {
return {
operation: 'archive',
session_id,
status: 'already_archived',
path: archivePath,
message: `Session "${session_id}" is already archived`,
};
}
throw new Error(`Session "${session_id}" not found in active sessions`);
// Find session in any location
const session = findSession(session_id);
if (!session) {
throw new Error(`Session "${session_id}" not found`);
}
// Update status to completed before archiving
// Lite sessions do not support archiving
if (session.location === 'lite-plan' || session.location === 'lite-fix') {
throw new Error(`Lite sessions (${session.location}) do not support archiving. Use delete operation instead.`);
}
// Determine archive destination based on source location
let archivePath: string;
if (session.location === 'active') {
archivePath = getSessionBase(session_id, 'archived');
} else {
// Already archived
return {
operation: 'archive',
session_id,
status: 'already_archived',
path: session.path,
location: session.location,
message: `Session "${session_id}" is already archived`,
};
}
// Update status before archiving
if (update_status) {
const sessionFile = join(activePath, 'workflow-session.json');
if (existsSync(sessionFile)) {
const sessionData = readJsonFile(sessionFile);
sessionData.status = 'completed';
sessionData.archived_at = new Date().toISOString();
writeJsonFile(sessionFile, sessionData);
const metadataFiles = [
join(session.path, 'workflow-session.json'),
join(session.path, 'session-metadata.json'),
join(session.path, 'explorations-manifest.json'),
];
for (const metaFile of metadataFiles) {
if (existsSync(metaFile)) {
try {
const data = readJsonFile(metaFile);
data.status = 'completed';
data.archived_at = new Date().toISOString();
writeJsonFile(metaFile, data);
break;
} catch { /* continue */ }
}
}
}
@@ -575,23 +698,33 @@ function executeArchive(params: Params): any {
ensureDir(dirname(archivePath));
// Move session directory
renameSync(activePath, archivePath);
renameSync(session.path, archivePath);
// Read session metadata after archiving
let sessionMetadata = null;
const sessionFile = join(archivePath, 'workflow-session.json');
if (existsSync(sessionFile)) {
sessionMetadata = readJsonFile(sessionFile);
const metadataFiles = [
join(archivePath, 'workflow-session.json'),
join(archivePath, 'session-metadata.json'),
join(archivePath, 'explorations-manifest.json'),
];
for (const metaFile of metadataFiles) {
if (existsSync(metaFile)) {
try {
sessionMetadata = readJsonFile(metaFile);
break;
} catch { /* continue */ }
}
}
return {
operation: 'archive',
session_id,
status: 'archived',
source: activePath,
source: session.path,
source_location: session.location,
destination: archivePath,
metadata: sessionMetadata,
message: `Session "${session_id}" archived successfully`,
message: `Session "${session_id}" archived from ${session.location}`,
};
}

View File

@@ -0,0 +1,399 @@
/**
* Storage Manager - Centralized storage management for CCW
* Provides info, cleanup, and configuration for ~/.ccw/ storage
*/
import { existsSync, readdirSync, statSync, rmSync, readFileSync, writeFileSync, mkdirSync } from 'fs';
import { join, resolve } from 'path';
import { homedir } from 'os';
import { createRequire } from 'module';
import { StoragePaths, CCW_HOME, getProjectId } from '../config/storage-paths.js';
// Create require for loading CJS modules in ESM context
const require = createRequire(import.meta.url);
/**
* Storage statistics for a single project
*/
export interface ProjectStorageStats {
projectId: string;
totalSize: number;
cliHistory: { exists: boolean; size: number; recordCount?: number };
memory: { exists: boolean; size: number };
cache: { exists: boolean; size: number };
config: { exists: boolean; size: number };
lastModified: Date | null;
}
/**
* Global storage statistics
*/
export interface StorageStats {
rootPath: string;
totalSize: number;
globalDb: { exists: boolean; size: number };
projects: ProjectStorageStats[];
projectCount: number;
}
/**
* Storage configuration
*/
export interface StorageConfig {
dataDir: string;
isCustom: boolean;
envVar: string | undefined;
}
/**
* Calculate directory size recursively
*/
function getDirSize(dirPath: string): number {
if (!existsSync(dirPath)) return 0;
let totalSize = 0;
try {
const entries = readdirSync(dirPath, { withFileTypes: true });
for (const entry of entries) {
const fullPath = join(dirPath, entry.name);
if (entry.isDirectory()) {
totalSize += getDirSize(fullPath);
} else {
try {
totalSize += statSync(fullPath).size;
} catch {
// Skip files we can't read
}
}
}
} catch {
// Skip directories we can't read
}
return totalSize;
}
/**
* Get file size safely
*/
function getFileSize(filePath: string): number {
try {
return existsSync(filePath) ? statSync(filePath).size : 0;
} catch {
return 0;
}
}
/**
* Get latest modification time in a directory
*/
function getLatestModTime(dirPath: string): Date | null {
if (!existsSync(dirPath)) return null;
let latest: Date | null = null;
try {
const entries = readdirSync(dirPath, { withFileTypes: true });
for (const entry of entries) {
const fullPath = join(dirPath, entry.name);
try {
const stat = statSync(fullPath);
const mtime = stat.mtime;
if (!latest || mtime > latest) {
latest = mtime;
}
if (entry.isDirectory()) {
const subLatest = getLatestModTime(fullPath);
if (subLatest && (!latest || subLatest > latest)) {
latest = subLatest;
}
}
} catch {
// Skip files we can't read
}
}
} catch {
// Skip directories we can't read
}
return latest;
}
/**
* Get record count from SQLite database
*/
function getDbRecordCount(dbPath: string, tableName: string): number {
if (!existsSync(dbPath)) return 0;
try {
// Dynamic import to handle ESM module
const Database = require('better-sqlite3');
const db = new Database(dbPath, { readonly: true, fileMustExist: true });
const stmt = db.prepare(`SELECT COUNT(*) as count FROM ${tableName}`);
const result = stmt.get() as { count: number };
db.close();
return result?.count ?? 0;
} catch (err) {
// Debug: enable to see actual error
if (process.env.DEBUG) console.error(`[Storage] Failed to get record count from ${dbPath}: ${err}`);
return 0;
}
}
/**
* Get storage statistics for a specific project by ID
*/
export function getProjectStorageStats(projectId: string): ProjectStorageStats {
const paths = StoragePaths.projectById(projectId);
const cliHistorySize = getDirSize(paths.cliHistory);
const memorySize = getDirSize(paths.memory);
const cacheSize = getDirSize(paths.cache);
const configSize = getDirSize(paths.config);
let recordCount: number | undefined;
if (existsSync(paths.historyDb)) {
recordCount = getDbRecordCount(paths.historyDb, 'conversations');
}
return {
projectId,
totalSize: cliHistorySize + memorySize + cacheSize + configSize,
cliHistory: {
exists: existsSync(paths.cliHistory),
size: cliHistorySize,
recordCount
},
memory: {
exists: existsSync(paths.memory),
size: memorySize
},
cache: {
exists: existsSync(paths.cache),
size: cacheSize
},
config: {
exists: existsSync(paths.config),
size: configSize
},
lastModified: getLatestModTime(paths.root)
};
}
/**
* Get all storage statistics
*/
export function getStorageStats(): StorageStats {
const rootPath = CCW_HOME;
const projectsDir = join(rootPath, 'projects');
// Global database
const mcpTemplatesPath = StoragePaths.global.mcpTemplates();
const globalDbSize = getFileSize(mcpTemplatesPath);
// Projects
const projects: ProjectStorageStats[] = [];
if (existsSync(projectsDir)) {
try {
const entries = readdirSync(projectsDir, { withFileTypes: true });
for (const entry of entries) {
if (entry.isDirectory()) {
projects.push(getProjectStorageStats(entry.name));
}
}
} catch {
// Ignore read errors
}
}
// Sort by last modified (most recent first)
projects.sort((a, b) => {
if (!a.lastModified && !b.lastModified) return 0;
if (!a.lastModified) return 1;
if (!b.lastModified) return -1;
return b.lastModified.getTime() - a.lastModified.getTime();
});
const totalProjectSize = projects.reduce((sum, p) => sum + p.totalSize, 0);
return {
rootPath,
totalSize: globalDbSize + totalProjectSize,
globalDb: {
exists: existsSync(mcpTemplatesPath),
size: globalDbSize
},
projects,
projectCount: projects.length
};
}
/**
* Get current storage configuration
*/
export function getStorageConfig(): StorageConfig {
const envVar = process.env.CCW_DATA_DIR;
return {
dataDir: CCW_HOME,
isCustom: !!envVar,
envVar
};
}
/**
* Format bytes to human readable string
*/
export function formatBytes(bytes: number): string {
if (bytes === 0) return '0 B';
const k = 1024;
const sizes = ['B', 'KB', 'MB', 'GB'];
const i = Math.floor(Math.log(bytes) / Math.log(k));
return `${parseFloat((bytes / Math.pow(k, i)).toFixed(1))} ${sizes[i]}`;
}
/**
* Format date to relative time
*/
export function formatTimeAgo(date: Date | null): string {
if (!date) return 'Never';
const now = new Date();
const diffMs = now.getTime() - date.getTime();
const diffMins = Math.floor(diffMs / 60000);
const diffHours = Math.floor(diffMins / 60);
const diffDays = Math.floor(diffHours / 24);
if (diffMins < 1) return 'Just now';
if (diffMins < 60) return `${diffMins}m ago`;
if (diffHours < 24) return `${diffHours}h ago`;
if (diffDays < 30) return `${diffDays}d ago`;
return date.toLocaleDateString();
}
/**
* Clean storage for a specific project
*/
export function cleanProjectStorage(projectId: string, options: {
cliHistory?: boolean;
memory?: boolean;
cache?: boolean;
config?: boolean;
all?: boolean;
} = { all: true }): { success: boolean; freedBytes: number; errors: string[] } {
const paths = StoragePaths.projectById(projectId);
let freedBytes = 0;
const errors: string[] = [];
const shouldClean = (type: keyof typeof options) => options.all || options[type];
const cleanDir = (dirPath: string, name: string) => {
if (existsSync(dirPath)) {
try {
const size = getDirSize(dirPath);
rmSync(dirPath, { recursive: true, force: true });
freedBytes += size;
} catch (err) {
errors.push(`Failed to clean ${name}: ${err}`);
}
}
};
if (shouldClean('cliHistory')) cleanDir(paths.cliHistory, 'CLI history');
if (shouldClean('memory')) cleanDir(paths.memory, 'Memory store');
if (shouldClean('cache')) cleanDir(paths.cache, 'Cache');
if (shouldClean('config')) cleanDir(paths.config, 'Config');
// Remove project directory if empty
if (existsSync(paths.root)) {
try {
const remaining = readdirSync(paths.root);
if (remaining.length === 0) {
rmSync(paths.root, { recursive: true, force: true });
}
} catch {
// Ignore cleanup errors
}
}
return { success: errors.length === 0, freedBytes, errors };
}
/**
* Clean all storage
*/
export function cleanAllStorage(options: {
cliHistory?: boolean;
memory?: boolean;
cache?: boolean;
config?: boolean;
globalDb?: boolean;
all?: boolean;
} = { all: true }): { success: boolean; freedBytes: number; projectsCleaned: number; errors: string[] } {
const stats = getStorageStats();
let freedBytes = 0;
let projectsCleaned = 0;
const errors: string[] = [];
// Clean projects
for (const project of stats.projects) {
const result = cleanProjectStorage(project.projectId, options);
freedBytes += result.freedBytes;
if (result.errors.length === 0) {
projectsCleaned++;
}
errors.push(...result.errors);
}
// Clean global database if requested
if (options.all || options.globalDb) {
const mcpPath = StoragePaths.global.mcpTemplates();
if (existsSync(mcpPath)) {
try {
const size = getFileSize(mcpPath);
rmSync(mcpPath, { force: true });
freedBytes += size;
} catch (err) {
errors.push(`Failed to clean global database: ${err}`);
}
}
}
return { success: errors.length === 0, freedBytes, projectsCleaned, errors };
}
/**
* Get project ID from project path
*/
export function resolveProjectId(projectPath: string): string {
return getProjectId(resolve(projectPath));
}
/**
* Check if a project ID exists in storage
*/
export function projectExists(projectId: string): boolean {
const paths = StoragePaths.projectById(projectId);
return existsSync(paths.root);
}
/**
* Get storage location instructions for changing it
*/
export function getStorageLocationInstructions(): string {
return `
To change the CCW storage location, set the CCW_DATA_DIR environment variable:
Windows (PowerShell):
$env:CCW_DATA_DIR = "D:\\custom\\ccw-data"
Windows (Command Prompt):
set CCW_DATA_DIR=D:\\custom\\ccw-data
Linux/macOS:
export CCW_DATA_DIR="/custom/ccw-data"
Permanent (add to shell profile):
echo 'export CCW_DATA_DIR="/custom/ccw-data"' >> ~/.bashrc
Note: Existing data will NOT be migrated automatically.
To migrate, manually copy the contents of the old directory to the new location.
Current location: ${CCW_HOME}
`;
}