{
"event": "PreToolUse",
"tool_name": "Write",
"tool_input": {
"file_path": "\/var\/www\/migration\/content-pipeline\/config\/settings.env.example",
"content": "# =============================================================================\n# Content Pipeline - Environment Configuration\n# =============================================================================\n# Copy this file to .env and adjust values for your environment\n# =============================================================================\n\n# =============================================================================\n# PATHS\n# =============================================================================\n\n# Directory containing documents to process\nPIPELINE_DOCUMENT_PATH=\/path\/to\/your\/documents\n\n# =============================================================================\n# DATABASE (MariaDB\/MySQL)\n# =============================================================================\n\nDB_HOST=localhost\nDB_NAME=content_pipeline\nDB_USER=pipeline\nDB_PASSWORD=your_secure_password\n\n# Optional: Separate logging database (defaults to main DB if not set)\n# DB_LOG_HOST=localhost\n# DB_LOG_NAME=pipeline_logs\n# DB_LOG_USER=pipeline\n# DB_LOG_PASSWORD=your_secure_password\n\n# =============================================================================\n# QDRANT VECTOR DATABASE\n# =============================================================================\n\nQDRANT_HOST=localhost\nQDRANT_PORT=6333\n\n# Collection names (customize as needed)\nQDRANT_COLLECTION_DOCUMENTS=documents\nQDRANT_COLLECTION_ENTITIES=entities\n\n# =============================================================================\n# OLLAMA (Local LLM)\n# =============================================================================\n\nOLLAMA_HOST=http:\/\/localhost:11434\n\n# Embedding model (must match EMBEDDING_DIMENSION)\nOLLAMA_EMBEDDING_MODEL=mxbai-embed-large\nEMBEDDING_DIMENSION=1024\n\n# Chat model for analysis\nOLLAMA_CHAT_MODEL=llama3.2:3b\n\n# =============================================================================\n# ANTHROPIC API (Optional - for higher quality analysis)\n# =============================================================================\n\n# Uncomment and set to enable Anthropic API\n# ANTHROPIC_API_KEY=sk-ant-...\n# ANTHROPIC_MODEL=claude-sonnet-4-20250514\n\n# =============================================================================\n# CHUNKING\n# =============================================================================\n\nCHUNK_OVERLAP_PERCENT=10\nMIN_CHUNK_SIZE=100\nMAX_CHUNK_SIZE=2000\n\n# =============================================================================\n# PIPELINE\n# =============================================================================\n\nPIPELINE_MAX_RETRIES=3\nPIPELINE_RETRY_BACKOFF=2\n\n# Semantic analysis mode\nSEMANTIC_SYNC=true\nSEMANTIC_USE_ANTHROPIC=false\nSEMANTIC_QUEUE_BATCH_SIZE=5\n\n# =============================================================================\n# OCR\n# =============================================================================\n\nOCR_ENABLED=true\nOCR_LANGUAGE=deu\n\n# =============================================================================\n# LOGGING\n# =============================================================================\n\nLOG_LEVEL=INFO\n"
}
}
{
"tool_response": {
"type": "create",
"filePath": "\/var\/www\/migration\/content-pipeline\/config\/settings.env.example",
"content": "# =============================================================================\n# Content Pipeline - Environment Configuration\n# =============================================================================\n# Copy this file to .env and adjust values for your environment\n# =============================================================================\n\n# =============================================================================\n# PATHS\n# =============================================================================\n\n# Directory containing documents to process\nPIPELINE_DOCUMENT_PATH=\/path\/to\/your\/documents\n\n# =============================================================================\n# DATABASE (MariaDB\/MySQL)\n# =============================================================================\n\nDB_HOST=localhost\nDB_NAME=content_pipeline\nDB_USER=pipeline\nDB_PASSWORD=your_secure_password\n\n# Optional: Separate logging database (defaults to main DB if not set)\n# DB_LOG_HOST=localhost\n# DB_LOG_NAME=pipeline_logs\n# DB_LOG_USER=pipeline\n# DB_LOG_PASSWORD=your_secure_password\n\n# =============================================================================\n# QDRANT VECTOR DATABASE\n# =============================================================================\n\nQDRANT_HOST=localhost\nQDRANT_PORT=6333\n\n# Collection names (customize as needed)\nQDRANT_COLLECTION_DOCUMENTS=documents\nQDRANT_COLLECTION_ENTITIES=entities\n\n# =============================================================================\n# OLLAMA (Local LLM)\n# =============================================================================\n\nOLLAMA_HOST=http:\/\/localhost:11434\n\n# Embedding model (must match EMBEDDING_DIMENSION)\nOLLAMA_EMBEDDING_MODEL=mxbai-embed-large\nEMBEDDING_DIMENSION=1024\n\n# Chat model for analysis\nOLLAMA_CHAT_MODEL=llama3.2:3b\n\n# =============================================================================\n# ANTHROPIC API (Optional - for higher quality analysis)\n# =============================================================================\n\n# Uncomment and set to enable Anthropic API\n# ANTHROPIC_API_KEY=sk-ant-...\n# ANTHROPIC_MODEL=claude-sonnet-4-20250514\n\n# =============================================================================\n# CHUNKING\n# =============================================================================\n\nCHUNK_OVERLAP_PERCENT=10\nMIN_CHUNK_SIZE=100\nMAX_CHUNK_SIZE=2000\n\n# =============================================================================\n# PIPELINE\n# =============================================================================\n\nPIPELINE_MAX_RETRIES=3\nPIPELINE_RETRY_BACKOFF=2\n\n# Semantic analysis mode\nSEMANTIC_SYNC=true\nSEMANTIC_USE_ANTHROPIC=false\nSEMANTIC_QUEUE_BATCH_SIZE=5\n\n# =============================================================================\n# OCR\n# =============================================================================\n\nOCR_ENABLED=true\nOCR_LANGUAGE=deu\n\n# =============================================================================\n# LOGGING\n# =============================================================================\n\nLOG_LEVEL=INFO\n",
"structuredPatch": [],
"originalFile": null
}
}