{
"event": "PreToolUse",
"tool_name": "Read",
"tool_input": {
"file_path": "\/opt\/scripts\/pipeline\/pipeline.py",
"offset": 325,
"limit": 40
}
}
{
"tool_response": {
"type": "text",
"file": {
"filePath": "\/opt\/scripts\/pipeline\/pipeline.py",
"content": " progress.add_log(f\"FEHLER bei {file_name}: {str(e)[:50]}\")\n\n if progress:\n progress.update_progress(\n processed=processed,\n failed=failed,\n chunks=total_chunks,\n embeddings=total_embeddings,\n )\n else:\n print(\"\\n[2\/3] No new documents to process\")\n if progress:\n progress.add_log(\"Keine neuen Dokumente gefunden\")\n\n # Phase 3: Embed remaining\n if progress:\n progress.update_step(\"embed\")\n progress.add_log(\"Verarbeite ausstehende Embeddings...\")\n\n print(\"\\n[3\/3] Embedding remaining chunks...\")\n embedded = embed_pending_chunks()\n print(f\"Embedded {embedded} chunks\")\n\n if progress and embedded > 0:\n progress.add_log(f\"{embedded} weitere Embeddings erstellt\")\n\n # Complete\n print(\"\\n\" + \"=\" * 50)\n print(\"Pipeline complete!\")\n\n if progress:\n progress.add_log(\"Pipeline erfolgreich abgeschlossen\")\n progress.complete(\"completed\")\n\n except Exception as e:\n db.log(\"ERROR\", f\"Pipeline error: {e}\")\n print(f\"Error: {e}\")\n if progress:\n progress.add_log(f\"FEHLER: {str(e)}\")\n progress.complete(\"failed\", str(e))",
"numLines": 40,
"startLine": 325,
"totalLines": 448
}
}
}