{
"event": "PreToolUse",
"tool_name": "Read",
"tool_input": {
"file_path": "\/var\/www\/scripts\/pipeline\/knowledge\/analyzer.py",
"offset": 90,
"limit": 80
}
}
{
"tool_response": {
"type": "text",
"file": {
"filePath": "\/var\/www\/scripts\/pipeline\/knowledge\/analyzer.py",
"content": " ontology = self.extract_ontology(entities, text, KnowledgeLevel.SECTION, section_id)\n taxonomy = self.extract_taxonomy(entities, text, KnowledgeLevel.SECTION, section_id)\n\n return {\n \"section_id\": section_id,\n \"entities\": len(entities),\n \"semantics\": len(semantics),\n \"ontology\": len(ontology),\n \"taxonomy\": len(taxonomy),\n }\n\n def analyze_document(self, document_id: int, text: str) -> dict:\n \"\"\"Vollständige Wissensanalyse für ein Dokument (konsolidiert).\"\"\"\n db.log(\"INFO\", f\"Starte Dokumentanalyse für document_id={document_id}\")\n\n entities = self.extract_entities(text, KnowledgeLevel.DOCUMENT, document_id)\n semantics = self.extract_semantics(entities, text, KnowledgeLevel.DOCUMENT, document_id)\n ontology = self.extract_ontology(entities, text, KnowledgeLevel.DOCUMENT, document_id)\n taxonomy = self.extract_taxonomy(entities, text, KnowledgeLevel.DOCUMENT, document_id)\n\n return {\n \"document_id\": document_id,\n \"entities\": len(entities),\n \"semantics\": len(semantics),\n \"ontology\": len(ontology),\n \"taxonomy\": len(taxonomy),\n }\n",
"numLines": 28,
"startLine": 90,
"totalLines": 117
}
}
}