{
"event": "PreToolUse",
"tool_name": "Read",
"tool_input": {
"file_path": "\/var\/www\/scripts\/pipeline\/knowledge.py",
"offset": 115,
"limit": 20
}
}
{
"tool_response": {
"type": "text",
"file": {
"filePath": "\/var\/www\/scripts\/pipeline\/quality_test.py",
"content": " \"\"\"Run prompt through Ollama model.\"\"\"\n start = time.time()\n try:\n response = requests.post(\n f\"{OLLAMA_HOST}\/api\/generate\",\n json={\n \"model\": model,\n \"prompt\": prompt,\n \"stream\": False,\n \"format\": \"json\",\n \"options\": {\"temperature\": 0.3, \"num_predict\": 2000},\n },\n timeout=timeout,\n )\n response.raise_for_status()\n data = response.json()\n elapsed = time.time() - start\n return {\n \"response\": data.get(\"response\", \"\"),\n \"tokens\": data.get(\"eval_count\", 0),",
"numLines": 20,
"startLine": 45,
"totalLines": 350
}
}
}