{
"event": "PreToolUse",
"tool_name": "Read",
"tool_input": {
"file_path": "\/var\/www\/dev.campus.systemische-tools.de\/crawler\/tests\/unit\/application\/test_process_url.py",
"offset": 65,
"limit": 35
}
}
{
"tool_response": {
"type": "text",
"file": {
"filePath": "\/var\/www\/dev.campus.systemische-tools.de\/crawler\/tests\/unit\/application\/test_process_url.py",
"content": "\n@pytest.mark.asyncio\nasync def test_process_url_persists_page_and_discovers_links() -> None:\n pages = FakePageRepo()\n urls = FakeUrlRepo()\n links = FakeLinkRepo()\n images = FakeImageRepo()\n structured = FakeStructuredDataRepo()\n hreflang = FakeHreflangRepo()\n queue = FakeUrlQueue()\n deps = ProcessUrlDeps(\n fetcher=_StubFetcher(),\n html_parser=_StubHtmlParser(),\n sd_parser=_StubSdParser(),\n pages=pages,\n urls=urls,\n links=links,\n images=images,\n structured=structured,\n hreflang=hreflang,\n queue=queue,\n classifier=LinkClassifier(registered_domain=\"example.com\"),\n normalizer=UrlNormalizer(),\n extractor=ExtractSeoDataUseCase(calculator=QualityFlagCalculator()),\n clock=FakeClock(datetime(2026, 4, 21, 12, 0)),\n )\n uc = ProcessUrlUseCase(deps)\n start = normalize(\"https:\/\/example.com\/\")\n\n page = await uc.execute(crawl_id=1, url=start)\n\n assert page.id == 1\n assert len(links.store) == 2\n assert len(images.store) == 1\n assert queue.pending_count(1) == 1",
"numLines": 35,
"startLine": 65,
"totalLines": 101
}
}
}