{
"event": "PreToolUse",
"tool_name": "Read",
"tool_input": {
"file_path": "\/var\/www\/dev.campus.systemische-tools.de\/crawler\/src\/interfaces\/cli\/wiring.py",
"offset": 70,
"limit": 55
}
}
{
"tool_response": {
"type": "text",
"file": {
"filePath": "\/var\/www\/dev.campus.systemische-tools.de\/crawler\/src\/interfaces\/cli\/wiring.py",
"content": " )\n\n\nclass Container:\n def __init__(self, conn: MariaDbConnectionFactory) -> None:\n self.conn = conn\n self.repos = _build_repositories(conn)\n self.clock = SystemClock()\n self.normalizer = UrlNormalizer(strip_tracking=False)\n self.queue = DbUrlQueue(conn)\n self.html_parser = Bs4HtmlParser()\n self.sd_parser = ExtructStructuredDataParser()\n self.calculator = QualityFlagCalculator()\n self.extractor = ExtractSeoDataUseCase(calculator=self.calculator)\n\n def start_crawl(self) -> StartCrawlUseCase:\n return StartCrawlUseCase(\n crawls=self.repos.crawls,\n urls=self.repos.urls,\n queue=self.queue,\n normalizer=self.normalizer,\n clock=self.clock,\n )\n\n def process_url(\n self,\n fetcher: HttpFetcherPort,\n *,\n registered_domain: str,\n render_mode: RenderMode = RenderMode.HTTP,\n ) -> ProcessUrlUseCase:\n deps = ProcessUrlDeps(\n fetcher=fetcher,\n html_parser=self.html_parser,\n sd_parser=self.sd_parser,\n pages=self.repos.pages,\n urls=self.repos.urls,\n links=self.repos.links,\n images=self.repos.images,\n structured=self.repos.structured,\n hreflang=self.repos.hreflang,\n queue=self.queue,\n classifier=LinkClassifier(registered_domain=registered_domain),\n normalizer=self.normalizer,\n extractor=self.extractor,\n clock=self.clock,\n render_mode=render_mode,\n )\n return ProcessUrlUseCase(deps)\n\n def finalize(self) -> FinalizeCrawlUseCase:\n return FinalizeCrawlUseCase(\n crawls=self.repos.crawls,\n errors=self.repos.errors,\n clock=self.clock,",
"numLines": 55,
"startLine": 70,
"totalLines": 159
}
}
}