{
"tool_response": {
"stdout": "29 | \"in\", \"im\", \"an\", \"am\", \"auf\", \"für\", \"fuer\", \"bei\", \"mit\", \"von\", \"vom\",\n |\n\nE501 Line too long (108 > 100)\n --> src\/domain\/services\/text_tokenizer.py:38:101\n |\n36 | \"warum\", \"wieso\", \"weshalb\", \"wann\", \"wo\", \"woher\", \"wohin\",\n37 | \"ja\", \"nein\", \"vielleicht\", \"immer\", \"nie\", \"manchmal\", \"oft\", \"selten\",\n38 | \"man\", \"jemand\", \"niemand\", \"etwas\", \"nichts\", \"the\", \"and\", \"or\", \"of\", \"to\", \"a\", \"for\", \"with\", \"on\",\n | ^^^^^^^^\n39 | \"at\", \"by\", \"is\", \"are\", \"were\", \"be\", \"been\", \"being\", \"have\",\n40 | \"has\", \"had\", \"do\", \"does\", \"did\", \"will\", \"would\", \"can\", \"could\",\n |\n\nFound 2 errors.",
"stderr": "",
"interrupted": false,
"isImage": false,
"noOutputExpected": false
}
}