{
"query": "this is a test",
"namespace": "02aaa563-baaa-43bf-b820-2dddbf405a51",
"messages": [
{
"role": "assistant",
"content": "Hallo! Ich bin hier, um Ihnen zu helfen. Wie kann ich Ihnen heute behilflich sein?\n\nM\u00f6chten Sie ein Problem im Zusammenhang mit unseren Produkten oder Angeboten beschreiben? \ud83e\udd14 Oder bevorzugen Sie vielleicht etwas anderes? Hier sind ein paar Optionen:\n\n1. **Produktproblem melden \ud83d\udc1e:** Haben Sie Probleme mit einem unserer Produkte? Erz\u00e4hlen Sie mir mehr!\n2. **Allgemeine Fragen stellen \u2753:** Haben Sie Fragen zu unseren Produkten, Dienstleistungen oder dem Unternehmen?\n3. **Produktinformationen erhalten \u2139\ufe0f:** Suchen Sie nach Informationen zu einem bestimmten Produkt?\n4. **Kaufberatung ben\u00f6tigen \ud83d\udecd\ufe0f:** Ben\u00f6tigen Sie Hilfe bei der Auswahl des richtigen Produkts f\u00fcr Ihre Bed\u00fcrfnisse?\n5. **Etwas anderes \ud83d\udca1:** Haben Sie eine andere Anfrage?\n\n\nBitte lassen Sie mich wissen, was Sie ben\u00f6tigen! \ud83d\ude0a\n"
},
{
"role": "user",
"content": "this is a test"
}
],
"stream": true,
"language_level": "",
"chat_channel": "kios:9e921dd6-f173-4b8a-9821-78013621a9e3:1743504030438",
"language": "German",
"tone": "neutral",
"writing_style": "standard",
"model": "gpt-4o-mini",
"knowledgebase": "ki-dev-large",
"seed": 1,
"client_id": 1743504030438,
"all_context": false,
"follow_up_for": null,
"knowledgebase_files_count": 163,
"override_command": null,
"disable_clarity_check": false,
"high_consistency": false,
"custom_primer": "You are a helpful virtual assistant for DigiRift GmbH, ready to assist users with their questions and needs.\r\n\r\nGeneral Guidelines:\r\n\r\nUse German in your responses.\r\n\r\nStart with a friendly greeting and a general question:\r\n\"How can I assist you today?\"\r\n\r\nHandling User Queries:\r\n\r\nIf the user seeks information:\r\n\r\nAsk clarifying questions to better understand their needs.\r\n\r\nProvide concise, accurate, and helpful information.\r\n\r\nPresent information in an easy-to-read format when appropriate.\r\n\r\nIf the user needs help with a task:\r\n\r\nBreak down complex processes into simple steps.\r\n\r\nProvide clear instructions and guidance.\r\n\r\nIf the user is looking for general conversation:\r\n\r\nEngage in a respectful and friendly manner.\r\n\r\nOffer thoughtful responses that encourage continued dialogue.\r\n\r\nSuggesting Products or Services:\r\n\r\nLook for the relevant page.\r\n\r\nProvide a detailed recommendation with an anchor tag link, formatted as follows:\r\n\r\n\r\n[replace this with the product name]\r\n \r\n\r\nEnsure all links open in a new tab (target=\"__blank\").\r\n\r\nDo not alter external URLs when translating to different languages.\r\n\r\nMultimedia Integration:\r\n\r\nIf relevant, suggest images, videos, or audio content related to the product, service, or topic, such as tutorials, testimonials, or usage guides.\r\n\r\nDisplaying Images:\r\n\r\n \r\n\r\nYouTube Videos:\r\n\r\n\r\n\r\nFor Google Drive videos, append /preview to the link and embed it:\r\n\r\n\r\n\r\nPublicly Hosted Videos:\r\n\r\n \r\n\r\nPlaying Audio:\r\n\r\n\r\n \r\n Your browser does not support the audio element.\r\n \r\n\r\nIf no multimedia content is available, skip this section.\r\n\r\nAdditional Guidelines:\r\n\r\nFocus on promoting DigiRift GmbH's offerings\u2014do not suggest competitors.\r\n\r\nProvide expert-backed general advice (e.g., best usage practices, common FAQs, industry tips).\r\n\r\nUse engaging elements like emojis where appropriate (e.g., \ud83d\uded2 for shopping, \u2b50 for quality, \ud83d\udd0d for searching).\r\n\r\nFor undecided users:\r\n\r\nOffer a few clear options to help them narrow down their needs.\r\n\r\nSuggest possible directions based on their initial query.\r\n\r\nDisclaimer:\r\n\r\nI'm designed to provide general assistance and information. For specialized advice in fields like medicine, law, finance, or other professional domains, please consult with qualified professionals.\n\n",
"logging": true,
"query_route": "list",
"web_search": false
}
QUERY ROUTE
Query Route: list
==================================================
**Elapsed Time: 0.14 seconds**
==================================================
RAG PARAMS
RAG Parameters: {'dynamically_expand': False, 'top_k': 300, 'actual_k': 120, 'satisfying_score': 0}
==================================================
**Elapsed Time: 0.00 seconds**
==================================================
VECTOR SEARCH RESULTS
Results: {'main_results': [{'id': 'c0b9c8d4-c80e-4742-a8d4-c6ef6efb91de',
'metadata': {'chunk': 0.0,
'file_name': 'gemini-quota-error.txt',
'is_dict': 'no',
'text': '429 error fires even when not doing load test\r\n'
'\r\n'
'2025-01-23 15:14:08 =======================\r\n'
'2025-01-23 15:14:08 CHAT RECEIVED\r\n'
'2025-01-23 15:14:08 =======================\r\n'
'2025-01-23 15:14:09 Retrying '
'langchain_google_vertexai.chat_models._completion_with_retry.._completion_with_retry_inner '
'in 4.0 seconds as it raised ResourceExhausted: 429 '
'Quota exceeded for '
'aiplatform.googleapis.com/generate_content_requests_per_minute_per_project_per_base_model '
'with base model: gemini-1.5-pro. Please submit a quota '
'increase request. '
'https://cloud.google.com/vertex-ai/docs/generative-ai/quotas-genai..\r\n'
'2025-01-23 15:14:14 Retrying '
'langchain_google_vertexai.chat_models._completion_with_retry.._completion_with_retry_inner '
'in 4.0 seconds as it raised ResourceExhausted: 429 '
'Quota exceeded for '
'aiplatform.googleapis.com/generate_content_requests_per_minute_per_project_per_base_model '
'with base model: gemini-1.5-pro. Please submit a quota '
'increase request. '
'https://cloud.google.com/vertex-ai/docs/generative-ai/quotas-genai..'},
'score': 0.300163597,
'values': []}, {'id': '500fbf44-69bb-4d60-8b0f-613311ead6d9',
'metadata': {'chunk': 0.0,
'file_name': 'link.txt',
'is_dict': 'no',
'text': "rebecca black's friday youtube link\r\n"
'https://www.youtube.com/watch?v=dQw4w9WgXcQ'},
'score': 0.269637913,
'values': []}, {'id': '7faee156-f57e-4930-b1a4-a0ea0e26d6e2',
'metadata': {'chunk': 1.0,
'file_name': 'apacare-primer%282%29.txt',
'is_dict': 'no',
'text': 'referrerpolicy="strict-origin-when-cross-origin"\r\n'
'allowfullscreen>\r\n'
'\r\n'
'\r\n'
'For Google Drive videos, append /preview to the link '
'and embed it:\r\n'
'\r\n'
'\r\n'
'For public URL video links, use the tag:\r\n'
' \r\n'
'\r\n'
'If no videos are available, skip the multimedia '
'content.\r\n'
'\r\n'
'Guidelines:\r\n'
'\r\n'
"Focus on promoting ApaCare's dental care products "
'onlyâ\x80\x94do not suggest competitors.\r\n'
"Suggest ApaCare's products as much as possible.\r\n"
"Suggest products from ApaCare's website's products "
'page.\r\n'
'\r\n'
'Answer general dental hygiene questions, offer '
'maintenance tips (e.g., replacing toothbrush heads, '
'flossing properly), or provide product-related FAQs.\r\n'
'Use engaging emojis where applicable (e.g., '
'ð\x9f¦·â\x9c¨ for whitening, ð\x9fª¥ for '
'toothbrushes).\r\n'
'Once a product is suggested:\r\n'
'\r\n'
'Avoid repeating the questionnaire unless the user '
'requests it.\r\n'
'Offer general dental care tips, product care advice, or '
'related topics to keep the conversation engaging.\r\n'
'For users who are unsure:\r\n'
'Provide a list of recommendations based on common '
'concerns:\r\n'
'\r\n'
'Whitening toothpaste\r\n'
'Electric toothbrushes for sensitive gums\r\n'
'\r\n'
"Don't tag your company's products links, videos, faq, "
'etc. as copyright infringement. \r\n'
'Give a disclaimer at the end of suggestions/advices '
'like "I am not a real doctor, but I can provide tips '
'based on expert recommendations. Please feel free to '
'contact our team anytime!" or similar.'},
'score': 0.243520498,
'values': []}, {'id': '38d46504-4583-43f8-9409-dcce0b7f2196',
'metadata': {'chunk': 1.0,
'file_name': 'apacare-primer%282%29.txt',
'is_dict': 'no',
'text': 'referrerpolicy="strict-origin-when-cross-origin"\r\n'
'allowfullscreen>\r\n'
'\r\n'
'\r\n'
'For Google Drive videos, append /preview to the link '
'and embed it:\r\n'
'\r\n'
'\r\n'
'For public URL video links, use the tag:\r\n'
' \r\n'
'\r\n'
'If no videos are available, skip the multimedia '
'content.\r\n'
'\r\n'
'Guidelines:\r\n'
'\r\n'
"Focus on promoting ApaCare's dental care products "
'onlyâ\x80\x94do not suggest competitors.\r\n'
"Suggest ApaCare's products as much as possible.\r\n"
"Suggest products from ApaCare's website's products "
'page.\r\n'
'\r\n'
'Answer general dental hygiene questions, offer '
'maintenance tips (e.g., replacing toothbrush heads, '
'flossing properly), or provide product-related FAQs.\r\n'
'Use engaging emojis where applicable (e.g., '
'ð\x9f¦·â\x9c¨ for whitening, ð\x9fª¥ for '
'toothbrushes).\r\n'
'Once a product is suggested:\r\n'
'\r\n'
'Avoid repeating the questionnaire unless the user '
'requests it.\r\n'
'Offer general dental care tips, product care advice, or '
'related topics to keep the conversation engaging.\r\n'
'For users who are unsure:\r\n'
'Provide a list of recommendations based on common '
'concerns:\r\n'
'\r\n'
'Whitening toothpaste\r\n'
'Electric toothbrushes for sensitive gums\r\n'
'\r\n'
"Don't tag your company's products links, videos, faq, "
'etc. as copyright infringement. \r\n'
'Give a disclaimer at the end of suggestions/advices '
'like "I am not a real doctor, but I can provide tips '
'based on expert recommendations. Please feel free to '
'contact our team anytime!" or similar.'},
'score': 0.243280068,
'values': []}, {'id': 'f3e19ec6-03e8-4217-9205-75ae8a73756f',
'metadata': {'chunk': 1.0,
'file_name': 'apacare-primer%282%29.txt',
'is_dict': 'no',
'text': 'referrerpolicy="strict-origin-when-cross-origin"\r\n'
'allowfullscreen>\r\n'
'\r\n'
'\r\n'
'For Google Drive videos, append /preview to the link '
'and embed it:\r\n'
'\r\n'
'\r\n'
'For public URL video links, use the tag:\r\n'
' \r\n'
'\r\n'
'If no videos are available, skip the multimedia '
'content.\r\n'
'\r\n'
'Guidelines:\r\n'
'\r\n'
"Focus on promoting ApaCare's dental care products "
'onlyâ\x80\x94do not suggest competitors.\r\n'
"Suggest ApaCare's products as much as possible.\r\n"
"Suggest products from ApaCare's website's products "
'page.\r\n'
'\r\n'
'Answer general dental hygiene questions, offer '
'maintenance tips (e.g., replacing toothbrush heads, '
'flossing properly), or provide product-related FAQs.\r\n'
'Use engaging emojis where applicable (e.g., '
'ð\x9f¦·â\x9c¨ for whitening, ð\x9fª¥ for '
'toothbrushes).\r\n'
'Once a product is suggested:\r\n'
'\r\n'
'Avoid repeating the questionnaire unless the user '
'requests it.\r\n'
'Offer general dental care tips, product care advice, or '
'related topics to keep the conversation engaging.\r\n'
'For users who are unsure:\r\n'
'Provide a list of recommendations based on common '
'concerns:\r\n'
'\r\n'
'Whitening toothpaste\r\n'
'Electric toothbrushes for sensitive gums\r\n'
'\r\n'
"Don't tag your company's products links, videos, faq, "
'etc. as copyright infringement. \r\n'
'Give a disclaimer at the end of suggestions/advices '
'like "I am not a real doctor, but I can provide tips '
'based on expert recommendations. Please feel free to '
'contact our team anytime!" or similar.'},
'score': 0.243178,
'values': []}, {'id': '0f436054-8a19-4bca-b593-8ac80e61b4db',
'metadata': {'chunk': 0.0,
'file_name': 'apacare-primer%282%29.txt',
'is_dict': 'no',
'text': 'You are a digital sales rep for ApaCare, a dental care '
'company. Please assist clients with their '
'dental-related questions.\r\n'
'Use German in your responses.\r\n'
'\r\n'
'Start by asking a general question:\r\n'
'"Are you looking for a specific type of dental product '
'or advice?"\r\n'
'\r\n'
'If they are looking for advice, proceed with a '
'questionnaire about their dental care needs:\r\n'
'Are they focusing on whitening, sensitivity, gum '
'health, or general hygiene?\r\n'
'Try to ask a questionnaire to have clients describe '
'their problems.\r\n'
'If they are looking for dental products:\r\n'
'give them a product suggestion from ApaCare only.\r\n'
'If they are not looking for dental products or advice, '
'skip to general suggestions or conversation.\r\n'
'\r\n'
'Once the questionnaire is complete:\r\n'
'Suggest a product and do not repeat the questionnaire '
'unless explicitly requested.\r\n'
'Format the questionnaire to be readable for the users, '
'like a list or similar.\r\n'
'\r\n'
'When suggesting a product:\r\n'
"Look for the relevant product's page in the context.\r\n"
'Provide a detailed suggestion with an anchor tag link. '
'Ensure the target attribute is set to "__blank" and use '
'this format:\r\n'
'\r\n'
'[replace this with the product name]\r\n'
' \r\n'
'\r\n'
'All links should have "__blank" target attribute.\r\n'
"Don't translate links href to German.\r\n"
'\r\n'
'Include related video suggestions:\r\n'
'\r\n'
'Search YouTube for videos about the product or topic '
'(e.g., how to use an electric toothbrush, flossing '
'techniques).\r\n'
'Embed the video in an iframe using this format:\r\n'
'
For Google Drive videos, append /preview to the link and embed it:
For public URL video links, use the tag:
If no videos are available, skip the multimedia content.
Guidelines:
Focus on promoting ApaCare's dental care products onlyâdo not suggest competitors.
Suggest ApaCare's products as much as possible.
Suggest products from ApaCare's website's products page.
Answer general dental hygiene questions, offer maintenance tips (e.g., replacing toothbrush heads, flossing properly), or provide product-related FAQs.
Use engaging emojis where applicable (e.g., ð¦·â¨ for whitening, 𪥠for toothbrushes).
Once a product is suggested:
Avoid repeating the questionnaire unless the user requests it.
Offer general dental care tips, product care advice, or related topics to keep the conversation engaging.
For users who are unsure:
Provide a list of recommendations based on common concerns:
Whitening toothpaste
Electric toothbrushes for sensitive gums
Don't tag your company's products links, videos, faq, etc. as copyright infringement.
Give a disclaimer at the end of suggestions/advices like "I am not a real doctor, but I can provide tips based on expert recommendations. Please feel free to contact our team anytime!" or similar.
####################
File: apacare-primer%282%29.txt
Page: 1
Context: You are a digital sales rep for ApaCare, a dental care company. Please assist clients with their dental-related questions.
Use German in your responses.
Start by asking a general question:
"Are you looking for a specific type of dental product or advice?"
If they are looking for advice, proceed with a questionnaire about their dental care needs:
Are they focusing on whitening, sensitivity, gum health, or general hygiene?
Try to ask a questionnaire to have clients describe their problems.
If they are looking for dental products:
give them a product suggestion from ApaCare only.
If they are not looking for dental products or advice, skip to general suggestions or conversation.
Once the questionnaire is complete:
Suggest a product and do not repeat the questionnaire unless explicitly requested.
Format the questionnaire to be readable for the users, like a list or similar.
When suggesting a product:
Look for the relevant product's page in the context.
Provide a detailed suggestion with an anchor tag link. Ensure the target attribute is set to "__blank" and use this format:
[replace this with the product name]
All links should have "__blank" target attribute.
Don't translate links href to German.
Include related video suggestions:
Search YouTube for videos about the product or topic (e.g., how to use an electric toothbrush, flossing techniques).
Embed the video in an iframe using this format:
####################
File: apacare-primer%282%29.txt
Page: 1
Context: You are a digital sales rep for ApaCare, a dental care company. Please assist clients with their dental-related questions.
Use German in your responses.
Start by asking a general question:
"Are you looking for a specific type of dental product or advice?"
If they are looking for advice, proceed with a questionnaire about their dental care needs:
Are they focusing on whitening, sensitivity, gum health, or general hygiene?
Try to ask a questionnaire to have clients describe their problems.
If they are looking for dental products:
give them a product suggestion from ApaCare only.
If they are not looking for dental products or advice, skip to general suggestions or conversation.
Once the questionnaire is complete:
Suggest a product and do not repeat the questionnaire unless explicitly requested.
Format the questionnaire to be readable for the users, like a list or similar.
When suggesting a product:
Look for the relevant product's page in the context.
Provide a detailed suggestion with an anchor tag link. Ensure the target attribute is set to "__blank" and use this format:
[replace this with the product name]
All links should have "__blank" target attribute.
Don't translate links href to German.
Include related video suggestions:
Search YouTube for videos about the product or topic (e.g., how to use an electric toothbrush, flossing techniques).
Embed the video in an iframe using this format:
####################
File: apacare-primer%282%29.txt
Page: 1
Context: You are a digital sales rep for ApaCare, a dental care company. Please assist clients with their dental-related questions.
Use German in your responses.
Start by asking a general question:
"Are you looking for a specific type of dental product or advice?"
If they are looking for advice, proceed with a questionnaire about their dental care needs:
Are they focusing on whitening, sensitivity, gum health, or general hygiene?
Try to ask a questionnaire to have clients describe their problems.
If they are looking for dental products:
give them a product suggestion from ApaCare only.
If they are not looking for dental products or advice, skip to general suggestions or conversation.
Once the questionnaire is complete:
Suggest a product and do not repeat the questionnaire unless explicitly requested.
Format the questionnaire to be readable for the users, like a list or similar.
When suggesting a product:
Look for the relevant product's page in the context.
Provide a detailed suggestion with an anchor tag link. Ensure the target attribute is set to "__blank" and use this format:
[replace this with the product name]
All links should have "__blank" target attribute.
Don't translate links href to German.
Include related video suggestions:
Search YouTube for videos about the product or topic (e.g., how to use an electric toothbrush, flossing techniques).
Embed the video in an iframe using this format:
####################
File: apacare-primer%281%29.txt
Page: 1
Context: referrerpolicy="strict-origin-when-cross-origin"
allowfullscreen>
For Google Drive videos, append /preview to the link and embed it:
For public URL video links, use the tag:
If no videos are available, skip the multimedia content.
Guidelines:
Focus on promoting ApaCare's dental care products onlyâdo not suggest competitors.
Suggest ApaCare's products as much as possible.
Suggest products from ApaCare's website's products page.
Answer general dental hygiene questions, offer maintenance tips (e.g., replacing toothbrush heads, flossing properly), or provide product-related FAQs.
Use engaging emojis where applicable (e.g., ð¦·â¨ for whitening, 𪥠for toothbrushes).
Once a product is suggested:
Avoid repeating the questionnaire unless the user requests it.
Offer general dental care tips, product care advice, or related topics to keep the conversation engaging.
For users who are unsure:
Provide a list of recommendations based on common concerns:
Whitening toothpaste
Electric toothbrushes for sensitive gums
Don't tag your company's products links, videos, faq, etc. as copyright infringement.
Give a disclaimer at the end of suggestions/advices like "I am not a real doctor, but I can provide tips based on expert recommendations. Please feel free to contact our team anytime!" or similar.
####################
File: apacare-primer%281%29.txt
Page: 1
Context: You are a digital sales rep for ApaCare, a dental care company. Please assist clients with their dental-related questions.
Use German in your responses.
Start by asking a general question:
"Are you looking for a specific type of dental product or advice?"
If they are looking for advice, proceed with a questionnaire about their dental care needs:
Are they focusing on whitening, sensitivity, gum health, or general hygiene?
Try to ask a questionnaire to have clients describe their problems.
If they are looking for dental products:
give them a product suggestion from ApaCare only.
If they are not looking for dental products or advice, skip to general suggestions or conversation.
Once the questionnaire is complete:
Suggest a product and do not repeat the questionnaire unless explicitly requested.
Format the questionnaire to be readable for the users, like a list or similar.
When suggesting a product:
Look for the relevant product's page in the context.
Provide a detailed suggestion with an anchor tag link. Ensure the target attribute is set to "__blank" and use this format:
[replace this with the product name]
All links should have "__blank" target attribute.
Don't translate links href to German.
Include related video suggestions:
Search YouTube for videos about the product or topic (e.g., how to use an electric toothbrush, flossing techniques).
Embed the video in an iframe using this format:
####################
File: apacare-primer.txt
Page: 1
Context: referrerpolicy="strict-origin-when-cross-origin"
allowfullscreen>
For Google Drive videos, append /preview to the link and embed it:
For public URL video links, use the tag:
If no videos are available, skip the multimedia content.
Guidelines:
Focus on promoting ApaCare's dental care products onlyâdo not suggest competitors.
Suggest ApaCare's products as much as possible.
Suggest products from ApaCare's website's products page.
Answer general dental hygiene questions, offer maintenance tips (e.g., replacing toothbrush heads, flossing properly), or provide product-related FAQs.
Use engaging emojis where applicable (e.g., ð¦·â¨ for whitening, 𪥠for toothbrushes).
Once a product is suggested:
Avoid repeating the questionnaire unless the user requests it.
Offer general dental care tips, product care advice, or related topics to keep the conversation engaging.
For users who are unsure:
Provide a list of recommendations based on common concerns:
Whitening toothpaste
Electric toothbrushes for sensitive gums
Don't tag your company's products links, videos, faq, etc. as copyright infringement.
Give a disclaimer at the end of suggestions/advices like "I am not a real doctor, but I can provide tips based on expert recommendations. Please feel free to contact our team anytime!" or similar.
####################
File: apacare-primer.txt
Page: 1
Context: You are a digital sales rep for ApaCare, a dental care company. Please assist clients with their dental-related questions.
Use German in your responses.
Start by asking a general question:
"Are you looking for a specific type of dental product or advice?"
If they are looking for advice, proceed with a questionnaire about their dental care needs:
Are they focusing on whitening, sensitivity, gum health, or general hygiene?
Try to ask a questionnaire to have clients describe their problems.
If they are looking for dental products:
give them a product suggestion from ApaCare only.
If they are not looking for dental products or advice, skip to general suggestions or conversation.
Once the questionnaire is complete:
Suggest a product and do not repeat the questionnaire unless explicitly requested.
Format the questionnaire to be readable for the users, like a list or similar.
When suggesting a product:
Look for the relevant product's page in the context.
Provide a detailed suggestion with an anchor tag link. Ensure the target attribute is set to "__blank" and use this format:
[replace this with the product name]
All links should have "__blank" target attribute.
Don't translate links href to German.
Include related video suggestions:
Search YouTube for videos about the product or topic (e.g., how to use an electric toothbrush, flossing techniques).
Embed the video in an iframe using this format:
####################
File: dupes.txt
Page: 1
Context: }
},
{
"stdClass":{
"id":37,
"knowledgebase_crawler_id":226,
"page_id":37,
"created_at":"2024-10-28 09:53:48.000000",
"updated_at":"2025-02-05 15:20:13.000000",
"uuid":"9d5a3e98-86b0-4ec0-a308-f417fda29c7a",
"page":"/bolg2/#/tag/semiotics",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fsemiotics-37.txt"
}
},
{
"stdClass":{
"id":38,
"knowledgebase_crawler_id":226,
"page_id":38,
"created_at":"2024-10-28 09:53:48.000000",
"updated_at":"2025-02-05 15:20:10.000000",
"uuid":"9d5a3e98-8f47-4c43-8c58-fce014133f34",
"page":"/bolg2/#/post/khidsPJEHTIwgD0pjFLc/Qui-tofu-hella,-four-dollar-toast-authentic-echo-park-in-shabby-chic-cornhole-deserunt-whatever.",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Fpost-2FkhidsPJEHTIwgD0pjFLc-2FQui-tofu-hella-2C-four-dollar-toast-authentic-echo-park-in-shabby-chic-cornhole-deserunt-whatever-38.txt"
}
},
{
####################
File: dupes.txt
Page: 1
Context: }
},
{
"stdClass":{
"id":29,
"knowledgebase_crawler_id":226,
"page_id":29,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:52.000000",
"uuid":"9d5a3e98-475c-4a56-a862-91220f6ff94d",
"page":"/bolg2/#/tag/DIY",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2FDIY-29.txt"
}
},
{
"stdClass":{
"id":30,
"knowledgebase_crawler_id":226,
"page_id":30,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:57.000000",
"uuid":"9d5a3e98-4f3d-4757-8df8-ed5b683c7a3b",
"page":"/bolg2/#/post/VYiIH8mV10vXP3x8s7Vi/Banh-mi-cold-pressed-hot-chicken-flannel-church-key-eiusmod-dolore-green-juice-actually-single-origin-coffee-meditation-air-plant.-",
"domain_id":2,
"is_viewed":0,
####################
File: dupes.txt
Page: 1
Context: "domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2FMollit-23.txt"
}
},
{
"stdClass":{
"id":25,
"knowledgebase_crawler_id":226,
"page_id":25,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:44.000000",
"uuid":"9d5a3e98-2745-42c7-a0c3-669804d74618",
"page":"/bolg2/#/tag/bushwick",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fbushwick-25.txt"
}
},
{
"stdClass":{
"id":26,
"knowledgebase_crawler_id":226,
"page_id":26,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:48.000000",
"uuid":"9d5a3e98-2f64-483e-8cc6-b22247e3f5e9",
"page":"/bolg2/#/post/eGaFHbZSNgCTmbn7UNmt/Snackwave-anim-small-batch-in-activated-charcoal-nisi.",
"domain_id":2,
"is_viewed":0,
####################
File: dupes.txt
Page: 1
Context: "domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Fpost-2FVYiIH8mV10vXP3x8s7Vi-2FBanh-mi-cold-pressed-hot-chicken-flannel-church-key-eiusmod-dolore-green-juice-actually-single-origin-coffee-meditation-air-plant-30.txt"
}
},
{
"stdClass":{
"id":31,
"knowledgebase_crawler_id":226,
"page_id":31,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:56.000000",
"uuid":"9d5a3e98-5751-491d-baeb-9f484fb8e5d4",
"page":"/bolg2/#/tag/mustache",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fmustache-31.txt"
}
},
{
"stdClass":{
"id":32,
"knowledgebase_crawler_id":226,
"page_id":32,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:59.000000",
"uuid":"9d5a3e98-5f3b-43ad-814f-854a84c627d9",
"page":"/bolg2/#/tag/lofi",
"domain_id":2,
"is_viewed":0,
####################
File: dupes.txt
Page: 1
Context: "updated_at":"2025-02-05 15:19:22.000000",
"uuid":"9d5a3e97-d202-4b78-ae70-1ab4b2b36159",
"page":"/bolg2/#/new-post",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Fnew-post-15.txt"
}
},
{
"stdClass":{
"id":16,
"knowledgebase_crawler_id":226,
"page_id":16,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:26.000000",
"uuid":"9d5a3e97-db24-48d3-886e-20dfd81362a3",
"page":"/bolg2/#/login",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Flogin-16.txt"
}
},
{
"stdClass":{
"id":17,
"knowledgebase_crawler_id":226,
"page_id":17,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:27.000000",
"uuid":"9d5a3e97-e396-4761-acb4-17231f629ec0",
"page":"/bolg2/#/post/oRqLD9jNpdFxjXZoL4Xc/New-Post-September-2021!!!",
"domain_id":2,
"is_viewed":0,
####################
File: dupes.txt
Page: 1
Context: "uuid":"9d5a3e97-f6b8-45fc-9e21-4adbd0cda3ea",
"page":"/bolg2/#/tag/taxidermy",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Ftaxidermy-19.txt"
}
},
{
"stdClass":{
"id":20,
"knowledgebase_crawler_id":226,
"page_id":20,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:35.000000",
"uuid":"9d5a3e97-ff51-45fb-bee7-2a7a7517ec13",
"page":"/bolg2/#/tag/disrupt",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fdisrupt-20.txt"
}
},
{
"stdClass":{
"id":21,
"knowledgebase_crawler_id":226,
"page_id":21,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:35.000000",
"uuid":"9d5a3e98-0843-471a-88fd-9770663f6e9d",
"page":"/bolg2/#/tag/ramps",
"domain_id":2,
"is_viewed":0,
####################
File: dupes.txt
Page: 1
Context: "updated_at":"2025-02-05 15:20:18.000000",
"uuid":"9d5a3e98-a7d1-4755-bd6f-f69c02efabb8",
"page":"/bolg2/#/tag/chic",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fchic-41.txt"
}
},
{
"stdClass":{
"id":42,
"knowledgebase_crawler_id":226,
"page_id":42,
"created_at":"2024-10-28 09:53:48.000000",
"updated_at":"2025-02-05 15:20:22.000000",
"uuid":"9d5a3e98-af2a-4c6b-b0bc-f71be13157cc",
"page":"/bolg2/#/post/xhbYdFYN0f28eysrAuQQ/Woke-gluten-free-tacos,-incididunt-et-meditation-affogato-man-braid-activated-charcoal-venmo.",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Fpost-2FxhbYdFYN0f28eysrAuQQ-2FWoke-gluten-free-tacos-2C-incididunt-et-meditation-affogato-man-braid-activated-charcoal-venmo-42.txt"
}
},
{
"stdClass":{
"id":43,
"knowledgebase_crawler_id":226,
"page_id":43,
"created_at":"2024-10-28 09:53:48.000000",
####################
File: dupes.txt
Page: 1
Context: "domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Flofi-32.txt"
}
},
{
"stdClass":{
"id":33,
"knowledgebase_crawler_id":226,
"page_id":33,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:20:01.000000",
"uuid":"9d5a3e98-6724-4fd7-96b5-ca34946da2d1",
"page":"/bolg2/#/tag/selvage",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fselvage-33.txt"
}
},
{
"stdClass":{
"id":34,
"knowledgebase_crawler_id":226,
"page_id":34,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:20:05.000000",
"uuid":"9d5a3e98-6e75-4984-ac79-78286209fcac",
"page":"/bolg2/#/post/vMDw80aFPZ2hHOWjIhV3/Voluptate-vexillologist-synth-direct-trade-heirloom-leggings.",
"domain_id":2,
"is_viewed":0,
####################
File: dupes.txt
Page: 1
Context: "domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Framps-21.txt"
}
},
{
"stdClass":{
"id":22,
"knowledgebase_crawler_id":226,
"page_id":22,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:40.000000",
"uuid":"9d5a3e98-0fb9-4152-a4a5-3d905e2aaeb7",
"page":"/bolg2/#/post/ggEuZ9Fdd5051xKzgp2l/Mollit-chillwave-chicharrones-freegan-ethical-art-party-sriracha-chia-paleo.",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Fpost-2FggEuZ9Fdd5051xKzgp2l-2FMollit-chillwave-chicharrones-freegan-ethical-art-party-sriracha-chia-paleo-22.txt"
}
},
{
"stdClass":{
"id":23,
"knowledgebase_crawler_id":226,
"page_id":23,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:39.000000",
"uuid":"9d5a3e98-17a1-4bfa-9712-57927e073b41",
"page":"/bolg2/#/tag/Mollit",
"domain_id":2,
"is_viewed":0,
####################
File: dupes.txt
Page: 1
Context: "domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Fpost-2FvMDw80aFPZ2hHOWjIhV3-2FVoluptate-vexillologist-synth-direct-trade-heirloom-leggings-34.txt"
}
},
{
"stdClass":{
"id":35,
"knowledgebase_crawler_id":226,
"page_id":35,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:20:05.000000",
"uuid":"9d5a3e98-76a3-49a4-a101-ac6edddce02c",
"page":"/bolg2/#/tag/adaptogen",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fadaptogen-35.txt"
}
},
{
"stdClass":{
"id":36,
"knowledgebase_crawler_id":226,
"page_id":36,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:20:09.000000",
"uuid":"9d5a3e98-7ee3-4177-8be6-c3dcb564de8b",
"page":"/bolg2/#/tag/craft",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fcraft-36.txt"
}
####################
File: dupes.txt
Page: 1
Context: }
},
{
"stdClass":{
"id":39,
"knowledgebase_crawler_id":226,
"page_id":39,
"created_at":"2024-10-28 09:53:48.000000",
"updated_at":"2025-02-05 15:20:14.000000",
"uuid":"9d5a3e98-978c-4686-98f2-0f4bd83ce2c0",
"page":"/bolg2/#/tag/incididunt",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fincididunt-39.txt"
}
},
{
"stdClass":{
"id":40,
"knowledgebase_crawler_id":226,
"page_id":40,
"created_at":"2024-10-28 09:53:48.000000",
"updated_at":"2025-02-05 15:20:17.000000",
"uuid":"9d5a3e98-9fc2-44fa-92fb-276e491f14e3",
"page":"/bolg2/#/tag/deserunt",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fdeserunt-40.txt"
}
},
{
"stdClass":{
"id":41,
"knowledgebase_crawler_id":226,
"page_id":41,
"created_at":"2024-10-28 09:53:48.000000",
####################
File: dupes.txt
Page: 1
Context: "domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Fpost-2FoRqLD9jNpdFxjXZoL4Xc-2FNew-Post-September-2021-21-21-21-17.txt"
}
},
{
"stdClass":{
"id":18,
"knowledgebase_crawler_id":226,
"page_id":18,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:31.000000",
"uuid":"9d5a3e97-ed2d-40fc-bd02-5baea8aaf5ea",
"page":"/bolg2/#/post/cFngsGqE4L71B1Q4fmZk/Bitters-la-croix-normcore-cred-church-key,-tilde-tote-bag-gluten-free-trust-fund-small-batch-selfies-ugh.",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Fpost-2FcFngsGqE4L71B1Q4fmZk-2FBitters-la-croix-normcore-cred-church-key-2C-tilde-tote-bag-gluten-free-trust-fund-small-batch-selfies-ugh-18.txt"
}
},
{
"stdClass":{
"id":19,
"knowledgebase_crawler_id":226,
"page_id":19,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:31.000000",
####################
File: dupes.txt
Page: 1
Context: "domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Fpost-2FeGaFHbZSNgCTmbn7UNmt-2FSnackwave-anim-small-batch-in-activated-charcoal-nisi-26.txt"
}
},
{
"stdClass":{
"id":27,
"knowledgebase_crawler_id":226,
"page_id":27,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:48.000000",
"uuid":"9d5a3e98-3768-49ae-b4fc-fea9d60e90d6",
"page":"/bolg2/#/tag/Dolore",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2FDolore-27.txt"
}
},
{
"stdClass":{
"id":28,
"knowledgebase_crawler_id":226,
"page_id":28,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:52.000000",
"uuid":"9d5a3e98-3f55-418b-98b6-d66d77d3f624",
"page":"/bolg2/#/tag/charcoal",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fcharcoal-28.txt"
}
####################
File: dupes.txt
Page: 1
Context: "updated_at":"2025-02-05 15:20:22.000000",
"uuid":"9d5a3e98-b777-4164-b0da-2b3fed2e1768",
"page":"/bolg2/#/tag/glossier",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fglossier-43.txt"
}
},
{
"stdClass":{
"id":44,
"knowledgebase_crawler_id":226,
"page_id":44,
"created_at":"2024-10-28 09:53:48.000000",
"updated_at":"2025-02-05 15:20:26.000000",
"uuid":"9d5a3e98-bee3-4170-91eb-375dab9b6ab9",
"page":"/bolg2/#/tag/schlitz",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fschlitz-44.txt"
}
},
{
"stdClass":{
"id":45,
"knowledgebase_crawler_id":226,
"page_id":45,
"created_at":"2024-10-28 09:53:48.000000",
"updated_at":"2025-02-05 15:20:26.000000",
"uuid":"9d5a3e98-c6b4-40be-b58b-942394adf716",
"page":"/bolg2/#/tag/hammock",
"domain_id":2,
"is_viewed":0,
####################
File: dupes.txt
Page: 1
Context: "domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fhammock-45.txt"
}
},
{
"stdClass":{
"id":24,
"knowledgebase_crawler_id":226,
"page_id":24,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:20:31.000000",
"uuid":"9d5a3e98-1f44-4700-8883-9e3841c448e1",
"page":"/bolg2/#/tag/pinterest",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-2Ftag-2Fpinterest-24.txt"
}
},
{
"stdClass":{
"id":46,
"knowledgebase_crawler_id":226,
"page_id":46,
"created_at":"2024-10-28 09:53:50.000000",
"updated_at":"2025-02-05 15:19:18.000000",
"uuid":"9d5a3e9c-0e90-42ff-83a3-10fd39d85244",
"page":"/Covid19Stats/#",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2FCovid19Stats-2F-23-46.txt"
}
},
{
"stdClass":{
"id":47,
"knowledgebase_crawler_id":226,
"page_id":47,
####################
File: dupes.txt
Page: 1
Context: "page_id":47,
"created_at":"2024-10-28 09:53:55.000000",
"updated_at":"2025-02-05 15:20:30.000000",
"uuid":"9d5a3ea3-6aca-4ca5-8a63-553d6af0cc55",
"page":"/RockPaperScissorsGame/#",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2FRockPaperScissorsGame-2F-23-47.txt"
}
},
{
"stdClass":{
"id":48,
"knowledgebase_crawler_id":226,
"page_id":48,
"created_at":"2024-10-28 09:53:55.000000",
"updated_at":"2025-02-05 15:20:34.000000",
"uuid":"9d5a3ea3-735b-4e8d-a8bd-3c74980e6a3e",
"page":"/RockPaperScissorsGame/#/paper",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2FRockPaperScissorsGame-2F-23-2Fpaper-48.txt"
}
},
{
"stdClass":{
"id":49,
"knowledgebase_crawler_id":226,
"page_id":49,
"created_at":"2024-10-28 09:53:55.000000",
"updated_at":"2025-02-05 15:20:34.000000",
"uuid":"9d5a3ea3-7bb7-4f3e-8aaf-ce47b17cdba1",
"page":"/RockPaperScissorsGame/#/scissors",
####################
File: dupes.txt
Page: 1
Context: {
"data":[
{
"stdClass":{
"id":2,
"knowledgebase_crawler_id":226,
"page_id":2,
"created_at":"2024-10-28 09:53:07.000000",
"updated_at":"2025-02-05 15:18:26.000000",
"uuid":"9d5a3e5b-155c-4259-890a-e990bb997f5b",
"page":"/",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2F-2.txt"
}
},
{
"stdClass":{
"id":3,
"knowledgebase_crawler_id":226,
"page_id":3,
"created_at":"2024-10-28 09:53:11.000000",
"updated_at":"2025-02-05 15:18:34.000000",
"uuid":"9d5a3e61-2c28-449a-92fb-2fe4d81b6e9e",
"page":"/#/work",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2F-23-2Fwork-3.txt"
}
},
{
"stdClass":{
"id":4,
"knowledgebase_crawler_id":226,
"page_id":4,
"created_at":"2024-10-28 09:53:11.000000",
"updated_at":"2025-02-05 15:18:33.000000",
"uuid":"9d5a3e61-356e-4948-8535-eb94619d3501",
"page":"/#/about",
####################
File: dupes.txt
Page: 1
Context: "page":"/dbxmp",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fdbxmp-12.txt"
}
},
{
"stdClass":{
"id":13,
"knowledgebase_crawler_id":226,
"page_id":13,
"created_at":"2024-10-28 09:53:19.000000",
"updated_at":"2025-02-05 15:18:58.000000",
"uuid":"9d5a3e6c-cacf-483e-85c7-5e42010a140e",
"page":"/gdmp",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fgdmp-13.txt"
}
},
{
"stdClass":{
"id":14,
"knowledgebase_crawler_id":226,
"page_id":14,
"created_at":"2024-10-28 09:53:47.000000",
"updated_at":"2025-02-05 15:19:23.000000",
"uuid":"9d5a3e97-c0ef-4c11-a00b-ce98292549f1",
"page":"/bolg2/#",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-2F-23-14.txt"
}
},
{
"stdClass":{
"id":15,
"knowledgebase_crawler_id":226,
"page_id":15,
"created_at":"2024-10-28 09:53:47.000000",
####################
File: dupes.txt
Page: 1
Context: }
},
{
"stdClass":{
"id":10,
"knowledgebase_crawler_id":226,
"page_id":10,
"created_at":"2024-10-28 09:53:19.000000",
"updated_at":"2025-02-05 15:18:50.000000",
"uuid":"9d5a3e6c-afb0-41cb-8238-098b84c40116",
"page":"/ip-address-tracker",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fip-address-tracker-10.txt"
}
},
{
"stdClass":{
"id":11,
"knowledgebase_crawler_id":226,
"page_id":11,
"created_at":"2024-10-28 09:53:19.000000",
"updated_at":"2025-02-05 15:18:54.000000",
"uuid":"9d5a3e6c-b94c-4773-839a-074fa1326a53",
"page":"/RockPaperScissorsGame",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2FRockPaperScissorsGame-11.txt"
}
},
{
"stdClass":{
"id":12,
"knowledgebase_crawler_id":226,
"page_id":12,
"created_at":"2024-10-28 09:53:19.000000",
"updated_at":"2025-02-05 15:18:56.000000",
"uuid":"9d5a3e6c-c25a-4340-9c94-a588043047ad",
####################
File: dupes.txt
Page: 1
Context: "page":"/#/about",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2F-23-2Fabout-4.txt"
}
},
{
"stdClass":{
"id":5,
"knowledgebase_crawler_id":226,
"page_id":5,
"created_at":"2024-10-28 09:53:11.000000",
"updated_at":"2025-02-05 15:18:36.000000",
"uuid":"9d5a3e61-40a1-47d1-bd7f-ac2047fb6b0a",
"page":"/#/contact",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2F-23-2Fcontact-5.txt"
}
},
{
"stdClass":{
"id":6,
"knowledgebase_crawler_id":226,
"page_id":6,
"created_at":"2024-10-28 09:53:19.000000",
"updated_at":"2025-02-05 15:18:42.000000",
"uuid":"9d5a3e6c-8a5d-430d-8b66-c5c58548e3ba",
"page":"/#",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2F-23-6.txt"
}
},
{
"stdClass":{
"id":7,
"knowledgebase_crawler_id":226,
"page_id":7,
"created_at":"2024-10-28 09:53:19.000000",
####################
File: dupes.txt
Page: 1
Context: "page":"/RockPaperScissorsGame/#/scissors",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2FRockPaperScissorsGame-2F-23-2Fscissors-49.txt"
}
},
{
"stdClass":{
"id":50,
"knowledgebase_crawler_id":226,
"page_id":50,
"created_at":"2024-10-28 09:53:55.000000",
"updated_at":"2025-02-05 15:20:37.000000",
"uuid":"9d5a3ea3-84ed-48db-814d-8b1d90e20319",
"page":"/RockPaperScissorsGame/#/rock",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2FRockPaperScissorsGame-2F-23-2Frock-50.txt"
}
}
]
}
####################
File: dupes.txt
Page: 1
Context: "updated_at":"2025-02-05 15:18:42.000000",
"uuid":"9d5a3e6c-92b5-4cb9-9a31-5430720eca62",
"page":"/cheesemax",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fcheesemax-7.txt"
}
},
{
"stdClass":{
"id":8,
"knowledgebase_crawler_id":226,
"page_id":8,
"created_at":"2024-10-28 09:53:19.000000",
"updated_at":"2025-02-05 15:18:49.000000",
"uuid":"9d5a3e6c-9c6a-4308-b0b1-4a76852f2904",
"page":"/bolg2",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2Fbolg2-8.txt"
}
},
{
"stdClass":{
"id":9,
"knowledgebase_crawler_id":226,
"page_id":9,
"created_at":"2024-10-28 09:53:19.000000",
"updated_at":"2025-02-05 15:18:46.000000",
"uuid":"9d5a3e6c-a7a1-407b-841e-b99e00d1366c",
"page":"/Covid19Stats",
"domain_id":2,
"is_viewed":0,
"txt_path":"crawler/page-content/jonasbros-github-io-2FCovid19Stats-9.txt"
}
},
{
"stdClass":{
####################
File: gpt-vector-dimension-error.txt
Page: 1
Context: 2025-01-23 15:23:13 =======================
2025-01-23 15:23:13 CHAT RECEIVED
2025-01-23 15:23:13 =======================
2025-01-23 15:23:18 INFO: 172.18.0.1:56228 - "POST /kios/knowledgebase/flexible-query/ HTTP/1.1" 500 Internal Server Error
2025-01-23 15:23:18 ERROR: Exception in ASGI application
2025-01-23 15:23:18 Traceback (most recent call last):
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/uvicorn/protocols/http/httptools_impl.py", line 426, in run_asgi
2025-01-23 15:23:18 result = await app( # type: ignore[func-returns-value]
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/uvicorn/middleware/proxy_headers.py", line 84, in __call__
2025-01-23 15:23:18 return await self.app(scope, receive, send)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/fastapi/applications.py", line 289, in __call__
2025-01-23 15:23:18 await super().__call__(scope, receive, send)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/starlette/applications.py", line 122, in __call__
2025-01-23 15:23:18 await self.middleware_stack(scope, receive, send)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/starlette/middleware/errors.py", line 184, in __call__
2025-01-23 15:23:18 raise exc
####################
File: gpt-vector-dimension-error.txt
Page: 1
Context: 2025-01-23 15:23:18 Reason: Bad Request
2025-01-23 15:23:18 HTTP response headers: HTTPHeaderDict({'content-type': 'application/json', 'Content-Length': '104', 'x-pinecone-request-latency-ms': '73', 'x-pinecone-request-id': '5221281482731493865', 'date': 'Thu, 23 Jan 2025 07:23:09 GMT', 'x-envoy-upstream-service-time': '17', 'server': 'envoy', 'Via': '1.1 google', 'Alt-Svc': 'h3=":443"; ma=2592000,h3-29=":443"; ma=2592000'})
2025-01-23 15:23:18 HTTP response body: {"code":3,"message":"Vector dimension 3072 does not match the dimension of the index 1536","details":[]}
####################
File: gpt-vector-dimension-error.txt
Page: 1
Context: 2025-01-23 15:23:18 result = context.run(func, *args)
2025-01-23 15:23:18 File "/app/main.py", line 2018, in kios_retrieve_information_via_chat
2025-01-23 15:23:18 search_results = docsearch.max_marginal_relevance_search_by_vector(res.data[0].embedding,k=actual_k, fetch_k=top_k)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/langchain_pinecone/vectorstores.py", line 307, in max_marginal_relevance_search_by_vector
2025-01-23 15:23:18 results = self._index.query(
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/pinecone/utils/error_handling.py", line 10, in inner_func
2025-01-23 15:23:18 return func(*args, **kwargs)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/pinecone/data/index.py", line 399, in query
2025-01-23 15:23:18 response = self._vector_api.query(
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/pinecone/core/client/api_client.py", line 772, in __call__
2025-01-23 15:23:18 return self.callable(self, *args, **kwargs)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/pinecone/core/client/api/data_plane_api.py", line 844, in __query
2025-01-23 15:23:18 return self.call_with_http_info(**kwargs)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/pinecone/core/client/api_client.py", line 834, in call_with_http_info
####################
File: gpt-vector-dimension-error.txt
Page: 1
Context: 2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/starlette/routing.py", line 276, in handle
2025-01-23 15:23:18 await self.app(scope, receive, send)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/starlette/routing.py", line 66, in app
2025-01-23 15:23:18 response = await func(request)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/fastapi/routing.py", line 273, in app
2025-01-23 15:23:18 raw_response = await run_endpoint_function(
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/fastapi/routing.py", line 192, in run_endpoint_function
2025-01-23 15:23:18 return await run_in_threadpool(dependant.call, **values)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/starlette/concurrency.py", line 41, in run_in_threadpool
2025-01-23 15:23:18 return await anyio.to_thread.run_sync(func, *args)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/anyio/to_thread.py", line 33, in run_sync
2025-01-23 15:23:18 return await get_asynclib().run_sync_in_worker_thread(
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/anyio/_backends/_asyncio.py", line 877, in run_sync_in_worker_thread
2025-01-23 15:23:18 return await future
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/anyio/_backends/_asyncio.py", line 807, in run
####################
File: gpt-vector-dimension-error.txt
Page: 1
Context: 2025-01-23 15:23:18 return self.api_client.call_api(
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/pinecone/core/client/api_client.py", line 409, in call_api
2025-01-23 15:23:18 return self.__call_api(resource_path, method,
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/pinecone/core/client/api_client.py", line 203, in __call_api
2025-01-23 15:23:18 raise e
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/pinecone/core/client/api_client.py", line 196, in __call_api
2025-01-23 15:23:18 response_data = self.request(
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/pinecone/core/client/api_client.py", line 455, in request
2025-01-23 15:23:18 return self.rest_client.POST(url,
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/pinecone/core/client/rest.py", line 302, in POST
2025-01-23 15:23:18 return self.request("POST", url,
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/pinecone/core/client/rest.py", line 261, in request
2025-01-23 15:23:18 raise PineconeApiException(http_resp=r)
2025-01-23 15:23:18 pinecone.core.client.exceptions.PineconeApiException: (400)
2025-01-23 15:23:18 Reason: Bad Request
####################
File: gpt-vector-dimension-error.txt
Page: 1
Context: 2025-01-23 15:23:18 raise exc
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/starlette/middleware/errors.py", line 162, in __call__
2025-01-23 15:23:18 await self.app(scope, receive, _send)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/starlette/middleware/cors.py", line 83, in __call__
2025-01-23 15:23:18 await self.app(scope, receive, send)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/starlette/middleware/exceptions.py", line 79, in __call__
2025-01-23 15:23:18 raise exc
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/starlette/middleware/exceptions.py", line 68, in __call__
2025-01-23 15:23:18 await self.app(scope, receive, sender)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/fastapi/middleware/asyncexitstack.py", line 20, in __call__
2025-01-23 15:23:18 raise e
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/fastapi/middleware/asyncexitstack.py", line 17, in __call__
2025-01-23 15:23:18 await self.app(scope, receive, send)
2025-01-23 15:23:18 File "/usr/local/lib/python3.10/dist-packages/starlette/routing.py", line 718, in __call__
2025-01-23 15:23:18 await route.handle(scope, receive, send)
####################
File: chatbot-error.txt
Page: 1
Context: INFO: 172.17.0.1:57890 - "POST /kios/knowledgebase/flexible-query/ HTTP/1.1" 500 Internal Server Error
ERROR: Exception in ASGI application
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/uvicorn/protocols/http/httptools_impl.py", line 426, in run_asgi
result = await app( # type: ignore[func-returns-value]
File "/usr/local/lib/python3.10/dist-packages/uvicorn/middleware/proxy_headers.py", line 84, in __call__
return await self.app(scope, receive, send)
File "/usr/local/lib/python3.10/dist-packages/fastapi/applications.py", line 289, in __call__
await super().__call__(scope, receive, send)
File "/usr/local/lib/python3.10/dist-packages/starlette/applications.py", line 122, in __call__
await self.middleware_stack(scope, receive, send)
File "/usr/local/lib/python3.10/dist-packages/starlette/middleware/errors.py", line 184, in __call__
raise exc
File "/usr/local/lib/python3.10/dist-packages/starlette/middleware/errors.py", line 162, in __call__
await self.app(scope, receive, _send)
File "/usr/local/lib/python3.10/dist-packages/starlette/middleware/cors.py", line 83, in __call__
await self.app(scope, receive, send)
File "/usr/local/lib/python3.10/dist-packages/starlette/middleware/exceptions.py", line 79, in __call__
raise exc
File "/usr/local/lib/python3.10/dist-packages/starlette/middleware/exceptions.py", line 68, in __call__
await self.app(scope, receive, sender)
File "/usr/local/lib/python3.10/dist-packages/fastapi/middleware/asyncexitstack.py", line 20, in __call__
raise e
####################
File: chatbot-error.txt
Page: 1
Context: raise e
File "/usr/local/lib/python3.10/dist-packages/fastapi/middleware/asyncexitstack.py", line 17, in __call__
await self.app(scope, receive, send)
File "/usr/local/lib/python3.10/dist-packages/starlette/routing.py", line 718, in __call__
await route.handle(scope, receive, send)
File "/usr/local/lib/python3.10/dist-packages/starlette/routing.py", line 276, in handle
await self.app(scope, receive, send)
File "/usr/local/lib/python3.10/dist-packages/starlette/routing.py", line 66, in app
response = await func(request)
File "/usr/local/lib/python3.10/dist-packages/fastapi/routing.py", line 273, in app
raw_response = await run_endpoint_function(
File "/usr/local/lib/python3.10/dist-packages/fastapi/routing.py", line 192, in run_endpoint_function
return await run_in_threadpool(dependant.call, **values)
File "/usr/local/lib/python3.10/dist-packages/starlette/concurrency.py", line 41, in run_in_threadpool
return await anyio.to_thread.run_sync(func, *args)
File "/usr/local/lib/python3.10/dist-packages/anyio/to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "/usr/local/lib/python3.10/dist-packages/anyio/_backends/_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
File "/usr/local/lib/python3.10/dist-packages/anyio/_backends/_asyncio.py", line 807, in run
result = context.run(func, *args)
File "/app/main.py", line 1020, in kios_retrieve_information
return process_single_query(
File "/app/main.py", line 1053, in process_single_query
provider = AIProviders(input.model, ws_connection)
File "/app/ai_providers.py", line 15, in __init__
####################
File: chatbot-error.txt
Page: 1
Context: File "/app/ai_providers.py", line 15, in __init__
self.set_provider(model)
File "/app/ai_providers.py", line 29, in set_provider
self.provider = Gemini(self.model)
File "/app/providers/gemini.py", line 10, in __init__
random_key = random.choice(config.settings.GEMINI_API_KEYS)
File "/usr/lib/python3.10/random.py", line 378, in choice
return seq[self._randbelow(len(seq))]
IndexError: list index out of range
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%285%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%284%29%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%283%29%20-%20Copy%281%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%283%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%281%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2810%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%284%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2814%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2822%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2824%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%283%29%20-%20Copy%282%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%282%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%282%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%284%29%20-%20Copy%282%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%286%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2817%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2816%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%284%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%281%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2813%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2819%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%281%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2818%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%285%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2826%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2825%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%281%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2821%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%281%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2823%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%283%29%20-%20Copy%284%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%283%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%282%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2812%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%283%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%283%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%284%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%282%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2815%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%283%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2820%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%283%29%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%281%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%282%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%287%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%282%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%289%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%2811%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%283%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%284%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%282%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%281%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%285%29%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%284%29%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy%20-%20Copy.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%288%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
####################
File: crawler-issues-19MAR2025%20-%20Copy%20%284%29%20-%20Copy%281%29.txt
Page: 1
Context: - if CrawlerJob fails statues will never update, import status wont update
(add failed() method -> create CrawlerProcess with failed status, record last process time??)
- if CrawlerProcessJob fails before recording last process time ("Cache::put($processCrawler->lastCrawlerProcessTimeCacheKey(), now());") the status will never upate
- importing failed Crawler pages still marked success
- if CrawlerFilesJob fails CrawlerProcess status wont update
- if CrawlerPrepareKnowledgebaseTrainingJob fails import status wont update
- CrawlerFilesProcessTrainingJob@handleProcessingError -- failed items are marked as processed/success.
should be markItemAsFailed() same as in CrawlerPageProcessTrainingJob?
- Finalizing Logic Duplication
The completion checking and finalization logic is duplicated across multiple jobs:
CrawlerPageProcessTrainingJob::checkCompletionAndFinalize
CrawlerFilesProcessTrainingJob::checkCompletionAndFinalize
CheckKnowledgebaseCrawlerImportCompletion::handle
Each has subtle differences, creating opportunities for inconsistent behavior.
- Unreliable S3 File Operations
File operations on S3 have minimal error handling:
$this->filesystem->put($s3Path, $newContent);
return $this->filesystem->url($s3Path);
If the S3 put operation fails silently, subsequent code would continue with a URL to a non-existent file.
- try using knowledgebase_crawler_imports table instead of cache for counting since it's already implemented?
update counts every x seconds instead of realtime updates?
- CrawlerFileProcessTrainingJob and/or CrawlerPageProcessTrainingJob failure not marking KnowledgebaseCrawler as fail
- KnowledgebaseCrawlerImport fails getting deleted after
##########
"""ANFRAGE: this is a test"""
Berücksichtigen Sie den Chatverlauf für relevante Informationen. Wenn die Anfrage bereits in der Historie gestellt wurde, überprüfen Sie die Richtigkeit Ihrer Antwort und korrigieren Sie gegebenenfalls Ihren vorherigen Fehler. Verwenden Sie so viele Tokens wie nötig, aber seien Sie gleichzeitig so effizient wie möglich.
Wichtig: Betrachten Sie die ANFRAGE und nur die ANFRAGE. Bitte versuchen Sie immer, die Frage der Anfrage zu beantworten. Wenn der Kunde nach einer Formatierungsstruktur fragt, folgen Sie seinem Rat.Aber wenn die Frage vage oder unklar ist, stellen Sie eine Folgefrage basierend auf dem Kontext.
==================================================
**Elapsed Time: 0.00 seconds**
==================================================
FINAL ANSWER
Final Answer: [Answer was streamed to client]
==================================================
**Elapsed Time: 3.66 seconds**
==================================================
SERVICES
Services: [{'type': 'chat_embeddings', 'model': 'text-embedding-3-large', 'input_tokens': 4, 'output_tokens': 0, 'total_tokens': 4}, {'type': 'chat', 'model': 'gpt-4o-mini', 'input_tokens': 44641, 'output_tokens': 7, 'total_tokens': 44648}]
==================================================
**Elapsed Time: 0.00 seconds**
==================================================