Michael Natkin
11/20/2024, 10:52 PMDEFAULT 2024-11-20T22:47:34.489866Z 22:47:34.490 | DEBUG | APILogWorkerThread | prefect._internal.concurrency - Encountered exception in call get(<dropped>)
{
"textPayload": "22:47:34.490 | DEBUG | APILogWorkerThread | prefect._internal.concurrency - Encountered exception in call get(<dropped>)",
"insertId": "673e67060007798ab687466f",
"resource": {
"type": "cloud_run_revision",
"labels": {
"service_name": "ringmaster",
"location": "us-central1",
"project_id": "michaeln-test-382306",
"configuration_name": "ringmaster",
"revision_name": "ringmaster-00004-gsf"
}
},
"timestamp": "2024-11-20T22:47:34.489866Z",
"labels": {
"instanceId": "004940b3b8bdb9d5a462c0ad04b92a01622f82b533a48e105fd08a4da715b51c769d7159c3af52c69abea8fa17bc09de6edfd151bbd66a8122ebf2f35c5c315c269a9307"
},
"logName": "projects/michaeln-test-382306/logs/run.googleapis.com%2Fstderr",
"receiveTimestamp": "2024-11-20T22:47:34.576867541Z"
}
ERROR 2024-11-20T22:47:34.489890Z Traceback (most recent call last): File "/usr/local/lib/python3.12/site-packages/prefect/_internal/concurrency/calls.py", line 346, in _run_sync result = self.fn(*self.args, **self.kwargs)
{
"textPayload": "Traceback (most recent call last):\n File \"/usr/local/lib/python3.12/site-packages/prefect/_internal/concurrency/calls.py\", line 346, in _run_sync\n result = self.fn(*self.args, **self.kwargs)",
"insertId": "673e6706000779a2ecd50a01",
"resource": {
"type": "cloud_run_revision",
"labels": {
"revision_name": "ringmaster-00004-gsf",
"project_id": "michaeln-test-382306",
"configuration_name": "ringmaster",
"service_name": "ringmaster",
"location": "us-central1"
}
},
"timestamp": "2024-11-20T22:47:34.489890Z",
"severity": "ERROR",
"labels": {
"instanceId": "004940b3b8bdb9d5a462c0ad04b92a01622f82b533a48e105fd08a4da715b51c769d7159c3af52c69abea8fa17bc09de6edfd151bbd66a8122ebf2f35c5c315c269a9307"
},
"logName": "projects/michaeln-test-382306/logs/run.googleapis.com%2Fstderr",
"receiveTimestamp": "2024-11-20T22:47:34.576867541Z",
"errorGroups": [
{
"id": "CNiO7oe-2uKcuwE"
}
]
}
DEFAULT 2024-11-20T22:47:34.489892Z ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
{
"textPayload": " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^",
"insertId": "673e6706000779a4d4358528",
"resource": {
"type": "cloud_run_revision",
"labels": {
"project_id": "michaeln-test-382306",
"revision_name": "ringmaster-00004-gsf",
"location": "us-central1",
"configuration_name": "ringmaster",
"service_name": "ringmaster"
}
},
"timestamp": "2024-11-20T22:47:34.489892Z",
"labels": {
"instanceId": "004940b3b8bdb9d5a462c0ad04b92a01622f82b533a48e105fd08a4da715b51c769d7159c3af52c69abea8fa17bc09de6edfd151bbd66a8122ebf2f35c5c315c269a9307"
},
"logName": "projects/michaeln-test-382306/logs/run.googleapis.com%2Fstderr",
"receiveTimestamp": "2024-11-20T22:47:34.576867541Z"
}
DEFAULT 2024-11-20T22:47:34.489895Z File "/usr/local/lib/python3.12/queue.py", line 179, in get
{
"textPayload": " File \"/usr/local/lib/python3.12/queue.py\", line 179, in get",
"insertId": "673e6706000779a75d59a4a1",
"resource": {
"type": "cloud_run_revision",
"labels": {
"location": "us-central1",
"service_name": "ringmaster",
"revision_name": "ringmaster-00004-gsf",
"project_id": "michaeln-test-382306",
"configuration_name": "ringmaster"
}
},
"timestamp": "2024-11-20T22:47:34.489895Z",
"labels": {
"instanceId": "004940b3b8bdb9d5a462c0ad04b92a01622f82b533a48e105fd08a4da715b51c769d7159c3af52c69abea8fa17bc09de6edfd151bbd66a8122ebf2f35c5c315c269a9307"
},
"logName": "projects/michaeln-test-382306/logs/run.googleapis.com%2Fstderr",
"receiveTimestamp": "2024-11-20T22:47:34.576867541Z"
}
DEFAULT 2024-11-20T22:47:34.489897Z raise Empty
{
"textPayload": " raise Empty",
"insertId": "673e6706000779a92b56f0db",
"resource": {
"type": "cloud_run_revision",
"labels": {
"configuration_name": "ringmaster",
"project_id": "michaeln-test-382306",
"location": "us-central1",
"service_name": "ringmaster",
"revision_name": "ringmaster-00004-gsf"
}
},
"timestamp": "2024-11-20T22:47:34.489897Z",
"labels": {
"instanceId": "004940b3b8bdb9d5a462c0ad04b92a01622f82b533a48e105fd08a4da715b51c769d7159c3af52c69abea8fa17bc09de6edfd151bbd66a8122ebf2f35c5c315c269a9307"
},
"logName": "projects/michaeln-test-382306/logs/run.googleapis.com%2Fstderr",
"receiveTimestamp": "2024-11-20T22:47:34.576867541Z"
}
DEFAULT 2024-11-20T22:47:34.489900Z _queue.Empty
{
"textPayload": "_queue.Empty",
"insertId": "673e6706000779ac6aadf983",
"resource": {
"type": "cloud_run_revision",
"labels": {
"revision_name": "ringmaster-00004-gsf",
"configuration_name": "ringmaster",
"location": "us-central1",
"service_name": "ringmaster",
"project_id": "michaeln-test-382306"
}
},
"timestamp": "2024-11-20T22:47:34.489900Z",
"labels": {
"instanceId": "004940b3b8bdb9d5a462c0ad04b92a01622f82b533a48e105fd08a4da715b51c769d7159c3af52c69abea8fa17bc09de6edfd151bbd66a8122ebf2f35c5c315c269a9307"
},
"logName": "projects/michaeln-test-382306/logs/run.googleapis.com%2Fstderr",
"receiveTimestamp": "2024-11-20T22:47:34.576867541Z"
}
Marvin
11/20/2024, 10:52 PMwhen runnning my container that uses .serve() on google cloud run, I see a lot of these errors```DEFAULT 2024-11-20T224734.489866Z 224734.490 | DEBUG | APILogWorkerThread | prefect._internal.concurrency - Encountered exception in call get(<dropped>) { "textPayload": "224734.490 | DEBUG | APILogWorkerThread | prefect._internal.concurrency - Encountered exception in call get(<dropped>)", "insertId": "673e67060007798ab687466f", "resource": { "type": "cloud_run_revision", "labels": { "service_name": "ringmaster", "location": "us-central1", "project_id": "michaeln-test-382306", "configuration_name": "ringmaster", "revision_name": "ringmaster-00004-gsf" } }, "timestamp": "2024-11-20T224734.489866Z", "labels": { "instanceId": "004940b3b8bdb9d5a462c0ad04b92a01622f82b533a48e105fd08a4da715b51c769d7159c3af52c69abe
Michael Natkin
11/20/2024, 10:53 PM@Marvin when runnning my container that uses .serve() on google cloud run, I see a lot of these errors```DEFAULT 2024-11-20T224734.489866Z 224734.490 | DEBUG | APILogWorkerThread | prefect._internal.concurrency - Encountered exception in call get(<dropped>) { "textPayload": "224734.490 | DEBUG | APILogWorkerThread | prefect._internal.concurrency - Encountered exception in call get(<dropped>)", "insertId": "673e67060007798ab687466f", "resource": { "type": "cloud_run_revision", "labels": { "service_name": "ringmaster", "location": "us-central1", "project_id": "michaeln-test-382306", "configuration_name": "ringmaster", "revision_name": "ringmaster-00004-gsf" } }, "timestamp": "2024-11-20T224734.489866Z", "labels": { "instanceId": "004940b3b8bdb9d5a462c0ad04b92a01622f82b533a48e105fd08a4da715b51c769d7159c3af52c69abe
Marvin
11/20/2024, 10:53 PM> when runnning my container that uses .serve() on google cloud run, I see a lot of these errors```DEFAULT 2024-11-20T224734.489866Z 224734.490 | DEBUG | APILogWorkerThread | prefect._internal.concurrency - Encountered exception in call get(<dropped>) { "textPayload": "224734.490 | DEBUG | APILogWorkerThread | prefect._internal.concurrency - Encountered exception in call get(<dropped>)", "insertId": "673e67060007798ab687466f", "resource": { "type": "cloud_run_revision", "labels": { "service_name": "ringmaster", "location": "us-central1", "project_id": "michaeln-test-382306", "configuration_name": "ringmaster", "revision_name": "ringmaster-00004-gsf" } }, "timestamp": "2024-11-20T224734.489866Z", "labels": { "instanceId": "004940b3b8bdb9d5a462c0ad04b92a01622f82b533a48e105fd08a4da715b51c769d7159c3af52
Michael Natkin
11/20/2024, 10:56 PMMarvin
11/20/2024, 10:57 PMget()
from a queue, leading to an _queue.Empty
exception. This generally indicates that the thread attempted to retrieve an item from a queue, but the queue was empty when the call was made.
To help you further, let me search the relevant Prefect documentation and GitHub issues to see if there are any similar issues or guidance for resolving this problem. This will also help identify if there are any specific version-related details.
I'll look into this and get back to you shortly!
It seems we hit a bit of a bump with the documentation tools just now, but I did find some relevant GitHub issues that might shine a light on your troubles.
There are documented cases where concurrency-related errors, like APILogWorkerThread
encountering _queue.Empty
, occur. One particular issue mentions that this kind of error might be linked with task runs going missing or concurrency challenges when the Prefect setup involves threading or the API logging worker is not able to process items as expected.
Here's what you might consider doing:
1. Version Check: First, ensure you're using a stable version of Prefect. There have been instances where similar errors were version-specific.
2. Concurrency Handling: You might want to ensure that the threads handling logging or other concurrent operations aren't being starved of necessary resources. This could mean re-evaluating how much concurrency (in terms of threads or processes) your environment can effectively sustain.
3. Investigate Worker Setup: If you're using workers in a Kubernetes setup or another orchestrator, ensure their configurations align with Prefect's best practices. Sometimes tweaking how workers or agents are configured can alleviate these issues.
4. Check for Updates: Keep an eye on Prefect’s release notes or GitHub issues for updates or patches regarding concurrency. Sometimes these issues are known and have fixes in newer releases.
For more details, you can explore the GitHub issue here.
I hope this helps! If the problem persists, or if you have more context to share, feel free to ask away. Remember, debugging is just another way of saying "You're learning magic!"