You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Sometimes queries may be get stuck "running" forever (until redis expiration) after worker failures in redash 10. Running the query again with no changes has no effect as redash thinks the job is already running.
Steps to Reproduce
I'm not sure if these steps are deterministic but I had great success in our pre-production environment.
Refresh a dashboard with several queries that takes more than a few seconds to run
Forcibly kill the worker process
Some of the queries on the dashboard are now in this stuck state.
Technical details:
In the network tab, you can see the redash pinging the job endpoint and getting the query in "Started" state. I double checked if the "remove_ghost_locks" was running and from the logs it didn't remove these queries. It seems the main issue is rq, looking at the recent changelogs, there a bunch of improvements to error handling. Upgrading rq to rq==1.10.1 seemed to fixed this issue.
Redash Version: 10.1
Browser/OS: Firefox/Linux
How did you install Redash: docker on kubernetes
The text was updated successfully, but these errors were encountered:
We plan to update rq to gather some of these benefits. Thank you for reporting your experience that updating to 1.10.1 seemed to fix the issue for you 👌
Issue Summary
Sometimes queries may be get stuck "running" forever (until redis expiration) after worker failures in redash 10. Running the query again with no changes has no effect as redash thinks the job is already running.
Steps to Reproduce
I'm not sure if these steps are deterministic but I had great success in our pre-production environment.
Some of the queries on the dashboard are now in this stuck state.
Technical details:
In the network tab, you can see the redash pinging the job endpoint and getting the query in "Started" state. I double checked if the "remove_ghost_locks" was running and from the logs it didn't remove these queries. It seems the main issue is
rq
, looking at the recent changelogs, there a bunch of improvements to error handling. Upgrading rq torq==1.10.1
seemed to fixed this issue.The text was updated successfully, but these errors were encountered: