We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
0.0.24
We just put 0.0.24 into production and it's starting to get into a state where river is logging hundreds of times a second with
river
{ "level": "error", "time": "2024-03-01T23:56:08.679305457Z", "notifier": { "err": { "error": "connection already established", "kind": "*errors.errorString", "stack": null } }, "subsystem": "river", "message": "error establishing connection from pool" }
via:
river/internal/notifier/notifier.go
Line 130 in 035ba59
river/riverdriver/riverpgxv5/river_pgx_v5_driver.go
Line 491 in 035ba59
(My apologies for the lack of a stack trace)
This is thrashing CPU significantly (the small bumps are the deployments that haven't rolled out 0.0.24 yet):
I managed to capture the lead up to ONE of these bursts and it looks like a TCP error on the PostgreSQL connection started the death spiral
{ "level": "info", "time": "2024-03-01T23:56:07.097802512Z", "num_completed_jobs": 40, "num_jobs_running": 0, "queue": "default", "subsystem": "river", "message": "producer: Heartbeat" } { "level": "error", "time": "2024-03-01T23:56:08.679126682Z", "notifier": { "err": { "error": "tls: failed to send closeNotify alert (but connection was closed anyway): write tcp 10.122.48.181:44034->10.122.30.240:5432: i/o timeout", "kind": "*fmt.wrapError", "stack": null } }, "subsystem": "river", "message": "error closing listener" } { "level": "error", "time": "2024-03-01T23:56:08.679305457Z", "notifier": { "err": { "error": "connection already established", "kind": "*errors.errorString", "stack": null } }, "subsystem": "river", "message": "error establishing connection from pool" }
I'm sorry I don't have anything more conclusive right now but wanted to at least get this in your hands in case something rings a bell.
The text was updated successfully, but these errors were encountered:
@dhermes thank you, I believe this is the same issue reported in #239 (comment) and it should be fixed by #246. Reviewing that now!
Sorry, something went wrong.
riverpgxv5
@dhermes Hey, sorry about the trouble here. We just cut v0.0.25, which should resolve the problem.
Thanks all!
Successfully merging a pull request may close this issue.
We just put
0.0.24
into production and it's starting to get into a state whereriver
is logging hundreds of times a second withvia:
river/internal/notifier/notifier.go
Line 130 in 035ba59
river/riverdriver/riverpgxv5/river_pgx_v5_driver.go
Line 491 in 035ba59
(My apologies for the lack of a stack trace)
This is thrashing CPU significantly (the small bumps are the deployments that haven't rolled out
0.0.24
yet):I managed to capture the lead up to ONE of these bursts and it looks like a TCP error on the PostgreSQL connection started the death spiral
I'm sorry I don't have anything more conclusive right now but wanted to at least get this in your hands in case something rings a bell.
The text was updated successfully, but these errors were encountered: