Skip to content

Machines running hydra nodes become unresponsive after long periods of intense traffic #1582

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
Quantumplation opened this issue Aug 21, 2024 · 1 comment
Milestone

Comments

@Quantumplation
Copy link
Contributor

Quantumplation commented Aug 21, 2024

Context & versions

Discovered during the hydra-doom demo at Rare Evo, using both the :doom and :doom-memory-hack tagged docker images built by @ch1bo

Steps to reproduce

(This may not be a minimal reproduce and requires further investigation)

Spin up 16 hydra nodes on an r5.8xlarge with 1TB of disk space

Submit between 35 and 250 transactions per second to each node for several hours.

Reach out to me if you want help reproducing the issue, I have a saved disk snapshot from the event.

Actual behavior

Eventually, the hydra heads disconnect from websocket connections, refuse new connections, and the host itself becomes unresponsive to SSH; The machine must be power cycled to regain access.

Initially this was believed to be because of the memory leak, but even with the hacks from #1572, this would occur.
It's also unlikely to be disk space, because several of the nodes were only at 60% disk used after rebooting.

Expected behavior

A hydra node should be able to operate indefinitely at high load, provided basic assumptions like enough disk space etc.

If this is an issue with something about the server provisioning, rather than the hydra node itself, a useful outcome of this story would be to document best practices for hosting the nodes and how to avoid this scenario.

@Quantumplation Quantumplation added the bug 🐛 Something isn't working label Aug 21, 2024
@github-project-automation github-project-automation bot moved this to In Progress 🕐 in ☕ Hydra Team Work Aug 21, 2024
@noonio noonio changed the title Machines running hydra nodes become responsive after long periods of intense traffic Machines running hydra nodes become unresponsive after long periods of intense traffic Aug 22, 2024
@noonio noonio moved this to 🚀 Planned in 🚢 Hydra Head Roadmap Feb 26, 2025
@noonio noonio moved this to Blocked ✋ in ☕ Hydra Team Work Mar 4, 2025
@noonio noonio moved this from Blocked ✋ to In progress 🕐 in ☕ Hydra Team Work Mar 10, 2025
@noonio noonio removed the bug 🐛 Something isn't working label Mar 11, 2025
@noonio
Copy link
Contributor

noonio commented Mar 14, 2025

Thanks for reporting this @Quantumplation ; I think I'll close for now and assume we've addressed it with some recent networking and memory changes.

Will re-investigate if further stress-testing shows issues.

@noonio noonio closed this as not planned Won't fix, can't repro, duplicate, stale Mar 14, 2025
@github-project-automation github-project-automation bot moved this from In progress 🕐 to Done ✔ in ☕ Hydra Team Work Mar 14, 2025
@noonio noonio added this to the 0.x.x milestone Mar 27, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
Archived in project
Archived in project
Development

No branches or pull requests

2 participants