Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ShrEx observed issues #1983

Closed
renaynay opened this issue Mar 29, 2023 · 3 comments
Closed

ShrEx observed issues #1983

renaynay opened this issue Mar 29, 2023 · 3 comments
Labels
area:shares Shares and samples

Comments

@renaynay
Copy link
Member

Documenting issues we've observed so far as a note:

  • DASer doesn’t max out the concurrency lim for full nodes — even when my catchup head is ~4k headers past my head of sampled chain, i only spawn two workers with about 50 headers each
  • Full node lags significantly behind network head (even when it was already synced up to network head) for sampling when chain is maxxed out (every block @8mb)
  • Re-DASing of headers
  • False alarm blacklisting of good peers
@renaynay renaynay added area:shares Shares and samples and removed needs:triage labels Mar 29, 2023
@walldiss
Copy link
Member

DASer doesn’t max out the concurrency lim for full nodes — even when my catchup head is ~4k headers past my head of sampled chain, i only spawn two workers with about 50 headers each

It is not an issue. DASer does not make any decision based on head of sampled chain. If catchup head is same as network head, it means all headers were sent to workers. Two workers situation could happen, if those 2 are working on header ranges that is significantly slower to sample. Meaning all headers after those ranges were already processed by DASer, but not included in head of sampled chain, because it still has gaps.

Full node lags significantly behind network head

Do we see increase of "request failed" logs with DEBUG logs enabled for "share/getters"?

Re-DASing of headers

Would need logs showing the issue to identify the cause.

False alarm blacklisting of good peers

Merge peer-manager metrics PR and plug it into grafana would help to identify if the blacklist happens inside peer-manger and the reason for blacklisting.

@renaynay
Copy link
Member Author

That's a fair point re 1.

Re 2.: we need to run metrics on a robusta full node for this

Re 3: Will report it next time it's observed

Re 4: Agree. @walldiss can you make an issue to track blacklisting stability?

@renaynay
Copy link
Member Author

renaynay commented Apr 4, 2023

Closing this issue in favour of #2016

@renaynay renaynay closed this as not planned Won't fix, can't repro, duplicate, stale Apr 4, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area:shares Shares and samples
Projects
None yet
Development

No branches or pull requests

2 participants