-
Notifications
You must be signed in to change notification settings - Fork 2.6k
Telemetry spam #7960
Comments
@cecton did you change the default buffer size with your pr or was it 16 before as well? Edit: Yeah it was 16 before. Maybe the new worker is too slow? |
@arkpar Do you know if this problem is recent, i.e. if it worked better before #7463? Item 4 seems like the lowest-hanging fruit, shall we make that the first priority item on the list? Item 2 seems pretty hard to do properly. cc @maciejhirsz |
@dvdplm Not sure. It wasn't there a couple of weeks ago, but I can't be sure at this point. |
I don't think it's recent (the spam that is, not the error reporting), Telemetry has a built in throttling for block updates during sync for that purpose so it doesn't affect the server that much, although the extra network traffic and json parsing we could definitely do without. |
I can say for the point 3 it is not possible to log at this place. This is why it is a eprintln. That being said I have 2 ideas:
|
Can we not handle this in the tracing layer we have for the telemetry? Check when we have printed this message the last time? |
Yes we have a state so that's possible to throttle. But the log will still use eprintln then. To make it print using log when need to log outside the layer. Ok that's option 3. |
Yeah, I understand that we must use |
Or could we push to the layers below us the error message? :D So that they will just print this? :P |
I won't try to use an internal API if that's what you had in mind 😁 But I think the solution with the async task in the background is the most maintainable. I'm not sure what would be the impact on the performance. To clarify my idea:
|
Another case. Seeing my testnet validator keeps synchronising but not finalising. After restart it does finalizing , but at the same time showing this might be it stopped
|
hmm, i still get it as well, saw also that telemetry page seems to be overload in general?
|
The message is now in trace log since 6ac86d5 so you shouldn't see it anymore @2075 To summarize the current status compared to the point raised by @arkpar :
It still send everything it can. Should we do something about this? Like set a maximum amount of messages per minute? This might lead to other issues where too many messages get lost.
The message is not throttled but the message is now in level
This has been fixed.
I could look into it. cc @bkchr |
thanks for the heads up, @cecton, will check again when we will merge substrate into our dev. telemetry itself seems a bit overwhelmed recently, though. |
Hey, is anyone still working on this? Due to the inactivity this issue has been automatically marked as stale. It will be closed if no further activity occurs. Thank you for your contributions. |
When syncing the blockchain with high enough speed (~600 bps) the console is spammed with
Telemetry buffer overflowed!
messages. I see ~50 of these per second. It also appears when runningimport-blocks
command. There's a few things wrong here:log
crate.import-blocks
should have telemetry disabled by default, as well as other network related services.The text was updated successfully, but these errors were encountered: