-
Notifications
You must be signed in to change notification settings - Fork 55
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Issue when processing data through logstash #22
Comments
Seems to be a time related and allowing for both the DB and cache to build up their relevant data. |
So as you can see below we are running at 95% established and only 6.9% RDAP lookup. Appreciate this is "only" a 15 min delay, however I was planning on using this domain create data to alert and potentially even sink-hole requests. |
Hi @The1WhoPrtNocks |
Email sent |
Hi Mark, Thank you for your time over the past few days. So it seems that domain_stats tops out at ~80 e/s (events per second) being sent, sometimes spiking to 100-120 e/s. |
Reopening while we try to fix this speed issue. Here are some notes: I've been publishing some updates to try and improve the speed. @The1WhoPrtNocks was initially consistently seen 80 e/s being processed. I disabled session logging and added an cache to a slow function (reduce_domain) and saw an improvement to 500 e/s with burst up to 1k. I've created another multiprocessing for to support multiple cores and a distributed cache across processors. @The1WhoPrtNocks has graciously agreed to test this alpha version of the code. We are working in the multiprocessed branch |
There is a drastic improvement running the multiprocessed branch. To clear a large back log it started and stayed pretty steady at 1k e/s . I will leave this to run for a while to update the database + cache, then conduct another test of clearing a large back log of events. Seems like the 5-8 min to grow to 1.8k e/s was where it topped out naturally, on second stress test it sent straight to 1.8 e/s and stayed there. |
Ok. I'm calling that a win. 1.8k is amazingly close to freq's 2k considering all of the network latency involved for those RDAP queries. I cleaned up the code a bit and added some features. Id appreaciate your help getting this one tested and I will kill the old version in favor of the multiprocessing. Here are the instructions for installing the multiprocessing version.. https://github.com/MarkBaggett/domain_stats/blob/multiprocessed/README.md New command "domain-stats-settings" lets you configure/reconfigure the daemons. You can change settings like the number of days for "established" tag. You can also enable freq_scores and set alert thresholds for each of the values. Set value for ip_address. Default=0.0.0.0 Current=0.0.0.0 (Enter to keep Current): The other new comaind "domain-stats" will launch the server. They both require you pass the path to a directory where the configs and database can be stored. |
Hi,
I am trying to send my networks data to both domain_stats and freq_server for analysis and enrichment.
I am sending on average 20 queries per second.
When this data is sent just to freq_server the VM idles and process the data like a dream.
However as soon as I also send that same data to domain_stats the ingesting of the data post processing by ELK drops of a cliff.
We have investigated CPU, memory and Disk IO and is all seems reasonable on both the VM running domain_stats and ELK.
The memory utilization for the caching is consistently under utilized based on sending the stats query.
Looking at the local logging for both domain_stats and log-stash they seem to be up-to-date.
Has anyone else come across this issue, i have tried multiple configs rdap vs ISC ?
The text was updated successfully, but these errors were encountered: