[Oisf-users] Inconsistent packet dropped behaviour with the same config on several nodes

Magmi A magmi.sec at gmail.com
Thu Sep 13 06:48:45 UTC 2018


Hi Peter,

Yes, of course.
Here you have stats.log for Node2:

Date:9/13/2018--06:37:42(uptime:6d,17h05m59s)
------------------------------------------------------------------------------------
Counter |TMName |Value
------------------------------------------------------------------------------------
capture.kernel_packets |Total |135374454
capture.kernel_drops |Total |11430384
decoder.pkts |Total |123946335
decoder.bytes |Total |122943123694
decoder.ipv4 |Total |123767936
decoder.ipv6 |Total |5036
decoder.ethernet |Total |123946335
decoder.tcp |Total |120082602
decoder.udp |Total |683567
decoder.icmpv4 |Total |114883
decoder.icmpv6 |Total |189
decoder.tered |Total |5
decoder.avg_pkt_size |Total |991
decoder.max_pkt_size |Total |1514
flow.tcp |Total |780742
flow.udp |Total |305951
flow.icmpv6 |Total |162
tcp.sessions |Total |727356
tcp.syn |Total |771112
tcp.synack |Total |720764
tcp.rst |Total |549359
tcp.stream_depth_reached |Total |454
tcp.reassembly_gap |Total |7722
tcp.overlap |Total |483624
detect.alert |Total |4
app_layer.flow.http |Total |108080
app_layer.tx.http |Total |262748
app_layer.flow.smtp |Total |7
app_layer.tx.smtp |Total |7
app_layer.flow.tls |Total |6612
app_layer.flow.ssh |Total |13
app_layer.flow.smb |Total |55361
app_layer.flow.dcerpc_tcp |Total |202204
app_layer.flow.dns_tcp |Total |10
app_layer.tx.dns_tcp |Total |10
app_layer.flow.failed_tcp |Total |274419
app_layer.flow.dcerpc_udp |Total |4
app_layer.flow.dns_udp |Total |139325
app_layer.tx.dns_udp |Total |239577
app_layer.flow.failed_udp |Total |166622
flow_mgr.closed_pruned |Total |684943
flow_mgr.new_pruned |Total |290500
flow_mgr.est_pruned |Total |110950
flow.spare |Total |10000
flow.tcp_reuse |Total |6055
flow_mgr.flows_checked |Total |12
flow_mgr.flows_notimeout |Total |5
flow_mgr.flows_timeout |Total |7
flow_mgr.flows_timeout_inuse |Total |1
flow_mgr.flows_removed |Total |6
flow_mgr.rows_checked |Total |65536
flow_mgr.rows_skipped |Total |65522
flow_mgr.rows_empty |Total |2
flow_mgr.rows_maxlen |Total |1
tcp.memuse |Total |4587520
tcp.reassembly_memuse |Total |6650304
dns.memuse |Total |16386
http.memuse |Total |12142178
flow.memuse |Total |7207360

And just for a reference stats.log for Node1:

Date:9/13/2018--06:35:39(uptime:6d,16h39m54s)
------------------------------------------------------------------------------------
Counter |TMName |Value
------------------------------------------------------------------------------------
capture.kernel_packets |Total |3577965800
capture.kernel_drops |Total |3416155
decoder.pkts |Total |3574589712
decoder.bytes |Total |3536139875210
decoder.invalid |Total |10070
decoder.ipv4 |Total |3571132083
decoder.ipv6 |Total |143756
decoder.ethernet |Total |3574589712
decoder.tcp |Total |3522243739
decoder.udp |Total |34827953
decoder.icmpv4 |Total |963831
decoder.icmpv6 |Total |33551
decoder.teredo |Total |1399
decoder.avg_pkt_size |Total |989
decoder.max_pkt_size |Total |1534
flow.tcp |Total |1144524
flow.udp |Total |202960
flow.icmpv6 |Total |439
decoder.ipv4.trunc_pkt |Total |10070
tcp.sessions |Total |341278
tcp.ssn_memcap_drop |Total |4446979
tcp.pseudo |Total |84
tcp.invalid_checksum |Total |4
tcp.syn |Total |6653717
tcp.synack |Total |2572744
tcp.rst |Total |1857715
tcp.segment_memcap_drop |Total |10
tcp.stream_depth_reached |Total |303
tcp.reassembly_gap |Total |95648
tcp.overlap |Total |3889304
tcp.insert_data_normal_fail |Total |3314483
detect.alert |Total |518
app_layer.flow.http |Total |34820
app_layer.tx.http |Total |60759
app_layer.flow.ftp |Total |20
app_layer.flow.smtp |Total |140
app_layer.tx.smtp |Total |177
app_layer.flow.tls |Total |43356
app_layer.flow.smb |Total |3430
app_layer.flow.dcerpc_tcp |Total |8894
app_layer.flow.dns_tcp |Total |48
app_layer.tx.dns_tcp |Total |46
app_layer.flow.failed_tcp |Total |107518
app_layer.flow.dcerpc_udp |Total |5
app_layer.flow.dns_udp |Total |114888
app_layer.tx.dns_udp |Total |482904
app_layer.flow.failed_udp |Total |88067
flow_mgr.closed_pruned |Total |259368
flow_mgr.new_pruned |Total |981024
flow_mgr.est_pruned |Total |107531
flow.spare |Total |10000
flow.tcp_reuse |Total |29932
flow_mgr.rows_checked |Total |65536
flow_mgr.rows_skipped |Total |65536
tcp.memuse |Total |4587520
tcp.reassembly_memuse |Total |655360
dns.memcap_global |Total |1086836
flow.memuse |Total |7074304

Thank you for any suggestions.

Best,
magmi

On Wed, 12 Sep 2018 at 16:16, Peter Manev <petermanev at gmail.com> wrote:

> On Wed, Sep 12, 2018 at 10:57 AM Magmi A <magmi.sec at gmail.com> wrote:
> >
> >
> >> > * Node1 receives ~ 500Mbps of traffic (it's 1Gbps interface), and
> gets in average 1-2% kernel packet dropped
> >> > while
> >> > * Node2 receives ~ 500kbps of traffic and gets in average 10% kernel
> packet dropped
> >>
> >> What is different between node 1 and node 2 ? (same config/same
> >> suricata/same HW/same rules...?)]
> >
> >
> > The nodes have the same HW, run the same config/ suricata version, have
> the same set of rules.
> >
> > The only difference is that they are exposed to different sources of
> traffic.
> > From Wireshark analysis the protocol hierarchies for both cases seem
> similar - there is no spectacular difference.
> >
> > So really the only difference is the captured traffic itself (MACs, IPs,
> partly protocols, data etc).
> >
> > That is why we have such a problem how to approach the problem and
> troubleshoot it.
>
>
> Can you share full update of the latest stats.log ?
>
>
>
> --
> Regards,
> Peter Manev
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openinfosecfoundation.org/pipermail/oisf-users/attachments/20180913/5c863ed3/attachment-0001.html>


More information about the Oisf-users mailing list