[Oisf-users] What does it means??

Victor Julien lists at inliniac.net
Wed Oct 9 12:39:09 UTC 2013


On 10/09/2013 02:36 PM, C. L. Martinez wrote:
> On Wed, Oct 9, 2013 at 12:31 PM, Victor Julien <lists at inliniac.net> wrote:
>> On 10/09/2013 02:28 PM, C. L. Martinez wrote:
>>> Hi all,
>>>
>>>  Recently, I have installed a FreeBSD 9.2 host with suricata 1.4.6 and
>>> returns me a lot of packets dropped by kernel:
>>>
>>> For example after 2 minutes up:
>>>
>>> Date: 10/9/2013 -- 12:19:50 (uptime: 0d, 00h 02m 58s)
>>> -------------------------------------------------------------------
>>> Counter                   | TM Name                   | Value
>>> -------------------------------------------------------------------
>>> capture.kernel_packets    | RxPcapem41                | 3137698
>>> capture.kernel_drops      | RxPcapem41                | 2415508
>>> capture.kernel_ifdrops    | RxPcapem41                | 0
>>>
>>> But tcp.ssn_memcap_drop and tcp.reassembly_gap:
>>>
>>> decoder.avg_pkt_size      | RxPcapem42                | 828
>>> decoder.max_pkt_size      | RxPcapem42                | 1514
>>> defrag.ipv4.fragments     | RxPcapem42                | 90
>>> defrag.ipv4.reassembled   | RxPcapem42                | 25
>>> defrag.ipv4.timeouts      | RxPcapem42                | 0
>>> defrag.ipv6.fragments     | RxPcapem42                | 0
>>> defrag.ipv6.reassembled   | RxPcapem42                | 0
>>> defrag.ipv6.timeouts      | RxPcapem42                | 0
>>> defrag.max_frag_hits      | RxPcapem42                | 0
>>> tcp.sessions              | RxPcapem42                | 308
>>> tcp.ssn_memcap_drop       | RxPcapem42                | 0
>>> tcp.pseudo                | RxPcapem42                | 23
>>> tcp.invalid_checksum      | RxPcapem42                | 0
>>> tcp.no_flow               | RxPcapem42                | 0
>>> tcp.reused_ssn            | RxPcapem42                | 0
>>> tcp.memuse                | RxPcapem42                | 6029312
>>> tcp.syn                   | RxPcapem42                | 1261
>>> tcp.synack                | RxPcapem42                | 702
>>> tcp.rst                   | RxPcapem42                | 565
>>> tcp.segment_memcap_drop   | RxPcapem42                | 0
>>> tcp.stream_depth_reached  | RxPcapem42                | 0
>>> tcp.reassembly_memuse     | RxPcapem42                | 11327048
>>> tcp.reassembly_gap        | RxPcapem42                | 23
>>
>> tcp.ssn_memcap_drop and tcp.reassembly_gap only related to memcaps, not
>> to packet loss.
>>
>>> I think the problem is with interrupts:
>>>
>>> interrupt                          total       rate
>>> irq1: atkbd0                           6          0
>>> irq10: em2 em3                   2320880       3453
>>> irq11: em0 em1 em4+              1256951       1870
>>> cpu0:timer                        148773        221
>>> cpu1:timer                        148310        220
>>> Total                            3877066       5769
>>
>> Not sure.
>>
>> What runmode are you using? Also, whats your max-pending-packets setting?
>>
> 
> I use runmode workers and for max-pending-packets 12288 ...

By using workers with multiple interfaces you get just one thread per
interface. There is no flow based load balancing in plain libpcap, so I
think runmode autofp may get you better results as then Suricata can use
more threads per interface.

-- 
---------------------------------------------
Victor Julien
http://www.inliniac.net/
PGP: http://www.inliniac.net/victorjulien.asc
---------------------------------------------




More information about the Oisf-users mailing list