[Oisf-users] Packet Loss

Peter Manev petermanev at gmail.com
Mon Jun 9 14:30:25 UTC 2014


On Mon, Jun 9, 2014 at 4:21 PM, Yasha Zislin <coolyasha at hotmail.com> wrote:
> So these two interfaces are kind of related. If traffic spikes on one, it
> will increase on the other.
> In fact, these two span ports see traffic before and after the firewall. So
> traffic is the same except for NATting piece. After the firewall IPs change
> to private addressing.
> Not sure if this scenario can have issues, BUT my packet drop was occurring
> when I was just trying to get Suricata to run with one interface.
> Important point is that CPUs dont hit 100% (it was doing that in the
> beginning until I started to offload everything to RAM).
>
> You are correct about memory consumption of threads. I've increased
> min_num_slots by running this:
> modprobe pf_ring transparent_mode=0 min_num_slots=400000 enable_tx_capture=0
>
> cat /proc/net/pf_ring/info
> PF_RING Version          : 6.0.2 ($Revision: exported$)
> Total rings              : 16
>
> Standard (non DNA) Options
> Ring slots               : 400000
> Slot version             : 15
> Capture TX               : No [RX only]
> IP Defragment            : No
> Socket Mode              : Standard
> Transparent mode         : Yes [mode 0]
> Total plugins            : 0
> Cluster Fragment Queue   : 3852
> Cluster Fragment Discard : 428212
>
> I could have increased it higher (it let me do that) but Free Num Slots
> stopped increased after 400000 value.
> I did notice if I set min_num_slots to default 65k number, Free Num Slots
> get to 0 faster and packet drop begins.
>
> Here is a stat for one of the threads:
>  cat /proc/net/pf_ring/6224-eth17.595
> Bound Device(s)    : eth17
> Active             : 1
> Breed              : Non-DNA
> Sampling Rate      : 1
> Capture Direction  : RX+TX
> Socket Mode        : RX+TX
> Appl. Name         : Suricata
> IP Defragment      : No
> BPF Filtering      : Disabled
> # Sw Filt. Rules   : 0
> # Hw Filt. Rules   : 0
> Poll Pkt Watermark : 128
> Num Poll Calls     : 6408432
> Channel Id Mask    : 0xFFFFFFFF
> Cluster Id         : 99
> Slot Version       : 15 [6.0.2]
> Min Num Slots      : 688290
> Bucket Len         : 1522
> Slot Len           : 1560 [bucket+header]
> Tot Memory         : 1073741824
> Tot Packets        : 902405618
> Tot Pkt Lost       : 79757335
> Tot Insert         : 822648289
> Tot Read           : 822648236
> Insert Offset      : 219035272
> Remove Offset      : 218997656
> TX: Send Ok        : 0
> TX: Send Errors    : 0
> Reflect: Fwd Ok    : 0
> Reflect: Fwd Errors: 0
> Num Free Slots     : 688237
>
>
> For NICs I have 10 gig Fiber HP nic (I think with Qlogic chip).
>
> BTW, I had to configure both PF_RING interfaces to be the same cluster ID.
> For some reason, setting them to different numbers would not work.
>
> You are correct about the behavior. It runs fine, Free Num Slots for ALL
> threads get to 0, packet drop starts, after some time Free Num Slots go back
> to almost 100% available and packet drop stops.
> It feels like it is getting choked on something and starts to fill up.
>
> Timeout values are as follows:
> flow-timeouts:
>
>   default:
>     new: 3
>     established: 30
>     closed: 0
>     emergency-new: 10
>     emergency-established: 10
>     emergency-closed: 0
>   tcp:
>     new: 6
>     established: 100
>     closed: 12
>     emergency-new: 1
>     emergency-established: 5
>     emergency-closed: 2
>   udp:
>     new: 3
>     established: 30
>     emergency-new: 3
>     emergency-established: 10
>   icmp:
>     new: 3
>     established: 30
>     emergency-new: 1
>     emergency-established: 10
>
> I set them just like one of these "10gig and beyond" articles said.
>
> Thank you for your help.
>

Thank you for the feedback.
Please keep the conversation on the list :)


-- 
Regards,
Peter Manev



More information about the Oisf-users mailing list