Mailing List Archive

Re: [Openstack-operators] UDP Buffer Filling
I know this thread is old but still wanted to post my finding which
may help other folks to understand issue.

I am dealing with same issue in my openstack network, we are media
company and dealing with lots of VoIP applications where we need to
handle high stream of udp packets, Virtio-net isn't meant to handl
high PPS rate, i ran couple of test and found no matter what txqueue
or multiqueue you set it will start dropping packet after 50kpps, I
have tried numa too but result is negative.

Finally i have decided to move and and try SR-IOV and now i am very
very happy, SR-IOV reduce my VM guest CPU load 50% and now my NIC can
handle 200kpps without dropping any packet.

I would say use "iptraf-ng" utility to find out packet rate and see if
its above ~40kpps.
On Fri, Jul 28, 2017 at 8:39 PM Eugene Nikanorov
<enikanorov@mirantis.com> wrote:
>
> John,
>
> multiqueue support will require qemu 2.5+
> I wonder why do you need this feature. It only will help in case of a really huge incoming pps or bandwidth.
> I'm not sure udp packet loss can be solved with this, but of course better try.
>
> my 2c.
>
> Thanks,
> Eugene.
>
> On Fri, Jul 28, 2017 at 5:00 PM, Liping Mao (limao) <limao@cisco.com> wrote:
>>
>> > We already tune these values in the VM. Would you suggest tuning them on the compute nodes as well?
>> No need on compute nodes.(AFAIK)
>>
>>
>> How much pps your vm need to handle?
>> You can monitor CPU usage ,especially si to see where may drop. If you see vhost almost reach to 100% CPU ,multi queue may help in some case.
>>
>> Thanks.
>>
>> Regards,
>> Liping Mao
>>
>> > ? 2017?7?28??22:45?John Petrini <jpetrini@coredial.com> ???
>> >
>> > We already tune these values in the VM. Would you suggest tuning them on the compute nodes as well?
>> _______________________________________________
>> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>> Post to : openstack@lists.openstack.org
>> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>
>
> _______________________________________________
> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
> Post to : openstack@lists.openstack.org
> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack

_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to : openstack@lists.openstack.org
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Re: [Openstack-operators] UDP Buffer Filling [ In reply to ]
Are you using ubutu, TripleO???

Thanks

> Il giorno 27 set 2018, alle ore 20:19, Satish Patel <satish.txt@gmail.com> ha scritto:
>
> I know this thread is old but still wanted to post my finding which
> may help other folks to understand issue.
>
> I am dealing with same issue in my openstack network, we are media
> company and dealing with lots of VoIP applications where we need to
> handle high stream of udp packets, Virtio-net isn't meant to handl
> high PPS rate, i ran couple of test and found no matter what txqueue
> or multiqueue you set it will start dropping packet after 50kpps, I
> have tried numa too but result is negative.
>
> Finally i have decided to move and and try SR-IOV and now i am very
> very happy, SR-IOV reduce my VM guest CPU load 50% and now my NIC can
> handle 200kpps without dropping any packet.
>
> I would say use "iptraf-ng" utility to find out packet rate and see if
> its above ~40kpps.
> On Fri, Jul 28, 2017 at 8:39 PM Eugene Nikanorov
> <enikanorov@mirantis.com> wrote:
>>
>> John,
>>
>> multiqueue support will require qemu 2.5+
>> I wonder why do you need this feature. It only will help in case of a really huge incoming pps or bandwidth.
>> I'm not sure udp packet loss can be solved with this, but of course better try.
>>
>> my 2c.
>>
>> Thanks,
>> Eugene.
>>
>>> On Fri, Jul 28, 2017 at 5:00 PM, Liping Mao (limao) <limao@cisco.com> wrote:
>>>
>>>> We already tune these values in the VM. Would you suggest tuning them on the compute nodes as well?
>>> No need on compute nodes.(AFAIK)
>>>
>>>
>>> How much pps your vm need to handle?
>>> You can monitor CPU usage ,especially si to see where may drop. If you see vhost almost reach to 100% CPU ,multi queue may help in some case.
>>>
>>> Thanks.
>>>
>>> Regards,
>>> Liping Mao
>>>
>>>> ? 2017?7?28??22:45?John Petrini <jpetrini@coredial.com> ???
>>>>
>>>> We already tune these values in the VM. Would you suggest tuning them on the compute nodes as well?
>>> _______________________________________________
>>> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>> Post to : openstack@lists.openstack.org
>>> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>
>>
>> _______________________________________________
>> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>> Post to : openstack@lists.openstack.org
>> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>
> _______________________________________________
> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
> Post to : openstack@lists.openstack.org
> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack


_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to : openstack@lists.openstack.org
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Re: [Openstack-operators] UDP Buffer Filling [ In reply to ]
I'm using openstack-ansible to deploy openstack on centos 7.5 ( we are 100% centos shop )

Sent from my iPhone

> On Sep 28, 2018, at 1:45 AM, Remo Mattei <remo@italy1.com> wrote:
>
> Are you using ubutu, TripleO???
>
> Thanks
>
>> Il giorno 27 set 2018, alle ore 20:19, Satish Patel <satish.txt@gmail.com> ha scritto:
>>
>> I know this thread is old but still wanted to post my finding which
>> may help other folks to understand issue.
>>
>> I am dealing with same issue in my openstack network, we are media
>> company and dealing with lots of VoIP applications where we need to
>> handle high stream of udp packets, Virtio-net isn't meant to handl
>> high PPS rate, i ran couple of test and found no matter what txqueue
>> or multiqueue you set it will start dropping packet after 50kpps, I
>> have tried numa too but result is negative.
>>
>> Finally i have decided to move and and try SR-IOV and now i am very
>> very happy, SR-IOV reduce my VM guest CPU load 50% and now my NIC can
>> handle 200kpps without dropping any packet.
>>
>> I would say use "iptraf-ng" utility to find out packet rate and see if
>> its above ~40kpps.
>> On Fri, Jul 28, 2017 at 8:39 PM Eugene Nikanorov
>> <enikanorov@mirantis.com> wrote:
>>>
>>> John,
>>>
>>> multiqueue support will require qemu 2.5+
>>> I wonder why do you need this feature. It only will help in case of a really huge incoming pps or bandwidth.
>>> I'm not sure udp packet loss can be solved with this, but of course better try.
>>>
>>> my 2c.
>>>
>>> Thanks,
>>> Eugene.
>>>
>>>>> On Fri, Jul 28, 2017 at 5:00 PM, Liping Mao (limao) <limao@cisco.com> wrote:
>>>>>
>>>>> We already tune these values in the VM. Would you suggest tuning them on the compute nodes as well?
>>>> No need on compute nodes.(AFAIK)
>>>>
>>>>
>>>> How much pps your vm need to handle?
>>>> You can monitor CPU usage ,especially si to see where may drop. If you see vhost almost reach to 100% CPU ,multi queue may help in some case.
>>>>
>>>> Thanks.
>>>>
>>>> Regards,
>>>> Liping Mao
>>>>
>>>>> ? 2017?7?28??22:45?John Petrini <jpetrini@coredial.com> ???
>>>>>
>>>>> We already tune these values in the VM. Would you suggest tuning them on the compute nodes as well?
>>>> _______________________________________________
>>>> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>> Post to : openstack@lists.openstack.org
>>>> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>
>>>
>>> _______________________________________________
>>> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>> Post to : openstack@lists.openstack.org
>>> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>
>> _______________________________________________
>> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>> Post to : openstack@lists.openstack.org
>> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>

_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to : openstack@lists.openstack.org
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack