As I wrote earlier, for me it is best to have 9000 on hosts and 8950 on instances. Then I have full speed between instances. With lower mtu on instances I have about 2-2.5 Gbps and I saw that vhost-net process on host is using 100 of 1 cpu core. I'm using libvirt with kvm - maybe You are using something else and it will be different on Your hosts.

Slawek Kaplonski


W dniu 22.01.2015 o 20:45, Pedro Sousa pisze:
Hi Slawek,

I've tried several options but that one that seems to work better is MTU
1450 on VM and MTU 1600 on the host. With MTU 1400 on the VM I would get
freezes and timeouts.

Still I get about 2.2Gbit/Sec while in the host I get 9 Gbit/Sec, do you
think is normal?

Thanks,
Pedro Sousa




On Thu, Jan 22, 2015 at 7:32 PM, Sławek Kapłoński <[email protected]
<mailto:[email protected]>> wrote:

    Hello,

    In dnsmasq file in neutron will be ok. It will then force option 26
    on vm.
    You can also manually change it on vms to tests.

    Slawek Kaplonski

    W dniu 22.01.2015 o 17:06, Pedro Sousa pisze:

        Hi Slawek,

        I'll test this, did you change the mtu on dnsmasq file in
        /etc/neutron/?
        Or do you need to change on other places too?

        Thanks,
        Pedro Sousa

        On Wed, Jan 21, 2015 at 4:26 PM, Sławek Kapłoński
        <[email protected] <mailto:[email protected]>
        <mailto:[email protected] <mailto:[email protected]>>> wrote:

             I have similar and I also got something like 2-2,5Gbps
        between vms.
             When I
             change it to 8950 on vms (so in neutron conf) (50 less then on
             hosts) then it
             is much better.
             You can check that probably when You make test between vms
        on host
             there is
             process called "vhost-net" (or something like that) and it
        uses 100%
             of one cpu
             core and that is imho bottleneck

             Slawek Kaplonski

             On Wed, Jan 21, 2015 at 04:12:02PM +0000, Pedro Sousa wrote:
              > Hi Slawek,
              >
              > I have dhcp-option-force=26,1400 in neutron-dnsmasq.conf and
             MTU=9000 on
              > network-interfaces in the operating system.
              >
              > Do I need to change somewhere else?
              >
              > Thanks,
              > Pedro Sousa
              >
              > On Wed, Jan 21, 2015 at 4:07 PM, Sławek Kapłoński
             <[email protected] <mailto:[email protected]>
        <mailto:[email protected] <mailto:[email protected]>>>

              > wrote:
              >
              > > Hello,
              > >
              > > Try to set bigger jumbo framse on hosts and vms. For
        example on
             hosts You
              > > can
              > > set 9000 and then 8950 and check then. It helps me
        with similar
             problem.
              > >
              > > Slawek Kaplonski
              > >
              > > On Wed, Jan 21, 2015 at 03:22:50PM +0000, Pedro Sousa
        wrote:
              > > > Hi all,
              > > >
              > > > is there a way to improve network performance on my
        instances
             with
              > > VXLAN? I
              > > > changed the MTU on physical interfaces to 1600, still
             performance it's
              > > > lower than in baremetal hosts:
              > > >
              > > > *On Instance:*
              > > >
              > > > [root@vms6-149a71e8-1f2a-4d6e-__bba4-e70dfa42b289
        ~]# iperf3 -s
              > > >
        ------------------------------__-----------------------------
              > > > Server listening on 5201
              > > >
        ------------------------------__-----------------------------
              > > > Accepted connection from 10.0.66.35, port 42900
              > > > [  5] local 10.0.66.38 port 5201 connected to
        10.0.66.35 port
             42901
              > > > [ ID] Interval           Transfer     Bandwidth
              > > > [  5]   0.00-1.00   sec   189 MBytes  1.59 Gbits/sec
              > > > [  5]   1.00-2.00   sec   245 MBytes  2.06 Gbits/sec
              > > > [  5]   2.00-3.00   sec   213 MBytes  1.78 Gbits/sec
              > > > [  5]   3.00-4.00   sec   227 MBytes  1.91 Gbits/sec
              > > > [  5]   4.00-5.00   sec   235 MBytes  1.97 Gbits/sec
              > > > [  5]   5.00-6.00   sec   235 MBytes  1.97 Gbits/sec
              > > > [  5]   6.00-7.00   sec   234 MBytes  1.96 Gbits/sec
              > > > [  5]   7.00-8.00   sec   235 MBytes  1.97 Gbits/sec
              > > > [  5]   8.00-9.00   sec   244 MBytes  2.05 Gbits/sec
              > > > [  5]   9.00-10.00  sec   234 MBytes  1.97 Gbits/sec
              > > > [  5]  10.00-10.04  sec  9.30 MBytes  1.97 Gbits/sec
              > > > - - - - - - - - - - - - - - - - - - - - - - - - -
              > > > [ ID] Interval           Transfer     Bandwidth
          Retr
              > > > [  5]   0.00-10.04  sec  2.25 GBytes  1.92
        Gbits/sec   43
              > >  sender
              > > > [  5]   0.00-10.04  sec  2.25 GBytes  1.92 Gbits/sec
              > > >  receiver
              > > >
              > > >
              > > > *On baremetal:*
              > > > iperf3 -s
              > > > warning: this system does not seem to support IPv6 -
        trying IPv4
              > > >
        ------------------------------__-----------------------------
              > > > Server listening on 5201
              > > >
        ------------------------------__-----------------------------
              > > > Accepted connection from 172.16.21.4, port 51408
              > > > [  5] local 172.16.21.5 port 5201 connected to
        172.16.21.4
             port 51409
              > > > [ ID] Interval           Transfer     Bandwidth
              > > > [  5]   0.00-1.00   sec  1.02 GBytes  8.76 Gbits/sec
              > > > [  5]   1.00-2.00   sec  1.07 GBytes  9.23 Gbits/sec
              > > > [  5]   2.00-3.00   sec  1.08 GBytes  9.29 Gbits/sec
              > > > [  5]   3.00-4.00   sec  1.08 GBytes  9.27 Gbits/sec
              > > > [  5]   4.00-5.00   sec  1.08 GBytes  9.27 Gbits/sec
              > > > [  5]   5.00-6.00   sec  1.08 GBytes  9.28 Gbits/sec
              > > > [  5]   6.00-7.00   sec  1.08 GBytes  9.28 Gbits/sec
              > > > [  5]   7.00-8.00   sec  1.08 GBytes  9.29 Gbits/sec
              > > > [  5]   8.00-9.00   sec  1.08 GBytes  9.28 Gbits/sec
              > > > [  5]   9.00-10.00  sec  1.08 GBytes  9.29 Gbits/sec
              > > > [  5]  10.00-10.04  sec  42.8 MBytes  9.31 Gbits/sec
              > > > - - - - - - - - - - - - - - - - - - - - - - - - -
              > > > [ ID] Interval           Transfer     Bandwidth
          Retr
              > > > [  5]   0.00-10.04  sec  10.8 GBytes  9.23
        Gbits/sec   95
              > >  sender
              > > > [  5]   0.00-10.04  sec  10.8 GBytes  9.22 Gbits/sec
              > > >  receiver
              > > >
              > > >
              > > > Thanks,
              > > > Pedro Sousa
              > >
              > > > _________________________________________________
              > > > OpenStack-operators mailing list
              > > > OpenStack-operators@lists.__openstack.org
        <mailto:[email protected]>
             <mailto:OpenStack-operators@__lists.openstack.org
        <mailto:[email protected]>>
              > > >
        
http://lists.openstack.org/__cgi-bin/mailman/listinfo/__openstack-operators
        
<http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators>
              > >
              > >



    --
    Pozdrawiam
    Sławek Kapłonski
    [email protected] <mailto:[email protected]>



--
Pozdrawiam
Sławek Kapłonski
[email protected]

_______________________________________________
OpenStack-operators mailing list
[email protected]
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators

Reply via email to