HI, Great to hear it! From the view of Rally team=) -Best regards, Roman Vasylets
On Sun, Aug 7, 2016 at 10:55 PM, Ricardo Rocha <[email protected]> wrote: > Hi Ton. > > I think we should. Also in cases where multiple volume types are available > (in our case with different iops) there would be additional parameters > required to select the volume type. I'll add it this week. > > It's a detail though, spawning container clusters with Magnum is now super > easy (and fast!). > > Cheers, > Ricardo > > On Fri, Aug 5, 2016 at 5:11 PM, Ton Ngo <[email protected]> wrote: > >> Hi Ricardo, >> For your question 1, you can modify the Heat template to not create the >> Cinder volume and tweak the call to >> configure-docker-storage.sh to use local storage. It should be fairly >> straightforward. You just need to make >> sure the local storage of the flavor is sufficient to host the containers >> in the benchmark. >> If you think this is a common scenario, we can open a blueprint for this >> option. >> Ton, >> >> [image: Inactive hide details for Ricardo Rocha ---08/05/2016 04:51:55 >> AM---Hi. Quick update is 1000 nodes and 7 million reqs/sec :) -]Ricardo >> Rocha ---08/05/2016 04:51:55 AM---Hi. Quick update is 1000 nodes and 7 >> million reqs/sec :) - and the number of >> >> From: Ricardo Rocha <[email protected]> >> To: "OpenStack Development Mailing List (not for usage questions)" < >> [email protected]> >> Date: 08/05/2016 04:51 AM >> >> Subject: Re: [openstack-dev] [magnum] 2 million requests / sec, 100s of >> nodes >> ------------------------------ >> >> >> >> Hi. >> >> Quick update is 1000 nodes and 7 million reqs/sec :) - and the number of >> requests should be higher but we had some internal issues. We have a >> submission for barcelona to provide a lot more details. >> >> But a couple questions came during the exercise: >> >> 1. Do we really need a volume in the VMs? On large clusters this is a >> burden, and local storage only should be enough? >> >> 2. We observe a significant delay (~10min, which is half the total time >> to deploy the cluster) on heat when it seems to be crunching the >> kube_minions nested stacks. Once it's done, it still adds new stacks >> gradually, so it doesn't look like it precomputed all the info in advance >> >> Anyone tried to scale Heat to stacks this size? We end up with a stack >> with: >> * 1000 nested stacks (depth 2) >> * 22000 resources >> * 47008 events >> >> And already changed most of the timeout/retrial values for rpc to get >> this working. >> >> This delay is already visible in clusters of 512 nodes, but 40% of the >> time in 1000 nodes seems like something we could improve. Any hints on Heat >> configuration optimizations for large stacks very welcome. >> >> Cheers, >> Ricardo >> >> On Sun, Jun 19, 2016 at 10:59 PM, Brad Topol <*[email protected]* >> <[email protected]>> wrote: >> >> Thanks Ricardo! This is very exciting progress! >> >> --Brad >> >> >> Brad Topol, Ph.D. >> IBM Distinguished Engineer >> OpenStack >> (919) 543-0646 >> Internet: *[email protected]* <[email protected]> >> Assistant: Kendra Witherspoon (919) 254-0680 >> >> [image: Inactive hide details for Ton Ngo---06/17/2016 12:10:33 >> PM---Thanks Ricardo for sharing the data, this is really encouraging! >> T]Ton >> Ngo---06/17/2016 12:10:33 PM---Thanks Ricardo for sharing the data, this >> is >> really encouraging! Ton, >> >> From: Ton Ngo/Watson/IBM@IBMUS >> To: "OpenStack Development Mailing List \(not for usage questions\)" < >> *[email protected]* >> <[email protected]>> >> Date: 06/17/2016 12:10 PM >> Subject: Re: [openstack-dev] [magnum] 2 million requests / sec, 100s >> of nodes >> >> >> ------------------------------ >> >> >> >> Thanks Ricardo for sharing the data, this is really encouraging! >> Ton, >> >> [image: Inactive hide details for Ricardo Rocha ---06/17/2016 >> 08:16:15 AM---Hi. Just thought the Magnum team would be happy to hear >> :)]Ricardo >> Rocha ---06/17/2016 08:16:15 AM---Hi. Just thought the Magnum team would >> be >> happy to hear :) >> >> From: Ricardo Rocha <*[email protected]* <[email protected]>> >> To: "OpenStack Development Mailing List (not for usage questions)" < >> *[email protected]* >> <[email protected]>> >> Date: 06/17/2016 08:16 AM >> Subject: [openstack-dev] [magnum] 2 million requests / sec, 100s of >> nodes >> ------------------------------ >> >> >> >> Hi. >> >> Just thought the Magnum team would be happy to hear :) >> >> We had access to some hardware the last couple days, and tried some >> tests with Magnum and Kubernetes - following an original blog post >> from the kubernetes team. >> >> Got a 200 node kubernetes bay (800 cores) reaching 2 million requests >> / sec. >> >> Check here for some details: >> >> >> *https://openstack-in-production.blogspot.ch/2016/06/scaling-magnum-and-kubernetes-2-million.html* >> >> <https://openstack-in-production.blogspot.ch/2016/06/scaling-magnum-and-kubernetes-2-million.html> >> >> We'll try bigger in a couple weeks, also using the Rally work from >> Winnie, Ton and Spyros to see where it breaks. Already identified a >> couple issues, will add bugs or push patches for those. If you have >> ideas or suggestions for the next tests let us know. >> >> Magnum is looking pretty good! >> >> Cheers, >> Ricardo >> >> ____________________________________________________________ >> ______________ >> OpenStack Development Mailing List (not for usage questions) >> Unsubscribe: >> *[email protected]?subject:unsubscribe* >> <http://[email protected]?subject:unsubscribe> >> *http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev* >> <http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev> >> >> >> >> ____________________________________________________________ >> ______________ >> OpenStack Development Mailing List (not for usage questions) >> Unsubscribe: >> *[email protected]?subject:unsubscribe* >> <http://[email protected]?subject:unsubscribe> >> *http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev* >> <http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev> >> >> >> >> >> ____________________________________________________________ >> ______________ >> OpenStack Development Mailing List (not for usage questions) >> Unsubscribe: >> *[email protected]?subject:unsubscribe* >> <http://[email protected]?subject:unsubscribe> >> *http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev* >> <http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev> >> >> >> ____________________________________________________________ >> ______________ >> OpenStack Development Mailing List (not for usage questions) >> Unsubscribe: [email protected]?subject:unsubscrib >> e >> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev >> >> >> >> >> ____________________________________________________________ >> ______________ >> OpenStack Development Mailing List (not for usage questions) >> Unsubscribe: [email protected]?subject:unsubscrib >> e >> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev >> >> > > __________________________________________________________________________ > OpenStack Development Mailing List (not for usage questions) > Unsubscribe: [email protected]?subject:unsubscribe > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev > >
__________________________________________________________________________ OpenStack Development Mailing List (not for usage questions) Unsubscribe: [email protected]?subject:unsubscribe http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
