Nice suggestion! Old fashioned UBC is real nightmare and was deprecated. On Fri, Nov 28, 2014 at 10:35 AM, CoolCold <coolthec...@gmail.com> wrote: > Hello! > I'd recommend set only ram/swap limits (via --ram/--swap) , letting other > settings be mostly unlimited (while ram/swap limits are not overflowed of > course) - http://wiki.openvz.org/VSwap > > On Fri, Nov 28, 2014 at 3:31 AM, Nipun Arora <nipunarora2...@gmail.com> > wrote: >> >> Nevermind, I figured it out by changing the fail counters in >> /proc/user_beans >> >> Thanks >> Nipun >> >> On Thu, Nov 27, 2014 at 7:14 PM, Nipun Arora <nipunarora2...@gmail.com> >> wrote: >>> >>> Thanks, the speed is improved by an order of magnitude :) >>> >>> btw. is there any benchmark, that you all have looked into for testing >>> how good/practical live migration is for real-world systems? >>> Additionally, I'm trying to run a java application(dacapo benchmark), but >>> keep having trouble in getting java to run.. >>> >>> java -version >>> >>> Error occurred during initialization of VM >>> >>> Could not reserve enough space for object heap >>> >>> Could not create the Java virtual machine. >>> >>> >>> I've put my vz conf file below, can anyone suggest what could be the >>> problem? >>> >>> Thanks >>> Nipun >>> >>> # UBC parameters (in form of barrier:limit) >>> >>> KMEMSIZE="14372700:14790164" >>> >>> LOCKEDPAGES="2048:2048" >>> >>> PRIVVMPAGES="65536:69632" >>> >>> SHMPAGES="21504:21504" >>> >>> NUMPROC="240:240" >>> >>> PHYSPAGES="0:131072" >>> >>> VMGUARPAGES="33792:unlimited" >>> >>> OOMGUARPAGES="26112:unlimited" >>> >>> NUMTCPSOCK="360:360" >>> >>> NUMFLOCK="188:206" >>> >>> NUMPTY="16:16" >>> >>> NUMSIGINFO="256:256" >>> >>> TCPSNDBUF="1720320:2703360" >>> >>> TCPRCVBUF="1720320:2703360" >>> >>> OTHERSOCKBUF="1126080:2097152" >>> >>> DGRAMRCVBUF="262144:262144" >>> >>> NUMOTHERSOCK="1200" >>> >>> DCACHESIZE="3409920:3624960" >>> >>> NUMFILE="9312:9312" >>> >>> AVNUMPROC="180:180" >>> >>> NUMIPTENT="128:128" >>> >>> >>> # Disk quota parameters (in form of softlimit:hardlimit) >>> >>> DISKSPACE="3145728:3145728" >>> >>> DISKINODES="131072:144179" >>> >>> QUOTATIME="0" >>> >>> >>> # CPU fair scheduler parameter >>> >>> CPUUNITS="1000" >>> >>> >>> NETFILTER="stateless" >>> >>> VE_ROOT="/vz/root/101" >>> >>> VE_PRIVATE="/vz/private/101" >>> >>> OSTEMPLATE="centos-6-x86_64" >>> >>> ORIGIN_SAMPLE="basic" >>> >>> HOSTNAME="test" >>> >>> IP_ADDRESS="192.168.1.101" >>> >>> NAMESERVER="8.8.8.8 8.8.4.4" >>> >>> CPULIMIT="25" >>> >>> SWAPPAGES="0:262144" >>> >>> >>> >>> On Mon, Nov 24, 2014 at 12:16 PM, Kir Kolyshkin <k...@openvz.org> wrote: >>>> >>>> >>>> On 11/23/2014 07:13 PM, Nipun Arora wrote: >>>> >>>> Thanks, I will try your suggestions, and get back to you. >>>> btw... any idea what could be used to share the base image on both >>>> containers? >>>> Like hardlink it in what way? Once both containers start, won't they >>>> have to write to different locations? >>>> >>>> >>>> ploop is composed as a set of stacked images, with all of them but the >>>> top one being read-only. >>>> >>>> >>>> I understand that some file systems have a copy on write mechanism, >>>> where after a snapshot all future writes are written to a additional linked >>>> disks. >>>> Does ploop operate in a similar way? >>>> >>>> >>>> yes >>>> >>>> >>>> http://wiki.qemu.org/Features/Snapshots >>>> >>>> >>>> http://openvz.livejournal.com/44508.html >>>> >>>> >>>> >>>> The cloning with a modified vzmigrate script helps. >>>> >>>> - Nipun >>>> >>>> On Sun, Nov 23, 2014 at 5:29 PM, Kir Kolyshkin <k...@openvz.org> wrote: >>>>> >>>>> >>>>> On 11/23/2014 04:59 AM, Nipun Arora wrote: >>>>> >>>>> Hi Kir, >>>>> >>>>> Thanks for the response, I'll update it, and tell you about the >>>>> results. >>>>> >>>>> 1. A follow up question... I found that the write I/O speed of >>>>> 500-1Mbps increased the suspend time to several minutes.(mostly pcopy >>>>> stage) >>>>> This seems extremely high for a relatively low I/O workload, which is >>>>> why I was wondering if there are any special things I need to take care >>>>> of. >>>>> (I ran fio (flexible i/o writer) with fixed throughput while doing live >>>>> migration) >>>>> >>>>> >>>>> Please retry with vzctl 4.8 and ploop 1.12.1 (make sure they are on >>>>> both sides). >>>>> There was a 5 second wait for the remote side to finish syncing >>>>> copied ploop data. It helped a case with not much I/O activity in >>>>> container, but >>>>> ruined the case you are talking about. >>>>> >>>>> Newer ploop and vzctl implement a feedback channel for ploop copy that >>>>> eliminates >>>>> that wait time. >>>>> >>>>> http://git.openvz.org/?p=ploop;a=commit;h=20d754c91079165b >>>>> http://git.openvz.org/?p=vzctl;a=commit;h=374b759dec45255d4 >>>>> >>>>> There are some other major improvements as well, such as async send for >>>>> ploop. >>>>> >>>>> http://git.openvz.org/?p=ploop;a=commit;h=a55e26e9606e0b >>>>> >>>>> >>>>> 2. For my purposes, I have modified the live migration script to allow >>>>> me to do cloning... i.e. I start both the containers instead of deleting >>>>> the >>>>> original. I need to do this "cloning" from time to time for the same >>>>> target >>>>> container... >>>>> >>>>> a. Which means that lets say we cloned container C1 to container >>>>> C2, and let both execute at time t0, this works with no apparent loss of >>>>> service. >>>>> >>>>> b. Now at time t1 I would like to again clone C1 to C2, and >>>>> would like to optimize the rsync process as most of the ploop file for C1 >>>>> and C2 should still be the same (i.e. less time to sync). Can anyone >>>>> suggest >>>>> what would be the best way to realize the second point? >>>>> >>>>> >>>>> You can create a ploop snapshot and use shared base image for both >>>>> containers >>>>> (instead of copying the base delta, hardlink it). This is not supported >>>>> by tools >>>>> (for example, since base delta is now shared you can't merge down to >>>>> it, but the >>>>> tools are not aware) so you need to figure it out by yourself and be >>>>> accurate >>>>> but it should work. >>>>> >>>>> >>>>> >>>>> >>>>> Thanks >>>>> Nipun >>>>> >>>>> On Sun, Nov 23, 2014 at 12:56 AM, Kir Kolyshkin <k...@openvz.org> wrote: >>>>>> >>>>>> >>>>>> On 11/22/2014 09:09 AM, Nipun Arora wrote: >>>>>> >>>>>> Hi All, >>>>>> >>>>>> I was wondering if anyone can suggest what is the most optimal way to >>>>>> do the following >>>>>> >>>>>> 1. Can anyone clarify if ploop is the best layout for minimum suspend >>>>>> time during live migration? >>>>>> >>>>>> >>>>>> Yes (due to ploop copy which only copies the modified blocks). >>>>>> >>>>>> >>>>>> 2. I tried migrating a ploop device where I increased the --diskspace >>>>>> to 5G, >>>>>> and found that the suspend time taken by live migration increased to >>>>>> 57 seconds >>>>>> (mainly undump and restore increased)... >>>>>> whereas a 2G diskspace was taking 2-3 seconds suspend time... Is this >>>>>> expected? >>>>>> >>>>>> >>>>>> No. Undump and restore times depends mostly on amount of RAM used by a >>>>>> container. >>>>>> >>>>>> Having said that, live migration stages influence each other, although >>>>>> it's less so >>>>>> in the latest vzctl release (I won't go into details here if you allow >>>>>> me -- just make sure >>>>>> you test with vzctl 4.8). >>>>>> >>>>>> >>>>>> 3. I tried running a write intensive workload, and found that beyond >>>>>> 100-150Kbps, >>>>>> the suspend time during live migration rapidly increased? Is this an >>>>>> expected trend? >>>>>> >>>>>> >>>>>> Sure. With increased writing speed, the amount of data that needs to >>>>>> be copied after CT >>>>>> is suspended increases. >>>>>> >>>>>> >>>>>> I am using vzctl 4.7, and ploop 1.11 in centos 6.5 >>>>>> >>>>>> >>>>>> You need to update vzctl and ploop and rerun your tests, there should >>>>>> be >>>>>> some improvement (in particular with respect to issue #3). >>>>>> >>>>>> >>>>>> Thanks >>>>>> Nipun >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Users mailing list >>>>>> Users@openvz.org >>>>>> https://lists.openvz.org/mailman/listinfo/users >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Users mailing list >>>>>> Users@openvz.org >>>>>> https://lists.openvz.org/mailman/listinfo/users >>>>>> >>>>> >>>>> >>>>> >>>>> _______________________________________________ >>>>> Users mailing list >>>>> Users@openvz.org >>>>> https://lists.openvz.org/mailman/listinfo/users >>>>> >>>>> >>>>> >>>>> _______________________________________________ >>>>> Users mailing list >>>>> Users@openvz.org >>>>> https://lists.openvz.org/mailman/listinfo/users >>>>> >>>> >>>> >>>> >>>> _______________________________________________ >>>> Users mailing list >>>> Users@openvz.org >>>> https://lists.openvz.org/mailman/listinfo/users >>>> >>>> >>>> >>>> _______________________________________________ >>>> Users mailing list >>>> Users@openvz.org >>>> https://lists.openvz.org/mailman/listinfo/users >>>> >>> >> >> >> _______________________________________________ >> Users mailing list >> Users@openvz.org >> https://lists.openvz.org/mailman/listinfo/users >> > > > > -- > Best regards, > [COOLCOLD-RIPN] > > _______________________________________________ > Users mailing list > Users@openvz.org > https://lists.openvz.org/mailman/listinfo/users >
-- Sincerely yours, Pavel Odintsov _______________________________________________ Users mailing list Users@openvz.org https://lists.openvz.org/mailman/listinfo/users