Hi, Jake Thanks for details. Also Cyrill asked about logs for P.Haul service. Unfortunately prlsrvctl report doesn't contain it. Could you provide it too?
And does plain vcztl suspend/resume on this container works fine? On 13:52 Sat 14 May , jjs - mainphrame wrote: > Hi Sergey, > > prlsrvctl report id: 99722287 > > Environment details (identical software stack on the 2 hosts): > virtuozzo-release-7.0.0-39.vz7.x86_64 > CentOS Linux release 7.2.1511 (Core) > > Just to reiterate - as of the previous update, live migration began > working properly; it then broke with the next day's update > > I've attached logs from both OVZ hosts (hachi = source, annie = target) > > Preparing to submit bug report now - > > Jake > > On Fri, May 13, 2016 at 2:13 AM, Sergey Bronnikov <serg...@openvz.org> wrote: > > Hi > > > > Thank you for report. > > > > Could submit a new bug to the bugs.openvz.org? > > and provide more details about your environment: > > - Vz7 version (rpm -qa virtuozzo-release) > > - problem report id (prlsrvctl problem-report --send) > > > > Sergey > > > > On 19:42 Thu 12 May , jjs - mainphrame wrote: > >> Aw shucks, I've spoken too soon. After today's updates, live migration > >> is back to not working: > >> > >> [root@hachi ~]# time vzmigrate annie 888 --online > >> Connection to destination node (annie) is successfully established > >> Moving/copying CT 888 -> CT 888, [], [] ... > >> locking 888 > >> Checking bindmounts > >> Check cluster ID > >> Checking keep dir for private area copy > >> Check of requires kernel modules > >> Checking technologies > >> Checking templates for CT > >> Checking IP addresses on destination node > >> Checking RATE parameters in config > >> Checking ploop format 2 > >> copy CT private /vz/private/888 > >> Live migration stage started > >> Phaul service failed to live migrate CT > >> Can't move/copy CT 888 -> CT 888, [], [] : Phaul service failed to > >> live migrate CT > >> > >> real 2m51.840s > >> user 0m0.385s > >> sys 0m3.457s > >> [root@hachi ~]# > >> > >> Jake > >> > >> On Wed, May 11, 2016 at 9:28 PM, jjs - mainphrame <j...@mainphrame.com> > >> wrote: > >> > Thank you Sergey - > >> > > >> > As of today's updates, I find that live container migration is working > >> > reliably and well here. > >> > > >> > Great work! > >> > > >> > Jake > >> > > >> > On Thu, Apr 28, 2016 at 8:29 AM, Sergey Bronnikov <serg...@openvz.org> > >> > wrote: > >> >> Hi everyone, > >> >> > >> >> As usual, I'd like to share with you the latest Virtuozzo 7 development > >> >> news. > >> >> > >> >> First and foremost, more features have become available in Virtuozzo 7: > >> >> > >> >> - Containers migration via vzmigrate in Vz7 [1] > >> >> - IP and MAC filtering for VMs [2] > >> >> - VM data stats supported switch to native KVM ones [3] > >> >> - AutoFS migration [4] > >> >> > >> >> All these features are available in the latests Virtuozzo 7 builds > >> >> https://download.openvz.org/virtuozzo/factory/x86_64/iso/ > >> >> > >> >> Also we have published new portions of Virtuozzo source code: > >> >> > >> >> - EZ templates: > >> >> * Ubuntu 15.10 x86_64 > >> >> * Fedora 23 x86_64 > >> >> * Debian 7.0 x86_64 > >> >> * VzLinux 6 x86_64 > >> >> * VzLinux 7 x86_64 > >> >> - DRS module for rmond [5] > >> >> > >> >> Events > >> >> ====== > >> >> > >> >> Earlier we participated in Linux Storage, Filesystem and Memory > >> >> Management > >> >> Conference. See notes from the conference [6] if you interested. > >> >> > >> >> Upcoming talks and conferences: > >> >> > >> >> - Ross Boucher will talk about running services cloning with Docker and > >> >> CRIU at > >> >> DockerCon 2016 [7] > >> >> - Sanidhya Kashyap will present an instant OS updater using CRIU at > >> >> USENIX > >> >> ATC'16 [8] > >> >> - Checkpoint-Restore micro-conference will be held 2-4 Nov at Santa-Fe, > >> >> NM at > >> >> Linux Plumbers 2016. The proposal for the conferense was accepted. [9] > >> >> > >> >> Links > >> >> ===== > >> >> > >> >> [1] https://lists.openvz.org/pipermail/users/2016-March/006833.html > >> >> [2] https://lists.openvz.org/pipermail/users/2016-April/006845.html > >> >> [3] https://lists.openvz.org/pipermail/users/2016-April/006850.html > >> >> [4] https://lists.openvz.org/pipermail/users/2016-April/006861.html > >> >> [5] https://github.com/OpenVZ/rmond-drs > >> >> [6] http://lwn.net/Articles/lsfmm2016/ > >> >> [7] https://blog.docker.com/2016/04/black-belt-talks-dockercon-2016/ > >> >> [8] > >> >> https://www.usenix.org/conference/atc16/technical-sessions/presentation/kashyap > >> >> [9] > >> >> https://www.linuxplumbersconf.org/2016/checkpoint-restore-microconference-accepted-into-2016-linux-plumbers-conference/ > >> >> > >> >> Sincerely, > >> >> Sergey > >> >> _______________________________________________ > >> >> Users mailing list > >> >> Users@openvz.org > >> >> https://lists.openvz.org/mailman/listinfo/users > >> _______________________________________________ > >> Users mailing list > >> Users@openvz.org > >> https://lists.openvz.org/mailman/listinfo/users > > > > -- > > sergeyb@ > [root@hachi ~]# time vzmigrate annie 888 --online > Connection to destination node (annie) is successfully established > Moving/copying CT 888 -> CT 888, [], [] ... > locking 888 > Checking bindmounts > Check cluster ID > Checking keep dir for private area copy > Check of requires kernel modules > Checking technologies > Checking templates for CT > Checking IP addresses on destination node > Checking RATE parameters in config > Checking ploop format 2 > copy CT private /vz/private/888 > Live migration stage started > Phaul service failed to live migrate CT > Can't move/copy CT 888 -> CT 888, [], [] : Phaul service failed to live > migrate CT > > real 2m54.298s > user 0m0.438s > sys 0m3.456s > [root@hachi ~]# > > > 20:26:06.144: 52839: > 20:26:06.144: 52839: > 20:26:06.144: 52839: > 20:26:06.144: 52839: Starting p.haul > 20:26:06.144: 52839: Use existing connections, fdrpc=9 fdmem=11 > fdfs=root.hdd/root.hds:13 > 20:26:06.335: 52839: Setting up local > 20:26:06.335: 52839: Loading config file from /etc/vz/conf/ > 20:26:06.337: 52839: Initialize ploop hauler > 20:26:06.337: 52839: `- /vz/private/888/root.hdd/root.hds > 20:26:06.338: 52839: Passing (ctl:10, data:8) pair to CRIU > 20:26:06.340: 52839: Setting up remote > 20:26:06.371: 52839: Start migration in live mode > 20:26:06.371: 52839: Checking CPU compatibility > 20:26:06.371: 52839: `- Dumping CPU info > 20:26:06.372: 52839: `- Sending CPU info > 20:26:06.382: 52839: `- Checking CPU info > 20:26:06.387: 52839: Checking criu version > 20:26:06.405: 52839: Checking for Dirty Tracking > 20:26:06.405: 52839: `- Explicitly enabled > 20:26:06.405: 52839: Preliminary FS migration > 20:28:40.114: 52839: Fs driver transfer 1503657984 bytes (~1434Mb) > 20:28:40.114: 52839: * Iteration 0 > 20:28:40.124: 52839: Making directory > /vz/dump/dmp-kFPTka-16.05.13-20.26/img/1 > 20:28:40.124: 52839: Issuing pre-dump command to service > 20:28:50.770: 52839: Dumped 28797 pages, 0 skipped > 20:28:50.770: 52839: Fs driver transfer 0 bytes > 20:28:50.770: 52839: Checking iteration progress: > 20:28:50.770: 52839: > Proceed to next iteration > 20:28:50.770: 52839: * Iteration 1 > 20:28:50.780: 52839: Making directory > /vz/dump/dmp-kFPTka-16.05.13-20.26/img/2 > 20:28:50.780: 52839: Issuing pre-dump command to service > 20:28:51.278: 52839: Dumped 134 pages, 28663 skipped > 20:28:51.278: 52839: Fs driver transfer 0 bytes > 20:28:51.279: 52839: Checking iteration progress: > 20:28:51.279: 52839: > Proceed to next iteration > 20:28:51.279: 52839: * Iteration 2 > 20:28:51.288: 52839: Making directory > /vz/dump/dmp-kFPTka-16.05.13-20.26/img/3 > 20:28:51.288: 52839: Issuing pre-dump command to service > 20:28:51.778: 52839: Dumped 121 pages, 28676 skipped > 20:28:51.778: 52839: Fs driver transfer 0 bytes > 20:28:51.779: 52839: Checking iteration progress: > 20:28:51.779: 52839: > Proceed to next iteration > 20:28:51.779: 52839: * Iteration 3 > 20:28:51.788: 52839: Making directory > /vz/dump/dmp-kFPTka-16.05.13-20.26/img/4 > 20:28:51.788: 52839: Issuing pre-dump command to service > 20:28:52.241: 52839: Dumped 105 pages, 28692 skipped > 20:28:52.241: 52839: Fs driver transfer 0 bytes > 20:28:52.241: 52839: Checking iteration progress: > 20:28:52.242: 52839: > Proceed to next iteration > 20:28:52.242: 52839: * Iteration 4 > 20:28:52.251: 52839: Making directory > /vz/dump/dmp-kFPTka-16.05.13-20.26/img/5 > 20:28:52.251: 52839: Issuing pre-dump command to service > 20:28:52.748: 52839: Dumped 123 pages, 28674 skipped > 20:28:52.748: 52839: Fs driver transfer 0 bytes > 20:28:52.749: 52839: Checking iteration progress: > 20:28:52.749: 52839: > Iteration grows > 20:28:52.749: 52839: Final dump and restore > 20:28:52.758: 52839: Making directory > /vz/dump/dmp-kFPTka-16.05.13-20.26/img/6 > 20:28:52.758: 52839: Issuing dump command to service > 20:28:52.760: 52839: Notify (pre-dump) > 20:28:52.766: 52839: Notify (network-lock) > 20:28:53.438: 52839: Final FS and images sync > 20:28:54.345: 52839: Sending images to target > 20:28:54.348: 52839: Pack > 20:28:54.386: 52839: Add htype images > 20:28:54.641: 52839: Asking target host to restore > 20:28:57.588: 52839: Remote exception > 20:28:57.588: 52839: Restore failed (17) > Traceback (most recent call last): > File "/usr/libexec/phaul/p.haul", line 47, in <module> > worker.start_migration() > File "/usr/lib/python2.7/site-packages/phaul/iters.py", line 159, in > start_migration > self.__start_live_migration() > File "/usr/lib/python2.7/site-packages/phaul/iters.py", line 229, in > __start_live_migration > self.target_host.restore_from_images() > File "/usr/lib/python2.7/site-packages/phaul/xem_rpc_client.py", line 26, > in __call__ > raise Exception(resp[1]) > Exception: Restore failed (17) > 17:26:06.333: 68411: > 17:26:06.333: 68411: > 17:26:06.334: 68411: > 17:26:06.334: 68411: Starting p.haul service > 17:26:06.334: 68411: Use existing connections, fdrpc=8 fdmem=10 > fdfs=root.hdd/root.hds:12 > 17:26:06.335: 68411: Connected > 17:26:06.342: 68411: Setting up service side ('vz', '888') > 17:26:06.347: 68411: Loading config file from /etc/vz/conf/ > 17:26:06.356: 68411: Initialize ploop receiver > 17:26:06.356: 68411: `- /vz/private/888/root.hdd/root.hds > 17:26:06.359: 68411: Passing (ctl:15, data:7) pair to CRIU > 17:26:06.375: 68411: Started images server > 17:26:06.379: 68411: Waiting for images to unpack > 17:26:06.383: 68411: Checking cpuinfo > 17:26:06.386: 68411: `- True > 17:26:06.391: 68411: Checking criu version > 17:26:06.404: 68411: `- 2.1.0.23 -> 2.1.0.23 > 17:26:06.404: 68411: `- True > 17:28:40.116: 68411: Making directory > /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/1 > 17:28:40.116: 68411: Starting page server for iter 1 > 17:28:40.117: 68411: Sending criu rpc req > 17:28:40.123: 68411: Page server started at 68999 > 17:28:50.772: 68411: Making directory > /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/2 > 17:28:50.773: 68411: Starting page server for iter 2 > 17:28:50.773: 68411: Sending criu rpc req > 17:28:50.779: 68411: Page server started at 69001 > 17:28:51.280: 68411: Making directory > /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/3 > 17:28:51.281: 68411: Starting page server for iter 3 > 17:28:51.281: 68411: Sending criu rpc req > 17:28:51.287: 68411: Page server started at 69003 > 17:28:51.780: 68411: Making directory > /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/4 > 17:28:51.781: 68411: Starting page server for iter 4 > 17:28:51.781: 68411: Sending criu rpc req > 17:28:51.787: 68411: Page server started at 69005 > 17:28:52.243: 68411: Making directory > /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/5 > 17:28:52.244: 68411: Starting page server for iter 5 > 17:28:52.244: 68411: Sending criu rpc req > 17:28:52.251: 68411: Page server started at 69008 > 17:28:52.750: 68411: Making directory > /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/6 > 17:28:52.751: 68411: Starting page server for iter 6 > 17:28:52.751: 68411: Sending criu rpc req > 17:28:52.758: 68411: Page server started at 69010 > 17:28:54.348: 68411: Started images server > 17:28:54.392: 68411: Waiting for images to unpack > 17:28:54.642: 68411: Restoring from images > 17:28:54.643: 68411: Starting vzctl restore > 17:28:57.581: 68411: > Restoring the Container ... > 17:28:57.582: 68411: > Mount image: /vz/private/888/root.hdd > 17:28:57.582: 68411: > Opening delta /vz/private/888/root.hdd/root.hds > 17:28:57.582: 68411: > Adding delta dev=/dev/ploop35223 > img=/vz/private/888/root.hdd/root.hds (rw) > 17:28:57.583: 68411: > Mounted /dev/ploop35223p1 at /vz/root/888 fstype=ext4 > data=',balloon_ino=12' > 17:28:57.583: 68411: > Container is mounted > 17:28:57.583: 68411: > cat: > /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/6/vz_clock_bootbased.img: No such file > or directory > 17:28:57.584: 68411: > cat: > /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/6/vz_clock_monotonic.img: No such file > or directory > 17:28:57.584: 68411: > cat: > /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/6/vz_aio_max_nr.img: No such file or > directory > 17:28:57.585: 68411: > The restore log was saved in > /vz/dump/rst-ee6Sr6-16.05.13-17.26/criu_restore.7.log > 17:28:57.585: 68411: > Unmount image: /vz/private/888/root.hdd > 17:28:57.585: 68411: > Unmounting file system at /vz/root/888 > 17:28:57.586: 68411: > Unmounting device /dev/ploop35223 > 17:28:57.586: 68411: > Container is unmounted > 17:28:57.586: 68411: > Failed to restore the Container > 17:28:57.636: 68411: Disconnected > 17:28:57.638: 68411: Closing images > 17:28:57.638: 68411: Keeping images > 17:28:57.638: 68411: Images are kept in /vz/dump/rst-ee6Sr6-16.05.13-17.26 > 17:28:57.775: 68411: Stop by 15 > 17:28:57.775: 68411: RPC Service stops > 17:28:57.776: 68411: Bye! > _______________________________________________ > Users mailing list > Users@openvz.org > https://lists.openvz.org/mailman/listinfo/users -- sergeyb@ _______________________________________________ Users mailing list Users@openvz.org https://lists.openvz.org/mailman/listinfo/users