Hi Sergey,

prlsrvctl report id: 99722287

Environment details (identical software stack on the 2 hosts):
virtuozzo-release-7.0.0-39.vz7.x86_64
CentOS Linux release 7.2.1511 (Core)

Just to reiterate - as of the previous update, live migration began
working properly; it then broke with the next day's update

I've attached logs from both OVZ hosts (hachi = source, annie = target)

Preparing to submit bug report now -

Jake

On Fri, May 13, 2016 at 2:13 AM, Sergey Bronnikov <serg...@openvz.org> wrote:
> Hi
>
> Thank you for report.
>
> Could submit a new bug to the bugs.openvz.org?
> and provide more details about your environment:
> - Vz7 version (rpm -qa virtuozzo-release)
> - problem report id (prlsrvctl problem-report --send)
>
> Sergey
>
> On 19:42 Thu 12 May , jjs - mainphrame wrote:
>> Aw shucks, I've spoken too soon. After today's updates, live migration
>> is back to not working:
>>
>> [root@hachi ~]# time vzmigrate annie 888 --online
>> Connection to destination node (annie) is successfully established
>> Moving/copying CT 888 -> CT 888, [], [] ...
>> locking 888
>> Checking bindmounts
>> Check cluster ID
>> Checking keep dir for private area copy
>> Check of requires kernel modules
>> Checking technologies
>> Checking templates for CT
>> Checking IP addresses on destination node
>> Checking RATE parameters in config
>> Checking ploop format 2
>> copy CT private /vz/private/888
>> Live migration stage started
>> Phaul service failed to live migrate CT
>> Can't move/copy CT 888 -> CT 888, [], [] : Phaul service failed to
>> live migrate CT
>>
>> real 2m51.840s
>> user 0m0.385s
>> sys 0m3.457s
>> [root@hachi ~]#
>>
>> Jake
>>
>> On Wed, May 11, 2016 at 9:28 PM, jjs - mainphrame <j...@mainphrame.com> 
>> wrote:
>> > Thank you Sergey -
>> >
>> > As of today's updates, I find that live container migration is working
>> > reliably and well here.
>> >
>> > Great work!
>> >
>> > Jake
>> >
>> > On Thu, Apr 28, 2016 at 8:29 AM, Sergey Bronnikov <serg...@openvz.org> 
>> > wrote:
>> >> Hi everyone,
>> >>
>> >> As usual, I'd like to share with you the latest Virtuozzo 7 development 
>> >> news.
>> >>
>> >> First and foremost, more features have become available in Virtuozzo 7:
>> >>
>> >> - Containers migration via vzmigrate in Vz7 [1]
>> >> - IP and MAC filtering for VMs [2]
>> >> - VM data stats supported switch to native KVM ones [3]
>> >> - AutoFS migration [4]
>> >>
>> >> All these features are available in the latests Virtuozzo 7 builds
>> >> https://download.openvz.org/virtuozzo/factory/x86_64/iso/
>> >>
>> >> Also we have published new portions of Virtuozzo source code:
>> >>
>> >> - EZ templates:
>> >>         * Ubuntu 15.10 x86_64
>> >>         * Fedora 23 x86_64
>> >>         * Debian 7.0 x86_64
>> >>         * VzLinux 6 x86_64
>> >>         * VzLinux 7 x86_64
>> >> - DRS module for rmond [5]
>> >>
>> >> Events
>> >> ======
>> >>
>> >> Earlier we participated in Linux Storage, Filesystem and Memory Management
>> >> Conference. See notes from the conference [6] if you interested.
>> >>
>> >> Upcoming talks and conferences:
>> >>
>> >> - Ross Boucher will talk about running services cloning with Docker and 
>> >> CRIU at
>> >> DockerCon 2016 [7]
>> >> - Sanidhya Kashyap will present an instant OS updater using CRIU at USENIX
>> >> ATC'16 [8]
>> >> - Checkpoint-Restore micro-conference will be held 2-4 Nov at Santa-Fe, 
>> >> NM at
>> >> Linux Plumbers 2016. The proposal for the conferense was accepted. [9]
>> >>
>> >> Links
>> >> =====
>> >>
>> >> [1] https://lists.openvz.org/pipermail/users/2016-March/006833.html
>> >> [2] https://lists.openvz.org/pipermail/users/2016-April/006845.html
>> >> [3] https://lists.openvz.org/pipermail/users/2016-April/006850.html
>> >> [4] https://lists.openvz.org/pipermail/users/2016-April/006861.html
>> >> [5] https://github.com/OpenVZ/rmond-drs
>> >> [6] http://lwn.net/Articles/lsfmm2016/
>> >> [7] https://blog.docker.com/2016/04/black-belt-talks-dockercon-2016/
>> >> [8] 
>> >> https://www.usenix.org/conference/atc16/technical-sessions/presentation/kashyap
>> >> [9] 
>> >> https://www.linuxplumbersconf.org/2016/checkpoint-restore-microconference-accepted-into-2016-linux-plumbers-conference/
>> >>
>> >> Sincerely,
>> >> Sergey
>> >> _______________________________________________
>> >> Users mailing list
>> >> Users@openvz.org
>> >> https://lists.openvz.org/mailman/listinfo/users
>> _______________________________________________
>> Users mailing list
>> Users@openvz.org
>> https://lists.openvz.org/mailman/listinfo/users
>
> --
> sergeyb@
[root@hachi ~]# time vzmigrate annie 888 --online
Connection to destination node (annie) is successfully established
Moving/copying CT 888 -> CT 888, [], [] ...
locking 888
Checking bindmounts
Check cluster ID
Checking keep dir for private area copy
Check of requires kernel modules
Checking technologies
Checking templates for CT
Checking IP addresses on destination node
Checking RATE parameters in config
Checking ploop format 2
copy CT private /vz/private/888
Live migration stage started
Phaul service failed to live migrate CT
Can't move/copy CT 888 -> CT 888, [], [] : Phaul service failed to live migrate 
CT

real    2m54.298s
user    0m0.438s
sys     0m3.456s
[root@hachi ~]# 


20:26:06.144: 52839: 
20:26:06.144: 52839: 
20:26:06.144: 52839: 
20:26:06.144: 52839: Starting p.haul
20:26:06.144: 52839: Use existing connections, fdrpc=9 fdmem=11 fdfs=root.hdd/root.hds:13
20:26:06.335: 52839: Setting up local
20:26:06.335: 52839: Loading config file from /etc/vz/conf/
20:26:06.337: 52839: Initialize ploop hauler
20:26:06.337: 52839: 	`- /vz/private/888/root.hdd/root.hds
20:26:06.338: 52839: Passing (ctl:10, data:8) pair to CRIU
20:26:06.340: 52839: Setting up remote
20:26:06.371: 52839: Start migration in live mode
20:26:06.371: 52839: Checking CPU compatibility
20:26:06.371: 52839: 	`- Dumping CPU info
20:26:06.372: 52839: 	`- Sending CPU info
20:26:06.382: 52839: 	`- Checking CPU info
20:26:06.387: 52839: Checking criu version
20:26:06.405: 52839: Checking for Dirty Tracking
20:26:06.405: 52839: 	`- Explicitly enabled
20:26:06.405: 52839: Preliminary FS migration
20:28:40.114: 52839: 	Fs driver transfer 1503657984 bytes (~1434Mb)
20:28:40.114: 52839: * Iteration 0
20:28:40.124: 52839: 	Making directory /vz/dump/dmp-kFPTka-16.05.13-20.26/img/1
20:28:40.124: 52839: 	Issuing pre-dump command to service
20:28:50.770: 52839: 	Dumped 28797 pages, 0 skipped
20:28:50.770: 52839: 	Fs driver transfer 0 bytes
20:28:50.770: 52839: Checking iteration progress:
20:28:50.770: 52839: 	> Proceed to next iteration
20:28:50.770: 52839: * Iteration 1
20:28:50.780: 52839: 	Making directory /vz/dump/dmp-kFPTka-16.05.13-20.26/img/2
20:28:50.780: 52839: 	Issuing pre-dump command to service
20:28:51.278: 52839: 	Dumped 134 pages, 28663 skipped
20:28:51.278: 52839: 	Fs driver transfer 0 bytes
20:28:51.279: 52839: Checking iteration progress:
20:28:51.279: 52839: 	> Proceed to next iteration
20:28:51.279: 52839: * Iteration 2
20:28:51.288: 52839: 	Making directory /vz/dump/dmp-kFPTka-16.05.13-20.26/img/3
20:28:51.288: 52839: 	Issuing pre-dump command to service
20:28:51.778: 52839: 	Dumped 121 pages, 28676 skipped
20:28:51.778: 52839: 	Fs driver transfer 0 bytes
20:28:51.779: 52839: Checking iteration progress:
20:28:51.779: 52839: 	> Proceed to next iteration
20:28:51.779: 52839: * Iteration 3
20:28:51.788: 52839: 	Making directory /vz/dump/dmp-kFPTka-16.05.13-20.26/img/4
20:28:51.788: 52839: 	Issuing pre-dump command to service
20:28:52.241: 52839: 	Dumped 105 pages, 28692 skipped
20:28:52.241: 52839: 	Fs driver transfer 0 bytes
20:28:52.241: 52839: Checking iteration progress:
20:28:52.242: 52839: 	> Proceed to next iteration
20:28:52.242: 52839: * Iteration 4
20:28:52.251: 52839: 	Making directory /vz/dump/dmp-kFPTka-16.05.13-20.26/img/5
20:28:52.251: 52839: 	Issuing pre-dump command to service
20:28:52.748: 52839: 	Dumped 123 pages, 28674 skipped
20:28:52.748: 52839: 	Fs driver transfer 0 bytes
20:28:52.749: 52839: Checking iteration progress:
20:28:52.749: 52839: 	> Iteration grows
20:28:52.749: 52839: Final dump and restore
20:28:52.758: 52839: 	Making directory /vz/dump/dmp-kFPTka-16.05.13-20.26/img/6
20:28:52.758: 52839: 	Issuing dump command to service
20:28:52.760: 52839: 		Notify (pre-dump)
20:28:52.766: 52839: 		Notify (network-lock)
20:28:53.438: 52839: Final FS and images sync
20:28:54.345: 52839: Sending images to target
20:28:54.348: 52839: 	Pack
20:28:54.386: 52839: 	Add htype images
20:28:54.641: 52839: Asking target host to restore
20:28:57.588: 52839: Remote exception
20:28:57.588: 52839: Restore failed (17)
Traceback (most recent call last):
  File "/usr/libexec/phaul/p.haul", line 47, in <module>
    worker.start_migration()
  File "/usr/lib/python2.7/site-packages/phaul/iters.py", line 159, in start_migration
    self.__start_live_migration()
  File "/usr/lib/python2.7/site-packages/phaul/iters.py", line 229, in __start_live_migration
    self.target_host.restore_from_images()
  File "/usr/lib/python2.7/site-packages/phaul/xem_rpc_client.py", line 26, in __call__
    raise Exception(resp[1])
Exception: Restore failed (17)
17:26:06.333: 68411: 
17:26:06.333: 68411: 
17:26:06.334: 68411: 
17:26:06.334: 68411: Starting p.haul service
17:26:06.334: 68411: Use existing connections, fdrpc=8 fdmem=10 fdfs=root.hdd/root.hds:12
17:26:06.335: 68411: Connected
17:26:06.342: 68411: Setting up service side ('vz', '888')
17:26:06.347: 68411: Loading config file from /etc/vz/conf/
17:26:06.356: 68411: Initialize ploop receiver
17:26:06.356: 68411: 	`- /vz/private/888/root.hdd/root.hds
17:26:06.359: 68411: Passing (ctl:15, data:7) pair to CRIU
17:26:06.375: 68411: Started images server
17:26:06.379: 68411: Waiting for images to unpack
17:26:06.383: 68411: Checking cpuinfo
17:26:06.386: 68411: 	`- True
17:26:06.391: 68411: Checking criu version
17:26:06.404: 68411: 	`- 2.1.0.23 -> 2.1.0.23
17:26:06.404: 68411: 	`- True
17:28:40.116: 68411: 	Making directory /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/1
17:28:40.116: 68411: Starting page server for iter 1
17:28:40.117: 68411: 	Sending criu rpc req
17:28:40.123: 68411: 	Page server started at 68999
17:28:50.772: 68411: 	Making directory /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/2
17:28:50.773: 68411: Starting page server for iter 2
17:28:50.773: 68411: 	Sending criu rpc req
17:28:50.779: 68411: 	Page server started at 69001
17:28:51.280: 68411: 	Making directory /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/3
17:28:51.281: 68411: Starting page server for iter 3
17:28:51.281: 68411: 	Sending criu rpc req
17:28:51.287: 68411: 	Page server started at 69003
17:28:51.780: 68411: 	Making directory /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/4
17:28:51.781: 68411: Starting page server for iter 4
17:28:51.781: 68411: 	Sending criu rpc req
17:28:51.787: 68411: 	Page server started at 69005
17:28:52.243: 68411: 	Making directory /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/5
17:28:52.244: 68411: Starting page server for iter 5
17:28:52.244: 68411: 	Sending criu rpc req
17:28:52.251: 68411: 	Page server started at 69008
17:28:52.750: 68411: 	Making directory /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/6
17:28:52.751: 68411: Starting page server for iter 6
17:28:52.751: 68411: 	Sending criu rpc req
17:28:52.758: 68411: 	Page server started at 69010
17:28:54.348: 68411: Started images server
17:28:54.392: 68411: Waiting for images to unpack
17:28:54.642: 68411: Restoring from images
17:28:54.643: 68411: Starting vzctl restore
17:28:57.581: 68411: 	> Restoring the Container ...
17:28:57.582: 68411: 	> Mount image: /vz/private/888/root.hdd
17:28:57.582: 68411: 	> Opening delta /vz/private/888/root.hdd/root.hds
17:28:57.582: 68411: 	> Adding delta dev=/dev/ploop35223 img=/vz/private/888/root.hdd/root.hds (rw)
17:28:57.583: 68411: 	> Mounted /dev/ploop35223p1 at /vz/root/888 fstype=ext4 data=',balloon_ino=12' 
17:28:57.583: 68411: 	> Container is mounted
17:28:57.583: 68411: 	> cat: /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/6/vz_clock_bootbased.img: No such file or directory
17:28:57.584: 68411: 	> cat: /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/6/vz_clock_monotonic.img: No such file or directory
17:28:57.584: 68411: 	> cat: /vz/dump/rst-ee6Sr6-16.05.13-17.26/img/6/vz_aio_max_nr.img: No such file or directory
17:28:57.585: 68411: 	> The restore log was saved in /vz/dump/rst-ee6Sr6-16.05.13-17.26/criu_restore.7.log
17:28:57.585: 68411: 	> Unmount image: /vz/private/888/root.hdd
17:28:57.585: 68411: 	> Unmounting file system at /vz/root/888
17:28:57.586: 68411: 	> Unmounting device /dev/ploop35223
17:28:57.586: 68411: 	> Container is unmounted
17:28:57.586: 68411: 	> Failed to restore the Container
17:28:57.636: 68411: Disconnected
17:28:57.638: 68411: Closing images
17:28:57.638: 68411: Keeping images
17:28:57.638: 68411: Images are kept in /vz/dump/rst-ee6Sr6-16.05.13-17.26
17:28:57.775: 68411: Stop by 15
17:28:57.775: 68411: RPC Service stops
17:28:57.776: 68411: Bye!
_______________________________________________
Users mailing list
Users@openvz.org
https://lists.openvz.org/mailman/listinfo/users

Reply via email to