[ https://issues.apache.org/jira/browse/CLOUDSTACK-3229?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741926#comment-13741926 ]
Thomas O'Dowd commented on CLOUDSTACK-3229: ------------------------------------------- Min - sorry for the long wait. I went to lunch. Following the other lines quoted above, I can see the following... [13872] 2013-08-16 03:29:34.137590 lock: acquired /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/sr [13872] 2013-08-16 03:29:34.138373 sr_scan {'sr_uuid': '9f3f9262-3f77-09cc-2df7-0d8475676260', 'subtask_of': 'DummyRef:|61e45046-c481-6607-89d4-d08e7134a759|SR.scan', 'args': [], 'host_ref': 'OpaqueRef:0d19f3fa-7f04-543a-d906-aaf478fc574c', 'session_ref': 'OpaqueRef:17f831b6-9115-37a2-8b81-e9a496c1a1f9', 'device_config': {'SRmaster': 'true', 'location': '/opt/storage/primary'}, 'command': 'sr_scan', 'sr_ref': 'OpaqueRef:75f9feeb-8309-b7b7-8455-9a1021a3baa4'} [13872] 2013-08-16 03:29:34.138570 ['/usr/bin/vhd-util', 'scan', '-f', '-c', '-m', '/opt/storage/primary/*.vhd'] [13872] 2013-08-16 03:29:34.151355 SUCCESS [13872] 2013-08-16 03:29:34.153339 ['ls', '/opt/storage/primary', '-1', '--color=never'] [13872] 2013-08-16 03:29:34.160103 SUCCESS [13872] 2013-08-16 03:29:34.160386 Scan found hidden leaf (ab85c9d3-56a1-45fa-b698-3410b77e83be), ignoring [13872] 2013-08-16 03:29:34.160454 Scan found hidden leaf (9fd77231-04a9-4b24-94b8-d61cee1ef9e0), ignoring [13872] 2013-08-16 03:29:34.160775 lock: tried lock /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/running, acquired: True (exists: True) [13872] 2013-08-16 03:29:34.160851 lock: released /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/running [13872] 2013-08-16 03:29:34.160892 Kicking GC <13872> 2013-08-16 03:29:34.160955 === SR 9f3f9262-3f77-09cc-2df7-0d8475676260: gc === <13876> 2013-08-16 03:29:34.164529 Will finish as PID [13877] <13872> 2013-08-16 03:29:34.165298 New PID [13876] [13872] 2013-08-16 03:29:34.165699 lock: closed /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/running [13872] 2013-08-16 03:29:34.230920 lock: released /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/sr <13877> 2013-08-16 03:29:34.236035 *~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~* <13877> 2013-08-16 03:29:34.236143 *********************** <13877> 2013-08-16 03:29:34.236191 * E X C E P T I O N * <13877> 2013-08-16 03:29:34.236235 *********************** <13877> 2013-08-16 03:29:34.236319 gc: EXCEPTION <type 'exceptions.OSError'>, [Errno 2] No such file or directory: '/var/run/sr-mount/9f3f9262-3f77-09cc-2df7-0d8475676260' <13877> 2013-08-16 03:29:34.236367 File "/usr/lib/xcp/sm/cleanup.py", line 2515, in gc _gc(None, srUuid, dryRun) File "/usr/lib/xcp/sm/cleanup.py", line 2415, in _gc sr.cleanupCache() File "/usr/lib/xcp/sm/cleanup.py", line 1867, in cleanupCache cacheFiles = filter(self._isCacheFileName, os.listdir(self.path)) <13877> 2013-08-16 03:29:34.236413 *~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~* <13877> 2013-08-16 03:29:34.236459 * * * * * SR 9f3f9262-3f77-09cc-2df7-0d8475676260: ERROR [13872] 2013-08-16 03:29:34.238564 lock: closed /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/sr [13910] 2013-08-16 03:30:15.889586 #### VMOPS enter gethostvmstats #### [13910] 2013-08-16 03:30:15.893586 #### VMOPS exit gethostvmstats #### [13917] 2013-08-16 03:30:18.591527 #### VMOPS enter gethostvmstats #### [13917] 2013-08-16 03:30:18.595713 #### VMOPS exit gethostvmstats #### [13927] 2013-08-16 03:30:34.465665 lock: acquired /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/sr [13927] 2013-08-16 03:30:34.465770 sr_scan {'sr_uuid': '9f3f9262-3f77-09cc-2df7-0d8475676260', 'subtask_of': 'DummyRef:|34268ce3-fc3a-ceb8-9821-9eafc05a6277|SR.scan', 'args': [], 'host_ref': 'OpaqueRef:0d19f3fa-7f04-543a-d906-aaf478fc574c', 'session_ref': 'OpaqueRef:0b1853b4-58c4-45de-c159-0482bccf6bab', 'device_config': {'SRmaster': 'true', 'location': '/opt/storage/primary'}, 'command': 'sr_scan', 'sr_ref': 'OpaqueRef:75f9feeb-8309-b7b7-8455-9a1021a3baa4'} [13927] 2013-08-16 03:30:34.465931 ['/usr/bin/vhd-util', 'scan', '-f', '-c', '-m', '/opt/storage/primary/*.vhd'] [13927] 2013-08-16 03:30:34.474920 SUCCESS [13927] 2013-08-16 03:30:34.478510 ['ls', '/opt/storage/primary', '-1', '--color=never'] [13927] 2013-08-16 03:30:34.483966 SUCCESS [13927] 2013-08-16 03:30:34.484193 Scan found hidden leaf (ab85c9d3-56a1-45fa-b698-3410b77e83be), ignoring [13927] 2013-08-16 03:30:34.484271 Scan found hidden leaf (9fd77231-04a9-4b24-94b8-d61cee1ef9e0), ignoring [13927] 2013-08-16 03:30:34.484581 lock: tried lock /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/running, acquired: True (exists: True) [13927] 2013-08-16 03:30:34.484662 lock: released /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/running [13927] 2013-08-16 03:30:34.484707 Kicking GC <13927> 2013-08-16 03:30:34.484772 === SR 9f3f9262-3f77-09cc-2df7-0d8475676260: gc === <13931> 2013-08-16 03:30:34.487964 Will finish as PID [13932] <13927> 2013-08-16 03:30:34.488619 New PID [13931] [13927] 2013-08-16 03:30:34.488937 lock: closed /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/running [13927] 2013-08-16 03:30:34.536340 lock: released /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/sr [13927] 2013-08-16 03:30:34.540301 lock: closed /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/sr <13932> 2013-08-16 03:30:34.561067 *~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~* <13932> 2013-08-16 03:30:34.561168 *********************** <13932> 2013-08-16 03:30:34.561212 * E X C E P T I O N * <13932> 2013-08-16 03:30:34.561255 *********************** <13932> 2013-08-16 03:30:34.561326 gc: EXCEPTION <type 'exceptions.OSError'>, [Errno 2] No such file or directory: '/var/run/sr-mount/9f3f9262-3f77-09cc-2df7-0d8475676260' <13932> 2013-08-16 03:30:34.561370 File "/usr/lib/xcp/sm/cleanup.py", line 2515, in gc _gc(None, srUuid, dryRun) File "/usr/lib/xcp/sm/cleanup.py", line 2415, in _gc sr.cleanupCache() File "/usr/lib/xcp/sm/cleanup.py", line 1867, in cleanupCache cacheFiles = filter(self._isCacheFileName, os.listdir(self.path)) <13932> 2013-08-16 03:30:34.561413 *~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~* <13932> 2013-08-16 03:30:34.561456 * * * * * SR 9f3f9262-3f77-09cc-2df7-0d8475676260: ERROR [13965] 2013-08-16 03:31:16.167244 #### VMOPS enter gethostvmstats #### [13965] 2013-08-16 03:31:16.171111 #### VMOPS exit gethostvmstats #### [13972] 2013-08-16 03:31:18.806609 #### VMOPS enter gethostvmstats #### [13972] 2013-08-16 03:31:18.810897 #### VMOPS exit gethostvmstats #### [13982] 2013-08-16 03:31:34.776573 lock: acquired /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/sr [13982] 2013-08-16 03:31:34.776677 sr_scan {'sr_uuid': '9f3f9262-3f77-09cc-2df7-0d8475676260', 'subtask_of': 'DummyRef:|7cfbd3bc-5082-6a5b-6c38-63456ef108b1|SR.scan', 'args': [], 'host_ref': 'OpaqueRef:0d19f3fa-7f04-543a-d906-aaf478fc574c', 'session_ref': 'OpaqueRef:1b3d55b9-a110-3c47-c890-147f475de984', 'device_config': {'SRmaster': 'true', 'location': '/opt/storage/primary'}, 'command': 'sr_scan', 'sr_ref': 'OpaqueRef:75f9feeb-8309-b7b7-8455-9a1021a3baa4'} [13982] 2013-08-16 03:31:34.776807 ['/usr/bin/vhd-util', 'scan', '-f', '-c', '-m', '/opt/storage/primary/*.vhd'] [13982] 2013-08-16 03:31:34.784780 SUCCESS [13982] 2013-08-16 03:31:34.788399 ['ls', '/opt/storage/primary', '-1', '--color=never'] [13982] 2013-08-16 03:31:34.793583 SUCCESS [13982] 2013-08-16 03:31:34.793794 Scan found hidden leaf (ab85c9d3-56a1-45fa-b698-3410b77e83be), ignoring [13982] 2013-08-16 03:31:34.793887 Scan found hidden leaf (9fd77231-04a9-4b24-94b8-d61cee1ef9e0), ignoring [13982] 2013-08-16 03:31:34.794182 lock: tried lock /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/running, acquired: True (exists: True) [13982] 2013-08-16 03:31:34.794262 lock: released /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/running [13982] 2013-08-16 03:31:34.794305 Kicking GC <13982> 2013-08-16 03:31:34.794366 === SR 9f3f9262-3f77-09cc-2df7-0d8475676260: gc === <13986> 2013-08-16 03:31:34.797329 Will finish as PID [13987] <13982> 2013-08-16 03:31:34.797952 New PID [13986] [13982] 2013-08-16 03:31:34.798274 lock: closed /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/running [13982] 2013-08-16 03:31:34.850358 lock: released /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/sr <13987> 2013-08-16 03:31:34.858440 *~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~* <13987> 2013-08-16 03:31:34.858535 *********************** <13987> 2013-08-16 03:31:34.858576 * E X C E P T I O N * <13987> 2013-08-16 03:31:34.858615 *********************** <13987> 2013-08-16 03:31:34.858680 gc: EXCEPTION <type 'exceptions.OSError'>, [Errno 2] No such file or directory: '/var/run/sr-mount/9f3f9262-3f77-09cc-2df7-0d8475676260' <13987> 2013-08-16 03:31:34.858721 File "/usr/lib/xcp/sm/cleanup.py", line 2515, in gc _gc(None, srUuid, dryRun) File "/usr/lib/xcp/sm/cleanup.py", line 2415, in _gc sr.cleanupCache() File "/usr/lib/xcp/sm/cleanup.py", line 1867, in cleanupCache cacheFiles = filter(self._isCacheFileName, os.listdir(self.path)) <13987> 2013-08-16 03:31:34.858761 *~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~* <13987> 2013-08-16 03:31:34.858801 * * * * * SR 9f3f9262-3f77-09cc-2df7-0d8475676260: ERROR [13982] 2013-08-16 03:31:34.859919 lock: closed /var/lock/sm/9f3f9262-3f77-09cc-2df7-0d8475676260/sr [14020] 2013-08-16 03:32:16.440440 #### VMOPS enter gethostvmstats #### [14020] 2013-08-16 03:32:16.444436 #### VMOPS exit gethostvmstats #### [14027] 2013-08-16 03:32:19.016155 #### VMOPS enter gethostvmstats #### [14027] 2013-08-16 03:32:19.020161 #### VMOPS exit gethostvmstats #### .... etc .... (seems to be repeating). If it helps I can attach the SMlog file. Let me know. Tom. > Object_Store_Refactor - Snapshot fails due to an internal error > --------------------------------------------------------------- > > Key: CLOUDSTACK-3229 > URL: https://issues.apache.org/jira/browse/CLOUDSTACK-3229 > Project: CloudStack > Issue Type: Bug > Security Level: Public(Anyone can view this level - this is the > default.) > Affects Versions: 4.2.0 > Environment: chrome on linux > devcloud > Cloudian or Amazon S3 Object store > Reporter: Thomas O'Dowd > Assignee: Min Chen > Priority: Blocker > Fix For: 4.2.0 > > Attachments: SMlog, SMlog.last_5000_lines.txt > > > Assuming initial devcloud state... > I added a cache for the S3 storage like this. > on devcloud machine as root: > # mkdir /opt/storage/cache > # vi /etc/exports (and append this line) > /opt/storage/cache *(rw,no_subtree_check,no_root_squash,fsid=9999) > # exportfs -a > On Mgmt server GUI: > 1. navigate to infrastructure -> secondary storage > 2. delete the NFS SS. > 3. add S3 storage for Cloudian (I used 60000 as the timeouts - assuming > millis). I used the /opt/storage/cache thing as the s3 cache. > 4. nav to templates > 5. register a new template (I uploaded tinyLinux again as "mytiny" (5.3 > 64bit)). > 6. confirm with s3cmd that 2 objects are now on S3. > --------- s3 objects ------- > template/tmpl/1/1/routing-1/acton-systemvm-02062012.vhd.bz2 > 2013-06-27T03:01:46.203Z None 140616708 "b533e7b65219439ee7fca0146ddd7ffa-27" > template/tmpl/2/201/201-2-ae9e9409-4c8e-3ad8-a62f-abec7a35fe26/tinylinux.vhd > 2013-06-27T03:04:06.730Z None 50430464 "4afac316e865adf74ca1a8039fae7399-10" > --------- s3 objects ------- > 7. I restarted the management server at this point which actually resulted in > another object on S3. > --------- the new s3 object ------- > template/tmpl/1/5/tiny Linux/ttylinux_pv.vhd 2013-06-27T03:43:26.494Z None > 50430464 "4afac316e865adf74ca1a8039fae7399-10" > --------- the new s3 object ------- > 8. Go to instance and create a new choosing the "mytiny" template which we > registered. > 9. launch it after selecting all defaults. > 10. wait until it starts. > 11. nav to storage. I see ROOT-8. Click on this to open. > 12. click the camera to take the snapshot. > after a pause I get a popup > "Failed to create snapshot due to an internal error creating snapshot > for volume 8" > Also on the mgmt terminal I get the following log entry (only 1): > INFO [user.snapshot.CreateSnapshotCmd] (Job-Executor-8:job-16) VOLSS: > createSnapshotCmd starts:1372321251009 > If I check the "view snapshots" button under storage, I can however see the > snapshot. It says its on primary. I'm expecting it to go to secondary storage > though. Nothing is in the S3 logs and no snapshots. > If I try to delete that snapshot from here I get this error in the logs: > ERROR [cloud.async.AsyncJobManagerImpl] (Job-Executor-12:job-20) Unexpected > exception while executing > org.apache.cloudstack.api.command.user.snapshot.DeleteSnapshotCmd > com.cloud.utils.exception.CloudRuntimeException: Failed to delete > snapshot:com.cloud.exception.InvalidParameterValueException: Can't delete > snapshotshot 4 due to it is not in BackedUp Status > at > com.cloud.storage.snapshot.SnapshotManagerImpl.deleteSnapshot(SnapshotManagerImpl.java:513) > at > com.cloud.utils.component.ComponentInstantiationPostProcessor$InterceptorDispatcher.intercept(ComponentInstantiationPostProcessor.java:125) > at > org.apache.cloudstack.api.command.user.snapshot.DeleteSnapshotCmd.execute(DeleteSnapshotCmd.java:96) > at com.cloud.api.ApiDispatcher.dispatch(ApiDispatcher.java:155) > at > com.cloud.async.AsyncJobManagerImpl$1.run(AsyncJobManagerImpl.java:437) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) > at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) > at java.util.concurrent.FutureTask.run(FutureTask.java:166) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:722) > If I navigate to instance, my instance, and try to take a vm snapshot from > here, I get a different pop-up which says: > "There is other active volume snapshot tasks on the instance to which the > volume is attached, please try again later" > And I get an exception: > ERROR [cloud.api.ApiServer] (352129314@qtp-2110413789-32:) unhandled > exception executing api command: createVMSnapshot > com.cloud.utils.exception.CloudRuntimeException: There is other active volume > snapshot tasks on the instance to which the volume is attached, please try > again later. > at > com.cloud.vm.snapshot.VMSnapshotManagerImpl.allocVMSnapshot(VMSnapshotManagerImpl.java:299) > at > org.apache.cloudstack.api.command.user.vmsnapshot.CreateVMSnapshotCmd.create(CreateVMSnapshotCmd.java:78) > at > com.cloud.api.ApiDispatcher.dispatchCreateCmd(ApiDispatcher.java:101) > at com.cloud.api.ApiServer.queueCommand(ApiServer.java:475) > at com.cloud.api.ApiServer.handleRequest(ApiServer.java:371) > at com.cloud.api.ApiServlet.processRequest(ApiServlet.java:304) > at com.cloud.api.ApiServlet.doGet(ApiServlet.java:66) > at javax.servlet.http.HttpServlet.service(HttpServlet.java:707) > at javax.servlet.http.HttpServlet.service(HttpServlet.java:820) > at > org.mortbay.jetty.servlet.ServletHolder.handle(ServletHolder.java:511) > at > org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:401) > at > org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216) > at > org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:182) > at > org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:766) > at > org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:450) > at > org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:230) > at > org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114) > at > org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152) > at org.mortbay.jetty.Server.handle(Server.java:326) > at > org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:542) > at > org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:928) > at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:549) > at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:212) > at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:404) > at > org.mortbay.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:410) > at > org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582) > There are no requests going to the S3 storage for the snap-shotting that I > can see and its the only secondary storage that I have setup. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira