[ 
https://issues.apache.org/jira/browse/PHOENIX-1263?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14152771#comment-14152771
 ] 

ramkrishna.s.vasudevan commented on PHOENIX-1263:
-------------------------------------------------

[~jeffreyz]
Thanks for the inputs.  Actually I had been running all my test cases in 
Windows using IDE. Even HBase. 
All things were working fine in Phoenix also until friday and all of a sudden 
after I took an update on the code and starting running test cases I started 
seeing this
{code}
014-09-30 09:10:33,946 DEBUG [IPC Server handler 0 on 56981] 
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager(2918): *BLOCK* 
NameNode.processIncrementalBlockReport: from DatanodeRegistration(127.0.0.1, 
datanodeUuid=ff801c90-0493-4f86-b3bf-4ea33ed80d55, infoPort=56993, 
ipcPort=56996, storageInfo=lv=-55;cid=testClusterID;nsid=2022730518;c=0) 
receiving: 1,  received: 0,  deleted: 0
2014-09-30 09:10:33,986 WARN  [PacketResponder: 
BP-2005792417-10.252.156.239-1412048430143:blk_1073741825_1001, 
type=LAST_IN_PIPELINE, downstreams=0:[]] 
org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder(1175): 
IOException in BlockReceiver.run(): 
java.io.IOException: Failed to move meta file for ReplicaBeingWritten, 
blk_1073741825_1001, RBW
  getNumBytes()     = 7
  getBytesOnDisk()  = 7
  getVisibleLength()= 7
  getVolume()       = 
C:\HBase\Phoenix\Apache_phoenix_git\apache-mirror\github-mirror\phoenix-1\phoenix-core\target\test-data\8e2cf486-2af2-407f-bdcf-cf76d36de4b8\dfscluster_07a1bc36-1653-4f73-b927-34fdcc299a57\dfs\data\data1\current
  getBlockFile()    = 
C:\HBase\Phoenix\Apache_phoenix_git\apache-mirror\github-mirror\phoenix-1\phoenix-core\target\test-data\8e2cf486-2af2-407f-bdcf-cf76d36de4b8\dfscluster_07a1bc36-1653-4f73-b927-34fdcc299a57\dfs\data\data1\current\BP-2005792417-10.252.156.239-1412048430143\current\rbw\blk_1073741825
  bytesAcked=7
  bytesOnDisk=7 from 
C:\HBase\Phoenix\Apache_phoenix_git\apache-mirror\github-mirror\phoenix-1\phoenix-core\target\test-data\8e2cf486-2af2-407f-bdcf-cf76d36de4b8\dfscluster_07a1bc36-1653-4f73-b927-34fdcc299a57\dfs\data\data1\current\BP-2005792417-10.252.156.239-1412048430143\current\rbw\blk_1073741825_1001.meta
 to 
C:\HBase\Phoenix\Apache_phoenix_git\apache-mirror\github-mirror\phoenix-1\phoenix-core\target\test-data\8e2cf486-2af2-407f-bdcf-cf76d36de4b8\dfscluster_07a1bc36-1653-4f73-b927-34fdcc299a57\dfs\data\data1\current\BP-2005792417-10.252.156.239-1412048430143\current\finalized\blk_1073741825_1001.meta
        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.moveBlockFiles(FsDatasetImpl.java:462)
        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LDir.addBlock(LDir.java:78)
        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LDir.addBlock(LDir.java:71)
        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice.addBlock(BlockPoolSlice.java:248)
        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsVolumeImpl.addBlock(FsVolumeImpl.java:199)
        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.finalizeReplica(FsDatasetImpl.java:958)
        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.finalizeBlock(FsDatasetImpl.java:939)
        at 
org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.finalizeBlock(BlockReceiver.java:1208)
        at 
org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1164)
        at java.lang.Thread.run(Thread.java:722)
Caused by: 3: The system cannot find the path specified.

        at org.apache.hadoop.io.nativeio.NativeIO.renameTo0(Native Method)
        at org.apache.hadoop.io.nativeio.NativeIO.renameTo(NativeIO.java:828)
        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.moveBlockFiles(FsDatasetImpl.java:460)
        ... 9 more
2014-09-30 09:10:33,990 WARN  [PacketResponder: 
BP-2005792417-10.252.156.239-1412048430143:blk_1073741825_1001, 
type=LAST_IN_PIPELINE, downstreams=0:[]] 
org.apache.hadoop.hdfs.server.datanode.DataNode(1334): checkDiskError: 
exception: 
java.io.IOException: Failed to move meta file for ReplicaBeingWritten, 
blk_1073741825_1001, RBW
  getNumBytes()     = 7
{code}

I remember i used to get this error while running test cases for Hbase on the 
trunk after hadoop 2.2.0 was updated.  Then we installed the natives for that 
and it started working. The same with hadoop 2.4.0.
But now Phoenix does not work with this natives.  Interesting thing is even now 
HBase test cases work with 2.4.0.  I even tried upgrading the Phoenix to point 
to 2.4.0 and still things does not work. I am trying out with this but with no 
benefits :(

> Only cache guideposts on physical PTable
> ----------------------------------------
>
>                 Key: PHOENIX-1263
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-1263
>             Project: Phoenix
>          Issue Type: Sub-task
>            Reporter: James Taylor
>            Assignee: ramkrishna.s.vasudevan
>         Attachments: Phoenix-1263_1.patch
>
>
> Rather than caching the guideposts on all tenant-specific tables, we should 
> cache them only on the physical table. On the client side, we should also 
> update the cache with the latest for the base multi-tenant table when we 
> update the cache for a tenant-specific table. Then when we lookup the 
> guideposts, we should ensure that we're getting them from the physical table.
> Otherwise, it'll be difficult to keep the guideposts cached on the PTable in 
> sync across all tenant-specific tables (not to mention using quite a bit of 
> memory).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to