[jira] [Created] (HADOOP-15887) Add an option to avoid writing data locally in Distcp

2018-10-29 Thread Tao Jie (JIRA)
Tao Jie created HADOOP-15887:


 Summary: Add an option to avoid writing data locally in Distcp
 Key: HADOOP-15887
 URL: https://issues.apache.org/jira/browse/HADOOP-15887
 Project: Hadoop Common
  Issue Type: Improvement
Affects Versions: 3.0.0, 2.8.2
Reporter: Tao Jie
Assignee: Tao Jie


When copying large amount of data from one cluster to another via Distcp, and 
the Distcp jobs run in the target cluster, the datanode local usage would be 
imbalanced. Because the default placement policy chooses the local node to 
store the first replication.

In https://issues.apache.org/jira/browse/HDFS-3702 we add a flag in DFSClient 
to avoid replicating to the local datanode.  We can make use of this flag in 
Distcp.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-15419) Should not obtain delegationTokens from all namenodes when using ViewFS

2018-04-26 Thread Tao Jie (JIRA)
Tao Jie created HADOOP-15419:


 Summary: Should not obtain delegationTokens from all namenodes 
when using ViewFS
 Key: HADOOP-15419
 URL: https://issues.apache.org/jira/browse/HADOOP-15419
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 3.0.2
Reporter: Tao Jie


Today when submit a job to a viewfs cluster, the client will try to obtain 
delegation token from all namenodes under the viewfs while only one namespace 
is actually used in this job. It would create many unnecessary rpc call to the 
whole cluster.
In viewfs situation, we can just obtain delegation token from specific namenode 
rather than all namenodes.




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-15253) Should update maxQueueSize when refresh call queue

2018-02-22 Thread Tao Jie (JIRA)
Tao Jie created HADOOP-15253:


 Summary: Should update maxQueueSize when refresh call queue
 Key: HADOOP-15253
 URL: https://issues.apache.org/jira/browse/HADOOP-15253
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Tao Jie
Assignee: Tao Jie


When calling {{dfsadmin -refreshCallQueue}} to update CallQueue instance, 
{{maxQueueSize}} should also be updated.
In case of changing CallQueue instance to FairCallQueue, the length of each 
queue in FairCallQueue would be 1/priorityLevels of original length of 
DefaultCallQueue. So it would be helpful for us to set the length of callqueue 
to a proper value.




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-15194) Some properties for FairCallQueue should be reloadable

2018-01-26 Thread Tao Jie (JIRA)
Tao Jie created HADOOP-15194:


 Summary: Some properties for FairCallQueue should be reloadable
 Key: HADOOP-15194
 URL: https://issues.apache.org/jira/browse/HADOOP-15194
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Tao Jie


When trying to make use of FairCallQueue for a large cluster, it is not easy to 
set the properties appropriately at the first time. User may adjust those 
properties frequently and restarting the namenode costs a lot for a large 
multi-user hadoop clluster. We should support to reload some of those 
properties in configuration without restarting.
Related properties:
ipc.8020.faircallqueue.multiplexer.weights
ipc.8020.faircallqueue.decay-scheduler.thresholds
ipc.8020.decay-scheduler.backoff.responsetime.thresholds




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-15121) Encounter NullPointerException when using DecayRpcScheduler

2017-12-14 Thread Tao Jie (JIRA)
Tao Jie created HADOOP-15121:


 Summary: Encounter NullPointerException when using 
DecayRpcScheduler
 Key: HADOOP-15121
 URL: https://issues.apache.org/jira/browse/HADOOP-15121
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Tao Jie


I set ipc.8020.scheduler.impl to org.apache.hadoop.ipc.DecayRpcScheduler, but 
got excetion in namenode:
{code}
2017-12-15 15:26:34,662 ERROR impl.MetricsSourceAdapter 
(MetricsSourceAdapter.java:getMetrics(202)) - Error getting metrics from source 
DecayRpcSchedulerMetrics2.ipc.8020
java.lang.NullPointerException
at 
org.apache.hadoop.ipc.DecayRpcScheduler$MetricsProxy.getMetrics(DecayRpcScheduler.java:781)
at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMetrics(MetricsSourceAdapter.java:199)
at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.updateJmxCache(MetricsSourceAdapter.java:182)
at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMBeanInfo(MetricsSourceAdapter.java:155)
at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getNewMBeanClassName(DefaultMBeanServerInterceptor.java:333)
at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerMBean(DefaultMBeanServerInterceptor.java:319)
at 
com.sun.jmx.mbeanserver.JmxMBeanServer.registerMBean(JmxMBeanServer.java:522)
at org.apache.hadoop.metrics2.util.MBeans.register(MBeans.java:66)
at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.startMBeans(MetricsSourceAdapter.java:222)
at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.start(MetricsSourceAdapter.java:100)
at 
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.registerSource(MetricsSystemImpl.java:268)
at 
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.register(MetricsSystemImpl.java:233)
at 
org.apache.hadoop.ipc.DecayRpcScheduler$MetricsProxy.registerMetrics2Source(DecayRpcScheduler.java:709)
at 
org.apache.hadoop.ipc.DecayRpcScheduler$MetricsProxy.(DecayRpcScheduler.java:685)
at 
org.apache.hadoop.ipc.DecayRpcScheduler$MetricsProxy.getInstance(DecayRpcScheduler.java:693)
at 
org.apache.hadoop.ipc.DecayRpcScheduler.(DecayRpcScheduler.java:236)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at 
org.apache.hadoop.ipc.CallQueueManager.createScheduler(CallQueueManager.java:102)
at 
org.apache.hadoop.ipc.CallQueueManager.(CallQueueManager.java:76)
at org.apache.hadoop.ipc.Server.(Server.java:2612)
at org.apache.hadoop.ipc.RPC$Server.(RPC.java:958)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server.(ProtobufRpcEngine.java:374)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine.getServer(ProtobufRpcEngine.java:349)
at org.apache.hadoop.ipc.RPC$Builder.build(RPC.java:800)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.(NameNodeRpcServer.java:415)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.createRpcServer(NameNode.java:755)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:697)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:905)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:884)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1610)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1678)
{code}
It seems that {{metricsProxy}} in DecayRpcScheduler should initiate its 
{{delegate}} field in its Initialization method




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-12569) ZKFC should stop namenode before itself quit in some circumstances

2015-11-13 Thread Tao Jie (JIRA)
Tao Jie created HADOOP-12569:


 Summary: ZKFC should stop namenode before itself quit in some 
circumstances
 Key: HADOOP-12569
 URL: https://issues.apache.org/jira/browse/HADOOP-12569
 Project: Hadoop Common
  Issue Type: Bug
  Components: ha
Affects Versions: 2.6.0
Reporter: Tao Jie


We have met such a HA scenario:
NN1(active) and zkfc1 on node1;
NN2(standby) and zkfc2 on node2.
1,Stop network on node1, NN2 becomes active. On node2, zkfc2 kills itself since 
it cannot connect to zookeeper, but leaving NN1 still running.
2,Several minutes later, network on node1 recovers. NN1 is running but out of 
control. NN1 and NN2 both run as active nn.
Maybe zkfc should stop nn before quit in such circumstances.




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HADOOP-12331) FSShell Delete and Rename should operate on symlinks rather than their target

2015-08-17 Thread Tao Jie (JIRA)
Tao Jie created HADOOP-12331:


 Summary: FSShell Delete and Rename should operate on symlinks 
rather than their target
 Key: HADOOP-12331
 URL: https://issues.apache.org/jira/browse/HADOOP-12331
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs
Affects Versions: 2.6.0
Reporter: Tao Jie
Priority: Minor


Currently,  we remove or rename a symlink in FSShell, symlink target is 
affected instead of symlink itself. However FileSystem#delete and 
FileSystem#rename can operate on symlinks rather than their targets.
FSShell should have consistent behavior as FileSystem on symlink remove and 
rename



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)