[jira] [Updated] (FLINK-9363) Bump up the Jackson version

2018-09-25 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-9363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated FLINK-9363:
--
Description: 
CVE's for Jackson :

CVE-2017-17485
CVE-2018-5968
CVE-2018-7489


We can upgrade to 2.9.5

  was:
CVE's for Jackson :

CVE-2017-17485
CVE-2018-5968
CVE-2018-7489

We can upgrade to 2.9.5


> Bump up the Jackson version
> ---
>
> Key: FLINK-9363
> URL: https://issues.apache.org/jira/browse/FLINK-9363
> Project: Flink
>  Issue Type: Improvement
>    Reporter: Ted Yu
>Assignee: vinoyang
>Priority: Major
>  Labels: security
>
> CVE's for Jackson :
> CVE-2017-17485
> CVE-2018-5968
> CVE-2018-7489
> We can upgrade to 2.9.5



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-21230) BackupUtils#checkTargetDir doesn't compose error message correctly

2018-09-25 Thread Ted Yu (JIRA)
Ted Yu created HBASE-21230:
--

 Summary: BackupUtils#checkTargetDir doesn't compose error message 
correctly
 Key: HBASE-21230
 URL: https://issues.apache.org/jira/browse/HBASE-21230
 Project: HBase
  Issue Type: Bug
  Components: backuprestore
Reporter: Ted Yu


Here is related code:
{code}
  String expMsg = e.getMessage();
  String newMsg = null;
  if (expMsg.contains("No FileSystem for scheme")) {
newMsg =
"Unsupported filesystem scheme found in the backup target url. 
Error Message: "
+ newMsg;
{code}
I think the intention was to concatenate expMsg at the end of newMsg.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-21230) BackupUtils#checkTargetDir doesn't compose error message correctly

2018-09-25 Thread Ted Yu (JIRA)
Ted Yu created HBASE-21230:
--

 Summary: BackupUtils#checkTargetDir doesn't compose error message 
correctly
 Key: HBASE-21230
 URL: https://issues.apache.org/jira/browse/HBASE-21230
 Project: HBase
  Issue Type: Bug
  Components: backuprestore
Reporter: Ted Yu


Here is related code:
{code}
  String expMsg = e.getMessage();
  String newMsg = null;
  if (expMsg.contains("No FileSystem for scheme")) {
newMsg =
"Unsupported filesystem scheme found in the backup target url. 
Error Message: "
+ newMsg;
{code}
I think the intention was to concatenate expMsg at the end of newMsg.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21221) Ineffective assertion in TestFromClientSide3#testMultiRowMutations

2018-09-25 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-21221:
---
   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 3.0.0
   Status: Resolved  (was: Patch Available)

Thanks for the review, Mingliang.

> Ineffective assertion in TestFromClientSide3#testMultiRowMutations
> --
>
> Key: HBASE-21221
> URL: https://issues.apache.org/jira/browse/HBASE-21221
> Project: HBase
>  Issue Type: Test
>    Reporter: Ted Yu
>    Assignee: Ted Yu
>Priority: Minor
> Fix For: 3.0.0
>
> Attachments: 21221.v10.txt, 21221.v11.txt, 21221.v12.txt, 
> 21221.v7.txt, 21221.v8.txt, 21221.v9.txt
>
>
> Observed the following in 
> org.apache.hadoop.hbase.util.TestFromClientSide3WoUnsafe-output.txt :
> {code}
> Caused by: 
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> java.io.IOException: Timed out waiting for lock for row: ROW-1 in region 
> 089bdfa75f44d88e596479038a6da18b
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:5816)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion$4.lockRowsAndBuildMiniBatch(HRegion.java:7432)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutate(HRegion.java:4008)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3982)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.mutateRowsWithLocks(HRegion.java:7424)
>   at 
> org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.mutateRows(MultiRowMutationEndpoint.java:116)
>   at 
> org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos$MultiRowMutationService.callMethod(MultiRowMutationProtos.java:2266)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8182)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2481)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2463)
> ...
> Exception in thread "pool-678-thread-1" java.lang.AssertionError: This cp 
> should fail because the target lock is blocked by previous put
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.hadoop.hbase.client.TestFromClientSide3.lambda$testMultiRowMutations$7(TestFromClientSide3.java:861)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> {code}
> Here is related code:
> {code}
>   cpService.execute(() -> {
> ...
> if (!threw) {
>   // Can't call fail() earlier because the catch would eat it.
>   fail("This cp should fail because the target lock is blocked by 
> previous put");
> }
> {code}
> Since the fail() call is executed by the cpService, the assertion had no 
> bearing on the outcome of the test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21228) Memory leak since AbstractFSWAL caches Thread object and never clean later

2018-09-25 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21228?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16627554#comment-16627554
 ] 

Ted Yu commented on HBASE-21228:


+1, pending QA

> Memory leak since AbstractFSWAL caches Thread object and never clean later
> --
>
> Key: HBASE-21228
> URL: https://issues.apache.org/jira/browse/HBASE-21228
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 2.1.0, 2.0.2, 1.4.7
>Reporter: Allan Yang
>Assignee: Allan Yang
>Priority: Major
> Attachments: HBASE-21228.branch-2.0.001.patch
>
>
> In AbstractFSWAL(FSHLog in branch-1), we have a map caches thread and 
> SyncFutures.
> {code}
> /**
>* Map of {@link SyncFuture}s keyed by Handler objects. Used so we reuse 
> SyncFutures.
>* 
>* TODO: Reuse FSWALEntry's rather than create them anew each time as we do 
> SyncFutures here.
>* 
>* TODO: Add a FSWalEntry and SyncFuture as thread locals on handlers 
> rather than have them get
>* them from this Map?
>*/
>   private final ConcurrentMap syncFuturesByHandler;
> {code}
> A colleague of mine find a memory leak case caused by this map.
> Every thread who writes WAL will be cached in this map, And no one will clean 
> the threads in the map even after the thread is dead. 
> In one of our customer's cluster, we noticed that even though there is no 
> requests, the heap of the RS is almost full and CMS GC was triggered every 
> second.
> We dumped the heap and then found out there were more than 30 thousands 
> threads with Terminated state. which are all cached in this map above. 
> Everything referenced in these threads were leaked. Most of the threads are:
> 1.PostOpenDeployTasksThread, which will write Open Region mark in WAL
> 2. hconnection-0x1f838e31-shared--pool, which are used to write index short 
> circuit(Phoenix), and WAL will be write and sync in these threads.
> 3.  Index writer thread(Phoenix), which referenced by 
> RegionCoprocessorHost$RegionEnvironment then by HRegion and finally been 
> referenced by PostOpenDeployTasksThread.
> We should turn this map into a thread local one, let JVM GC the terminated 
> thread for us. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (KAFKA-5863) Potential null dereference in DistributedHerder#reconfigureConnector()

2018-09-24 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/KAFKA-5863?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated KAFKA-5863:
--
Description: 
Here is the call chain:
{code}
RestServer.httpRequest(reconfigUrl, "POST", 
taskProps, null);
{code}
In httpRequest():
{code}
} else if (responseCode >= 200 && responseCode < 300) {
InputStream is = connection.getInputStream();
T result = JSON_SERDE.readValue(is, responseFormat);
{code}
For readValue():
{code}
public  T readValue(InputStream src, TypeReference valueTypeRef)
throws IOException, JsonParseException, JsonMappingException
{
return (T) _readMapAndClose(_jsonFactory.createParser(src), 
_typeFactory.constructType(valueTypeRef));
{code}
Then there would be NPE in constructType():
{code}
public JavaType constructType(TypeReference typeRef)
{
// 19-Oct-2015, tatu: Simpler variant like so should work
return _fromAny(null, typeRef.getType(), EMPTY_BINDINGS);
{code}

  was:
Here is the call chain:

{code}
RestServer.httpRequest(reconfigUrl, "POST", 
taskProps, null);
{code}
In httpRequest():
{code}
} else if (responseCode >= 200 && responseCode < 300) {
InputStream is = connection.getInputStream();
T result = JSON_SERDE.readValue(is, responseFormat);
{code}
For readValue():
{code}
public  T readValue(InputStream src, TypeReference valueTypeRef)
throws IOException, JsonParseException, JsonMappingException
{
return (T) _readMapAndClose(_jsonFactory.createParser(src), 
_typeFactory.constructType(valueTypeRef));
{code}
Then there would be NPE in constructType():
{code}
public JavaType constructType(TypeReference typeRef)
{
// 19-Oct-2015, tatu: Simpler variant like so should work
return _fromAny(null, typeRef.getType(), EMPTY_BINDINGS);
{code}


> Potential null dereference in DistributedHerder#reconfigureConnector()
> --
>
> Key: KAFKA-5863
> URL: https://issues.apache.org/jira/browse/KAFKA-5863
> Project: Kafka
>  Issue Type: Bug
>  Components: KafkaConnect
>Reporter: Ted Yu
>Priority: Minor
>
> Here is the call chain:
> {code}
> RestServer.httpRequest(reconfigUrl, "POST", 
> taskProps, null);
> {code}
> In httpRequest():
> {code}
> } else if (responseCode >= 200 && responseCode < 300) {
> InputStream is = connection.getInputStream();
> T result = JSON_SERDE.readValue(is, responseFormat);
> {code}
> For readValue():
> {code}
> public  T readValue(InputStream src, TypeReference valueTypeRef)
> throws IOException, JsonParseException, JsonMappingException
> {
> return (T) _readMapAndClose(_jsonFactory.createParser(src), 
> _typeFactory.constructType(valueTypeRef));
> {code}
> Then there would be NPE in constructType():
> {code}
> public JavaType constructType(TypeReference typeRef)
> {
> // 19-Oct-2015, tatu: Simpler variant like so should work
> return _fromAny(null, typeRef.getType(), EMPTY_BINDINGS);
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21221) Ineffective assertion in TestFromClientSide3#testMultiRowMutations

2018-09-24 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-21221:
---
Attachment: 21221.v12.txt

> Ineffective assertion in TestFromClientSide3#testMultiRowMutations
> --
>
> Key: HBASE-21221
> URL: https://issues.apache.org/jira/browse/HBASE-21221
> Project: HBase
>  Issue Type: Test
>    Reporter: Ted Yu
>    Assignee: Ted Yu
>Priority: Minor
> Attachments: 21221.v10.txt, 21221.v11.txt, 21221.v12.txt, 
> 21221.v7.txt, 21221.v8.txt, 21221.v9.txt
>
>
> Observed the following in 
> org.apache.hadoop.hbase.util.TestFromClientSide3WoUnsafe-output.txt :
> {code}
> Caused by: 
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> java.io.IOException: Timed out waiting for lock for row: ROW-1 in region 
> 089bdfa75f44d88e596479038a6da18b
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:5816)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion$4.lockRowsAndBuildMiniBatch(HRegion.java:7432)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutate(HRegion.java:4008)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3982)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.mutateRowsWithLocks(HRegion.java:7424)
>   at 
> org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.mutateRows(MultiRowMutationEndpoint.java:116)
>   at 
> org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos$MultiRowMutationService.callMethod(MultiRowMutationProtos.java:2266)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8182)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2481)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2463)
> ...
> Exception in thread "pool-678-thread-1" java.lang.AssertionError: This cp 
> should fail because the target lock is blocked by previous put
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.hadoop.hbase.client.TestFromClientSide3.lambda$testMultiRowMutations$7(TestFromClientSide3.java:861)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> {code}
> Here is related code:
> {code}
>   cpService.execute(() -> {
> ...
> if (!threw) {
>   // Can't call fail() earlier because the catch would eat it.
>   fail("This cp should fail because the target lock is blocked by 
> previous put");
> }
> {code}
> Since the fail() call is executed by the cpService, the assertion had no 
> bearing on the outcome of the test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AMBARI-14163) zookeeper session timeout for hbase should take zookeeper tickTime into account

2018-09-24 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/AMBARI-14163?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated AMBARI-14163:

Description: 
With tickTime=2000 in zoo.cfg, I tried to set zookeeper.session.timeout value 
of 1 min 40 seconds.
The change was accepted.

However, such timeout is not reachable (it is > 20 times tickTime).


Ambari should detect such scenario and warn user.

  was:
With tickTime=2000 in zoo.cfg, I tried to set zookeeper.session.timeout value 
of 1 min 40 seconds.
The change was accepted.

However, such timeout is not reachable (it is > 20 times tickTime).

Ambari should detect such scenario and warn user.


> zookeeper session timeout for hbase should take zookeeper tickTime into 
> account
> ---
>
> Key: AMBARI-14163
> URL: https://issues.apache.org/jira/browse/AMBARI-14163
> Project: Ambari
>  Issue Type: Bug
>Reporter: Ted Yu
>Priority: Major
>
> With tickTime=2000 in zoo.cfg, I tried to set zookeeper.session.timeout value 
> of 1 min 40 seconds.
> The change was accepted.
> However, such timeout is not reachable (it is > 20 times tickTime).
> Ambari should detect such scenario and warn user.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AMBARI-17346) Dependent components should be shutdown before stopping hdfs

2018-09-24 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/AMBARI-17346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated AMBARI-17346:

Description: 
Sometimes admin shuts down hdfs first, then hbase. 

By the time hbase is shutdown, no data can be persisted (including metadata). 
This results in large number of inconsistencies when hbase cluster is brought 
back up.


Before hdfs is shutdown, the components dependent on hdfs should be shutdown 
first.

  was:
Sometimes admin shuts down hdfs first, then hbase. 


By the time hbase is shutdown, no data can be persisted (including metadata). 
This results in large number of inconsistencies when hbase cluster is brought 
back up.


Before hdfs is shutdown, the components dependent on hdfs should be shutdown 
first.


> Dependent components should be shutdown before stopping hdfs
> 
>
> Key: AMBARI-17346
> URL: https://issues.apache.org/jira/browse/AMBARI-17346
> Project: Ambari
>  Issue Type: Bug
>    Reporter: Ted Yu
>Priority: Major
>
> Sometimes admin shuts down hdfs first, then hbase. 
> By the time hbase is shutdown, no data can be persisted (including metadata). 
> This results in large number of inconsistencies when hbase cluster is brought 
> back up.
> Before hdfs is shutdown, the components dependent on hdfs should be shutdown 
> first.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21219) Hbase incremental backup fails with null pointer exception

2018-09-24 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21219?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16626634#comment-16626634
 ] 

Ted Yu commented on HBASE-21219:


{code}
307   LOG.warn("Known hosts (from newestTimestamps):");
308   for (String s: newestTimestamps.keySet()) {
309 LOG.warn(s);
{code}
The above seems to be for debug purpose. Use LOG.debug ?

> Hbase incremental backup fails with null pointer exception
> --
>
> Key: HBASE-21219
> URL: https://issues.apache.org/jira/browse/HBASE-21219
> Project: HBase
>  Issue Type: Bug
>  Components: backuprestore
>Reporter: Vladimir Rodionov
>Assignee: Vladimir Rodionov
>Priority: Major
> Fix For: 3.0.0
>
> Attachments: HBASE-21219-v1.patch
>
>
> hbase backup create incremental hdfs:///bkpHbase_Test/bkpHbase_Test2 -t 
> bkpHbase_Test2 
> 2018-09-21 15:35:31,421 INFO [main] impl.TableBackupClient: Backup 
> backup_1537524313995 started at 1537524331419. 2018-09-21 15:35:31,454 INFO 
> [main] impl.IncrementalBackupManager: Execute roll log procedure for 
> incremental backup ... 2018-09-21 15:35:32,985 ERROR [main] 
> impl.TableBackupClient: Unexpected Exception : java.lang.NullPointerException 
> java.lang.NullPointerException at 
> org.apache.hadoop.hbase.backup.impl.IncrementalBackupManager.getLogFilesForNewBackup(IncrementalBackupManager.java:309)
>  at 
> org.apache.hadoop.hbase.backup.impl.IncrementalBackupManager.getIncrBackupLogFileMap(IncrementalBackupManager.java:103)
>  at 
> org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient.execute(IncrementalTableBackupClient.java:276)
>  at 
> org.apache.hadoop.hbase.backup.impl.BackupAdminImpl.backupTables(BackupAdminImpl.java:601)
>  at 
> org.apache.hadoop.hbase.backup.impl.BackupCommands$CreateCommand.execute(BackupCommands.java:347)
>  at 
> org.apache.hadoop.hbase.backup.BackupDriver.parseAndRun(BackupDriver.java:138)
>  at org.apache.hadoop.hbase.backup.BackupDriver.doWork(BackupDriver.java:171) 
> at org.apache.hadoop.hbase.backup.BackupDriver.run(BackupDriver.java:204) at 
> org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76) at 
> org.apache.hadoop.hbase.backup.BackupDriver.main(BackupDriver.java:179) 
> 2018-09-21 15:35:32,989 ERROR [main] impl.TableBackupClient: 
> BackupId=backup_1537524313995,startts=1537524331419,failedts=1537524332989,failedphase=PREPARE_INCREMENTAL,failedmessage=null
>  2018-09-21 15:35:57,167 ERROR [main] impl.TableBackupClient: Backup 
> backup_1537524313995 failed. 
> Backup session finished. Status: FAILURE 2018-09-21 15:35:57,175 ERROR [main] 
> backup.BackupDriver: Error running 
> command-line tool java.io.IOException: java.lang.NullPointerException at 
> org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient.execute(IncrementalTableBackupClient.java:281)
>  at 
> org.apache.hadoop.hbase.backup.impl.BackupAdminImpl.backupTables(BackupAdminImpl.java:601)
>  at 
> org.apache.hadoop.hbase.backup.impl.BackupCommands$CreateCommand.execute(BackupCommands.java:347)
>  at 
> org.apache.hadoop.hbase.backup.BackupDriver.parseAndRun(BackupDriver.java:138)
>  at org.apache.hadoop.hbase.backup.BackupDriver.doWork(BackupDriver.java:171) 
> at org.apache.hadoop.hbase.backup.BackupDriver.run(BackupDriver.java:204) at 
> org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76) at 
> org.apache.hadoop.hbase.backup.BackupDriver.main(BackupDriver.java:179) 
> Caused by: java.lang.NullPointerException at 
> org.apache.hadoop.hbase.backup.impl.IncrementalBackupManager.getLogFilesForNewBackup(IncrementalBackupManager.java:309)
>  at 
> org.apache.hadoop.hbase.backup.impl.IncrementalBackupManager.getIncrBackupLogFileMap(IncrementalBackupManager.java:103)
>  at 
> org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient.execute(IncrementalTableBackupClient.java:276)
>  ... 7 more



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21221) Ineffective assertion in TestFromClientSide3#testMultiRowMutations

2018-09-24 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-21221:
---
Attachment: 21221.v11.txt

> Ineffective assertion in TestFromClientSide3#testMultiRowMutations
> --
>
> Key: HBASE-21221
> URL: https://issues.apache.org/jira/browse/HBASE-21221
> Project: HBase
>  Issue Type: Test
>    Reporter: Ted Yu
>    Assignee: Ted Yu
>Priority: Minor
> Attachments: 21221.v10.txt, 21221.v11.txt, 21221.v7.txt, 
> 21221.v8.txt, 21221.v9.txt
>
>
> Observed the following in 
> org.apache.hadoop.hbase.util.TestFromClientSide3WoUnsafe-output.txt :
> {code}
> Caused by: 
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> java.io.IOException: Timed out waiting for lock for row: ROW-1 in region 
> 089bdfa75f44d88e596479038a6da18b
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:5816)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion$4.lockRowsAndBuildMiniBatch(HRegion.java:7432)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutate(HRegion.java:4008)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3982)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.mutateRowsWithLocks(HRegion.java:7424)
>   at 
> org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.mutateRows(MultiRowMutationEndpoint.java:116)
>   at 
> org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos$MultiRowMutationService.callMethod(MultiRowMutationProtos.java:2266)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8182)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2481)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2463)
> ...
> Exception in thread "pool-678-thread-1" java.lang.AssertionError: This cp 
> should fail because the target lock is blocked by previous put
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.hadoop.hbase.client.TestFromClientSide3.lambda$testMultiRowMutations$7(TestFromClientSide3.java:861)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> {code}
> Here is related code:
> {code}
>   cpService.execute(() -> {
> ...
> if (!threw) {
>   // Can't call fail() earlier because the catch would eat it.
>   fail("This cp should fail because the target lock is blocked by 
> previous put");
> }
> {code}
> Since the fail() call is executed by the cpService, the assertion had no 
> bearing on the outcome of the test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21221) Ineffective assertion in TestFromClientSide3#testMultiRowMutations

2018-09-24 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21221?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16626563#comment-16626563
 ] 

Ted Yu commented on HBASE-21221:


bq. the ex in RecordingEndpoint should be safely published

The custom endpoint is removed in patch v10. So the above is not needed.

> Ineffective assertion in TestFromClientSide3#testMultiRowMutations
> --
>
> Key: HBASE-21221
> URL: https://issues.apache.org/jira/browse/HBASE-21221
> Project: HBase
>  Issue Type: Test
>    Reporter: Ted Yu
>    Assignee: Ted Yu
>Priority: Minor
> Attachments: 21221.v10.txt, 21221.v7.txt, 21221.v8.txt, 21221.v9.txt
>
>
> Observed the following in 
> org.apache.hadoop.hbase.util.TestFromClientSide3WoUnsafe-output.txt :
> {code}
> Caused by: 
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> java.io.IOException: Timed out waiting for lock for row: ROW-1 in region 
> 089bdfa75f44d88e596479038a6da18b
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:5816)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion$4.lockRowsAndBuildMiniBatch(HRegion.java:7432)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutate(HRegion.java:4008)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3982)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.mutateRowsWithLocks(HRegion.java:7424)
>   at 
> org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.mutateRows(MultiRowMutationEndpoint.java:116)
>   at 
> org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos$MultiRowMutationService.callMethod(MultiRowMutationProtos.java:2266)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8182)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2481)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2463)
> ...
> Exception in thread "pool-678-thread-1" java.lang.AssertionError: This cp 
> should fail because the target lock is blocked by previous put
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.hadoop.hbase.client.TestFromClientSide3.lambda$testMultiRowMutations$7(TestFromClientSide3.java:861)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> {code}
> Here is related code:
> {code}
>   cpService.execute(() -> {
> ...
> if (!threw) {
>   // Can't call fail() earlier because the catch would eat it.
>   fail("This cp should fail because the target lock is blocked by 
> previous put");
> }
> {code}
> Since the fail() call is executed by the cpService, the assertion had no 
> bearing on the outcome of the test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21221) Ineffective assertion in TestFromClientSide3#testMultiRowMutations

2018-09-24 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-21221:
---
Attachment: 21221.v10.txt

> Ineffective assertion in TestFromClientSide3#testMultiRowMutations
> --
>
> Key: HBASE-21221
> URL: https://issues.apache.org/jira/browse/HBASE-21221
> Project: HBase
>  Issue Type: Test
>    Reporter: Ted Yu
>    Assignee: Ted Yu
>Priority: Minor
> Attachments: 21221.v10.txt, 21221.v7.txt, 21221.v8.txt, 21221.v9.txt
>
>
> Observed the following in 
> org.apache.hadoop.hbase.util.TestFromClientSide3WoUnsafe-output.txt :
> {code}
> Caused by: 
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> java.io.IOException: Timed out waiting for lock for row: ROW-1 in region 
> 089bdfa75f44d88e596479038a6da18b
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:5816)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion$4.lockRowsAndBuildMiniBatch(HRegion.java:7432)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutate(HRegion.java:4008)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3982)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.mutateRowsWithLocks(HRegion.java:7424)
>   at 
> org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.mutateRows(MultiRowMutationEndpoint.java:116)
>   at 
> org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos$MultiRowMutationService.callMethod(MultiRowMutationProtos.java:2266)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8182)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2481)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2463)
> ...
> Exception in thread "pool-678-thread-1" java.lang.AssertionError: This cp 
> should fail because the target lock is blocked by previous put
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.hadoop.hbase.client.TestFromClientSide3.lambda$testMultiRowMutations$7(TestFromClientSide3.java:861)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> {code}
> Here is related code:
> {code}
>   cpService.execute(() -> {
> ...
> if (!threw) {
>   // Can't call fail() earlier because the catch would eat it.
>   fail("This cp should fail because the target lock is blocked by 
> previous put");
> }
> {code}
> Since the fail() call is executed by the cpService, the assertion had no 
> bearing on the outcome of the test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (KYLIN-3590) Missing argument to exception message String in ModelController

2018-09-24 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3590?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16626212#comment-16626212
 ] 

Ted Yu commented on KYLIN-3590:
---

There is similar issue in ProjectController#saveProject

> Missing argument to exception message String in ModelController
> ---
>
> Key: KYLIN-3590
> URL: https://issues.apache.org/jira/browse/KYLIN-3590
> Project: Kylin
>  Issue Type: Bug
>    Reporter: Ted Yu
>Priority: Major
>
> Here is related code:
> {code}
> String.format(Locale.ROOT,
> "Invalid model name %s, only letters, numbers and 
> underscore " + "supported."),
> {code}
> There is no argument for '%s'



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3590) Missing argument to exception message String in ModelController

2018-09-24 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3590:
-

 Summary: Missing argument to exception message String in 
ModelController
 Key: KYLIN-3590
 URL: https://issues.apache.org/jira/browse/KYLIN-3590
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


Here is related code:
{code}
String.format(Locale.ROOT,
"Invalid model name %s, only letters, numbers and 
underscore " + "supported."),
{code}
There is no argument for '%s'



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3590) Missing argument to exception message String in ModelController

2018-09-24 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3590:
-

 Summary: Missing argument to exception message String in 
ModelController
 Key: KYLIN-3590
 URL: https://issues.apache.org/jira/browse/KYLIN-3590
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


Here is related code:
{code}
String.format(Locale.ROOT,
"Invalid model name %s, only letters, numbers and 
underscore " + "supported."),
{code}
There is no argument for '%s'



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3589) Different type is used to check presence in Collection in AclPermissionUtil

2018-09-24 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3589:
-

 Summary: Different type is used to check presence in Collection in 
AclPermissionUtil
 Key: KYLIN-3589
 URL: https://issues.apache.org/jira/browse/KYLIN-3589
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


Here is related code:
{code}
public static List transformAuthorities(Collection authorities) {
List ret = new ArrayList();
for (GrantedAuthority auth : authorities) {
if (!authorities.contains(auth.getAuthority())) {
{code}
auth.getAuthority() shouldn't be used to check presence in the Collection since 
it is not of GrantedAuthority type.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3589) Different type is used to check presence in Collection in AclPermissionUtil

2018-09-24 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3589:
-

 Summary: Different type is used to check presence in Collection in 
AclPermissionUtil
 Key: KYLIN-3589
 URL: https://issues.apache.org/jira/browse/KYLIN-3589
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


Here is related code:
{code}
public static List transformAuthorities(Collection authorities) {
List ret = new ArrayList();
for (GrantedAuthority auth : authorities) {
if (!authorities.contains(auth.getAuthority())) {
{code}
auth.getAuthority() shouldn't be used to check presence in the Collection since 
it is not of GrantedAuthority type.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3588) Potentially duplicate put to RemoveBlackoutRealizationsRule#filters map

2018-09-24 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3588:
-

 Summary: Potentially duplicate put to 
RemoveBlackoutRealizationsRule#filters map
 Key: KYLIN-3588
 URL: https://issues.apache.org/jira/browse/KYLIN-3588
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


Here is related code:
{code}
IRealizationFilter filter = filters.get(conf);
if (filter == null) {
try {
Class clz = 
ClassUtil.forName(conf.getQueryRealizationFilter(), IRealizationFilter.class);
filter = 
clz.getConstructor(KylinConfig.class).newInstance(conf);
} catch (Exception e) {
throw new RuntimeException(e);
}
filters.put(conf, filter);
{code}
Two threads may find out there is no filter for conf, they would proceed with 
creation of IRealizationFilter instances and try to put into the map.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3588) Potentially duplicate put to RemoveBlackoutRealizationsRule#filters map

2018-09-24 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3588:
-

 Summary: Potentially duplicate put to 
RemoveBlackoutRealizationsRule#filters map
 Key: KYLIN-3588
 URL: https://issues.apache.org/jira/browse/KYLIN-3588
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


Here is related code:
{code}
IRealizationFilter filter = filters.get(conf);
if (filter == null) {
try {
Class clz = 
ClassUtil.forName(conf.getQueryRealizationFilter(), IRealizationFilter.class);
filter = 
clz.getConstructor(KylinConfig.class).newInstance(conf);
} catch (Exception e) {
throw new RuntimeException(e);
}
filters.put(conf, filter);
{code}
Two threads may find out there is no filter for conf, they would proceed with 
creation of IRealizationFilter instances and try to put into the map.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3587) Null value is used in IllegalStateException message

2018-09-24 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3587:
-

 Summary: Null value is used in IllegalStateException message
 Key: KYLIN-3587
 URL: https://issues.apache.org/jira/browse/KYLIN-3587
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


In ProjectManager :
{code}
TableDesc table = metaMgr.getTableDesc(tableId, projectName);
if (table == null) {
throw new IllegalStateException("Cannot find table '" + 
table + "' in metadata manager");
{code}
Since table is null, the exception message is not helpful.
It seems including tableId in the message is better.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3587) Null value is used in IllegalStateException message

2018-09-24 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3587:
-

 Summary: Null value is used in IllegalStateException message
 Key: KYLIN-3587
 URL: https://issues.apache.org/jira/browse/KYLIN-3587
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


In ProjectManager :
{code}
TableDesc table = metaMgr.getTableDesc(tableId, projectName);
if (table == null) {
throw new IllegalStateException("Cannot find table '" + 
table + "' in metadata manager");
{code}
Since table is null, the exception message is not helpful.
It seems including tableId in the message is better.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (AMBARI-13109) Verify that max logs parameter wouldn't cause premature region flush

2018-09-24 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/AMBARI-13109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16120333#comment-16120333
 ] 

Ted Yu edited comment on AMBARI-13109 at 9/24/18 3:27 PM:
--

The max logs parameter plays an important role in preventing fragmentation in 
memstore .


was (Author: yuzhih...@gmail.com):
The max logs parameter plays an important role in preventing fragmentation in 
memstore.

> Verify that max logs parameter wouldn't cause premature region flush
> 
>
> Key: AMBARI-13109
> URL: https://issues.apache.org/jira/browse/AMBARI-13109
> Project: Ambari
>  Issue Type: Improvement
>    Reporter: Ted Yu
>Priority: Major
>
> For hbase, the following relationship should be maintained:
> dfs.blocksize * 0.95 *  hbase.regionserver.maxlogs (default 32) > 
> hbase.regionserver.global.memstore.upperLimit * HBASE_HEAPSIZE
> The rationale is to avoid premature region flush due to log roll which is 
> triggered by insufficiently high max logs.
> When user changes the values for max logs, memstore upper limit or max heap 
> size, we should verify that the above relationship is satisfied.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (HBASE-16627) AssignmentManager#isDisabledorDisablingRegionInRIT should check whether table exists

2018-09-24 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-16627?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu resolved HBASE-16627.

Resolution: Later

> AssignmentManager#isDisabledorDisablingRegionInRIT should check whether table 
> exists
> 
>
> Key: HBASE-16627
> URL: https://issues.apache.org/jira/browse/HBASE-16627
> Project: HBase
>  Issue Type: Bug
>    Reporter: Ted Yu
>Assignee: Stephen Yuan Jiang
>Priority: Minor
>
> [~stack] first reported this issue when he played with backup feature.
> The following exception can be observed in backup unit tests:
> {code}
> 2016-09-13 16:21:57,661 ERROR [ProcedureExecutor-3] 
> master.TableStateManager(134): Unable to get table hbase:backup state
> org.apache.hadoop.hbase.TableNotFoundException: hbase:backup
> at 
> org.apache.hadoop.hbase.master.TableStateManager.getTableState(TableStateManager.java:174)
> at 
> org.apache.hadoop.hbase.master.TableStateManager.isTableState(TableStateManager.java:131)
> at 
> org.apache.hadoop.hbase.master.AssignmentManager.isDisabledorDisablingRegionInRIT(AssignmentManager.java:1221)
> at 
> org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:739)
> at 
> org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:1567)
> at 
> org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:1546)
> at 
> org.apache.hadoop.hbase.util.ModifyRegionUtils.assignRegions(ModifyRegionUtils.java:254)
> at 
> org.apache.hadoop.hbase.master.procedure.CreateTableProcedure.assignRegions(CreateTableProcedure.java:430)
> at 
> org.apache.hadoop.hbase.master.procedure.CreateTableProcedure.executeFromState(CreateTableProcedure.java:127)
> at 
> org.apache.hadoop.hbase.master.procedure.CreateTableProcedure.executeFromState(CreateTableProcedure.java:57)
> at 
> org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:119)
> at 
> org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:452)
> at 
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1066)
> at 
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:855)
> at 
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:808)
> {code}
> AssignmentManager#isDisabledorDisablingRegionInRIT should take table 
> existence into account.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (HBASE-16627) AssignmentManager#isDisabledorDisablingRegionInRIT should check whether table exists

2018-09-24 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-16627?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu resolved HBASE-16627.

Resolution: Later

> AssignmentManager#isDisabledorDisablingRegionInRIT should check whether table 
> exists
> 
>
> Key: HBASE-16627
> URL: https://issues.apache.org/jira/browse/HBASE-16627
> Project: HBase
>  Issue Type: Bug
>    Reporter: Ted Yu
>Assignee: Stephen Yuan Jiang
>Priority: Minor
>
> [~stack] first reported this issue when he played with backup feature.
> The following exception can be observed in backup unit tests:
> {code}
> 2016-09-13 16:21:57,661 ERROR [ProcedureExecutor-3] 
> master.TableStateManager(134): Unable to get table hbase:backup state
> org.apache.hadoop.hbase.TableNotFoundException: hbase:backup
> at 
> org.apache.hadoop.hbase.master.TableStateManager.getTableState(TableStateManager.java:174)
> at 
> org.apache.hadoop.hbase.master.TableStateManager.isTableState(TableStateManager.java:131)
> at 
> org.apache.hadoop.hbase.master.AssignmentManager.isDisabledorDisablingRegionInRIT(AssignmentManager.java:1221)
> at 
> org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:739)
> at 
> org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:1567)
> at 
> org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:1546)
> at 
> org.apache.hadoop.hbase.util.ModifyRegionUtils.assignRegions(ModifyRegionUtils.java:254)
> at 
> org.apache.hadoop.hbase.master.procedure.CreateTableProcedure.assignRegions(CreateTableProcedure.java:430)
> at 
> org.apache.hadoop.hbase.master.procedure.CreateTableProcedure.executeFromState(CreateTableProcedure.java:127)
> at 
> org.apache.hadoop.hbase.master.procedure.CreateTableProcedure.executeFromState(CreateTableProcedure.java:57)
> at 
> org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:119)
> at 
> org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:452)
> at 
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1066)
> at 
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:855)
> at 
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:808)
> {code}
> AssignmentManager#isDisabledorDisablingRegionInRIT should take table 
> existence into account.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-18580) Allow gdb to attach to selected process in docker

2018-09-24 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-18580?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-18580:
---
Labels: debugger  (was: )

> Allow gdb to attach to selected process in docker
> -
>
> Key: HBASE-18580
> URL: https://issues.apache.org/jira/browse/HBASE-18580
> Project: HBase
>  Issue Type: Sub-task
>    Reporter: Ted Yu
>Priority: Major
>  Labels: debugger
>
> In the current docker image, if you want to attach gdb to a process, you 
> would see:
> bq. ptrace: Operation not permitted
> We should provide better support for gdb in docker.
> This article is a start:
> https://thirld.com/blog/2016/08/15/war-stories-debugging-julia-gdb-docker/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (KYLIN-3561) Upgrade to JUnit 5

2018-09-23 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3561?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated KYLIN-3561:
--
Description: 
JUnit 5 brings multiple useful features so tests are easier to read and write.

We can bump up the dependency version and create new tests with JUnit 5 
features.


Relevant features of JUnit 5: dynamic test, nested tests, parameterized tests
https://twitter.com/nipafx/status/1027095088059559936

  was:
JUnit 5 brings multiple useful features so tests are easier to read and write.

We can bump up the dependency version and create new tests with JUnit 5 
features.

Relevant features of JUnit 5: dynamic test, nested tests, parameterized tests
https://twitter.com/nipafx/status/1027095088059559936


> Upgrade to JUnit 5
> --
>
> Key: KYLIN-3561
> URL: https://issues.apache.org/jira/browse/KYLIN-3561
> Project: Kylin
>  Issue Type: Task
>    Reporter: Ted Yu
>Priority: Major
>
> JUnit 5 brings multiple useful features so tests are easier to read and write.
> We can bump up the dependency version and create new tests with JUnit 5 
> features.
> Relevant features of JUnit 5: dynamic test, nested tests, parameterized tests
> https://twitter.com/nipafx/status/1027095088059559936



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-18547) Provide better cleanup of WAL related entries in backup table

2018-09-23 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-18547?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-18547:
---
Labels: backup  (was: )

> Provide better cleanup of WAL related entries in backup table
> -
>
> Key: HBASE-18547
> URL: https://issues.apache.org/jira/browse/HBASE-18547
> Project: HBase
>  Issue Type: Bug
>    Reporter: Ted Yu
>Priority: Major
>  Labels: backup
>
> I was discussing the design around incremental backup with Vladimir.
> Currently if WAL file is recorded in backup table, BackupLogCleaner would 
> signal that the file can be cleaned up.
> However, cleaner chore doesn't notify BackupLogCleaner whether the file is 
> actually deleted.
> This means that potentially large number of WAL file entries would stay in 
> backup table.
> This issue is to investigate better cleanup strategy for the WAL entries.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-18610) Provide capability to activate chaos monkey

2018-09-23 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-18610?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-18610:
---
Attachment: (was: 18610.v1.txt)

> Provide capability to activate chaos monkey
> ---
>
> Key: HBASE-18610
> URL: https://issues.apache.org/jira/browse/HBASE-18610
> Project: HBase
>  Issue Type: Sub-task
>    Reporter: Ted Yu
>Priority: Major
>  Labels: fault-tolerance
> Attachments: 18610.v4.txt, 18610.v5.txt, 18610.v6.txt, 18610.v7.txt, 
> 18610.v8.txt, 18610.v9.txt
>
>
> Currently load-client runs against a cluster where region servers are stable.
> We need to introduce chaos monkey so that wider coverage for read path is 
> exercised.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (GORA-545) Boxing/unboxing to parse a primitive is suboptimal

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created GORA-545:
---

 Summary: Boxing/unboxing to parse a primitive is suboptimal
 Key: GORA-545
 URL: https://issues.apache.org/jira/browse/GORA-545
 Project: Apache Gora
  Issue Type: Bug
Reporter: Ted Yu


An example comes from InfinispanStore.java :
{code}
  splitSize = Integer.valueOf(
  properties.getProperty( BUFFER_LIMIT_READ_NAME,
{code}
splitSize, an int, should be assigned by calling Integer.parseInt()



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3586) Boxing/unboxing to parse a primitive is suboptimal

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3586:
-

 Summary: Boxing/unboxing to parse a primitive is suboptimal
 Key: KYLIN-3586
 URL: https://issues.apache.org/jira/browse/KYLIN-3586
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


An example is from HBaseLookupRowEncoder:
{code}
int valIdx = Integer.valueOf(Bytes.toString(qualifier));
{code}
valueOf returns an Integer object which would be unboxed and assigned to valIdx.

Integer.parseInt() should be used instead.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3586) Boxing/unboxing to parse a primitive is suboptimal

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3586:
-

 Summary: Boxing/unboxing to parse a primitive is suboptimal
 Key: KYLIN-3586
 URL: https://issues.apache.org/jira/browse/KYLIN-3586
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


An example is from HBaseLookupRowEncoder:
{code}
int valIdx = Integer.valueOf(Bytes.toString(qualifier));
{code}
valueOf returns an Integer object which would be unboxed and assigned to valIdx.

Integer.parseInt() should be used instead.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3585) Ineffective declaration of volatile field in SparkFactDistinct

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3585:
-

 Summary: Ineffective declaration of volatile field in 
SparkFactDistinct
 Key: KYLIN-3585
 URL: https://issues.apache.org/jira/browse/KYLIN-3585
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


In CuboidStatCalculator :
{code}
private volatile HLLCounter[] cuboidsHLL;
{code}
Though the array is declared volatile, the array elements are not.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3585) Ineffective declaration of volatile field in SparkFactDistinct

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3585:
-

 Summary: Ineffective declaration of volatile field in 
SparkFactDistinct
 Key: KYLIN-3585
 URL: https://issues.apache.org/jira/browse/KYLIN-3585
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


In CuboidStatCalculator :
{code}
private volatile HLLCounter[] cuboidsHLL;
{code}
Though the array is declared volatile, the array elements are not.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3584) Incorrect lazy initialization of DefaultScheduler.INSTANCE

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3584:
-

 Summary: Incorrect lazy initialization of DefaultScheduler.INSTANCE
 Key: KYLIN-3584
 URL: https://issues.apache.org/jira/browse/KYLIN-3584
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


Here is related code:
{code}
public static DefaultScheduler getInstance() {
if (INSTANCE == null) {
INSTANCE = createInstance();
{code}
The check against may not be effective.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3584) Incorrect lazy initialization of DefaultScheduler.INSTANCE

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3584:
-

 Summary: Incorrect lazy initialization of DefaultScheduler.INSTANCE
 Key: KYLIN-3584
 URL: https://issues.apache.org/jira/browse/KYLIN-3584
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


Here is related code:
{code}
public static DefaultScheduler getInstance() {
if (INSTANCE == null) {
INSTANCE = createInstance();
{code}
The check against may not be effective.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3583) Integer multiplication is performed where Long result is expected

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3583:
-

 Summary: Integer multiplication is performed where Long result is 
expected
 Key: KYLIN-3583
 URL: https://issues.apache.org/jira/browse/KYLIN-3583
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


In RecordEventTimeDetail :
{code}
long timeStampForWeekBegin = timeStamp;
timeStampForWeekBegin -= 360 * 24 * 
(calendar.get(Calendar.DAY_OF_WEEK) - 1);
{code}
Casting to long should be done first before multiplication.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3583) Integer multiplication is performed where Long result is expected

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3583:
-

 Summary: Integer multiplication is performed where Long result is 
expected
 Key: KYLIN-3583
 URL: https://issues.apache.org/jira/browse/KYLIN-3583
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


In RecordEventTimeDetail :
{code}
long timeStampForWeekBegin = timeStamp;
timeStampForWeekBegin -= 360 * 24 * 
(calendar.get(Calendar.DAY_OF_WEEK) - 1);
{code}
Casting to long should be done first before multiplication.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3582) String comparison should not use == in RecordEvent

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3582:
-

 Summary: String comparison should not use == in RecordEvent
 Key: KYLIN-3582
 URL: https://issues.apache.org/jira/browse/KYLIN-3582
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


Here is related code:
{code}
for (RecordReserveKeyEnum reserveKey : 
RecordReserveKeyEnum.values()) {
if (reserveKey.reserveKey == key) {
{code}
equals() should be used to compare Strings.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3582) String comparison should not use == in RecordEvent

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3582:
-

 Summary: String comparison should not use == in RecordEvent
 Key: KYLIN-3582
 URL: https://issues.apache.org/jira/browse/KYLIN-3582
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


Here is related code:
{code}
for (RecordReserveKeyEnum reserveKey : 
RecordReserveKeyEnum.values()) {
if (reserveKey.reserveKey == key) {
{code}
equals() should be used to compare Strings.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3581) compareTo method should be paired with equals method

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3581:
-

 Summary: compareTo method should be paired with equals method
 Key: KYLIN-3581
 URL: https://issues.apache.org/jira/browse/KYLIN-3581
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


DoggedCubeBuilder$MergeSlot and InMemCubeBuilder$CuboidTask define compareTo 
method.

They should define equals method as well.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3581) compareTo method should be paired with equals method

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3581:
-

 Summary: compareTo method should be paired with equals method
 Key: KYLIN-3581
 URL: https://issues.apache.org/jira/browse/KYLIN-3581
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


DoggedCubeBuilder$MergeSlot and InMemCubeBuilder$CuboidTask define compareTo 
method.

They should define equals method as well.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3580) CuboidStatsUtil#complementRowCountForMandatoryCuboids uses entrySet to add elements

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3580:
-

 Summary: CuboidStatsUtil#complementRowCountForMandatoryCuboids 
uses entrySet to add elements
 Key: KYLIN-3580
 URL: https://issues.apache.org/jira/browse/KYLIN-3580
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


Here is related code:
{code}
sortedStatsSet.addAll(statistics.entrySet());
{code}
The entrySet() method is allowed to return a view of the underlying Map in 
which a single Entry object is reused and returned during the iteration.
When iterating through such a Map, the Entry value is only valid until you 
advance to the next iteration.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3579) entrySet iterator should be used in BPUSCalculator

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3579:
-

 Summary: entrySet iterator should be used in BPUSCalculator 
 Key: KYLIN-3579
 URL: https://issues.apache.org/jira/browse/KYLIN-3579
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


BPUSCalculator.calculateBenefitTotal and BPUSCalculator.initCuboidAggCostMap 
use keySet iterator where entrySet iterator should be used.
{code}
for (Long cuboid : cuboidAggCostMapCopy.keySet()) {
if (cuboidAggCostMapCopy.get(cuboid) < 
processCuboidAggCostMap.get(cuboid)) {
{code}
CubeCodeSystem.newMetricsAggregators has similar issue.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3578) Do not synchronize on the intrinsic locks of high-level concurrency objects

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3578:
-

 Summary: Do not synchronize on the intrinsic locks of high-level 
concurrency objects
 Key: KYLIN-3578
 URL: https://issues.apache.org/jira/browse/KYLIN-3578
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


>From 
>core-common/src/main/java/org/apache/kylin/common/util/MemoryBudgetController.java
> :
{code}
private final ReentrantLock lock = new ReentrantLock();
...
synchronized (lock) {
{code}
See the following for why such practice is to be avoided:

https://wiki.sei.cmu.edu/confluence/display/java/LCK03-J.+Do+not+synchronize+on+the+intrinsic+locks+of+high-level+concurrency+objects



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3578) Do not synchronize on the intrinsic locks of high-level concurrency objects

2018-09-23 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3578:
-

 Summary: Do not synchronize on the intrinsic locks of high-level 
concurrency objects
 Key: KYLIN-3578
 URL: https://issues.apache.org/jira/browse/KYLIN-3578
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


>From 
>core-common/src/main/java/org/apache/kylin/common/util/MemoryBudgetController.java
> :
{code}
private final ReentrantLock lock = new ReentrantLock();
...
synchronized (lock) {
{code}
See the following for why such practice is to be avoided:

https://wiki.sei.cmu.edu/confluence/display/java/LCK03-J.+Do+not+synchronize+on+the+intrinsic+locks+of+high-level+concurrency+objects



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21221) Ineffective assertion in TestFromClientSide3#testMultiRowMutations

2018-09-23 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-21221:
---
Attachment: 21221.v9.txt

> Ineffective assertion in TestFromClientSide3#testMultiRowMutations
> --
>
> Key: HBASE-21221
> URL: https://issues.apache.org/jira/browse/HBASE-21221
> Project: HBase
>  Issue Type: Test
>    Reporter: Ted Yu
>    Assignee: Ted Yu
>Priority: Minor
> Attachments: 21221.v7.txt, 21221.v8.txt, 21221.v9.txt
>
>
> Observed the following in 
> org.apache.hadoop.hbase.util.TestFromClientSide3WoUnsafe-output.txt :
> {code}
> Caused by: 
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> java.io.IOException: Timed out waiting for lock for row: ROW-1 in region 
> 089bdfa75f44d88e596479038a6da18b
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:5816)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion$4.lockRowsAndBuildMiniBatch(HRegion.java:7432)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutate(HRegion.java:4008)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3982)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.mutateRowsWithLocks(HRegion.java:7424)
>   at 
> org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.mutateRows(MultiRowMutationEndpoint.java:116)
>   at 
> org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos$MultiRowMutationService.callMethod(MultiRowMutationProtos.java:2266)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8182)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2481)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2463)
> ...
> Exception in thread "pool-678-thread-1" java.lang.AssertionError: This cp 
> should fail because the target lock is blocked by previous put
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.hadoop.hbase.client.TestFromClientSide3.lambda$testMultiRowMutations$7(TestFromClientSide3.java:861)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> {code}
> Here is related code:
> {code}
>   cpService.execute(() -> {
> ...
> if (!threw) {
>   // Can't call fail() earlier because the catch would eat it.
>   fail("This cp should fail because the target lock is blocked by 
> previous put");
> }
> {code}
> Since the fail() call is executed by the cpService, the assertion had no 
> bearing on the outcome of the test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Issue Comment Deleted] (FLINK-8037) Missing cast in integer arithmetic in TransactionalIdsGenerator#generateIdsToAbort

2018-09-23 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-8037?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated FLINK-8037:
--
Comment: was deleted

(was: Please rebase PR.)

> Missing cast in integer arithmetic in 
> TransactionalIdsGenerator#generateIdsToAbort
> --
>
> Key: FLINK-8037
> URL: https://issues.apache.org/jira/browse/FLINK-8037
> Project: Flink
>  Issue Type: Bug
>    Reporter: Ted Yu
>Assignee: Greg Hogan
>Priority: Minor
>  Labels: kafka, kafka-connect
>
> {code}
>   public Set generateIdsToAbort() {
> Set idsToAbort = new HashSet<>();
> for (int i = 0; i < safeScaleDownFactor; i++) {
>   idsToAbort.addAll(generateIdsToUse(i * poolSize * 
> totalNumberOfSubtasks));
> {code}
> The operands are integers where generateIdsToUse() expects long parameter.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21221) Ineffective assertion in TestFromClientSide3#testMultiRowMutations

2018-09-23 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-21221:
---
Assignee: Ted Yu
  Status: Patch Available  (was: Open)

> Ineffective assertion in TestFromClientSide3#testMultiRowMutations
> --
>
> Key: HBASE-21221
> URL: https://issues.apache.org/jira/browse/HBASE-21221
> Project: HBase
>  Issue Type: Test
>    Reporter: Ted Yu
>    Assignee: Ted Yu
>Priority: Minor
> Attachments: 21221.v7.txt, 21221.v8.txt
>
>
> Observed the following in 
> org.apache.hadoop.hbase.util.TestFromClientSide3WoUnsafe-output.txt :
> {code}
> Caused by: 
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> java.io.IOException: Timed out waiting for lock for row: ROW-1 in region 
> 089bdfa75f44d88e596479038a6da18b
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:5816)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion$4.lockRowsAndBuildMiniBatch(HRegion.java:7432)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutate(HRegion.java:4008)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3982)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.mutateRowsWithLocks(HRegion.java:7424)
>   at 
> org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.mutateRows(MultiRowMutationEndpoint.java:116)
>   at 
> org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos$MultiRowMutationService.callMethod(MultiRowMutationProtos.java:2266)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8182)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2481)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2463)
> ...
> Exception in thread "pool-678-thread-1" java.lang.AssertionError: This cp 
> should fail because the target lock is blocked by previous put
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.hadoop.hbase.client.TestFromClientSide3.lambda$testMultiRowMutations$7(TestFromClientSide3.java:861)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> {code}
> Here is related code:
> {code}
>   cpService.execute(() -> {
> ...
> if (!threw) {
>   // Can't call fail() earlier because the catch would eat it.
>   fail("This cp should fail because the target lock is blocked by 
> previous put");
> }
> {code}
> Since the fail() call is executed by the cpService, the assertion had no 
> bearing on the outcome of the test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21221) Ineffective assertion in TestFromClientSide3#testMultiRowMutations

2018-09-23 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-21221:
---
Attachment: 21221.v8.txt

> Ineffective assertion in TestFromClientSide3#testMultiRowMutations
> --
>
> Key: HBASE-21221
> URL: https://issues.apache.org/jira/browse/HBASE-21221
> Project: HBase
>  Issue Type: Test
>    Reporter: Ted Yu
>Priority: Minor
> Attachments: 21221.v7.txt, 21221.v8.txt
>
>
> Observed the following in 
> org.apache.hadoop.hbase.util.TestFromClientSide3WoUnsafe-output.txt :
> {code}
> Caused by: 
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> java.io.IOException: Timed out waiting for lock for row: ROW-1 in region 
> 089bdfa75f44d88e596479038a6da18b
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:5816)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion$4.lockRowsAndBuildMiniBatch(HRegion.java:7432)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutate(HRegion.java:4008)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3982)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.mutateRowsWithLocks(HRegion.java:7424)
>   at 
> org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.mutateRows(MultiRowMutationEndpoint.java:116)
>   at 
> org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos$MultiRowMutationService.callMethod(MultiRowMutationProtos.java:2266)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8182)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2481)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2463)
> ...
> Exception in thread "pool-678-thread-1" java.lang.AssertionError: This cp 
> should fail because the target lock is blocked by previous put
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.hadoop.hbase.client.TestFromClientSide3.lambda$testMultiRowMutations$7(TestFromClientSide3.java:861)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> {code}
> Here is related code:
> {code}
>   cpService.execute(() -> {
> ...
> if (!threw) {
>   // Can't call fail() earlier because the catch would eat it.
>   fail("This cp should fail because the target lock is blocked by 
> previous put");
> }
> {code}
> Since the fail() call is executed by the cpService, the assertion had no 
> bearing on the outcome of the test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21221) Ineffective assertion in TestFromClientSide3#testMultiRowMutations

2018-09-23 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21221?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16625060#comment-16625060
 ] 

Ted Yu commented on HBASE-21221:


Patch v8 doesn't wait for the captured exception to be non-null.
Instead, check captured exception toward the end of the test.

> Ineffective assertion in TestFromClientSide3#testMultiRowMutations
> --
>
> Key: HBASE-21221
> URL: https://issues.apache.org/jira/browse/HBASE-21221
> Project: HBase
>  Issue Type: Test
>    Reporter: Ted Yu
>Priority: Minor
> Attachments: 21221.v7.txt, 21221.v8.txt
>
>
> Observed the following in 
> org.apache.hadoop.hbase.util.TestFromClientSide3WoUnsafe-output.txt :
> {code}
> Caused by: 
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> java.io.IOException: Timed out waiting for lock for row: ROW-1 in region 
> 089bdfa75f44d88e596479038a6da18b
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:5816)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion$4.lockRowsAndBuildMiniBatch(HRegion.java:7432)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutate(HRegion.java:4008)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3982)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.mutateRowsWithLocks(HRegion.java:7424)
>   at 
> org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.mutateRows(MultiRowMutationEndpoint.java:116)
>   at 
> org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos$MultiRowMutationService.callMethod(MultiRowMutationProtos.java:2266)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8182)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2481)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2463)
> ...
> Exception in thread "pool-678-thread-1" java.lang.AssertionError: This cp 
> should fail because the target lock is blocked by previous put
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.hadoop.hbase.client.TestFromClientSide3.lambda$testMultiRowMutations$7(TestFromClientSide3.java:861)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> {code}
> Here is related code:
> {code}
>   cpService.execute(() -> {
> ...
> if (!threw) {
>   // Can't call fail() earlier because the catch would eat it.
>   fail("This cp should fail because the target lock is blocked by 
> previous put");
> }
> {code}
> Since the fail() call is executed by the cpService, the assertion had no 
> bearing on the outcome of the test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21221) Ineffective assertion in TestFromClientSide3#testMultiRowMutations

2018-09-22 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21221?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16624975#comment-16624975
 ] 

Ted Yu commented on HBASE-21221:


By subclassing MultiRowMutationEndpoint, I was able to capture the exception 
encountered by mutateRows.

However, the testMultiRowMutations tears down before the captured exception can 
be examined ...

> Ineffective assertion in TestFromClientSide3#testMultiRowMutations
> --
>
> Key: HBASE-21221
> URL: https://issues.apache.org/jira/browse/HBASE-21221
> Project: HBase
>  Issue Type: Test
>    Reporter: Ted Yu
>Priority: Minor
> Attachments: 21221.v7.txt
>
>
> Observed the following in 
> org.apache.hadoop.hbase.util.TestFromClientSide3WoUnsafe-output.txt :
> {code}
> Caused by: 
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> java.io.IOException: Timed out waiting for lock for row: ROW-1 in region 
> 089bdfa75f44d88e596479038a6da18b
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:5816)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion$4.lockRowsAndBuildMiniBatch(HRegion.java:7432)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutate(HRegion.java:4008)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3982)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.mutateRowsWithLocks(HRegion.java:7424)
>   at 
> org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.mutateRows(MultiRowMutationEndpoint.java:116)
>   at 
> org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos$MultiRowMutationService.callMethod(MultiRowMutationProtos.java:2266)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8182)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2481)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2463)
> ...
> Exception in thread "pool-678-thread-1" java.lang.AssertionError: This cp 
> should fail because the target lock is blocked by previous put
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.hadoop.hbase.client.TestFromClientSide3.lambda$testMultiRowMutations$7(TestFromClientSide3.java:861)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> {code}
> Here is related code:
> {code}
>   cpService.execute(() -> {
> ...
> if (!threw) {
>   // Can't call fail() earlier because the catch would eat it.
>   fail("This cp should fail because the target lock is blocked by 
> previous put");
> }
> {code}
> Since the fail() call is executed by the cpService, the assertion had no 
> bearing on the outcome of the test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21221) Ineffective assertion in TestFromClientSide3#testMultiRowMutations

2018-09-22 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-21221:
---
Attachment: 21221.v7.txt

> Ineffective assertion in TestFromClientSide3#testMultiRowMutations
> --
>
> Key: HBASE-21221
> URL: https://issues.apache.org/jira/browse/HBASE-21221
> Project: HBase
>  Issue Type: Test
>    Reporter: Ted Yu
>Priority: Minor
> Attachments: 21221.v7.txt
>
>
> Observed the following in 
> org.apache.hadoop.hbase.util.TestFromClientSide3WoUnsafe-output.txt :
> {code}
> Caused by: 
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> java.io.IOException: Timed out waiting for lock for row: ROW-1 in region 
> 089bdfa75f44d88e596479038a6da18b
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:5816)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion$4.lockRowsAndBuildMiniBatch(HRegion.java:7432)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutate(HRegion.java:4008)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3982)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.mutateRowsWithLocks(HRegion.java:7424)
>   at 
> org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.mutateRows(MultiRowMutationEndpoint.java:116)
>   at 
> org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos$MultiRowMutationService.callMethod(MultiRowMutationProtos.java:2266)
>   at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8182)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2481)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2463)
> ...
> Exception in thread "pool-678-thread-1" java.lang.AssertionError: This cp 
> should fail because the target lock is blocked by previous put
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.hadoop.hbase.client.TestFromClientSide3.lambda$testMultiRowMutations$7(TestFromClientSide3.java:861)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> {code}
> Here is related code:
> {code}
>   cpService.execute(() -> {
> ...
> if (!threw) {
>   // Can't call fail() earlier because the catch would eat it.
>   fail("This cp should fail because the target lock is blocked by 
> previous put");
> }
> {code}
> Since the fail() call is executed by the cpService, the assertion had no 
> bearing on the outcome of the test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-9824) Support IPv6 literal

2018-09-22 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-9824?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated FLINK-9824:
--
Description: 
Currently we use colon as separator when parsing host and port.


We should support the usage of IPv6 literals in parsing.

  was:
Currently we use colon as separator when parsing host and port.

We should support the usage of IPv6 literals in parsing.


> Support IPv6 literal
> 
>
> Key: FLINK-9824
> URL: https://issues.apache.org/jira/browse/FLINK-9824
> Project: Flink
>  Issue Type: Bug
>  Components: Network
>    Reporter: Ted Yu
>Assignee: vinoyang
>Priority: Minor
>
> Currently we use colon as separator when parsing host and port.
> We should support the usage of IPv6 literals in parsing.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-20654) Expose regions in transition thru JMX

2018-09-22 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-20654?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-20654:
---
Resolution: Won't Fix
Status: Resolved  (was: Patch Available)

> Expose regions in transition thru JMX
> -
>
> Key: HBASE-20654
> URL: https://issues.apache.org/jira/browse/HBASE-20654
> Project: HBase
>  Issue Type: Improvement
>    Reporter: Ted Yu
>Assignee: liubangchen
>Priority: Major
> Attachments: 1.png, HBASE-20654-1.patch, HBASE-20654-2.patch, 
> HBASE-20654-3.patch
>
>
> Currently only the count of regions in transition is exposed thru JMX.
> Here is a sample snippet of the /jmx output:
> {code}
> {
>   "beans" : [ {
> ...
>   }, {
> "name" : "Hadoop:service=HBase,name=Master,sub=AssignmentManager",
> "modelerType" : "Master,sub=AssignmentManager",
> "tag.Context" : "master",
> ...
> "ritCount" : 3
> {code}
> It would be desirable to expose region name, state for the regions in 
> transition as well.
> We can place configurable upper bound on the number of entries returned in 
> case there're a lot of regions in transition.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21141) Enable MOB in backup / restore test involving incremental backup

2018-09-22 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21141?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-21141:
---
Labels: mob  (was: )

> Enable MOB in backup / restore test involving incremental backup
> 
>
> Key: HBASE-21141
> URL: https://issues.apache.org/jira/browse/HBASE-21141
> Project: HBase
>  Issue Type: Test
>  Components: backuprestore
>    Reporter: Ted Yu
>Priority: Major
>  Labels: mob
>
> Currently we only have one test (TestRemoteBackup) where MOB feature is 
> enabled. The test only performs full backup.
> This issue is to enable MOB in backup / restore test(s) involving incremental 
> backup.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-21221) Ineffective assertion in TestFromClientSide3#testMultiRowMutations

2018-09-22 Thread Ted Yu (JIRA)
Ted Yu created HBASE-21221:
--

 Summary: Ineffective assertion in 
TestFromClientSide3#testMultiRowMutations
 Key: HBASE-21221
 URL: https://issues.apache.org/jira/browse/HBASE-21221
 Project: HBase
  Issue Type: Test
Reporter: Ted Yu


Observed the following in 
org.apache.hadoop.hbase.util.TestFromClientSide3WoUnsafe-output.txt :
{code}
Caused by: 
org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
java.io.IOException: Timed out waiting for lock for row: ROW-1 in region 
089bdfa75f44d88e596479038a6da18b
  at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:5816)
  at 
org.apache.hadoop.hbase.regionserver.HRegion$4.lockRowsAndBuildMiniBatch(HRegion.java:7432)
  at 
org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutate(HRegion.java:4008)
  at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3982)
  at 
org.apache.hadoop.hbase.regionserver.HRegion.mutateRowsWithLocks(HRegion.java:7424)
  at 
org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.mutateRows(MultiRowMutationEndpoint.java:116)
  at 
org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos$MultiRowMutationService.callMethod(MultiRowMutationProtos.java:2266)
  at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8182)
  at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2481)
  at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2463)
...
Exception in thread "pool-678-thread-1" java.lang.AssertionError: This cp 
should fail because the target lock is blocked by previous put
  at org.junit.Assert.fail(Assert.java:88)
  at 
org.apache.hadoop.hbase.client.TestFromClientSide3.lambda$testMultiRowMutations$7(TestFromClientSide3.java:861)
  at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
  at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
{code}
Here is related code:
{code}
  cpService.execute(() -> {
...
if (!threw) {
  // Can't call fail() earlier because the catch would eat it.
  fail("This cp should fail because the target lock is blocked by 
previous put");
}
{code}
Since the fail() call is executed by the cpService, the assertion had no 
bearing on the outcome of the test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-21221) Ineffective assertion in TestFromClientSide3#testMultiRowMutations

2018-09-22 Thread Ted Yu (JIRA)
Ted Yu created HBASE-21221:
--

 Summary: Ineffective assertion in 
TestFromClientSide3#testMultiRowMutations
 Key: HBASE-21221
 URL: https://issues.apache.org/jira/browse/HBASE-21221
 Project: HBase
  Issue Type: Test
Reporter: Ted Yu


Observed the following in 
org.apache.hadoop.hbase.util.TestFromClientSide3WoUnsafe-output.txt :
{code}
Caused by: 
org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
java.io.IOException: Timed out waiting for lock for row: ROW-1 in region 
089bdfa75f44d88e596479038a6da18b
  at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:5816)
  at 
org.apache.hadoop.hbase.regionserver.HRegion$4.lockRowsAndBuildMiniBatch(HRegion.java:7432)
  at 
org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutate(HRegion.java:4008)
  at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3982)
  at 
org.apache.hadoop.hbase.regionserver.HRegion.mutateRowsWithLocks(HRegion.java:7424)
  at 
org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.mutateRows(MultiRowMutationEndpoint.java:116)
  at 
org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos$MultiRowMutationService.callMethod(MultiRowMutationProtos.java:2266)
  at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8182)
  at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2481)
  at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2463)
...
Exception in thread "pool-678-thread-1" java.lang.AssertionError: This cp 
should fail because the target lock is blocked by previous put
  at org.junit.Assert.fail(Assert.java:88)
  at 
org.apache.hadoop.hbase.client.TestFromClientSide3.lambda$testMultiRowMutations$7(TestFromClientSide3.java:861)
  at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
  at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
{code}
Here is related code:
{code}
  cpService.execute(() -> {
...
if (!threw) {
  // Can't call fail() earlier because the catch would eat it.
  fail("This cp should fail because the target lock is blocked by 
previous put");
}
{code}
Since the fail() call is executed by the cpService, the assertion had no 
bearing on the outcome of the test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-8554) Upgrade AWS SDK

2018-09-22 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-8554?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated FLINK-8554:
--
Description: 
AWS SDK 1.11.271 fixes a lot of bugs.

One of which would exhibit the following:

{code}
Caused by: java.lang.NullPointerException
at com.amazonaws.metrics.AwsSdkMetrics.getRegion(AwsSdkMetrics.java:729)
at com.amazonaws.metrics.MetricAdmin.getRegion(MetricAdmin.java:67)
at sun.reflect.GeneratedMethodAccessor132.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
{code}

AWS SDK 1.11.375 has been released.

 https://aws.amazon.com/about-aws/whats-new/2018/03/longer-role-sessions/

  was:
AWS SDK 1.11.271 fixes a lot of bugs.

One of which would exhibit the following:

{code}
Caused by: java.lang.NullPointerException
at com.amazonaws.metrics.AwsSdkMetrics.getRegion(AwsSdkMetrics.java:729)
at com.amazonaws.metrics.MetricAdmin.getRegion(MetricAdmin.java:67)
at sun.reflect.GeneratedMethodAccessor132.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
{code}


> Upgrade AWS SDK
> ---
>
> Key: FLINK-8554
> URL: https://issues.apache.org/jira/browse/FLINK-8554
> Project: Flink
>  Issue Type: Improvement
>  Components: Build System
>    Reporter: Ted Yu
>Assignee: vinoyang
>Priority: Minor
>
> AWS SDK 1.11.271 fixes a lot of bugs.
> One of which would exhibit the following:
> {code}
> Caused by: java.lang.NullPointerException
>   at com.amazonaws.metrics.AwsSdkMetrics.getRegion(AwsSdkMetrics.java:729)
>   at com.amazonaws.metrics.MetricAdmin.getRegion(MetricAdmin.java:67)
>   at sun.reflect.GeneratedMethodAccessor132.invoke(Unknown Source)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
> {code}
> AWS SDK 1.11.375 has been released.
>  https://aws.amazon.com/about-aws/whats-new/2018/03/longer-role-sessions/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (KAFKA-6904) DynamicBrokerReconfigurationTest#testAdvertisedListenerUpdate is flaky

2018-09-22 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KAFKA-6904?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16607291#comment-16607291
 ] 

Ted Yu edited comment on KAFKA-6904 at 9/22/18 7:06 PM:


Haven't seen this failure lately .


was (Author: yuzhih...@gmail.com):
Haven't seen this failure lately.

> DynamicBrokerReconfigurationTest#testAdvertisedListenerUpdate is flaky
> --
>
> Key: KAFKA-6904
> URL: https://issues.apache.org/jira/browse/KAFKA-6904
> Project: Kafka
>  Issue Type: Test
>  Components: unit tests
>    Reporter: Ted Yu
>Priority: Major
>
> From 
> https://builds.apache.org/job/kafka-pr-jdk10-scala2.12/820/testReport/junit/kafka.server/DynamicBrokerReconfigurationTest/testAdvertisedListenerUpdate/
>  :
> {code}
> kafka.server.DynamicBrokerReconfigurationTest.testAdvertisedListenerUpdate
> Failing for the past 1 build (Since Failed#820 )
> Took 21 sec.
> Error Message
> java.util.concurrent.ExecutionException: 
> org.apache.kafka.common.errors.NotLeaderForPartitionException: This server is 
> not the leader for that topic-partition.
> Stacktrace
> java.util.concurrent.ExecutionException: 
> org.apache.kafka.common.errors.NotLeaderForPartitionException: This server is 
> not the leader for that topic-partition.
>   at 
> org.apache.kafka.clients.producer.internals.FutureRecordMetadata.valueOrError(FutureRecordMetadata.java:94)
>   at 
> org.apache.kafka.clients.producer.internals.FutureRecordMetadata.get(FutureRecordMetadata.java:77)
>   at 
> org.apache.kafka.clients.producer.internals.FutureRecordMetadata.get(FutureRecordMetadata.java:29)
>   at 
> kafka.server.DynamicBrokerReconfigurationTest.$anonfun$verifyProduceConsume$3(DynamicBrokerReconfigurationTest.scala:996)
>   at 
> scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:234)
>   at scala.collection.Iterator.foreach(Iterator.scala:944)
>   at scala.collection.Iterator.foreach$(Iterator.scala:944)
>   at scala.collection.AbstractIterator.foreach(Iterator.scala:1432)
>   at scala.collection.IterableLike.foreach(IterableLike.scala:71)
>   at scala.collection.IterableLike.foreach$(IterableLike.scala:70)
>   at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
>   at scala.collection.TraversableLike.map(TraversableLike.scala:234)
>   at scala.collection.TraversableLike.map$(TraversableLike.scala:227)
>   at scala.collection.AbstractTraversable.map(Traversable.scala:104)
>   at 
> kafka.server.DynamicBrokerReconfigurationTest.verifyProduceConsume(DynamicBrokerReconfigurationTest.scala:996)
>   at 
> kafka.server.DynamicBrokerReconfigurationTest.testAdvertisedListenerUpdate(DynamicBrokerReconfigurationTest.scala:742)
> {code}
> The above happened with jdk 10.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (KYLIN-3431) Avoid FileInputStream/FileOutputStream

2018-09-22 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3431?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16624744#comment-16624744
 ] 

Ted Yu commented on KYLIN-3431:
---

Thanks, Long.

> Avoid FileInputStream/FileOutputStream
> --
>
> Key: KYLIN-3431
> URL: https://issues.apache.org/jira/browse/KYLIN-3431
> Project: Kylin
>  Issue Type: Improvement
>    Reporter: Ted Yu
>Priority: Major
> Fix For: Backlog
>
>
> They rely on finalizers (before Java 11), which create unnecessary GC load.
> The alternatives, {{Files.newInputStream}}, are as easy to use and don't have 
> this issue.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (KYLIN-3417) Consider replacing ReentrantReadWriteLock with StampedLock

2018-09-21 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3417?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16624443#comment-16624443
 ] 

Ted Yu commented on KYLIN-3417:
---

For phase I, we don't need to use Optimistic read lock

> Consider replacing ReentrantReadWriteLock with StampedLock
> --
>
> Key: KYLIN-3417
> URL: https://issues.apache.org/jira/browse/KYLIN-3417
> Project: Kylin
>  Issue Type: Improvement
>    Reporter: Ted Yu
>Assignee: jiatao.tao
>Priority: Major
> Fix For: Backlog
>
>
> ReentrantReadWriteLock's are only the right solution when there is long hold 
> time due to expensive I/O.
> It is expensive for readers.
> We should see if the lighter {{StampedLock}} can be used instead.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (GORA-544) Upgrade zookeeper to 3.4.13

2018-09-21 Thread Ted Yu (JIRA)
Ted Yu created GORA-544:
---

 Summary: Upgrade zookeeper to 3.4.13
 Key: GORA-544
 URL: https://issues.apache.org/jira/browse/GORA-544
 Project: Apache Gora
  Issue Type: Task
Reporter: Ted Yu


zookeeper 3.4.13 is being released with the following fixes:

ZOOKEEPER-2959 fixes data loss when observer is used

ZOOKEEPER-2184 allows ZooKeeper Java clients to work in dynamic IP (container / 
cloud)
environment



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (GORA-542) Upgrade to JUnit 5

2018-09-21 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/GORA-542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated GORA-542:

Description: 
JUnit 5 brings multiple useful features so tests are easier to read and write.

We can bump up the dependency version and create new tests with JUnit 5 
features.

Relevant features of JUnit 5: dynamic test, nested tests, parameterized tests
https://twitter.com/nipafx/status/1027095088059559936

  was:
JUnit 5 brings multiple useful features so tests are easier to read and write.

We can bump up the dependency version and create new tests with JUnit 5 
features.

Relevant features of JUnit 5: dynamic test, nested tests, parameterized tests

https://twitter.com/nipafx/status/1027095088059559936


> Upgrade to JUnit 5
> --
>
> Key: GORA-542
> URL: https://issues.apache.org/jira/browse/GORA-542
> Project: Apache Gora
>  Issue Type: Task
>    Reporter: Ted Yu
>Priority: Major
>
> JUnit 5 brings multiple useful features so tests are easier to read and write.
> We can bump up the dependency version and create new tests with JUnit 5 
> features.
> Relevant features of JUnit 5: dynamic test, nested tests, parameterized tests
> https://twitter.com/nipafx/status/1027095088059559936



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-10391) MillisOfDay is used in place of instant for LocalTime ctor in AvroKryoSerializerUtils

2018-09-21 Thread Ted Yu (JIRA)
Ted Yu created FLINK-10391:
--

 Summary: MillisOfDay is used in place of instant for LocalTime 
ctor in AvroKryoSerializerUtils
 Key: FLINK-10391
 URL: https://issues.apache.org/jira/browse/FLINK-10391
 Project: Flink
  Issue Type: Bug
Reporter: Ted Yu


>From the JodaLocalTimeSerializer#write, we serialize getMillisOfDay() value 
>from LocalTime.
For read method:
{code}
  final int time = input.readInt(true);
  return new LocalTime(time, 
ISOChronology.getInstanceUTC().withZone(DateTimeZone.UTC));
{code}
It seems 
http://joda-time.sourceforge.net/apidocs/org/joda/time/LocalTime.html#fromMillisOfDay(long,%20org.joda.time.Chronology)
 should be used instead.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (KYLIN-3447) Upgrade zookeeper to 3.4.13

2018-09-21 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated KYLIN-3447:
--
Description: 
zookeeper 3.4.13 is being released with the following fixes:

ZOOKEEPER-2959 fixes data loss when observer is used

ZOOKEEPER-2184 allows ZooKeeper Java clients to work in dynamic IP (container / 
cloud)
environment

  was:
zookeeper 3.4.13 is being released with the following fixes:

ZOOKEEPER-2959 fixes data loss when observer is used
ZOOKEEPER-2184 allows ZooKeeper Java clients to work in dynamic IP (container / 
cloud)
environment


> Upgrade zookeeper to 3.4.13
> ---
>
> Key: KYLIN-3447
> URL: https://issues.apache.org/jira/browse/KYLIN-3447
> Project: Kylin
>  Issue Type: Improvement
>    Reporter: Ted Yu
>Priority: Major
>
> zookeeper 3.4.13 is being released with the following fixes:
> ZOOKEEPER-2959 fixes data loss when observer is used
> ZOOKEEPER-2184 allows ZooKeeper Java clients to work in dynamic IP (container 
> / cloud)
> environment



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (KYLIN-3484) Update Hadoop version to 2.7.7

2018-09-21 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated KYLIN-3484:
--
Description: We should upgrade the Hadoop 2.7 dependency to 2.7.7, to pick 
up bug and security fixes .  (was: We should upgrade the Hadoop 2.7 dependency 
to 2.7.7, to pick up bug and security fixes.)

> Update Hadoop version to 2.7.7
> --
>
> Key: KYLIN-3484
> URL: https://issues.apache.org/jira/browse/KYLIN-3484
> Project: Kylin
>  Issue Type: Task
>    Reporter: Ted Yu
>Priority: Minor
>
> We should upgrade the Hadoop 2.7 dependency to 2.7.7, to pick up bug and 
> security fixes .



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (KYLIN-3519) Upgrade Jacoco version to 0.8.2

2018-09-21 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3519?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated KYLIN-3519:
--
Description: 
Jacoco 0.8.2 adds Java 11 support:

   https://github.com/jacoco/jacoco/releases/tag/v0.8.2

Java 11 RC1 is out.
We should consider upgrading Jacoco.

  was:
Jacoco 0.8.2 adds Java 11 support:

https://github.com/jacoco/jacoco/releases/tag/v0.8.2

Java 11 RC1 is out.
We should consider upgrading Jacoco.


> Upgrade Jacoco version to 0.8.2
> ---
>
> Key: KYLIN-3519
> URL: https://issues.apache.org/jira/browse/KYLIN-3519
> Project: Kylin
>  Issue Type: Improvement
>    Reporter: Ted Yu
>Priority: Minor
>
> Jacoco 0.8.2 adds Java 11 support:
>https://github.com/jacoco/jacoco/releases/tag/v0.8.2
> Java 11 RC1 is out.
> We should consider upgrading Jacoco.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (KYLIN-3559) Use Splitter for splitting String

2018-09-21 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3559?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated KYLIN-3559:
--
Description: See http://errorprone.info/bugpattern/StringSplitter for why 
Splitter is preferred .  (was: See 
http://errorprone.info/bugpattern/StringSplitter for why Splitter is preferred.)

> Use Splitter for splitting String
> -
>
> Key: KYLIN-3559
> URL: https://issues.apache.org/jira/browse/KYLIN-3559
> Project: Kylin
>  Issue Type: Task
>    Reporter: Ted Yu
>Priority: Major
> Fix For: v2.6.0
>
>
> See http://errorprone.info/bugpattern/StringSplitter for why Splitter is 
> preferred .



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HBASE-20578) Support region server group in target cluster

2018-09-21 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-20578?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16605078#comment-16605078
 ] 

Ted Yu edited comment on HBASE-20578 at 9/21/18 8:54 PM:
-

Ping [~albertlee...@gmail.com]

Do you have any further question ?


was (Author: yuzhih...@gmail.com):
Ping [~albertlee...@gmail.com]

> Support region server group in target cluster
> -
>
> Key: HBASE-20578
> URL: https://issues.apache.org/jira/browse/HBASE-20578
> Project: HBase
>  Issue Type: Improvement
>  Components: Replication, rsgroup
>    Reporter: Ted Yu
>Assignee: Albert Lee
>Priority: Major
> Fix For: 3.0.0
>
> Attachments: HBASE-20578-001.patch
>
>
> When source tables belong to non-default region server group(s) and there are 
> region server group counterpart in the target cluster, we should support 
> replicating to target cluster using the region server group mapping.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-10389) TaskManagerServicesConfiguration ctor contains self assignment

2018-09-21 Thread Ted Yu (JIRA)
Ted Yu created FLINK-10389:
--

 Summary: TaskManagerServicesConfiguration ctor contains self 
assignment
 Key: FLINK-10389
 URL: https://issues.apache.org/jira/browse/FLINK-10389
 Project: Flink
  Issue Type: Task
Reporter: Ted Yu


TaskManagerServicesConfiguration has:
{code}
this.systemResourceMetricsEnabled = systemResourceMetricsEnabled;
{code}
There is no systemResourceMetricsEnabled parameter to the ctor.

This was reported by findbugs.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-10389) TaskManagerServicesConfiguration ctor contains self assignment

2018-09-21 Thread Ted Yu (JIRA)
Ted Yu created FLINK-10389:
--

 Summary: TaskManagerServicesConfiguration ctor contains self 
assignment
 Key: FLINK-10389
 URL: https://issues.apache.org/jira/browse/FLINK-10389
 Project: Flink
  Issue Type: Task
Reporter: Ted Yu


TaskManagerServicesConfiguration has:
{code}
this.systemResourceMetricsEnabled = systemResourceMetricsEnabled;
{code}
There is no systemResourceMetricsEnabled parameter to the ctor.

This was reported by findbugs.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-10388) RestClientTest sometimes fails with AssertionError

2018-09-21 Thread Ted Yu (JIRA)
Ted Yu created FLINK-10388:
--

 Summary: RestClientTest sometimes fails with AssertionError
 Key: FLINK-10388
 URL: https://issues.apache.org/jira/browse/FLINK-10388
 Project: Flink
  Issue Type: Test
Reporter: Ted Yu


Running the test on Linux I got:
{code}
testConnectionTimeout(org.apache.flink.runtime.rest.RestClientTest)  Time 
elapsed: 1.918 sec  <<< FAILURE!
java.lang.AssertionError:
Expected: an instance of 
org.apache.flink.shaded.netty4.io.netty.channel.ConnectTimeoutException
 but: 
 is a 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannel$AnnotatedSocketException
  at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
  at org.junit.Assert.assertThat(Assert.java:956)
  at org.junit.Assert.assertThat(Assert.java:923)
  at 
org.apache.flink.runtime.rest.RestClientTest.testConnectionTimeout(RestClientTest.java:69)
{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-10388) RestClientTest sometimes fails with AssertionError

2018-09-21 Thread Ted Yu (JIRA)
Ted Yu created FLINK-10388:
--

 Summary: RestClientTest sometimes fails with AssertionError
 Key: FLINK-10388
 URL: https://issues.apache.org/jira/browse/FLINK-10388
 Project: Flink
  Issue Type: Test
Reporter: Ted Yu


Running the test on Linux I got:
{code}
testConnectionTimeout(org.apache.flink.runtime.rest.RestClientTest)  Time 
elapsed: 1.918 sec  <<< FAILURE!
java.lang.AssertionError:
Expected: an instance of 
org.apache.flink.shaded.netty4.io.netty.channel.ConnectTimeoutException
 but: 
 is a 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannel$AnnotatedSocketException
  at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
  at org.junit.Assert.assertThat(Assert.java:956)
  at org.junit.Assert.assertThat(Assert.java:923)
  at 
org.apache.flink.runtime.rest.RestClientTest.testConnectionTimeout(RestClientTest.java:69)
{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3575) Unclosed Connection in DriverTest

2018-09-21 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3575:
-

 Summary: Unclosed Connection in DriverTest
 Key: KYLIN-3575
 URL: https://issues.apache.org/jira/browse/KYLIN-3575
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


In testCalciteProps :
{code}
KylinConnection conn = (KylinConnection) 
driver.connect("jdbc:kylin:test_url/test_db", props);
{code}
The connection should be closed upon return.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3575) Unclosed Connection in DriverTest

2018-09-21 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3575:
-

 Summary: Unclosed Connection in DriverTest
 Key: KYLIN-3575
 URL: https://issues.apache.org/jira/browse/KYLIN-3575
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


In testCalciteProps :
{code}
KylinConnection conn = (KylinConnection) 
driver.connect("jdbc:kylin:test_url/test_db", props);
{code}
The connection should be closed upon return.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3574) Unclosed lookup table in DictionaryGeneratorCLI#processSegment

2018-09-21 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3574:
-

 Summary: Unclosed lookup table in 
DictionaryGeneratorCLI#processSegment
 Key: KYLIN-3574
 URL: https://issues.apache.org/jira/browse/KYLIN-3574
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


Here is related code:
{code}
JoinDesc join = 
cubeSeg.getModel().getJoinsTree().getJoinByPKSide(lookup);
cubeMgr.getLookupTable(cubeSeg, join);
{code}
The returned lookup table should be closed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3574) Unclosed lookup table in DictionaryGeneratorCLI#processSegment

2018-09-21 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3574:
-

 Summary: Unclosed lookup table in 
DictionaryGeneratorCLI#processSegment
 Key: KYLIN-3574
 URL: https://issues.apache.org/jira/browse/KYLIN-3574
 Project: Kylin
  Issue Type: Bug
Reporter: Ted Yu


Here is related code:
{code}
JoinDesc join = 
cubeSeg.getModel().getJoinsTree().getJoinByPKSide(lookup);
cubeMgr.getLookupTable(cubeSeg, join);
{code}
The returned lookup table should be closed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (AMBARI-24674) Upgrade checkstyle version to 8.9

2018-09-21 Thread Ted Yu (JIRA)
Ted Yu created AMBARI-24674:
---

 Summary: Upgrade checkstyle version to 8.9
 Key: AMBARI-24674
 URL: https://issues.apache.org/jira/browse/AMBARI-24674
 Project: Ambari
  Issue Type: Task
Reporter: Ted Yu


Currently we use checkstyle 6.19 which is very old.

We should upgrade to 8.9 release.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


Re: Can MOB be enabled for existing table?

2018-09-21 Thread Ted Yu
bq. region server crashes after ttl archiver runs

This is the first time I hear of such problem. Can you give us more
information (stack trace, server log snippet for that period) ?

Thanks

On Thu, Sep 20, 2018 at 11:14 PM meirav.malka 
wrote:

> Hi,Since you probably don't have enough space for backup for such a large
> table, test it on a small one and rowcount before and after.We work
> extensively with mob and there are several issues with it that we had to
> work around.1.Mob and ttl - region server crashes after ttl archiver
> runs.2.Mob and snapshot - didn't work in cdh 5.7.13.Mob compaction-
> currently performed only by the master.if your flushes aren't big enough ,
> compaction will never end.4.Deletes - major mob compaction is the only one
> doing mob delete,it is also done by the master, on the entire data( it is
> not scheduled )If you have deletes, the will never actually be deleted.
> Regards,Meirav
> Sent from my Samsung Galaxy smartphone.
>  Original message From: Jean-Marc Spaggiari <
> jean-m...@spaggiari.org> Date: 9/21/18  01:09  (GMT+02:00) To: user <
> user@hbase.apache.org> Subject: Re: Can MOB be enabled for existing
> table?
> Hi Andrea,
>
> I will say that the easiest way to know is to try it... I'm guessing that
> it should work. When HBase will compact the file, it will figure that some
> fields are bigger than the configures MOB threshold and will move them
> under the MOB umbrella. But this has to be tested. Let us know.
>
> Thanks,
>
> JMS
>
> Le jeu. 20 sept. 2018 à 08:26, Andrea Fonti  a
> écrit :
>
> > Hi,
> >
> > I’m currently running HBASE on a Cloudera CDH 5.8.4 distribution
> > (hbase-1.2.0+cdh5.8.5+263), is it possible to enable MOB for a column of
> an
> > existing multi terabyte table?
> >
> > If the answer is affirmative what will hbase do with the existing data?
> > Will it rewrite with MOB files the next major compaction?
> >
> > Thank you
> >
> > Andrea
>


[jira] [Commented] (HBASE-21149) TestIncrementalBackupWithBulkLoad may fail due to file copy failure

2018-09-21 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21149?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623691#comment-16623691
 ] 

Ted Yu commented on HBASE-21149:


>From 
>https://builds.apache.org/job/HBase%20Nightly/job/master/503//testReport/junit/org.apache.hadoop.hbase.backup/TestIncrementalBackupWithBulkLoad/health_checks___yetus_jdk8_hadoop3_checks___TestIncBackupDeleteTable/
> :
{code}
java.io.IOException: Failed copy from 
hdfs://localhost:40355/user/jenkins/test-data/6d0386bb-d9d8-ae4e-e2bf-c423554ea898/data/default/test-1537528002870/711c2cc8826d7170be6051afbe0b5ca0/f/1afeb42e839a40afa6389ac80c82e1ed_SeqId_205_,hdfs://localhost:40355/user/jenkins/test-data/6d0386bb-d9d8-ae4e-e2bf-c423554ea898/data/default/test-1537528002870/711c2cc8826d7170be6051afbe0b5ca0/f/2087ea2ec81a449ea5d35c405767f970_SeqId_205_
 to hdfs://localhost:40355/backupUT/backup_1537528036126
{code}
There were some output truncated.

I searched for 2087ea2ec81a449ea5d35c405767f970_SeqId_205_ in the output but 
didn't find much clue.

> TestIncrementalBackupWithBulkLoad may fail due to file copy failure
> ---
>
> Key: HBASE-21149
> URL: https://issues.apache.org/jira/browse/HBASE-21149
> Project: HBase
>  Issue Type: Test
>  Components: backuprestore
>Reporter: Ted Yu
>Assignee: Vladimir Rodionov
>Priority: Major
> Attachments: HBASE-21149-v1.patch
>
>
> From 
> https://builds.apache.org/job/HBase%20Nightly/job/master/471/testReport/junit/org.apache.hadoop.hbase.backup/TestIncrementalBackupWithBulkLoad/TestIncBackupDeleteTable/
>  :
> {code}
> 2018-09-03 11:54:30,526 ERROR [Time-limited test] 
> impl.TableBackupClient(235): Unexpected Exception : Failed copy from 
> hdfs://localhost:53075/user/jenkins/test-data/ecd40bd0-cb93-91e0-90b5-7bfd5bb2c566/data/default/test-1535975627781/773f5709b645b46bd3840f9cfb549c5a/f/0f626c66493649daaf84057b8dd71a30_SeqId_205_,hdfs://localhost:53075/user/jenkins/test-data/ecd40bd0-cb93-91e0-90b5-7bfd5bb2c566/data/default/test-1535975627781/773f5709b645b46bd3840f9cfb549c5a/f/ad8df6415bd9459d9b3df76c588d79df_SeqId_205_
>  to hdfs://localhost:53075/backupUT/backup_1535975655488
> java.io.IOException: Failed copy from 
> hdfs://localhost:53075/user/jenkins/test-data/ecd40bd0-cb93-91e0-90b5-7bfd5bb2c566/data/default/test-1535975627781/773f5709b645b46bd3840f9cfb549c5a/f/0f626c66493649daaf84057b8dd71a30_SeqId_205_,hdfs://localhost:53075/user/jenkins/test-data/ecd40bd0-cb93-91e0-90b5-7bfd5bb2c566/data/default/test-1535975627781/773f5709b645b46bd3840f9cfb549c5a/f/ad8df6415bd9459d9b3df76c588d79df_SeqId_205_
>  to hdfs://localhost:53075/backupUT/backup_1535975655488
>   at 
> org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient.incrementalCopyHFiles(IncrementalTableBackupClient.java:351)
>   at 
> org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient.copyBulkLoadedFiles(IncrementalTableBackupClient.java:219)
>   at 
> org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient.handleBulkLoad(IncrementalTableBackupClient.java:198)
>   at 
> org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient.execute(IncrementalTableBackupClient.java:320)
>   at 
> org.apache.hadoop.hbase.backup.impl.BackupAdminImpl.backupTables(BackupAdminImpl.java:605)
>   at 
> org.apache.hadoop.hbase.backup.TestIncrementalBackupWithBulkLoad.TestIncBackupDeleteTable(TestIncrementalBackupWithBulkLoad.java:104)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> {code}
> However, some part of the test output was lost:
> {code}
> 2018-09-03 11:53:36,793 DEBUG [RS:0;765c9ca5ea28:36357] regions
> ...[truncated 398396 chars]...
> 8)
>   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
>   at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (KAFKA-6303) Potential lack of synchronization in NioEchoServer#AcceptorThread

2018-09-20 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KAFKA-6303?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16333251#comment-16333251
 ] 

Ted Yu edited comment on KAFKA-6303 at 9/21/18 1:50 AM:


+1 from me .


was (Author: yuzhih...@gmail.com):
+1 from me.

> Potential lack of synchronization in NioEchoServer#AcceptorThread
> -
>
> Key: KAFKA-6303
> URL: https://issues.apache.org/jira/browse/KAFKA-6303
> Project: Kafka
>  Issue Type: Bug
>  Components: network
>    Reporter: Ted Yu
>Assignee: siva santhalingam
>Priority: Minor
>
> In the run() method:
> {code}
> SocketChannel socketChannel = 
> ((ServerSocketChannel) key.channel()).accept();
> socketChannel.configureBlocking(false);
> newChannels.add(socketChannel);
> {code}
> Modification to newChannels should be protected by synchronized block.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-16458) Shorten backup / restore test execution time

2018-09-20 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-16458?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-16458:
---
Fix Version/s: 3.0.0

> Shorten backup / restore test execution time
> 
>
> Key: HBASE-16458
> URL: https://issues.apache.org/jira/browse/HBASE-16458
> Project: HBase
>  Issue Type: Test
>    Reporter: Ted Yu
>Assignee: Vladimir Rodionov
>Priority: Major
>  Labels: backup
> Fix For: 3.0.0
>
> Attachments: 16458-v1.patch, 16458.HBASE-7912.v3.txt, 
> 16458.HBASE-7912.v4.txt, 16458.v1.txt, 16458.v2.txt, 16458.v2.txt, 
> 16458.v3.txt, 16458.v4.txt, 16458.v5.txt, HBASE-16458-v1.patch, 
> HBASE-16458-v2.patch
>
>
> Below was timing information for all the backup / restore tests (today's 
> result):
> {code}
> Running org.apache.hadoop.hbase.backup.TestIncrementalBackup
> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 576.273 sec - 
> in org.apache.hadoop.hbase.backup.TestIncrementalBackup
> Running org.apache.hadoop.hbase.backup.TestBackupBoundaryTests
> Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 124.67 sec - 
> in org.apache.hadoop.hbase.backup.TestBackupBoundaryTests
> Running org.apache.hadoop.hbase.backup.TestBackupStatusProgress
> Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 102.34 sec - 
> in org.apache.hadoop.hbase.backup.TestBackupStatusProgress
> Running org.apache.hadoop.hbase.backup.TestBackupAdmin
> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 490.251 sec - 
> in org.apache.hadoop.hbase.backup.TestBackupAdmin
> Running org.apache.hadoop.hbase.backup.TestHFileArchiving
> Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 84.323 sec - 
> in org.apache.hadoop.hbase.backup.TestHFileArchiving
> Running org.apache.hadoop.hbase.backup.TestSystemTableSnapshot
> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 65.492 sec - 
> in org.apache.hadoop.hbase.backup.TestSystemTableSnapshot
> Running org.apache.hadoop.hbase.backup.TestBackupDescribe
> Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 93.758 sec - 
> in org.apache.hadoop.hbase.backup.TestBackupDescribe
> Running org.apache.hadoop.hbase.backup.TestBackupLogCleaner
> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 109.187 sec - 
> in org.apache.hadoop.hbase.backup.TestBackupLogCleaner
> Running org.apache.hadoop.hbase.backup.TestIncrementalBackupNoDataLoss
> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 330.539 sec - 
> in org.apache.hadoop.hbase.backup.TestIncrementalBackupNoDataLoss
> Running org.apache.hadoop.hbase.backup.TestRemoteBackup
> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 84.371 sec - 
> in org.apache.hadoop.hbase.backup.TestRemoteBackup
> Running org.apache.hadoop.hbase.backup.TestBackupSystemTable
> Tests run: 15, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 67.893 sec - 
> in org.apache.hadoop.hbase.backup.TestBackupSystemTable
> Running org.apache.hadoop.hbase.backup.TestRestoreBoundaryTests
> Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 120.779 sec - 
> in org.apache.hadoop.hbase.backup.TestRestoreBoundaryTests
> Running org.apache.hadoop.hbase.backup.TestFullBackupSetRestoreSet
> Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 117.815 sec - 
> in org.apache.hadoop.hbase.backup.TestFullBackupSetRestoreSet
> Running org.apache.hadoop.hbase.backup.TestBackupShowHistory
> Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 136.517 sec - 
> in org.apache.hadoop.hbase.backup.TestBackupShowHistory
> Running org.apache.hadoop.hbase.backup.TestRemoteRestore
> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 91.799 sec - 
> in org.apache.hadoop.hbase.backup.TestRemoteRestore
> Running org.apache.hadoop.hbase.backup.TestFullRestore
> Tests run: 12, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 317.711 sec 
> - in org.apache.hadoop.hbase.backup.TestFullRestore
> Running org.apache.hadoop.hbase.backup.TestFullBackupSet
> Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 87.045 sec - 
> in org.apache.hadoop.hbase.backup.TestFullBackupSet
> Running org.apache.hadoop.hbase.backup.TestBackupDelete
> Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 86.214 sec - 
> in org.apache.hadoop.hbase.backup.TestBackupDelete
> Running org.apache.hadoop.hbase.backup.TestBackupDeleteRestore
> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 77.631 sec - 
> in org.apache.hadoop.hbase.backup.TestBackup

[jira] [Created] (HBASE-21216) TestSnapshotFromMaster#testSnapshotHFileArchiving is flaky

2018-09-20 Thread Ted Yu (JIRA)
Ted Yu created HBASE-21216:
--

 Summary: TestSnapshotFromMaster#testSnapshotHFileArchiving is flaky
 Key: HBASE-21216
 URL: https://issues.apache.org/jira/browse/HBASE-21216
 Project: HBase
  Issue Type: Test
Reporter: Ted Yu


>From 
>https://builds.apache.org/job/HBase-Flaky-Tests/job/branch-2/794/testReport/junit/org.apache.hadoop.hbase.master.cleaner/TestSnapshotFromMaster/testSnapshotHFileArchiving/
> :
{code}
java.lang.AssertionError: Archived hfiles [] and table hfiles 
[9ca09392705f425f9c916beedc10d63c] is missing snapshot 
file:6739a09747e54189a4112a6d8f37e894
at 
org.apache.hadoop.hbase.master.cleaner.TestSnapshotFromMaster.testSnapshotHFileArchiving(TestSnapshotFromMaster.java:370)
{code}
The file appeared in archive dir before hfile cleaners were run:
{code}
2018-09-20 10:38:53,187 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|-archive/
2018-09-20 10:38:53,188 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|data/
2018-09-20 10:38:53,189 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|---default/
2018-09-20 10:38:53,190 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|--test/
2018-09-20 10:38:53,191 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|-1237d57b63a7bdf067a930441a02514a/
2018-09-20 10:38:53,192 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|recovered.edits/
2018-09-20 10:38:53,193 DEBUG [Time-limited test] util.CommonFSUtils(774): 
|---4.seqid
2018-09-20 10:38:53,193 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|-29e1700e09b51223ad2f5811105a4d51/
2018-09-20 10:38:53,194 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|fam/
2018-09-20 10:38:53,195 DEBUG [Time-limited test] util.CommonFSUtils(774): 
|---2c66a18f6c1a4074b84ffbb3245268c4
2018-09-20 10:38:53,196 DEBUG [Time-limited test] util.CommonFSUtils(774): 
|---45bb396c6a5e49629e45a4d56f1e9b14
2018-09-20 10:38:53,196 DEBUG [Time-limited test] util.CommonFSUtils(774): 
|---6739a09747e54189a4112a6d8f37e894
{code}
However, the archive dir became empty after hfile cleaners were run:
{code}
2018-09-20 10:38:53,312 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|-archive/
2018-09-20 10:38:53,313 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|-corrupt/
{code}
Leading to the assertion failure.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-21216) TestSnapshotFromMaster#testSnapshotHFileArchiving is flaky

2018-09-20 Thread Ted Yu (JIRA)
Ted Yu created HBASE-21216:
--

 Summary: TestSnapshotFromMaster#testSnapshotHFileArchiving is flaky
 Key: HBASE-21216
 URL: https://issues.apache.org/jira/browse/HBASE-21216
 Project: HBase
  Issue Type: Test
Reporter: Ted Yu


>From 
>https://builds.apache.org/job/HBase-Flaky-Tests/job/branch-2/794/testReport/junit/org.apache.hadoop.hbase.master.cleaner/TestSnapshotFromMaster/testSnapshotHFileArchiving/
> :
{code}
java.lang.AssertionError: Archived hfiles [] and table hfiles 
[9ca09392705f425f9c916beedc10d63c] is missing snapshot 
file:6739a09747e54189a4112a6d8f37e894
at 
org.apache.hadoop.hbase.master.cleaner.TestSnapshotFromMaster.testSnapshotHFileArchiving(TestSnapshotFromMaster.java:370)
{code}
The file appeared in archive dir before hfile cleaners were run:
{code}
2018-09-20 10:38:53,187 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|-archive/
2018-09-20 10:38:53,188 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|data/
2018-09-20 10:38:53,189 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|---default/
2018-09-20 10:38:53,190 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|--test/
2018-09-20 10:38:53,191 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|-1237d57b63a7bdf067a930441a02514a/
2018-09-20 10:38:53,192 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|recovered.edits/
2018-09-20 10:38:53,193 DEBUG [Time-limited test] util.CommonFSUtils(774): 
|---4.seqid
2018-09-20 10:38:53,193 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|-29e1700e09b51223ad2f5811105a4d51/
2018-09-20 10:38:53,194 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|fam/
2018-09-20 10:38:53,195 DEBUG [Time-limited test] util.CommonFSUtils(774): 
|---2c66a18f6c1a4074b84ffbb3245268c4
2018-09-20 10:38:53,196 DEBUG [Time-limited test] util.CommonFSUtils(774): 
|---45bb396c6a5e49629e45a4d56f1e9b14
2018-09-20 10:38:53,196 DEBUG [Time-limited test] util.CommonFSUtils(774): 
|---6739a09747e54189a4112a6d8f37e894
{code}
However, the archive dir became empty after hfile cleaners were run:
{code}
2018-09-20 10:38:53,312 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|-archive/
2018-09-20 10:38:53,313 DEBUG [Time-limited test] util.CommonFSUtils(771): 
|-corrupt/
{code}
Leading to the assertion failure.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (FLINK-7795) Utilize error-prone to discover common coding mistakes

2018-09-20 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-7795?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16345955#comment-16345955
 ] 

Ted Yu edited comment on FLINK-7795 at 9/20/18 4:30 PM:


error-prone has JDK 8 dependency .


was (Author: yuzhih...@gmail.com):
error-prone has JDK 8 dependency.

> Utilize error-prone to discover common coding mistakes
> --
>
> Key: FLINK-7795
> URL: https://issues.apache.org/jira/browse/FLINK-7795
> Project: Flink
>  Issue Type: Improvement
>  Components: Build System
>    Reporter: Ted Yu
>Priority: Major
>
> http://errorprone.info/ is a tool which detects common coding mistakes.
> We should incorporate into Flink build process.
> Here are the dependencies:
> {code}
> 
>   com.google.errorprone
>   error_prone_annotation
>   ${error-prone.version}
>   provided
> 
> 
>   
>   com.google.auto.service
>   auto-service
>   1.0-rc3
>   true
> 
> 
>   com.google.errorprone
>   error_prone_check_api
>   ${error-prone.version}
>   provided
>   
> 
>   com.google.code.findbugs
>   jsr305
> 
>   
> 
> 
>   com.google.errorprone
>   javac
>   9-dev-r4023-3
>   provided
> 
>   
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (FLINK-6105) Properly handle InterruptedException in HadoopInputFormatBase

2018-09-20 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-6105?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16307281#comment-16307281
 ] 

Ted Yu edited comment on FLINK-6105 at 9/20/18 4:29 PM:


In 
flink-connectors/flink-connector-filesystem/src/main/java/org/apache/flink/streaming/connectors/fs/RollingSink.java
 :
{code}
  try {
Thread.sleep(500);
  } catch (InterruptedException e1) {
// ignore it
  }
{code}
Interrupt status should be restored, or throw InterruptedIOException.


was (Author: yuzhih...@gmail.com):
In 
flink-connectors/flink-connector-filesystem/src/main/java/org/apache/flink/streaming/connectors/fs/RollingSink.java
 :
{code}
  try {
Thread.sleep(500);
  } catch (InterruptedException e1) {
// ignore it
  }
{code}
Interrupt status should be restored, or throw InterruptedIOException .

> Properly handle InterruptedException in HadoopInputFormatBase
> -
>
> Key: FLINK-6105
> URL: https://issues.apache.org/jira/browse/FLINK-6105
> Project: Flink
>  Issue Type: Bug
>  Components: DataStream API
>    Reporter: Ted Yu
>Assignee: zhangminglei
>Priority: Major
>
> When catching InterruptedException, we should throw InterruptedIOException 
> instead of IOException.
> The following example is from HadoopInputFormatBase :
> {code}
> try {
>   splits = this.mapreduceInputFormat.getSplits(jobContext);
> } catch (InterruptedException e) {
>   throw new IOException("Could not get Splits.", e);
> }
> {code}
> There may be other places where IOE is thrown.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (GORA-543) Upgrade Apache parent POM to version 21

2018-09-20 Thread Ted Yu (JIRA)
Ted Yu created GORA-543:
---

 Summary: Upgrade Apache parent POM to version 21
 Key: GORA-543
 URL: https://issues.apache.org/jira/browse/GORA-543
 Project: Apache Gora
  Issue Type: Task
Reporter: Ted Yu


Apache parent POM 21 has just been released. It generates .sha512 checksums 
during release with Maven Release Plugin.





--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3573) Upgrade Apache parent POM to version 21

2018-09-20 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3573:
-

 Summary: Upgrade Apache parent POM to version 21
 Key: KYLIN-3573
 URL: https://issues.apache.org/jira/browse/KYLIN-3573
 Project: Kylin
  Issue Type: Task
Reporter: Ted Yu


Apache parent POM 21 has just been released. It generates .sha512 checksums 
during release with Maven Release Plugin.

We should consider upgrading parent POM.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3573) Upgrade Apache parent POM to version 21

2018-09-20 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3573:
-

 Summary: Upgrade Apache parent POM to version 21
 Key: KYLIN-3573
 URL: https://issues.apache.org/jira/browse/KYLIN-3573
 Project: Kylin
  Issue Type: Task
Reporter: Ted Yu


Apache parent POM 21 has just been released. It generates .sha512 checksums 
during release with Maven Release Plugin.

We should consider upgrading parent POM.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21204) NPE when scan raw DELETE_FAMILY_VERSION and codec is not set

2018-09-20 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21204?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-21204:
---
   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: (was: 2.0.0)
   3.0.0
   Status: Resolved  (was: Patch Available)

Thanks for the patch, jingyun

> NPE when scan raw DELETE_FAMILY_VERSION and codec is not set
> 
>
> Key: HBASE-21204
> URL: https://issues.apache.org/jira/browse/HBASE-21204
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 2.1.0, 2.0.0, 2.2.0
>Reporter: Jingyun Tian
>Assignee: Jingyun Tian
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 2.1.0
>
> Attachments: HBASE-21204.branch-2.001.patch, 
> HBASE-21204.branch-2.002.patch, HBASE-21204.master.001.patch, 
> HBASE-21204.master.002.patch, HBASE-21204.master.003.patch
>
>
> There are 7 types of our Cell,
> Minimum((byte)0),
> Put((byte)4),
> Delete((byte)8),
> DeleteFamilyVersion((byte)10),
> DeleteColumn((byte)12),
> DeleteFamily((byte)14),
> Maximum((byte)255);
> But there are only 6 types of our CellType protobuf definition:
> enum CellType {
> MINIMUM = 0;
> PUT = 4;
> DELETE = 8;
> DELETE_FAMILY_VERSION = 10;
> DELETE_COLUMN = 12;
> DELETE_FAMILY = 14;
> MAXIMUM = 255;
> }
> Thus if we scan raw data which is DELETE_FAMILY_VERSION,it will throw NPE.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-9924) Upgrade zookeeper to 3.4.13

2018-09-19 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-9924?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated FLINK-9924:
--
Description: 
zookeeper 3.4.13 is being released.

ZOOKEEPER-2959 fixes data loss when observer is used

ZOOKEEPER-2184 allows ZooKeeper Java clients to work in dynamic IP (container / 
cloud) environment

  was:
zookeeper 3.4.13 is being released.

ZOOKEEPER-2959 fixes data loss when observer is used
ZOOKEEPER-2184 allows ZooKeeper Java clients to work in dynamic IP (container / 
cloud) environment


> Upgrade zookeeper to 3.4.13
> ---
>
> Key: FLINK-9924
> URL: https://issues.apache.org/jira/browse/FLINK-9924
> Project: Flink
>  Issue Type: Task
>    Reporter: Ted Yu
>Assignee: vinoyang
>Priority: Major
>
> zookeeper 3.4.13 is being released.
> ZOOKEEPER-2959 fixes data loss when observer is used
> ZOOKEEPER-2184 allows ZooKeeper Java clients to work in dynamic IP (container 
> / cloud) environment



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3572) Upgrade commons-dbcp to latest commons-dbcp2

2018-09-19 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3572:
-

 Summary: Upgrade commons-dbcp to latest commons-dbcp2
 Key: KYLIN-3572
 URL: https://issues.apache.org/jira/browse/KYLIN-3572
 Project: Kylin
  Issue Type: Task
Reporter: Ted Yu


There have been many improvements and a major release in commons-dbcp. The 
latest version is 2.5.0.

This Jira proposes to upgrade the version of commons-dbcp to commons-dbcp2 
version 2.5.0 in order to benefit from the improvements/additions. This will 
involve some code changes in addition to the updated dependencies, as the API 
has changed slightly to use different terminology 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (KYLIN-3572) Upgrade commons-dbcp to latest commons-dbcp2

2018-09-19 Thread Ted Yu (JIRA)
Ted Yu created KYLIN-3572:
-

 Summary: Upgrade commons-dbcp to latest commons-dbcp2
 Key: KYLIN-3572
 URL: https://issues.apache.org/jira/browse/KYLIN-3572
 Project: Kylin
  Issue Type: Task
Reporter: Ted Yu


There have been many improvements and a major release in commons-dbcp. The 
latest version is 2.5.0.

This Jira proposes to upgrade the version of commons-dbcp to commons-dbcp2 
version 2.5.0 in order to benefit from the improvements/additions. This will 
involve some code changes in addition to the updated dependencies, as the API 
has changed slightly to use different terminology 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-20952) Re-visit the WAL API

2018-09-19 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-20952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16621260#comment-16621260
 ] 

Ted Yu commented on HBASE-20952:


Thanks for the patience.

I have modified the google doc link above for a high level design.

Thanks to Josh and Sergey for internal reviews.

> Re-visit the WAL API
> 
>
> Key: HBASE-20952
> URL: https://issues.apache.org/jira/browse/HBASE-20952
> Project: HBase
>  Issue Type: Sub-task
>  Components: wal
>Reporter: Josh Elser
>Priority: Major
> Attachments: 20952.v1.txt
>
>
> Take a step back from the current WAL implementations and think about what an 
> HBase WAL API should look like. What are the primitive calls that we require 
> to guarantee durability of writes with a high degree of performance?
> The API needs to take the current implementations into consideration. We 
> should also have a mind for what is happening in the Ratis LogService (but 
> the LogService should not dictate what HBase's WAL API looks like RATIS-272).
> Other "systems" inside of HBase that use WALs are replication and 
> backup Replication has the use-case for "tail"'ing the WAL which we 
> should provide via our new API. B doesn't do anything fancy (IIRC). We 
> should make sure all consumers are generally going to be OK with the API we 
> create.
> The API may be "OK" (or OK in a part). We need to also consider other methods 
> which were "bolted" on such as {{AbstractFSWAL}} and 
> {{WALFileLengthProvider}}. Other corners of "WAL use" (like the 
> {{WALSplitter}} should also be looked at to use WAL-APIs only).
> We also need to make sure that adequate interface audience and stability 
> annotations are chosen.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (AMBARI-18952) Register BackupObserver and BackupHFileCleaner

2018-09-19 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/AMBARI-18952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15884466#comment-15884466
 ] 

Ted Yu edited comment on AMBARI-18952 at 9/19/18 5:40 PM:
--

Note:

There is cost when BackupObserver is registered - it would poll backup table 
(at the end of bulk load per region) for whether the underlying table has gone 
thru full backup.



was (Author: yuzhih...@gmail.com):
Note:

There is cost when BackupObserver is registered - it would poll backup table 
(at the end of bulk load per region) for whether the underlying table has gone 
thru full backup .


> Register BackupObserver and BackupHFileCleaner
> --
>
> Key: AMBARI-18952
> URL: https://issues.apache.org/jira/browse/AMBARI-18952
> Project: Ambari
>  Issue Type: Improvement
>    Reporter: Ted Yu
>Priority: Major
>
> Over in HBASE-14417, two new classes are added.
> org.apache.hadoop.hbase.backup.BackupHFileCleaner should be registered 
> through hbase.master.hfilecleaner.plugins . It is responsible for keeping 
> bulk loaded hfiles so that incremental backup can pick them up.
> org.apache.hadoop.hbase.backup.BackupObserver should be registered through 
> hbase.coprocessor.region.classes
> It is notified when bulk load completes and writes records into hbase:backup 
> table.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (AMBARI-18528) Need a way to escape config values that contain $

2018-09-19 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/AMBARI-18528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15603928#comment-15603928
 ] 

Ted Yu edited comment on AMBARI-18528 at 9/19/18 5:40 PM:
--

Escaping the dollar sign automatically is a desirable solution .


was (Author: yuzhih...@gmail.com):
Escaping the dollar sign automatically is a desirable solution.

> Need a way to escape config values that contain $
> -
>
> Key: AMBARI-18528
> URL: https://issues.apache.org/jira/browse/AMBARI-18528
> Project: Ambari
>  Issue Type: Bug
>    Reporter: Ted Yu
>Priority: Major
>  Labels: zookeeper
>
> We tried specifying auth_to_local in zookeeper env input box:
> {code}
> export SERVER_JVMFLAGS="$SERVER_JVMFLAGS 
> -Dzookeeper.security.auth_to_local=RULE:[2:$1@$0](hb...@c.net)s/.*/hbase/  
> -Djava.security.auth.login.config={{zk_server_jaas_file}}"
> {code}
> However, when zookeeper quorum starts, the rule got interrupted with 
> zkServer.sh
> We should add the capability of specifying auth_to_local in Ambari UI



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AMBARI-22965) Boolean#getBoolean is used to parse value

2018-09-19 Thread Ted Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/AMBARI-22965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated AMBARI-22965:

Description: 
In StormTimelineMetricsReporter.java :

{code}
  if (cf.containsKey(SET_INSTANCE_ID_PROPERTY)) {
setInstanceId = 
Boolean.getBoolean(cf.get(SET_INSTANCE_ID_PROPERTY).toString());
{code}
According to 
https://docs.oracle.com/javase/7/docs/api/java/lang/Boolean.html#getBoolean(java.lang.String):
{code}
Returns true if and only if the system property named by the argument exists 
and is equal to the string "true"
{code}

This was not the intention of the quoted code.

FlumeTimelineMetricsSink.java may have the same issue.

  was:
In StormTimelineMetricsReporter.java :
{code}
  if (cf.containsKey(SET_INSTANCE_ID_PROPERTY)) {
setInstanceId = 
Boolean.getBoolean(cf.get(SET_INSTANCE_ID_PROPERTY).toString());
{code}
According to 
https://docs.oracle.com/javase/7/docs/api/java/lang/Boolean.html#getBoolean(java.lang.String):
{code}
Returns true if and only if the system property named by the argument exists 
and is equal to the string "true"
{code}

This was not the intention of the quoted code.

FlumeTimelineMetricsSink.java may have the same issue.


> Boolean#getBoolean is used to parse value
> -
>
> Key: AMBARI-22965
> URL: https://issues.apache.org/jira/browse/AMBARI-22965
> Project: Ambari
>  Issue Type: Bug
>Reporter: Ted Yu
>Priority: Major
>
> In StormTimelineMetricsReporter.java :
> {code}
>   if (cf.containsKey(SET_INSTANCE_ID_PROPERTY)) {
> setInstanceId = 
> Boolean.getBoolean(cf.get(SET_INSTANCE_ID_PROPERTY).toString());
> {code}
> According to 
> https://docs.oracle.com/javase/7/docs/api/java/lang/Boolean.html#getBoolean(java.lang.String):
> {code}
> Returns true if and only if the system property named by the argument exists 
> and is equal to the string "true"
> {code}
> This was not the intention of the quoted code.
> FlumeTimelineMetricsSink.java may have the same issue.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


Re: from_csv

2018-09-19 Thread Ted Yu
+1
 Original message From: Dongjin Lee  Date: 
9/19/18  7:20 AM  (GMT-08:00) To: dev  Subject: Re: 
from_csv 
Another +1.
I already experienced this case several times.

On Mon, Sep 17, 2018 at 11:03 AM Hyukjin Kwon  wrote:
+1 for this idea since text parsing in CSV/JSON is quite common.
One thing is about schema inference likewise with JSON functionality. In case 
of JSON, we added schema_of_json for it and same thing should be able to apply 
to CSV too.
If we see some more needs for it, we can consider a function like schema_of_csv 
as well.

2018년 9월 16일 (일) 오후 4:41, Maxim Gekk 님이 작성:
Hi Reynold,
> i'd make this as consistent as to_json / from_json as possible
Sure, new function from_csv() has the same signature as from_json().
> how would this work in sql? i.e. how would passing options in work?
The options are passed to the function via map, for example:select 
from_csv('26/08/2015', 'time Timestamp', map('timestampFormat', 'dd/MM/'))

On Sun, Sep 16, 2018 at 7:01 AM Reynold Xin  wrote:
makes sense - i'd make this as consistent as to_json / from_json as possible. 
how would this work in sql? i.e. how would passing options in work?
--excuse the brevity and lower case due to wrist injury

On Sat, Sep 15, 2018 at 2:58 AM Maxim Gekk  wrote:
Hi All,
I would like to propose new function from_csv() for parsing columns containing 
strings in CSV format. Here is my PR: https://github.com/apache/spark/pull/22379
An use case is loading a dataset from an external storage, dbms or systems like 
Kafka to where CSV content was dumped as one of columns/fields. Other columns 
could contain related information like timestamps, ids, sources of data and 
etc. The column with CSV strings can be parsed by existing method csv() of 
DataFrameReader but in that case we have to "clean up" dataset and remove other 
columns since the csv() method requires Dataset[String]. Joining back result of 
parsing and original dataset by positions is expensive and not convenient. 
Instead users parse CSV columns by string functions. The approach is usually 
error prone especially for quoted values and other special cases.
The proposed in the PR methods should make a better user experience in parsing 
CSV-like columns. Please, share your thoughts.
-- 

Maxim Gekk
Technical Solutions LeadDatabricks inc.maxim.g...@databricks.comdatabricks.com 






-- 
Dongjin Lee
A hitchhiker in the mathematical world.
github: github.com/dongjinleekrlinkedin: 
kr.linkedin.com/in/dongjinleekrslideshare: www.slideshare.net/dongjinleekr


[jira] [Commented] (HBASE-21102) ServerCrashProcedure should select target server where no other replicas exist for the current region

2018-09-18 Thread Ted Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21102?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16620116#comment-16620116
 ] 

Ted Yu commented on HBASE-21102:


I will commit tomorrow. 

The failing test is in hbase-rsgroups, a different module. So qa didn't run 
that test.

> ServerCrashProcedure should select target server where no other replicas 
> exist for the current region
> -
>
> Key: HBASE-21102
> URL: https://issues.apache.org/jira/browse/HBASE-21102
> Project: HBase
>  Issue Type: Bug
>  Components: Region Assignment
>Affects Versions: 3.0.0, 2.2.0
>Reporter: ramkrishna.s.vasudevan
>Assignee: ramkrishna.s.vasudevan
>Priority: Major
> Attachments: 21102.addendum2.txt, HBASE-21102_1.patch, 
> HBASE-21102_2.patch, HBASE-21102_3.patch, HBASE-21102_4.patch, 
> HBASE-21102_addendum.patch, HBASE-21102_addendum.patch, 
> HBASE-21102_addendum.patch, HBASE-21102_branch-2.1.patch, 
> HBASE-21102_branch-2.1.patch, HBASE-21102_initial.patch
>
>
> Currently when a server with region replica crashes, when the target server 
> is created for the replica region assignment there is no guarentee that a 
> server is selected where there is no other replica for the current region 
> getting assigned. It so happens that currently we do an assignment randomly 
> and later the LB comes and identifies these cases and again does MOVE for 
> such regions. It will be better if we can identify target servers at least 
> minimally ensuring that replicas are not colocated.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


<    4   5   6   7   8   9   10   11   12   13   >