bq. Caused by: java.net.SocketException: Too many open files Please adjust ulimit.
FYI On Wed, Jul 5, 2017 at 1:33 PM, Jyotirmoy Sundi <sundi...@gmail.com> wrote: > Hi Folks , > > Any idea why the build is failing in release-2.0.0 , i did "mvn clean > package" > > > *Trace* > > [INFO] Running org.apache.beam.sdk.io.hbase.HBaseResultCoderTest > > [INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: > 0.461 s - in org.apache.beam.sdk.io.hbase.HBaseResultCoderTest > > [INFO] Running org.apache.beam.sdk.io.hbase.HBaseIOTest > > [ERROR] Tests run: 17, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: > 4.504 s <<< FAILURE! - in org.apache.beam.sdk.io.hbase.HBaseIOTest > > [ERROR] testReadingWithKeyRange(org.apache.beam.sdk.io.hbase.HBaseIOTest) > Time > elapsed: 4.504 s <<< ERROR! > > java.lang.RuntimeException: > > org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after > attempts=1, exceptions: > > Wed Jul 05 13:31:23 PDT 2017, > RpcRetryingCaller{globalStartTime=1499286683193, pause=100, retries=1}, > java.net.SocketException: Too many open files > > > at > org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult. > waitUntilFinish(DirectRunner.java:330) > > at > org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult. > waitUntilFinish(DirectRunner.java:292) > > at org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:200) > > at org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:63) > > at org.apache.beam.sdk.Pipeline.run(Pipeline.java:295) > > at org.apache.beam.sdk.Pipeline.run(Pipeline.java:281) > > at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:340) > > at > org.apache.beam.sdk.io.hbase.HBaseIOTest.runReadTestLength( > HBaseIOTest.java:418) > > at > org.apache.beam.sdk.io.hbase.HBaseIOTest.testReadingWithKeyRange( > HBaseIOTest.java:253) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java: > 62) > > at > sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > > at java.lang.reflect.Method.invoke(Method.java:498) > > at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall( > FrameworkMethod.java:50) > > at > org.junit.internal.runners.model.ReflectiveCallable.run( > ReflectiveCallable.java:12) > > at > org.junit.runners.model.FrameworkMethod.invokeExplosively( > FrameworkMethod.java:47) > > at > org.junit.internal.runners.statements.InvokeMethod. > evaluate(InvokeMethod.java:17) > > at > org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:321) > > at > org.junit.rules.ExpectedException$ExpectedExceptionStatement. > evaluate(ExpectedException.java:239) > > at org.junit.rules.RunRules.evaluate(RunRules.java:20) > > at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325) > > at > org.junit.runners.BlockJUnit4ClassRunner.runChild( > BlockJUnit4ClassRunner.java:78) > > at > org.junit.runners.BlockJUnit4ClassRunner.runChild( > BlockJUnit4ClassRunner.java:57) > > at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) > > at > org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:393) > > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > at > java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1142) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:617) > > at java.lang.Thread.run(Thread.java:745) > > Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedException: > Failed > after attempts=1, exceptions: > > Wed Jul 05 13:31:23 PDT 2017, > RpcRetryingCaller{globalStartTime=1499286683193, pause=100, retries=1}, > java.net.SocketException: Too many open files > > > at > org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries( > RpcRetryingCaller.java:157) > > at > org.apache.hadoop.hbase.client.ResultBoundedCompletionService > $QueueingFuture.run(ResultBoundedCompletionService.java:65) > > ... 3 more > > Caused by: java.net.SocketException: Too many open files > > at sun.nio.ch.Net.socket0(Native Method) > > at sun.nio.ch.Net.socket(Net.java:411) > > at sun.nio.ch.Net.socket(Net.java:404) > > at sun.nio.ch.SocketChannelImpl.<init>(SocketChannelImpl.java:105) > > at > sun.nio.ch.SelectorProviderImpl.openSocketChannel( > SelectorProviderImpl.java:60) > > at java.nio.channels.SocketChannel.open(SocketChannel.java:145) > > at > org.apache.hadoop.net.StandardSocketFactory.createSocket( > StandardSocketFactory.java:62) > > at > org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection. > setupConnection(RpcClientImpl.java:410) > > at > org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection. > setupIOstreams(RpcClientImpl.java:722) > > at > org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection. > writeRequest(RpcClientImpl.java:906) > > at > org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest( > RpcClientImpl.java:873) > > at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1241) > > at > org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod( > AbstractRpcClient.java:227) > > at > org.apache.hadoop.hbase.ipc.AbstractRpcClient$ > BlockingRpcChannelImplementation.callBlockingMethod( > AbstractRpcClient.java:336) > > at > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$ > BlockingStub.scan(ClientProtos.java:34094) > > at > org.apache.hadoop.hbase.client.ClientSmallReversedScanner$ > SmallReversedScannerCallable.call(ClientSmallReversedScanner.java:298) > > at > org.apache.hadoop.hbase.client.ClientSmallReversedScanner$ > SmallReversedScannerCallable.call(ClientSmallReversedScanner.java:276) > > at > org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries( > RpcRetryingCaller.java:210) > > at > org.apache.hadoop.hbase.client.ScannerCallableWithReplicas$ > RetryingRPC.call(ScannerCallableWithReplicas.java:364) > > at > org.apache.hadoop.hbase.client.ScannerCallableWithReplicas$ > RetryingRPC.call(ScannerCallableWithReplicas.java:338) > > at > org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries( > RpcRetryingCaller.java:136) > > ... 4 more > > > [INFO] Running org.apache.beam.sdk.io.hbase.SerializableScanTest > > [INFO] Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: > 0.368 s - in org.apache.beam.sdk.io.hbase.SerializableScanTest > > [INFO] Running org.apache.beam.sdk.io.hbase.HBaseMutationCoderTest > > [INFO] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: > 0.466 s - in org.apache.beam.sdk.io.hbase.HBaseMutationCoderTest > > [INFO] > > [INFO] Results: > > [INFO] > > [ERROR] Errors: > > [ERROR] HBaseIOTest.testReadingWithKeyRange:253->runReadTestLength:418 ยป > Runtime org.a... > > [INFO] > > [ERROR] Tests run: 21, Failures: 0, Errors: 1, Skipped: 0 > > [INFO] > > [INFO] > ------------------------------------------------------------------------ > > [INFO] Reactor Summary: > > [INFO] > > [INFO] Apache Beam :: Parent .............................. SUCCESS [ > 2.100 > s] > > [INFO] Apache Beam :: SDKs :: Java :: Build Tools ......... SUCCESS [ > 2.459 > s] > > [INFO] Apache Beam :: SDKs ................................ SUCCESS [ > 0.407 > s] > > [INFO] Apache Beam :: SDKs :: Common ...................... SUCCESS [ > 0.224 > s] > > [INFO] Apache Beam :: SDKs :: Common :: Fn API ............ SUCCESS [ > 5.013 > s] > > [INFO] Apache Beam :: SDKs :: Common :: Runner API ........ SUCCESS [ > 1.707 > s] > > [INFO] Apache Beam :: SDKs :: Java ........................ SUCCESS [ > 0.342 > s] > > [INFO] Apache Beam :: SDKs :: Java :: Core ................ SUCCESS [ > 54.456 s] > > [INFO] Apache Beam :: SDKs :: Java :: IO .................. SUCCESS [ > 0.344 > s] > > [INFO] Apache Beam :: SDKs :: Java :: IO :: Common ........ SUCCESS [ > 1.860 > s] > > [INFO] Apache Beam :: Runners ............................. SUCCESS [ > 0.329 > s] > > [INFO] Apache Beam :: Runners :: Core Construction Java ... SUCCESS [ > 6.296 > s] > > [INFO] Apache Beam :: Runners :: Core Java ................ SUCCESS [ > 18.418 s] > > [INFO] Apache Beam :: Runners :: Direct Java .............. SUCCESS [ > 25.190 s] > > [INFO] Apache Beam :: SDKs :: Java :: IO :: Elasticsearch . SUCCESS [ > 13.994 s] > > [INFO] Apache Beam :: SDKs :: Java :: Extensions .......... SUCCESS [ > 0.285 > s] > > [INFO] Apache Beam :: SDKs :: Java :: Extensions :: Google Cloud Platform > Core SUCCESS [ 14.985 s] > > [INFO] Apache Beam :: SDKs :: Java :: Extensions :: Protobuf SUCCESS [ > 5.334 > s] > > [INFO] Apache Beam :: SDKs :: Java :: IO :: Google Cloud Platform SUCCESS [ > 29.326 s] > > [INFO] Apache Beam :: SDKs :: Java :: IO :: Hadoop Common . SUCCESS [ > 5.403 > s] > > [INFO] Apache Beam :: SDKs :: Java :: IO :: Hadoop File System SUCCESS [ > 14.641 s] > > [INFO] Apache Beam :: SDKs :: Java :: IO :: Hadoop ........ SUCCESS [ > 1.484 > s] > > [INFO] Apache Beam :: SDKs :: Java :: IO :: Hadoop :: input-format SUCCESS > [ > 6.105 s] > > [INFO] Apache Beam :: SDKs :: Java :: IO :: Hadoop :: jdk1.8-tests SUCCESS > [ 39.188 s] > > [INFO] Apache Beam :: SDKs :: Java :: IO :: HBase ......... FAILURE [ > 26.018 s] > > [INFO] Apache Beam :: SDKs :: Java :: IO :: JDBC .......... SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: IO :: JMS ........... SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: IO :: Kafka ......... SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: IO :: Kinesis ....... SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: IO :: MongoDB ....... SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: IO :: MQTT .......... SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: IO :: XML ........... SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: Maven Archetypes .... SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: Maven Archetypes :: Starter SKIPPED > > [INFO] Apache Beam :: Examples ............................ SKIPPED > > [INFO] Apache Beam :: Examples :: Java .................... SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: Maven Archetypes :: Examples SKIPPED > > [INFO] Apache Beam :: Examples :: Java 8 .................. SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: Maven Archetypes :: Examples - Java 8 > SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: Extensions :: Jackson SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: Extensions :: Join library SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: Extensions :: Sorter SKIPPED > > [INFO] Apache Beam :: Runners :: Google Cloud Dataflow .... SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: Harness ............. SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: Java 8 Tests ........ SKIPPED > > [INFO] Apache Beam :: SDKs :: Python ...................... SKIPPED > > [INFO] Apache Beam :: Runners :: Flink .................... SKIPPED > > [INFO] Apache Beam :: Runners :: Spark .................... SKIPPED > > [INFO] Apache Beam :: Runners :: Apex ..................... SKIPPED > > [INFO] Apache Beam :: SDKs :: Java :: Aggregated Javadoc .. SKIPPED > > [INFO] > ------------------------------------------------------------------------ > > [INFO] BUILD FAILURE > > [INFO] > ------------------------------------------------------------------------ > > [INFO] Total time: 04:36 min > > [INFO] Finished at: 2017-07-05T13:31:29-07:00 > > [INFO] Final Memory: 195M/1978M > > [INFO] > ------------------------------------------------------------------------ > > [ERROR] Failed to execute goal > org.apache.maven.plugins:maven-surefire-plugin:2.20:test (default-test) on > project beam-sdks-java-io-hbase: There are test failures. > > [ERROR] > > [ERROR] Please refer to > /Users/jsundi/git/beam/sdks/java/io/hbase/target/surefire-reports for the > individual test results. > > [ERROR] Please refer to dump files (if any exist) [date]-jvmRun[N].dump, > [date].dumpstream and [date]-jvmRun[N].dumpstream. > > [ERROR] -> [Help 1] > > [ERROR] > > [ERROR] To see the full stack trace of the errors, re-run Maven with the -e > switch. > > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > > [ERROR] > > [ERROR] For more information about the errors and possible solutions, > please read the following articles: > > [ERROR] [Help 1] > http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException > > [ERROR] > > [ERROR] After correcting the problems, you can resume the build with the > command > > [ERROR] mvn <goals> -rf :beam-sdks-java-io-hbase > > MTVL14996a7cf:beam jsundi$ git branch > > master > > * release-2.0.0 > > MTVL14996a7cf:beam jsundi$ > > -- > Best Regards, > Jyotirmoy Sundi >