thanks

the machine where spark job was being submitted had SPARK_HOME pointing old
2.1.1 directory.



On Wed, Dec 6, 2017 at 1:35 PM, Qiao, Richard <richard.q...@capitalone.com>
wrote:

> Are you now building your app using spark 2.2 or 2.1?
>
>
>
> Best Regards
>
> Richard
>
>
>
>
>
> *From: *Imran Rajjad <raj...@gmail.com>
> *Date: *Wednesday, December 6, 2017 at 2:45 AM
> *To: *"user @spark" <user@spark.apache.org>
> *Subject: *unable to connect to connect to cluster 2.2.0
>
>
>
> Hi,
>
>
>
> Recently upgraded from 2.1.1 to 2.2.0. My Streaming job seems to have
> broken. The submitted application is unable to connect to the cluster, when
> all is running.
>
>
>
> below is my stack trace
>
> Spark Master:spark://192.168.10.207:7077
> Job Arguments:
> -appName orange_watch -directory /u01/watch/stream/
> Spark Configuration:
> [spark.executor.memory, spark.driver.memory, spark.app.name,
> spark.executor.cores]:6g
> [spark.executor.memory, spark.driver.memory, spark.app.name,
> spark.executor.cores]:4g
> [spark.executor.memory, spark.driver.memory, spark.app.name,
> spark.executor.cores]:orange_watch
> [spark.executor.memory, spark.driver.memory, spark.app.name,
> spark.executor.cores]:2
>
> Spark Arguments:
> [--packages]:graphframes:graphframes:0.5.0-spark2.1-s_2.11
>
> Using properties file: /home/my_user/spark-2.2.0-bin-
> hadoop2.7/conf/spark-defaults.conf
> Adding default property: spark.jars.packages=
> graphframes:graphframes:0.5.0-spark2.1-s_2.11
> Parsed arguments:
>   master                  spark://192.168.10.207:7077
>   deployMode              null
>   executorMemory          6g
>   executorCores           2
>   totalExecutorCores      null
>   propertiesFile          /home/my_user/spark-2.2.0-bin-
> hadoop2.7/conf/spark-defaults.conf
>   driverMemory            4g
>   driverCores             null
>   driverExtraClassPath    null
>   driverExtraLibraryPath  null
>   driverExtraJavaOptions  null
>   supervise               false
>   queue                   null
>   numExecutors            null
>   files                   null
>   pyFiles                 null
>   archives                null
>   mainClass               com.my_user.MainClassWatch
>   primaryResource         file:/home/my_user/cluster-testing/job.jar
>   name                    orange_watch
>   childArgs               [-watchId 3199 -appName orange_watch -directory
> /u01/watch/stream/]
>   jars                    null
>   packages                graphframes:graphframes:0.5.0-spark2.1-s_2.11
>   packagesExclusions      null
>   repositories            null
>   verbose                 true
>
> Spark properties used, including those specified through
>  --conf and those from the properties file /home/my_user/spark-2.2.0-bin-
> hadoop2.7/conf/spark-defaults.conf:
>   (spark.driver.memory,4g)
>   (spark.executor.memory,6g)
>   (spark.jars.packages,graphframes:graphframes:0.5.0-spark2.1-s_2.11)
>   (spark.app.name,orange_watch)
>   (spark.executor.cores,2)
>
>
> Ivy Default Cache set to: /home/my_user/.ivy2/cache
> The jars for the packages stored in: /home/my_user/.ivy2/jars
> :: loading settings :: url = jar:file:/home/my_user/spark-
> 2.2.0-bin-hadoop2.7/jars/ivy-2.4.0.jar!/org/apache/ivy/
> core/settings/ivysettings.xml
> graphframes#graphframes added as a dependency
> :: resolving dependencies :: org.apache.spark#spark-submit-parent;1.0
>         confs: [default]
>         found graphframes#graphframes;0.5.0-spark2.1-s_2.11 in spark-list
>         found com.typesafe.scala-logging#scala-logging-api_2.11;2.1.2 in
> central
>         found com.typesafe.scala-logging#scala-logging-slf4j_2.11;2.1.2
> in central
>         found org.scala-lang#scala-reflect;2.11.0 in central
>         found org.slf4j#slf4j-api;1.7.7 in spark-list
> :: resolution report :: resolve 191ms :: artifacts dl 5ms
>         :: modules in use:
>         com.typesafe.scala-logging#scala-logging-api_2.11;2.1.2 from
> central in [default]
>         com.typesafe.scala-logging#scala-logging-slf4j_2.11;2.1.2 from
> central in [default]
>         graphframes#graphframes;0.5.0-spark2.1-s_2.11 from spark-list in
> [default]
>         org.scala-lang#scala-reflect;2.11.0 from central in [default]
>         org.slf4j#slf4j-api;1.7.7 from spark-list in [default]
>         ------------------------------------------------------------
> ---------
>         |                  |            modules            ||
> artifacts   |
>         |       conf       | number| search|dwnlded|evicted||
> number|dwnlded|
>         ------------------------------------------------------------
> ---------
>         |      default     |   5   |   0   |   0   |   0   ||   5   |
> 0   |
>         ------------------------------------------------------------
> ---------
> :: retrieving :: org.apache.spark#spark-submit-parent
>         confs: [default]
>         0 artifacts copied, 5 already retrieved (0kB/7ms)
> Main class:
> com.my_user.MainClassWatch
> Arguments:
> -watchId
> 3199
> -appName
> orange_watch
> -directory
> /u01/watch/stream/
> System properties:
> (spark.executor.memory,6g)
> (spark.driver.memory,4g)
> (SPARK_SUBMIT,true)
> (spark.jars.packages,graphframes:graphframes:0.5.0-spark2.1-s_2.11)
> (spark.app.name,orange_watch)
> (spark.jars,file:/home/my_user/.ivy2/jars/graphframes_
> graphframes-0.5.0-spark2.1-s_2.11.jar,file:/home/my_user/.
> ivy2/jars/com.typesafe.scala-logging_scala-logging-api_2.
> 11-2.1.2.jar,file:/home/my_user/.ivy2/jars/com.typesafe.
> scala-logging_scala-logging-slf4j_2.11-2.1.2.jar,file:/
> home/my_user/.ivy2/jars/org.scala-lang_scala-reflect-2.11.
> 0.jar,file:/home/my_user/.ivy2/jars/org.slf4j_slf4j-api-
> 1.7.7.jar,file:/home/my_user/cluster-testing/job.jar)
> (spark.submit.deployMode,client)
> (spark.master,spark://192.168.10.207:7077)
> (spark.executor.cores,2)
> Classpath elements:
> file:/home/my_user/cluster-testing/job.jar
> /home/my_user/.ivy2/jars/graphframes_graphframes-0.5.0-spark2.1-s_2.11.jar
> /home/my_user/.ivy2/jars/com.typesafe.scala-logging_scala-
> logging-api_2.11-2.1.2.jar
> /home/my_user/.ivy2/jars/com.typesafe.scala-logging_scala-
> logging-slf4j_2.11-2.1.2.jar
> /home/my_user/.ivy2/jars/org.scala-lang_scala-reflect-2.11.0.jar
> /home/my_user/.ivy2/jars/org.slf4j_slf4j-api-1.7.7.jar
>
>
> Using Spark's default log4j profile: org/apache/spark/log4j-
> defaults.properties
> 17/12/06 12:36:50 INFO SparkContext: Running Spark version 2.2.0
> 17/12/06 12:36:50 INFO SparkContext: Submitted application: orange_watch
> 17/12/06 12:36:50 INFO SecurityManager: Changing view acls to: my_user
> 17/12/06 12:36:50 INFO SecurityManager: Changing modify acls to: my_user
> 17/12/06 12:36:50 INFO SecurityManager: Changing view acls groups to:
> 17/12/06 12:36:50 INFO SecurityManager: Changing modify acls groups to:
> 17/12/06 12:36:50 INFO SecurityManager: SecurityManager: authentication
> disabled; ui acls disabled; users  with view permissions: Set(my_user);
> groups with view permissions: Set(); users  with modify permissions:
> Set(my_user); groups with modify permissions: Set()
> 17/12/06 12:36:50 INFO Utils: Successfully started service 'sparkDriver'
> on port 37329.
> 17/12/06 12:36:50 INFO SparkEnv: Registering MapOutputTracker
> 17/12/06 12:36:50 INFO SparkEnv: Registering BlockManagerMaster
> 17/12/06 12:36:50 INFO BlockManagerMasterEndpoint: Using
> org.apache.spark.storage.DefaultTopologyMapper for getting topology
> information
> 17/12/06 12:36:50 INFO BlockManagerMasterEndpoint:
> BlockManagerMasterEndpoint up
> 17/12/06 12:36:50 INFO DiskBlockManager: Created local directory at
> /tmp/blockmgr-f441dcc1-71c8-437e-ad7d-1057ab2b0f87
> 17/12/06 12:36:50 INFO MemoryStore: MemoryStore started with capacity
> 2004.6 MB
> 17/12/06 12:36:50 INFO SparkEnv: Registering OutputCommitCoordinator
> 17/12/06 12:36:50 INFO Utils: Successfully started service 'SparkUI' on
> port 4040.
> 17/12/06 12:36:50 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started at
> http://192.168.10.207:4040
> 17/12/06 12:36:50 INFO SparkContext: Added JAR
> file:/home/my_user/.ivy2/jars/graphframes_graphframes-0.5.0-spark2.1-s_2.11.jar
> at spark://192.168.10.207:37329/jars/graphframes_graphframes-
> 0.5.0-spark2.1-s_2.11.jar with timestamp 1512545810823
> 17/12/06 12:36:50 INFO SparkContext: Added JAR
> file:/home/my_user/.ivy2/jars/com.typesafe.scala-logging_
> scala-logging-api_2.11-2.1.2.jar at spark://192.168.10.207:37329/
> jars/com.typesafe.scala-logging_scala-logging-api_2.11-2.1.2.jar with
> timestamp 1512545810824
> 17/12/06 12:36:50 INFO SparkContext: Added JAR
> file:/home/my_user/.ivy2/jars/com.typesafe.scala-logging_
> scala-logging-slf4j_2.11-2.1.2.jar at spark://192.168.10.207:37329/
> jars/com.typesafe.scala-logging_scala-logging-slf4j_2.11-2.1.2.jar with
> timestamp 1512545810824
> 17/12/06 12:36:50 INFO SparkContext: Added JAR
> file:/home/my_user/.ivy2/jars/org.scala-lang_scala-reflect-2.11.0.jar at
> spark://192.168.10.207:37329/jars/org.scala-lang_scala-reflect-2.11.0.jar
> with timestamp 1512545810824
> 17/12/06 12:36:50 INFO SparkContext: Added JAR
> file:/home/my_user/.ivy2/jars/org.slf4j_slf4j-api-1.7.7.jar at spark://
> 192.168.10.207:37329/jars/org.slf4j_slf4j-api-1.7.7.jar with timestamp
> 1512545810824
> 17/12/06 12:36:50 INFO SparkContext: Added JAR 
> file:/home/my_user/cluster-testing/job.jar
> at spark://192.168.10.207:37329/jars/job.jar with timestamp 1512545810824
> 17/12/06 12:36:50 INFO StandaloneAppClient$ClientEndpoint: Connecting to
> master spark://192.168.10.207:7077...
> 17/12/06 12:36:50 INFO TransportClientFactory: Successfully created
> connection to /192.168.10.207:7077 after 16 ms (0 ms spent in bootstraps)
> 17/12/06 12:36:50 WARN StandaloneAppClient$ClientEndpoint: Failed to
> connect to master 192.168.10.207:7077
> org.apache.spark.SparkException: Exception thrown in awaitResult:
>         at org.apache.spark.util.ThreadUtils$.awaitResult(
> ThreadUtils.scala:205)
>         at org.apache.spark.rpc.RpcTimeout.awaitResult(
> RpcTimeout.scala:75)
>         at org.apache.spark.rpc.RpcEnv.setupEndpointRefByURI(RpcEnv.
> scala:100)
>         at org.apache.spark.rpc.RpcEnv.setupEndpointRef(RpcEnv.scala:108)
>         at org.apache.spark.deploy.client.StandaloneAppClient$
> ClientEndpoint$$anonfun$tryRegisterAllMasters$1$$anon$
> 1.run(StandaloneAppClient.scala:106)
>         at java.util.concurrent.Executors$RunnableAdapter.
> call(Executors.java:511)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>         at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1149)
>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:624)
>         at java.lang.Thread.run(Thread.java:748)
> Caused by: java.lang.RuntimeException: java.io.StreamCorruptedException:
> invalid stream header: 01000E31
>         at java.io.ObjectInputStream.readStreamHeader(
> ObjectInputStream.java:857)
>         at java.io.ObjectInputStream.<init>(ObjectInputStream.java:349)
>         at org.apache.spark.serializer.JavaDeserializationStream$$
> anon$1.<init>(JavaSerializer.scala:63)
>         at org.apache.spark.serializer.JavaDeserializationStream.<
> init>(JavaSerializer.scala:63)
>         at org.apache.spark.serializer.JavaSerializerInstance.
> deserializeStream(JavaSerializer.scala:122)
>         at org.apache.spark.serializer.JavaSerializerInstance.
> deserialize(JavaSerializer.scala:107)
>         at org.apache.spark.rpc.netty.NettyRpcEnv$$anonfun$
> deserialize$1$$anonfun$apply$1.apply(NettyRpcEnv.scala:259)
>         at scala.util.DynamicVariable.withValue(DynamicVariable.scala:58)
>         at org.apache.spark.rpc.netty.NettyRpcEnv.deserialize(
> NettyRpcEnv.scala:308)
>         at org.apache.spark.rpc.netty.NettyRpcEnv$$anonfun$
> deserialize$1.apply(NettyRpcEnv.scala:258)
>         at scala.util.DynamicVariable.withValue(DynamicVariable.scala:58)
>         at org.apache.spark.rpc.netty.NettyRpcEnv.deserialize(
> NettyRpcEnv.scala:257)
>         at org.apache.spark.rpc.netty.NettyRpcHandler.
> internalReceive(NettyRpcEnv.scala:577)
>         at org.apache.spark.rpc.netty.NettyRpcHandler.receive(
> NettyRpcEnv.scala:562)
>         at org.apache.spark.network.server.TransportRequestHandler.
> processRpcRequest(TransportRequestHandler.java:159)
>         at org.apache.spark.network.server.TransportRequestHandler.handle(
> TransportRequestHandler.java:107)
>         at org.apache.spark.network.server.TransportChannelHandler.
> channelRead(TransportChannelHandler.java:118)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:346)
>         at io.netty.handler.timeout.IdleStateHandler.channelRead(
> IdleStateHandler.java:266)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:346)
>         at io.netty.handler.codec.MessageToMessageDecoder.channelRead(
> MessageToMessageDecoder.java:102)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:346)
>         at org.apache.spark.network.util.TransportFrameDecoder.
> channelRead(TransportFrameDecoder.java:85)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:346)
>         at io.netty.channel.DefaultChannelPipeline$
> HeadContext.channelRead(DefaultChannelPipeline.java:1294)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.DefaultChannelPipeline.fireChannelRead(
> DefaultChannelPipeline.java:911)
>         at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(
> AbstractNioByteChannel.java:131)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKey(
> NioEventLoop.java:652)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(
> NioEventLoop.java:575)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKeys(
> NioEventLoop.java:489)
>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:451)
>         at io.netty.util.concurrent.SingleThreadEventExecutor$2.
> run(SingleThreadEventExecutor.java:140)
>         at io.netty.util.concurrent.DefaultThreadFactory$
> DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
>         at java.lang.Thread.run(Thread.java:748)
>
>         at org.apache.spark.network.client.TransportResponseHandler.
> handle(TransportResponseHandler.java:207)
>         at org.apache.spark.network.server.TransportChannelHandler.
> channelRead(TransportChannelHandler.java:120)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:336)
>         at io.netty.handler.timeout.IdleStateHandler.channelRead(
> IdleStateHandler.java:287)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:336)
>         at io.netty.handler.codec.MessageToMessageDecoder.channelRead(
> MessageToMessageDecoder.java:102)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:336)
>         at org.apache.spark.network.util.TransportFrameDecoder.
> channelRead(TransportFrameDecoder.java:85)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:336)
>         at io.netty.channel.DefaultChannelPipeline$
> HeadContext.channelRead(DefaultChannelPipeline.java:1294)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.DefaultChannelPipeline.fireChannelRead(
> DefaultChannelPipeline.java:911)
>         at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(
> AbstractNioByteChannel.java:131)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKey(
> NioEventLoop.java:643)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(
> NioEventLoop.java:566)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKeys(
> NioEventLoop.java:480)
>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
>         at io.netty.util.concurrent.SingleThreadEventExecutor$2.
> run(SingleThreadEventExecutor.java:131)
>         at io.netty.util.concurrent.DefaultThreadFactory$
> DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
>         ... 1 more
> 17/12/06 12:37:10 INFO StandaloneAppClient$ClientEndpoint: Connecting to
> master spark://192.168.10.207:7077...
> 17/12/06 12:37:10 WARN StandaloneAppClient$ClientEndpoint: Failed to
> connect to master 192.168.10.207:7077
> org.apache.spark.SparkException: Exception thrown in awaitResult:
>         at org.apache.spark.util.ThreadUtils$.awaitResult(
> ThreadUtils.scala:205)
>         at org.apache.spark.rpc.RpcTimeout.awaitResult(
> RpcTimeout.scala:75)
>         at org.apache.spark.rpc.RpcEnv.setupEndpointRefByURI(RpcEnv.
> scala:100)
>         at org.apache.spark.rpc.RpcEnv.setupEndpointRef(RpcEnv.scala:108)
>         at org.apache.spark.deploy.client.StandaloneAppClient$
> ClientEndpoint$$anonfun$tryRegisterAllMasters$1$$anon$
> 1.run(StandaloneAppClient.scala:106)
>         at java.util.concurrent.Executors$RunnableAdapter.
> call(Executors.java:511)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>         at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1149)
>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:624)
>         at java.lang.Thread.run(Thread.java:748)
> Caused by: java.lang.RuntimeException: java.io.StreamCorruptedException:
> invalid stream header: 01000E31
>         at java.io.ObjectInputStream.readStreamHeader(
> ObjectInputStream.java:857)
>         at java.io.ObjectInputStream.<init>(ObjectInputStream.java:349)
>         at org.apache.spark.serializer.JavaDeserializationStream$$
> anon$1.<init>(JavaSerializer.scala:63)
>         at org.apache.spark.serializer.JavaDeserializationStream.<
> init>(JavaSerializer.scala:63)
>         at org.apache.spark.serializer.JavaSerializerInstance.
> deserializeStream(JavaSerializer.scala:122)
>         at org.apache.spark.serializer.JavaSerializerInstance.
> deserialize(JavaSerializer.scala:107)
>         at org.apache.spark.rpc.netty.NettyRpcEnv$$anonfun$
> deserialize$1$$anonfun$apply$1.apply(NettyRpcEnv.scala:259)
>         at scala.util.DynamicVariable.withValue(DynamicVariable.scala:58)
>         at org.apache.spark.rpc.netty.NettyRpcEnv.deserialize(
> NettyRpcEnv.scala:308)
>         at org.apache.spark.rpc.netty.NettyRpcEnv$$anonfun$
> deserialize$1.apply(NettyRpcEnv.scala:258)
>         at scala.util.DynamicVariable.withValue(DynamicVariable.scala:58)
>         at org.apache.spark.rpc.netty.NettyRpcEnv.deserialize(
> NettyRpcEnv.scala:257)
>         at org.apache.spark.rpc.netty.NettyRpcHandler.
> internalReceive(NettyRpcEnv.scala:577)
>         at org.apache.spark.rpc.netty.NettyRpcHandler.receive(
> NettyRpcEnv.scala:562)
>         at org.apache.spark.network.server.TransportRequestHandler.
> processRpcRequest(TransportRequestHandler.java:159)
>         at org.apache.spark.network.server.TransportRequestHandler.handle(
> TransportRequestHandler.java:107)
>         at org.apache.spark.network.server.TransportChannelHandler.
> channelRead(TransportChannelHandler.java:118)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:346)
>         at io.netty.handler.timeout.IdleStateHandler.channelRead(
> IdleStateHandler.java:266)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:346)
>         at io.netty.handler.codec.MessageToMessageDecoder.channelRead(
> MessageToMessageDecoder.java:102)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:346)
>         at org.apache.spark.network.util.TransportFrameDecoder.
> channelRead(TransportFrameDecoder.java:85)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:346)
>         at io.netty.channel.DefaultChannelPipeline$
> HeadContext.channelRead(DefaultChannelPipeline.java:1294)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.DefaultChannelPipeline.fireChannelRead(
> DefaultChannelPipeline.java:911)
>         at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(
> AbstractNioByteChannel.java:131)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKey(
> NioEventLoop.java:652)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(
> NioEventLoop.java:575)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKeys(
> NioEventLoop.java:489)
>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:451)
>         at io.netty.util.concurrent.SingleThreadEventExecutor$2.
> run(SingleThreadEventExecutor.java:140)
>         at io.netty.util.concurrent.DefaultThreadFactory$
> DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
>         at java.lang.Thread.run(Thread.java:748)
>
>         at org.apache.spark.network.client.TransportResponseHandler.
> handle(TransportResponseHandler.java:207)
>         at org.apache.spark.network.server.TransportChannelHandler.
> channelRead(TransportChannelHandler.java:120)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:336)
>         at io.netty.handler.timeout.IdleStateHandler.channelRead(
> IdleStateHandler.java:287)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:336)
>         at io.netty.handler.codec.MessageToMessageDecoder.channelRead(
> MessageToMessageDecoder.java:102)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:336)
>         at org.apache.spark.network.util.TransportFrameDecoder.
> channelRead(TransportFrameDecoder.java:85)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:336)
>         at io.netty.channel.DefaultChannelPipeline$
> HeadContext.channelRead(DefaultChannelPipeline.java:1294)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.DefaultChannelPipeline.fireChannelRead(
> DefaultChannelPipeline.java:911)
>         at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(
> AbstractNioByteChannel.java:131)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKey(
> NioEventLoop.java:643)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(
> NioEventLoop.java:566)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKeys(
> NioEventLoop.java:480)
>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
>         at io.netty.util.concurrent.SingleThreadEventExecutor$2.
> run(SingleThreadEventExecutor.java:131)
>         at io.netty.util.concurrent.DefaultThreadFactory$
> DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
>         ... 1 more
> 17/12/06 12:37:30 INFO StandaloneAppClient$ClientEndpoint: Connecting to
> master spark://192.168.10.207:7077...
> 17/12/06 12:37:30 WARN StandaloneAppClient$ClientEndpoint: Failed to
> connect to master 192.168.10.207:7077
> org.apache.spark.SparkException: Exception thrown in awaitResult:
>         at org.apache.spark.util.ThreadUtils$.awaitResult(
> ThreadUtils.scala:205)
>         at org.apache.spark.rpc.RpcTimeout.awaitResult(
> RpcTimeout.scala:75)
>         at org.apache.spark.rpc.RpcEnv.setupEndpointRefByURI(RpcEnv.
> scala:100)
>         at org.apache.spark.rpc.RpcEnv.setupEndpointRef(RpcEnv.scala:108)
>         at org.apache.spark.deploy.client.StandaloneAppClient$
> ClientEndpoint$$anonfun$tryRegisterAllMasters$1$$anon$
> 1.run(StandaloneAppClient.scala:106)
>         at java.util.concurrent.Executors$RunnableAdapter.
> call(Executors.java:511)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>         at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1149)
>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:624)
>         at java.lang.Thread.run(Thread.java:748)
> Caused by: java.lang.RuntimeException: java.io.StreamCorruptedException:
> invalid stream header: 01000E31
>         at java.io.ObjectInputStream.readStreamHeader(
> ObjectInputStream.java:857)
>         at java.io.ObjectInputStream.<init>(ObjectInputStream.java:349)
>         at org.apache.spark.serializer.JavaDeserializationStream$$
> anon$1.<init>(JavaSerializer.scala:63)
>         at org.apache.spark.serializer.JavaDeserializationStream.<
> init>(JavaSerializer.scala:63)
>         at org.apache.spark.serializer.JavaSerializerInstance.
> deserializeStream(JavaSerializer.scala:122)
>         at org.apache.spark.serializer.JavaSerializerInstance.
> deserialize(JavaSerializer.scala:107)
>         at org.apache.spark.rpc.netty.NettyRpcEnv$$anonfun$
> deserialize$1$$anonfun$apply$1.apply(NettyRpcEnv.scala:259)
>         at scala.util.DynamicVariable.withValue(DynamicVariable.scala:58)
>         at org.apache.spark.rpc.netty.NettyRpcEnv.deserialize(
> NettyRpcEnv.scala:308)
>         at org.apache.spark.rpc.netty.NettyRpcEnv$$anonfun$
> deserialize$1.apply(NettyRpcEnv.scala:258)
>         at scala.util.DynamicVariable.withValue(DynamicVariable.scala:58)
>         at org.apache.spark.rpc.netty.NettyRpcEnv.deserialize(
> NettyRpcEnv.scala:257)
>         at org.apache.spark.rpc.netty.NettyRpcHandler.
> internalReceive(NettyRpcEnv.scala:577)
>         at org.apache.spark.rpc.netty.NettyRpcHandler.receive(
> NettyRpcEnv.scala:562)
>         at org.apache.spark.network.server.TransportRequestHandler.
> processRpcRequest(TransportRequestHandler.java:159)
>         at org.apache.spark.network.server.TransportRequestHandler.handle(
> TransportRequestHandler.java:107)
>         at org.apache.spark.network.server.TransportChannelHandler.
> channelRead(TransportChannelHandler.java:118)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:346)
>         at io.netty.handler.timeout.IdleStateHandler.channelRead(
> IdleStateHandler.java:266)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:346)
>         at io.netty.handler.codec.MessageToMessageDecoder.channelRead(
> MessageToMessageDecoder.java:102)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:346)
>         at org.apache.spark.network.util.TransportFrameDecoder.
> channelRead(TransportFrameDecoder.java:85)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:346)
>         at io.netty.channel.DefaultChannelPipeline$
> HeadContext.channelRead(DefaultChannelPipeline.java:1294)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:367)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:353)
>         at io.netty.channel.DefaultChannelPipeline.fireChannelRead(
> DefaultChannelPipeline.java:911)
>         at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(
> AbstractNioByteChannel.java:131)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKey(
> NioEventLoop.java:652)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(
> NioEventLoop.java:575)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKeys(
> NioEventLoop.java:489)
>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:451)
>         at io.netty.util.concurrent.SingleThreadEventExecutor$2.
> run(SingleThreadEventExecutor.java:140)
>         at io.netty.util.concurrent.DefaultThreadFactory$
> DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
>         at java.lang.Thread.run(Thread.java:748)
>
>         at org.apache.spark.network.client.TransportResponseHandler.
> handle(TransportResponseHandler.java:207)
>         at org.apache.spark.network.server.TransportChannelHandler.
> channelRead(TransportChannelHandler.java:120)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:336)
>         at io.netty.handler.timeout.IdleStateHandler.channelRead(
> IdleStateHandler.java:287)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:336)
>         at io.netty.handler.codec.MessageToMessageDecoder.channelRead(
> MessageToMessageDecoder.java:102)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:336)
>         at org.apache.spark.network.util.TransportFrameDecoder.
> channelRead(TransportFrameDecoder.java:85)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(
> AbstractChannelHandlerContext.java:336)
>         at io.netty.channel.DefaultChannelPipeline$
> HeadContext.channelRead(DefaultChannelPipeline.java:1294)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:357)
>         at io.netty.channel.AbstractChannelHandlerContext.
> invokeChannelRead(AbstractChannelHandlerContext.java:343)
>         at io.netty.channel.DefaultChannelPipeline.fireChannelRead(
> DefaultChannelPipeline.java:911)
>         at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(
> AbstractNioByteChannel.java:131)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKey(
> NioEventLoop.java:643)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(
> NioEventLoop.java:566)
>         at io.netty.channel.nio.NioEventLoop.processSelectedKeys(
> NioEventLoop.java:480)
>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
>         at io.netty.util.concurrent.SingleThreadEventExecutor$2.
> run(SingleThreadEventExecutor.java:131)
>         at io.netty.util.concurrent.DefaultThreadFactory$
> DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
>         ... 1 more
> 17/12/06 12:37:50 ERROR StandaloneSchedulerBackend: Application has been
> killed. Reason: All masters are unresponsive! Giving up.
> 17/12/06 12:37:50 WARN StandaloneSchedulerBackend: Application ID is not
> initialized yet.
> 17/12/06 12:37:50 INFO Utils: Successfully started service
> 'org.apache.spark.network.netty.NettyBlockTransferService' on port 42644.
> 17/12/06 12:37:50 INFO NettyBlockTransferService: Server created on
> 192.168.10.207:42644
> 17/12/06 12:37:50 INFO BlockManager: Using 
> org.apache.spark.storage.RandomBlockReplicationPolicy
> for block replication policy
> 17/12/06 12:37:50 INFO SparkUI: Stopped Spark web UI at
> http://192.168.10.207:4040
> 17/12/06 12:37:50 INFO BlockManagerMaster: Registering BlockManager
> BlockManagerId(driver, 192.168.10.207, 42644, None)
> 17/12/06 12:37:50 INFO StandaloneSchedulerBackend: Shutting down all
> executors
> 17/12/06 12:37:50 INFO BlockManagerMasterEndpoint: Registering block
> manager 192.168.10.207:42644 with 2004.6 MB RAM, BlockManagerId(driver,
> 192.168.10.207, 42644, None)
> 17/12/06 12:37:50 INFO CoarseGrainedSchedulerBackend$DriverEndpoint:
> Asking each executor to shut down
> 17/12/06 12:37:50 INFO BlockManagerMaster: Registered BlockManager
> BlockManagerId(driver, 192.168.10.207, 42644, None)
> 17/12/06 12:37:50 INFO BlockManager: Initialized BlockManager:
> BlockManagerId(driver, 192.168.10.207, 42644, None)
> 17/12/06 12:37:50 WARN StandaloneAppClient$ClientEndpoint: Drop
> UnregisterApplication(null) because has not yet connected to master
> 17/12/06 12:37:50 INFO MapOutputTrackerMasterEndpoint:
> MapOutputTrackerMasterEndpoint stopped!
> 17/12/06 12:37:50 INFO MemoryStore: MemoryStore cleared
> 17/12/06 12:37:50 INFO BlockManager: BlockManager stopped
> 17/12/06 12:37:50 INFO BlockManagerMaster: BlockManagerMaster stopped
> 17/12/06 12:37:50 INFO OutputCommitCoordinator$
> OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
> 17/12/06 12:37:50 ERROR SparkContext: Error initializing SparkContext.
> java.lang.NullPointerException
>         at org.apache.spark.SparkContext.<init>(SparkContext.scala:567)
>         at org.apache.spark.SparkContext$.getOrCreate(SparkContext.
> scala:2509)
>         at org.apache.spark.sql.SparkSession$Builder$$anonfun$
> 6.apply(SparkSession.scala:909)
>         at org.apache.spark.sql.SparkSession$Builder$$anonfun$
> 6.apply(SparkSession.scala:901)
>         at scala.Option.getOrElse(Option.scala:121)
>         at org.apache.spark.sql.SparkSession$Builder.
> getOrCreate(SparkSession.scala:901)
>         at com.my_user.MainClass.setCluster(MainClass.java:150)
>         at com.my_user.MainClass.initSpark(MainClass.java:69)
>         at com.my_user.MainClassWatch.main(MainClassWatch.java:25)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:62)
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:498)
>         at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> deploy$SparkSubmit$$runMain(SparkSubmit.scala:755)
>         at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> SparkSubmit.scala:180)
> Finished! Exit code:1
>         at org.apache.spark.deploy.SparkSubmit$.submit(
> SparkSubmit.scala:205)
>         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
> scala:119)
>         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> 17/12/06 12:37:50 INFO SparkContext: SparkContext already stopped.
> 17/12/06 12:37:50 INFO SparkContext: Successfully stopped SparkContext
> Exception in thread "main" java.lang.NullPointerException
>         at org.apache.spark.SparkContext.<init>(SparkContext.scala:567)
>         at org.apache.spark.SparkContext$.getOrCreate(SparkContext.
> scala:2509)
>         at org.apache.spark.sql.SparkSession$Builder$$anonfun$
> 6.apply(SparkSession.scala:909)
>         at org.apache.spark.sql.SparkSession$Builder$$anonfun$
> 6.apply(SparkSession.scala:901)
>         at scala.Option.getOrElse(Option.scala:121)
>         at org.apache.spark.sql.SparkSession$Builder.
> getOrCreate(SparkSession.scala:901)
>         at com.my_user.MainClass.setCluster(MainClass.java:150)
>         at com.my_user.MainClass.initSpark(MainClass.java:69)
>         at com.my_user.MainClassWatch.main(MainClassWatch.java:25)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:62)
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:498)
>         at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> deploy$SparkSubmit$$runMain(SparkSubmit.scala:755)
>         at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> SparkSubmit.scala:180)
>         at org.apache.spark.deploy.SparkSubmit$.submit(
> SparkSubmit.scala:205)
>         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
> scala:119)
>         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> 17/12/06 12:37:50 INFO ShutdownHookManager: Shutdown hook called
> 17/12/06 12:37:50 INFO ShutdownHookManager: Deleting directory
> /tmp/spark-f5329cbc-7de3-40ab-a4a2-9942e7b33815
>
>
>
> Reverting back to 2.1.1 remove the issue
>
> --
>
> I.R
>
> ------------------------------
>
> The information contained in this e-mail is confidential and/or
> proprietary to Capital One and/or its affiliates and may only be used
> solely in performance of work or services for Capital One. The information
> transmitted herewith is intended only for use by the individual or entity
> to which it is addressed. If the reader of this message is not the intended
> recipient, you are hereby notified that any review, retransmission,
> dissemination, distribution, copying or other use of, or taking of any
> action in reliance upon this information is strictly prohibited. If you
> have received this communication in error, please contact the sender and
> delete the material from your computer.
>



-- 
I.R

Reply via email to