[ 
https://issues.apache.org/jira/browse/BEAM-11224?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17548945#comment-17548945
 ] 

Danny McCormick commented on BEAM-11224:
----------------------------------------

This issue has been migrated to https://github.com/apache/beam/issues/20555

> Spark runner side inputs causing SIGNAL TERM
> --------------------------------------------
>
>                 Key: BEAM-11224
>                 URL: https://issues.apache.org/jira/browse/BEAM-11224
>             Project: Beam
>          Issue Type: Bug
>          Components: runner-spark
>    Affects Versions: 2.25.0
>         Environment: Beam: 2.25
> SparkRunner: 2.25
> Java version: 11.0.9-ea
> Maven Compiler Source: 1.8
> Maven Compiler Target: 1.8
> Spark version: 2.4.7
>            Reporter: Thomas Li Fredriksen
>            Priority: P3
>              Labels: sideinput, spark, spark-runner
>          Time Spent: 10m
>  Remaining Estimate: 0h
>
> I wish to use side-inputs in order to pass some configuration to my pipeline, 
> however the driver commands a shutdown after the `PCollectionView` has been 
> created when running on my local spark-cluster (spark version 2.4.7, 1 
> master, 1 worker, running on localhost). This however works perfectly on the 
> DirectRunner.
> I have attempted to strip the code to its bare essentials (see below). Still 
> the issue persists when running on the spark cluster. DirectRunner still 
> works fine.
> The spark-cluster does accept jobs, and I have sucessfully run a 
> "hello-world" pipeline that completed without issue.
> What is happening here?
> Logs pasted below.
> {code}
>     // Pipeline
>     private static PipelineResult runPipeline(PipelineOptions options) {
>         Pipeline p = Pipeline.create(options);
>         PCollectionView<String> schema = p
>                 .apply("Dummy tabular schema builder", Create.of("This is a 
> string"))
>                 .apply("Collect", View.asSingleton());
>         p
>                 .apply("Hello world", Create.of("Hello world"))
>                 .apply("Side input test", 
> ParDo.of(DummyFn.builder().setSchemaView(schema).build()).withSideInput("schema",
>  schema))
>                 .apply(ConsoleIO.create());
>         return p.run();
>     }
> {code}
> {code}
> // Simple FN that prints the side input
> @AutoValue
> public abstract class DummyFn extends DoFn<String, String> {
>     private final static Logger LOG = LoggerFactory.getLogger(DummyFn.class);
>     public static Builder builder() {
>         return new org.odp.beam.io.fn.AutoValue_DummyFn.Builder();
>     }
>     public abstract PCollectionView<String> getSchemaView();
>     @ProcessElement
>     public void processElement(@Element String element,
>                                OutputReceiver<String> out,
>                                ProcessContext context) {
>         String schema = context.sideInput(getSchemaView());
>         LOG.warn(schema.toString());
>         out.output(element.toUpperCase());
>     }
>     @AutoValue.Builder
>     public abstract static class Builder {
>         public abstract Builder setSchemaView(PCollectionView<String> value);
>         public abstract DummyFn build();
>     }
> }
> {code}
> {code}
> // Simple PTransform that prints the output of the toString-method
> public class ConsoleIO<T> extends PTransform<PCollection<T>, PDone> {
>     public static <T> ConsoleIO<T> create() {
>         return new ConsoleIO();
>     }
>     @Override
>     public PDone expand(PCollection<T> input) {
>         input
>                 .apply("Print elements", ParDo.of(new PrintElementFn<T>()));
>         return PDone.in(input.getPipeline());
>     }
>     static class PrintElementFn<T> extends DoFn<T, Void> {
>         @DoFn.ProcessElement
>         public void processElement(@Element T element, ProcessContext 
> context) throws Exception {
>             System.out.println(element.toString());
>         }
>     }
> }
> {code}
> spark-submit output
> {code}
> $ spark-submit \
> --class org.odp.beam.extractors.CsvToCdfRawExtractor \
> --verbose \
> --driver-memory 4G \
> --executor-memory 4G \
> --total-executor-cores 4 \
> --deploy-mode client \
> --supervise \
> --conf spark.dynamicAllocation.enabled=false \
> --conf spark.network.timeout=420000 \
> --master spark://192.168.10.172:7077 \
> target/beam-poc-0.1-shaded.jar \
> --runner=SparkRunner
> Using properties file: null
> 20/11/10 15:46:44 WARN Utils: Your hostname, localhost.localdomain resolves 
> to a loopback address: 127.0.0.1; using 192.168.10.172 instead (on interface 
> enp7s0)
> 20/11/10 15:46:44 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to 
> another address
> WARNING: An illegal reflective access operation has occurred
> WARNING: Illegal reflective access by org.apache.spark.unsafe.Platform 
> (file:/home/tom/app/spark/spark-2.4.7-bin-hadoop2.7/jars/spark-unsafe_2.11-2.4.7.jar)
>  to method java.nio.Bits.unaligned()
> WARNING: Please consider reporting this to the maintainers of 
> org.apache.spark.unsafe.Platform
> WARNING: Use --illegal-access=warn to enable warnings of further illegal 
> reflective access operations
> WARNING: All illegal access operations will be denied in a future release
> Parsed arguments:
>   master                  spark://192.168.10.172:7077
>   deployMode              client
>   executorMemory          4G
>   executorCores           null
>   totalExecutorCores      4
>   propertiesFile          null
>   driverMemory            4G
>   driverCores             null
>   driverExtraClassPath    null
>   driverExtraLibraryPath  null
>   driverExtraJavaOptions  null
>   supervise               true
>   queue                   null
>   numExecutors            null
>   files                   null
>   pyFiles                 null
>   archives                null
>   mainClass               org.odp.beam.extractors.CsvToCdfRawExtractor
>   primaryResource         
> file:/home/tom/project/odf/beam-poc/target/beam-poc-0.1-shaded.jar
>   name                    org.odp.beam.extractors.CsvToCdfRawExtractor
>   childArgs               [--runner=SparkRunner]
>   jars                    null
>   packages                null
>   packagesExclusions      null
>   repositories            null
>   verbose                 true
> Spark properties used, including those specified through
>  --conf and those from the properties file null:
>   (spark.network.timeout,420000)
>   (spark.driver.memory,4G)
>   (spark.dynamicAllocation.enabled,false)
>     
> 20/11/10 15:46:45 WARN NativeCodeLoader: Unable to load native-hadoop library 
> for your platform... using builtin-java classes where applicable
> Main class:
> org.odp.beam.extractors.CsvToCdfRawExtractor
> Arguments:
> --runner=SparkRunner
> Spark config:
> (spark.jars,file:/home/tom/project/odf/beam-poc/target/beam-poc-0.1-shaded.jar)
> (spark.app.name,org.odp.beam.extractors.CsvToCdfRawExtractor)
> (spark.cores.max,4)
> (spark.network.timeout,420000)
> (spark.driver.memory,4G)
> (spark.submit.deployMode,client)
> (spark.master,spark://192.168.10.172:7077)
> (spark.executor.memory,4G)
> (spark.dynamicAllocation.enabled,false)
> Classpath elements:
> file:/home/tom/project/odf/beam-poc/target/beam-poc-0.1-shaded.jar
> log4j:WARN No appenders could be found for logger 
> (org.apache.beam.sdk.options.PipelineOptionsFactory).
> log4j:WARN Please initialize the log4j system properly.
> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more 
> info.
> Using Spark's default log4j profile: 
> org/apache/spark/log4j-defaults.properties
> 20/11/10 15:46:46 INFO SparkContext: Running Spark version 2.4.7
> 20/11/10 15:46:47 INFO SparkContext: Submitted application: 
> CsvToCdfRawExtractor
> 20/11/10 15:46:47 INFO SecurityManager: Changing view acls to: tom
> 20/11/10 15:46:47 INFO SecurityManager: Changing modify acls to: tom
> 20/11/10 15:46:47 INFO SecurityManager: Changing view acls groups to: 
> 20/11/10 15:46:47 INFO SecurityManager: Changing modify acls groups to: 
> 20/11/10 15:46:47 INFO SecurityManager: SecurityManager: authentication 
> disabled; ui acls disabled; users  with view permissions: Set(tom); groups 
> with view permissions: Set(); users  with modify permissions: Set(tom); 
> groups with modify permissions: Set()
> 20/11/10 15:46:47 INFO Utils: Successfully started service 'sparkDriver' on 
> port 35103.
> 20/11/10 15:46:47 INFO SparkEnv: Registering MapOutputTracker
> 20/11/10 15:46:47 INFO SparkEnv: Registering BlockManagerMaster
> 20/11/10 15:46:47 INFO BlockManagerMasterEndpoint: Using 
> org.apache.spark.storage.DefaultTopologyMapper for getting topology 
> information
> 20/11/10 15:46:47 INFO BlockManagerMasterEndpoint: BlockManagerMasterEndpoint 
> up
> 20/11/10 15:46:47 INFO DiskBlockManager: Created local directory at 
> /tmp/blockmgr-58419068-d0ad-45c9-b90b-92b659dee1c3
> 20/11/10 15:46:47 INFO MemoryStore: MemoryStore started with capacity 2.2 GB
> 20/11/10 15:46:47 INFO SparkEnv: Registering OutputCommitCoordinator
> 20/11/10 15:46:47 INFO Utils: Successfully started service 'SparkUI' on port 
> 4040.
> 20/11/10 15:46:47 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started at 
> http://fedora:4040
> 20/11/10 15:46:47 INFO SparkContext: Added JAR 
> file:/home/tom/project/odf/beam-poc/target/beam-poc-0.1-shaded.jar at 
> spark://fedora:35103/jars/beam-poc-0.1-shaded.jar with timestamp 1605019607514
> 20/11/10 15:46:47 INFO StandaloneAppClient$ClientEndpoint: Connecting to 
> master spark://192.168.10.172:7077...
> 20/11/10 15:46:47 INFO TransportClientFactory: Successfully created 
> connection to /192.168.10.172:7077 after 25 ms (0 ms spent in bootstraps)
> 20/11/10 15:46:47 INFO StandaloneSchedulerBackend: Connected to Spark cluster 
> with app ID app-20201110154647-0020
> 20/11/10 15:46:47 INFO StandaloneAppClient$ClientEndpoint: Executor added: 
> app-20201110154647-0020/0 on worker-20201109144752-192.168.10.172-45535 
> (192.168.10.172:45535) with 4 core(s)
> 20/11/10 15:46:47 INFO StandaloneSchedulerBackend: Granted executor ID 
> app-20201110154647-0020/0 on hostPort 192.168.10.172:45535 with 4 core(s), 
> 4.0 GB RAM
> 20/11/10 15:46:47 INFO Utils: Successfully started service 
> 'org.apache.spark.network.netty.NettyBlockTransferService' on port 33169.
> 20/11/10 15:46:47 INFO NettyBlockTransferService: Server created on 
> fedora:33169
> 20/11/10 15:46:47 INFO BlockManager: Using 
> org.apache.spark.storage.RandomBlockReplicationPolicy for block replication 
> policy
> 20/11/10 15:46:47 INFO StandaloneAppClient$ClientEndpoint: Executor updated: 
> app-20201110154647-0020/0 is now RUNNING
> 20/11/10 15:46:47 INFO BlockManagerMaster: Registering BlockManager 
> BlockManagerId(driver, fedora, 33169, None)
> 20/11/10 15:46:47 INFO BlockManagerMasterEndpoint: Registering block manager 
> fedora:33169 with 2.2 GB RAM, BlockManagerId(driver, fedora, 33169, None)
> 20/11/10 15:46:47 INFO BlockManagerMaster: Registered BlockManager 
> BlockManagerId(driver, fedora, 33169, None)
> 20/11/10 15:46:47 INFO BlockManager: Initialized BlockManager: 
> BlockManagerId(driver, fedora, 33169, None)
> 20/11/10 15:46:47 INFO StandaloneSchedulerBackend: SchedulerBackend is ready 
> for scheduling beginning after reached minRegisteredResourcesRatio: 0.0
> 20/11/10 15:46:48 INFO SparkRunner$Evaluator: Entering directly-translatable 
> composite transform: 
> 'Collect/Combine.GloballyAsSingletonView/Combine.globally(Singleton)'
> 20/11/10 15:46:48 INFO MetricsAccumulator: Instantiated metrics accumulator: 
> MetricQueryResults()
> 20/11/10 15:46:48 INFO AggregatorsAccumulator: Instantiated aggregators 
> accumulator: 
> 20/11/10 15:46:48 INFO SparkRunner$Evaluator: Evaluating Read(CreateSource)
> 20/11/10 15:46:48 INFO SparkRunner$Evaluator: Entering directly-translatable 
> composite transform: 
> 'Collect/Combine.GloballyAsSingletonView/Combine.globally(Singleton)'
> 20/11/10 15:46:48 INFO SparkRunner$Evaluator: Evaluating 
> Combine.globally(Singleton)
> 20/11/10 15:46:48 INFO SparkContext: Starting job: aggregate at 
> GroupCombineFunctions.java:107
> 20/11/10 15:46:48 INFO DAGScheduler: Got job 0 (aggregate at 
> GroupCombineFunctions.java:107) with 1 output partitions
> 20/11/10 15:46:48 INFO DAGScheduler: Final stage: ResultStage 0 (aggregate at 
> GroupCombineFunctions.java:107)
> 20/11/10 15:46:48 INFO DAGScheduler: Parents of final stage: List()
> 20/11/10 15:46:48 INFO DAGScheduler: Missing parents: List()
> 20/11/10 15:46:48 INFO DAGScheduler: Submitting ResultStage 0 (Dummy tabular 
> schema builder/Read(CreateSource).out Bounded[0] at RDD at 
> SourceRDD.java:80), which has no missing parents
> 20/11/10 15:46:48 INFO MemoryStore: Block broadcast_0 stored as values in 
> memory (estimated size 16.2 KB, free 2.2 GB)
> 20/11/10 15:46:48 INFO MemoryStore: Block broadcast_0_piece0 stored as bytes 
> in memory (estimated size 6.8 KB, free 2.2 GB)
> 20/11/10 15:46:48 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory 
> on fedora:33169 (size: 6.8 KB, free: 2.2 GB)
> 20/11/10 15:46:48 INFO SparkContext: Created broadcast 0 from broadcast at 
> DAGScheduler.scala:1184
> 20/11/10 15:46:48 INFO DAGScheduler: Submitting 1 missing tasks from 
> ResultStage 0 (Dummy tabular schema builder/Read(CreateSource).out Bounded[0] 
> at RDD at SourceRDD.java:80) (first 15 tasks are for partitions Vector(0))
> 20/11/10 15:46:48 INFO TaskSchedulerImpl: Adding task set 0.0 with 1 tasks
> 20/11/10 15:46:49 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: 
> Registered executor NettyRpcEndpointRef(spark-client://Executor) 
> (192.168.10.172:48382) with ID 0
> 20/11/10 15:46:49 INFO TaskSetManager: Starting task 0.0 in stage 0.0 (TID 0, 
> 192.168.10.172, executor 0, partition 0, PROCESS_LOCAL, 8546 bytes)
> 20/11/10 15:46:49 INFO BlockManagerMasterEndpoint: Registering block manager 
> 192.168.10.172:43781 with 2.2 GB RAM, BlockManagerId(0, 192.168.10.172, 
> 43781, None)
> 20/11/10 15:46:50 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory 
> on 192.168.10.172:43781 (size: 6.8 KB, free: 2.2 GB)
> 20/11/10 15:46:51 INFO TaskSetManager: Finished task 0.0 in stage 0.0 (TID 0) 
> in 2056 ms on 192.168.10.172 (executor 0) (1/1)
> 20/11/10 15:46:51 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose tasks 
> have all completed, from pool 
> 20/11/10 15:46:51 INFO DAGScheduler: ResultStage 0 (aggregate at 
> GroupCombineFunctions.java:107) finished in 3.091 s
> 20/11/10 15:46:51 INFO DAGScheduler: Job 0 finished: aggregate at 
> GroupCombineFunctions.java:107, took 3.132405 s
> 20/11/10 15:46:51 INFO SparkRunner$Evaluator: Evaluating 
> org.apache.beam.sdk.transforms.View$VoidKeyToMultimapMaterialization$VoidKeyToMultimapMaterializationDoFn@14924f41
> 20/11/10 15:46:51 INFO SparkRunner$Evaluator: Evaluating 
> View.CreatePCollectionView
> 20/11/10 15:46:51 INFO SparkContext: Invoking stop() from shutdown hook
> 20/11/10 15:46:51 INFO SparkUI: Stopped Spark web UI at http://fedora:4040
> 20/11/10 15:46:51 INFO StandaloneSchedulerBackend: Shutting down all executors
> 20/11/10 15:46:51 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: Asking 
> each executor to shut down
> 20/11/10 15:46:51 INFO MapOutputTrackerMasterEndpoint: 
> MapOutputTrackerMasterEndpoint stopped!
> 20/11/10 15:46:51 INFO MemoryStore: MemoryStore cleared
> 20/11/10 15:46:51 INFO BlockManager: BlockManager stopped
> 20/11/10 15:46:51 INFO BlockManagerMaster: BlockManagerMaster stopped
> 20/11/10 15:46:51 INFO 
> OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: 
> OutputCommitCoordinator stopped!
> 20/11/10 15:46:51 INFO SparkContext: Successfully stopped SparkContext
> 20/11/10 15:46:51 INFO ShutdownHookManager: Shutdown hook called
> 20/11/10 15:46:51 INFO ShutdownHookManager: Deleting directory 
> /tmp/spark-665a903f-22db-497e-989f-a5ca3e0635e2
> 20/11/10 15:46:51 INFO ShutdownHookManager: Deleting directory 
> /tmp/spark-d4b5a04f-f6a3-48ff-b229-4eb966151d86
> {code}
> stderr from spark worker:
> {code}
> Spark Executor Command: 
> "/usr/lib/jvm/java-11-openjdk-11.0.9.6-0.0.ea.fc33.x86_64/bin/java" "-cp" 
> "/home/tom/app/spark/spark/conf/:/home/tom/app/spark/spark/jars/*" 
> "-Xmx4096M" "-Dspark.driver.port=35103" "-Dspark.network.timeout=420000" 
> "org.apache.spark.executor.CoarseGrainedExecutorBackend" "--driver-url" 
> "spark://CoarseGrainedScheduler@fedora:35103" "--executor-id" "0" 
> "--hostname" "192.168.10.172" "--cores" "4" "--app-id" 
> "app-20201110154647-0020" "--worker-url" "spark://Worker@192.168.10.172:45535"
> ========================================
> Using Spark's default log4j profile: 
> org/apache/spark/log4j-defaults.properties
> 20/11/10 15:46:48 INFO CoarseGrainedExecutorBackend: Started daemon with 
> process name: 360353@localhost.localdomain
> 20/11/10 15:46:48 INFO SignalUtils: Registered signal handler for TERM
> 20/11/10 15:46:48 INFO SignalUtils: Registered signal handler for HUP
> 20/11/10 15:46:48 INFO SignalUtils: Registered signal handler for INT
> 20/11/10 15:46:48 WARN Utils: Your hostname, localhost.localdomain resolves 
> to a loopback address: 127.0.0.1; using 192.168.10.172 instead (on interface 
> enp7s0)
> 20/11/10 15:46:48 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to 
> another address
> WARNING: An illegal reflective access operation has occurred
> WARNING: Illegal reflective access by org.apache.spark.unsafe.Platform 
> (file:/home/tom/app/spark/spark-2.4.7-bin-hadoop2.7/jars/spark-unsafe_2.11-2.4.7.jar)
>  to method java.nio.Bits.unaligned()
> WARNING: Please consider reporting this to the maintainers of 
> org.apache.spark.unsafe.Platform
> WARNING: Use --illegal-access=warn to enable warnings of further illegal 
> reflective access operations
> WARNING: All illegal access operations will be denied in a future release
> 20/11/10 15:46:48 WARN NativeCodeLoader: Unable to load native-hadoop library 
> for your platform... using builtin-java classes where applicable
> 20/11/10 15:46:48 INFO SecurityManager: Changing view acls to: tom
> 20/11/10 15:46:48 INFO SecurityManager: Changing modify acls to: tom
> 20/11/10 15:46:48 INFO SecurityManager: Changing view acls groups to: 
> 20/11/10 15:46:48 INFO SecurityManager: Changing modify acls groups to: 
> 20/11/10 15:46:48 INFO SecurityManager: SecurityManager: authentication 
> disabled; ui acls disabled; users  with view permissions: Set(tom); groups 
> with view permissions: Set(); users  with modify permissions: Set(tom); 
> groups with modify permissions: Set()
> 20/11/10 15:46:49 INFO TransportClientFactory: Successfully created 
> connection to fedora/192.168.10.172:35103 after 54 ms (0 ms spent in 
> bootstraps)
> 20/11/10 15:46:49 INFO SecurityManager: Changing view acls to: tom
> 20/11/10 15:46:49 INFO SecurityManager: Changing modify acls to: tom
> 20/11/10 15:46:49 INFO SecurityManager: Changing view acls groups to: 
> 20/11/10 15:46:49 INFO SecurityManager: Changing modify acls groups to: 
> 20/11/10 15:46:49 INFO SecurityManager: SecurityManager: authentication 
> disabled; ui acls disabled; users  with view permissions: Set(tom); groups 
> with view permissions: Set(); users  with modify permissions: Set(tom); 
> groups with modify permissions: Set()
> 20/11/10 15:46:49 INFO TransportClientFactory: Successfully created 
> connection to fedora/192.168.10.172:35103 after 4 ms (0 ms spent in 
> bootstraps)
> 20/11/10 15:46:49 INFO DiskBlockManager: Created local directory at 
> /tmp/spark-0e47fa97-8714-4e8e-950e-b1032fe36995/executor-e7667d04-198d-4144-8897-ddada0bfd1de/blockmgr-019262b3-4d3e-4158-b984-ff85c0846191
> 20/11/10 15:46:49 INFO MemoryStore: MemoryStore started with capacity 2.2 GB
> 20/11/10 15:46:49 INFO CoarseGrainedExecutorBackend: Connecting to driver: 
> spark://CoarseGrainedScheduler@fedora:35103
> 20/11/10 15:46:49 INFO WorkerWatcher: Connecting to worker 
> spark://Worker@192.168.10.172:45535
> 20/11/10 15:46:49 INFO TransportClientFactory: Successfully created 
> connection to /192.168.10.172:45535 after 2 ms (0 ms spent in bootstraps)
> 20/11/10 15:46:49 INFO WorkerWatcher: Successfully connected to 
> spark://Worker@192.168.10.172:45535
> 20/11/10 15:46:49 INFO CoarseGrainedExecutorBackend: Successfully registered 
> with driver
> 20/11/10 15:46:49 INFO Executor: Starting executor ID 0 on host 192.168.10.172
> 20/11/10 15:46:49 INFO Utils: Successfully started service 
> 'org.apache.spark.network.netty.NettyBlockTransferService' on port 43781.
> 20/11/10 15:46:49 INFO NettyBlockTransferService: Server created on 
> 192.168.10.172:43781
> 20/11/10 15:46:49 INFO BlockManager: Using 
> org.apache.spark.storage.RandomBlockReplicationPolicy for block replication 
> policy
> 20/11/10 15:46:49 INFO BlockManagerMaster: Registering BlockManager 
> BlockManagerId(0, 192.168.10.172, 43781, None)
> 20/11/10 15:46:49 INFO BlockManagerMaster: Registered BlockManager 
> BlockManagerId(0, 192.168.10.172, 43781, None)
> 20/11/10 15:46:49 INFO BlockManager: Initialized BlockManager: 
> BlockManagerId(0, 192.168.10.172, 43781, None)
> 20/11/10 15:46:49 INFO CoarseGrainedExecutorBackend: Got assigned task 0
> 20/11/10 15:46:49 INFO Executor: Running task 0.0 in stage 0.0 (TID 0)
> 20/11/10 15:46:49 INFO Executor: Fetching 
> spark://fedora:35103/jars/beam-poc-0.1-shaded.jar with timestamp 1605019607514
> 20/11/10 15:46:49 INFO TransportClientFactory: Successfully created 
> connection to fedora/192.168.10.172:35103 after 2 ms (0 ms spent in 
> bootstraps)
> 20/11/10 15:46:49 INFO Utils: Fetching 
> spark://fedora:35103/jars/beam-poc-0.1-shaded.jar to 
> /tmp/spark-0e47fa97-8714-4e8e-950e-b1032fe36995/executor-e7667d04-198d-4144-8897-ddada0bfd1de/spark-62556d02-a044-4c2c-8f97-c7f25ef3e337/fetchFileTemp6325880319900581024.tmp
> 20/11/10 15:46:49 INFO Utils: Copying 
> /tmp/spark-0e47fa97-8714-4e8e-950e-b1032fe36995/executor-e7667d04-198d-4144-8897-ddada0bfd1de/spark-62556d02-a044-4c2c-8f97-c7f25ef3e337/2058038551605019607514_cache
>  to 
> /home/tom/app/spark/spark-2.4.7-bin-hadoop2.7/work/app-20201110154647-0020/0/./beam-poc-0.1-shaded.jar
> 20/11/10 15:46:50 INFO Executor: Adding 
> file:/home/tom/app/spark/spark-2.4.7-bin-hadoop2.7/work/app-20201110154647-0020/0/./beam-poc-0.1-shaded.jar
>  to class loader
> 20/11/10 15:46:50 INFO TorrentBroadcast: Started reading broadcast variable 0
> 20/11/10 15:46:50 INFO TransportClientFactory: Successfully created 
> connection to fedora/192.168.10.172:33169 after 2 ms (0 ms spent in 
> bootstraps)
> 20/11/10 15:46:50 INFO MemoryStore: Block broadcast_0_piece0 stored as bytes 
> in memory (estimated size 6.8 KB, free 2.2 GB)
> 20/11/10 15:46:50 INFO TorrentBroadcast: Reading broadcast variable 0 took 
> 112 ms
> 20/11/10 15:46:50 INFO MemoryStore: Block broadcast_0 stored as values in 
> memory (estimated size 16.2 KB, free 2.2 GB)
> 20/11/10 15:46:51 INFO Executor: Finished task 0.0 in stage 0.0 (TID 0). 6312 
> bytes result sent to driver
> 20/11/10 15:46:51 INFO CoarseGrainedExecutorBackend: Driver commanded a 
> shutdown
> 20/11/10 15:46:51 ERROR CoarseGrainedExecutorBackend: RECEIVED SIGNAL TERM
> tdown
> {code}



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

Reply via email to