Re: Driver vs master

2019-10-07 Thread Andrew Melo
On Mon, Oct 7, 2019 at 20:49 ayan guha wrote: > HI > > I think you are mixing terminologies here. Loosely speaking, Master > manages worker machines. Each worker machine can run one or more processes. > A process can be a driver or executor. You submit applications to the > master. Each

Re: Driver vs master

2019-10-07 Thread ayan guha
HI I think you are mixing terminologies here. Loosely speaking, Master manages worker machines. Each worker machine can run one or more processes. A process can be a driver or executor. You submit applications to the master. Each application will have driver and executors. Master will decide

Re: Driver vs master

2019-10-07 Thread Andrew Melo
Hi On Mon, Oct 7, 2019 at 19:20 Amit Sharma wrote: > Thanks Andrew but I am asking specific to driver memory not about > executors memory. We have just one master and if each jobs driver.memory=4g > and master nodes total memory is 16gb then we can not execute more than 4 > jobs at a time. I

Re: Driver vs master

2019-10-07 Thread Amit Sharma
Thanks Andrew but I am asking specific to driver memory not about executors memory. We have just one master and if each jobs driver.memory=4g and master nodes total memory is 16gb then we can not execute more than 4 jobs at a time. On Monday, October 7, 2019, Andrew Melo wrote: > Hi Amit > > On

Re: Driver vs master

2019-10-07 Thread Andrew Melo
Hi Amit On Mon, Oct 7, 2019 at 18:33 Amit Sharma wrote: > Can you please help me understand this. I believe driver programs runs on > master node If we are running 4 spark job and driver memory config is 4g then total 16 > 6b would be used of master node. This depends on what master/deploy

Driver vs master

2019-10-07 Thread Amit Sharma
Can you please help me understand this. I believe driver programs runs on master node. If we are running 4 spark job and driver memory config is 4g then total 16 6b would be used of master node. So if we will run more jobs then we need more memory on master. Please correct me if I am wrong.

Re: pandas_udf throws "Unsupported class file major version 56"

2019-10-07 Thread Lian Jiang
I figured out. Thanks. On Mon, Oct 7, 2019 at 9:55 AM Lian Jiang wrote: > Hi, > > from pyspark.sql.functions import pandas_udf, PandasUDFType > import pyspark > from pyspark.sql import SparkSession > spark = SparkSession.builder.getOrCreate() > > df = spark.createDataFrame( > [(1, True,

pandas_udf throws "Unsupported class file major version 56"

2019-10-07 Thread Lian Jiang
Hi, from pyspark.sql.functions import pandas_udf, PandasUDFType import pyspark from pyspark.sql import SparkSession spark = SparkSession.builder.getOrCreate() df = spark.createDataFrame( [(1, True, 1.0, 'aa'), (1, False, 2.0, 'aa'), (2, True, 3.0, 'aa'), (2, True, 5.0, 'aa'), (2, True, 10.0,

RE: [External Sender] Spark Executor pod not getting created on kubernetes cluster

2019-10-07 Thread Rao, Abhishek (Nokia - IN/Bangalore)
Hi Manish, Is this issue resolved? If not, please check the overlay network of your cluster once. We had faced similar issues when we had problems with overlay networking. In our case, executor had spawned, but the communication with driver and executor had failed (due to issues with overlay