Re: [pyspark delta] [delta][Spark SQL]: Getting an Analysis Exception. The associated location (path) is not empty

2022-08-01 Thread Kumba Janga
Thanks Sean! That was a simple fix. I changed it to "Create or Replace Table" but now I am getting the following error. I am still researching solutions but so far no luck. ParseException: mismatched input '' expecting {'ADD', 'AFTER', 'ALL', 'ALTER', 'ANALYZE', 'AND', 'ANTI', 'ANY', 'ARCHIVE',

Re: [pyspark delta] [delta][Spark SQL]: Getting an Analysis Exception. The associated location (path) is not empty

2022-08-01 Thread Sean Owen
Pretty much what it says? you are creating a table over a path that already has data in it. You can't do that without mode=overwrite at least, if that's what you intend. On Mon, Aug 1, 2022 at 7:29 PM Kumba Janga wrote: > > >- Component: Spark Delta, Spark SQL >- Level: Beginner >-

[pyspark delta] [delta][Spark SQL]: Getting an Analysis Exception. The associated location (path) is not empty

2022-08-01 Thread Kumba Janga
- Component: Spark Delta, Spark SQL - Level: Beginner - Scenario: Debug, How-to *Python in Jupyter:* import pyspark import pyspark.sql.functions from pyspark.sql import SparkSession spark = ( SparkSession .builder .appName("programming") .master("local")

Re: WARN: netlib.BLAS

2022-08-01 Thread Sean Owen
Hm, I think the problem is either that you need to build the spark-ganglia-lgpl module in your Spark distro, or the pomOnly() part of your build. You need the code in your app. Yes you need openblas too. On Mon, Aug 1, 2022 at 7:36 AM 陈刚 wrote: > Dear export, > > > I'm using spark-3.1.1 mllib,

WARN: netlib.BLAS

2022-08-01 Thread 陈刚
Dear export, I'm using spark-3.1.1 mllib, and I got this on CentOS 7.6: 22/08/01 09:42:34 WARN netlib.BLAS: Failed to load implementation from: com.github.fommil.netlib.NativeSystemBLAS 22/08/01 09:42:34 WARN netlib.BLAS: Failed to load implementation from:

Re: Use case idea

2022-08-01 Thread pengyh
* streaming handler is still useful for spark, though there is flink as alternative * RDD is also useful for transform especially for non-structure data * there are many SQL products in market like Drill/Impala, but spark is more powerful for distributed deployment as far as I know * we never

Re: Use case idea

2022-08-01 Thread Gourav Sengupta
Hi, my comments were for purposes of SQL, also most of other technologies like snowflake, and Redshift, and using KSQL directly to other sinks quite easily, without massive engineering, infact databricks is trying to play a catchup game in this market by coming out with GIU based ETL tools :)

Re: unsubscribe

2022-08-01 Thread pengyh
you could be able to unsubscribe yourself by using the signature below. To unsubscribe e-mail: user-unsubscr...@spark.apache.org - To unsubscribe e-mail: user-unsubscr...@spark.apache.org

unsubscribe

2022-08-01 Thread Martin Soch
- To unsubscribe e-mail: user-unsubscr...@spark.apache.org