Re: About Spark executs sqlscript

2021-05-24 Thread Mich Talebzadeh
Apologies I missed your two points My question: #1 If there are 10 tables or more tables, do I need to read each table into memory though Spark bases on memory compution? Every table will be read as I described above. It is lazy read by Spark. The computation happens when there is an action on

Re: About Spark executs sqlscript

2021-05-24 Thread Mich Talebzadeh
Well, Spark to BigQuery API is very efficient in doing what it needs to do. Personally I have never found a JDBC connection to BigQuery that works under all circumstances . In a typical environment you need to set-up your connection variable to BigQuery from Spark. These are my recommended ones

Re: About Spark executs sqlscript

2021-05-24 Thread Wenchen Fan
It's not possible to load everything into memory. We should use a big query connector (should be existing already?) and register table B and C and temp views in Spark. On Fri, May 14, 2021 at 8:50 AM bo zhao wrote: > Hi Team, > > I've followed Spark community for several years. This is my first

About Spark executs sqlscript

2021-05-13 Thread bo zhao
Hi Team, I've followed Spark community for several years. This is my first time for asking help. I hope you guys can give some experience. I want to develop a spark application with processing a sqlscript file. The data is on BigQuery. For example, the sqlscript is: delete from tableA; insert