You are reusing HiveDF for two vars and it ends up ambiguous. Just rename one.
On Thu, Oct 1, 2020, 5:02 PM Mich Talebzadeh <mich.talebza...@gmail.com> wrote: > Hi, > > > Spark version 2.3.3 on Google Dataproc > > > I am trying to use databricks to other databases > > > https://spark.apache.org/docs/latest/sql-data-sources-jdbc.html > > > to read from Hive table on Prem using Spark in Cloud > > > This works OK without a Try enclosure. > > > import spark.implicits._ > > import scala.util.{Try, Success, Failure} > > val HiveDF = Try(spark.read. > > format("jdbc"). > > option("url", jdbcUrl). > > option("dbtable", HiveSchema+"."+HiveTable). > > option("user", HybridServerUserName). > > option("password", HybridServerPassword). > > load()) match { > > case Success(HiveDF) => HiveDF > > case Failure(e) => throw new Exception("Error > Encountered reading Hive table") > > } > > However, with Try I am getting the following error > > > <console>:66: error: recursive value HiveDF needs type > > case Success(HiveDF) => HiveDF > > Wondering what is causing this. I have used it before (say reading from an > XML file) and it worked the, > > Thanks > > > > > > *Disclaimer:* Use it at your own risk. Any and all responsibility for any > loss, damage or destruction of data or any other property which may arise > from relying on this email's technical content is explicitly disclaimed. > The author will in no case be liable for any monetary damages arising from > such loss, damage or destruction. > > >