[ https://issues.apache.org/jira/browse/SPARK-11173?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Sean Owen resolved SPARK-11173. ------------------------------- Resolution: Not A Problem This does not appear to be an issue from Spark. > Cannot save data via MSSQL JDBC > ------------------------------- > > Key: SPARK-11173 > URL: https://issues.apache.org/jira/browse/SPARK-11173 > Project: Spark > Issue Type: Bug > Components: Java API, PySpark, SQL > Affects Versions: 1.5.1 > Environment: Windows 7 sp1 x64, java version "1.8.0_60", Spark 1.5.1, > hadoop 2.6, microsoft jdbc 4.2, pyspark > Reporter: Gianluca Salvo > Labels: patch > > Hello, > I'm experiencing an issue in writing dataframe via JBDC. My code is > {code:title=Example.python|borderStyle=solid} > from pyspark import SparkContext > from pyspark.sql import SQLContext > import sys > sc=SparkContext(appName="SQL Query") > sqlctx=SQLContext(sc) > serverName="SQLIPAddress" > serverPort="SQL Port" > serverUsername="username" > serverPassword="password" > serverDatabase="database" > ######################################################################### > connString="jdbc:sqlserver://{SERVER}:{PORT};user={USER};password={PASSWORD};databasename={DATABASENAME}" > connString=connString.format(SERVER=serverName,PORT=serverPort,USER=serverUsername,PASSWORD=serverPassword,DATABASENAME=serverDatabase) > df=sqlctx.read.format("jdbc").options(url=connString,dbtable="(select * from > TestTable) as test_Table").load() > df.show() > try: > df.write.jdbc(connString,"Test_Target","append") > print("saving completed") > except: > print("Error in saving data",sys.exc_info()[0]) > sc.stop() > {code} > Even if i specify *append*, the code throws an exception saying it is trying > to create the table *Test_Target* but the table is already present. > If I target the script to MariaDB, all is fine > {code:title=New Connection string|borderStyle=solid} > connString="jdbc:mysql://{SERVER}:{PORT}/{DATABASENAME}?user={USER}&password={PASSWORD}"; > {code} > The problem seems to be the Microsoft JDBC driver. Can you suggest or > implement same workaround? > Best regards -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org