[
https://issues.apache.org/jira/browse/BAHIR-123?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16181612#comment-16181612
]
ASF GitHub Bot commented on BAHIR-123:
--------------------------------------
Github user ckadner commented on the issue:
https://github.com/apache/bahir/pull/50
@emlaver
> `CloudantChangesDFSuite`:
> `- save dataframe to database using createDBOnSave=true option FAILED`
**Test failures:**
```
14:10:05 [INFO] --- scalatest-maven-plugin:1.0:test (test) @
spark-sql-cloudant_2.11 ---
14:10:05 Discovery starting.
14:10:05
14:10:05 Sql-cloudant tests that require Cloudant databases have been
enabled by
14:10:05 the environment variables CLOUDANT_USER and CLOUDANT_PASSWORD.
14:10:05
14:10:05 Discovery completed in 187 milliseconds.
14:10:05 Run starting. Expected test count is: 22
14:10:05 CloudantOptionSuite:
14:10:09 - invalid api receiver option throws an error message
14:10:09 - empty username option throws an error message
14:10:09 - empty password option throws an error message
14:10:10 - empty databaseName throws an error message
14:10:10 ClientSparkFunSuite:
14:10:10 CloudantChangesDFSuite:
14:10:34 - load and save data from Cloudant database
14:10:36 - load and count data from Cloudant search index
14:10:52 - load data and verify deleted doc is not in results
14:11:12 - load data and count rows in filtered dataframe
14:11:52 - save filtered dataframe to database
14:12:12 - save dataframe to database using createDBOnSave=true option ***
FAILED ***
14:12:12 org.apache.bahir.cloudant.common.CloudantException: Database
airportcodemapping_df create error: {"error":"file_exists","reason":"The
database could not be created, the file already exists."}
14:12:12 at
org.apache.bahir.cloudant.common.JsonStoreDataAccess.createDB(JsonStoreDataAccess.scala:143)
14:12:12 at
org.apache.bahir.cloudant.CloudantReadWriteRelation.insert(DefaultSource.scala:72)
14:12:12 at
org.apache.bahir.cloudant.DefaultSource.createRelation(DefaultSource.scala:172)
14:12:12 at
org.apache.bahir.cloudant.DefaultSource.createRelation(DefaultSource.scala:86)
14:12:12 at
org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:472)
14:12:12 at
org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:48)
14:12:12 at
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
14:12:12 at
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
14:12:12 at
org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
14:12:12 at
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:117)
14:12:12 ...
14:12:13 - load and count data from view
14:12:13 - load data from view with MapReduce function
14:12:53 - load data and verify total count of selector, filter, and view
option
14:12:53 CloudantSparkSQLSuite:
14:12:56 - verify results from temp view of database n_airportcodemapping
14:12:59 - verify results from temp view of index in n_flight
14:13:00 CloudantAllDocsDFSuite:
14:13:03 - load and save data from Cloudant database
14:13:04 - load and count data from Cloudant search index
14:13:04 - load data and count rows in filtered dataframe
14:13:06 - save filtered dataframe to database
14:13:07 - save dataframe to database using createDBOnSave=true option ***
FAILED ***
14:13:07 org.apache.bahir.cloudant.common.CloudantException: Database
airportcodemapping_df create error: {"error":"file_exists","reason":"The
database could not be created, the file already exists."}
14:13:07 at
org.apache.bahir.cloudant.common.JsonStoreDataAccess.createDB(JsonStoreDataAccess.scala:143)
14:13:07 at
org.apache.bahir.cloudant.CloudantReadWriteRelation.insert(DefaultSource.scala:72)
14:13:07 at
org.apache.bahir.cloudant.DefaultSource.createRelation(DefaultSource.scala:172)
14:13:07 at
org.apache.bahir.cloudant.DefaultSource.createRelation(DefaultSource.scala:86)
14:13:07 at
org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:472)
14:13:07 at
org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:48)
14:13:07 at
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
14:13:07 at
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
14:13:07 at
org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
14:13:07 at
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:117)
14:13:07 ...
14:13:07 - load and count data from view
14:13:07 - load data from view with MapReduce function
14:13:08 Run completed in 3 minutes, 2 seconds.
14:13:08 Total number of tests run: 22
14:13:08 Suites: completed 6, aborted 0
14:13:08 Tests: succeeded 20, failed 2, canceled 0, ignored 0, pending 0
14:13:08 *** 2 TESTS FAILED ***
14:13:08 [INFO]
------------------------------------------------------------------------
...
14:13:08 [INFO] Apache Bahir - Spark SQL Cloudant DataSource .......
FAILURE [03:24 min]
...
```
> Fix errors to support the latest version of Play JSON library for sql-cloudant
> ------------------------------------------------------------------------------
>
> Key: BAHIR-123
> URL: https://issues.apache.org/jira/browse/BAHIR-123
> Project: Bahir
> Issue Type: Task
> Reporter: Esteban Laver
> Assignee: Esteban Laver
> Priority: Minor
>
> The latest version is 2.6.2. Error during mvn install -pl sql-cloudant after
> updating play-json to 2.6.2 in sql-cloudant/pom.xml:
> [ERROR]
> /Users/estebanmlaver/emlaver-bahir/sql-cloudant/src/main/scala/org/apache/bahir/cloudant/common/JsonStoreConfigManager.scala:19:
> object typesafe is not a member of package com
> [ERROR] import com.typesafe.config.ConfigFactory
> [ERROR] ^
> [ERROR]
> /Users/estebanmlaver/emlaver-bahir/sql-cloudant/src/main/scala/org/apache/bahir/cloudant/common/JsonStoreConfigManager.scala:52:
> not found: value ConfigFactory
> [ERROR] private val configFactory = ConfigFactory.load()
> [ERROR] ^
> [ERROR] two errors found
> Maven compile dependencies need to be added to pom.xml that existed in
> play-json 2.5.9 but were removed in 2.6.2.
> Additional info. from Patrick Titzler between play-json versions 2.5.x and
> 2.6.x:
> Looks like the parameter data type has been changed from `Seq[JsValue]`
> (https://www.playframework.com/documentation/2.5.x/api/scala/index.html#play.api.libs.json.JsArray)
> to `IndexedSeq[JsValue]`
> https://playframework.com/documentation/2.6.x/api/scala/index.html#play.api.libs.json.JsArray
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)