Have you tried adding:

    .mode(SaveMode.Overwrite)

On Tue, May 17, 2016 at 8:55 PM, Todd <bit1...@163.com> wrote:

> scala> records.groupBy("name").count().write.trigger(ProcessingTime("30
> seconds")).option("checkpointLocation",
> "file:///home/hadoop/jsoncheckpoint").startStream("file:///home/hadoop/jsonresult")
> org.apache.spark.sql.AnalysisException: Aggregations are not supported on
> streaming DataFrames/Datasets in Append output mode. Consider changing
> output mode to Update.;
>   at
> org.apache.spark.sql.catalyst.analysis.UnsupportedOperationChecker$.org$apache$spark$sql$catalyst$analysis$UnsupportedOperationChecker$$throwError(UnsupportedOperationChecker.scala:142)
>   at
> org.apache.spark.sql.catalyst.analysis.UnsupportedOperationChecker$$anonfun$checkForStreaming$1.apply(UnsupportedOperationChecker.scala:59)
>   at
> org.apache.spark.sql.catalyst.analysis.UnsupportedOperationChecker$$anonfun$checkForStreaming$1.apply(UnsupportedOperationChecker.scala:46)
>   at
> org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:125)
>   at
> org.apache.spark.sql.catalyst.analysis.UnsupportedOperationChecker$.checkForStreaming(UnsupportedOperationChecker.scala:46)
>   at
> org.apache.spark.sql.ContinuousQueryManager.startQuery(ContinuousQueryManager.scala:190)
>   at
> org.apache.spark.sql.DataFrameWriter.startStream(DataFrameWriter.scala:351)
>   at
> org.apache.spark.sql.DataFrameWriter.startStream(DataFrameWriter.scala:279)
>
>
> I brief the spark code, looks like there is no way to change output mode
> to Update?
>

Reply via email to