Repository: spark Updated Branches: refs/heads/branch-2.0 f17ffef38 -> 03008e049
[SPARK-16256][DOCS] Fix window operation diagram Author: Tathagata Das <tathagata.das1...@gmail.com> Closes #14001 from tdas/SPARK-16256-2. (cherry picked from commit 5d00a7bc19ddeb1b5247733b55095a03ee7b1a30) Signed-off-by: Tathagata Das <tathagata.das1...@gmail.com> Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/03008e04 Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/03008e04 Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/03008e04 Branch: refs/heads/branch-2.0 Commit: 03008e049a366bc7a63b3915b42ee50320ac6f34 Parents: f17ffef Author: Tathagata Das <tathagata.das1...@gmail.com> Authored: Thu Jun 30 14:01:34 2016 -0700 Committer: Tathagata Das <tathagata.das1...@gmail.com> Committed: Thu Jun 30 14:01:56 2016 -0700 ---------------------------------------------------------------------- docs/img/structured-streaming-late-data.png | Bin 138931 -> 138226 bytes docs/img/structured-streaming-window.png | Bin 128930 -> 132875 bytes docs/img/structured-streaming.pptx | Bin 1105315 -> 1105413 bytes docs/structured-streaming-programming-guide.md | 2 +- 4 files changed, 1 insertion(+), 1 deletion(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/spark/blob/03008e04/docs/img/structured-streaming-late-data.png ---------------------------------------------------------------------- diff --git a/docs/img/structured-streaming-late-data.png b/docs/img/structured-streaming-late-data.png index 5276b47..2283f67 100644 Binary files a/docs/img/structured-streaming-late-data.png and b/docs/img/structured-streaming-late-data.png differ http://git-wip-us.apache.org/repos/asf/spark/blob/03008e04/docs/img/structured-streaming-window.png ---------------------------------------------------------------------- diff --git a/docs/img/structured-streaming-window.png b/docs/img/structured-streaming-window.png index be9d3fb..c1842b1 100644 Binary files a/docs/img/structured-streaming-window.png and b/docs/img/structured-streaming-window.png differ http://git-wip-us.apache.org/repos/asf/spark/blob/03008e04/docs/img/structured-streaming.pptx ---------------------------------------------------------------------- diff --git a/docs/img/structured-streaming.pptx b/docs/img/structured-streaming.pptx index c278323..6aad2ed 100644 Binary files a/docs/img/structured-streaming.pptx and b/docs/img/structured-streaming.pptx differ http://git-wip-us.apache.org/repos/asf/spark/blob/03008e04/docs/structured-streaming-programming-guide.md ---------------------------------------------------------------------- diff --git a/docs/structured-streaming-programming-guide.md b/docs/structured-streaming-programming-guide.md index 5932566..7949396 100644 --- a/docs/structured-streaming-programming-guide.md +++ b/docs/structured-streaming-programming-guide.md @@ -620,7 +620,7 @@ df.groupBy("type").count() ### Window Operations on Event Time Aggregations over a sliding event-time window are straightforward with Structured Streaming. The key idea to understand about window-based aggregations are very similar to grouped aggregations. In a grouped aggregation, aggregate values (e.g. counts) are maintained for each unique value in the user-specified grouping column. In case of window-based aggregations, aggregate values are maintained for each window the event-time of a row falls into. Let's understand this with an illustration. -Imagine the quick example is modified and the stream contains lines along with the time when the line was generated. Instead of running word counts, we want to count words within 10 minute windows, updating every 5 minutes. That is, word counts in words received between 10 minute windows 12:00 - 12:10, 12:05 - 12:15, 12:10 - 12:20, etc. Note that 12:00 - 12:10 means data that arrived after 12:00 but before 12:10. Now, consider a word that was received at 12:07. This word should increment the counts corresponding to two windows 12:00 - 12:10 and 12:05 - 12:15. So the counts will be indexed by both, the grouping key (i.e. the word) and the window (can be calculated from the event-time). +Imagine our quick example is modified and the stream now contains lines along with the time when the line was generated. Instead of running word counts, we want to count words within 10 minute windows, updating every 5 minutes. That is, word counts in words received between 10 minute windows 12:00 - 12:10, 12:05 - 12:15, 12:10 - 12:20, etc. Note that 12:00 - 12:10 means data that arrived after 12:00 but before 12:10. Now, consider a word that was received at 12:07. This word should increment the counts corresponding to two windows 12:00 - 12:10 and 12:05 - 12:15. So the counts will be indexed by both, the grouping key (i.e. the word) and the window (can be calculated from the event-time). The result tables would look something like the following. --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org