[
https://issues.apache.org/jira/browse/SPARK-20708?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz resolved SPARK-20708.
-
Resolution: Fixed
Assignee: Dongjoon Hyun
Fix Version/s: 2.3.0
> M
[
https://issues.apache.org/jira/browse/SPARK-20708?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16032475#comment-16032475
]
Burak Yavuz commented on SPARK-20708:
-
Resolved by https://github.com/apache/spark/pull/17947
> M
Hi Paul,
>From what you're describing, it seems that stream1 is possibly generating
tons of small files and stream2 is OOMing because it tries to maintain an
in-memory list of files. Some notes/questions:
1. Parquet files are splittable, therefore having large parquet files
shouldn't be a
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: 3a556caa9429ae024a9ef84fb7abf147adf146f3
https://github.com/phpmyadmin/localized_docs/commit/3a556caa9429ae024a9ef84fb7abf147adf146f3
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
Hi Kant,
>
>
> 1. Can we use Spark Structured Streaming for stateless transformations
> just like we would do with DStreams or Spark Structured Streaming is only
> meant for stateful computations?
>
Of course you can do stateless transformations. Any map, filter, select,
type of transformation
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: d74c4b4691e26179f1de914d76b81d25c93ea214
https://github.com/phpmyadmin/localized_docs/commit/d74c4b4691e26179f1de914d76b81d25c93ea214
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/phpmyadmin
Commit: 37b73a2457778862b6539cdd8fa52511aa9132db
https://github.com/phpmyadmin/phpmyadmin/commit/37b73a2457778862b6539cdd8fa52511aa9132db
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date: 2017
Branch: refs/heads/QA_4_7
Home: https://github.com/phpmyadmin/phpmyadmin
Commit: 445677ecd9de799ab4d0e3b695ccf6a72f1cfe3d
https://github.com/phpmyadmin/phpmyadmin/commit/445677ecd9de799ab4d0e3b695ccf6a72f1cfe3d
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date: 2017
[
https://issues.apache.org/jira/browse/SPARK-20140?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz resolved SPARK-20140.
-
Resolution: Fixed
Fix Version/s: 2.3.0
2.2.1
> Remove hardco
[
https://issues.apache.org/jira/browse/SPARK-20140?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz reassigned SPARK-20140:
---
Assignee: Yash Sharma
> Remove hardcoded kinesis retry wait and max retr
[
https://issues.apache.org/jira/browse/SPARK-20140?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16013191#comment-16013191
]
Burak Yavuz commented on SPARK-20140:
-
resolved by https://github.com/apache/spark/pull/17467
Burak Yavuz created SPARK-20775:
---
Summary: from_json should also have an API where the schema is
specified with a string
Key: SPARK-20775
URL: https://issues.apache.org/jira/browse/SPARK-20775
Project
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: 252c862dd57bb83b08a72a76939986154bb43350
https://github.com/phpmyadmin/localized_docs/commit/252c862dd57bb83b08a72a76939986154bb43350
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
tly the same schema, but
> one side support null and the other doesn't, this exception (in union
> dataset) will be thrown?
>
>
>
> 2017-05-08 16:41 GMT-03:00 Burak Yavuz <brk...@gmail.com>:
>
>> I also want to add that generally these may be caused by the
>> `nu
I also want to add that generally these may be caused by the `nullability`
field in the schema.
On Mon, May 8, 2017 at 12:25 PM, Shixiong(Ryan) Zhu wrote:
> This is because RDD.union doesn't check the schema, so you won't see the
> problem unless you run RDD and hit
[
https://issues.apache.org/jira/browse/SPARK-20571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15998680#comment-15998680
]
Burak Yavuz commented on SPARK-20571:
-
Thanks!
> Flaky SparkR StructuredStreaming te
[
https://issues.apache.org/jira/browse/SPARK-20441?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz resolved SPARK-20441.
-
Resolution: Fixed
Resolved with https://github.com/apache/spark/pull/17735
> Within the s
[
https://issues.apache.org/jira/browse/SPARK-20432?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz closed SPARK-20432.
---
Resolution: Duplicate
> Unioning two identical Streaming DataFrames fails during attrib
[
https://issues.apache.org/jira/browse/SPARK-20571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15994185#comment-15994185
]
Burak Yavuz commented on SPARK-20571:
-
cc [~felixcheung]
> Flaky SparkR StructuredStreaming te
Burak Yavuz created SPARK-20571:
---
Summary: Flaky SparkR StructuredStreaming tests
Key: SPARK-20571
URL: https://issues.apache.org/jira/browse/SPARK-20571
Project: Spark
Issue Type: Test
Burak Yavuz created SPARK-20549:
---
Summary: java.io.CharConversionException: Invalid UTF-32 in
JsonToStructs
Key: SPARK-20549
URL: https://issues.apache.org/jira/browse/SPARK-20549
Project: Spark
[
https://issues.apache.org/jira/browse/SPARK-20496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz resolved SPARK-20496.
-
Resolution: Fixed
Fix Version/s: 2.2.0
2.1.2
Resolved with https
[
https://issues.apache.org/jira/browse/SPARK-20496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz reassigned SPARK-20496:
---
Assignee: Bill Chambers
> KafkaWriter Uses Unanalyzed Logical P
Burak Yavuz created SPARK-20432:
---
Summary: Unioning two identical Streaming DataFrames fails during
attribute resolution
Key: SPARK-20432
URL: https://issues.apache.org/jira/browse/SPARK-20432
Project
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/phpmyadmin
Commit: 39fc41d15c6b2bfd566a7944d1a762b98f443d19
https://github.com/phpmyadmin/phpmyadmin/commit/39fc41d15c6b2bfd566a7944d1a762b98f443d19
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date: 2017
Burak Yavuz created SPARK-20301:
---
Summary: Flakiness in StreamingAggregationSuite
Key: SPARK-20301
URL: https://issues.apache.org/jira/browse/SPARK-20301
Project: Spark
Issue Type: Test
[
https://issues.apache.org/jira/browse/SPARK-20301?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz updated SPARK-20301:
Labels: flaky-test (was: )
> Flakiness in StreamingAggregationSu
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: b339ecdf2d4b08ba0946d8a58f23f805f8f031ef
https://github.com/phpmyadmin/localized_docs/commit/b339ecdf2d4b08ba0946d8a58f23f805f8f031ef
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
Burak Yavuz created SPARK-20230:
---
Summary: FetchFailedExceptions should invalidate file caches in
MapOutputTracker even if newer stages are launched
Key: SPARK-20230
URL: https://issues.apache.org/jira/browse/SPARK
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: 27e1d79fc1be067ef3ab402398eaba7b1fdc96f5
https://github.com/phpmyadmin/localized_docs/commit/27e1d79fc1be067ef3ab402398eaba7b1fdc96f5
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: e0075726ad73bae46849758407ed6c1aacddbe51
https://github.com/phpmyadmin/localized_docs/commit/e0075726ad73bae46849758407ed6c1aacddbe51
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: 9e9bd81d8d9a69034f5cbb5af9ee22955f3b7d3d
https://github.com/phpmyadmin/localized_docs/commit/9e9bd81d8d9a69034f5cbb5af9ee22955f3b7d3d
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
[
https://issues.apache.org/jira/browse/SPARK-19911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz resolved SPARK-19911.
-
Resolution: Fixed
Assignee: Adam Budde
Fix Version/s: 2.2.0
Target
Hi Everett,
IIRC we added unionAll in Spark 2.0 which is the same implementation as rdd
union. The union in DataFrames with Spark 2.0 does dedeuplication, and
that's why you should be seeing the slowdown.
Best,
Burak
On Thu, Mar 16, 2017 at 4:14 PM, Everett Anderson
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: ea2f5f4500513e0856a4cb56f5954f46b01fb230
https://github.com/phpmyadmin/localized_docs/commit/ea2f5f4500513e0856a4cb56f5954f46b01fb230
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
Burak Yavuz created SPARK-19886:
---
Summary: reportDataLoss cause != null check is wrong for
Structured Streaming KafkaSource
Key: SPARK-19886
URL: https://issues.apache.org/jira/browse/SPARK-19886
[
https://issues.apache.org/jira/browse/SPARK-19813?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz resolved SPARK-19813.
-
Resolution: Fixed
Fix Version/s: 2.2.0
2.1.1
> maxFilesPerTrig
[
https://issues.apache.org/jira/browse/SPARK-19304?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz resolved SPARK-19304.
-
Resolution: Fixed
Fix Version/s: 2.2.0
Target Version/s: 2.2.0
Resolved
[
https://issues.apache.org/jira/browse/SPARK-19304?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz reassigned SPARK-19304:
---
Assignee: Gaurav Shah
> Kinesis checkpoint recovery is 10x s
[
https://issues.apache.org/jira/browse/SPARK-19595?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz resolved SPARK-19595.
-
Resolution: Fixed
Fix Version/s: 2.2.0
Resolved by https://github.com/apache/spark/pull
[
https://issues.apache.org/jira/browse/SPARK-19595?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz reassigned SPARK-19595:
---
Assignee: Hyukjin Kwon
> from_json produces only a single row when input is a json ar
Burak Yavuz created SPARK-19813:
---
Summary: maxFilesPerTrigger combo latestFirst may miss old files
in combination with maxFileAge in FileStreamSource
Key: SPARK-19813
URL: https://issues.apache.org/jira/browse
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/phpmyadmin
Commit: b17a0f91cc23bf8800c4aa715d5a7c9a050de41a
https://github.com/phpmyadmin/phpmyadmin/commit/b17a0f91cc23bf8800c4aa715d5a7c9a050de41a
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date: 2017
Burak Yavuz created SPARK-19774:
---
Summary: StreamExecution should call stop() on sources when a
stream fails
Key: SPARK-19774
URL: https://issues.apache.org/jira/browse/SPARK-19774
Project: Spark
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: bb684c9e7108bb04ac7421b6126f1f0b05209c94
https://github.com/phpmyadmin/localized_docs/commit/bb684c9e7108bb04ac7421b6126f1f0b05209c94
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
[
https://issues.apache.org/jira/browse/SPARK-19405?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz resolved SPARK-19405.
-
Resolution: Fixed
Assignee: Adam Budde
Fix Version/s: 2.2.0
Resolved with: https
Burak Yavuz created SPARK-19637:
---
Summary: add to_json APIs to SQL
Key: SPARK-19637
URL: https://issues.apache.org/jira/browse/SPARK-19637
Project: Spark
Issue Type: New Feature
Congrats Takuya!
On Mon, Feb 13, 2017 at 2:17 PM, Dilip Biswal wrote:
> Congratulations, Takuya!
>
> Regards,
> Dilip Biswal
> Tel: 408-463-4980 <(408)%20463-4980>
> dbis...@us.ibm.com
>
>
>
> - Original message -
> From: Takeshi Yamamuro
>
[
https://issues.apache.org/jira/browse/SPARK-19542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz resolved SPARK-19542.
-
Resolution: Fixed
Fix Version/s: 2.2.0
2.1.1
> Delete the t
Burak Yavuz created SPARK-19543:
---
Summary: from_json fails when the input row is empty
Key: SPARK-19543
URL: https://issues.apache.org/jira/browse/SPARK-19543
Project: Spark
Issue Type: Bug
le. How can I
> do that?
>
> 2017-02-06 14:25 GMT-08:00 Burak Yavuz <brk...@gmail.com>:
>
>> Hi Egor,
>>
>> Structured Streaming handles all of its metadata itself, which files are
>> actually valid, etc. You may use the "create table" syntax in SQL
Hi Egor,
Structured Streaming handles all of its metadata itself, which files are
actually valid, etc. You may use the "create table" syntax in SQL to treat
it like a hive table, but it will handle all partitioning information in
its own metadata log. Is there a specific reason that you want to
Branch: refs/heads/QA_4_7
Home: https://github.com/phpmyadmin/phpmyadmin
Commit: 98ad19576a8fb0fa588a6b060b88a7ab514b23e9
https://github.com/phpmyadmin/phpmyadmin/commit/98ad19576a8fb0fa588a6b060b88a7ab514b23e9
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date: 2017
ount happens most likely rdd.isCheckpointed
> will be false, and the count will be on the rdd before it was checkpointed.
> what is the benefit of that?
>
>
> On Thu, Jan 26, 2017 at 11:19 PM, Burak Yavuz <brk...@gmail.com> wrote:
>
>> Hi,
>>
>> One of the goal
Hi,
One of the goals of checkpointing is to cut the RDD lineage. Otherwise you
run into StackOverflowExceptions. If you eagerly checkpoint, you basically
cut the lineage there, and the next operations all depend on the
checkpointed DataFrame. If you don't checkpoint, you continue to build the
[
https://issues.apache.org/jira/browse/SPARK-18218?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz updated SPARK-18218:
Assignee: Weichen Xu
> Optimize BlockMatrix multiplication, which may cause OOM and
[
https://issues.apache.org/jira/browse/SPARK-18218?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz resolved SPARK-18218.
-
Resolution: Implemented
Fix Version/s: 2.2.0
Resolved by https://github.com/apache/spark
[
https://issues.apache.org/jira/browse/SPARK-19378?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz updated SPARK-19378:
Description:
If you have a StreamingDataFrame with an aggregation, we report a metric called
Burak Yavuz created SPARK-19378:
---
Summary: StateOperator metrics should still return the total
number of rows in state even if there was no data for a trigger
Key: SPARK-19378
URL: https://issues.apache.org/jira
Hi,
Have you tried creating more column blocks?
BlockMatrix matrix = cmatrix.toBlockMatrix(100, 100);
for example.
Is your data randomly spread out, or do you generally have clusters of
data points together?
On Wed, Jan 25, 2017 at 4:23 AM, Petr Shestov wrote:
> Hi
[
https://issues.apache.org/jira/browse/SPARK-18020?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz updated SPARK-18020:
Assignee: Takeshi Yamamuro
> Kinesis receiver does not snapshot when shard comple
[
https://issues.apache.org/jira/browse/SPARK-18020?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz resolved SPARK-18020.
-
Resolution: Fixed
Fix Version/s: 2.2.0
Resolved by https://github.com/apache/spark/pull
deshpande <deshpandesh...@gmail.com>
wrote:
> Thanks Burak. But with BloomFilter, won't I be getting a false poisitve?
>
> On Wed, Jan 25, 2017 at 11:28 AM, Burak Yavuz <brk...@gmail.com> wrote:
>
>> I noticed that 1 wouldn't be a problem, because you'll save t
gave me 2 solutions
> 1. Bloom filter --> problem in repopulating the bloom filter on restarts
> 2. keeping the state of the unique ids
>
> Please elaborate on 2.
>
>
>
> On Wed, Jan 25, 2017 at 10:53 AM, Burak Yavuz <brk...@gmail.com> wrote:
>
>> I don't
gt; Thanks
>
> On Wed, Jan 25, 2017 at 9:13 AM, Burak Yavuz <brk...@gmail.com> wrote:
>
>> Off the top of my head... (Each may have it's own issues)
>>
>> If upstream you add a uniqueId to all your records, then you may use a
>> BloomFilter to appro
Off the top of my head... (Each may have it's own issues)
If upstream you add a uniqueId to all your records, then you may use a
BloomFilter to approximate if you've seen a row before.
The problem I can see with that approach is how to repopulate the bloom
filter on restarts.
If you are certain
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/phpmyadmin
Commit: 5fbd21a0bd453f1c5393cc5f143a28bf36280e02
https://github.com/phpmyadmin/phpmyadmin/commit/5fbd21a0bd453f1c5393cc5f143a28bf36280e02
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date: 2017
Thank you very much everyone! Hoping to help out the community as much as I
can!
Best,
Burak
On Tue, Jan 24, 2017 at 2:29 PM, Jacek Laskowski wrote:
> Wow! At long last. Congrats Burak and Holden!
>
> p.s. I was a bit worried that the process of accepting new committers
> is
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/phpmyadmin
Commit: 60b77eea6e73c67199a1f11e7d83c9c6212f08c3
https://github.com/phpmyadmin/phpmyadmin/commit/60b77eea6e73c67199a1f11e7d83c9c6212f08c3
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date: 2017
Branch: refs/heads/QA_4_6
Home: https://github.com/phpmyadmin/phpmyadmin
Commit: d121c692265078fb75b19e9b1f6eb49aae54c9ab
https://github.com/phpmyadmin/phpmyadmin/commit/d121c692265078fb75b19e9b1f6eb49aae54c9ab
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date: 2017
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: 9e6df7a487803450187d60e21d3c8fe1341fca8b
https://github.com/phpmyadmin/localized_docs/commit/9e6df7a487803450187d60e21d3c8fe1341fca8b
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
Hi Maciej,
I believe it would be useful to either fix the documentation or fix the
implementation. I'll leave it to the community to comment on. The code
right now disallows intervals provided in months and years, because they
are not a "consistently" fixed amount of time. A month can be 28, 29,
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: 8004d4ceb72c7446a3dccae369a0912c8948b72c
https://github.com/phpmyadmin/localized_docs/commit/8004d4ceb72c7446a3dccae369a0912c8948b72c
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: d580dc246ce2e159cf587637bc0d3cf5b06ad8b6
https://github.com/phpmyadmin/localized_docs/commit/d580dc246ce2e159cf587637bc0d3cf5b06ad8b6
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: c5fa88dc04a4a141da1be9af881a6f9b86575fdd
https://github.com/phpmyadmin/localized_docs/commit/c5fa88dc04a4a141da1be9af881a6f9b86575fdd
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: c2fa72dd06b4bd698b2d915b49c0745ef219f594
https://github.com/phpmyadmin/localized_docs/commit/c2fa72dd06b4bd698b2d915b49c0745ef219f594
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/phpmyadmin
Commit: 6fe9e68e762e257c7dd3ca3997fcef59e148789f
https://github.com/phpmyadmin/phpmyadmin/commit/6fe9e68e762e257c7dd3ca3997fcef59e148789f
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date: 2016
[
https://issues.apache.org/jira/browse/SPARK-18952?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz updated SPARK-18952:
Summary: regex strings not properly escaped in codegen for aggregations
(was: regex strings
Burak Yavuz created SPARK-18952:
---
Summary: regex strings not properly escaped in codegen
Key: SPARK-18952
URL: https://issues.apache.org/jira/browse/SPARK-18952
Project: Spark
Issue Type: Bug
Burak Yavuz created SPARK-18927:
---
Summary: MemorySink for StructuredStreaming can't recover from
checkpoint if location is provided in conf
Key: SPARK-18927
URL: https://issues.apache.org/jira/browse/SPARK-18927
Burak Yavuz created SPARK-18900:
---
Summary: Flaky Test: StateStoreSuite.maintenance
Key: SPARK-18900
URL: https://issues.apache.org/jira/browse/SPARK-18900
Project: Spark
Issue Type: Test
[
https://issues.apache.org/jira/browse/SPARK-1?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz updated SPARK-1:
Affects Version/s: (was: 2.1.0)
2.0.2
> partitionBy in DataStreamWri
Burak Yavuz created SPARK-1:
---
Summary: partitionBy in DataStreamWriter in Python throws _to_seq
not defined
Key: SPARK-1
URL: https://issues.apache.org/jira/browse/SPARK-1
Project: Spark
Burak Yavuz created SPARK-18868:
---
Summary: Flaky Test: StreamingQueryListenerSuite
Key: SPARK-18868
URL: https://issues.apache.org/jira/browse/SPARK-18868
Project: Spark
Issue Type: Test
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/localized_docs
Commit: 917856aa5d2fde4a7be01719d802a920df96cb4a
https://github.com/phpmyadmin/localized_docs/commit/917856aa5d2fde4a7be01719d802a920df96cb4a
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date:
Branch: refs/heads/master
Home: https://github.com/phpmyadmin/phpmyadmin
Commit: e73a59f18e949c7ee8ea2e15bbd95cfe785db674
https://github.com/phpmyadmin/phpmyadmin/commit/e73a59f18e949c7ee8ea2e15bbd95cfe785db674
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date: 2016
Branch: refs/heads/QA_4_6
Home: https://github.com/phpmyadmin/phpmyadmin
Commit: 2e561a296e29df9c2f57d959a7b8c519921dcd25
https://github.com/phpmyadmin/phpmyadmin/commit/2e561a296e29df9c2f57d959a7b8c519921dcd25
Author: Burak Yavuz <hitowerdi...@hotmail.com>
Date: 2016
Burak Yavuz created SPARK-18811:
---
Summary: Stream Source resolution should happen in StreamExecution
thread, not main thread
Key: SPARK-18811
URL: https://issues.apache.org/jira/browse/SPARK-18811
Hi Daniela,
This is trivial with Structured Streaming. If your Kafka cluster is 0.10.0
or above, you may use Spark 2.0.2 to create a Streaming DataFrame from
Kafka, and then also create a DataFrame using the JDBC connection, and you
may join those. In Spark 2.1, there's support for a function
[
https://issues.apache.org/jira/browse/SPARK-18475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706788#comment-15706788
]
Burak Yavuz commented on SPARK-18475:
-
I'd be happy to share performance results. You're right, I
[
https://issues.apache.org/jira/browse/SPARK-18634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz updated SPARK-18634:
Description:
There are some weird issues with exploding Python UDFs in SparkSQL.
There are 2
[
https://issues.apache.org/jira/browse/SPARK-18634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz updated SPARK-18634:
Description:
There are some weird issues with exploding Python UDFs in SparkSQL.
There are 2
[
https://issues.apache.org/jira/browse/SPARK-18634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz updated SPARK-18634:
Summary: Corruption and Correctness issues with exploding Python UDFs
(was: Issues with exploding
[
https://issues.apache.org/jira/browse/SPARK-18634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz updated SPARK-18634:
Description:
There are some weird issues with exploding Python UDFs in SparkSQL.
There are 2
Burak Yavuz created SPARK-18634:
---
Summary: Issues with exploding Python UDFs
Key: SPARK-18634
URL: https://issues.apache.org/jira/browse/SPARK-18634
Project: Spark
Issue Type: Bug
[
https://issues.apache.org/jira/browse/SPARK-18407?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15696337#comment-15696337
]
Burak Yavuz commented on SPARK-18407:
-
This is also resolved as part of
https://issues.apache.org
[
https://issues.apache.org/jira/browse/SPARK-18510?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15681677#comment-15681677
]
Burak Yavuz commented on SPARK-18510:
-
No. Working on a separate fix
> Partition schema infere
[
https://issues.apache.org/jira/browse/SPARK-18510?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15680328#comment-15680328
]
Burak Yavuz commented on SPARK-18510:
-
cc [~r...@databricks.com] I marked this as a blocker
[
https://issues.apache.org/jira/browse/SPARK-18510?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Burak Yavuz updated SPARK-18510:
Description:
Not sure if this is a regression from 2.0 to 2.1. I was investigating
Burak Yavuz created SPARK-18510:
---
Summary: Partition schema inference corrupts data
Key: SPARK-18510
URL: https://issues.apache.org/jira/browse/SPARK-18510
Project: Spark
Issue Type: Bug
401 - 500 of 1060 matches
Mail list logo