EdisonWang created SPARK-27033:
--
Summary: Add rule to optimize binary comparisons to its push down
format
Key: SPARK-27033
URL: https://issues.apache.org/jira/browse/SPARK-27033
Project: Spark
[
https://issues.apache.org/jira/browse/SPARK-27033?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-27033:
---
Description: _emphasized text_
> Add rule to optimize binary comparisons to its push down format
> -
[
https://issues.apache.org/jira/browse/SPARK-27033?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-27033:
---
Description: Currently, filters like this "select * from table where a + 1
>= 3" cannot be pushed do
[
https://issues.apache.org/jira/browse/SPARK-27033?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-27033:
---
Affects Version/s: (was: 3.0.0)
2.4.0
> Add rule to optimize binary compa
EdisonWang created SPARK-27079:
--
Summary: Fix typo & Remove useless imports
Key: SPARK-27079
URL: https://issues.apache.org/jira/browse/SPARK-27079
Project: Spark
Issue Type: Improvement
EdisonWang created SPARK-27202:
--
Summary: update comments to keep according with code
Key: SPARK-27202
URL: https://issues.apache.org/jira/browse/SPARK-27202
Project: Spark
Issue Type: Improveme
EdisonWang created SPARK-27232:
--
Summary: Skip to get file block location if locality is ignored
Key: SPARK-27232
URL: https://issues.apache.org/jira/browse/SPARK-27232
Project: Spark
Issue Type
[
https://issues.apache.org/jira/browse/SPARK-27232?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-27232:
---
Description:
`InMemoryFileIndex` needs to request file block location information in order
to do lo
[
https://issues.apache.org/jira/browse/SPARK-27232?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-27232:
---
Summary: Ignore file locality in InMemoryFileIndex if spark.locality.wait
is set to (was: Skip to g
[
https://issues.apache.org/jira/browse/SPARK-26500?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang closed SPARK-26500.
--
> Add conf to support ignore hdfs data locality
> -
>
>
EdisonWang created SPARK-26500:
--
Summary: Add conf to support ignore hdfs data locality
Key: SPARK-26500
URL: https://issues.apache.org/jira/browse/SPARK-26500
Project: Spark
Issue Type: Improve
[
https://issues.apache.org/jira/browse/SPARK-26500?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang resolved SPARK-26500.
Resolution: Not A Problem
> Add conf to support ignore hdfs data locality
> --
EdisonWang created SPARK-26544:
--
Summary: the string serialized from map/array type is not a valid
json (while hive is)
Key: SPARK-26544
URL: https://issues.apache.org/jira/browse/SPARK-26544
Project: Sp
[
https://issues.apache.org/jira/browse/SPARK-26544?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-26544:
---
Summary: escape string when serialize map/array make it a valid json (keep
alignment with hive) (wa
[
https://issues.apache.org/jira/browse/SPARK-26544?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-26544:
---
Summary: escape string when serialize map/array to make it a valid json
(keep alignment with hive)
EdisonWang created SPARK-27789:
--
Summary: Use stopEarly in codegen of ColumnarBatchScan
Key: SPARK-27789
URL: https://issues.apache.org/jira/browse/SPARK-27789
Project: Spark
Issue Type: Improve
EdisonWang created SPARK-28257:
--
Summary: Use ConfigEntry for hardcoded configs in SQL module
Key: SPARK-28257
URL: https://issues.apache.org/jira/browse/SPARK-28257
Project: Spark
Issue Type: I
[
https://issues.apache.org/jira/browse/SPARK-28123?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16899349#comment-16899349
]
EdisonWang commented on SPARK-28123:
seems it is the same with trim() on both sides?
EdisonWang created SPARK-29343:
--
Summary: Eliminate sorts without limit in the subquery of
Join/Aggregation
Key: SPARK-29343
URL: https://issues.apache.org/jira/browse/SPARK-29343
Project: Spark
[
https://issues.apache.org/jira/browse/SPARK-27789?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang resolved SPARK-27789.
Resolution: Not A Problem
> Use stopEarly in codegen of ColumnarBatchScan
> --
EdisonWang created SPARK-29918:
--
Summary: RecordBinaryComparator should check endianness when
compared by long
Key: SPARK-29918
URL: https://issues.apache.org/jira/browse/SPARK-29918
Project: Spark
[
https://issues.apache.org/jira/browse/SPARK-29918?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-29918:
---
Labels: correctness (was: )
> RecordBinaryComparator should check endianness when compared by long
EdisonWang created SPARK-30088:
--
Summary: Adaptive execution should convert SortMergeJoin to
BroadcastJoin when plan generates empty result
Key: SPARK-30088
URL: https://issues.apache.org/jira/browse/SPARK-30088
[
https://issues.apache.org/jira/browse/SPARK-28332?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17002148#comment-17002148
]
EdisonWang commented on SPARK-28332:
I've taken it [~cloud_fan]
> SQLMetric wrong i
EdisonWang created SPARK-32559:
--
Summary: Fix the trim logic in UTF8String.toInt/toLong did't
handle Chinese characters correctly
Key: SPARK-32559
URL: https://issues.apache.org/jira/browse/SPARK-32559
P
[
https://issues.apache.org/jira/browse/SPARK-32559?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-32559:
---
Description:
The trim logic in Cast expression introduced in
[https://github.com/apache/spark/pull/
[
https://issues.apache.org/jira/browse/SPARK-31952?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-31952:
---
Summary: The metric of MemoryBytesSpill is incorrect when doing Aggregate
(was: Fix incorrect Memor
EdisonWang created SPARK-31952:
--
Summary: Fix incorrect MemoryBytesSpill metric when doing Aggregate
Key: SPARK-31952
URL: https://issues.apache.org/jira/browse/SPARK-31952
Project: Spark
Issue
[
https://issues.apache.org/jira/browse/SPARK-31952?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-31952:
---
Attachment: image-2020-06-10-16-35-58-002.png
> The metric of MemoryBytesSpill is incorrect when doi
[
https://issues.apache.org/jira/browse/SPARK-31952?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-31952:
---
Description:
When doing Aggregate and spill occurs, the Spill(memory) metric is zero while
Spill(di
EdisonWang created SPARK-30806:
--
Summary: Evaluate once per group in UnboundedWindowFunctionFrame
Key: SPARK-30806
URL: https://issues.apache.org/jira/browse/SPARK-30806
Project: Spark
Issue Typ
EdisonWang created SPARK-33306:
--
Summary: TimezoneID is needed when there cast from Date to String
Key: SPARK-33306
URL: https://issues.apache.org/jira/browse/SPARK-33306
Project: Spark
Issue Ty
[
https://issues.apache.org/jira/browse/SPARK-33306?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-33306:
---
Description:
A simple way to reproduce this is
```
spark-shell --conf spark.sql.legacy.typeCoerci
[
https://issues.apache.org/jira/browse/SPARK-33306?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-33306:
---
Description:
A simple way to reproduce this is
```
spark-shell --conf spark.sql.legacy.typeCoerci
EdisonWang created SPARK-34633:
--
Summary: Self-join with script transformation failed to resolve
attribute correctly
Key: SPARK-34633
URL: https://issues.apache.org/jira/browse/SPARK-34633
Project: Spark
EdisonWang created SPARK-34634:
--
Summary: Self-join with script transformation failed to resolve
attribute correctly
Key: SPARK-34634
URL: https://issues.apache.org/jira/browse/SPARK-34634
Project: Spark
[
https://issues.apache.org/jira/browse/SPARK-34634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-34634:
---
Description:
To reproduce,
{code:java}
// code placeholder
create temporary view t as select * from
[
https://issues.apache.org/jira/browse/SPARK-34634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-34634:
---
Description:
To reproduce,
{code:java}
// code placeholder
create temporary view t as select * from
[
https://issues.apache.org/jira/browse/SPARK-34634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-34634:
---
Description:
To reproduce,
```
create temporary view t as select * from values 0, 1, 2 as t(a);
W
[
https://issues.apache.org/jira/browse/SPARK-34634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-34634:
---
Affects Version/s: 2.4.0
2.4.1
2.4.2
EdisonWang created SPARK-34819:
--
Summary: MapType supports orderable semantics
Key: SPARK-34819
URL: https://issues.apache.org/jira/browse/SPARK-34819
Project: Spark
Issue Type: Bug
Co
[
https://issues.apache.org/jira/browse/SPARK-34819?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-34819:
---
Issue Type: New Feature (was: Bug)
> MapType supports orderable semantics
> ---
EdisonWang created SPARK-39002:
--
Summary: StringEndsWith/Contains support push down to Parquet so
that we can leverage dictionary filter
Key: SPARK-39002
URL: https://issues.apache.org/jira/browse/SPARK-39002
EdisonWang created SPARK-39249:
--
Summary: Improve subexpression elimination for conditional
expressions
Key: SPARK-39249
URL: https://issues.apache.org/jira/browse/SPARK-39249
Project: Spark
Is
[
https://issues.apache.org/jira/browse/SPARK-39249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-39249:
---
Description: Currently we can do subexpression elimination for conditional
expressions when the sube
[
https://issues.apache.org/jira/browse/SPARK-37593?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-37593:
---
Summary: Optimize HeapMemoryAllocator to avoid memory waste in humongous
allocation when using G1GC
EdisonWang created SPARK-37593:
--
Summary: Optimize HeapMmeoryAllocator to avoid memory waste in
humongous allocation when using G1GC
Key: SPARK-37593
URL: https://issues.apache.org/jira/browse/SPARK-37593
[
https://issues.apache.org/jira/browse/SPARK-37593?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-37593:
---
Description:
Spark's tungsten memory model usually tries to allocate memory by one `page`
each time
[
https://issues.apache.org/jira/browse/SPARK-37593?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
EdisonWang updated SPARK-37593:
---
Summary: Optimize HeapMemoryAllocator to avoid memory waste when using G1GC
(was: Optimize HeapMemo
EdisonWang created SPARK-38160:
--
Summary: Shuffle by rand could lead to incorrect answers when
ShuffleFetchFailed happend
Key: SPARK-38160
URL: https://issues.apache.org/jira/browse/SPARK-38160
Project:
EdisonWang created SPARK-39476:
--
Summary: Disable Unwrap cast optimize when casting from Long to
Float/ Double or from Integer to Float
Key: SPARK-39476
URL: https://issues.apache.org/jira/browse/SPARK-39476
EdisonWang created SPARK-40035:
--
Summary: Avoid apply filter twice when listing files
Key: SPARK-40035
URL: https://issues.apache.org/jira/browse/SPARK-40035
Project: Spark
Issue Type: Improveme
EdisonWang created SPARK-41200:
--
Summary: BytesToBytesMap's longArray size can be up to MAX_CAPACITY
Key: SPARK-41200
URL: https://issues.apache.org/jira/browse/SPARK-41200
Project: Spark
Issue
53 matches
Mail list logo