Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17465
cc @ericl, @bogdanrdc, @adrian-ionescu, @cloud-fan
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
GitHub user rxin opened a pull request:
https://github.com/apache/spark/pull/17465
[SPARK-20136][SQL] Add num files and metadata operation timing to scan
operator metrics
## What changes were proposed in this pull request?
This patch adds explicit metadata operation timing
ide. This
patch introduces a new SQLMetrics.postDriverMetricUpdates function to do that,
and adds documentation to make it more obvious.
## How was this patch tested?
Updated a test case to use this method.
Author: Reynold Xin <r...@databricks.com>
Closes #17464 from rxin/SPARK-20134.
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17464
Merging in master/branch-2.1.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
ver side. This
patch introduces a new SQLMetrics.postDriverMetricUpdates function to do that,
and adds documentation to make it more obvious.
## How was this patch tested?
Updated a test case to use this method.
Author: Reynold Xin <r...@databricks.com>
Closes #17464 from rxin/SPARK-20134.
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17464#discussion_r108600240
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/ui/SQLListenerSuite.scala
---
@@ -477,9 +477,11 @@ private case class MyPlan(sc
GitHub user rxin opened a pull request:
https://github.com/apache/spark/pull/17464
[SPARK-20134][SQL] SQLMetrics.postDriverMetricUpdates to simplify driver
side metric updates
## What changes were proposed in this pull request?
It is not super intuitive how to update SQLMetric
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17424
Hm - so this would require us to update the test suite every time there is
an update to the docs?
---
If your project is set up for it, you can reply to this email and have your
reply appear
Repository: spark
Updated Branches:
refs/heads/master f88f56b83 -> 0a6c50711
[SPARK-20070][SQL] Fix 2.10 build
## What changes were proposed in this pull request?
Commit
https://github.com/apache/spark/commit/91fa80fe8a2480d64c430bd10f97b3d44c007bcc
broke the build for scala 2.10. The
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17420
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Repository: spark
Updated Branches:
refs/heads/master e011004be -> f88f56b83
[DOCS] Clarify round mode for format_number & round functions
## What changes were proposed in this pull request?
Updated the description for the `format_number` description to indicate that it
uses `HALF_EVEN`
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17399
Thanks - merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Repository: spark
Updated Branches:
refs/heads/master b5c5bd98e -> e011004be
[SPARK-19846][SQL] Add a flag to disable constraint propagation
## What changes were proposed in this pull request?
Constraint propagation can be computation expensive and block the driver
execution for long time.
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17186
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Repository: spark
Updated Branches:
refs/heads/master 91fa80fe8 -> b5c5bd98e
Disable generate codegen since it fails my workload.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/b5c5bd98
Tree:
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17399
@roxannemoslehi can you fix the title? We can then merge this.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17399
Yea we definitely need a better title. Thanks for contributing though.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Repository: spark
Updated Branches:
refs/heads/master b70c03a42 -> b0ae6a38a
Typo fixup in comment
## What changes were proposed in this pull request?
Fixup typo in comment.
## How was this patch tested?
Don't need.
Author: Ye Yin
Closes #17396 from hustcat/fix.
Project:
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17397
LGTM
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17396
Merging in master. Thanks.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17312
That would be pretty confusing wouldn't it? The table has 3 entries and the
title says only 2.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17312
Your screenshot had 3 executors. Why does it say 2?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17359
Why do we want this? Seems extremely low usage on this function in the wild.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Repository: spark
Updated Branches:
refs/heads/master a8877bdbb -> a04dcde8c
clarify array_contains function description
## What changes were proposed in this pull request?
The description in the comment for array_contains is vague/incomplete (i.e.,
doesn't mention that it returns `null` if
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17380
Thanks - merging in master/branch-2.1.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Repository: spark
Updated Branches:
refs/heads/branch-2.1 5c18b6c31 -> 9dfdd2adf
clarify array_contains function description
## What changes were proposed in this pull request?
The description in the comment for array_contains is vague/incomplete (i.e.,
doesn't mention that it returns
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17343
Can you add some documentation inline so in the future we'd know why
specific implementations were chosen?
---
If your project is set up for it, you can reply to this email and have your
reply
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17312
Can you put a screenshot here? Might actually be useful to have.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17318
Can you put the after exception in the pr description as well?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Repository: spark
Updated Branches:
refs/heads/branch-2.1 5fb70831b -> 780f6060c
[SQL][MINOR] Fix scaladoc for UDFRegistration
## What changes were proposed in this pull request?
Fix scaladoc for UDFRegistration
## How was this patch tested?
local build
Author: Jacek Laskowski
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17337
Merging in master/branch-2.1.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Repository: spark
Updated Branches:
refs/heads/master 3783539d7 -> 6326d406b
[SQL][MINOR] Fix scaladoc for UDFRegistration
## What changes were proposed in this pull request?
Fix scaladoc for UDFRegistration
## How was this patch tested?
local build
Author: Jacek Laskowski
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17330#discussion_r106758290
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/subquery.scala
---
@@ -61,6 +63,36 @@ abstract class SubqueryExpression
Repository: spark
Updated Branches:
refs/heads/master 376d78216 -> bfdeea5c6
[SPARK-18847][GRAPHX] PageRank gives incorrect results for graphs with sinks
## What changes were proposed in this pull request?
Graphs with sinks (vertices with no outgoing edges) don't have the expected
rank sum
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16483
Merging in master. Thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
GitHub user rxin opened a pull request:
https://github.com/apache/spark/pull/17322
[SPARK-19987][SQL] Pass all filters into FileIndex
## What changes were proposed in this pull request?
This is a tiny teeny refactoring to pass data filters also to the
FileIndex, so FileIndex
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17191
I personally have run into this issue and was surprised that we didn't
support it ... it's pretty verbose to retype everything.
If Postgres and MySQL both support it, I think we should do
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17303
Yes it'd be nice to have some benchmark on this.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Repository: spark
Updated Branches:
refs/heads/master 97cc5e5a5 -> 54a3697f1
[MINOR][CORE] Fix a info message of `prunePartitions`
## What changes were proposed in this pull request?
`PrunedInMemoryFileIndex.prunePartitions` shows `pruned NaN% partitions` for
the following case.
```scala
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17273
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Repository: spark
Updated Branches:
refs/heads/master 02c274eab -> 97cc5e5a5
[SPARK-19960][CORE] Move `SparkHadoopWriter` to `internal/io/`
## What changes were proposed in this pull request?
This PR introduces the following changes:
1. Move `SparkHadoopWriter` to `core/internal/io/`, so
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17304
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user rxin closed the pull request at:
https://github.com/apache/spark/pull/17301
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17166
hm it might be useful to have details, but it'd also be useful to have this
in the overview page without having to drill down. iiuc, the pr already has the
information in task list page, doesn't
GitHub user rxin opened a pull request:
https://github.com/apache/spark/pull/17301
[SPARK-19944][SQL] Move SQLConf from sql/core to sql/catalyst (branch-2.1)
## What changes were proposed in this pull request?
This patch moves SQLConf from sql/core to sql/catalyst. To minimize
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17273
I'd fix the log msg instead.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17292#discussion_r106093910
--- Diff: core/src/test/scala/org/apache/spark/SparkContextSuite.scala ---
@@ -537,6 +539,21 @@ class SparkContextSuite extends SparkFunSuite
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17264
In the future can we put the perf result in PR descriptions?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17285#discussion_r105976759
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SimpleCatalystConf.scala
---
@@ -0,0 +1,48 @@
+/*
+ * Licensed to the Apache
GitHub user rxin opened a pull request:
https://github.com/apache/spark/pull/17285
[SPARK-19944][SQL] Move SQLConf from sql/core to sql/catalyst
## What changes were proposed in this pull request?
This patch moves SQLConf from sql/core to sql/catalyst. To minimize the
changes
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16541
I didn't look into the details here, but very often scanning data twice
doesn't necessarily slow things down, especially in the case of sequential
scan.
---
If your project is set up for it, you
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16826#discussion_r105506911
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/internal/SessionState.scala ---
@@ -17,43 +17,70 @@
package org.apache.spark.sql.internal
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17241#discussion_r105453191
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Analyzer.scala
---
@@ -595,6 +594,11 @@ class Analyzer(
case view
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17241
SGTM
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17244
LGTM
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17220
I don't think you understand this. This value is here so if at some point
some user picked tungsten-sort, we won't break it. In recent versions of Spark
the default sort manager accomplishes the thing
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17220
If anything, we should just update the file to add a line of comment to
make sure people don't delete this in the future.
---
If your project is set up for it, you can reply to this email and have
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17220
Is this change even correct? This is here for backward compatibility.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17202#discussion_r104983300
--- Diff: sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala ---
@@ -576,6 +576,11 @@ class Dataset[T] private[sql](
val parsedDelay
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17202#discussion_r104983221
--- Diff: sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala ---
@@ -563,7 +563,7 @@ class Dataset[T] private[sql](
* @param eventTime the name
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17205
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Repository: spark
Updated Branches:
refs/heads/master 9a6ac7226 -> e420fd459
[SPARK-19843][SQL][FOLLOWUP] Classdoc for `IntWrapper` and `LongWrapper`
## What changes were proposed in this pull request?
This is as per suggestion by rxin at :
https://github.com/apache/spark/pull/17
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17205
LGTM too
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17184#discussion_r104845661
--- Diff:
common/unsafe/src/main/java/org/apache/spark/unsafe/types/UTF8String.java ---
@@ -897,41 +898,52 @@ public long toLong() {
break
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17184
I believe IBM J9 actually improved this specific case (their JIT handles
tons of exceptions better). Oh well -- if only JIT is perfect.
---
If your project is set up for it, you can reply
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17184#discussion_r104841789
--- Diff:
common/unsafe/src/main/java/org/apache/spark/unsafe/types/UTF8String.java ---
@@ -897,41 +898,52 @@ public long toLong() {
break
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17184#discussion_r104841761
--- Diff:
common/unsafe/src/main/java/org/apache/spark/unsafe/types/UTF8String.java ---
@@ -897,41 +898,52 @@ public long toLong() {
break
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17184#discussion_r104841735
--- Diff:
common/unsafe/src/main/java/org/apache/spark/unsafe/types/UTF8String.java ---
@@ -850,26 +850,27 @@ public UTF8String translate(Map<Charac
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17196#discussion_r104804384
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FilePartitionStrategy.scala
---
@@ -0,0 +1,156 @@
+/*
+ * Licensed
Github user rxin closed the pull request at:
https://github.com/apache/spark/pull/16958
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17196#discussion_r104798525
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FilePartitionStrategy.scala
---
@@ -0,0 +1,156 @@
+/*
+ * Licensed
GitHub user rxin opened a pull request:
https://github.com/apache/spark/pull/17196
[SPARK-19855][SQL] Create an internal FilePartitionStrategy interface
## What changes were proposed in this pull request?
The way we currently do file partitioning strategy is hard coded
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17166#discussion_r104595706
--- Diff: core/src/main/scala/org/apache/spark/SparkContext.scala ---
@@ -2250,6 +2250,25 @@ class SparkContext(config: SparkConf) extends
Logging
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17166#discussion_r104593920
--- Diff:
core/src/main/scala/org/apache/spark/scheduler/cluster/CoarseGrainedClusterMessage.scala
---
@@ -40,7 +40,8 @@ private[spark] object
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17166#discussion_r104593825
--- Diff: core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala
---
@@ -732,6 +732,13 @@ class DAGScheduler
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17166#discussion_r104593790
--- Diff: core/src/main/scala/org/apache/spark/executor/Executor.scala ---
@@ -158,7 +158,8 @@ private[spark] class Executor(
threadPool.execute(tr
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17166#discussion_r104593710
--- Diff: core/src/main/scala/org/apache/spark/SparkContext.scala ---
@@ -2250,6 +2250,25 @@ class SparkContext(config: SparkConf) extends
Logging
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17166#discussion_r104593724
--- Diff: core/src/main/scala/org/apache/spark/SparkContext.scala ---
@@ -2250,6 +2250,25 @@ class SparkContext(config: SparkConf) extends
Logging
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/15928
What do you mean? The improvement was small?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17114
Put the test case in a sql file?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17099#discussion_r103501851
--- Diff: sql/core/src/test/resources/sql-tests/inputs/inner-join.sql ---
@@ -0,0 +1,25 @@
+CREATE TEMPORARY VIEW t1 AS SELECT * FROM VALUES (1
Repository: spark
Updated Branches:
refs/heads/master a920a4369 -> 3e40f6c3d
[SPARK-17495][SQL] Add more tests for hive hash
## What changes were proposed in this pull request?
This PR adds tests hive-hash by comparing the outputs generated against Hive
1.2.1. Following datatypes are
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17049
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17053#discussion_r102889140
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalog.scala
---
@@ -251,7 +251,8 @@ abstract class ExternalCatalog
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17049
Looks good except that comment.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17049#discussion_r102881054
--- Diff:
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/HashExpressionsSuite.scala
---
@@ -71,6 +75,242 @@ class HashExpressionsSuite
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17002
Yea @gatorsmile be careful in the future and check the commit hash.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17002#discussion_r102070142
--- Diff: sql/core/src/main/scala/org/apache/spark/sql/SparkSession.scala
---
@@ -95,16 +95,26 @@ class SparkSession private(
/**
* State
GitHub user rxin opened a pull request:
https://github.com/apache/spark/pull/17002
[SPARK-19669][SQL] Open up visibility for sharedState, sessionState, and a
few other functions
## What changes were proposed in this pull request?
To ease debugging, most of Spark SQL internals
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16977
Are tests flaky right now? Otherwise it seems like this has introduced
legitimate issue with the test timing out. Three times in a row.
---
If your project is set up for it, you can reply
Repository: spark
Updated Branches:
refs/heads/master 729ce3703 -> b486ffc86
[SPARK-19447] Make Range operator generate "recordsRead" metric
## What changes were proposed in this pull request?
The Range was modified to produce "recordsRead" metric instead of "generated
rows". The tests were
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16960
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16960
cc @hvanhovell if you have a min to review this ...
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16960#discussion_r101575264
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/metric/SQLMetricsSuite.scala
---
@@ -309,4 +314,84 @@ class SQLMetricsSuite extends
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16960#discussion_r101575199
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/metric/SQLMetricsSuite.scala
---
@@ -309,4 +314,84 @@ class SQLMetricsSuite extends
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16958
So nice when I got two LGTMs and then Jenkins disagreed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16826
What's WIP about this?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16611
For SQL, rather than "array", can we follow Python, e.g.
```
CREATE TEMPORARY TABLE tableA USING csv
OPTIONS (nullValue ['NA', 'null'], ...)
```
---
If your project
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16611#discussion_r101553890
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala ---
@@ -97,6 +99,15 @@ class DataFrameReader private[sql](sparkSession
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16534
Change looks good to me but I didn't look super carefully.
@holdenk can you take a look at this?
---
If your project is set up for it, you can reply to this email and have your
reply appear
901 - 1000 of 19261 matches
Mail list logo