Repository: spark Updated Branches: refs/heads/master 1a7e747ce -> 2949a835f
[SPARK-25063][SQL] Rename class KnowNotNull to KnownNotNull ## What changes were proposed in this pull request? Correct the class name typo checked in through SPARK-24891 ## How was this patch tested? Passed all existing tests. Closes #22049 from maryannxue/known-not-null. Authored-by: maryannxue <maryann...@apache.org> Signed-off-by: Xiao Li <gatorsm...@gmail.com> Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/2949a835 Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/2949a835 Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/2949a835 Branch: refs/heads/master Commit: 2949a835fae3f4ac6e3dae6f18cd8b6543b74601 Parents: 1a7e747 Author: maryannxue <maryann...@apache.org> Authored: Thu Aug 9 08:11:30 2018 -0700 Committer: Xiao Li <gatorsm...@gmail.com> Committed: Thu Aug 9 08:11:30 2018 -0700 ---------------------------------------------------------------------- .../org/apache/spark/sql/catalyst/analysis/Analyzer.scala | 4 ++-- .../spark/sql/catalyst/expressions/constraintExpressions.scala | 2 +- .../org/apache/spark/sql/catalyst/analysis/AnalysisSuite.scala | 6 +++--- 3 files changed, 6 insertions(+), 6 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/spark/blob/2949a835/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Analyzer.scala ---------------------------------------------------------------------- diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Analyzer.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Analyzer.scala index d23d43b..a7cd96e 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Analyzer.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Analyzer.scala @@ -2157,7 +2157,7 @@ class Analyzer( // trust the `nullable` information. // (cls, expr) => cls.isPrimitive && expr.nullable val needsNullCheck = (cls: Class[_], expr: Expression) => - cls.isPrimitive && !expr.isInstanceOf[KnowNotNull] + cls.isPrimitive && !expr.isInstanceOf[KnownNotNull] val inputsNullCheck = parameterTypes.zip(inputs) .filter { case (cls, expr) => needsNullCheck(cls, expr) } .map { case (_, expr) => IsNull(expr) } @@ -2167,7 +2167,7 @@ class Analyzer( // branch of `If` will be called if any of these checked inputs is null. Thus we can // prevent this rule from being applied repeatedly. val newInputs = parameterTypes.zip(inputs).map{ case (cls, expr) => - if (needsNullCheck(cls, expr)) KnowNotNull(expr) else expr } + if (needsNullCheck(cls, expr)) KnownNotNull(expr) else expr } inputsNullCheck .map(If(_, Literal.create(null, udf.dataType), udf.copy(children = newInputs))) .getOrElse(udf) http://git-wip-us.apache.org/repos/asf/spark/blob/2949a835/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/constraintExpressions.scala ---------------------------------------------------------------------- diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/constraintExpressions.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/constraintExpressions.scala index 53936aa..2917b0b 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/constraintExpressions.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/constraintExpressions.scala @@ -21,7 +21,7 @@ import org.apache.spark.sql.catalyst.InternalRow import org.apache.spark.sql.catalyst.expressions.codegen.{CodegenContext, ExprCode, FalseLiteral} import org.apache.spark.sql.types.DataType -case class KnowNotNull(child: Expression) extends UnaryExpression { +case class KnownNotNull(child: Expression) extends UnaryExpression { override def nullable: Boolean = false override def dataType: DataType = child.dataType http://git-wip-us.apache.org/repos/asf/spark/blob/2949a835/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/analysis/AnalysisSuite.scala ---------------------------------------------------------------------- diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/analysis/AnalysisSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/analysis/AnalysisSuite.scala index ba44484..a1c976d 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/analysis/AnalysisSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/analysis/AnalysisSuite.scala @@ -319,7 +319,7 @@ class AnalysisSuite extends AnalysisTest with Matchers { // only primitive parameter needs special null handling val udf2 = ScalaUDF((s: String, d: Double) => "x", StringType, string :: double :: Nil) val expected2 = - If(IsNull(double), nullResult, udf2.copy(children = string :: KnowNotNull(double) :: Nil)) + If(IsNull(double), nullResult, udf2.copy(children = string :: KnownNotNull(double) :: Nil)) checkUDF(udf2, expected2) // special null handling should apply to all primitive parameters @@ -327,7 +327,7 @@ class AnalysisSuite extends AnalysisTest with Matchers { val expected3 = If( IsNull(short) || IsNull(double), nullResult, - udf3.copy(children = KnowNotNull(short) :: KnowNotNull(double) :: Nil)) + udf3.copy(children = KnownNotNull(short) :: KnownNotNull(double) :: Nil)) checkUDF(udf3, expected3) // we can skip special null handling for primitive parameters that are not nullable @@ -339,7 +339,7 @@ class AnalysisSuite extends AnalysisTest with Matchers { val expected4 = If( IsNull(short), nullResult, - udf4.copy(children = KnowNotNull(short) :: double.withNullability(false) :: Nil)) + udf4.copy(children = KnownNotNull(short) :: double.withNullability(false) :: Nil)) // checkUDF(udf4, expected4) } --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org