comphead commented on code in PR #2614:
URL: https://github.com/apache/datafusion-comet/pull/2614#discussion_r2449272476
##########
fuzz-testing/src/main/scala/org/apache/comet/fuzz/Meta.scala:
##########
@@ -35,100 +61,261 @@ object Meta {
(DataTypes.createDecimalType(10, 2), 0.2),
(DataTypes.DateType, 0.2),
(DataTypes.TimestampType, 0.2),
- // TimestampNTZType only in Spark 3.4+
- // (DataTypes.TimestampNTZType, 0.2),
+ (DataTypes.TimestampNTZType, 0.2),
(DataTypes.StringType, 0.2),
(DataTypes.BinaryType, 0.1))
- val stringScalarFunc: Seq[Function] = Seq(
- Function("substring", 3),
- Function("coalesce", 1),
- Function("starts_with", 2),
- Function("ends_with", 2),
- Function("contains", 2),
- Function("ascii", 1),
- Function("bit_length", 1),
- Function("octet_length", 1),
- Function("upper", 1),
- Function("lower", 1),
- Function("chr", 1),
- Function("init_cap", 1),
- Function("trim", 1),
- Function("ltrim", 1),
- Function("rtrim", 1),
- Function("string_space", 1),
- Function("rpad", 2),
- Function("rpad", 3), // rpad can have 2 or 3 arguments
- Function("hex", 1),
- Function("unhex", 1),
- Function("xxhash64", 1),
- Function("sha1", 1),
- // Function("sha2", 1), -- needs a second argument for number of bits
- Function("substring", 3),
- Function("btrim", 1),
- Function("concat_ws", 2),
- Function("repeat", 2),
- Function("length", 1),
- Function("reverse", 1),
- Function("instr", 2),
- Function("replace", 2),
- Function("translate", 2))
-
- val dateScalarFunc: Seq[Function] =
- Seq(Function("year", 1), Function("hour", 1), Function("minute", 1),
Function("second", 1))
+ private def createFunctionWithInputs(name: String, inputs: Seq[SparkType]):
Function = {
+ Function(name, Seq(FunctionSignature(inputs)))
+ }
+ private def createFunctions(name: String, signatures:
Seq[FunctionSignature]): Function = {
+ Function(name, signatures)
+ }
+
+ private def createUnaryStringFunction(name: String): Function = {
+ createFunctionWithInputs(name, Seq(SparkStringType))
+ }
+
+ private def createUnaryNumericFunction(name: String): Function = {
+ createFunctionWithInputs(name, Seq(SparkNumericType))
+ }
+
+ // Math expressions (corresponds to mathExpressions in QueryPlanSerde)
val mathScalarFunc: Seq[Function] = Seq(
- Function("abs", 1),
- Function("acos", 1),
- Function("asin", 1),
- Function("atan", 1),
- Function("Atan2", 1),
- Function("Cos", 1),
- Function("Exp", 2),
- Function("Ln", 1),
- Function("Log10", 1),
- Function("Log2", 1),
- Function("Pow", 2),
- Function("Round", 1),
- Function("Signum", 1),
- Function("Sin", 1),
- Function("Sqrt", 1),
- Function("Tan", 1),
- Function("Ceil", 1),
- Function("Floor", 1),
- Function("bool_and", 1),
- Function("bool_or", 1),
- Function("bitwise_not", 1))
+ createUnaryNumericFunction("abs"),
+ createUnaryNumericFunction("acos"),
+ createUnaryNumericFunction("asin"),
+ createUnaryNumericFunction("atan"),
+ createFunctionWithInputs("atan2", Seq(SparkNumericType, SparkNumericType)),
+ createUnaryNumericFunction("cos"),
+ createUnaryNumericFunction("exp"),
+ createUnaryNumericFunction("expm1"),
+ createFunctionWithInputs("log", Seq(SparkNumericType, SparkNumericType)),
+ createUnaryNumericFunction("log10"),
+ createUnaryNumericFunction("log2"),
+ createFunctionWithInputs("pow", Seq(SparkNumericType, SparkNumericType)),
+ createFunctionWithInputs("remainder", Seq(SparkNumericType,
SparkNumericType)),
+ createFunctions(
+ "round",
+ Seq(
+ FunctionSignature(Seq(SparkNumericType)),
+ FunctionSignature(Seq(SparkNumericType, SparkIntType)))),
+ createUnaryNumericFunction("signum"),
+ createUnaryNumericFunction("sin"),
+ createUnaryNumericFunction("sqrt"),
+ createUnaryNumericFunction("tan"),
+ createUnaryNumericFunction("ceil"),
+ createUnaryNumericFunction("floor"),
+ createFunctionWithInputs("unary_minus", Seq(SparkNumericType)))
+
+ // Hash expressions (corresponds to hashExpressions in QueryPlanSerde)
+ val hashScalarFunc: Seq[Function] = Seq(
+ createFunctionWithInputs("md5", Seq(SparkAnyType)),
+ createFunctionWithInputs("murmur3_hash", Seq(SparkAnyType)), // TODO can
take multiple columns
+ createFunctionWithInputs("sha2", Seq(SparkAnyType, SparkIntType)))
+ // String expressions (corresponds to stringExpressions in QueryPlanSerde)
+ val stringScalarFunc: Seq[Function] = Seq(
+ createUnaryStringFunction("ascii"),
+ createUnaryStringFunction("bit_length"),
+ createUnaryStringFunction("chr"),
+ createFunctionWithInputs("concat_ws", Seq(SparkStringType,
SparkStringType)),
Review Comment:
we should be supporting `concat` with strings input in 50.3.0 #2604 so need
to add it there.
Btw @andygrove `concat` supports string or arrays as input, looks like this
design supports it?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]