Hi,

Re-writing the test in the following manner works. But I am not sure if
this is the correct way.

def testAggregationWithNull(): Unit = {

    val env = ExecutionEnvironment.getExecutionEnvironment
    val dataSet = env.fromElements[(Integer, String)]((123, "a"), (234,
"b"), (345, "c"), (0, "d"))

    implicit val rowInfo: TypeInformation[Row] = new RowTypeInfo(
      Seq(BasicTypeInfo.INT_TYPE_INFO, BasicTypeInfo.STRING_TYPE_INFO),
Seq("id", "name"))

    val rowDataSet = dataSet.map {
      entry =>
        val row = new Row(2)
        val amount = if(entry._1<100) null else entry._1
        row.setField(0, amount)
        row.setField(1, entry._2)
        row
    }

    val total =
rowDataSet.toTable.select('id.sum).collect().head.productElement(0)
    assertEquals(total, 702)
  }



On Sun, Jun 14, 2015 at 11:42 PM, Shiti Saxena <ssaxena....@gmail.com>
wrote:

> Hi,
>
> For
>
> val table = env.fromElements[(Integer, String)]((123, "a"), (234, "b"),
> (345, "c"), (null, "d")).toTable
>
> I get the following error,
>
> Error translating node 'Data Source "at
> org.apache.flink.api.scala.ExecutionEnvironment.fromElements(ExecutionEnvironment.scala:505)
> (org.apache.flink.api.java.io.CollectionInputFormat)" : NONE [[
> GlobalProperties [partitioning=RANDOM_PARTITIONED] ]] [[ LocalProperties
> [ordering=null, grouped=null, unique=null] ]]': null
> org.apache.flink.optimizer.CompilerException: Error translating node 'Data
> Source "at
> org.apache.flink.api.scala.ExecutionEnvironment.fromElements(ExecutionEnvironment.scala:505)
> (org.apache.flink.api.java.io.CollectionInputFormat)" : NONE [[
> GlobalProperties [partitioning=RANDOM_PARTITIONED] ]] [[ LocalProperties
> [ordering=null, grouped=null, unique=null] ]]': null
> at
> org.apache.flink.optimizer.plantranslate.JobGraphGenerator.preVisit(JobGraphGenerator.java:360)
> at
> org.apache.flink.optimizer.plantranslate.JobGraphGenerator.preVisit(JobGraphGenerator.java:103)
> at
> org.apache.flink.optimizer.plan.SourcePlanNode.accept(SourcePlanNode.java:87)
> at
> org.apache.flink.optimizer.plan.SingleInputPlanNode.accept(SingleInputPlanNode.java:199)
> at
> org.apache.flink.optimizer.plan.SingleInputPlanNode.accept(SingleInputPlanNode.java:199)
> at
> org.apache.flink.optimizer.plan.SingleInputPlanNode.accept(SingleInputPlanNode.java:199)
> at
> org.apache.flink.optimizer.plan.SingleInputPlanNode.accept(SingleInputPlanNode.java:199)
> at
> org.apache.flink.optimizer.plan.SingleInputPlanNode.accept(SingleInputPlanNode.java:199)
> at
> org.apache.flink.optimizer.plan.SingleInputPlanNode.accept(SingleInputPlanNode.java:199)
> at
> org.apache.flink.optimizer.plan.OptimizedPlan.accept(OptimizedPlan.java:127)
> at
> org.apache.flink.optimizer.plantranslate.JobGraphGenerator.compileJobGraph(JobGraphGenerator.java:170)
> at
> org.apache.flink.test.util.TestEnvironment.execute(TestEnvironment.java:52)
> at
> org.apache.flink.api.java.ExecutionEnvironment.execute(ExecutionEnvironment.java:789)
> at
> org.apache.flink.api.scala.ExecutionEnvironment.execute(ExecutionEnvironment.scala:576)
> at org.apache.flink.api.scala.DataSet.collect(DataSet.scala:544)
> at
> org.apache.flink.api.scala.table.test.AggregationsITCase.testAggregationWithNull(AggregationsITCase.scala:135)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
> at
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> at
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
> at
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> at
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> at
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
> at org.junit.rules.RunRules.evaluate(RunRules.java:20)
> at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
> at
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
> at
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
> at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
> at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
> at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
> at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
> at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
> at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
> at org.junit.runners.Suite.runChild(Suite.java:127)
> at org.junit.runners.Suite.runChild(Suite.java:26)
> at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
> at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
> at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
> at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
> at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
> at
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> at
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
> at org.junit.runner.JUnitCore.run(JUnitCore.java:160)
> at
> com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:78)
> at
> com.intellij.rt.execution.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:212)
> at com.intellij.rt.execution.junit.JUnitStarter.main(JUnitStarter.java:68)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at com.intellij.rt.execution.application.AppMain.main(AppMain.java:140)
> Caused by: java.lang.NullPointerException
> at
> org.apache.flink.api.common.typeutils.base.IntSerializer.serialize(IntSerializer.java:63)
> at
> org.apache.flink.api.common.typeutils.base.IntSerializer.serialize(IntSerializer.java:27)
> at
> org.apache.flink.api.scala.typeutils.CaseClassSerializer.serialize(CaseClassSerializer.scala:89)
> at
> org.apache.flink.api.scala.typeutils.CaseClassSerializer.serialize(CaseClassSerializer.scala:29)
> at
> org.apache.flink.api.java.io.CollectionInputFormat.writeObject(CollectionInputFormat.java:88)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at java.io.ObjectStreamClass.invokeWriteObject(ObjectStreamClass.java:988)
> at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1493)
> at
> java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1429)
> at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1175)
> at
> java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1541)
> at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1506)
> at
> java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1429)
> at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1175)
> at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:347)
> at
> org.apache.flink.util.InstantiationUtil.serializeObject(InstantiationUtil.java:314)
> at
> org.apache.flink.util.InstantiationUtil.writeObjectToConfig(InstantiationUtil.java:268)
> at
> org.apache.flink.runtime.operators.util.TaskConfig.setStubWrapper(TaskConfig.java:273)
> at
> org.apache.flink.optimizer.plantranslate.JobGraphGenerator.createDataSourceVertex(JobGraphGenerator.java:853)
> at
> org.apache.flink.optimizer.plantranslate.JobGraphGenerator.preVisit(JobGraphGenerator.java:260)
> ... 55 more
>
>
> Does this mean that the collect method is being called before doing the
> aggregation? Is this because base serializers do not handle null values
> like POJOSerializer? And is that why fromCollection does not support
> collections with null values?
>
> Or I could write the test using a file load if thats alright.
>
>
> On Sun, Jun 14, 2015 at 11:11 PM, Aljoscha Krettek <aljos...@apache.org>
> wrote:
>
>> Hi,
>> sorry, my mail client sent before I was done.
>>
>> I think the problem is that the Scala compiler derives a wrong type for
>> this statement:
>> val table = env.fromElements((123, "a"), (234, "b"), (345, "c"), (null,
>> "d")).toTable
>>
>> Because of the null value it derives (Any, String) as the type if you do
>> it like this, I think it should work:
>> val table = env.fromElements[(Integer, String)]((123, "a"), (234, "b"),
>> (345, "c"), (null, "d")).toTable
>>
>> I used Integer instead of Int because Scala will complain that null is
>> not a valid value for Int otherwise.
>>
>> Cheers,
>> Aljoscha
>>
>>
>> On Sun, 14 Jun 2015 at 19:34 Aljoscha Krettek <aljos...@apache.org>
>> wrote:
>>
>>> Hi,
>>> I think the problem is that the Scala compiler derives a wrong type for
>>> this statement:
>>>
>>>
>>>
>>> On Sun, 14 Jun 2015 at 18:28 Shiti Saxena <ssaxena....@gmail.com> wrote:
>>>
>>>> Hi Aljoscha,
>>>>
>>>> I created the issue FLINK-2210
>>>> <https://issues.apache.org/jira/browse/FLINK-2210> for aggregate on
>>>> null. I made changes to ExpressionAggregateFunction to handle ignore null
>>>> values. But I am unable to create a Table with null values in tests.
>>>>
>>>> The code I used is,
>>>>
>>>> def testAggregationWithNull(): Unit = {
>>>>
>>>>     val env = ExecutionEnvironment.getExecutionEnvironment
>>>>     val table = env.fromElements((123, "a"), (234, "b"), (345, "c"),
>>>> (null, "d")).toTable
>>>>
>>>>     val total = table.select('_1.sum).collect().head.productElement(0)
>>>>     assertEquals(total, 702)
>>>>   }
>>>>
>>>> and the error i get is,
>>>>
>>>> org.apache.flink.api.table.ExpressionException: Invalid expression
>>>> "('_1).sum": Unsupported type GenericType<java.lang.Object> for aggregation
>>>> ('_1).sum. Only numeric data types supported.
>>>> at
>>>> org.apache.flink.api.table.expressions.analysis.TypeCheck.apply(TypeCheck.scala:50)
>>>> at
>>>> org.apache.flink.api.table.expressions.analysis.TypeCheck.apply(TypeCheck.scala:31)
>>>> at
>>>> org.apache.flink.api.table.trees.Analyzer$$anonfun$analyze$1.apply(Analyzer.scala:34)
>>>> at
>>>> org.apache.flink.api.table.trees.Analyzer$$anonfun$analyze$1.apply(Analyzer.scala:31)
>>>> at scala.collection.immutable.List.foreach(List.scala:318)
>>>> at org.apache.flink.api.table.trees.Analyzer.analyze(Analyzer.scala:31)
>>>> at org.apache.flink.api.table.Table$$anonfun$1.apply(Table.scala:59)
>>>> at org.apache.flink.api.table.Table$$anonfun$1.apply(Table.scala:59)
>>>> at
>>>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
>>>> at
>>>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
>>>> at
>>>> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
>>>> at scala.collection.mutable.WrappedArray.foreach(WrappedArray.scala:34)
>>>> at scala.collection.TraversableLike$class.map(TraversableLike.scala:244)
>>>> at scala.collection.AbstractTraversable.map(Traversable.scala:105)
>>>> at org.apache.flink.api.table.Table.select(Table.scala:59)
>>>> at
>>>> org.apache.flink.api.scala.table.test.AggregationsITCase.testAggregationWithNull(AggregationsITCase.scala:135)
>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>> at
>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>> at
>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>> at
>>>> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
>>>> at
>>>> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>>>> at
>>>> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
>>>> at
>>>> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>>>> at
>>>> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
>>>> at
>>>> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>>>> at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
>>>> at org.junit.rules.RunRules.evaluate(RunRules.java:20)
>>>> at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
>>>> at
>>>> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
>>>> at
>>>> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>>>> at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
>>>> at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
>>>> at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
>>>> at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
>>>> at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
>>>> at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>>>> at org.junit.runners.Suite.runChild(Suite.java:127)
>>>> at org.junit.runners.Suite.runChild(Suite.java:26)
>>>> at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
>>>> at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
>>>> at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
>>>> at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
>>>> at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
>>>> at
>>>> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
>>>> at
>>>> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>>>> at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>>>> at org.junit.runner.JUnitCore.run(JUnitCore.java:160)
>>>> at
>>>> com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:78)
>>>> at
>>>> com.intellij.rt.execution.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:212)
>>>> at
>>>> com.intellij.rt.execution.junit.JUnitStarter.main(JUnitStarter.java:68)
>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>> at
>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>> at com.intellij.rt.execution.application.AppMain.main(AppMain.java:140)
>>>>
>>>>
>>>> The ExecutionEnvironment.fromCollection method also throws an error
>>>> when the collection contains a null.
>>>>
>>>> Could you please point out what I am doing wrong? How do we create a
>>>> Table with null values?
>>>>
>>>> In our application, we load a file and transform each line into a Row
>>>> resulting in a DataSet[Row]. This DataSet[Row] is then converted into
>>>> Table. Should I use the same approach for the test case?
>>>>
>>>>
>>>> Thanks,
>>>> Shiti
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> On Sun, Jun 14, 2015 at 4:10 PM, Shiti Saxena <ssaxena....@gmail.com>
>>>> wrote:
>>>>
>>>>> I'll do the fix
>>>>>
>>>>> On Sun, Jun 14, 2015 at 12:42 AM, Aljoscha Krettek <
>>>>> aljos...@apache.org> wrote:
>>>>>
>>>>>> I merged your PR for the RowSerializer. Teaching the aggregators to
>>>>>> deal with null values should be a very simple fix in
>>>>>> ExpressionAggregateFunction.scala. There it is simply always aggregating
>>>>>> the values without checking whether they are null. If you want you can 
>>>>>> also
>>>>>> fix that or I can quickly fix it.
>>>>>>
>>>>>> On Thu, 11 Jun 2015 at 10:40 Aljoscha Krettek <aljos...@apache.org>
>>>>>> wrote:
>>>>>>
>>>>>>> Cool, good to hear.
>>>>>>>
>>>>>>> The PojoSerializer already handles null fields. The RowSerializer
>>>>>>> can be modified in pretty much the same way. So you should start by 
>>>>>>> looking
>>>>>>> at the copy()/serialize()/deserialize() methods of PojoSerializer and 
>>>>>>> then
>>>>>>> modify RowSerializer in a similar way.
>>>>>>>
>>>>>>> You can also send me a private mail if you want more in-depth
>>>>>>> explanations.
>>>>>>>
>>>>>>> On Thu, 11 Jun 2015 at 09:33 Till Rohrmann <trohrm...@apache.org>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Hi Shiti,
>>>>>>>>
>>>>>>>> here is the issue [1].
>>>>>>>>
>>>>>>>> Cheers,
>>>>>>>> Till
>>>>>>>>
>>>>>>>> [1] https://issues.apache.org/jira/browse/FLINK-2203
>>>>>>>>
>>>>>>>> On Thu, Jun 11, 2015 at 8:42 AM Shiti Saxena <ssaxena....@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Hi Aljoscha,
>>>>>>>>>
>>>>>>>>> Could you please point me to the JIRA tickets? If you could
>>>>>>>>> provide some guidance on how to resolve these, I will work on them and
>>>>>>>>> raise a pull-request.
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>> Shiti
>>>>>>>>>
>>>>>>>>> On Thu, Jun 11, 2015 at 11:31 AM, Aljoscha Krettek <
>>>>>>>>> aljos...@apache.org> wrote:
>>>>>>>>>
>>>>>>>>>> Hi,
>>>>>>>>>> yes, I think the problem is that the RowSerializer does not
>>>>>>>>>> support null-values. I think we can add support for this, I will 
>>>>>>>>>> open a
>>>>>>>>>> Jira issue.
>>>>>>>>>>
>>>>>>>>>> Another problem I then see is that the aggregations can not
>>>>>>>>>> properly deal with null-values. This would need separate support.
>>>>>>>>>>
>>>>>>>>>> Regards,
>>>>>>>>>> Aljoscha
>>>>>>>>>>
>>>>>>>>>> On Thu, 11 Jun 2015 at 06:41 Shiti Saxena <ssaxena....@gmail.com>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi,
>>>>>>>>>>>
>>>>>>>>>>> In our project, we are using the Flink Table API and are facing
>>>>>>>>>>> the following issues,
>>>>>>>>>>>
>>>>>>>>>>> We load data from a CSV file and create a DataSet[Row]. The CSV
>>>>>>>>>>> file can also have invalid entries in some of the fields which we 
>>>>>>>>>>> replace
>>>>>>>>>>> with null when building the DataSet[Row].
>>>>>>>>>>>
>>>>>>>>>>> This DataSet[Row] is later on transformed to Table whenever
>>>>>>>>>>> required and specific operation such as select or aggregate, etc are
>>>>>>>>>>> performed.
>>>>>>>>>>>
>>>>>>>>>>> When a null value is encountered, we get a null pointer
>>>>>>>>>>> exception and the whole job fails. (We can see this by calling 
>>>>>>>>>>> collect on
>>>>>>>>>>> the resulting DataSet).
>>>>>>>>>>>
>>>>>>>>>>> The error message is similar to,
>>>>>>>>>>>
>>>>>>>>>>> Job execution failed.
>>>>>>>>>>> org.apache.flink.runtime.client.JobExecutionException: Job
>>>>>>>>>>> execution failed.
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:315)
>>>>>>>>>>> at
>>>>>>>>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
>>>>>>>>>>> at
>>>>>>>>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
>>>>>>>>>>> at
>>>>>>>>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:43)
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29)
>>>>>>>>>>> at
>>>>>>>>>>> scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29)
>>>>>>>>>>> at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:94)
>>>>>>>>>>> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
>>>>>>>>>>> at akka.actor.ActorCell.invoke(ActorCell.scala:487)
>>>>>>>>>>> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
>>>>>>>>>>> at akka.dispatch.Mailbox.run(Mailbox.scala:221)
>>>>>>>>>>> at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
>>>>>>>>>>> at
>>>>>>>>>>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>>>>>>>>>>> at
>>>>>>>>>>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>>>>>>>>>>> at
>>>>>>>>>>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>>>>>>>>>>> at
>>>>>>>>>>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>>>>>>>>>>> Caused by: java.lang.NullPointerException
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.api.common.typeutils.base.IntSerializer.serialize(IntSerializer.java:63)
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.api.common.typeutils.base.IntSerializer.serialize(IntSerializer.java:27)
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.api.table.typeinfo.RowSerializer.serialize(RowSerializer.scala:80)
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.api.table.typeinfo.RowSerializer.serialize(RowSerializer.scala:28)
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.runtime.plugable.SerializationDelegate.write(SerializationDelegate.java:51)
>>>>>>>>>>> at org.apache.flink.runtime.io
>>>>>>>>>>> .network.api.serialization.SpanningRecordSerializer.addRecord(SpanningRecordSerializer.java:76)
>>>>>>>>>>> at org.apache.flink.runtime.io
>>>>>>>>>>> .network.api.writer.RecordWriter.emit(RecordWriter.java:83)
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.runtime.operators.shipping.OutputCollector.collect(OutputCollector.java:65)
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.runtime.operators.chaining.ChainedMapDriver.collect(ChainedMapDriver.java:78)
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.runtime.operators.chaining.ChainedMapDriver.collect(ChainedMapDriver.java:78)
>>>>>>>>>>> at
>>>>>>>>>>> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:177)
>>>>>>>>>>> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
>>>>>>>>>>> at java.lang.Thread.run(Thread.java:724)
>>>>>>>>>>>
>>>>>>>>>>> Could this be because the RowSerializer does not support null
>>>>>>>>>>> values? (Similar to Flink-629
>>>>>>>>>>> <https://issues.apache.org/jira/browse/FLINK-629> )
>>>>>>>>>>>
>>>>>>>>>>> Currently, to overcome this issue, we are ignoring all the rows
>>>>>>>>>>> which may have null values. For example, we have a method cleanData 
>>>>>>>>>>> defined
>>>>>>>>>>> as,
>>>>>>>>>>>
>>>>>>>>>>> def cleanData(table:Table, relevantColumns:Seq[String]):Table = {
>>>>>>>>>>>     val whereClause: String = relevantColumns.map{
>>>>>>>>>>>         cName=>
>>>>>>>>>>>             s"$cName.isNotNull"
>>>>>>>>>>>     }.mkString(" && ")
>>>>>>>>>>>
>>>>>>>>>>>     val result :Table =
>>>>>>>>>>> table.select(relevantColumns.mkString(",")).where(whereClause)
>>>>>>>>>>>     result
>>>>>>>>>>> }
>>>>>>>>>>>
>>>>>>>>>>> Before operating on any Table, we use this method and then
>>>>>>>>>>> continue with task.
>>>>>>>>>>>
>>>>>>>>>>> Is this the right way to handle this? If not please let me know
>>>>>>>>>>> how to go about it.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Thanks,
>>>>>>>>>>> Shiti
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>
>>>>>
>>>>
>

Reply via email to