[ https://issues.apache.org/jira/browse/SPARK-38812?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
gaokui updated SPARK-38812: --------------------------- Description: when id do clean data,one rdd according one value(>or <) filter data, and then generate two different set,one is error data file, another is errorless data file. Now I use filter, but this filter must have two spark dag job, that cost too much. exactly some code like iterator.span(preidicate) and then return one tuple(iter1,iter2) one dataset will be spilted tow dataset in one rule data clean progress. i hope compute once not twice. was: when id do clean data,one rdd according one value(>or <) filter data, and then generate two different set,one is error data file, another is errorless data file. Now I use filter, but this filter must have two spark dag job, that cost too much. exactly some code like iterator.span(preidicate) and then return one tuple(iter1,iter2) > when i clean data ,I hope one rdd spill two rdd according clean data rule > ------------------------------------------------------------------------- > > Key: SPARK-38812 > URL: https://issues.apache.org/jira/browse/SPARK-38812 > Project: Spark > Issue Type: New Feature > Components: Spark Core > Affects Versions: 3.2.1 > Reporter: gaokui > Priority: Major > > when id do clean data,one rdd according one value(>or <) filter data, and > then generate two different set,one is error data file, another is errorless > data file. > Now I use filter, but this filter must have two spark dag job, that cost too > much. > exactly some code like iterator.span(preidicate) and then return one > tuple(iter1,iter2) > one dataset will be spilted tow dataset in one rule data clean progress. > i hope compute once not twice. -- This message was sent by Atlassian Jira (v8.20.1#820001) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org