Thanks for the extra context, Marco. I thought you were trying to propose a solution.
On Thu, Dec 13, 2018 at 2:45 AM Marco Gaido <marcogaid...@gmail.com> wrote: > Hi Ryan, > > My goal with this email thread is to discuss with the community if there > are better ideas (as I was told many other people tried to address this). > I'd consider this as a brainstorming email thread. Once we have a good > proposal, then we can go ahead with a SPIP. > > Thanks, > Marco > > Il giorno mer 12 dic 2018 alle ore 19:13 Ryan Blue <rb...@netflix.com> ha > scritto: > >> Marco, >> >> I'm actually asking for a design doc that clearly states the problem and >> proposes a solution. This is a substantial change and probably should be an >> SPIP. >> >> I think that would be more likely to generate discussion than referring >> to PRs or a quick paragraph on the dev list, because the only people that >> are looking at it now are the ones already familiar with the problem. >> >> rb >> >> On Wed, Dec 12, 2018 at 2:05 AM Marco Gaido <marcogaid...@gmail.com> >> wrote: >> >>> Thank you all for your answers. >>> >>> @Ryan Blue <rb...@netflix.com> sure, let me state the problem more >>> clearly: imagine you have 2 dataframes with a common lineage (for instance >>> one is derived from the other by some filtering or anything you prefer). >>> And imagine you want to join these 2 dataframes. Currently, there is a fix >>> by Reynold which deduplicates the join condition in case the condition is >>> an equality one (please notice that in this case, it doesn't matter which >>> one is on the left and which one on the right). But if the condition >>> involves other comparisons, such as a ">" or a "<", this would result in an >>> analysis error, because the attributes on both sides are the same (eg. you >>> have the same id#3 attribute on both sides), and you cannot deduplicate >>> them blindly as which one is on a specific side matters. >>> >>> @Reynold Xin <r...@databricks.com> my proposal was to add a dataset id >>> in the metadata of each attribute, so that in this case we can distinguish >>> from which dataframe the attribute is coming from, ie. having the >>> DataFrames `df1` and `df2` where `df2` is derived from `df1`, >>> `df1.join(df2, df1("a") > df2("a"))` could be resolved because we would >>> know that the first attribute is taken from `df1` and so it has to be >>> resolved using it and the same for the other. But I am open to any approach >>> to this problem, if other people have better ideas/suggestions. >>> >>> Thanks, >>> Marco >>> >>> Il giorno mar 11 dic 2018 alle ore 18:31 Jörn Franke < >>> jornfra...@gmail.com> ha scritto: >>> >>>> I don’t know your exact underlying business problem, but maybe a graph >>>> solution, such as Spark Graphx meets better your requirements. Usually >>>> self-joins are done to address some kind of graph problem (even if you >>>> would not describe it as such) and is for these kind of problems much more >>>> efficient. >>>> >>>> Am 11.12.2018 um 12:44 schrieb Marco Gaido <marcogaid...@gmail.com>: >>>> >>>> Hi all, >>>> >>>> I'd like to bring to the attention of a more people a problem which has >>>> been there for long, ie, self joins. Currently, we have many troubles with >>>> them. This has been reported several times to the community and seems to >>>> affect many people, but as of now no solution has been accepted for it. >>>> >>>> I created a PR some time ago in order to address the problem ( >>>> https://github.com/apache/spark/pull/21449), but Wenchen mentioned he >>>> tried to fix this problem too but so far no attempt was successful because >>>> there is no clear semantic ( >>>> https://github.com/apache/spark/pull/21449#issuecomment-393554552). >>>> >>>> So I'd like to propose to discuss here which is the best approach for >>>> tackling this issue, which I think would be great to fix for 3.0.0, so if >>>> we decide to introduce breaking changes in the design, we can do that. >>>> >>>> Thoughts on this? >>>> >>>> Thanks, >>>> Marco >>>> >>>> >> >> -- >> Ryan Blue >> Software Engineer >> Netflix >> > -- Ryan Blue Software Engineer Netflix