Thanks Stephan.
Using flink-dist jar solves the issue.

Thanks and Regards
Amit Pawar


On Thu, Apr 23, 2015 at 2:02 PM, Stephan Ewen <se...@apache.org> wrote:

> Hi!
>
> What you describe sounds pretty much like a version mixup - NoSuchMethod
> indicates one part of the code is out of sync with the other. Can you make
> sure that you have all jars from the same Flink version in the classpath?
>
> For the Optimizer Exception: The cause may be a similar issue (version
> mixup) or a completely missing jar file. If you use the big jar file from
> flink-dist in version 0.9, does that error occur?
>
> Greetings,
> Stephan
>
>
> On Thu, Apr 23, 2015 at 2:01 AM, Amit Pawar <amitpawar5...@gmail.com>
> wrote:
>
> > Many thanks Stephan.
> > I followed your instructions and it was working fine when I had the
> > required flink projects in the build path,
> > later when I substituted it by adding respective dependencies with the
> > snapshots in pom, I am getting the below exception at
> >
> > OptimizedPlan opPlan = op.compile(env.createProgramPlan());
> >
> > Exception in thread "main" org.apache.flink.optimizer.CompilerException:
> > Class 'org.apache.flink.compiler.postpass.JavaApiPostPass' is not an
> > optimizer post-pass.
> > at
> >
> >
> org.apache.flink.optimizer.Optimizer.getPostPassFromPlan(Optimizer.java:573)
> > at org.apache.flink.optimizer.Optimizer.compile(Optimizer.java:402)
> > at thesis.examples.SampleTest.main(SampleTest.java:189)
> > Caused by: java.lang.ClassCastException: class
> > org.apache.flink.compiler.postpass.JavaApiPostPass
> > at java.lang.Class.asSubclass(Class.java:3208)
> > at
> >
> >
> org.apache.flink.optimizer.Optimizer.getPostPassFromPlan(Optimizer.java:557)
> > ... 2 more
> >
> > I can work around that by having the necessary flink projects in
> > eclipse/build path, but then I face  different issue of scala, No such
> > method exception on env.execute();
> >
> > Please advise.
> >
> > Thanks and Regards
> > Amit Pawar
> >
> >
> > On Mon, Mar 30, 2015 at 6:26 PM, Stephan Ewen <se...@apache.org> wrote:
> >
> > > Hi Amit!
> > >
> > > The DataSet API is basically a fluent builder for the internal DAG of
> > > operations, the "Plan". This plan is build when you call
> "env.execute()".
> > >
> > > You can directly get the Plan by calling
> > > ExecutionEnvironment#createProgramPlan()
> > >
> > > The JSON plan has in addition the information inserted by the Optimizer
> > > (what partitioning to use where, what keys to use). This is called the
> > > "OptimizedPlan".
> > > To obtain that, you have to push the Plan through the Optimizer:
> > > "OptimizedPlan op = new Optimizer(new DataStaristics(), new
> > > DefaultCostEstimator()).compile(plan)"
> > >
> > > That optimized plan has everything in information for the execution.
> The
> > > JSON is created from that OptimizedPlan via "new
> > > PlanJSONDumpGenerator().getOptimizerPlanAsJSON(optimizedPlan)"
> > >
> > > Note: These classnames and instructions refer to Flink 0.9. For version
> > > 0.8, the names are a bit different.
> > >
> > > Greetings,
> > > Stephan
> > >
> > >
> > >
> > > On Mon, Mar 30, 2015 at 5:22 PM, Amit Pawar <amitpawar5...@gmail.com>
> > > wrote:
> > >
> > > > Hi
> > > >
> > > > I am trying to extract/retrieve the Flink execution plan. I managed
> to
> > > get
> > > > it as JSON string in following ways:
> > > > 1. Using JAR - via PackagedProgram using getPreviewPlan() ; or
> > > > 2. Directly in program - via ExecutionEnvironment's
> getExecutionPlan()
> > > >
> > > > My question is - Is it possible to retrieve directly the Plan object?
> > > > I tried for this but was not successful as submitting the jar takes
> us
> > > into
> > > > interactive mode, and in order to use the other mode,
> > programEntryPoint,
> > > > the main class needs to implement Program interface with getPlan
> > method.
> > > >
> > > > Even if we manage to get the execution plan as a Plan object, will it
> > be
> > > > different from what we have using JSON string? like in terms of -
> > > > 1. What are the datatypes used in the dataset's tuple
> > > > 2. On what key is the Join taking place
> > > > 3. Filtering predicate
> > > > 4. Field for Distinct and so on
> > > > (JSON plan does have the operator tree but the contents field points
> to
> > > the
> > > > line of code in the class, which is not that helpful)
> > > >
> > > > If not, is it possible (by some other way) to get the above details
> > just
> > > by
> > > > using the Flink job/jar as an input?
> > > >
> > > >
> > > > Thanks and Regards
> > > > Amit Pawar
> > > >
> > >
> >
>

Reply via email to