Hi! What you describe sounds pretty much like a version mixup - NoSuchMethod indicates one part of the code is out of sync with the other. Can you make sure that you have all jars from the same Flink version in the classpath?
For the Optimizer Exception: The cause may be a similar issue (version mixup) or a completely missing jar file. If you use the big jar file from flink-dist in version 0.9, does that error occur? Greetings, Stephan On Thu, Apr 23, 2015 at 2:01 AM, Amit Pawar <amitpawar5...@gmail.com> wrote: > Many thanks Stephan. > I followed your instructions and it was working fine when I had the > required flink projects in the build path, > later when I substituted it by adding respective dependencies with the > snapshots in pom, I am getting the below exception at > > OptimizedPlan opPlan = op.compile(env.createProgramPlan()); > > Exception in thread "main" org.apache.flink.optimizer.CompilerException: > Class 'org.apache.flink.compiler.postpass.JavaApiPostPass' is not an > optimizer post-pass. > at > > org.apache.flink.optimizer.Optimizer.getPostPassFromPlan(Optimizer.java:573) > at org.apache.flink.optimizer.Optimizer.compile(Optimizer.java:402) > at thesis.examples.SampleTest.main(SampleTest.java:189) > Caused by: java.lang.ClassCastException: class > org.apache.flink.compiler.postpass.JavaApiPostPass > at java.lang.Class.asSubclass(Class.java:3208) > at > > org.apache.flink.optimizer.Optimizer.getPostPassFromPlan(Optimizer.java:557) > ... 2 more > > I can work around that by having the necessary flink projects in > eclipse/build path, but then I face different issue of scala, No such > method exception on env.execute(); > > Please advise. > > Thanks and Regards > Amit Pawar > > > On Mon, Mar 30, 2015 at 6:26 PM, Stephan Ewen <se...@apache.org> wrote: > > > Hi Amit! > > > > The DataSet API is basically a fluent builder for the internal DAG of > > operations, the "Plan". This plan is build when you call "env.execute()". > > > > You can directly get the Plan by calling > > ExecutionEnvironment#createProgramPlan() > > > > The JSON plan has in addition the information inserted by the Optimizer > > (what partitioning to use where, what keys to use). This is called the > > "OptimizedPlan". > > To obtain that, you have to push the Plan through the Optimizer: > > "OptimizedPlan op = new Optimizer(new DataStaristics(), new > > DefaultCostEstimator()).compile(plan)" > > > > That optimized plan has everything in information for the execution. The > > JSON is created from that OptimizedPlan via "new > > PlanJSONDumpGenerator().getOptimizerPlanAsJSON(optimizedPlan)" > > > > Note: These classnames and instructions refer to Flink 0.9. For version > > 0.8, the names are a bit different. > > > > Greetings, > > Stephan > > > > > > > > On Mon, Mar 30, 2015 at 5:22 PM, Amit Pawar <amitpawar5...@gmail.com> > > wrote: > > > > > Hi > > > > > > I am trying to extract/retrieve the Flink execution plan. I managed to > > get > > > it as JSON string in following ways: > > > 1. Using JAR - via PackagedProgram using getPreviewPlan() ; or > > > 2. Directly in program - via ExecutionEnvironment's getExecutionPlan() > > > > > > My question is - Is it possible to retrieve directly the Plan object? > > > I tried for this but was not successful as submitting the jar takes us > > into > > > interactive mode, and in order to use the other mode, > programEntryPoint, > > > the main class needs to implement Program interface with getPlan > method. > > > > > > Even if we manage to get the execution plan as a Plan object, will it > be > > > different from what we have using JSON string? like in terms of - > > > 1. What are the datatypes used in the dataset's tuple > > > 2. On what key is the Join taking place > > > 3. Filtering predicate > > > 4. Field for Distinct and so on > > > (JSON plan does have the operator tree but the contents field points to > > the > > > line of code in the class, which is not that helpful) > > > > > > If not, is it possible (by some other way) to get the above details > just > > by > > > using the Flink job/jar as an input? > > > > > > > > > Thanks and Regards > > > Amit Pawar > > > > > >