Hey! Could you share the source code with us? It makes the debugging much much simpler ;-)
Stephan On Thu, Dec 4, 2014 at 1:40 PM, Maximilian Alber <[email protected] > wrote: > Could you reproduce the error now? > > > On Mon, Dec 1, 2014 at 5:12 PM, Maximilian Alber < > [email protected]> wrote: > >> Sure. >> Those are the params (in_file and random_file need to be set, out_file >> you can choose a name.): >> >> flink 'run', '-v', >> '/media/alber_disk/alber/Arbeit/Uni/Kurse/MA/repo/python/implementations/../../scripts/../flink/bump_boost/target/bump_boost-0.1.jar', >> '-c', 'bumpboost.Job', 'in_file=/tmp/tmpxj8LsC', 'out_file=/tmp/tmpuDSlIn', >> 'random_file=/tmp/tmpp_dFYE', 'dimensions=1', 'N=100', 'iterations=30', >> 'multi_bump_boost=1', 'gradient_descent_iterations=30', 'cache=False', >> 'start_width=1.0', 'min_width=-4', 'max_width=6', 'min_width_update=1e-08', >> 'max_width_update=10' >> >> On Mon, Dec 1, 2014 at 4:47 PM, Stephan Ewen <[email protected]> wrote: >> >>> Hi! >>> >>> I have been trying to re-implement the program you sent here (the code >>> is incomplete), but I cannot trigger the exception. Can you send us the >>> complete example? >>> >>> Stephan >>> >>> >>> On Fri, Nov 28, 2014 at 3:18 PM, Maximilian Alber < >>> [email protected]> wrote: >>> >>>> Hi Flinksters! >>>> >>>> I try to write a BulkIteration. Somehow I get a cryptic error message, >>>> at least I have no clue what's wrong: >>>> >>>> Code: >>>> >>>> var width = >>>> env.fromCollection[Vector](Seq(Vector.ones(config.dimensions) * >>>> config.startWidth)) map {x => new Vector(0, x.values)} >>>> var update = >>>> env.fromCollection[Vector](Seq(Vector.ones(config.dimensions) * 0.01F)) map >>>> {x => new Vector(1, x.values)} >>>> var lastGradient = >>>> env.fromCollection[Vector](Seq(Vector.zeros(config.dimensions))) map {x => >>>> new Vector(2, x.values)} >>>> >>>> var stepSet = width union update union lastGradient >>>> stepSet = stepSet.iterate(config.gradientDescentIterations){ >>>> stepSet => >>>> var width = stepSet filter {_.id == 0} >>>> var update = stepSet filter {_.id == 1} >>>> var lastGradient = stepSet filter {_.id == 2} >>>> >>>> val gradient = getGradient(X, residual, center, width) >>>> val term = gradient * lastGradient >>>> lastGradient = gradient >>>> >>>> update = update.map(new RichMapFunction[Vector, Vector]{ >>>> var term: Vector = null >>>> val minWidthUpdate = 0.00000001F >>>> val maxWidthUpdate = 10.0F >>>> override def open(config: Configuration) = { >>>> term = getRuntimeContext.getBroadcastVariable("term").toList.head >>>> } >>>> >>>> def map(x: Vector) = {x.condMul(term.isLess(0), >>>> 0.5F).condMul(term.isGreater(0), 1.2F).clip(minWidthUpdate, >>>> maxWidthUpdate)} >>>> }).withBroadcastSet(term, "term") >>>> /* >>>> width = width.map(new RichMapFunction[Vector, Vector]{ >>>> var update: Vector = null >>>> var gradient: Vector = null >>>> override def open(config: Configuration) = { >>>> update = >>>> getRuntimeContext.getBroadcastVariable("update").toList.head >>>> gradient = >>>> getRuntimeContext.getBroadcastVariable("gradient").toList.head >>>> } >>>> >>>> def map(x: Vector) = {(x + update * (gradient >>>> sign)).clip(config.minWidth, config.maxWidth)} >>>> }).withBroadcastSet(update, "update")withBroadcastSet(gradient, >>>> "gradient") >>>> */ >>>> >>>> width union update union lastGradient >>>> } >>>> >>>> >>>> Error: >>>> >>>> java.lang.IllegalStateException >>>> at >>>> org.apache.flink.compiler.dag.BulkPartialSolutionNode.setCandidateProperties(BulkPartialSolutionNode.java:50) >>>> at >>>> org.apache.flink.compiler.dag.BulkIterationNode.instantiateCandidate(BulkIterationNode.java:292) >>>> at >>>> org.apache.flink.compiler.dag.SingleInputNode.addLocalCandidates(SingleInputNode.java:367) >>>> at >>>> org.apache.flink.compiler.dag.SingleInputNode.getAlternativePlans(SingleInputNode.java:315) >>>> at >>>> org.apache.flink.compiler.dag.SingleInputNode.getAlternativePlans(SingleInputNode.java:258) >>>> at >>>> org.apache.flink.compiler.dag.SingleInputNode.getAlternativePlans(SingleInputNode.java:258) >>>> at >>>> org.apache.flink.compiler.dag.SingleInputNode.getAlternativePlans(SingleInputNode.java:258) >>>> at >>>> org.apache.flink.compiler.dag.BinaryUnionNode.getAlternativePlans(BinaryUnionNode.java:105) >>>> at >>>> org.apache.flink.compiler.dag.BinaryUnionNode.getAlternativePlans(BinaryUnionNode.java:104) >>>> at >>>> org.apache.flink.compiler.dag.SingleInputNode.getAlternativePlans(SingleInputNode.java:258) >>>> at >>>> org.apache.flink.compiler.dag.DataSinkNode.getAlternativePlans(DataSinkNode.java:194) >>>> at org.apache.flink.compiler.PactCompiler.compile(PactCompiler.java:561) >>>> at org.apache.flink.compiler.PactCompiler.compile(PactCompiler.java:466) >>>> at >>>> org.apache.flink.client.program.Client.getOptimizedPlan(Client.java:197) >>>> at >>>> org.apache.flink.client.program.Client.getOptimizedPlan(Client.java:210) >>>> at org.apache.flink.client.program.Client.run(Client.java:288) >>>> at org.apache.flink.client.program.Client.run(Client.java:231) >>>> at >>>> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:347) >>>> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:334) >>>> at >>>> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1001) >>>> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1025) >>>> >>>> Thanks! >>>> Cheers, >>>> Max >>>> >>> >>> >> >
