Hi David,
First, can you please ask "technical support" questions on Gremlin-Users?
Second, please look at the documentation section around Spark and notice how a
TraversalSource is generated for GraphComputers.
http://tinkerpop.incubator.apache.org/docs/3.0.0.M9-incubating/#sparkgraphcomputer
(first line of the first code block )
Third, to your :remote question, please look at the second code block, third
line.
Good luck,
Marko.
http://markorodriguez.com
On Jun 24, 2015, at 11:37 AM, David Robinson <[email protected]> wrote:
> 1) In the very first example under OLTP Hadoop-Gemlin, is that example now
> missing a "traversal" call ?
>
> The following is what worked for me...but does not match the example in the
> docs:
>
> gremlin>
> graph=GraphFactory.open('/full/path/conf/hadoop/hadoop-gryo.properties')
> ==>hadoopgraph[gryoinputformat->gryooutputformat]
> *gremlin> g=graph.traversal()*
> ==>graphtraversalsource[hadoopgraph[gryoinputformat->gryooutputformat],
> standard]
> gremlin> g.V().count()
> ==>6
>
> 2) How do I verify that SparkGraphComputer is being used ?
>
> The example implies that I will see the following if using Spark:
>
> ==>hadoopgraph[gryoinputformat->gryooutputformat*[sparkgraphcomputer]]*
>
> However, even though the following is set in the properties file for
> opening the graph:
>
> gremlin.hadoop.defaultGraphComputer=SparkGraphComputer
>
> - set in the hadoop-gryo.properties - what I see is this:
>
> ==>hadoopgraph[gryoinputformat->gryooutputformat]
>
>
> 3) Has the syntax of the :remote command changed ?
>
> I cannot get the :remote example shown for Spark to work.
> The remote command fails with
>
> gremlin> :remote connect tinkerpop.hadoop
> './conf/hadoop/hadoop-gryo.properties'
> No such property: './conf/hadoop/hadoop-gryo.properties' for class:
> groovy.lang.Binding
> Display stack trace? [yN] n
>
> gremlin> :remote connect tinkerpop.hadoop "./conf/spark-gryo.properties"
> No such property: "./conf/spark-gryo.properties" for class:
> groovy.lang.Binding
> Display stack trace? [yN] n
>
> I also tried a full path and created a bogus sparkp-gryo.properties file.
>
> Can you please provide an example of the remote command for this situation
> ?
>
>
> 4) Can you please paste example contents of a spark-gryo.properties which
> is shown in the example ?
> Do I need both a hadoop-gryo.properties AND a spark-gryo.properties when
> doing remote commands ?
>
> There did not appear to be a sample in the source or build.
> The few sample files I found included a lot of Giraph stuff, which I
> am not using, and
> it isn't clear what is needed for Spark vs other GraphComputers.
>
> 5) If the Spark master has a URL like this: spark://test.machine.com:7077 ,
> what is the appropriate value for spark.master in the
> hadoop-gryo.properties and/or spark-gryo.properties ?
>
> I don't understand what spark.master=local[4] is supposed to mean.
>
> 6) Can you please explain what this is doing in the examples:
> g.engine(computer)
>
> "computer" doesn't ever appear to be set in the examples
> and this command doesn't work in my environment.
>
> gremlin> g.engine(computer)
> No such property: computer for class: groovysh_evaluate
> Display stack trace? [yN] n
>
> What is "g.engine" doing that isn't already done via a config file ?
>
> 7) Is the GA release candidate shipping with MapReduceGraphComputer undefined
> ?
>
> *COMING SOON*
>
>
> Thanks !