Thank you for the pointers. I will take a look.
On Wed, Jun 24, 2015 at 4:58 PM, Marko Rodriguez <[email protected]> wrote: > Hi David, > > First, can you please ask "technical support" questions on Gremlin-Users? > > Second, please look at the documentation section around Spark and notice > how a TraversalSource is generated for GraphComputers. > > > http://tinkerpop.incubator.apache.org/docs/3.0.0.M9-incubating/#sparkgraphcomputer > (first line of the first code block ) > > Third, to your :remote question, please look at the second code block, > third line. > > Good luck, > Marko. > > http://markorodriguez.com > > On Jun 24, 2015, at 11:37 AM, David Robinson <[email protected]> wrote: > > > 1) In the very first example under OLTP Hadoop-Gemlin, is that example > now > > missing a "traversal" call ? > > > > The following is what worked for me...but does not match the example in > the > > docs: > > > > gremlin> > > graph=GraphFactory.open('/full/path/conf/hadoop/hadoop-gryo.properties') > > ==>hadoopgraph[gryoinputformat->gryooutputformat] > > *gremlin> g=graph.traversal()* > > ==>graphtraversalsource[hadoopgraph[gryoinputformat->gryooutputformat], > > standard] > > gremlin> g.V().count() > > ==>6 > > > > 2) How do I verify that SparkGraphComputer is being used ? > > > > The example implies that I will see the following if using Spark: > > > > ==>hadoopgraph[gryoinputformat->gryooutputformat*[sparkgraphcomputer]]* > > > > However, even though the following is set in the properties file for > > opening the graph: > > > > gremlin.hadoop.defaultGraphComputer=SparkGraphComputer > > > > - set in the hadoop-gryo.properties - what I see is this: > > > > ==>hadoopgraph[gryoinputformat->gryooutputformat] > > > > > > 3) Has the syntax of the :remote command changed ? > > > > I cannot get the :remote example shown for Spark to work. > > The remote command fails with > > > > gremlin> :remote connect tinkerpop.hadoop > > './conf/hadoop/hadoop-gryo.properties' > > No such property: './conf/hadoop/hadoop-gryo.properties' for class: > > groovy.lang.Binding > > Display stack trace? [yN] n > > > > gremlin> :remote connect tinkerpop.hadoop "./conf/spark-gryo.properties" > > No such property: "./conf/spark-gryo.properties" for class: > > groovy.lang.Binding > > Display stack trace? [yN] n > > > > I also tried a full path and created a bogus sparkp-gryo.properties file. > > > > Can you please provide an example of the remote command for this > situation > > ? > > > > > > 4) Can you please paste example contents of a spark-gryo.properties which > > is shown in the example ? > > Do I need both a hadoop-gryo.properties AND a spark-gryo.properties when > > doing remote commands ? > > > > There did not appear to be a sample in the source or build. > > The few sample files I found included a lot of Giraph stuff, which I > > am not using, and > > it isn't clear what is needed for Spark vs other GraphComputers. > > > > 5) If the Spark master has a URL like this: spark:// > test.machine.com:7077 , > > what is the appropriate value for spark.master in the > > hadoop-gryo.properties and/or spark-gryo.properties ? > > > > I don't understand what spark.master=local[4] is supposed to mean. > > > > 6) Can you please explain what this is doing in the examples: > > g.engine(computer) > > > > "computer" doesn't ever appear to be set in the examples > > and this command doesn't work in my environment. > > > > gremlin> g.engine(computer) > > No such property: computer for class: groovysh_evaluate > > Display stack trace? [yN] n > > > > What is "g.engine" doing that isn't already done via a config file ? > > > > 7) Is the GA release candidate shipping with MapReduceGraphComputer > undefined ? > > > > *COMING SOON* > > > > > > Thanks ! > >
