Hi Marko,

I'm actually trying to run all the unit tests. I'd like to start from a
point that passes all the unit tests so that after I make my changes I can
confirm that it didn't break anything.

I tried to see if compiling on my mac would circumvent the NFS issue, but
now I get this instead:

shouldSupportHDFSMethods(org.apache.tinkerpop.gremlin.hadoop.groovy.plugin.HadoopGremlinPluginCheck)
Time elapsed: 0.502 sec <<< ERROR!
java.lang.IllegalArgumentException: java.net.URISyntaxException: Relative
path in absolute URI: .bash_profile.macports-saved_2015-10-30_at_21:51:10
at java.net.URI.checkPath(URI.java:1823)
at java.net.URI.<init>(URI.java:745)
at org.apache.hadoop.fs.Path.initialize(Path.java:202)
at org.apache.hadoop.fs.Path.<init>(Path.java:171)
at org.apache.hadoop.fs.Path.<init>(Path.java:93)
at org.apache.hadoop.fs.Globber.glob(Globber.java:241)
at org.apache.hadoop.fs.FileSystem.globStatus(FileSystem.java:1655)
at
org.codehaus.groovy.vmplugin.v7.IndyInterface.selectMethod(IndyInterface.java:215)

at
org.apache.tinkerpop.gremlin.hadoop.groovy.plugin.HadoopLoader$_load_closure2.doCall(HadoopLoader.groovy:58)

at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)

at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

at java.lang.reflect.Method.invoke(Method.java:497)
at org.codehaus.groovy.reflection.CachedMethod.invoke(CachedMethod.java:90)
at
org.codehaus.groovy.runtime.metaclass.ClosureMetaMethod.invoke(ClosureMetaMethod.java:81)

at groovy.lang.MetaMethod.doMethodInvoke(MetaMethod.java:324)
at
org.codehaus.groovy.runtime.callsite.PojoMetaMethodSite.invoke(PojoMetaMethodSite.java:48)

at
org.codehaus.groovy.runtime.callsite.PojoMetaMethodSite.call(PojoMetaMethodSite.java:53)

at
org.codehaus.groovy.runtime.callsite.CallSiteArray.defaultCall(CallSiteArray.java:45)

at
org.codehaus.groovy.runtime.callsite.AbstractCallSite.call(AbstractCallSite.java:110)

at
org.codehaus.groovy.runtime.callsite.AbstractCallSite.call(AbstractCallSite.java:114)

at groovysh_evaluate.run(groovysh_evaluate:3)
at
org.codehaus.groovy.vmplugin.v7.IndyInterface.selectMethod(IndyInterface.java:215)

at
org.codehaus.groovy.tools.shell.Interpreter.evaluate(Interpreter.groovy:69)
at
org.codehaus.groovy.vmplugin.v7.IndyInterface.selectMethod(IndyInterface.java:215)

at org.codehaus.groovy.tools.shell.Groovysh.execute(Groovysh.groovy:185)
at
org.apache.tinkerpop.gremlin.groovy.util.TestableConsolePluginAcceptor.eval(TestableConsolePluginAcceptor.java:67)

at
org.apache.tinkerpop.gremlin.hadoop.groovy.plugin.HadoopGremlinPluginCheck.shouldSupportHDFSMethods(HadoopGremlinPluginCheck.java:108)

On Wed, Jan 6, 2016 at 6:00 AM Marko Rodriguez <okramma...@gmail.com> wrote:

> Hi,
>
> As a short term fix, are you wanting to run tests or just build the
> project? If the latter, then just do "mvn clean install
> -Dmaven.test.skip=true"
>
> HTH,
> Marko.
>
> http://markorodriguez.com
>
> On Jan 6, 2016, at 6:05 AM, Stephen Mallette <spmalle...@gmail.com> wrote:
>
> > I don't know what the problem could be here.  I'm not aware of others
> > having problems building master atm.
> >
> > On Mon, Jan 4, 2016 at 11:41 PM, Jonathan Ellithorpe <
> j...@cs.stanford.edu>
> > wrote:
> >
> >> Just tried that, but I'm still getting this "Output directory
> >> target/test-output/~traversers already exists" error above.
> >>
> >> I noticed the following in the logs:
> >>
> >> [INFO] Slf4jLogger$anonfun$receive$1$anonfun$applyOrElse$3 - Shutting
> down
> >> remote daemon.
> >> [INFO] Slf4jLogger$anonfun$receive$1$anonfun$applyOrElse$3 - Remote
> daemon
> >> shut down; proceeding with flushing remote transports.
> >> [INFO] Slf4jLogger$anonfun$receive$1$anonfun$applyOrElse$3 - Remoting
> shut
> >> down.
> >> [WARN] HadoopGremlinPlugin - Be sure to set the environmental variable:
> >> HADOOP_GREMLIN_LIBS
> >> [WARN] FileUtil - Failed to delete file or dir
> >>
> [/home/jdellit/tmp/incubator-tinkerpop/spark-gremlin/target/test-output/m]:
> >> it still exists.
> >> [WARN] FileUtil - Failed to delete file or dir
> >>
> >>
> [/home/jdellit/tmp/incubator-tinkerpop/spark-gremlin/target/test-output/~traversers/.nfs00000000066612b4000000c2]:
> >> it still exists.
> >> [WARN] FileUtil - Failed to delete file or dir
> >>
> >>
> [/home/jdellit/tmp/incubator-tinkerpop/spark-gremlin/target/test-output/~traversers/.nfs00000000066612b2000000c6]:
> >> it still exists.
> >> [WARN] FileUtil - Failed to delete file or dir
> >>
> >>
> [/home/jdellit/tmp/incubator-tinkerpop/spark-gremlin/target/test-output/~traversers/.nfs00000000066612b0000000bf]:
> >> it still exists.
> >> [WARN] FileUtil - Failed to delete file or dir
> >>
> >>
> [/home/jdellit/tmp/incubator-tinkerpop/spark-gremlin/target/test-output/~traversers/.nfs00000000066612b5000000c5]:
> >> it still exists.
> >> [WARN] FileUtil - Failed to delete file or dir
> >>
> >>
> [/home/jdellit/tmp/incubator-tinkerpop/spark-gremlin/target/test-output/~traversers/.nfs00000000066612b3000000c4]:
> >> it still exists.
> >> [WARN] FileUtil - Failed to delete file or dir
> >>
> >>
> [/home/jdellit/tmp/incubator-tinkerpop/spark-gremlin/target/test-output/~traversers/.nfs00000000066612b6000000c3]:
> >> it still exists.
> >> [WARN] FileUtil - Failed to delete file or dir
> >>
> >>
> [/home/jdellit/tmp/incubator-tinkerpop/spark-gremlin/target/test-output/~traversers/.nfs00000000066612b1000000c1]:
> >> it still exists.
> >> [WARN] FileUtil - Failed to delete file or dir
> >>
> >>
> [/home/jdellit/tmp/incubator-tinkerpop/spark-gremlin/target/test-output/~traversers/.nfs00000000066612af000000c0]:
> >> it still exists.
> >> [INFO] Logging$class - Running Spark version 1.5.1
> >> [INFO] Logging$class - Changing view acls to: jdellit
> >> [INFO] Logging$class - Changing modify acls to: jdellit
> >>
> >> The appearance of the .nfsXXX files leads me to suspect that files are
> >> being deleted by one thread that are still open on another thread. If
> that
> >> other thread attempts a write then an .nfsXXX file is generated.
> >>
> >> Are other folks running these tests on NFS?
> >>
> >> Jonathan
> >>
> >>
> >> On Mon, Jan 4, 2016 at 4:00 PM Stephen Mallette <spmalle...@gmail.com>
> >> wrote:
> >>
> >>> how about good ol' mvn clean install?  that works for me (and others)
> and
> >>> travis on master.
> >>>
> >>> i guess "mvn clean compile" and "mvn test" fails because gremlin-shaded
> >>> doesn't get built as the core of its work is bound to the "package"
> >> phase,
> >>> so you get all those errors. it seems that you have to minimally
> execute
> >>> the "package" phase of the maven life cycle
> >>>
> >>> On Mon, Jan 4, 2016 at 4:47 PM, Jonathan Ellithorpe <
> j...@cs.stanford.edu
> >>>
> >>> wrote:
> >>>
> >>>> Hi Stephen, thanks for the heads up. I remember getting stuck trying
> to
> >>>> find a point in the tree from which to base my changes. Either there
> >> was
> >>> an
> >>>> error compiling tinkerpop, or the unit tests did not pass. For example
> >> on
> >>>> tag 3.1.0-incubating I get the following error when executing mvn
> clean
> >>>> compile:
> >>>>
> >>>> [ERROR] COMPILATION ERROR :
> >>>> [INFO] -------------------------------------------------------------
> >>>> [ERROR]
> >>>>
> >>>>
> >>>
> >>
> /home/jdellit/tmp/incubator-tinkerpop/gremlin-core/src/main/java/org/apache/tinkerpop/gremlin/structure/io/graphson/GraphSONReader.java:[41,53]
> >>>> package org.apache.tinkerpop.shaded.jackson.core.type does not exist
> >>>> .
> >>>> .
> >>>> .
> >>>>
> >>>> "mvn clean package -DskipTests" does work.
> >>>>
> >>>> Then when I run "mvn test" I get:
> >>>>
> >>>> shouldGetVersion(org.apache.tinkerpop.gremlin.util.GremlinTest) Time
> >>>> elapsed: 0.016 sec <<< ERROR!
> >>>> java.lang.ExceptionInInitializerError: null
> >>>> at com.jcabi.manifests.Manifests.read(Manifests.java:274)
> >>>> at org.apache.tinkerpop.gremlin.util.Gremlin.<clinit>(Gremlin.java:32)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.tinkerpop.gremlin.util.GremlinTest.shouldGetVersion(GremlinTest.java:39)
> >>>>
> >>>> And several other errors while running tests in Gremlin Core.
> >> Strangely,
> >>>> however, running "mvn package" at this point does not produce those
> >>> errors,
> >>>> even though its running the same tests. It encounters a different
> error
> >>> for
> >>>> Spark Gremlin:
> >>>>
> >>>>
> >>>>
> >>>
> >>
> shouldGracefullyHandleBadGremlinHadoopLibs(org.apache.tinkerpop.gremlin.hadoop.groovy.plugin.HadoopGremlinPluginTest)
> >>>> Time elapsed: 3.318 sec <<< ERROR!
> >>>>
> >>>> org.apache.tinkerpop.gremlin.groovy.plugin.RemoteException:
> >>>> java.util.concurrent.ExecutionException:
> >>>> org.apache.hadoop.mapred.FileAlreadyExistsException: Output directory
> >>>> target/test-output/~traversers already exists
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
> >>>> at
> >>> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.tinkerpop.gremlin.hadoop.groovy.plugin.HadoopRemoteAcceptor.submit(HadoopRemoteAcceptor.java:99)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.tinkerpop.gremlin.hadoop.groovy.plugin.HadoopGremlinPluginTest.shouldGracefullyHandleBadGremlinHadoopLibs(HadoopGremlinPluginTest.java:169)
> >>>> Caused by: org.apache.hadoop.mapred.FileAlreadyExistsException: Output
> >>>> directory target/test-output/~traversers already exists
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.checkOutputSpecs(FileOutputFormat.java:146)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.spark.rdd.PairRDDFunctions$anonfun$saveAsNewAPIHadoopDataset$1.apply$mcV$sp(PairRDDFunctions.scala:1011)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.spark.rdd.PairRDDFunctions$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:998)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.spark.rdd.PairRDDFunctions$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:998)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:108)
> >>>> at org.apache.spark.rdd.RDD.withScope(RDD.scala:306)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopDataset(PairRDDFunctions.scala:998)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.spark.rdd.PairRDDFunctions$anonfun$saveAsNewAPIHadoopFile$2.apply$mcV$sp(PairRDDFunctions.scala:938)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.spark.rdd.PairRDDFunctions$anonfun$saveAsNewAPIHadoopFile$2.apply(PairRDDFunctions.scala:930)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.spark.rdd.PairRDDFunctions$anonfun$saveAsNewAPIHadoopFile$2.apply(PairRDDFunctions.scala:930)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:108)
> >>>> at org.apache.spark.rdd.RDD.withScope(RDD.scala:306)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopFile(PairRDDFunctions.scala:930)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.spark.api.java.JavaPairRDD.saveAsNewAPIHadoopFile(JavaPairRDD.scala:809)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.tinkerpop.gremlin.spark.process.computer.SparkExecutor.saveMapReduceRDD(SparkExecutor.java:208)
> >>>> at
> >>>>
> >>>>
> >>>
> >>
> org.apache.tinkerpop.gremlin.spark.process.computer.SparkGraphComputer.lambda$submit$21(SparkGraphComputer.java:211)
> >>>>
> >>>> I am confused as to why mvn package results in the success of certain
> >>> tests
> >>>> that fail for mvn package.
> >>>>
> >>>> Jonathan
> >>>>
> >>>> On Mon, Jan 4, 2016 at 7:08 AM Stephen Mallette <spmalle...@gmail.com
> >
> >>>> wrote:
> >>>>
> >>>>> Jonathan, just wanted to throw a heads up your way so that you're
> >> aware
> >>>> of
> >>>>> our expecting timing.  If all goes as planned, we will head into code
> >>>>> freeze for 3.1.1-incubating in about three weeks.  If you are still
> >>>>> planning on submitting PRs for:
> >>>>>
> >>>>> https://issues.apache.org/jira/browse/TINKERPOP3-997
> >>>>> https://issues.apache.org/jira/browse/TINKERPOP3-998
> >>>>>
> >>>>> we'd need to see it in that time frame.  I don't mean to apply
> >>> pressure,
> >>>> I
> >>>>> just don't want to miss the chance to get these fixes in for
> >>>>> 3.1.1-incubating.
> >>>>>
> >>>>>
> >>>>>
> >>>>> On Wed, Dec 9, 2015 at 1:24 AM, Jonathan Ellithorpe <
> >>> j...@cs.stanford.edu
> >>>>>
> >>>>> wrote:
> >>>>>
> >>>>>> Hi Stephen, working on that now, thanks for pinging me on this.
> >>>>>>
> >>>>>> On Tue, Dec 8, 2015 at 4:48 PM Stephen Mallette <
> >>> spmalle...@gmail.com>
> >>>>>> wrote:
> >>>>>>
> >>>>>>> Hi Jonathan, just wondering if you still plan to look at offering
> >>> PRs
> >>>>>> for:
> >>>>>>>
> >>>>>>> https://issues.apache.org/jira/browse/TINKERPOP-998
> >>>>>>> https://issues.apache.org/jira/browse/TINKERPOP-997
> >>>>>>>
> >>>>>>> I'll stay away from those, if you think you will be working on
> >>> them.
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>> On Mon, Nov 30, 2015 at 1:39 PM, Stephen Mallette <
> >>>>> spmalle...@gmail.com>
> >>>>>>> wrote:
> >>>>>>>
> >>>>>>>>> Using VertexProperyFeatures.FEATURE_{ADD, REMOVE}_PROPERTY
> >>>> perhaps
> >>>>>>>> would be more consistent with the logic used everywhere else...
> >>>>>>>>
> >>>>>>>> yeah - i'm +1 for this approach. it makes more sense given
> >>>> ADD/REMOVE
> >>>>>>>> already being the pattern for graph Element instances.
> >>>>>>>>
> >>>>>>>> On Mon, Nov 30, 2015 at 1:31 PM, Jonathan Ellithorpe <
> >>>>>>> j...@cs.stanford.edu>
> >>>>>>>> wrote:
> >>>>>>>>
> >>>>>>>>> I think it's either that or change FEATURE_META_PROPERTY to a
> >>>>>> symmetric
> >>>>>>>>> VertexFeatures.FEATURE_{ADD, REMOVE}_METAPROPERTY to pair with
> >>>>>>>>> VertexFeatures.FEATURE_{ADD, REMOVE}_PROPERTY.
> >>>>>>>>>
> >>>>>>>>> Using VertexProperyFeatures.FEATURE_{ADD, REMOVE}_PROPERTY
> >>> perhaps
> >>>>>> would
> >>>>>>>>> be
> >>>>>>>>> more consistent with the logic used everywhere else...
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> On Mon, Nov 30, 2015 at 6:30 AM Stephen Mallette <
> >>>>>> spmalle...@gmail.com>
> >>>>>>>>> wrote:
> >>>>>>>>>
> >>>>>>>>>> ugh - mess.  maybe we should just keep the add/remove
> >> symmetry
> >>>> and
> >>>>>>>>>> deprecate FEATURE_META_PROPERTY then.
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> On Sat, Nov 28, 2015 at 4:11 PM, Jonathan Ellithorpe <
> >>>>>>>>> j...@cs.stanford.edu>
> >>>>>>>>>> wrote:
> >>>>>>>>>>
> >>>>>>>>>>> 1) Yes, I can submit a PR for fixing the SIMPLE feature
> >>>>>> requirement
> >>>>>>>>> set.
> >>>>>>>>>>> 2) I also agree with deprecating
> >>>>>>>>>>> VertexPropertyFeatures.FEATURE_ADD_PROPERTY, but looking
> >> at
> >>>> the
> >>>>>>> code I
> >>>>>>>>>>> think I realized there is a slight complication here. That
> >>> is,
> >>>>>> what
> >>>>>>>>> to do
> >>>>>>>>>>> with VertexPropertyFeatures.FEATURE_REMOVE_PROPERTY. Does
> >>>>>>>>>>> VertexFeatures.FEATURE_META_PROPERTIES imply both ADD and
> >>>>> REMOVE,
> >>>>>> or
> >>>>>>>>> only
> >>>>>>>>>>> ADD? In the later case, we would need to leave
> >>>>>>>>>>> VertexPropertyFeatures.FEATURE_REMOVE_PROPERTIES.
> >>> Personally,
> >>>>>> seeing
> >>>>>>>>> as
> >>>>>>>>>> how
> >>>>>>>>>>> VertexFeatures, extending ElementFeatures, has a
> >>>>>>> FEATURE_ADD_PROPERTY
> >>>>>>>>> and
> >>>>>>>>>>> FEATURE_REMOVE_PROPERTY, that the FEATURE_META_PROPERTIES
> >> be
> >>>>>> changed
> >>>>>>>>> to
> >>>>>>>>>>> FEATURE_ADD_METAPROPERTY and FEATURE_REMOVE_METAPROPERTY.
> >>>>>>>>>>>
> >>>>>>>>>>> Jonathan
> >>>>>>>>>>>
> >>>>>>>>>>> On Fri, Nov 27, 2015 at 4:55 AM Stephen Mallette <
> >>>>>>>>> spmalle...@gmail.com>
> >>>>>>>>>>> wrote:
> >>>>>>>>>>>
> >>>>>>>>>>>> ...damn - hot key sent my post too soon - trying again:
> >>>>>>>>>>>>
> >>>>>>>>>>>> Hi Jonathan, thanks for bringing this up.  It would be
> >>> nice
> >>>> if
> >>>>>> we
> >>>>>>>>> could
> >>>>>>>>>>>> expand coverage of our test suite by simply improving
> >> the
> >>>> way
> >>>>> in
> >>>>>>>>> which
> >>>>>>>>>>>> features are applied.  I was about to suggest a big set
> >> of
> >>>>>> changes
> >>>>>>>>>> when I
> >>>>>>>>>>>> realized that FeatureRequirementSet.SIMPLE is just
> >> defined
> >>>>>> wrong.
> >>>>>>>>> It
> >>>>>>>>>>>> shouldn't have this entry:
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>
> >>>>
> >>>
> >>
> addFeatureRequirement.Factory.create(Graph.Features.VertexPropertyFeatures.FEATURE_ADD_PROPERTY,
> >>>>>>>>>>>> Graph.Features.VertexPropertyFeatures.class));
> >>>>>>>>>>>>
> >>>>>>>>>>>> it should just be:
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>
> >>>>
> >>>
> >>
> add(FeatureRequirement.Factory.create(Graph.Features.VertexFeatures.FEATURE_ADD_PROPERTY,
> >>>>>>>>>>>> Graph.Features.VertexFeatures.class));
> >>>>>>>>>>>>
> >>>>>>>>>>>> I've created an issue for that to track things:
> >>>>>>>>>>>>
> >>>>>>>>>>>> https://issues.apache.org/jira/browse/TINKERPOP3-997
> >>>>>>>>>>>>
> >>>>>>>>>>>> because it is a "breaking" change as it will open up
> >> tests
> >>>> and
> >>>>>>>>> possibly
> >>>>>>>>>>>> cause existing implementations to fail.  If you'd like
> >> to
> >>>>>> submit a
> >>>>>>>>> PR
> >>>>>>>>>> for
> >>>>>>>>>>>> this little fix, as you were the reporter for it and as
> >>>>> someone
> >>>>>>> who
> >>>>>>>>> can
> >>>>>>>>>>>> test it in a way that is currently failing for them,
> >> just
> >>>> let
> >>>>> me
> >>>>>>>>> know.
> >>>>>>>>>>>>
> >>>>>>>>>>>> As for the this issue:
> >>>>>>>>>>>>
> >> Graph.Features.VertexPropertyFeatures.FEATURE_ADD_PROPERTY
> >>>>>>>>>>>> <==>
> >>> Graph.Features.VertexFeatures.FEATURE_META_PROPERTIES -
> >>>>>> yeah
> >>>>>>> -
> >>>>>>>>> we
> >>>>>>>>>>> need
> >>>>>>>>>>>> to deprecate one of those as they are the same thing.
> >> Not
> >>>>> sure
> >>>>>> if
> >>>>>>>>>> anyone
> >>>>>>>>>>>> has any preferences on that.  in one sense,
> >>>>> FEATURE_ADD_PROPERTY
> >>>>>>> is
> >>>>>>>>>>> better
> >>>>>>>>>>>> because it matches the approach for Vertex/Edge.
> >>>>>>>>>>>
> >>>>>>>>>>> On the other hand, the
> >>>>>>>>>>>> documentation refers to this feature as
> >> "meta-properties".
> >>>> I
> >>>>>>> guess
> >>>>>>>>> i
> >>>>>>>>>>> would
> >>>>>>>>>>>> go with keeping FEATURE_META_PROPERTIES and deprecating
> >>>>>>>>>>>> FEATURE_ADD_PROPERTY.  I've created an issue as such:
> >>>>>>>>>>>>
> >>>>>>>>>>>> https://issues.apache.org/jira/browse/TINKERPOP3-998
> >>>>>>>>>>>
> >>>>>>>>>>> If no one has any objections in the next 72 hours (Monday,
> >>>>>> November
> >>>>>>>>> 30,
> >>>>>>>>>>>> 2015 at 7:45am) I'll assume lazy consensus and we can
> >> move
> >>>>>> forward
> >>>>>>>>> with
> >>>>>>>>>>>> this one.
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>> On Fri, Nov 27, 2015 at 7:35 AM, Stephen Mallette <
> >>>>>>>>>> spmalle...@gmail.com>
> >>>>>>>>>>>> wrote:
> >>>>>>>>>>>>
> >>>>>>>>>>>>> Hi Jonathan, thanks for bringing this up.  It would be
> >>>> nice
> >>>>> if
> >>>>>>> we
> >>>>>>>>>> could
> >>>>>>>>>>>>> expand coverage of our test suite by simply improving
> >>> the
> >>>>> way
> >>>>>> in
> >>>>>>>>>> which
> >>>>>>>>>>>>> features are applied.  I was about to suggest a big
> >> set
> >>> of
> >>>>>>> changes
> >>>>>>>>>>> when I
> >>>>>>>>>>>>> realized that FeatureRequirementSet.SIMPLE is just
> >>> defined
> >>>>>>>>> wrong.  It
> >>>>>>>>>>>>> shouldn't have
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>
> >>>>
> >>>
> >>
> addFeatureRequirement.Factory.create(Graph.Features.VertexPropertyFeatures.FEATURE_ADD_PROPERTY,
> >>>>>>>>>>>>> Graph.Features.VertexPropertyFeatures.class));
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> it should just be:
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> On Mon, Nov 23, 2015 at 7:39 PM, Jonathan Ellithorpe <
> >>>>>>>>>>>> j...@cs.stanford.edu>
> >>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>> Hello all,
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> I am currently working on an experimental
> >>> implementation
> >>>> of
> >>>>>>>>>> TinkerPop3
> >>>>>>>>>>>> on
> >>>>>>>>>>>>>> an in-memory key-value store called RAMCloud. In the
> >>>>> process
> >>>>>> of
> >>>>>>>>>>> running
> >>>>>>>>>>>>>> the
> >>>>>>>>>>>>>> unit tests I noticed that turning on support for
> >>>>> persistence
> >>>>>>> did
> >>>>>>>>> not
> >>>>>>>>>>>>>> trigger any new unit tests in GraphTests. Looking
> >> into
> >>>> the
> >>>>>>>>> matter, I
> >>>>>>>>>>>> found
> >>>>>>>>>>>>>> that the unit test that tests this,
> >>> shouldPersistOnClose,
> >>>>> was
> >>>>>>> not
> >>>>>>>>>>>>>> executing
> >>>>>>>>>>>>>> because meta properties support is included in its
> >>>> feature
> >>>>>>>>>>> requirements,
> >>>>>>>>>>>>>> but I do not have support for meta properties. Oddly,
> >>>>> though,
> >>>>>>>>> this
> >>>>>>>>>>>>>> features
> >>>>>>>>>>>>>> requirement seems to be superfluous, since the test
> >>> does
> >>>>> not
> >>>>>>>>> utilize
> >>>>>>>>>>>> meta
> >>>>>>>>>>>>>> properties.
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> An orthogonal issue seems to be that
> >>>>>>>>>>>>>>
> >>>> Graph.Features.VertexPropertyFeatures.FEATURE_ADD_PROPERTY
> >>>>>> <==>
> >>>>>>>>>>>>>> Graph.Features.VertexFeatures.FEATURE_META_PROPERTIES
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> Best,
> >>>>>>>>>>>>>> Jonathan
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>
> >>>>
> >>>
> >>
>
>

Reply via email to