> > can you rebase your branc to test the "JS intelligence" PR
Done. On Thu, Mar 8, 2018 at 7:35 PM, Stephen Mallette <spmalle...@gmail.com> wrote: > wow - works for me too. so, (1) why did this ever work on any system and > (2), kuppitz, can you rebase your branc to test the "JS intelligence" PR > > On Thu, Mar 8, 2018 at 5:13 PM, Robert Dale <robd...@gmail.com> wrote: > > > Committed. > > > > Robert Dale > > > > On Thu, Mar 8, 2018 at 3:13 PM, Florian Hockmann <f...@florian-hockmann.de > > > > wrote: > > > > > Wow, that actually worked. Now I'm finally able to generate the > > > documentation! Thanks a lot, Robert! > > > > > > It's still a bit strange to me that it worked for you guys even without > > > this patch, but anyway, at least it works now. > > > > > > Can you commit this now that we know that it fixes the problem? > > > > > > Am 08.03.2018 um 15:52 schrieb Robert Dale: > > > > Sorry, shouldn't need to rebuild both images. Just apply the patch > and > > > > docker/build.sh -d > > > > > > > > Robert Dale > > > > > > > > On Thu, Mar 8, 2018 at 9:51 AM, Robert Dale <robd...@gmail.com> > wrote: > > > > > > > >> Try applying this patch. Remove and rebuild both tinkerpop:base and > > > >> tinkerpop:hadoop-* docker images. Let me know how it goes. > > > >> > > > >> diff --git a/docs/src/recipes/olap-spark-yarn.asciidoc > > > >> b/docs/src/recipes/olap-spark-yarn.asciidoc > > > >> index 634adeb1ab..154382918e 100644 > > > >> --- a/docs/src/recipes/olap-spark-yarn.asciidoc > > > >> +++ b/docs/src/recipes/olap-spark-yarn.asciidoc > > > >> @@ -105,6 +105,7 @@ conf.setProperty('spark. > executor.extraClassPath', > > > >> "./$archive/*:$hadoopConfDir") > > > >> conf.setProperty('spark.driver.extraLibraryPath', > > > >> "$hadoop/lib/native:$hadoop/lib/native/Linux-amd64-64") > > > >> conf.setProperty('spark.executor.extraLibraryPath', > > > >> "$hadoop/lib/native:$hadoop/lib/native/Linux-amd64-64") > > > >> conf.setProperty('gremlin.spark.persistContext', 'true') > > > >> +hdfs.copyFromLocal('data/tinkerpop-modern.kryo', > > > 'tinkerpop-modern.kryo') > > > >> graph = GraphFactory.open(conf) > > > >> g = graph.traversal().withComputer(SparkGraphComputer) > > > >> g.V().group().by(values('name')).by(both().count()) > > > >> > > > >> > > > >> > > > >> > > > >> Robert Dale > > > >> > > > >> On Mon, Mar 5, 2018 at 5:11 PM, Florian Hockmann < > > > f...@florian-hockmann.de> > > > >> wrote: > > > >> > > > >>> You can save images as a tar file with docker save: > > > >>> https://docs.docker.com/engine/reference/commandline/save/ > > > >>> > > > >>> |docker save -o tinkerpop.tar tinkerpop| > > > >>> > > > >>> Since I also have the problem where the build fails at > > > >>> olap-spark-yarn.asciidoc, I just exported my images and uploaded > them > > > >>> here: > > > >>> > > > >>> https://uploadfiles.io/eew3d > > > >>> > > > >>> The images were created by running ./docker/build.sh -d on master > > after > > > >>> deleting all images I had locally. > > > >>> > > > >>> > > > >>> Am 05.03.2018 um 22:05 schrieb Stephen Mallette: > > > >>>> I'm up to try anything to solve this so, sure, happy to try to > > share. > > > >>> Sorry > > > >>>> though, but what's the "easy" way to share an image? > > > >>>> > > > >>>> On Fri, Mar 2, 2018 at 3:03 PM, Robert Dale <robd...@gmail.com> > > > wrote: > > > >>>> > > > >>>>> is it possible to share your images and see if it's an image > issue > > or > > > >>>>> a > > > >>>>> docker environment issue? > > > >>>>> > > > >>>>> Robert Dale > > > >>>>> > > > >>>>> On Wed, Feb 28, 2018 at 3:34 PM, Stephen Mallette < > > > >>> spmalle...@gmail.com> > > > >>>>> wrote: > > > >>>>> > > > >>>>>> I pushed the fix for master, so I don't have the console problem > > > >>> anymore. > > > >>>>>> So now, it's just the issue I guess I've always had..... > > > >>>>>> > > > >>>>>> On Wed, Feb 28, 2018 at 12:18 PM, Robert Dale < > robd...@gmail.com> > > > >>> wrote: > > > >>>>>>> That works for me. > > > >>>>>>> > > > >>>>>>> Robert Dale > > > >>>>>>> > > > >>>>>>> On Wed, Feb 28, 2018 at 11:23 AM, Florian Hockmann < > > > >>>>>> f...@florian-hockmann.de > > > >>>>>>> wrote: > > > >>>>>>> > > > >>>>>>>> I also ran into the exact same problem in my feature branch > for > > > the > > > >>>>>>>> docker images which is why haven't added my own vote for the > PR > > > >>>>>>>> yet. > > > >>>>>> But > > > >>>>>>>> it's good to see that it's really completely unrelated to my > > > >>> changes. > > > >>>>>>>> > > > >>>>>>>> Am 28.02.2018 um 17:15 schrieb Stephen Mallette: > > > >>>>>>>>> I'm still a bust - same kind of error I keep having - > > > >>>>>>>>> > > > >>>>>>>>> * source: /usr/src/tinkerpop/docs/src/ > > > recipes/olap-spark-yarn. > > > >>>>>>>> asciidoc > > > >>>>>>>>> target: > > > >>>>>>>>> /usr/src/tinkerpop/target/postprocess-asciidoc/recipes/ > > > >>>>>>>> olap-spark-yarn.asciidoc > > > >>>>>>>>> progress: > > > >>>>>>>>> [=========================================================== > > > >>>>> ======> > > > >>>>>>>>> ] 65%java.io.IOException: No input > > > paths > > > >>>>>>>>> specified in job > > > >>>>>>>>> Type ':help' or ':h' for help. > > > >>>>>>>>> Display stack trace? [yN]pb(94); '----' > > > >>>>>>>>> > > > >>>>>>>>> > > > >>>>>>>>> Last 10 lines of > > > >>>>>>>>> /usr/src/tinkerpop/target/postprocess-asciidoc/recipes/ > > > >>>>>>>> olap-spark-yarn.asciidoc: > > > >>>>>>>>> gremlin> conf.setProperty('spark.executor.extraLibraryPath', > > > >>>>>>>>> "$hadoop/lib/native:$hadoop/lib/native/Linux-amd64-64") > > > >>>>>>>>> ==>null > > > >>>>>>>>> gremlin> conf.setProperty('gremlin.spark.persistContext', > > > 'true') > > > >>>>>>>>> ==>null > > > >>>>>>>>> gremlin> graph = GraphFactory.open(conf) > > > >>>>>>>>> ==>hadoopgraph[gryoinputformat->gryooutputformat] > > > >>>>>>>>> gremlin> g = graph.traversal().withComputer( > > SparkGraphComputer) > > > >>>>>>>>> ==>graphtraversalsource[hadoopgraph[gryoinputformat-> > > > >>>>>>> gryooutputformat], > > > >>>>>>>>> sparkgraphcomputer] > > > >>>>>>>>> gremlin> g.V().group().by(values('name')).by(both().count()) > > > >>>>>>>>> gremlin> > > > >>>>>>>>> > > > >>>>>>>>> xargs: /usr/src/tinkerpop/docs/preprocessor/preprocess-file. > > sh: > > > >>>>>> exited > > > >>>>>>>> with > > > >>>>>>>>> status 255; aborting > > > >>>>>>>>> > > > >>>>>>>>> > > > >>>>>>>>> > > > >>>>>>>>> > > > >>>>>>>>> On Wed, Feb 28, 2018 at 10:46 AM, Robert Dale < > > robd...@gmail.com > > > > > > > >>>>>>> wrote: > > > >>>>>>>>>> Yup, it's a step in the release docs. Once updated, master > > > >>>>>>>>>> builds > > > >>>>>>> docs > > > >>>>>>>> for > > > >>>>>>>>>> me. > > > >>>>>>>>>> > > > >>>>>>>>>> Robert Dale > > > >>>>>>>>>> > > > >>>>>>>>>> On Wed, Feb 28, 2018 at 10:42 AM, Stephen Mallette < > > > >>>>>>>> spmalle...@gmail.com> > > > >>>>>>>>>> wrote: > > > >>>>>>>>>> > > > >>>>>>>>>>> ah - i forgot to do that step.....i'm running tp32 now to > see > > > if > > > >>>>>> that > > > >>>>>>>>>>> works, but i'll fix that issue on master. > > > >>>>>>>>>>> > > > >>>>>>>>>>> On Wed, Feb 28, 2018 at 10:41 AM, Robert Dale < > > > robd...@gmail.com > > > >>>>>>>> wrote: > > > >>>>>>>>>>>> Could it be that 'bin/gremlin.sh' is linked to a specific > > > >>>>> version? > > > >>>>>>>> Does > > > >>>>>>>>>>>> this have to be updated every release? > > > >>>>>>>>>>>> > > > >>>>>>>>>>>> $ ll gremlin-console/bin/gremlin.sh > > > >>>>>>>>>>>> gremlin-console/bin/gremlin.sh -> > > > >>>>>>>>>>>> ../target/apache-tinkerpop-gremlin-console-3.3.2- > > > >>>>>>>>>>> SNAPSHOT-standalone/bin/ > > > >>>>>>>>>>>> gremlin.sh > > > >>>>>>>>>>>> > > > >>>>>>>>>>>> Robert Dale > > > >>>>>>>>>>>> > > > >>>>>>>>>>>> On Wed, Feb 28, 2018 at 10:37 AM, Robert Dale < > > > >>>>> robd...@gmail.com> > > > >>>>>>>>>> wrote: > > > >>>>>>>>>>>>> I got it too. tp32: good. tp33: good. master: bad. > > > >>>>>>>>>>>>> > > > >>>>>>>>>>>>> Robert Dale > > > >>>>>>>>>>>>> > > > >>>>>>>>>>>>> On Wed, Feb 28, 2018 at 10:20 AM, Stephen Mallette < > > > >>>>>>>>>>> spmalle...@gmail.com > > > >>>>>>>>>>>>> wrote: > > > >>>>>>>>>>>>> > > > >>>>>>>>>>>>>> I'm having no success generating docs with Docker. On > > master > > > >>>>> i'm > > > >>>>>>>>>>>> currently > > > >>>>>>>>>>>>>> getting: > > > >>>>>>>>>>>>>> > > > >>>>>>>>>>>>>> Starting namenodes on [localhost] > > > >>>>>>>>>>>>>> localhost: Warning: Permanently added 'localhost' > (ECDSA) > > to > > > >>>>> the > > > >>>>>>>>>> list > > > >>>>>>>>>>> of > > > >>>>>>>>>>>>>> known hosts. > > > >>>>>>>>>>>>>> localhost: starting namenode, logging to > > > >>>>>>>>>>>>>> /usr/local/lib/hadoop-2.7.2/logs/hadoop-root-namenode- > > > >>>>>>>>>>> 61cfd8f63c77.out > > > >>>>>>>>>>>>>> localhost: Warning: Permanently added 'localhost' > (ECDSA) > > to > > > >>>>> the > > > >>>>>>>>>> list > > > >>>>>>>>>>> of > > > >>>>>>>>>>>>>> known hosts. > > > >>>>>>>>>>>>>> localhost: starting datanode, logging to > > > >>>>>>>>>>>>>> /usr/local/lib/hadoop-2.7.2/logs/hadoop-root-datanode- > > > >>>>>>>>>>> 61cfd8f63c77.out > > > >>>>>>>>>>>>>> Starting secondary namenodes [0.0.0.0] > > > >>>>>>>>>>>>>> 0.0.0.0: Warning: Permanently added '0.0.0.0' (ECDSA) > to > > > the > > > >>>>>> list > > > >>>>>>>>>> of > > > >>>>>>>>>>>>>> known > > > >>>>>>>>>>>>>> hosts. > > > >>>>>>>>>>>>>> 0.0.0.0: starting secondarynamenode, logging to > > > >>>>>>>>>>>>>> /usr/local/lib/hadoop-2.7.2/logs/hadoop-root- > > > secondarynameno > > > >>>>>>>>>>>>>> de-61cfd8f63c77.out > > > >>>>>>>>>>>>>> starting yarn daemons > > > >>>>>>>>>>>>>> starting resourcemanager, logging to > > > >>>>>>>>>>>>>> /usr/local/lib/hadoop-2.7.2/logs/yarn-root- > > resourcemanager- > > > >>>>>>>>>>>>>> 61cfd8f63c77.out > > > >>>>>>>>>>>>>> localhost: Warning: Permanently added 'localhost' > (ECDSA) > > to > > > >>>>> the > > > >>>>>>>>>> list > > > >>>>>>>>>>> of > > > >>>>>>>>>>>>>> known hosts. > > > >>>>>>>>>>>>>> localhost: starting nodemanager, logging to > > > >>>>>>>>>>>>>> /usr/local/lib/hadoop-2.7.2/logs/yarn-root-nodemanager- > > > >>>>>>>>>>> 61cfd8f63c77.out > > > >>>>>>>>>>>>>> Gremlin REPL is not available. Cannot preprocess > AsciiDoc > > > >>>>> files. > > > >>>>>>>>>>>>>> Untagged: tinkerpop:build-1519830232 > > > >>>>>>>>>>>>>> > > > >>>>>>>>>>>>>> On that particular run, I'd just deleted all my Docker > > > images > > > >>>>>>>>>>> (including > > > >>>>>>>>>>>>>> the linux image) and it still failed. Any clues as to > what > > > >>>>> might > > > >>>>>>> be > > > >>>>>>>>>>>> wrong? > > > >>>>>>>>>>>>>> am i the only person with a problem here? > > > >>>>>>>>>>>>>> > > > >>> > > > > > > > > > > > >