Hi Jason, Yes, that would be the page to use for setting up schedule import jobs for RDF or other structured data sources, can also use http://virtuoso.openlinksw.com/dataspace/doc/dav/wiki/Main/VirtSetCrawlerJobsGuideSemanticSitemaps <http://virtuoso.openlinksw.com/dataspace/doc/dav/wiki/Main/VirtSetCrawlerJobsGuideSemanticSitemaps>
Both need updating to reflect current UI which I have requested be done … Best Regards Hugh Williams Professional Services OpenLink Software, Inc. // http://www.openlinksw.com/ Weblog -- http://www.openlinksw.com/blogs/ LinkedIn -- http://www.linkedin.com/company/openlink-software/ Twitter -- http://twitter.com/OpenLink Google+ -- http://plus.google.com/100570109519069333827/ Facebook -- http://www.facebook.com/OpenLinkSoftware Universal Data Access, Integration, and Management Technology Providers > On 28 Sep 2015, at 22:20, Haag, Jason <jhaa...@gmail.com> wrote: > > What would the steps/instructions be to set up an automatic import for > 7.2.1? The instructions and screens here don't match the new interface > and field options: > http://docs.openlinksw.com/virtuoso/rdfinsertmethods.html#rdfinsertmethodvirtuosocrawler > > For example, there is no longer a field for "Local WebDAV Identifier" > which was previously required. > ------------------------------------------------------- > +1.850.266.7100(office) > +1.850.471.1300 (mobile) > jhaag75 (skype) > http://jasonhaag.com (Web) > http://twitter.com/mobilejson (Twitter) > http://linkedin.com/in/jasonhaag (LinkedIn) > > > > On Sat, Sep 26, 2015 at 5:39 PM, Paul Houle <ontolo...@gmail.com> wrote: >> I like the cloud solution of creating a new virtuoso system, doing the >> load, having plenty of time to test it, then replacing the production >> instance with the new instance and retiring the production instance. >> >> The main advantage here is that there is no way a screw-up in the load >> procedure can trash the production system -- even if Virtuoso was entirely >> reliable, as the data sources grow the rate of exceptional events (say you >> fill the disk) goes up. The temporary server approach eliminates a lot of >> headaches and it is good cloud economics. (if you run a server at AMZN for >> 1 hour a day to update, the cost of your system only goes up by %4). >> >> I was having good luck with this approach until Virtuoso 7.2.0 came along >> and since then I've had problems similar in severity to what the N.I.H. was >> reporting, it really looked like massive corruption of the data structures, >> 7.2.1 did not help. >> >> I don't know if these issues are fixed in the current TRUNK but if they are >> it would be nice to get an official release. >> >> On Fri, Sep 25, 2015 at 1:31 PM, Haag, Jason <jhaa...@gmail.com> wrote: >>> >>> >>> Hi Users, >>> >>> I'm trying to determine the best option for my situation for importing RDF >>> data into Virtuoso. Here's my situation: >>> >>> I currently have several RDF datasets available on my server. Each data >>> set has an RDF dump available as RDF/XML, JSON-LD, and Turtle. These dumps >>> are generated automatically without virtuoso from an HTML page marked up >>> using RDFa. >>> >>> What is the best option for automating the import of this data on a >>> regular basis into the virtuoso DB? The datasets may grow so it should not >>> just import the data once, but import on a regular basis, perhaps daily or >>> weekly. >>> >>> Based on what I've read in the documentation, this crawler option seems >>> like the most appropriate option for my situation: >>> http://virtuoso.openlinksw.com/dataspace/doc/dav/wiki/Main/VirtSetCrawlerJobsGuideDirectories >>> >>> Can anyone verify if this would be the best approach? Does anyone know if >>> the crawler supports RDFa/HTML or should it point to a specific directory >>> with only the RDF dump files? >>> >>> Thanks in advance! >>> >>> J Haag >>> >>> >>> ------------------------------------------------------------------------------ >>> >>> _______________________________________________ >>> Virtuoso-users mailing list >>> Virtuoso-users@lists.sourceforge.net >>> https://lists.sourceforge.net/lists/listinfo/virtuoso-users >>> >> >> >> >> -- >> Paul Houle >> >> Applying Schemas for Natural Language Processing, Distributed Systems, >> Classification and Text Mining and Data Lakes >> >> (607) 539 6254 paul.houle on Skype ontolo...@gmail.com >> >> :BaseKB -- Query Freebase Data With SPARQL >> http://basekb.com/gold/ >> >> Legal Entity Identifier Lookup >> https://legalentityidentifier.info/lei/lookup/ >> >> Join our Data Lakes group on LinkedIn >> https://www.linkedin.com/grp/home?gid=8267275 >> > > ------------------------------------------------------------------------------ > _______________________________________________ > Virtuoso-users mailing list > Virtuoso-users@lists.sourceforge.net > https://lists.sourceforge.net/lists/listinfo/virtuoso-users
------------------------------------------------------------------------------
_______________________________________________ Virtuoso-users mailing list Virtuoso-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/virtuoso-users