On 10/20/15 11:26 AM, Haag, Jason wrote: > Wanted to touch base on this again. I just checked my cartridges and noticed > that I don't have a "HTML (and variants)" option. Is it called > something else or definitely called "HTML" (and variants)? > > Also, in conductor I navigated to "Linked Data" > "Sponger" and do not > see a HTML cartridge to configure. However, I do see a RDFa, RDFa (no > translation, and xHTML in the list. Were you referring to one of this > or am I somehow missing the HTML extractor cartridge? > > If I'm missing it is there a way to add it? > > > Finally, I just noticed while in the Sponger section I clicked on the > "Entity URIs" tab just to see what was there and received the > following error page: > > Error S0002 > > SQ200: No table DB.DBA.RDF_ENTITY_URI_CARTRIDGE_MODES > > My VOS instance is a clean install on Ubuntu. I'm running 07.20.3214. > The following packages are installed: > > conductor dav 1.00.8752 > fct dav 1.13.79 > iSPARQL dav 1.29.29 > rdb2rdf dav 1.00.0003 > rdf_mappers dav 1.34.74
Jason, You can use the commercial edition cartridges_dav.vad package against your VOS instance. That's the quickest route forward as the version bundled with VOS is out of date, and requires integration work to bring it back up to date. Kingsley > > >> ------------------------------ >> >> Message: 3 >> Date: Thu, 15 Oct 2015 12:45:14 -0400 >> From: Kingsley Idehen <kide...@openlinksw.com> >> Subject: Re: [Virtuoso-users] RDF Mapper Options in Conductor >> To: virtuoso-users@lists.sourceforge.net >> Message-ID: <561fd81a.7090...@openlinksw.com> >> Content-Type: text/plain; charset="windows-1252" >> >> On 10/15/15 12:09 PM, Haag, Jason wrote: >>> Hi All, >>> >>> Rather than wait for someone to tell me the right options based on all >>> of the different options that are available for a crawl job targeting >>> RDFa I will post which options I selected. Perhaps someone can tell me >>> what I'm missing or did wrong. I'm running Version: 07.20.3214. >>> >>> Here is what I entered under "Web Application Server > Content Imports": >>> >>> Target Description: ADL Verbs (RDFa/HTML) >>> Target URL: http://xapi.vocab.pub/datasets/adl/verbs/index.html >> http://xapi.vocab.pub/datasets/adl/verbs/ -- if you want everything. >> >> >>> Copy to Local DAV collection: DAV/home/dba/rdf_sink >> /DAV/home/dba/rdf_sink/ >> >>> Number of redirects to follow: 1 >>> Update interval: 10 >>> Checked the following: >>> X Run Sponger >>> X Store Metadata >>> >>> Cartridges Selected: >>> X RDFa >> Select HTML (and variants) -- but note that via the "Linked Data" menu's >> "Sponger" section you need to goto "Extractor Cartridges" section to >> select and configure the HTML with the following: >> >> add-html-meta=yes >> get-feeds=no >> preview-length=512 >> fallback-mode=no >> rdfa=yes >> reify_html5md=0 >> reify_rdfa=0 >> reify_jsonld=0 >> reify_all_grddl=0 >> reify_html=0 >> passthrough_mode=yes >> loose=yes >> reify_html_misc=no >> reify_turtle=no >> >> I know this seems awkward, but this is the best solution we could come >> up with due to the problems posed by text/html content-type overloading >> re. HTML+Microdata and RDFa etc.. >> >> >>> After I created the crawl job, I went to "import queues" and clicked "run" >>> >>> I received the following message: >>> >>> Results for xapi.vocab.pub >>> errors while retrieving target. Select "reset" to return initial state >>> Total URLs processed : 1 >>> Download finished >>> >>> I also checked "retrieved sites" and 0/1 were downloaded. >>> >>> Where do I find out the error that was encountered while retrieving >>> target? Thanks! >> Click on the "Edit" button aligned with your crawler job. >> >> Another bit of quirky UI to be fixed. >>> Also, I'm not sure if this is a bug, but I noticed when I specify the >>> Local DAV collection of DAV/home/dba/rdf_sink/ and then go to edit the >>> job, it changes to /DAV/DAV/home/dba/rdf_sink/ by adding a new >>> directory 'DAV' >> You initial value should have been: /DAV/home/dba/rdf_sink/. >> >> We are going to look into some of these quirks, in due course. >> >>> If I use /home/dba/rdf_sink/ as the the local path when creating the >>> crawl job it won't let me without adding /DAV/ path in front of it. So >>> it seems it is creating a sub directory /DAV/ under /DAV/ when it >>> shouldn't be. >> See my comment above. >> >> Kingsley > ------------------------------------------------------------------------------ > _______________________________________________ > Virtuoso-users mailing list > Virtuoso-users@lists.sourceforge.net > https://lists.sourceforge.net/lists/listinfo/virtuoso-users > -- Regards, Kingsley Idehen Founder & CEO OpenLink Software Company Web: http://www.openlinksw.com Personal Weblog 1: http://kidehen.blogspot.com Personal Weblog 2: http://www.openlinksw.com/blog/~kidehen Twitter Profile: https://twitter.com/kidehen Google+ Profile: https://plus.google.com/+KingsleyIdehen/about LinkedIn Profile: http://www.linkedin.com/in/kidehen Personal WebID: http://kingsley.idehen.net/dataspace/person/kidehen#this
smime.p7s
Description: S/MIME Cryptographic Signature
------------------------------------------------------------------------------
_______________________________________________ Virtuoso-users mailing list Virtuoso-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/virtuoso-users