Wanted to touch base on this again. I just checked my cartridges and noticed
that I don't have a "HTML (and variants)" option. Is it called
something else or definitely called "HTML" (and variants)?

Also, in conductor I navigated to "Linked Data" > "Sponger" and do not
see a HTML cartridge to configure. However, I do see a RDFa, RDFa (no
translation, and xHTML in the list. Were you referring to one of this
or am I somehow missing the HTML extractor cartridge?

If I'm missing it is there a way to add it?


Finally, I just noticed while in the Sponger section I clicked on the
"Entity URIs" tab just to see what was there and received the
following error page:

Error S0002

SQ200: No table DB.DBA.RDF_ENTITY_URI_CARTRIDGE_MODES

My VOS instance is a clean install on Ubuntu. I'm running  07.20.3214.
The following packages are installed:

 conductor dav 1.00.8752
 fct dav 1.13.79
 iSPARQL dav 1.29.29
 rdb2rdf dav 1.00.0003
 rdf_mappers dav 1.34.74


> ------------------------------
>
> Message: 3
> Date: Thu, 15 Oct 2015 12:45:14 -0400
> From: Kingsley Idehen <kide...@openlinksw.com>
> Subject: Re: [Virtuoso-users] RDF Mapper Options in Conductor
> To: virtuoso-users@lists.sourceforge.net
> Message-ID: <561fd81a.7090...@openlinksw.com>
> Content-Type: text/plain; charset="windows-1252"
>
> On 10/15/15 12:09 PM, Haag, Jason wrote:
>> Hi All,
>>
>> Rather than wait for someone to tell me the right options based on all
>> of the different options that are available for a crawl job targeting
>> RDFa I will post which options I selected. Perhaps someone can tell me
>> what I'm missing or did wrong. I'm running Version: 07.20.3214.
>>
>> Here is what I entered under "Web Application Server > Content Imports":
>>
>> Target Description: ADL Verbs (RDFa/HTML)
>> Target URL: http://xapi.vocab.pub/datasets/adl/verbs/index.html
>
> http://xapi.vocab.pub/datasets/adl/verbs/ -- if you want everything.
>
>
>> Copy to Local DAV collection: DAV/home/dba/rdf_sink
> /DAV/home/dba/rdf_sink/
>
>> Number of redirects to follow: 1
>> Update interval: 10
>> Checked the following:
>> X Run Sponger
>> X Store Metadata
>>
>> Cartridges Selected:
>> X RDFa
>
> Select HTML (and variants) -- but note that via the "Linked Data" menu's
> "Sponger"  section you need to goto "Extractor Cartridges" section to
> select and configure the HTML with the following:
>
> add-html-meta=yes
> get-feeds=no
> preview-length=512
> fallback-mode=no
> rdfa=yes
> reify_html5md=0
> reify_rdfa=0
> reify_jsonld=0
> reify_all_grddl=0
> reify_html=0
> passthrough_mode=yes
> loose=yes
> reify_html_misc=no
> reify_turtle=no
>
> I know this seems awkward, but this is the best solution we could come
> up with due to the problems posed by text/html content-type overloading
> re. HTML+Microdata and RDFa etc..
>
>
>>
>> After I created the crawl job, I went to "import queues" and clicked "run"
>>
>> I received the following message:
>>
>> Results for xapi.vocab.pub
>> errors while retrieving target. Select "reset" to return initial state
>> Total URLs processed : 1
>> Download finished
>>
>> I also checked "retrieved sites" and 0/1 were downloaded.
>>
>> Where do I find out the error that was encountered while retrieving
>> target? Thanks!
>
> Click on the "Edit" button aligned with your crawler job.
>
> Another bit of quirky UI to be fixed.
>>
>> Also, I'm not sure if this is a bug, but I noticed when I specify the
>> Local DAV collection of DAV/home/dba/rdf_sink/ and then go to edit the
>> job, it changes to /DAV/DAV/home/dba/rdf_sink/ by adding a new
>> directory 'DAV'
>
> You initial value should have been: /DAV/home/dba/rdf_sink/.
>
> We are going to look into some of these quirks, in due course.
>
>>
>> If I use /home/dba/rdf_sink/ as the the local path when creating the
>> crawl job it won't let me without adding /DAV/ path in front of it. So
>> it seems it is creating a sub directory /DAV/ under /DAV/ when it
>> shouldn't be.
>
> See my comment above.
>
> Kingsley

------------------------------------------------------------------------------
_______________________________________________
Virtuoso-users mailing list
Virtuoso-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/virtuoso-users

Reply via email to