On 10/20/15 11:26 AM, Haag, Jason wrote:
> Wanted to touch base on this again. I just checked my cartridges and noticed
> that I don't have a "HTML (and variants)" option. Is it called
> something else or definitely called "HTML" (and variants)?
>
> Also, in conductor I navigated to "Linked Data" > "Sponger" and do not
> see a HTML cartridge to configure. However, I do see a RDFa, RDFa (no
> translation, and xHTML in the list. Were you referring to one of this
> or am I somehow missing the HTML extractor cartridge?
>
> If I'm missing it is there a way to add it?
>
>
> Finally, I just noticed while in the Sponger section I clicked on the
> "Entity URIs" tab just to see what was there and received the
> following error page:
>
> Error S0002
>
> SQ200: No table DB.DBA.RDF_ENTITY_URI_CARTRIDGE_MODES
>
> My VOS instance is a clean install on Ubuntu. I'm running  07.20.3214.
> The following packages are installed:
>
>  conductor dav 1.00.8752
>  fct dav 1.13.79
>  iSPARQL dav 1.29.29
>  rdb2rdf dav 1.00.0003
>  rdf_mappers dav 1.34.74

Jason,

You can use the commercial edition cartridges_dav.vad package against
your VOS instance. That's the quickest route forward as the version
bundled with VOS is out of date, and requires integration work to bring
it back up to date.


Kingsley

>
>
>> ------------------------------
>>
>> Message: 3
>> Date: Thu, 15 Oct 2015 12:45:14 -0400
>> From: Kingsley Idehen <kide...@openlinksw.com>
>> Subject: Re: [Virtuoso-users] RDF Mapper Options in Conductor
>> To: virtuoso-users@lists.sourceforge.net
>> Message-ID: <561fd81a.7090...@openlinksw.com>
>> Content-Type: text/plain; charset="windows-1252"
>>
>> On 10/15/15 12:09 PM, Haag, Jason wrote:
>>> Hi All,
>>>
>>> Rather than wait for someone to tell me the right options based on all
>>> of the different options that are available for a crawl job targeting
>>> RDFa I will post which options I selected. Perhaps someone can tell me
>>> what I'm missing or did wrong. I'm running Version: 07.20.3214.
>>>
>>> Here is what I entered under "Web Application Server > Content Imports":
>>>
>>> Target Description: ADL Verbs (RDFa/HTML)
>>> Target URL: http://xapi.vocab.pub/datasets/adl/verbs/index.html
>> http://xapi.vocab.pub/datasets/adl/verbs/ -- if you want everything.
>>
>>
>>> Copy to Local DAV collection: DAV/home/dba/rdf_sink
>> /DAV/home/dba/rdf_sink/
>>
>>> Number of redirects to follow: 1
>>> Update interval: 10
>>> Checked the following:
>>> X Run Sponger
>>> X Store Metadata
>>>
>>> Cartridges Selected:
>>> X RDFa
>> Select HTML (and variants) -- but note that via the "Linked Data" menu's
>> "Sponger"  section you need to goto "Extractor Cartridges" section to
>> select and configure the HTML with the following:
>>
>> add-html-meta=yes
>> get-feeds=no
>> preview-length=512
>> fallback-mode=no
>> rdfa=yes
>> reify_html5md=0
>> reify_rdfa=0
>> reify_jsonld=0
>> reify_all_grddl=0
>> reify_html=0
>> passthrough_mode=yes
>> loose=yes
>> reify_html_misc=no
>> reify_turtle=no
>>
>> I know this seems awkward, but this is the best solution we could come
>> up with due to the problems posed by text/html content-type overloading
>> re. HTML+Microdata and RDFa etc..
>>
>>
>>> After I created the crawl job, I went to "import queues" and clicked "run"
>>>
>>> I received the following message:
>>>
>>> Results for xapi.vocab.pub
>>> errors while retrieving target. Select "reset" to return initial state
>>> Total URLs processed : 1
>>> Download finished
>>>
>>> I also checked "retrieved sites" and 0/1 were downloaded.
>>>
>>> Where do I find out the error that was encountered while retrieving
>>> target? Thanks!
>> Click on the "Edit" button aligned with your crawler job.
>>
>> Another bit of quirky UI to be fixed.
>>> Also, I'm not sure if this is a bug, but I noticed when I specify the
>>> Local DAV collection of DAV/home/dba/rdf_sink/ and then go to edit the
>>> job, it changes to /DAV/DAV/home/dba/rdf_sink/ by adding a new
>>> directory 'DAV'
>> You initial value should have been: /DAV/home/dba/rdf_sink/.
>>
>> We are going to look into some of these quirks, in due course.
>>
>>> If I use /home/dba/rdf_sink/ as the the local path when creating the
>>> crawl job it won't let me without adding /DAV/ path in front of it. So
>>> it seems it is creating a sub directory /DAV/ under /DAV/ when it
>>> shouldn't be.
>> See my comment above.
>>
>> Kingsley
> ------------------------------------------------------------------------------
> _______________________________________________
> Virtuoso-users mailing list
> Virtuoso-users@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/virtuoso-users
>


-- 
Regards,

Kingsley Idehen       
Founder & CEO 
OpenLink Software     
Company Web: http://www.openlinksw.com
Personal Weblog 1: http://kidehen.blogspot.com
Personal Weblog 2: http://www.openlinksw.com/blog/~kidehen
Twitter Profile: https://twitter.com/kidehen
Google+ Profile: https://plus.google.com/+KingsleyIdehen/about
LinkedIn Profile: http://www.linkedin.com/in/kidehen
Personal WebID: http://kingsley.idehen.net/dataspace/person/kidehen#this


Attachment: smime.p7s
Description: S/MIME Cryptographic Signature

------------------------------------------------------------------------------
_______________________________________________
Virtuoso-users mailing list
Virtuoso-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/virtuoso-users

Reply via email to