Re: [CODE4LIB] CODE4LIB Digest - 12 Sep 2013 to 13 Sep 2013 (#2013-237)

2013-09-16 Thread aj...@virginia.edu
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

I'd suggest that perhaps the confusion arises because This instance is (not) 
'valid' according to that ontology. might be inferred from an instance and an 
ontology (under certain conditions), and that's the soul of what we're asking 
when we define constraints on the data. Perhaps OWL can be used to express 
conditions of validity, as long as we represent the quality valid for use in 
inferences.

- ---
A. Soroka
The University of Virginia Library

On Sep 13, 2013, at 11:00 PM, CODE4LIB automatic digest system wrote:

 Also, remember that OWL does NOT constrain your data, it constrains only the 
 inferences that you can make about your data. OWL operates at the ontology 
 level, not the data level. (The OWL 2 documentation makes this more clear, in 
 my reading of it. I agree that the example you cite sure looks like a 
 constraint on the data... it's very confusing.)

-BEGIN PGP SIGNATURE-
Version: GnuPG/MacGPG2 v2.0.19 (Darwin)
Comment: GPGTools - http://gpgtools.org

iQEcBAEBAgAGBQJSNwe2AAoJEATpPYSyaoIkwLcIAK+sMzy1XkqLStg94F2I40pe
0DepjqVhdPnaDS1Msg7pd7c7iC0L5NhCWd9BxzdvRgeMRr123zZ3EmKDSy8XZiGf
uQyXlA9cOqpCxdQLj2zXv5VHrOdlsA1UAGprwhYrxOz/v3xQ7b2nXusRoZRfDlts
iadvWx5DhLEb2+uVl9geteeymLIVUTzm8WnUITEE7by2HAQf9VlT9CrQSVQ21wLC
hvmk47Nt8WIGyPwRh1qOhvIXLDLxD9rkBSC1G01RhzwvctDy88Tmt2Ut47ZREScP
YUz/bf/qxITzX2L7tE35s2w+RUIFIFc4nJa3Xhp0wMoTAz5UYMiWIcXZ38qfGlY=
=PJTS
-END PGP SIGNATURE-


Re: [CODE4LIB] CODE4LIB Digest - 12 Sep 2013 to 13 Sep 2013 (#2013-237)

2013-09-16 Thread Karen Coyle

On 9/16/13 6:29 AM, aj...@virginia.edu wrote:

-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

I'd suggest that perhaps the confusion arises because This instance is (not) 'valid' 
according to that ontology. might be inferred from an instance and an ontology (under certain 
conditions), and that's the soul of what we're asking when we define constraints on the data. 
Perhaps OWL can be used to express conditions of validity, as long as we represent the quality 
valid for use in inferences.


Based on the results of the RDF Validation workshop [1], validation is 
being expressed today as SPARQL rules. If you express the rules in OWL 
then unfortunately you affect downstream re-use of your ontology, and 
that can create a mess for inferencing and can add a burden onto any 
reasoners, which are supposed to apply the OWL declarations.


One participant at the workshop demonstrated a system that used the OWL 
constraints as constraints, but only in a closed system. I think that 
the use of SPARQL is superior because it does not affect the semantics 
of the classes and properties, only the instance data, and that means 
that the same properties can be validated differently for different 
applications or under different contexts. As an example, one community 
may wish to say that their metadata can have one and only one dc:title, 
while others may allow more than one. You do not want to constrain 
dc:title throughout the Web, only your own use of it. (Tom Baker and I 
presented a solution to this on the second day as Application Profiles 
[2], as defined by the DC community).


kc
[1] https://www.w3.org/2012/12/rdf-val/agenda
[2] 
http://www.w3.org/2001/sw/wiki/images/e/ef/Baker-dc-abstract-model-revised.pdf




- ---
A. Soroka
The University of Virginia Library

On Sep 13, 2013, at 11:00 PM, CODE4LIB automatic digest system wrote:


Also, remember that OWL does NOT constrain your data, it constrains only the 
inferences that you can make about your data. OWL operates at the ontology 
level, not the data level. (The OWL 2 documentation makes this more clear, in 
my reading of it. I agree that the example you cite sure looks like a 
constraint on the data... it's very confusing.)

-BEGIN PGP SIGNATURE-
Version: GnuPG/MacGPG2 v2.0.19 (Darwin)
Comment: GPGTools - http://gpgtools.org

iQEcBAEBAgAGBQJSNwe2AAoJEATpPYSyaoIkwLcIAK+sMzy1XkqLStg94F2I40pe
0DepjqVhdPnaDS1Msg7pd7c7iC0L5NhCWd9BxzdvRgeMRr123zZ3EmKDSy8XZiGf
uQyXlA9cOqpCxdQLj2zXv5VHrOdlsA1UAGprwhYrxOz/v3xQ7b2nXusRoZRfDlts
iadvWx5DhLEb2+uVl9geteeymLIVUTzm8WnUITEE7by2HAQf9VlT9CrQSVQ21wLC
hvmk47Nt8WIGyPwRh1qOhvIXLDLxD9rkBSC1G01RhzwvctDy88Tmt2Ut47ZREScP
YUz/bf/qxITzX2L7tE35s2w+RUIFIFc4nJa3Xhp0wMoTAz5UYMiWIcXZ38qfGlY=
=PJTS
-END PGP SIGNATURE-


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] CODE4LIB Digest - 12 Sep 2013 to 13 Sep 2013 (#2013-237)

2013-09-16 Thread Ethan Gruber
Using SPARQL to validate seems like tremendous overhead.  From the Gerber
abstract: A total of 55 rules have been defined representing the
constraints and requirements of the OA Specification and Ontology. For each
rule we have defined a SPARQL query to check compliance. I hope this isn't
55 SPARQL queries per RDF resource.

Europeana's review of schematron indicated what I pointed out earlier, that
it confines one to using RDF/XML, which is sub-optimal in their own
words.  One could accept RDF in any serialization and then run it through
an RDF processor, like rapper (http://librdf.org/raptor/rapper.html), into
XML and then validate.  Eventually, when XPath/XSLT 3 supports JSON and
other non-XML data models, theoretically, schematron might then be able to
validate other serializations of RDF.  Ditto for XForms, which we are using
to validate RDF/XML.  Obviously, this is sub-optimal because our workflow
doesn't yet account for non-XML data.  We will probably go with the rapper
intermediary process until XForms 2 is released.

Ethan


On Mon, Sep 16, 2013 at 10:22 AM, Karen Coyle li...@kcoyle.net wrote:

 On 9/16/13 6:29 AM, aj...@virginia.edu wrote:

 -BEGIN PGP SIGNED MESSAGE-
 Hash: SHA1

 I'd suggest that perhaps the confusion arises because This instance is
 (not) 'valid' according to that ontology. might be inferred from an
 instance and an ontology (under certain conditions), and that's the soul of
 what we're asking when we define constraints on the data. Perhaps OWL can
 be used to express conditions of validity, as long as we represent the
 quality valid for use in inferences.


 Based on the results of the RDF Validation workshop [1], validation is
 being expressed today as SPARQL rules. If you express the rules in OWL then
 unfortunately you affect downstream re-use of your ontology, and that can
 create a mess for inferencing and can add a burden onto any reasoners,
 which are supposed to apply the OWL declarations.

 One participant at the workshop demonstrated a system that used the OWL
 constraints as constraints, but only in a closed system. I think that the
 use of SPARQL is superior because it does not affect the semantics of the
 classes and properties, only the instance data, and that means that the
 same properties can be validated differently for different applications or
 under different contexts. As an example, one community may wish to say that
 their metadata can have one and only one dc:title, while others may allow
 more than one. You do not want to constrain dc:title throughout the Web,
 only your own use of it. (Tom Baker and I presented a solution to this on
 the second day as Application Profiles [2], as defined by the DC community).

 kc
 [1] 
 https://www.w3.org/2012/12/**rdf-val/agendahttps://www.w3.org/2012/12/rdf-val/agenda
 [2] http://www.w3.org/2001/sw/**wiki/images/e/ef/Baker-dc-**
 abstract-model-revised.pdfhttp://www.w3.org/2001/sw/wiki/images/e/ef/Baker-dc-abstract-model-revised.pdf


  - ---
 A. Soroka
 The University of Virginia Library

 On Sep 13, 2013, at 11:00 PM, CODE4LIB automatic digest system wrote:

  Also, remember that OWL does NOT constrain your data, it constrains only
 the inferences that you can make about your data. OWL operates at the
 ontology level, not the data level. (The OWL 2 documentation makes this
 more clear, in my reading of it. I agree that the example you cite sure
 looks like a constraint on the data... it's very confusing.)

 -BEGIN PGP SIGNATURE-
 Version: GnuPG/MacGPG2 v2.0.19 (Darwin)
 Comment: GPGTools - http://gpgtools.org

 iQEcBAEBAgAGBQJSNwe2AAoJEATpPY**SyaoIkwLcIAK+**sMzy1XkqLStg94F2I40pe
 0DepjqVhdPnaDS1Msg7pd7c7iC0L5N**hCWd9BxzdvRgeMRr123zZ3EmKDSy8X**ZiGf
 uQyXlA9cOqpCxdQLj2zXv5VHrOdlsA**1UAGprwhYrxOz/**v3xQ7b2nXusRoZRfDlts
 iadvWx5DhLEb2+**uVl9geteeymLIVUTzm8WnUITEE7by2**HAQf9VlT9CrQSVQ21wLC
 hvmk47Nt8WIGyPwRh1qOhvIXLDLxD9**rkBSC1G01RhzwvctDy88Tmt2Ut47ZR**EScP
 YUz/bf/qxITzX2L7tE35s2w+**RUIFIFc4nJa3Xhp0wMoTAz5UYMiWIc**XZ38qfGlY=
 =PJTS
 -END PGP SIGNATURE-


 --
 Karen Coyle
 kco...@kcoyle.net http://kcoyle.net
 m: 1-510-435-8234
 skype: kcoylenet



Re: [CODE4LIB] CODE4LIB Digest - 12 Sep 2013 to 13 Sep 2013 (#2013-237)

2013-09-16 Thread Karen Coyle
Ethan, if you are interested in dialoguing about this topic, I suspect 
this isn't the forum for it. I don't think that W3C has yet set up a 
public list on rdf validation (the meeting participants need to form an 
actual W3C group for that to happen, and I hope that won't take too 
long), but there should be one soon. It's really rather useless to keep 
telling *me* this, since I'm not arguing for any particular technology, 
just reporting what I've learned in the last few weeks about what others 
are doing.


That is, if you are interested in having an exchange of ideas about this 
topic rather than repeatedly trying to convince me that what I'm saying 
is wrong. It's like you're trying to convince me that I really did not 
hear what I did. But I did hear it. Maybe all of those people are wrong; 
maybe you could explain to them that they are wrong. But if you care 
about this, then you need to be talking to them.


kc


On 9/16/13 7:40 AM, Ethan Gruber wrote:

Using SPARQL to validate seems like tremendous overhead.  From the Gerber
abstract: A total of 55 rules have been defined representing the
constraints and requirements of the OA Specification and Ontology. For each
rule we have defined a SPARQL query to check compliance. I hope this isn't
55 SPARQL queries per RDF resource.

Europeana's review of schematron indicated what I pointed out earlier, that
it confines one to using RDF/XML, which is sub-optimal in their own
words.  One could accept RDF in any serialization and then run it through
an RDF processor, like rapper (http://librdf.org/raptor/rapper.html), into
XML and then validate.  Eventually, when XPath/XSLT 3 supports JSON and
other non-XML data models, theoretically, schematron might then be able to
validate other serializations of RDF.  Ditto for XForms, which we are using
to validate RDF/XML.  Obviously, this is sub-optimal because our workflow
doesn't yet account for non-XML data.  We will probably go with the rapper
intermediary process until XForms 2 is released.

Ethan


On Mon, Sep 16, 2013 at 10:22 AM, Karen Coyle li...@kcoyle.net wrote:


On 9/16/13 6:29 AM, aj...@virginia.edu wrote:


-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

I'd suggest that perhaps the confusion arises because This instance is
(not) 'valid' according to that ontology. might be inferred from an
instance and an ontology (under certain conditions), and that's the soul of
what we're asking when we define constraints on the data. Perhaps OWL can
be used to express conditions of validity, as long as we represent the
quality valid for use in inferences.


Based on the results of the RDF Validation workshop [1], validation is
being expressed today as SPARQL rules. If you express the rules in OWL then
unfortunately you affect downstream re-use of your ontology, and that can
create a mess for inferencing and can add a burden onto any reasoners,
which are supposed to apply the OWL declarations.

One participant at the workshop demonstrated a system that used the OWL
constraints as constraints, but only in a closed system. I think that the
use of SPARQL is superior because it does not affect the semantics of the
classes and properties, only the instance data, and that means that the
same properties can be validated differently for different applications or
under different contexts. As an example, one community may wish to say that
their metadata can have one and only one dc:title, while others may allow
more than one. You do not want to constrain dc:title throughout the Web,
only your own use of it. (Tom Baker and I presented a solution to this on
the second day as Application Profiles [2], as defined by the DC community).

kc
[1] 
https://www.w3.org/2012/12/**rdf-val/agendahttps://www.w3.org/2012/12/rdf-val/agenda
[2] http://www.w3.org/2001/sw/**wiki/images/e/ef/Baker-dc-**
abstract-model-revised.pdfhttp://www.w3.org/2001/sw/wiki/images/e/ef/Baker-dc-abstract-model-revised.pdf


  - ---

A. Soroka
The University of Virginia Library

On Sep 13, 2013, at 11:00 PM, CODE4LIB automatic digest system wrote:

  Also, remember that OWL does NOT constrain your data, it constrains only

the inferences that you can make about your data. OWL operates at the
ontology level, not the data level. (The OWL 2 documentation makes this
more clear, in my reading of it. I agree that the example you cite sure
looks like a constraint on the data... it's very confusing.)


-BEGIN PGP SIGNATURE-
Version: GnuPG/MacGPG2 v2.0.19 (Darwin)
Comment: GPGTools - http://gpgtools.org

iQEcBAEBAgAGBQJSNwe2AAoJEATpPY**SyaoIkwLcIAK+**sMzy1XkqLStg94F2I40pe
0DepjqVhdPnaDS1Msg7pd7c7iC0L5N**hCWd9BxzdvRgeMRr123zZ3EmKDSy8X**ZiGf
uQyXlA9cOqpCxdQLj2zXv5VHrOdlsA**1UAGprwhYrxOz/**v3xQ7b2nXusRoZRfDlts
iadvWx5DhLEb2+**uVl9geteeymLIVUTzm8WnUITEE7by2**HAQf9VlT9CrQSVQ21wLC
hvmk47Nt8WIGyPwRh1qOhvIXLDLxD9**rkBSC1G01RhzwvctDy88Tmt2Ut47ZR**EScP
YUz/bf/qxITzX2L7tE35s2w+**RUIFIFc4nJa3Xhp0wMoTAz5UYMiWIc**XZ38qfGlY=
=PJTS
-END PGP SIGNATURE-


--
Karen Coyle