Re: [CODE4LIB] LCSH, Bisac, facets, hierarchy?

2016-04-14 Thread Karen Coyle
 content in one term.  Searching is exploring through overlapping
terms, and compiling a bibliography from the pearls found in the process.
This interface makes clearer what the related terms may be, given a borad
term like practical theology.  And it’s so nice that it combines the
classification structure with the subject headings.

Cindy Harper
@vts.edu<http://vts.edu>

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Kent Fitch
Sent: Wednesday, April 13, 2016 8:17 PM
To:CODE4LIB@LISTSERV.ND.EDU<mailto:CODE4LIB@LISTSERV.ND.EDU>
Subject: Re: [CODE4LIB] LCSH, Bisac, facets, hierarchy?
About ten years ago, I was wondering how to make the structure in LCSH,

or

at least how it was encoded in MARC subject tags more useful, so when
implementing a prototype for a new library catalogue at the National
Library of Australia, I tried using the subject tag contents to

represent a

hierarchy, then counted the number of hits against parts of that

hierarchy

for a given search and then represented the subject tags in a hierarchy
with hit counts.   One of the motivations was to help expose to the
searcher how works relevant to their search may have been
LCSH-subject-catalogued.

I'm a programmer, not a UI person, so the formatting of theresults were
fairly primitive, but that prototype from ten years ago ("Library Labs")

is

still running.

For example, search results for /ancient egypt/




http://ll01.nla.gov.au/search.jsp?searchTerm=ancient+egypt=0.5=0.05=12.0=9.0=9.0=9.0=4.0=3.0=3.0=3.0=18.0=15.0

/computer art/




http://ll01.nla.gov.au/search.jsp?searchTerm=computer+art=0.5=0.05=12.0=9.0=9.0=9.0=4.0=3.0=3.0=3.0=18.0=15.0

/history of utah/




http://ll01.nla.gov.au/search.jsp?searchTerm=history+of+utah=0.5=0.05=12.0=9.0=9.0=9.0=4.0=3.0=3.0=3.0=18.0=15.0

This prototype also explored a subject hierarchy which had been of
interest to the NLA's Assistant Director-General, Dr Warwick Cathro, over
many years, the RLG "Conspectus" hierarchy, which I guess was not unlike
BISAC in its aims.  It is shown further down the right-hand column.

Both the subject hierarchy and Conspectus were interesting, but neither
made it into the eventual production search system, Trove, implemented at
the NLA, in which subject faceting or hierarchy is absent from results
display:

http://trove.nla.gov.au/book/result?q=ancient+egypt
http://trove.nla.gov.au/book/result?q=computer+art
http://trove.nla.gov.au/book/result?q=history+of+utah

The "Library Labs" prototype is running on a small VM, so searching may

be

slow, and it hasnt been updated with any content since 2006..  But maybe
the way it attempted to provide subject grouping and encourage narrowing

of

search by LCSH or exploring using LCSH rather than the provided search
terms may trigger some more experiments.

Kent Fitch

On Thu, Apr 14, 2016 at 3:11 AM, Mark Watkins <m...@thehawaiiproject.com
<mailto:m...@thehawaiiproject.com>>
wrote:


 :)

sounds like there is a lot of useful metadata but somewhat scattered
amongst various fields, depending on when the item was cataloged or

tagged.

Which seems to correspond to anecdotal surfing of the Harvard data.

I guess my new task is to build something that aggregates and
reconciles portions of LCSH, LCFGT, and GSAFD :).

Thanks for the additional perspective!





--
Karen Coyle
kco...@kcoyle.net  http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] LCSH, Bisac, facets, hierarchy?

2016-04-14 Thread Karen Coyle

On 4/14/16 7:40 AM, Joseph Montibello wrote:

Real use of LCSH
would search the reference vocabulary as well as the preferred term
headings which get into bib records.  Working with LCSH bib headings alone
misses the point of a sophisticated controlled vocabulary, where much of
the terminological and semantic richness for searching is contained in
"see" and "see also" references, complex references and scope and other
kinds of notes.  The controlled vocabulary itself needs to be integrated
into search results so that searches call up not only bib records with a
matching heading but vocabulary records which can expand the user's search
vocabulary and point to related controlled terms outside those generated by
the retrieved bib records' themselves.

Are there discovery systems out there that attempt this? It would be great to use all the 
work that has gone into these vocabs to improve end-user experience, not by telling them 
to click a "see also" link but by doing that work for them in some way.


We tried that in an early version of the U of Calif's MELVYL system (I'm 
talking early 80's here). The difficulty is in trying to coordinate a 
keyword search, that can bring up a wide variety of headings in very 
different "hierarchies", and some direct linking that would make sense 
to the user. The best we could do was to combine the keywords from the 
authoritative and non-authoritative headings for retrieval, thus 
increasing both the desired retrieval but also the false drops. If you 
could isolate the subject "graphs" and present them it would be cleaner, 
but in some cases the number of different graphs retrieved would make 
for a very difficult presentation for the user.


The disconnect between keyword searching and headings is something that 
needs more analysis.


kc


Joe Montibello, MLIS
Library Systems Manager
Dartmouth College
603.646.9394
joseph.montibe...@dartmouth.edu


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] LCSH, Bisac, facets, hierarchy?

2016-04-14 Thread Karen Coyle
 content in one term.  Searching is exploring through overlapping
terms, and compiling a bibliography from the pearls found in the process.
This interface makes clearer what the related terms may be, given a borad
term like practical theology.  And it’s so nice that it combines the
classification structure with the subject headings.

Cindy Harper
@vts.edu<http://vts.edu>

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Kent Fitch
Sent: Wednesday, April 13, 2016 8:17 PM
To: CODE4LIB@LISTSERV.ND.EDU<mailto:CODE4LIB@LISTSERV.ND.EDU>
Subject: Re: [CODE4LIB] LCSH, Bisac, facets, hierarchy?
About ten years ago, I was wondering how to make the structure in LCSH,

or

at least how it was encoded in MARC subject tags more useful, so when
implementing a prototype for a new library catalogue at the National
Library of Australia, I tried using the subject tag contents to

represent a

hierarchy, then counted the number of hits against parts of that

hierarchy

for a given search and then represented the subject tags in a hierarchy
with hit counts.   One of the motivations was to help expose to the
searcher how works relevant to their search may have been
LCSH-subject-catalogued.

I'm a programmer, not a UI person, so the formatting of theresults were
fairly primitive, but that prototype from ten years ago ("Library Labs")

is

still running.

For example, search results for /ancient egypt/




http://ll01.nla.gov.au/search.jsp?searchTerm=ancient+egypt=0.5=0.05=12.0=9.0=9.0=9.0=4.0=3.0=3.0=3.0=18.0=15.0

/computer art/




http://ll01.nla.gov.au/search.jsp?searchTerm=computer+art=0.5=0.05=12.0=9.0=9.0=9.0=4.0=3.0=3.0=3.0=18.0=15.0

/history of utah/




http://ll01.nla.gov.au/search.jsp?searchTerm=history+of+utah=0.5=0.05=12.0=9.0=9.0=9.0=4.0=3.0=3.0=3.0=18.0=15.0

This prototype also explored a subject hierarchy which had been of
interest to the NLA's Assistant Director-General, Dr Warwick Cathro, over
many years, the RLG "Conspectus" hierarchy, which I guess was not unlike
BISAC in its aims.  It is shown further down the right-hand column.

Both the subject hierarchy and Conspectus were interesting, but neither
made it into the eventual production search system, Trove, implemented at
the NLA, in which subject faceting or hierarchy is absent from results
display:

http://trove.nla.gov.au/book/result?q=ancient+egypt
http://trove.nla.gov.au/book/result?q=computer+art
http://trove.nla.gov.au/book/result?q=history+of+utah

The "Library Labs" prototype is running on a small VM, so searching may

be

slow, and it hasnt been updated with any content since 2006..  But maybe
the way it attempted to provide subject grouping and encourage narrowing

of

search by LCSH or exploring using LCSH rather than the provided search
terms may trigger some more experiments.

Kent Fitch

On Thu, Apr 14, 2016 at 3:11 AM, Mark Watkins <m...@thehawaiiproject.com
<mailto:m...@thehawaiiproject.com>>
wrote:


 :)

sounds like there is a lot of useful metadata but somewhat scattered
amongst various fields, depending on when the item was cataloged or

tagged.

Which seems to correspond to anecdotal surfing of the Harvard data.

I guess my new task is to build something that aggregates and
reconciles portions of LCSH, LCFGT, and GSAFD :).

Thanks for the additional perspective!








--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Google can give you answers, but librarians give you the right answers

2016-04-08 Thread Karen Coyle

On 4/6/16 9:51 AM, Kyle Banerjee wrote:

On Wed, Apr 6, 2016 at 7:42 AM, Karen Coyle <li...@kcoyle.net> wrote:


... Libraries "do" it, but our user interfaces ignore it (honestly, does
anyone NOT think that the whole BT/NT relationship in LCSH is completely
wasted in today's systems?).  Google searches "work" best on proper nouns
that are nearly unique. You cannot do concept searches, and you cannot see
relationships between concepts. It's great for named people, organizations
and products, but not great for anything else.[1]...


Conceptually, I like the idea of using the relationships in LCSH. However,
I don't hold out much hope that anyone will make hay out of that.

The percentage of things that have decent LCSH assigned to them is small
and shrinking for the simple reason that a fewer and fewer humans have to
manage more resources.


I'm not sure what you are saying here -- that there are fewer headings 
being assigned, or that they are not as "good" as ones assigned in the 
past? Or is it that many of our resources aren't covered by library 
cataloging rules?



Automation could help (getting the needed data from
publishers might be tricky), but the only benefit I can think of for using
LCSH for automated applications is to maximize relationships with older
materials -- possibly at the expense of the "findability" of the newer
stuff.

LCSH is relatively flat, the rules for constructing headings are so
Byzantine that they stymie even experienced catalogers (which contributes
to inconsistent application in terms of quality, level of analysis, and
completeness), and its ability to express concepts at all is highly
variable as it is designed by a committee on an enumerative basis.


?? Sorry, what's this "enumerative basis"?


  Add to
this that concepts in records frequently must be expressed across multiple
headings and subheadings, any type of automated assignment is going to
result in really "dirty" relationships so I can't blame ILS designers for
limiting their use of LCSH primarily to controlled keyword access.


Well, actually, there's nothing at all "controlled" about keyword 
access. It's pretty much a pot shot, or, as I've called it before, a 
form of dumpster diving for information. There is a huge disconnect 
between the results of keyword searching and the expected functionality 
(read: user service) of controlled left-anchored headings, and I 
continue to be amazed that we've been living with this disconnect for 
decades without ever coming to an agreement that we need a solution.[1] 
Instead, great effort goes into modifying the descriptive cataloging 
rules, while no major changes have been made in subject access. I find 
this to be... well, stunning, in the sense that I'm pretty much stunned 
that this is the case.


kc

[1] See Lorcan Dempsey's approach in this slide deck: 
http://www.slideshare.net/lisld/irish-studies-making-library-data-work-harder 
. It has to do with "things", which is in good part what I first found 
interesting about LoD.


kyle


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Google can give you answers, but librarians give you the right answers

2016-04-06 Thread Karen Coyle

On 4/6/16 4:04 AM, Eric Lease Morgan wrote:

Instead, I think the problem to solve surrounds assisting the reader in using & 
understanding the stuff they find.


I'd like to see innovation a step before find, but I think in a sense 
we're on the same wavelength. My take is that bibliographic information 
should be the end of a process, not the beginning. Before arriving at 
bib data, there's a lot of understanding and context that needs to be 
clarified. Some of this involves authors and subjects, but not as they 
are currently represented (mainly as text strings, and without 
relationships). I think that one of the main questions a user has at the 
catalog is "Where am I?" - where am I in the knowledge universe, where 
am I in the collection of this library?


Note that Google does not give users an answer to this question because 
there is no larger context, no inherent organization. Google does not do 
knowledge organization. Libraries "do" it, but our user interfaces 
ignore it (honestly, does anyone NOT think that the whole BT/NT 
relationship in LCSH is completely wasted in today's systems?).  Google 
searches "work" best on proper nouns that are nearly unique. You cannot 
do concept searches, and you cannot see relationships between concepts. 
It's great for named people, organizations and products, but not great 
for anything else.[1] Also, without the links that fuel pagerank, the 
ranking is very unsatisfactory - cf. Google Book searches, which are 
often very unsatisfying -- and face it, if Google can't make it work, 
what are the odds that we can? We do have knowledge organization 
potential; it's a bit out of date, it hasn't been made truly actionable, 
but it's there.


kc

[1] Except where there's a Wikipedia article using the concept term. 
Basically Wikipedia provides the only knowledge organization that Google has


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Google can give you answers, but librarians give you the right answers

2016-04-05 Thread Karen Coyle
Eric, there were studies done a few decades ago using factual questions. 
Here's a critical round-up of some of the studies: 
http://www.jstor.org/stable/25828215


Basically, 40-60% correct, but possibly the questions were not 
representative -- so possibly the results are really worse :(


kc

On 4/5/16 1:11 PM, Eric Lease Morgan wrote:

  I sincerely wonder to what extent librarians give the reader
(patrons) the right -- correct -- answer to a (reference) question.
Such is a hypothesis that can be tested and measured. Please show me
non-antidotal evidence one way or the other. --ELM


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Structured Data Markup on library web sites

2016-03-29 Thread Karen Coyle
I'm using OpenLink structured data sniffer[1]. No idea how it intuits 
schema that isn't there...


kc
[1] http://osds.openlinksw.com/

On 3/29/16 9:46 AM, Kevin Ford wrote:

Huh. I didn't look at "How to read..." but I did look at the other two.

Just so we're on the same page, here are the two I found in which I 
could not detect any schema.org markup:


http://bmcr.brynmawr.edu/2014/2014-02-18.html
http://bmcr.brynmawr.edu/2014/2014-01-48.html

Interestingly, the "how to read" doesn't appear to have schema.org 
either:


http://bmcr.brynmawr.edu/2014/2014-01-07.html

FWIW, I'm looking at the HTML itself.  You may be using a tool that is 
generating a little but of schema.  Is that accurate?


If you look at the HTML of the 2016 item you sent along, you can see 
the schema.org vocab embedded in the code:


view-source:http://bmcr.brynmawr.edu/2016/2016-03-30.html

That's probably why it generated more output.  The 2014 reviews do not 
actually include schema.org markup.


Best,
Kevin


On 03/29/2016 11:36 AM, Karen Coyle wrote:

That's odd. I haven't done a large survey, but every recent item that
I've looked at has had the code.

http://bmcr.brynmawr.edu/2016/2016-03-30.html

Statement Collection #1
Entity http://bmcr.brynmawr.edu/2016/2016-03-30.html
Attributes
rdfa:usesVocabulary <http://www.w3.org/ns/rdfa#usesVocabulary>
schema: <http://schema.org/>



Statement Collection #2
Entity http://bmcr.brynmawr.edu/2016/2016-03-30.html#review_text
Attributes
rdf:type <http://www.w3.org/1999/02/22-rdf-syntax-ns#type> schema:Review
<http://schema.org/Review>
schema:name <http://schema.org/name> Bryn Mawr Classical Review
2016.03.30
schema:itemReviewed <http://schema.org/itemReviewed>
http://bmcr.brynmawr.edu/2016/2016-03-30.html#review_item
schema:author <http://schema.org/author> Robert W. Wallace
schema:reviewBody <http://schema.org/reviewBody> This book is cause
for celebration. Notwithstanding the complexity of his many source
traditions, Solon is our best attested historical figure from archaic
Greece, as sophos, poet, statesman, lawgiver, and the subject of a
biography by Plutarch.


It's possible that earlier data wasn't coded sufficiently to pick up the
details. Here's the "latin poem" one:

Entity http://bmcr.brynmawr.edu/2014/2014-01-07.html#this
Attributes
schema:description <http://schema.org/description> William
Fitzgerald, How to Read a Latin Poem: If You Can’t Read Latin Yet.
Oxford; New York: Oxford University Press, 2013. Pp. ix, 278. ISBN
9780199657865. $35.00.


This is indeed why I wanted a "before and after" test - to see if schema
did add SEO. Now we don't know.

kc

On 3/29/16 7:48 AM, Kevin Ford wrote:

Hi Karen,

I took a look at those bryn mawr hits and I don't see the schema.org
used in the page.  Am I missing it?  Perhaps I found the wrong thing.

If indeed it's not there, it just goes to show how using schema is not
a panacea.  Loads of factors go into search ranking, relevancy, and
display.

Yours,
Kevin

On 03/24/2016 09:28 PM, Karen Coyle wrote:

I worked on the addition of schema.org data to the Bryn Mawr Classical
Reviews. Although I advised doing a "before and after" test to see how
it affected retrieval, I lost touch with the folks before that could
happen. However, their reviews do show up fairly high in Google, 
around

the 3-5th place on page one. Try these searches:

how to read a latin poem
/From Listeners to Viewers:/
/Butrint 4: The Archaeology and Histories of an Ionian Town

kc

/
On 3/22/16 5:44 PM, Jennifer DeJonghe wrote:

Hello,

I'm looking for examples of library web sites or university web sites
that are using Structured Data / schema.org to mark up books,
locations, events, etc, on their public web sites or blogs. I'm NOT
really looking for huge linked data projects where large record sets
are marked up, but more simple SEO practices for displaying rich
snippets in search engine results.

If you have examples of library or university websites doing this,
please send me a link!

Thank you,
Jennifer

Jennifer DeJonghe
Librarian and Professor
Library and Information Services
Metropolitan State University
St. Paul, MN






--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Structured Data Markup on library web sites

2016-03-29 Thread Karen Coyle
That's odd. I haven't done a large survey, but every recent item that 
I've looked at has had the code.


http://bmcr.brynmawr.edu/2016/2016-03-30.html

Statement Collection #1 
Entity  http://bmcr.brynmawr.edu/2016/2016-03-30.html
Attributes  
rdfa:usesVocabulary <http://www.w3.org/ns/rdfa#usesVocabulary> 	schema: 
<http://schema.org/>




Statement Collection #2 
Entity  http://bmcr.brynmawr.edu/2016/2016-03-30.html#review_text
Attributes  
rdf:type <http://www.w3.org/1999/02/22-rdf-syntax-ns#type> 
schema:Review <http://schema.org/Review>

schema:name <http://schema.org/name>  Bryn Mawr Classical Review 2016.03.30
schema:itemReviewed <http://schema.org/itemReviewed> 
http://bmcr.brynmawr.edu/2016/2016-03-30.html#review_item

schema:author <http://schema.org/author>  Robert W. Wallace
schema:reviewBody <http://schema.org/reviewBody> 	This book is cause for 
celebration. Notwithstanding the complexity of his many source 
traditions, Solon is our best attested historical figure from archaic 
Greece, as sophos, poet, statesman, lawgiver, and the subject of a 
biography by Plutarch.



It's possible that earlier data wasn't coded sufficiently to pick up the 
details. Here's the "latin poem" one:


Entity  http://bmcr.brynmawr.edu/2014/2014-01-07.html#this
Attributes  
schema:description <http://schema.org/description> 	William Fitzgerald, 
How to Read a Latin Poem: If You Can’t Read Latin Yet. Oxford; New York: 
Oxford University Press, 2013. Pp. ix, 278. ISBN 9780199657865. $35.00.



This is indeed why I wanted a "before and after" test - to see if schema 
did add SEO. Now we don't know.


kc

On 3/29/16 7:48 AM, Kevin Ford wrote:

Hi Karen,

I took a look at those bryn mawr hits and I don't see the schema.org 
used in the page.  Am I missing it?  Perhaps I found the wrong thing.


If indeed it's not there, it just goes to show how using schema is not 
a panacea.  Loads of factors go into search ranking, relevancy, and 
display.


Yours,
Kevin

On 03/24/2016 09:28 PM, Karen Coyle wrote:

I worked on the addition of schema.org data to the Bryn Mawr Classical
Reviews. Although I advised doing a "before and after" test to see how
it affected retrieval, I lost touch with the folks before that could
happen. However, their reviews do show up fairly high in Google, around
the 3-5th place on page one. Try these searches:

how to read a latin poem
/From Listeners to Viewers:/
/Butrint 4: The Archaeology and Histories of an Ionian Town

kc

/
On 3/22/16 5:44 PM, Jennifer DeJonghe wrote:

Hello,

I'm looking for examples of library web sites or university web sites
that are using Structured Data / schema.org to mark up books,
locations, events, etc, on their public web sites or blogs. I'm NOT
really looking for huge linked data projects where large record sets
are marked up, but more simple SEO practices for displaying rich
snippets in search engine results.

If you have examples of library or university websites doing this,
please send me a link!

Thank you,
Jennifer

Jennifer DeJonghe
Librarian and Professor
Library and Information Services
Metropolitan State University
St. Paul, MN




--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Structured Data Markup on library web sites

2016-03-24 Thread Karen Coyle
I worked on the addition of schema.org data to the Bryn Mawr Classical 
Reviews. Although I advised doing a "before and after" test to see how 
it affected retrieval, I lost touch with the folks before that could 
happen. However, their reviews do show up fairly high in Google, around 
the 3-5th place on page one. Try these searches:


how to read a latin poem
/From Listeners to Viewers:/
/Butrint 4: The Archaeology and Histories of an Ionian Town

kc

/
On 3/22/16 5:44 PM, Jennifer DeJonghe wrote:

Hello,

I'm looking for examples of library web sites or university web sites that are 
using Structured Data / schema.org to mark up books, locations, events, etc, on 
their public web sites or blogs. I'm NOT really looking for huge linked data 
projects where large record sets are marked up, but more simple SEO practices 
for displaying rich snippets in search engine results.

If you have examples of library or university websites doing this, please send 
me a link!

Thank you,
Jennifer

Jennifer DeJonghe
Librarian and Professor
Library and Information Services
Metropolitan State University
St. Paul, MN


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] DPLA and Pop Up Archive partner to make audiovisual collections across the U.S. searchable

2015-11-18 Thread Karen Coyle
open platform 
for developers. The model supports or establishes local 
collaborations, professional networks, metadata globalization, and 
long-term sustainability. It ensures that even the smallest 
institutions have an on-ramp to participation in DPLA.


/Pop Up Archive is funded by the John S. and James L. Knight 
Foundation, the National Endowment for the Humanities, and the 
Institute for Museum and Library Services. Pop Up Archive’s partner 
collections include the American Archive of Public Broadcasting, the 
Studs Terkel Radio Archive from the WFMT Radio Network, and tens of 
thousands of hours of audio from across the United States collected by 
StoryCorps, the New York Public Library, and numerous public media, 
storytelling, and oral history organizations. Learn more at 
//www.popuparchive.org/ 
<http://us4.list-manage.com/track/click?u=e1490d1305c4b651f3ad0ace4=ea65a70553=9ed97c7d64>/./


/The Digital Public Library of America is generously supported by a 
number of foundations and government agencies, including the Alfred P. 
Sloan Foundation, the Andrew W. Mellon Foundation, an anonymous donor, 
the Arcadia Fund, the Bill & Melinda Gates Foundation, the Institute 
of Museum and Library Services, the John S. and James L. Knight 
Foundation, the Whiting Foundation, and the National Endowment for the 
Humanities. To find out more about DPLA, visit //http://dp.la// 
<http://us4.list-manage.com/track/click?u=e1490d1305c4b651f3ad0ace4=dbfca60bba=9ed97c7d64>/./


dp.la 
<http://us4.list-manage.com/track/click?u=e1490d1305c4b651f3ad0ace4=4f2923a5f4=9ed97c7d64> 

dp.la 
<http://us4.list-manage.com/track/click?u=e1490d1305c4b651f3ad0ace4=b30e84203e=9ed97c7d64> 



Facebook 
<http://us4.list-manage2.com/track/click?u=e1490d1305c4b651f3ad0ace4=902402ff96=9ed97c7d64> 

Facebook 
<http://us4.list-manage.com/track/click?u=e1490d1305c4b651f3ad0ace4=4ec9793366=9ed97c7d64> 



Twitter 
<http://us4.list-manage.com/track/click?u=e1490d1305c4b651f3ad0ace4=46ecea48a4=9ed97c7d64> 

Twitter 
<http://us4.list-manage.com/track/click?u=e1490d1305c4b651f3ad0ace4=07f1b6aab1=9ed97c7d64> 



Tumblr 
<http://us4.list-manage.com/track/click?u=e1490d1305c4b651f3ad0ace4=3ed4f9ad16=9ed97c7d64> 

Tumblr 
<http://us4.list-manage2.com/track/click?u=e1490d1305c4b651f3ad0ace4=aba4d3124a=9ed97c7d64> 



Pinterest 
<http://us4.list-manage.com/track/click?u=e1490d1305c4b651f3ad0ace4=2de196dad5=9ed97c7d64> 

Pinterest 
<http://us4.list-manage.com/track/click?u=e1490d1305c4b651f3ad0ace4=f637ce1c18=9ed97c7d64> 



Email <mailto:i...@dp.la>
Email <mailto:i...@dp.la>

Donate 
<http://us4.list-manage1.com/track/click?u=e1490d1305c4b651f3ad0ace4=2671428bea=9ed97c7d64> 

Donate 
<http://us4.list-manage.com/track/click?u=e1490d1305c4b651f3ad0ace4=1265c30f54=9ed97c7d64> 



You are receiving this message because you are subscribed to the 
Digital Public Library of America (DPLA) News List.


Follow on Twitter 
<http://us4.list-manage.com/track/click?u=e1490d1305c4b651f3ad0ace4=90a3a7cbdd=9ed97c7d64> | 
Friend on Facebook 
<http://us4.list-manage1.com/track/click?u=e1490d1305c4b651f3ad0ace4=de9809cf29=9ed97c7d64> | 
Forward to a friend 
<http://us4.list-manage1.com/track/click?u=e1490d1305c4b651f3ad0ace4=aeab5358ca=9ed97c7d64>








This email was sent to li...@kcoyle.net <mailto:li...@kcoyle.net>
/why did I get this?/ 
<http://us4.list-manage1.com/about?u=e1490d1305c4b651f3ad0ace4=1ee988aadc=9ed97c7d64=2031452ad8> 
unsubscribe from this list 
<http://us4.list-manage1.com/unsubscribe?u=e1490d1305c4b651f3ad0ace4=1ee988aadc=9ed97c7d64=2031452ad8> 
update subscription preferences 
<http://us4.list-manage.com/profile?u=e1490d1305c4b651f3ad0ace4=1ee988aadc=9ed97c7d64> 

Digital Public Library of America · Digital Public Library of America 
c/o Boston Public Library · 700 Boylston St. · Boston, MA 02116 · USA




--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] SWIB15 – Registration Open (Semantic Web in Libraries, Hamburg, 23-25 November)

2015-08-11 Thread Karen Coyle
I just got there through the link. I did have some problems with another 
email that used https and the SWIB site was rejected by Chrome (but 
not Firefox). Try again?


kc (on the SWIB org committee)

On 8/11/15 7:27 AM, Justin Coyne wrote:

The links you provided aren't resolving.

Not Found

The requested URL /swib15/programme.php was not found on this server.


-Justin

On Tue, Aug 11, 2015 at 9:18 AM, Neubert, Joachim j.neub...@zbw.eu wrote:


We are happy to invite you to this year's SWIB (Semantic Web in Libraries)
conference in Hamburg, 23 - 25 November 2015.

Take a look at the programme for SWIB15 here:
http://swib.org/swib15/programme.php.

You can register for the conference at
http://swib.org/swib15/registration.php.

Further information and contact:

Joachim Neubert
German National Library of Economics
Leibniz Information Centre for Economics (ZBW)
Phone +49 40 428 34-462
E-Mail: j.neubert(at) zbw.eu

Adrian Pohl
North Rhine-Westphalian Library Service Center (hbz)
Phone +49 221 400 75-235
E-Mail: swib(at)hbz-nrw.de

  Website:
http://swib.org/swib15

Twitter:
http://twitter.com/swibcon

Twitter Hashtag:
#swib15

Looking forward to meeting you in Hamburg,
Joachim



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


[CODE4LIB] Fwd: new W3C CSV on the Web specs, now at Candidate Recommendation stage - please implement!

2015-08-03 Thread Karen Coyle
For those who have to work with output from spreadsheets. And I wonder 
if this doesn't interact with OpenRefine output as well.


kc

 Forwarded Message 
Subject: new W3C CSV on the Web specs, now at Candidate Recommendation 
stage -  please implement!

Resent-Date: Tue, 28 Jul 2015 14:33:21 +
Resent-From: public-voc...@w3.org
Date: Tue, 28 Jul 2015 15:32:52 +0100
From: Dan Brickley dan...@google.com
To: public-...@w3.org, semantic-...@w3.org Web semantic-...@w3.org, 
W3C Web Schemas Task Force public-voc...@w3.org


Hi! Short version: Please see
http://www.w3.org/blog/news/archives/4830 for the Candidate
Recommendation specs from W3C's CSV on the Web group -
https://www.w3.org/2013/csvw/wiki/Main_Page

Long version:

These are the 4 docs,

Model for Tabular Data and Metadata on the Web—an abstract model for
tabular data, and how to locate metadata that enables users to better
understand what the data holds; this specification also contains
non-normative guidance on how to parse CSV files
http://www.w3.org/TR/2015/CR-tabular-data-model-20150716/

Metadata Vocabulary for Tabular Data—a JSON-based format for
expressing metadata about tabular data to inform validation,
conversion, display and data entry for tabular data
http://www.w3.org/TR/2015/CR-tabular-metadata-20150716/

Generating JSON from Tabular Data on the Web—how to convert tabular
data into JSON
http://www.w3.org/TR/2015/CR-csv2json-20150716/

Generating RDF from Tabular Data on the Web—how to convert tabular
data into RDF
http://www.w3.org/TR/2015/CR-csv2rdf-20150716/

See the blog post for more links including an extensive set of test
cases, our GitHub repo and the mailing list for feedback. Also note
that the approach takes CSV as its central stereotypical use case but
should apply to many other tabular data-sharing approaches too (e.g.
most obviously tab separated). So if you prefer tab-separated files to
comma-separated, do please take a look! The Model spec defines that
common model, the metadata document defines terminology for talking
about instances of that model, and the last two specs apply this
approach to the problem of mapping tables into JSON and/or RDF.

The group expects to satisfy the implementation goals (i.e., at least
two, independent implementations for each of the test cases) by
October 30, 2015. Please take a look, and pass this along to other
groups who may be interested.

cheers,

Dan

for the CSVW WG


p.s. since I'm writing I'll indulge myself and share my personal
favourite part, which is the ability (in the csv2rdf doc) to map from
rows in a table via templates into RDF triples. This is a particularly
interesting/important facility and worth some attention. Normally I
wouldn't enthuse over (yet another) new RDF syntax but the ability to
map tabular data into triples via out-of-band mappings is very
powerful. BTW the group gave some serious consideration to applying
R2RML here (see docs and github/wiki for details), however given the
subtle differences between SQL and CSV environments we have taken a
different approach. Anyway please take a look!


Re: [CODE4LIB] Indie Preserves

2015-07-30 Thread Karen Coyle
I recommend a look at Pop Up Archive [1] - digital archiving for the 
non-archivist. It's heavily based on the archiving of sound files.


kc
[1] https://www.popuparchive.com/

On 7/29/15 9:13 PM, Scott Carlson wrote:

Apologies for any cross-posting, and please excuse the shameless self-promotion... 
Norie Guthrie (an archivist/special collections librarian) and myself have started 
a website/blog to help DIY  born-digital music labels with the 
digital/physical preservation of their materials. We hope to provide practical 
archiving tips and solutions to those putting out music on a shoestring budget.

This past spring, we conducted a survey to understand what types of materials 
record labels were saving and how they were saving them. We hope to formally 
present on this data some time in the future.

If you have time, please stop by the blog: http://www.indiepreserves.info/

Feel free to look us up on Twitter as well: https://twitter.com/IndiePreserves

Thanks,
Scott Carlson
Metadata Coordinator
Rice University, Fondren Library
scarl...@rice.edu


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Auto discovery of Dewey, UDC

2015-06-14 Thread Karen Coyle
Oh, dear. That's not good news:  There is no current projected date for 
the return of Dewey.info right now. We apologize for the inconvenience. 
Dated June 9, 2015.


kc

On 6/12/15 10:54 AM, LeVan,Ralph wrote:

This is as close to an official statement as I can find:

http://www.oclc.org/developer/news/2015/dewey-down.en.html

I've asked around, but can't add anything to that.

Ralph

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Karen 
Coyle
Sent: Friday, June 12, 2015 12:35 PM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: Auto discovery of Dewey, UDC

Hi. I tweeted this last month and got a reply that dewey.info is indeed 
currently down for major work, but is intended to return. That said, it was 
intended to return a month or two ago, so the usual coding project delays are 
in action here. ;-)

kc

On 6/12/15 7:08 AM, Sergio Letuche wrote:

dewey.info

seems to be dead, we have also checked this.

2015-06-12 16:57 GMT+03:00 Péter Király kirun...@gmail.com:


Hi Sergio,

As part of eXtensible Catalog we developed a Dewey module for Drupal,
which takes a Dewey number, and use OCLC's dewey.info to fetch the
textual description of the part. When it was created the service
contained only 3 levels of the classification system, since then they
went ahead, and now it is deeper.

You can find the sorce here:
http://cgit.drupalcode.org/xc/tree/xc_dewey/xc_dewey.module?h=7.x-1.x

Maybe it helps you.

Regarding to UDC: it is much a harder task, and when I worked with it,
I run into a blocking problem, which is that UDC was not licenced as
freely usable, and I was not able to get a licence to use it in an
open source project. There were some other problems as well: UDC
changed from time to time, and sometimes it means, that a given
classification code means this thing in a given point of time, and
that thing some years later. The MARC catalog I worked with did not
contain any information about the UDC versions, so the accuracy of the
tool was not guaranted (of course you can do some intelligent
guessing). And the last problem was, that on contrary to the Dewey
classification UDC contains sometime very lengthy descriptions instead
of one or two words. Semantically it is OK, but makes the UI design a
little bit hard, and if you want to search for the textual
description, you'll end up sometimes with a noisy result set.
Otherwise to handle the operators, the subclasses, and all the nice
things UDC provides is a very interesting challange.

Cheers,
Péter


2015-06-12 12:59 GMT+02:00 Sergio Letuche code4libus...@gmail.com:

thank you very much for your quick reply, dear Stefano,

i appreciate it

2015-06-12 13:47 GMT+03:00 Stefano Bargioni bargi...@pusc.it:


Hi, Sergio:
maybe this article [1 abstract] [2 English text] can give you some basic
ideas. We added a lot of DDC info in our Koha catalog two years ago.
HTH. Stefano

[1] http://leo.cineca.it/index.php/jlis/article/view/8766
[2] http://leo.cineca.it/index.php/jlis/article/view/8766/8060

On 12/giu/2015, at 12:03, Sergio Letuche code4libus...@gmail.com

wrote:

hello community!

we are facing this challenging issue. We need to complete for a vast

amount

of records, the dewey, UDC info, has anyone had any experience with

this?

We need some way (via modeling? mahout?) to try and discover these

values,

based on some text, found in the records' metadata, and then auto

complete

these values.

I would appreciate any feedback, if there is any opensource tool you

have

used for this purpose, or if you are aware of any best practice for

doing

this task.

Best


__
Il tuo 5x1000 al Patronato di San Girolamo della Carita' e' un gesto
semplice ma di grande valore.
Una tua firma aiutera' i sacerdoti ad essere piu' vicini alle esigenze

di

tutti noi.
Aiutaci a formare sacerdoti e seminaristi provenienti dai 5 continenti
indicando nella dichiarazione dei redditi il codice fiscale 97023980580.



--
Péter Király
software developer
GWDG, Göttingen - Europeana - eXtensible Catalog - The Code4Lib Journal
http://linkedin.com/in/peterkiraly



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Distribution of collections by DDC or UDC?

2015-06-12 Thread Karen Coyle
Christina, I was hoping that someone with more info would reply, but to 
my knowledge the services that do these statistical surveys are 
pay-fer so this might be considered proprietary info. Presumably 
libraries that have used these services received the results, but may 
not be allowed to share them. You might, however, have better luck on a 
list that has a higher percentage of collection development librarians. 
Unfortunately, I don't know what list that would be. Anyone?


kc

On 5/14/15 6:47 AM, Pikas, Christina K. wrote:

This might be a bizarre question, but can anyone point to some analysis for a 
large general library, consortium, or even like WorldCat, a distribution of 
materials by class?  So say for example 10% of the collection is in the 700s, 
and half of that is in the 741s, a quarter is in 746.432...
This table:
Table 4: Subject breakdown, nonfiction print books
History and auxiliary sciences

8 percent

Engineering and technology

7 percent

Business and economics

7 percent

Language, linguistics, and literature

6 percent

Philosophy and religion

5 percent

Health and medicine

5 percent

Art and architecture

3 percent

Law

3 percent

Sociology

3 percent

Education

3 percent

Other

15 percent

Unknown

35 percent

 From http://www.dlib.org/dlib/november09/lavoie/11lavoie.html isn't really 
granular enough.
Thanks!
--
Christina K. Pikas
Librarian
The Johns Hopkins University Applied Physics Laboratory
Baltimore: 443.778.4812
D.C.: 240.228.4812
christina.pi...@jhuapl.edu


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] hathitrust research center workset browser

2015-06-01 Thread Karen Coyle
Right. Which is why *someone* copied all of the Google digitized books 
to the Internet Archive -- someone not associated with the library 
partners. So generally if you cannot download from HT you can find the 
same scan via openlibrary.org. Unfortunately that doesn't help with 
using the tool that ELM has alerted us to.


kc

On 6/1/15 2:19 PM, Jimmy Ghaphery wrote:

I think we are in agreement (especially about the utility of all things
HathiTrust). My one point is that any restrictions on digitized public
domain works, as I understand it, are not related to copyright.

On Mon, Jun 1, 2015 at 5:00 PM, Terry Reese ree...@gmail.com wrote:


However, the digitizing agency cannot dictate any copyright
restrictions on the digitized copies once released to the public

The digital objects have not, and as far as I understand, cannot be made
available to the public if digitized as part of the google books
digitization project.  Most institutions got very limited use, and
generally these were tied to their specific, immediate, communities.
Though, with that said each institution has slightly different terms.  For
what it's worth, the research center does not make the digital copies
available for download -- it provides tools for working with data in
aggregate (worksets) and provides a proof of concept environment
demonstrating the feasibility of creating a secured data repository with I
believe the long-term goal of providing data mining for the entire
hathitrust resources (both within and outside of the public domain).  But
even as it stands now, the tool has become a fantastic teaching tool when
talking to instructors and graduate students looking for large data sets to
work with, that also includes some pretty interesting research algori!
  thms for working with the data.

--tr

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of
Jimmy Ghaphery
Sent: Monday, June 1, 2015 4:47 PM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] hathitrust research center workset browser

Thanks Eric for posting the webinar in the other thread.

I am pretty sure that digitizing something in the public domain does not
change its copyright status, at least in the U.S. The digitizing agency
certainly has the right to sell, restrict access, watermark, or even keep
the scans locked up on a thumb drive in a closet. They are not obligated to
share or to provide the digital files in a re-usable format. However, the
digitizing agency cannot dictate any copyright restrictions on the
digitized copies once released to the public.

#iamnotalawyer and welcome correction

best,

Jimmy



On Mon, Jun 1, 2015 at 12:12 PM, Eric Lease Morgan emor...@nd.edu wrote:


On Jun 1, 2015, at 10:58 AM, davesgonechina davesgonech...@gmail.com
wrote:


They just informed me I need a .edu address. Having trouble
understanding the use of the term public domain here.

   Gung fhpx, naq fbhaqf ernyyl fbeg bs fghcvq!! --RYZ




--
Jimmy Ghaphery
Head, Digital Technologies
VCU Libraries
804-827-3551






--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] hathitrust research center workset browser

2015-05-27 Thread Karen Coyle
Eric, what happens if you access this from a non-HT institution? When I 
go to HT I am often unable to download public domain titles because they 
aren't available to members of the general public.


kc

On 5/26/15 8:30 AM, Eric Lease Morgan wrote:

In my copious spare time I have hacked together a thing I’m calling the 
HathiTrust Research Center Workset Browser, a (fledgling) tool for doing 
“distant reading” against corpora from the HathiTrust. [1]

The idea is to: 1) create, refine, or identify a HathiTrust Research Center 
workset of interest — your corpus, 2) feed the workset’s rsync file to the 
Browser, 3) have the Browser download, index, and analyze the corpus, and 4) 
enable to reader to search, browse, and interact with the result of the 
analysis. With varying success, I have done this with a number of worksets 
ranging on topics from literature, philosophy, Rome, and cookery. The best 
working examples are the ones from Thoreau and Austen. [2, 3] The others are 
still buggy.

As a further example, the Browser can/will create reports describing the corpus 
as a whole. This analysis includes the size of a corpus measured in pages as 
well as words, date ranges, word frequencies, and selected items of interest 
based on pre-set “themes” — usage of color words, name of “great” authors, and 
a set of timeless ideas. [4] This report is based on more fundamental reports 
such as frequency tables, a “catalog”, and lists of unique words. [5, 6, 7, 8]

The whole thing is written in a combination of shell and Python scripts. It 
should run on just about any out-of-the-box Linux or Macintosh computer. Take a 
look at the code. [9] No special libraries needed. (“Famous last words.”) In 
its current state, it is very Unix-y. Everything is done from the command line. 
Lot’s of plain text files and the exploitation of STDIN and STDOUT. Like a 
Renaissance cartoon, the Browser, in its current state, is only a sketch. Only 
later will a more full-bodied, Web-based interface be created.

The next steps are numerous and listed in no priority order: putting the whole 
thing on GitHub, outputting the reports in generic formats so other things can 
easily read them, improving the terminal-based search interface, implementing a 
Web-based search interface, writing advanced programs in R that chart and graph 
analysis, provide a means for comparing  contrasting two or more items from a 
corpus, indexing the corpus with a (real) indexer such as Solr, writing a 
“cookbook” describing how to use the browser to to “kewl” things, making the 
metadata of corpora available as Linked Data, etc.

'Want to give it a try? For a limited period of time, go to the HathiTrust 
Research Center Portal, create (refine or identify) a collection of personal 
interest, use the Algorithms tool to export the collection's rsync file, and 
send the file to me. I will feed the rsync file to the Browser, and then send 
you the URL pointing to the results. [10] Let’s see what happens.

Fun with public domain content, text mining, and the definition of 
librarianship.

Links

[1] HTRC Workset Browser - http://bit.ly/workset-browser
[2] Thoreau - http://bit.ly/browser-thoreau
[3] Austen - http://bit.ly/browser-austen
[4] Thoreau report - http://ntrda.me/1LD3xds
[5] Thoreau dictionary (frequency list) - http://bit.ly/thoreau-dictionary
[6] usage of color words in Thoreau — http://bit.ly/thoreau-colors
[7] unique words in the corpus - http://bit.ly/thoreau-unique
[8] Thoreau “catalog” — http://bit.ly/thoreau-catalog
[9] source code - http://ntrda.me/1Q8pPoI
   [10] HathiTrust Research Center - https://sharc.hathitrust.org

—
Eric Lease Morgan, Librarian
University of Notre Dame


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Library Hours

2015-05-06 Thread Karen Coyle
Charlie, I don't know of any libraries that have used schema.org for 
their web site - perhaps others do. If it is used, it should be picked 
up the next time the search engines index the site. What the search 
engines do with schema.org is not guaranteed, but can be observed. It is 
not guaranteed because none of the search engines will say what they do, 
as that is considered a trade secret (especially from each other).


However, as locations and hours are important for their commercial 
customers (stores, restaurants, etc.) I would expect that to be picked 
up as a matter of course. Note that already locations and hours for some 
businesses do show in the search engines, and that is for sites that are 
not yet using schema.org, so the engines have some way of picking that 
up from the HTML. The Google side-bar knowledge graph for my local 
libraries shows  Hours 
https://www.google.com/search?sa=Xbiw=1299bih=561q=san+francisco+public+library+larkin+street+hoursstick=H4sIAGOovnz8BQMDgzYHnxCXfq6-gVlZhbF5sZZ0drKVfk5-cmJJZn4enGGVkV9aVBzLKeznIsHxlTMy2S10V0iJwvZlMgBPWBDOSAei=qhlKVcKWJ8b7oQS65oCQCAved=0CJgBEOgTMBA: 


Open today · 9:00 am – 8:00 pm javascript:void(0)
 but I have no idea where that comes from.

kc

On 5/6/15 5:22 AM, Charlie Morris wrote:

I'm curious, Karen, Ethan or anyone else, do you know of any examples of
libraries that have implemented schema.org or RDFa for hours data and have
noticed that Google or some other search engine has picked it up (i.e.,
correctly displaying that data as part of the search results)?  And if so,
how quickly will Google or the like pickup on changes to hours (i.e.,
shifting between semesters or unplanned changes)?

On Wed, May 6, 2015 at 8:15 AM, Ethan Gruber ewg4x...@gmail.com wrote:


+1 on the RDFa and schema.org. For those that don't know the library URL
off-hand, it is much easier to find a library website by Googling than it
is to go through the central university portal, and the hours will show up
at the top of the page after having been harvested by search engines.

On Tue, May 5, 2015 at 6:54 PM, Karen Coyle li...@kcoyle.net wrote:


Note that library hours is one of the possible bits of information that
could be encoded as RDFa in the library web site, thus making it possible
to derive library hours directly from the listing of hours on the web

site

rather than keeping a separate list. Schema.org does have the elements

such

that hours can be encoded. This would mean that hours could show in the
display of the library's catalog entry on Google, Yahoo and Bing. Being
available directly through the search engines might be sufficient, not
necessitating creating yet-another-database for that data.

Schema.org uses a restaurant as its opening hours example, but much of

the

data would be the same for a library:

div vocab=http://schema.org/; typeof=Restaurant
   span property=nameGreatFood/span
   div property=aggregateRating  typeof=AggregateRating
 span property=ratingValue4/span stars -
 based on span property=reviewCount250/span reviews
   /div
   div property=address  typeof=PostalAddress
 span property=streetAddress1901 Lemur Ave/span
 span property=addressLocalitySunnyvale/span,
 span property=addressRegionCA/span span
property=postalCode94086/span
   /div
   span property=telephone(408) 714-1489/span
   a property=url href=http://www.dishdash.com;www.greatfood.com/a
   Hours:
   meta property=openingHours content=Mo-Sa 11:00-14:30Mon-Sat 11am

-

2:30pm
   meta property=openingHours content=Mo-Th 17:00-21:30Mon-Thu 5pm -
9:30pm
   meta property=openingHours content=Fr-Sa 17:00-22:00Fri-Sat 5pm -
10:00pm
   Categories:
   span property=servesCuisine
 Middle Eastern
   /span,
   span property=servesCuisine
 Mediterranean
   /span
   Price Range: span property=priceRange$$/span
   Takes Reservations: Yes
/div

It seems to me that using schema.org would get more bang for the buck --
it would get into the search engines and could also be aggregated into
whatever database is needed. As we've seen with OCLC, having a separate
listing is likely to mean that the data will be out of date.

kc

On 5/5/15 2:19 PM, nitin arora wrote:


I can't see they distinguished between public libraries and other types

on

their campaign page.

They say  all libraries as far as I can see.
So I suppose then that this is true for all libraries:
Libraries offer a space anyone can enter, where money isn't exchanged,
and
documentation doesn't have to be shown.
Who knew fines and library/student-IDs were a thing of the past?

The only data sets I can find where they got the 17,000 number is for
public libraries:
http://www.imls.gov/research/pls_data_files.aspx
Maybe I missed something.
There is an hours field on one of the CSVs I downloaded, etc for 2012

data

(the most recent I could find).

Asking 10k for something targeted for completion in June and without a
grasp on what types of libraries there are and how volatile the hours

Re: [CODE4LIB] Library Hours

2015-05-06 Thread Karen Coyle
The search engine may not pick it up quickly enough, but the emergency 
services in the area could get it from the RDFa as soon as it hits the web.


kc

On 5/6/15 6:45 AM, nitin arora wrote:

I think both creating a one-off list and schema.org approaches pose
problems within the context of the original fund raising campaign's pitch.
I don't think every library can necessarily implement the latter for a
variety of reasons, not always technical.

 From the pov that a library can be a community center in a time of crisis,
I'm wondering not only how quickly a search engine would pick that up but
also, in such moments, how prioritized updating that data would be in the
first place.

On Wed, May 6, 2015 at 8:22 AM, Charlie Morris cdmorri...@gmail.com wrote:


I'm curious, Karen, Ethan or anyone else, do you know of any examples of
libraries that have implemented schema.org or RDFa for hours data and have
noticed that Google or some other search engine has picked it up (i.e.,
correctly displaying that data as part of the search results)?  And if so,
how quickly will Google or the like pickup on changes to hours (i.e.,
shifting between semesters or unplanned changes)?

On Wed, May 6, 2015 at 8:15 AM, Ethan Gruber ewg4x...@gmail.com wrote:


+1 on the RDFa and schema.org. For those that don't know the library URL
off-hand, it is much easier to find a library website by Googling than it
is to go through the central university portal, and the hours will show

up

at the top of the page after having been harvested by search engines.

On Tue, May 5, 2015 at 6:54 PM, Karen Coyle li...@kcoyle.net wrote:


Note that library hours is one of the possible bits of information that
could be encoded as RDFa in the library web site, thus making it

possible

to derive library hours directly from the listing of hours on the web

site

rather than keeping a separate list. Schema.org does have the elements

such

that hours can be encoded. This would mean that hours could show in the
display of the library's catalog entry on Google, Yahoo and Bing. Being
available directly through the search engines might be sufficient, not
necessitating creating yet-another-database for that data.

Schema.org uses a restaurant as its opening hours example, but much of

the

data would be the same for a library:

div vocab=http://schema.org/; typeof=Restaurant
   span property=nameGreatFood/span
   div property=aggregateRating  typeof=AggregateRating
 span property=ratingValue4/span stars -
 based on span property=reviewCount250/span reviews
   /div
   div property=address  typeof=PostalAddress
 span property=streetAddress1901 Lemur Ave/span
 span property=addressLocalitySunnyvale/span,
 span property=addressRegionCA/span span
property=postalCode94086/span
   /div
   span property=telephone(408) 714-1489/span
   a property=url href=http://www.dishdash.com;www.greatfood.com

/a

   Hours:
   meta property=openingHours content=Mo-Sa 11:00-14:30Mon-Sat

11am

-

2:30pm
   meta property=openingHours content=Mo-Th 17:00-21:30Mon-Thu

5pm -

9:30pm
   meta property=openingHours content=Fr-Sa 17:00-22:00Fri-Sat

5pm -

10:00pm
   Categories:
   span property=servesCuisine
 Middle Eastern
   /span,
   span property=servesCuisine
 Mediterranean
   /span
   Price Range: span property=priceRange$$/span
   Takes Reservations: Yes
/div

It seems to me that using schema.org would get more bang for the buck

--

it would get into the search engines and could also be aggregated into
whatever database is needed. As we've seen with OCLC, having a separate
listing is likely to mean that the data will be out of date.

kc

On 5/5/15 2:19 PM, nitin arora wrote:


I can't see they distinguished between public libraries and other

types

on

their campaign page.

They say  all libraries as far as I can see.
So I suppose then that this is true for all libraries:
Libraries offer a space anyone can enter, where money isn't

exchanged,

and
documentation doesn't have to be shown.
Who knew fines and library/student-IDs were a thing of the past?

The only data sets I can find where they got the 17,000 number is for
public libraries:
http://www.imls.gov/research/pls_data_files.aspx
Maybe I missed something.
There is an hours field on one of the CSVs I downloaded, etc for 2012

data

(the most recent I could find).

Asking 10k for something targeted for completion in June and without a
grasp on what types of libraries there are and how volatile the hours
information is (especially in crisis) ...
Sounds naive at best, sketchy at worst.

The flexible funding button says this campaign will receive all

funds

raised even if it does not reach its goals.

The value of these places for youth cannot be underestimated.
So is the value of a quick buck ...

On Tue, May 5, 2015 at 4:53 PM, McCanna, Terran 
tmcca...@georgialibraries.org wrote:

  I'm not at all surprised that this doesn't already exist, and even if

OCLC's was available, I'd be willing to bet

Re: [CODE4LIB] Library Hours

2015-05-06 Thread Karen Coyle
Right, but I don't think that meets any particular standard, which means 
that Google is doing a lot of text analysis when it indexes pages, 
looking for a pattern that looks like opening hours. That takes more 
cycles than having it all neatly wrapped in some known RDFa.


kc

On 5/6/15 6:54 AM, Tajoli Zeno wrote:

Hi

Open today · 9:00 am – 8:00 pm javascript:void(0)
 but I have no idea where that comes from.


probably because the web page http://sfpl.org/index.php?pg=010101
insert library hours inside

div id=library-hours /div

Bye
Zeno Tajoli



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Library Hours

2015-05-06 Thread Karen Coyle
Tom, Google will not tell you. The entirety of how Google search works 
is a trade secret. We don't know the algorithm for ranking, and we don't 
know what information they glean from web pages -- and they are unlikely 
to tell. It is a constant on the schema.org discussion list that 
developers want to know what Google/Bing/Yahoo/Yandex will do with 
specific information in the web pages, and it is a constant that the 
reps there reply: we cannot tell you that. The only way to find out is 
to code and observe.


kc

On 5/6/15 7:00 AM, Tom Keays wrote:

I'd like to find out how and why Google is parsing this information. If you
go to the the SFPL hours page (first link in the Google results), and look
at the source code, this is all you find.
http://sfpl.org/index.php?pg=010101
Is the ID in the DIV sufficient?  It would be nice to have a set of use
cases to work from.

Currently, I'm generating a weekly hours box by pulling JSONP from the
hours API of LibCal. I could easily output this in schema.org format (and
probably will now), but can Google pick up the information from the DOM if
it is delivered as JSON and transformed into HTML?

div id=library-hours
   h2Hours/h2
   table class=hours cellpadding=0 cellspacing=0
 tr
   thSun/th
   thMon/th
   thTue/th
   th class=todayWed/th
   thThu/th
   thFri/th
   thSat/th
 /tr
 tr
   td12-5/td
   td10-6/td
   td9-8/td
   td class=today9-8/td
   td9-8/td
   td12-6/td
   td10-6/td
 /tr
   /table
/div


On Wed, May 6, 2015 at 9:47 AM, Karen Coyle li...@kcoyle.net wrote:


Charlie, I don't know of any libraries that have used schema.org for
their web site - perhaps others do. If it is used, it should be picked up
the next time the search engines index the site. What the search engines do
with schema.org is not guaranteed, but can be observed. It is not
guaranteed because none of the search engines will say what they do, as
that is considered a trade secret (especially from each other).

However, as locations and hours are important for their commercial
customers (stores, restaurants, etc.) I would expect that to be picked up
as a matter of course. Note that already locations and hours for some
businesses do show in the search engines, and that is for sites that are
not yet using schema.org, so the engines have some way of picking that up
from the HTML. The Google side-bar knowledge graph for my local libraries
shows  Hours 
https://www.google.com/search?sa=Xbiw=1299bih=561q=san+francisco+public+library+larkin+street+hoursstick=H4sIAGOovnz8BQMDgzYHnxCXfq6-gVlZhbF5sZZ0drKVfk5-cmJJZn4enGGVkV9aVBzLKeznIsHxlTMy2S10V0iJwvZlMgBPWBDOSAei=qhlKVcKWJ8b7oQS65oCQCAved=0CJgBEOgTMBA:

Open today · 9:00 am – 8:00 pm javascript:void(0)
 but I have no idea where that comes from.

kc


On 5/6/15 5:22 AM, Charlie Morris wrote:


I'm curious, Karen, Ethan or anyone else, do you know of any examples of
libraries that have implemented schema.org or RDFa for hours data and
have
noticed that Google or some other search engine has picked it up (i.e.,
correctly displaying that data as part of the search results)?  And if so,
how quickly will Google or the like pickup on changes to hours (i.e.,
shifting between semesters or unplanned changes)?

On Wed, May 6, 2015 at 8:15 AM, Ethan Gruber ewg4x...@gmail.com wrote:

  +1 on the RDFa and schema.org. For those that don't know the library URL

off-hand, it is much easier to find a library website by Googling than it
is to go through the central university portal, and the hours will show
up
at the top of the page after having been harvested by search engines.

On Tue, May 5, 2015 at 6:54 PM, Karen Coyle li...@kcoyle.net wrote:

  Note that library hours is one of the possible bits of information that

could be encoded as RDFa in the library web site, thus making it
possible
to derive library hours directly from the listing of hours on the web


site


rather than keeping a separate list. Schema.org does have the elements


such


that hours can be encoded. This would mean that hours could show in the
display of the library's catalog entry on Google, Yahoo and Bing. Being
available directly through the search engines might be sufficient, not
necessitating creating yet-another-database for that data.

Schema.org uses a restaurant as its opening hours example, but much of


the


data would be the same for a library:

div vocab=http://schema.org/; typeof=Restaurant
span property=nameGreatFood/span
div property=aggregateRating  typeof=AggregateRating
  span property=ratingValue4/span stars -
  based on span property=reviewCount250/span reviews
/div
div property=address  typeof=PostalAddress
  span property=streetAddress1901 Lemur Ave/span
  span property=addressLocalitySunnyvale/span,
  span property=addressRegionCA/span span
property=postalCode94086/span
/div
span property=telephone(408) 714-1489/span
a property=url

Re: [CODE4LIB] Library Hours

2015-05-06 Thread Karen Coyle
Yes, it definitely does. Which actually is a problem for Wikipedia 
because it encourages people/companies to try to get entries into WP for 
SEO purposes and so that the sidebox will show up. I spend a lot of time 
on the articles for deletion pages of WP trying to get these 
promotional pages out of the encyclopedia. A big success is when I see 
them disappear from search results. (BTW, the various ways that 
self-published authors of written crap game the system is truly 
astonishing. A+ for effort, and their skill in PR is way beyond their 
literary skills.)


kc

On 5/6/15 8:33 AM, Bigwood, David wrote:

I have heard that at least part of the sidebox is constructed using data from 
Wikipedia, especially the structured info in the infobox there.

Dave

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Karen 
Coyle
Sent: Wednesday, May 06, 2015 9:21 AM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Library Hours

Tom, Google will not tell you. The entirety of how Google search works is a 
trade secret. We don't know the algorithm for ranking, and we don't know what 
information they glean from web pages -- and they are unlikely to tell. It is a 
constant on the schema.org discussion list that developers want to know what 
Google/Bing/Yahoo/Yandex will do with specific information in the web pages, 
and it is a constant that the reps there reply: we cannot tell you that. The 
only way to find out is to code and observe.

kc


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Library Hours

2015-05-06 Thread Karen Coyle
I generally find that Bing makes better use of RDFa/schema.org than 
Google does.


kc

On 5/6/15 7:33 AM, Megan O'Neill Kudzia wrote:

Hi all,

I've been experimenting with schema.org OpeningHoursSpecification, and
currently Bing is scraping our hours, but Google isn't. I am using
RDFa-lite and I've validated it using a linter (thanks Jason Ronallo!), so
I'm scratching my head as to why our hours *still* don't show up on a
google search.

I suspect part of it for us might be that we're re-branding away from
Stockwell-Mudd Libraries to Albion College Library, as it's much more
explanatory, but neither search through Google yields a nice box with hours
in it like the SFPL.

If and when I figure out the problem I'd be happy to send you an update of
what we did and what caused it to finally work properly.

On Wed, May 6, 2015 at 10:21 AM, Karen Coyle li...@kcoyle.net wrote:


Tom, Google will not tell you. The entirety of how Google search works is
a trade secret. We don't know the algorithm for ranking, and we don't know
what information they glean from web pages -- and they are unlikely to
tell. It is a constant on the schema.org discussion list that developers
want to know what Google/Bing/Yahoo/Yandex will do with specific
information in the web pages, and it is a constant that the reps there
reply: we cannot tell you that. The only way to find out is to code and
observe.

kc


On 5/6/15 7:00 AM, Tom Keays wrote:


I'd like to find out how and why Google is parsing this information. If
you
go to the the SFPL hours page (first link in the Google results), and look
at the source code, this is all you find.
http://sfpl.org/index.php?pg=010101
Is the ID in the DIV sufficient?  It would be nice to have a set of use
cases to work from.

Currently, I'm generating a weekly hours box by pulling JSONP from the
hours API of LibCal. I could easily output this in schema.org format (and
probably will now), but can Google pick up the information from the DOM if
it is delivered as JSON and transformed into HTML?

div id=library-hours
h2Hours/h2
table class=hours cellpadding=0 cellspacing=0
  tr
thSun/th
thMon/th
thTue/th
th class=todayWed/th
thThu/th
thFri/th
thSat/th
  /tr
  tr
td12-5/td
td10-6/td
td9-8/td
td class=today9-8/td
td9-8/td
td12-6/td
td10-6/td
  /tr
/table
/div


On Wed, May 6, 2015 at 9:47 AM, Karen Coyle li...@kcoyle.net wrote:

  Charlie, I don't know of any libraries that have used schema.org for

their web site - perhaps others do. If it is used, it should be picked
up
the next time the search engines index the site. What the search engines
do
with schema.org is not guaranteed, but can be observed. It is not
guaranteed because none of the search engines will say what they do, as
that is considered a trade secret (especially from each other).

However, as locations and hours are important for their commercial
customers (stores, restaurants, etc.) I would expect that to be picked up
as a matter of course. Note that already locations and hours for some
businesses do show in the search engines, and that is for sites that are
not yet using schema.org, so the engines have some way of picking that
up
from the HTML. The Google side-bar knowledge graph for my local
libraries
shows  Hours 

https://www.google.com/search?sa=Xbiw=1299bih=561q=san+francisco+public+library+larkin+street+hoursstick=H4sIAGOovnz8BQMDgzYHnxCXfq6-gVlZhbF5sZZ0drKVfk5-cmJJZn4enGGVkV9aVBzLKeznIsHxlTMy2S10V0iJwvZlMgBPWBDOSAei=qhlKVcKWJ8b7oQS65oCQCAved=0CJgBEOgTMBA

:

Open today · 9:00 am – 8:00 pm javascript:void(0)
 but I have no idea where that comes from.

kc


On 5/6/15 5:22 AM, Charlie Morris wrote:

  I'm curious, Karen, Ethan or anyone else, do you know of any examples of

libraries that have implemented schema.org or RDFa for hours data and
have
noticed that Google or some other search engine has picked it up (i.e.,
correctly displaying that data as part of the search results)?  And if
so,
how quickly will Google or the like pickup on changes to hours (i.e.,
shifting between semesters or unplanned changes)?

On Wed, May 6, 2015 at 8:15 AM, Ethan Gruber ewg4x...@gmail.com
wrote:

   +1 on the RDFa and schema.org. For those that don't know the library
URL


off-hand, it is much easier to find a library website by Googling than
it
is to go through the central university portal, and the hours will show
up
at the top of the page after having been harvested by search engines.

On Tue, May 5, 2015 at 6:54 PM, Karen Coyle li...@kcoyle.net wrote:

   Note that library hours is one of the possible bits of information
that


could be encoded as RDFa in the library web site, thus making it
possible
to derive library hours directly from the listing of hours on the web

  site

  rather than keeping a separate list. Schema.org does have the elements

  such

  that hours can be encoded

Re: [CODE4LIB] Library Hours

2015-05-05 Thread Karen Coyle
 and hours of operation for all libraries
in the US. We are going to track down the hours of operation for all 17,000
libraries and make that information available -- in Range and for other
developers who may want to use it.
https://www.indiegogo.com/projects/range-food-and-safe-places-for-youth

Are the hours of public libraries really not available?

Sincerely,
David Bigwood
dbigw...@gmail.commailto:dbigw...@gmail.com
Lunar and Planetary Institute
@LPI_Library
https://www.flickr.com/photos/lunarandplanetaryinstitute/





--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Protagonists

2015-04-14 Thread Karen Coyle

On 4/14/15 7:18 AM, Joel Marchesoni wrote:

  It is a little disappointing that I couldn't find the IMDB of the literary 
world.


Which reflects an interesting difference between books and movies -- 
movies have credits that list characters and actors. There are a few 
books that include character lists, but not many. You can get a list of 
Harry Potter characters from the movie because that list exists in the 
movie. The movie credits can also be considered authoritative in the 
same way that the title page of a book with the title and author is. Any 
list of characters that is not from the book itself has something of an 
authority problem.


kc

--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Native MarcEdit for MacOSX

2015-04-06 Thread Karen Coyle
I wish I could teleport my rarely used iMac, but in keeping with current 
technology capabilities, I'll just donate digital monetary value. This 
is definitely a worthy cause.


kc

On 4/6/15 12:31 PM, Francis Kayiwa wrote:
Terry Reese (thanks Terry if you are reading this) has offered to 
write a Object-C version of MarcEdit. In order to this he needed 
access to a Apple Hardware. While my initial proposal on Go Fund me 
below was for a Macbook Pro, we've since realized it need not be a 
portable device. My current arithmetic puts the price of this at 
~US$1300 as opposed to the listed ~US$2400



Thanks for boosting this signal if you cannot otherwise help fund 
porting MarcEdit for the Mac OSX platform.


http://www.gofundme.com/qtbzq4


Cheers,
./fxk


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Native MarcEdit for MacOSX

2015-04-06 Thread Karen Coyle
Have you had one? I started out with one a while back because it was 
indeed the cheapest Mac. It seemed underpowered at the time. Plus you 
have to add a screen to it, and a keyboard, and a mouse so the price 
goes up. On the plus size, it doesn't take up much room... without the 
screen, the keyboard...


kc

On 4/6/15 2:36 PM, Roy Tennant wrote:

Better yet, get a brand-new Mac mini for what you've already raised:

http://store.apple.com/us/buy-mac/mac-mini

Roy

On Mon, Apr 6, 2015 at 1:18 PM, Cary Gordon listu...@chillco.com wrote:


You might want to consider picking up a MacBook on Craigslist, or
something like that. You can get a lot of only slightly vintage computer
for $5-700.

Cary

Random ad:

13 MacBook ProTurbo 2.9GHz i5 Dual-core ϟ Thunderbolt! + 500GB Storage -
$499 (Los Angeles)
make / manufacturer: Apple MacBook Pro
size / dimensions: 13
Up for sale my 13-inch Macbook Pro with free $40 case
2.3GHz Turbo 2.9Ghz
500GB Storage
4GB Ram DDR3
Thunderbolt
USB
FireWire


On Apr 6, 2015, at 12:31 PM, Francis Kayiwa kay...@pobox.com wrote:

Terry Reese (thanks Terry if you are reading this) has offered to write

a Object-C version of MarcEdit. In order to this he needed access to a
Apple Hardware. While my initial proposal on Go Fund me below was for a
Macbook Pro, we've since realized it need not be a portable device. My
current arithmetic puts the price of this at ~US$1300 as opposed to the
listed ~US$2400


Thanks for boosting this signal if you cannot otherwise help fund

porting MarcEdit for the Mac OSX platform.

http://www.gofundme.com/qtbzq4


Cheers,
./fxk
--
Your analyst has you mixed up with another patient.  Don't believe a
thing he tells you.


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] PBCore RDF Ontology Hackathon Wiki page

2015-01-05 Thread Karen Coyle

Casey, I have some suggestions:

1) Everyone should read at least the first chapters of the Allemang 
book, Semantic Web for the Working Ontologist:

http://www.worldcat.org/title/semantic-web-for-the-working-ontologist-effective-modeling-in-rdfs-and-owl/oclc/73393667

2) Everyone should understand the RDF meaning of classes, properties, 
domain and range before beginning. (cf: 
http://kcoyle.blogspot.com/2014/11/classes-in-rdf.html)


3) Don't lean too heavily on Protege. Protege is very OWL-oriented and 
can lead one far astray. It's easy to click on check boxes without 
knowing what they really mean. Do as much development as you can without 
using Protege, and do your development in RDFS not OWL. Later you can 
use Protege to check your work, or to complete the code.


4) Develop in ntriples or turtle but NOT rdf/xml. RDF differs from XML 
in some fundamental ways that are not obvious, and developing in rdf/xml 
masks these differences and often leads to the development of not very 
good ontologies.


kc


On 1/5/15 9:35 AM, Casey Davis wrote:

Hi Code4Libbers,

In case you are interested in attending or staying in the know about the PBCore RDF 
ontology hackathon happening on February 7  8 before Code4Lib, check out the 
Wiki page: http://wiki.code4lib.org/PBCore_RDF_Hackathon. This is also where we 
will provide links to all documentation created during the hackathon. In case you 
are unable to join us, we will be using the hashtag #PBCoreRDF15 on the days of the 
event.

Best,
Casey

Casey E. Davis, MLIS | Project Manager, American Archive of Public Broadcasting
WGBH Media Library and Archives | WGBH Educational Foundation
casey_da...@wgbh.orgmailto:casey_da...@wgbh.org | 617-300-5921 | One Guest 
Street | Boston, MA 02135

Subscribe to the American Archive bloghttp://americanarchivepb.wordpress.com/
Follow the #AmericanArchive @amarchivepub


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] linked data and open access

2014-12-23 Thread Karen Coyle

Off the top of my head:

http://www.epsiplatform.eu/content/what-linked-open-government-data
http://aims.fao.org/agris
http://data.gov.uk/location
http://datos.bne.es/
http://statistics.data.gov.uk/
http://europeana.eu/
etc.

What linked and open provide is exactly what it says - linked=able 
to be used in combination with data from other Web resources; 
open=anyone can use the data. There are projects that are using CSV or 
XSL files, but those function as self-contained bits of data, without 
the linking, even if they are openly available.


kc

On 12/22/14 7:30 PM, Jonathan Rochkind wrote:

And as has already been pointed out, no one has really show an impressive end 
user use for linked data, which American decision making tends to be more 
driven by.

Well, that raises an important question -- whether an 'end user use', or other 
use, do people have examples of neat/important/useful things done with linked 
data in Europe, especially that would have been harder or less likely without 
the data being modelled/distributed as linked data?


From: Code for Libraries [CODE4LIB@LISTSERV.ND.EDU] on behalf of Brent Hanner 
[behan...@mediumaevum.com]
Sent: Monday, December 22, 2014 6:11 PM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] linked data and open access

There are deeper issues at work here than just the kind of obvious surface 
issues.

One of the reason Europe embraced rdf triples and linked data was timing.  The 
EU was forming its centralized information institutions the same time the idea 
of linked data to solve certain problem came about.  So they took it and ran 
with it.  In the US we have been primarily driven by the big data movement that 
gained steam shortly after.  And as has already been pointed out, no one has 
really show an impressive end user use for linked data, which American decision 
making tends to be more driven by.


Europeans can think about data and databases differently than we can here in 
the US.  In Europe a database is intellectual property, in the US only parts of 
the database that fall under copyright law are intellectual property, which for 
most databases isn't much.  You can’t copyright a fact.  So in the US once you 
release the data into the wild its usually public domain.


As for government data, the Federal and most state governments are in need of 
an overhaul that would make it possible.  If you don’t have the systems or 
people in place who can make it happen it won’t happen.  Heck the federal 
government can’t even get a single set of accounting software and what not.


So it isn’t just a lack of leadership or will, there are other things at work 
as well.



Brent






Sent from Windows Mail





From: Karen Coyle
Sent: ‎Friday‎, ‎December‎ ‎19‎, ‎2014 ‎10‎:‎32‎ ‎AM
To: CODE4LIB@LISTSERV.ND.EDU





Yep, yep, and yep.

Plus I'd add that the lack of centralization of library direction (read:
states) is also a hindrance here. Having national leadership would be
great. Being smaller also wouldn't hurt.

kc

On 12/19/14 6:48 AM, Eric Lease Morgan wrote:

I don’t know about y’all, but it seems to me that things like linked data and 
open access are larger trends in Europe than here in the United States. Is 
there are larger commitment to sharing in Europe when compared to the United 
States? If so, is this a factor based on the nonexistence of a national library 
in the United States? Is this your perception too? —Eric Morgan

--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] linked data and open access

2014-12-19 Thread Karen Coyle

Yep, yep, and yep.

Plus I'd add that the lack of centralization of library direction (read: 
states) is also a hindrance here. Having national leadership would be 
great. Being smaller also wouldn't hurt.


kc

On 12/19/14 6:48 AM, Eric Lease Morgan wrote:

I don’t know about y’all, but it seems to me that things like linked data and 
open access are larger trends in Europe than here in the United States. Is 
there are larger commitment to sharing in Europe when compared to the United 
States? If so, is this a factor based on the nonexistence of a national library 
in the United States? Is this your perception too? —Eric Morgan


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Wednesday afternoon reverie

2014-11-06 Thread Karen Coyle

Cynthia, it's been a while but I wanted to give you feedback...

Ranking on importance based on library ownership and/or circulation is 
something that I've seen discussed but not implemented -- mainly due to 
the difficulty of gathering the data from library systems. But it seems 
like an obvious way to rank results, IMO.


Too bad that one has to pay for BISAC headings. They tend to mirror the 
headings in bookstores (and ebook stores) that people might be familiar 
with. They capture fiction topics, especially, in a way to resonates 
with some users (topics like Teen Paranormal Romance).


kc

On 10/22/14 1:25 PM, Harper, Cynthia wrote:

So I'm deleting all the Bisac subject headings (650_7|2bisacsh) from our ebook 
records - they were deemed not to be useful, especially as it would entail a 
for-fee indexing change to make them clickable.  But I'm thinking if we someday 
have a discovery system, they'll be useful as a means for broader-to-narrower 
term browsing that won't require translation to English, as would call number 
ranges.

As I watch the system slowly chunk through them, I think about how library 
collections and catalogs facilitate jumping to the most specific subjects, but 
browsing is something of an afterthought.

What if we could set a ranking score for the importance of an item in browsing, based on 
circulation data - authors ranked by the relative circulation of all their works, same for series, latest 
edition of a multi-edition work given higher ranking, etc.?  Then have a means to set the threshold 
importance value you want to look at, and browse through these general Bisac terms, or the classification?  
Or have a facet for importance threshold.  I see Bisac sometimes has a broadness/narrowness facet 
(overview) - wonder how consistently that's applied, enough to be useful?

Guess those rankings would be very expensive in compute time.

Well, back to the deletions.

Cindy Harper
Electronic Services and Serials Librarian
Virginia Theological Seminary
3737 Seminary Road
Alexandria VA 22304
703-461-1794
char...@vts.edu


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] code4lib stickers

2014-11-05 Thread Karen Coyle
and for those of us with mushy brains who can't remember if they 
responded previously? ... kc


On 11/4/14 4:29 PM, Riley Childs wrote:

This is a reminder to fill out the code4lib stickers response form if you are 
interested (yeah, I haven't given up yet...), I am waiting to order until I 
have 50 people on the list (or enough where I can be sure I won't be left with 
leftovers). The google form is here: 
https://docs.google.com/a/tfsgeo.com/forms/d/1k-bQVSduKyOVMkXpJ_xOwk9SDjjEoX7QnQ4JTyp2BqI/viewform
I am leaning towards a diecut sticker depending on further response. Filling 
out this form is not a commitment to purchase. Stickers will be printed by 
StickerMule and will be on demand following the initial batch. Those who 
purchase from the first batch will receive a 30% or more (depending on number 
of people who order) discount over the standard price.

Thanks
//Riley

--
Riley Childs
Senior
Charlotte United Christian Academy
IT Services Administrator
Library Services Administrator
https://rileychilds.net
cell: +1 (704) 497-2086
office: +1 (704) 537-0331x101
twitter: @rowdychildren
Checkout our new Online Library Catalog: https://catalog.cucawarriors.com

Proudly sent in plain text


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Official #teamharpy Statement on the case of Joseph Hawley Murphy vs. nina de jesus and Lisa Rabey

2014-09-21 Thread Karen Coyle
Ed - I don't know why, but that sounds quite defensive. If your CV 
emphasizes speaking engagements and writings, then what you have written 
is relevant. But please look at the notability suggestions for Wikipedia 
for people and for academics before focusing on that one piece of 
information. I believe that I carefully explained that this was only one 
piece of information, that it wasn't absolute, and that notability is 
complex. Sheesh.


kc


On 9/20/14, 4:21 PM, Edward M. Corrado wrote:

I don't know anything about the lawsuit or what has transpired to cause it,
but since when does an H-index score make one a notable librarian? Many
notable librarians don't publish anything at all.

Edward

On Sat, Sep 20, 2014 at 2:12 PM, Karen Coyle li...@kcoyle.net wrote:


Lisa, I hadn't know about this so I just spend some time reading the items
you list below. I was primarily motivated to do so because I had never
heard of this famous librarian, Joe Murphy. (I must be in a different
conference circuit than he is.)

I also was interested because I've recently joined the hardworking group
of Wikipedians who work to distinguish between notable persons and able
self-promoters. In doing so, I've learned a lot about how self-promotion
works, especially in social media. In Wikipedia, to be considered notable,
there needs to be some reliable proof - that is, third-party references,
not provided by the individual in question. In terms of accomplishments,
for example for academics, there is a list of measures, albeit not
measurable in the scientific sense. [1]

Just for a lark, look at the Google scholar profiles for Joe Murphy, RoyT,
and for myself:

http://scholar.google.com/citations?user=zW1lb04Jhl=enoi=ao
http://scholar.google.com/citations?user=LJw73cAJhl=en
http://scholar.google.com/citations?user=m4Tx73QJhl=enoi=ao

The h-index, while imprecise, is about as close as you get to something
one can cite as a measure. It's not a decision, but it is an indication.

I put this forward not as proof of anything, but to offer that reputation
is extremely hard to quantify, but should be looked at with a critical eye
and not taken for granted. It also fits in with what we already know, which
is that men promote themselves in the workplace more aggressively than
women do. In fact, in the Wikipedia group, we mainly find articles about
men whose notability is over-stated. (You can see my blog post on the
problems of notability for women. [2])

I greatly admire your stand for free speech. Beyond this, I will contact
you offline with other thoughts.

kc
[1] http://en.wikipedia.org/wiki/Wikipedia:Notability_%28academics%29
[2] http://kcoyle.blogspot.com/2014/09/wpnotability-and-women.html


On 9/20/14, 9:16 AM, Lisa Rabey wrote:


Friends:


I know many of you have already been boosting the signal, and we thank
you profusely for the help.

For those who do not know, Joe Murphy is currently suing nina and I in
$1.25M defamation case because

  From our official statement
(http://teamharpy.wordpress.com/why-are-we-being-sued/)

Mr. Murphy claims that Ms. Rabey “posted the following false,
libelous and highly damaging tweet accusing the plaintiff of being a
‘sexual predator'”3. He further claims that Ms. de jesus wrote a blog
post that “makes additional false, libelous, highly damaging,
outrageous, malicious statements against the plaintiff alleging the
commission of sexual harassment and sexual abuse of women and other
forms of criminal and unlawful behaviour”4.

Both Ms. Rabey and Ms. de jesus maintain that our comments are fair
and are truthful, which we intend to establish in our defense. Neither
of us made the claims maliciously nor with any intent to damage Mr.
Murphy’s reputation.

Right now we need the following most importantly:

1. We have a call out for additional witnesses
(http://teamharpy.wordpress.com/call-for-witnesses/), which have
started to filter in more accounts of harrassment. Please, PLEASE, if
you know/seen/heard anything about the plaintiff, or know someone who
might -- please have them get in touch.

2. Share our site (http://teamharpy.wordpress.com) which includes
details of the case and updates. Please help us get the word out to as
many people as possible about the plaintiff's attempt to silence those
speaking up against sexual harassment and why you won't stand for it.

3.
onations: Many, many of you have asked to help donate to fund our
mounting legal costs. We will have a donation page up soon. Even if
you cannot help financially, please share across your social networks.

We will not be silenced. We will not be shamed.

Thank you again. The outpouring of support that has been happening has
made this all very much worth while.

Best,
Lisa








--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Official #teamharpy Statement on the case of Joseph Hawley Murphy vs. nina de jesus and Lisa Rabey

2014-09-20 Thread Karen Coyle
Lisa, I hadn't know about this so I just spend some time reading the 
items you list below. I was primarily motivated to do so because I had 
never heard of this famous librarian, Joe Murphy. (I must be in a 
different conference circuit than he is.)


I also was interested because I've recently joined the hardworking group 
of Wikipedians who work to distinguish between notable persons and able 
self-promoters. In doing so, I've learned a lot about how self-promotion 
works, especially in social media. In Wikipedia, to be considered 
notable, there needs to be some reliable proof - that is, third-party 
references, not provided by the individual in question. In terms of 
accomplishments, for example for academics, there is a list of 
measures, albeit not measurable in the scientific sense. [1]


Just for a lark, look at the Google scholar profiles for Joe Murphy, 
RoyT, and for myself:


http://scholar.google.com/citations?user=zW1lb04Jhl=enoi=ao
http://scholar.google.com/citations?user=LJw73cAJhl=en
http://scholar.google.com/citations?user=m4Tx73QJhl=enoi=ao

The h-index, while imprecise, is about as close as you get to 
something one can cite as a measure. It's not a decision, but it is an 
indication.


I put this forward not as proof of anything, but to offer that 
reputation is extremely hard to quantify, but should be looked at with a 
critical eye and not taken for granted. It also fits in with what we 
already know, which is that men promote themselves in the workplace more 
aggressively than women do. In fact, in the Wikipedia group, we mainly 
find articles about men whose notability is over-stated. (You can see my 
blog post on the problems of notability for women. [2])


I greatly admire your stand for free speech. Beyond this, I will contact 
you offline with other thoughts.


kc
[1] http://en.wikipedia.org/wiki/Wikipedia:Notability_%28academics%29
[2] http://kcoyle.blogspot.com/2014/09/wpnotability-and-women.html

On 9/20/14, 9:16 AM, Lisa Rabey wrote:

Friends:


I know many of you have already been boosting the signal, and we thank
you profusely for the help.

For those who do not know, Joe Murphy is currently suing nina and I in
$1.25M defamation case because

 From our official statement
(http://teamharpy.wordpress.com/why-are-we-being-sued/)

Mr. Murphy claims that Ms. Rabey “posted the following false,
libelous and highly damaging tweet accusing the plaintiff of being a
‘sexual predator'”3. He further claims that Ms. de jesus wrote a blog
post that “makes additional false, libelous, highly damaging,
outrageous, malicious statements against the plaintiff alleging the
commission of sexual harassment and sexual abuse of women and other
forms of criminal and unlawful behaviour”4.

Both Ms. Rabey and Ms. de jesus maintain that our comments are fair
and are truthful, which we intend to establish in our defense. Neither
of us made the claims maliciously nor with any intent to damage Mr.
Murphy’s reputation.

Right now we need the following most importantly:

1. We have a call out for additional witnesses
(http://teamharpy.wordpress.com/call-for-witnesses/), which have
started to filter in more accounts of harrassment. Please, PLEASE, if
you know/seen/heard anything about the plaintiff, or know someone who
might -- please have them get in touch.

2. Share our site (http://teamharpy.wordpress.com) which includes
details of the case and updates. Please help us get the word out to as
many people as possible about the plaintiff's attempt to silence those
speaking up against sexual harassment and why you won't stand for it.

3.
onations: Many, many of you have asked to help donate to fund our
mounting legal costs. We will have a donation page up soon. Even if
you cannot help financially, please share across your social networks.

We will not be silenced. We will not be shamed.

Thank you again. The outpouring of support that has been happening has
made this all very much worth while.

Best,
Lisa








--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Official #teamharpy Statement on the case of Joseph Hawley Murphy vs. nina de jesus and Lisa Rabey

2014-09-20 Thread Karen Coyle
Roy, this is pretty much in line with what I said, below: shameless 
self-promotion can be seen as a plus or at least not a minus for men, 
but few women would have replied the way you did. It does make you 
wonder how much more visible women would be if self-promotion were more 
culturally acceptable for them. Again, I refer to my blog post, which 
you may not have had time to read before posting.


kc

On 9/20/14, 12:41 PM, Roy Tennant wrote:

Just a minor point. Despite my relative index score it doesn't necessarily mean 
that I am not a shameless self-promoter, it may just mean I've been at it 
longer. Just sayin'.
Roy


On Sep 20, 2014, at 2:12 PM, Karen Coyle li...@kcoyle.net wrote:

Lisa, I hadn't know about this so I just spend some time reading the items you list 
below. I was primarily motivated to do so because I had never heard of this 
famous librarian, Joe Murphy. (I must be in a different conference circuit 
than he is.)

I also was interested because I've recently joined the hardworking group of Wikipedians 
who work to distinguish between notable persons and able self-promoters. In doing so, 
I've learned a lot about how self-promotion works, especially in social media. In 
Wikipedia, to be considered notable, there needs to be some reliable proof - that is, 
third-party references, not provided by the individual in question. In terms of 
accomplishments, for example for academics, there is a list of measures, 
albeit not measurable in the scientific sense. [1]

Just for a lark, look at the Google scholar profiles for Joe Murphy, RoyT, and 
for myself:

http://scholar.google.com/citations?user=zW1lb04Jhl=enoi=ao
http://scholar.google.com/citations?user=LJw73cAJhl=en
http://scholar.google.com/citations?user=m4Tx73QJhl=enoi=ao

The h-index, while imprecise, is about as close as you get to something one 
can cite as a measure. It's not a decision, but it is an indication.

I put this forward not as proof of anything, but to offer that reputation is 
extremely hard to quantify, but should be looked at with a critical eye and not 
taken for granted. It also fits in with what we already know, which is that men 
promote themselves in the workplace more aggressively than women do. In fact, 
in the Wikipedia group, we mainly find articles about men whose notability is 
over-stated. (You can see my blog post on the problems of notability for women. 
[2])

I greatly admire your stand for free speech. Beyond this, I will contact you 
offline with other thoughts.

kc
[1] http://en.wikipedia.org/wiki/Wikipedia:Notability_%28academics%29
[2] http://kcoyle.blogspot.com/2014/09/wpnotability-and-women.html


On 9/20/14, 9:16 AM, Lisa Rabey wrote:
Friends:


I know many of you have already been boosting the signal, and we thank
you profusely for the help.

For those who do not know, Joe Murphy is currently suing nina and I in
$1.25M defamation case because

 From our official statement
(http://teamharpy.wordpress.com/why-are-we-being-sued/)

Mr. Murphy claims that Ms. Rabey “posted the following false,
libelous and highly damaging tweet accusing the plaintiff of being a
‘sexual predator'”3. He further claims that Ms. de jesus wrote a blog
post that “makes additional false, libelous, highly damaging,
outrageous, malicious statements against the plaintiff alleging the
commission of sexual harassment and sexual abuse of women and other
forms of criminal and unlawful behaviour”4.

Both Ms. Rabey and Ms. de jesus maintain that our comments are fair
and are truthful, which we intend to establish in our defense. Neither
of us made the claims maliciously nor with any intent to damage Mr.
Murphy’s reputation.

Right now we need the following most importantly:

1. We have a call out for additional witnesses
(http://teamharpy.wordpress.com/call-for-witnesses/), which have
started to filter in more accounts of harrassment. Please, PLEASE, if
you know/seen/heard anything about the plaintiff, or know someone who
might -- please have them get in touch.

2. Share our site (http://teamharpy.wordpress.com) which includes
details of the case and updates. Please help us get the word out to as
many people as possible about the plaintiff's attempt to silence those
speaking up against sexual harassment and why you won't stand for it.

3.
onations: Many, many of you have asked to help donate to fund our
mounting legal costs. We will have a donation page up soon. Even if
you cannot help financially, please share across your social networks.

We will not be silenced. We will not be shamed.

Thank you again. The outpouring of support that has been happening has
made this all very much worth while.

Best,
Lisa

--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Privacy audits Re: [CODE4LIB] Library Privacy, RIP

2014-08-20 Thread Karen Coyle
Thanks, Bill. I actually didn't see this (must be a tweeter I don't 
follow? Or I'm running behind).


I'll see what I can add.

kc

On 8/20/14, 8:29 AM, William Denton wrote:

On 16 August 2014, Karen Coyle wrote:

I think a Code4lib guide to library privacy or something of that 
nature would be a valuable contribution. I'd be happy to work with 
folks on it.


I saw on Twitter that this is being built up:

  https://www.piratepad.ca/p/Library_Privacy_Audit

I'm sure many of you knew about it but thought I'd mention it.

Bill


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] metadata for free ebook repositories

2014-08-19 Thread Karen Coyle
Authors in OL have already been linked to Wikipedia, and Wikipedia has 
been linked to VIAF, and the OCLC number, when present, has been taken 
from the MARC record. Therefore the OL record in some cases already has 
these connections.


kc


On 8/18/14, 8:21 PM, Stuart Yeates wrote:

I think what I'm looking for is a crowd-sourcing platform to add:

https://en.wikipedia.org/wiki/Willa_Cather
http://viaf.org/viaf/182113193/

https://en.wikipedia.org/wiki/My_%C3%81ntonia
http://www.worldcat.org/title/my-antonia/oclc/809034

...

to

https://archive.org/download/myantonia00cathrich/myantonia00cathrich_marc.xml 



cheers
stuart


On 19/08/14 11:57, Karen Coyle wrote:

About 1/3 of the 1M ebooks on OpenLibrary.org have full MARC records,
and you can retrieve the record via the API. There is also a secret
record format that returns not the full MARC for the hard copy (which is
what the records represent because these are digitized books) but a
record that has been modified to represent the ebook.

The MARC records for the hard copy follow the pattern:

https://archive.org/download/[archive identifier]/[archive
identifier]_marc.[xml|mrc]

Download MARC XML
https://archive.org/download/myantonia00cathrich/myantonia00cathrich_marc.xml 



Download MARC binary
https://www.archive.org/download/myantonia00cathrich/myantonia00cathrich_meta.mrc 

https://archive.org/download/myantonia00cathrich/myantonia00cathrich_meta.mrc 





To get the one that represents the ebook, do:

https://archive.org/download/[archive identifier]/[archive
identifier]_archive_marc.xml

https://archive.org/download/myantonia00cathrich/myantonia00cathrich_archive_marc.xml 




This one has an 007, the 245 $h, and a few other things.

Tom Morris did some code that helps you search for books by author and
title and retrieve a MARC record. I don't recall where his github
archive is, but I'll find out and post it here. The code is open source.
We used it for a project that added ebook records to a public library
catalog.

You can also use the OPenLibrary API to select all open access ebooks.
What I'd like to see is a way to create a list or bibliography in OL
that then is imported into a program that will find MARC records for
those books. The list function is still under development, though.

kc

On 8/18/14, 3:04 PM, Stuart Yeates wrote:

There are a stack of great free ebook repositories available on the
web, things like https://unglue.it/ http://www.gutenberg.org/
https://en.wikibooks.org/wiki/Main_Page http://www.gutenberg.net.au/
https://www.smashwords.com/books/category/1/newest/0/free/any etc, etc

What there doesn't appear to be, is high-quality AACR2 / RDA records
available for these. There are things like
https://ebooks.adelaide.edu.au/meta/pg/ which are elaborate dublin
core to MARC converters, but these lack standardisation of names,
authority control (people, entities, places, etc), interlinking, etc.

It seems to me that quality metadata would greatly increase the value
/ findability / use of these projects and thus their visibility and
available sources.

Are there any projects working in this space already? Are there
suitable tools available?

cheers
stuart




--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] metadata for free ebook repositories

2014-08-19 Thread Karen Coyle
It's hard to explain, but the Web-based mind of the Internet Archive 
just can't seem to grasp the role of individual library catalogs in 
helping people find books. The attitude seems to be: well, they should 
just search OL. The project I mentioned is an IMLS experiment to test 
ebook usage of these works through the addition of catalog records. 
There are lots of reasons why this is a good idea (selection, service to 
users...) but folks who don't use libraries don't and probably won't get it.


There is an on-again off-again Internet Archive project to fill in the 
missing MARC records. I was told that if someone can create a file of 
MARC records with links to IA books, that they could be loaded. If 
anyone wants to work on this, I can put you in touch with the 
appropriate person.


kc

On 8/18/14, 5:32 PM, Dana Pearson wrote:

Karen,

It seems to me that the Open Library would want to broaden use of this
great collection as much as possible.  Yet, MARC records for the 1/3  or so
items in the collection cannot be downloaded so that they could be imported
into local library systems.

Lots of users searching local libraries who might well use google and Open
Library, Internet Archive for finding ebooks less frequently.

I'll look at Tom Morris's code to see if I might automate record selection
of Open Library records compared with element of MARCXML records of this
last group of Guterberg Project additions.  Thanks for that information.

regards,
dana


On Mon, Aug 18, 2014 at 6:57 PM, Karen Coyle li...@kcoyle.net wrote:


About 1/3 of the 1M ebooks on OpenLibrary.org have full MARC records, and
you can retrieve the record via the API. There is also a secret record
format that returns not the full MARC for the hard copy (which is what the
records represent because these are digitized books) but a record that has
been modified to represent the ebook.

The MARC records for the hard copy follow the pattern:

https://archive.org/download/[archive identifier]/[archive
identifier]_marc.[xml|mrc]

Download MARC XML https://archive.org/download/myantonia00cathrich/
myantonia00cathrich_marc.xml
Download MARC binary https://www.archive.org/download/myantonia00cathrich/
myantonia00cathrich_meta.mrc https://archive.org/download/
myantonia00cathrich/myantonia00cathrich_meta.mrc


To get the one that represents the ebook, do:

https://archive.org/download/[archive identifier]/[archive
identifier]_archive_marc.xml

https://archive.org/download/myantonia00cathrich/
myantonia00cathrich_archive_marc.xml

This one has an 007, the 245 $h, and a few other things.

Tom Morris did some code that helps you search for books by author and
title and retrieve a MARC record. I don't recall where his github archive
is, but I'll find out and post it here. The code is open source. We used it
for a project that added ebook records to a public library catalog.

You can also use the OPenLibrary API to select all open access ebooks.
What I'd like to see is a way to create a list or bibliography in OL that
then is imported into a program that will find MARC records for those
books. The list function is still under development, though.

kc


On 8/18/14, 3:04 PM, Stuart Yeates wrote:


There are a stack of great free ebook repositories available on the web,
things like https://unglue.it/ http://www.gutenberg.org/
https://en.wikibooks.org/wiki/Main_Page http://www.gutenberg.net.au/
https://www.smashwords.com/books/category/1/newest/0/free/any etc, etc

What there doesn't appear to be, is high-quality AACR2 / RDA records
available for these. There are things like https://ebooks.adelaide.edu.
au/meta/pg/ which are elaborate dublin core to MARC converters, but
these lack standardisation of names, authority control (people, entities,
places, etc), interlinking, etc.

It seems to me that quality metadata would greatly increase the value /
findability / use of these projects and thus their visibility and available
sources.

Are there any projects working in this space already? Are there suitable
tools available?

cheers
stuart


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600






--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] metadata for free ebook repositories

2014-08-19 Thread Karen Coyle

On 8/19/14, 2:30 PM, Stuart Yeates wrote:

Authors in OL have already been linked to Wikipedia, and Wikipedia has
been linked to VIAF, and the OCLC number, when present, has been taken
from the MARC record. Therefore the OL record in some cases already has
these connections.

It's not just about authors. It's also about the work (+manifestation, +... ), 
the subject (particularly when the subject is separate work), the publisher, 
the illustrator, etc, etc.


Stuart, yes, of course. But there are few entities that we have that are 
clearly encoded beyond authors. We have never developed identifiers for 
works or manifestations (although hopefully we will); subjects, at least 
in terms of LCSH, are a mess because we don't have a viable list of 
post-coordinated subject terms. I'm not quite sure what you are getting 
at, but I would suggest that low hanging fruit is not a bad place to 
begin. Nor that we should reject some links because we can't achieve all.


kc



cheers
stuart


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] metadata for free ebook repositories

2014-08-18 Thread Karen Coyle
About 1/3 of the 1M ebooks on OpenLibrary.org have full MARC records, 
and you can retrieve the record via the API. There is also a secret 
record format that returns not the full MARC for the hard copy (which is 
what the records represent because these are digitized books) but a 
record that has been modified to represent the ebook.


The MARC records for the hard copy follow the pattern:

https://archive.org/download/[archive identifier]/[archive 
identifier]_marc.[xml|mrc]


Download MARC XML 
https://archive.org/download/myantonia00cathrich/myantonia00cathrich_marc.xml 

Download MARC binary 
https://www.archive.org/download/myantonia00cathrich/myantonia00cathrich_meta.mrc 
https://archive.org/download/myantonia00cathrich/myantonia00cathrich_meta.mrc 




To get the one that represents the ebook, do:

https://archive.org/download/[archive identifier]/[archive 
identifier]_archive_marc.xml


https://archive.org/download/myantonia00cathrich/myantonia00cathrich_archive_marc.xml

This one has an 007, the 245 $h, and a few other things.

Tom Morris did some code that helps you search for books by author and 
title and retrieve a MARC record. I don't recall where his github 
archive is, but I'll find out and post it here. The code is open source. 
We used it for a project that added ebook records to a public library 
catalog.


You can also use the OPenLibrary API to select all open access ebooks. 
What I'd like to see is a way to create a list or bibliography in OL 
that then is imported into a program that will find MARC records for 
those books. The list function is still under development, though.


kc

On 8/18/14, 3:04 PM, Stuart Yeates wrote:
There are a stack of great free ebook repositories available on the 
web, things like https://unglue.it/ http://www.gutenberg.org/ 
https://en.wikibooks.org/wiki/Main_Page http://www.gutenberg.net.au/ 
https://www.smashwords.com/books/category/1/newest/0/free/any etc, etc


What there doesn't appear to be, is high-quality AACR2 / RDA records 
available for these. There are things like 
https://ebooks.adelaide.edu.au/meta/pg/ which are elaborate dublin 
core to MARC converters, but these lack standardisation of names, 
authority control (people, entities, places, etc), interlinking, etc.


It seems to me that quality metadata would greatly increase the value 
/ findability / use of these projects and thus their visibility and 
available sources.


Are there any projects working in this space already? Are there 
suitable tools available?


cheers
stuart


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Library Privacy, RIP (Was: Canvas Fingerprinting by AddThis)

2014-08-17 Thread Karen Coyle
:-) Well, I don't know that I would use the word adorable, but it does 
warm my heart.


I found, to my pleasure, that libraries were shredding the paper 
computer sign-up sheets every evening (or when they filled up). That was 
good. But then I found, to my displeasure, that they had a box on the 
table in the childrens' room where summer reading program kids wrote 
their name, school, and age, and that the box was not secured in any way 
from scrutiny by others. Gulp! So it's a mixed bag in most libraries. 
Plus, there's always a hoarder or two who will not get rid of obsolete 
records. One value of an audit is that timely record destruction becomes 
a *policy*.


kc


On 8/17/14, 11:54 AM, Debra Shapiro wrote:

Conversation between 2 instructional staff at a library school:

Staff 1, “Say, I went down to our departmental library, and had to use the little 
paper slip to take out a book, because it’s summer and after hours. You have to 
fill in the book title, book bar code, and your own name  ID barcode. The fold 
the paper in half and stick it in a box. It’s got a little disclaimer on the bottom 
that the slip of paper will be destroyed as soon as the infor is entered into the 
system.”

Staff 2, “That’s adorable.”




On Aug 15, 2014, at 5:02 PM, Jason Bengtson j.bengtson...@gmail.com wrote:


Generally speaking, I think  surveillance is wretched stuff. But there is a
point at which the hand wringing becomes a bit much.

dsshap...@wisc.edu
Debra Shapiro
UW-Madison SLIS
Helen C. White Hall, Rm. 4282
600 N. Park St.
Madison WI 53706
608 262 9195
mobile 608 712 6368
FAX 608 263 4849


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


[CODE4LIB] Privacy audits Re: [CODE4LIB] Library Privacy, RIP

2014-08-16 Thread Karen Coyle
Some years ago I did a few privacy audits for local public libraries, 
where they went through all of their data-gathering points (circulation, 
summer reading programs, vendors, patron letters to the library 
director...) [1] It was very useful for them to discover where data 
might leak. At the time, none of the libraries I was working with was 
terribly involved with social networking.


I think it would be good to provide libraries with more information 
about what data is gathered via social networks, along with an analysis 
of where they are putting their patrons' privacy at risk. My guess is 
that many librarians are unaware of the data gathering done behind 
something like Google Analytics -- they just see a service that they 
need. And there's no use us complaining about it if we can't give good, 
solid information about 1) what data is gathered 2) what is the 
alternative.


I think a Code4lib guide to library privacy or something of that 
nature would be a valuable contribution. I'd be happy to work with folks 
on it.


kc
[1] http://kcoyle.net/privacy_audit.html


On 8/16/14, 10:12 AM, Riley Childs wrote:

I think that pretty much sums up the situation ;)

Sent from my Windows Phone

From: Eric Hellmanmailto:e...@hellman.net
Sent: ‎8/‎16/‎2014 1:06 PM
To: CODE4LIB@LISTSERV.ND.EDUmailto:CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Library Privacy, RIP  (Was: Canvas Fingerprinting by 
AddThis)

I think what we want is http://socialitejs.com/

On Aug 16, 2014, at 12:52 PM, Riley Childs rchi...@cucawarriors.com wrote:


Another question for someone who utilizes these services: What analytics does 
this provide and are the social analytics worth losing  your user's privacy? (I 
think not)
Can't we make our own non dynamic share links




Sent from my Windows Phone

From: Eric Hellmanmailto:e...@hellman.net
Sent: ‎8/‎16/‎2014 12:25 PM
To: CODE4LIB@LISTSERV.ND.EDUmailto:CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Library Privacy, RIP  (Was: Canvas Fingerprinting by 
AddThis)

So, 2 points worth discussing here.

1. I'll bet you most proxy servers are not proxying AddThis.com or 
Sharethis.com. So there wouldn't be any effect of proxying on the user tracking 
they do.

2. It really doesn't matter if you identify yourself to the catalog or not. 
You're being tracked across sites all over the internet. If you identify 
yourself to one of them, you can be identified. Note that the main concern here 
is if you use your own device to access the library's catalog.


On Aug 15, 2014, at 5:52 PM, Karen Coyle li...@kcoyle.net wrote:


On 8/15/14, 12:07 PM, Eric Hellman wrote:

AddThis and ShareThis, on the other hand have TOS that let them use tracking 
for advertising, and that's what their business is. So, hypothetically, a teen 
could look at library catalog records for books about childbirth, and as a 
result, later be shown ads for pregnancy tests, and that would be something the 
library has permitted.

Eric, I'm wondering about the full scenario that you are envisioning. Many 
libraries use proxy servers, so individual users are not identified. (Meaning 
that an 80-yr-old man may get the ad for the pregnancy test, not the teen.) In 
addition, in many cases the machine wipes itself clean daily, replacing all 
potential user files. (Someone else can explain this MUCH better than I just 
did.)

In my public library, I do not identify myself to the use the catalog on site 
-- not even to use journal article databases, because 1) authentication takes 
place in the library system 2) the proxy server's IP is my identity for those 
services. I have no idea what exits the library when I hook my laptop to the 
open network. Shouldn't all of these factors be taken into account? Can anyone 
articulate them from the point of view of a public library?

Note: At the university here at Berkeley, no network use is allowed without an 
account, so there is no anonymous use, at least on the human side of any proxy 
server that they run. But at the public library there is no log-on. So what is 
AddThis getting in those two situations?

kc

--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


[CODE4LIB] Parking, was: Re: [CODE4LIB] Hiring strategy for a library programmer with tight budget - thoughts?

2014-08-15 Thread Karen Coyle

On 8/15/14, 10:18 AM, Joe Hourcle wrote:

  If you can offer them reduced tuition or parking (matters at some campuses)


Have you heard what it takes to get your own parking space on Berkeley 
campus? A Nobel prize. Yep, you get a parking space with a Nobel, and 
every time there's a new Nobel winner on campus, he (so far) always gets 
interviewed by the school paper about how it feels to have a guaranteed 
parking place. [1]


BTW, they are marked on campus with the letters NL ONLY - Nobel 
Laureate only.[2]


kc
[1] 
http://www.dailycal.org/2013/12/12/uc-berkeley-professor-nobel-laureate-winner-randy-schekman-answers-reddit-questions/

[2] http://makezine.com/2010/02/08/uc-berkeley-has-nobel-laureate-only/

--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Parking, was: Re: [CODE4LIB] Hiring strategy for a library programmer with tight budget - thoughts?

2014-08-15 Thread Karen Coyle
Oh, they each have their own spot nearest their building, and they know 
what it is.


However, I'd pay $$ to watch two NL's duke it out over a parking space. 
Seriously. Physics vs. chemistry? Biology vs. mathematics? Mathematics 
vs. economics (!)? Wow.


kc


On 8/15/14, 10:47 AM, Jim Gilbert(WTPL) wrote:

Please tell me there is a policy about how the arrival of two NL's at the same 
spatial and temporal coordinates occupied by a NL-spot shall be handled!! :-D

James Gilbert, BS, MLIS
Systems Librarian
Whitehall Township Public Library
3700 Mechanicsville Road
Whitehall, PA 18052
610-432-4339 ext: 203


-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Karen 
Coyle
Sent: Friday, August 15, 2014 1:40 PM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: [CODE4LIB] Parking, was: Re: [CODE4LIB] Hiring strategy for a library 
programmer with tight budget - thoughts?

On 8/15/14, 10:18 AM, Joe Hourcle wrote:

   If you can offer them reduced tuition or parking (matters at some
campuses)

Have you heard what it takes to get your own parking space on Berkeley campus? 
A Nobel prize. Yep, you get a parking space with a Nobel, and every time 
there's a new Nobel winner on campus, he (so far) always gets interviewed by 
the school paper about how it feels to have a guaranteed parking place. [1]

BTW, they are marked on campus with the letters NL ONLY - Nobel Laureate 
only.[2]

kc
[1]
http://www.dailycal.org/2013/12/12/uc-berkeley-professor-nobel-laureate-winner-randy-schekman-answers-reddit-questions/
[2] http://makezine.com/2010/02/08/uc-berkeley-has-nobel-laureate-only/

--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Library Privacy, RIP (Was: Canvas Fingerprinting by AddThis)

2014-08-15 Thread Karen Coyle

On 8/15/14, 12:07 PM, Eric Hellman wrote:

AddThis and ShareThis, on the other hand have TOS that let them use tracking 
for advertising, and that's what their business is. So, hypothetically, a teen 
could look at library catalog records for books about childbirth, and as a 
result, later be shown ads for pregnancy tests, and that would be something the 
library has permitted.


Eric, I'm wondering about the full scenario that you are envisioning. 
Many libraries use proxy servers, so individual users are not 
identified. (Meaning that an 80-yr-old man may get the ad for the 
pregnancy test, not the teen.) In addition, in many cases the machine 
wipes itself clean daily, replacing all potential user files. (Someone 
else can explain this MUCH better than I just did.)


In my public library, I do not identify myself to the use the catalog on 
site -- not even to use journal article databases, because 1) 
authentication takes place in the library system 2) the proxy server's 
IP is my identity for those services. I have no idea what exits the 
library when I hook my laptop to the open network. Shouldn't all of 
these factors be taken into account? Can anyone articulate them from the 
point of view of a public library?


Note: At the university here at Berkeley, no network use is allowed 
without an account, so there is no anonymous use, at least on the human 
side of any proxy server that they run. But at the public library there 
is no log-on. So what is AddThis getting in those two situations?


kc

--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Canvas Fingerprinting by AddThis

2014-08-14 Thread Karen Coyle
Bill (others), are you running PrivacyBadger alongside AdBlock? I'm 
concerned about the confluence of decisions there, although tempted to 
try anyway.


Thanks,
kc

On 8/13/14, 2:08 PM, William Denton wrote:

On 13 August 2014, Karen Coyle wrote:

*ps - I had a great cookie manager for a while, but it's no longer 
around. Cookie control in browsers actually was easier a decade ago - 
they've obviously been discouraged from including that software. If 
anyone knows of a good cookie program or plugin, I'd like to hear 
about it.


I use Cookie Monster [0] and like it.

Related:  on my work box I'm trying out the EFF's Privacy Badger [1], 
which I hope will be a success.  At home I use Disconnect [2], which 
blocks entire domains.  It's great for cutting out cookies and junk 
like AddThis, but cripes, I hadn't realized how many people pull in 
Javascript libraries from Google or Yahoo. That's a harder way of 
tracking to avoid.


Bill

[0] https://addons.mozilla.org/en-US/firefox/addon/cookie-monster/
[1] https://www.eff.org/privacybadger
[2] https://disconnect.me/disconnect



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Canvas Fingerprinting by AddThis

2014-08-13 Thread Karen Coyle
I think this would bother me more if I thought that there were a 
significant number of users who either do not use cookies at all, or who 
had some kind of effective cookie manager. I suspect that the actual 
number is very close to zero, in the .n range at best.*


I have a problem with libraries using social media at all, actually, 
since it has turned out to be such a privacy disaster. When I go to what 
I think of as a benign site (my local library, DPLA, EFF!) and see that 
they've got a FB page that gathers likes it just chills me. I realize 
that all of these organizations need to maintain a level of visibility, 
and Facebook or Tumblr or whatever is a way to do that. However, there 
is no use of social media that can be argued as being privacy-neutral. 
It's a dilemma, I know. But I hate to see libraries and others seemingly 
promoting its use.


As for web use, only Tor, and perhaps not even Tor, can give you 
something close to anonymity (except, perhaps, with the NSA). But it 
requires certain tech chops and an effort way beyond that of clearing 
out your cookies now and again, something that most people do not do.


kc
*ps - I had a great cookie manager for a while, but it's no longer 
around. Cookie control in browsers actually was easier a decade ago - 
they've obviously been discouraged from including that software. If 
anyone knows of a good cookie program or plugin, I'd like to hear about it.



On 8/13/14, 10:22 AM, Eric Hellman wrote:

It seems that Code4Lib hasn't discussed this., though the news is 2 weeks old. It seems that 
there are libraries using social share tools from AddThis, a company that has been using 
a technology called Canvas Fingerprinting to track users.

In other words, it looks like libraries are giving away the user-privacy store.

For example, AddThis is used by my public library's Polaris catalog (BCCLS).

I'd be interested to learn how widespread this is.

Here's the article from ProPublica.
http://www.propublica.org/article/meet-the-online-tracking-device-that-is-virtually-impossible-to-block

And a follow-on discussion from Princeton CITP
https://freedom-to-tinker.com/blog/englehardt/the-hidden-perils-of-cookie-syncing/

The research article:
https://securehomes.esat.kuleuven.be/~gacar/persistent/the_web_never_forgets.pdf

Techdirt:
https://www.techdirt.com/articles/20140721/14523127960/tons-sites-including-whitehousegov-experiment-with-tracking-technology-that-is-difficult-to-block.shtml


Eric


Eric Hellman
President, Gluejar.Inc.
Founder, Unglue.it https://unglue.it/
http://go-to-hellman.blogspot.com/
twitter: @gluejar


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] ACTION REQUESTED: Volunteer for 2015 Conference Committees

2014-08-12 Thread Karen Coyle

Hi, Tom. I guess we've got OSU as the place, but what are the dates? - kc

On 8/12/14, 3:35 PM, Tom Johnson wrote:

The 2015 Code4Lib annual conference may seem like a long way away, but it's
already time for various volunteer committees to start work on the
planning. As you know, the conference happens each year thanks to the work
of the community at large; please take the time to sign up for confrence
committees here:

*http://wiki.code4lib.org/index.php/2015_Conference_Committees
http://wiki.code4lib.org/index.php/2015_Conference_Committees*

*We especially need people for the Keynote Speakers and Sponsorship
committees.* These groups need to begin work ASAP and are short on
membership.

Thanks to everyone who has already signed up to help!

--
Tom Johnson
on behalf of the C4L15 PDX Team


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: +1-510-435-8234
skype: kcoylenet/+1-510-984-3600


Re: [CODE4LIB] Creating a Linked Data Service

2014-08-08 Thread Karen Coyle
Also, note that Martin Malmsten, of the Norwegian National Library 
(whose catalog is a linked data catalog) always states that one of the 
advantages of LD is that the difference between inside resources and 
outside resources disappears. It's all just linked resources. It makes 
sense to start with an inside resource, since you know more about it, 
but the same technology should work for any linking, anywhere, any time.


kc

On 8/8/14, 7:07 AM, Mark A. Matienzo wrote:

Per Laura's message, and what I think was the underlying idea behind Mike's
post, I think there's still a great opportunity to learn something new.
Perhaps you might want to look at WebSocket [0], and Jason Ronallo's
presentation from Code4lib 2014 [1] was a great intro. It seems like this
might be a good candidate for showing real-time availability information.

[0] https://www.websocket.org/
[1] http://jronallo.github.io/presentations/code4lib-2014-websockets/

Cheers,
Mark


--
Mark A. Matienzo m...@matienzo.org
Director of Technology, Digital Public Library of America


On Fri, Aug 8, 2014 at 9:23 AM, Shaun Ellis sha...@princeton.edu wrote:


I don't understand the publish it and they will come mentality when it
comes to linked data.  If you can't define a clear use case for your own
data infrastructure, then I can't see how you would justify the time spent.

The making data available to the world at large is a nice byproduct, but
you can't write a use case for unknown users with unknown goals.  So,
if you have no plans to use the data in some productive way, then I'm sure
you have more pressing things to do with your time.

-Shaun


On 8/7/14 9:48 AM, Scott Prater wrote:


Echoing others... the use case for linked data appears to be making data
available to the world at large, unknown consumers, who may find a use
for it that you never imagined.

Name authority services (like VIAF), catalogs of public resources, map
data -- all these are good candidates for a linked data approach.

Hardware availability at your library?  Not so much.  It's hard to
imagine a case where that information would be useful outside your walls.

-- Scott

On 08/07/2014 08:09 AM, Ethan Gruber wrote:


I agree with others saying linked data is overkill here. If you don't
have
an audience in mind or a specific purpose for implementing linked data,
it's not worth it.


On Thu, Aug 7, 2014 at 9:07 AM, Jason Stirnaman jstirna...@kumc.edu
wrote:

  Mike,

Check out
http://json-ld.org/,
http://json-ld.org/primer/latest/, and
https://github.com/digitalbazaar/pyld

But, if you haven't yet sketched out a model for *your* data, then
the LD
stuff will just be a distraction. The information on Linked Data seems
overly complex because trying to represent data for the Semantic Web
gets
complex - and verbose.

As others have suggested, it's never a bad idea to just do the simplest
thing that could possibly work.[1] Mark recommended writing a simple
API.
That would be a good start to understanding your data model and to
eventually serving LD. And, you may find that it's enough for now.

1. http://www.xprogramming.com/Practices/PracSimplest.html

Jason

Jason Stirnaman
Lead, Library Technology Services
University of Kansas Medical Center
jstirna...@kumc.edu
913-588-7319

On Aug 6, 2014, at 1:45 PM, Michael Beccaria mbecca...@paulsmiths.edu
wrote:

  I have recently had the opportunity to create a new library web page

and


host it on my own servers. One of the elements of the new page that I
want
to improve upon is providing live or near live information on technology
availability (10 of 12 laptops available, etc.). That data resides on my
ILS server and I thought it might be a good time to upgrade the
bubble gum
and duct tape solution I now have to creating a real linked data service
that would provide that availability information to the web server.


The problem is there is a lot of overly complex and complicated


information out there onlinked data and RDF and the semantic web etc.
and
I'm looking for a simple guide to creating a very simple linked data
service with php or python or whatever. Does such a resource exist? Any
advice on where to start?


Thanks,

Mike Beccaria
Systems Librarian
Head of Digital Initiative
Paul Smith's College
518.327.6376
mbecca...@paulsmiths.edu
Become a friend of Paul Smith's Library on Facebook today!






--
Shaun Ellis
User Interface Developer, Digital Initiatives
Princeton University Library
609.258.1698

“Any darn fool can get complicated. It takes genius to attain simplicity.”
-Pete Seeger



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Have file will share

2014-07-17 Thread Karen Coyle

https://web.archive.org/web/20070209042706/http://www4.infotrieve.com/ariel/downloads.asp

??
kc


On 7/17/14, 7:05 AM, Francis Kayiwa wrote:

-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

Heya,

I am looking for the Ariel Receive and Full file found at the 404'ing
file in the URI below.

http://www4.infotrieve.com/ariel/downloads.asp

I've sent a message to @infotrieve.com that seems to have gone to a
bitbucket. Thanks for sending this file my way.

What is it with Library Vendors? Ugh!

./fxk

- -- 
Don't stop to stomp ants when the elephants are stampeding.

-BEGIN PGP SIGNATURE-
Version: GnuPG v2.0.22 (MingW32)
Comment: Using GnuPG with Thunderbird - http://www.enigmail.net/

iQEcBAEBAgAGBQJTx9geAAoJEOptrq/fXk6Mp3MH/jd60zLIpzDTz/IjOEuzlfDs
zihdbdvpmbAvwqgpDM12lsaGYE9NALi9GMW6RDVpXphjimfhGn222iOLTCB/B68R
MzKto02pNSk6O3lfTKZuqfBrOJyjIEgUuSscwwurKRHxlfg3RBo9dcBMcmJoqG98
kJFzcOzmdq9SCT6SVu223nFwHArYnGgEUa0iveVnG9LtHCax5G8oEiBc2C9NWR9U
V3sNKAW1APrr+87sDQ86JEsucJ4V3UtsSvSzWLOvWxU74Mo4Oms0wXD+bNl6u9Di
uh95r2Pq8emxhAqTfVB9nrdGwcnfDCdCUkclXjW/Tan82wTBFx7+FNg9yxmwkG0=
=ausY
-END PGP SIGNATURE-


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


[CODE4LIB] Zoia Horn - RIP

2014-07-15 Thread Karen Coyle

From her family:

Zoia died on July 12th, Saturday, around 7 am.

She was in hospice and her family was with her.

I'll do the sad update to her wikipedia page.

kc


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Is ISNI / ISO 27729:2012 a name identifier or an entity identifier?

2014-06-20 Thread Karen Coyle
I concur with Richard's analysis[1]. Each identifier type serves a 
different community. In particular, ORCID identifiers will tend to 
identify faculty and researchers whose sole output is journal articles 
-- thus who would not normally appear in a library authority file. The 
ISNI is sometimes seen as an interloper from the publishing community, 
but most likely is integrated into the publisher workflow (e.g. writing 
checks to authors).


Like Richard, I don't see anything to worry about. You can use one, 
some, or all of the identifiers based on your needs. So a faculty 
digital repository may need to used ORCIDs because there are authors who 
are only identified by those. (Repositories are beginning to require 
ORCIDs for deposit.) The same repository can also use LCNA for some 
authors -- you are in no way limited to one identifier per person or 
system. If you are hoping to pull in author data in your library catalog 
from wiki/DB/pedia, then you might favor the VIAF identifier, since this 
is being linked to the pedia world.


This seems to me to be quite similar to other data and metadata choices 
that we make: define your use case, then choose the data that meets that 
need.


kc
[1] One possible difference is that I would consider ORCID a viable URI 
for linked data purposes, although at the moment ORCID does not export 
its data in RDF. All of the identifiers listed below are HTTP URIs.


On 6/20/14, 7:56 AM, Richard Wallis wrote:

Hi Eric,

What distinguishes one from another?

The communities behind them, the [often overlapping] communities they
are intended to serve, and the technical implementation.

As a librarian, why should I care?

I would, as a non-librarian, suggest that once you are happy with
the ‘authority’ of them, you shouldn’t have to care. Ideally, we are not
there yet, systems should be flexible and accommodating enough to link to
any appropriate authority.

I will probably get flamed for over generalisation here but - VIAF is
an aggregation of National Libraries Authority files.  - ISNI is a more
publisher focused but similar effort.  - OCID comes from and and tries to
serve individual academic institutions, their researchers and falsity
authors.


   authority control |simple identifier |Linked Data capability
  +-+--+--+
   VIAF   |X|X |  X   |
  +-+--+--+
   ORCID  | | X|  |
  +-+--+--+
ISNI  |X| X|X |
  +-+--+--+

~Richard


On 20 June 2014 15:42, Eric Lease Morgan emor...@nd.edu wrote:


On Jun 20, 2014, at 10:31 AM, Richard Wallis 
richard.wal...@dataliberate.com wrote:


ISNI has a suite of programs that detects pseudonyms coded as name

variants

and changes them into related name and generates related identity

records.

It is a while since it was run and will be re-run in the next few weeks.
This should change Currer Bell into a related name of Charlotte Brontë .


Please humor me as I ask this question again. What is the difference
between ISNI and other identifiers systems (like ORCID, etc.)? What
distinguishes one from another? As a librarian, why should I care? Was as a
faculty member/scholar, why should I care? Under what context is one
identifier expected to be used instead of another? Maybe a picture/graph is
in order:

   authority control simple pointer
  +-+--+
   VIAF   |X|  |
  +-+--+
   ORCID  | | X|
  +-+--+
ISNI  | |  |
  +-+--+

—
Eric Lease Morgan






--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Is ISNI / ISO 27729:2012 a name identifier or an entity identifier?

2014-06-20 Thread Karen Coyle

On 6/20/14, 11:38 AM, Richard Wallis wrote:

In what ways does ISNI support linked data?

See: http://www.isni.org/how-isni-works#HowItWorks_LinkedData


 accessible by a persistent URI in the form 
isni-url.oclc.nl/isni/000134596520 (for example)  and soon also in 
the form isni.org/isni/000134596520. 


Odd. I assume that whoever wrote that on their page just forgot the 
http://; part of those strings. Right?


kc




~Richard


On 20 June 2014 18:57, Eric Lease Morgan emor...@nd.edu wrote:


On Jun 20, 2014, at 10:56 AM, Richard Wallis 
richard.wal...@dataliberate.com wrote:


  authority control|simple identifier |Linked Data capability
 +-+--+--+
  VIAF   |X|X |  X   |
 +-+--+--+
  ORCID  | |X |  |
 +-+--+--+
   ISNI  |X|X |  X   |
 +-+--+--+


Increasingly I like linked data, and consequently, here is clarification
and a question. ORCID does support RDF, but only barely. It can output
FOAF-like data, but not bibliographic. Moreover, it is experimental, at
best:

   curl -L -H 'accept: application/rdf+xml'
http://orcid.org/-0002-9952-7800

In what ways does ISNI support linked data?

---
Eric Morgan






--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Is ISNI / ISO 27729:2012 a name identifier or an entity identifier?

2014-06-20 Thread Karen Coyle

On 6/20/14, 1:49 PM, Joe Hourcle wrote:



Now, it's possible that this whole we don't need to bother with
http://; thing has spilled into the CMS building community, and
they're actively stripping it out.


I actually had the editors of an ALA publication remove http://; 
whenever it preceded www because they were convinced that you only 
needed http://; when there was no www at the front of the domain 
name. I had to fight to get the http://; put back in.  (I believe the 
excuse was that it took up space.)



 From their page, I think they're
using Drupal, but the horrible block of HTML that this was in is
blatantly MS Word's 'save as HTML' foulness:

   h2span lang=EN-USa name=HowItWorks_LinkedData/aLinked 
Data/span/h2
   p class=MsoNormalspan lang=EN-USLinked data is part of the ISNI-IA’s strategy to make ISNIs 
freely available and widely diffused.nbsp; Each assigned ISNI is accessible by a persistent URI in the form 
isni-url.oclc.nl/isni/000134596520 (for example) nbsp;and soon also in the form 
isni.org/isni/000134596520.nbsp;/span/p
   p class=MsoNormalspan lang=EN-USComing soon:nbsp; ISNI core metadata in RDF 
triples.nbsp; The RDF triples will be embedded in the public web pages and the format will be available via the persistent URI 
and the SRU search API./span/p
   p class=MsoNormalspan lang=EN-USnbsp;/span/p


Which is so foul that Dreamweaver has a Clean up Word HTML command in 
its menu. But you still end up with pretty bad HTML.


kc



-Joe




On 20 June 2014 18:57, Eric Lease Morgan emor...@nd.edu wrote:


On Jun 20, 2014, at 10:56 AM, Richard Wallis 
richard.wal...@dataliberate.com wrote:


  authority control|simple identifier |Linked Data capability
 +-+--+--+
  VIAF   |X|X |  X   |
 +-+--+--+
  ORCID  | |X |  |
 +-+--+--+
   ISNI  |X|X |  X   |
 +-+--+--+

Increasingly I like linked data, and consequently, here is clarification
and a question. ORCID does support RDF, but only barely. It can output
FOAF-like data, but not bibliographic. Moreover, it is experimental, at
best:

   curl -L -H 'accept: application/rdf+xml'
http://orcid.org/-0002-9952-7800

In what ways does ISNI support linked data?

---
Eric Morgan




--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] software for a glossary

2014-06-20 Thread Karen Coyle

On 6/20/14, 1:16 PM, Tom Keays wrote:


Crossing the thread over to linked author data, this item made me laugh.
http://w3cmemes.tumblr.com/post/76273506486/dave-started-reviewing-open-annotations-today


The only thing worse would be:
OMG IT'S FULL OF OWL

kc



On Fri, Jun 20, 2014 at 2:28 PM, Eric Lease Morgan emor...@nd.edu wrote:

Do you know of a Web-based tool or piece of desktop software that would let a 
professor post a text in a frame, then highlight words or phrases and link them 
to a glossary? A quick-and-dirty web page (possibly attached) and link below 
illustrates the idea:

   http://dh.crc.nd.edu/tmp/glossary.html

—
Eric Morgan


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Does 'Freedom to Read' require us to systematically privilege HTTPS over HTTP?

2014-06-16 Thread Karen Coyle

On 6/16/14, 1:49 PM, Galen Charlton wrote:
However, I think that's only part of the picture for ILSs. Other parts 
would include: * staff training on handling patron and circulation 
data * ensuring that the ILS has the ability to control (and let users 
control) how much circulation and search history data gets retained * 
ensuring that the ILS backup policy strikes the correct balance 
between having enough for disaster recovery while not keeping 
individually identifiable circ history forever * ensuring that 
contracts with ILS hosting providers and services that access patron 
data from the ILS have appropriate language concerning data retention 
and notification of subpoenas. Regards, Galen 
Echoing Galen, staff training is very important. One way to begin this 
is by having the staff do a privacy audit, where they make sure that the 
library understands the reality of its practices, and makes changes 
where it should and can. I have examples and materials at:

  http://kcoyle.net/privacy_audit.html
although these were developed mainly for public libraries.

Part of the process is setting up a chain of command for privacy issues. 
For US libraries, Mary Minow has given talks to librarians on what to do 
if law enforcement shows up at your door. According to her experience, 
they often try to find a library staff member who has access to systems 
but who isn't at a management level, and they tend to try to (and mostly 
succeed to) intimidate. Knowing the law makes a difference. So for US 
libraries, there is:

  http://librarylaw.com/Privacy.html

kc


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] mailing list administratativia

2014-06-09 Thread Karen Coyle
Mine is bibliounitarian -- for those who believe in a single record 
model for bibliographic data, as opposed to biblioquatritarians 
(FRBR-ites) or biblioduotarians (Bibframers).


kc

On 6/9/14, 3:30 PM, Eric Hellman wrote:

I assume you've discovered the word is usually spelled administrivia.

It's fun making up words. My latest is bibliopotheosis

On May 29, 2014, at 10:23 AM, Eric Lease Morgan emor...@nd.edu wrote:


* Apparently “administratativia” is a word of my own design because a search of 
it in Google returns only postings I’ve written. No wonder my spell checker 
doesn’t like it.


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] orcid and researcherid and scopus, oh my

2014-06-06 Thread Karen Coyle

On 6/4/14, 8:01 PM, Dave Caroline wrote:


You seem to have multiple so none are unique :)


Just in case this was the confusion, the U in URI stands for Uniform. 
As Jodi says, there is no requirement for uniqueness of identifiers on 
the web. In fact, non-uniqueness is a feature, not a bug. It mirrors 
reality, where you have an SSN, a library card number, a driver's 
license number, etc., and each identifies you in some context.


kc





couple of instances where these sort of identifiers are being put into MARC

and what happens when one of the servers falls off the net or some
boss decides to modify a service, lots of data to fixup, yet again

Dave Caroline, not convinced of the linked world prophesies yet


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] orcid and researcherid and scopus, oh my

2014-06-06 Thread Karen Coyle
My understand was that ISNI comes out of the publishing world -- 
primarily commercial publishing. The Wikipedia article puts it:


 The ISNI allows a single identity (such as an author's pseudonym 
https://en.wikipedia.org/wiki/Pseudonym or the imprint 
https://en.wikipedia.org/wiki/Imprint used by a publisher 
https://en.wikipedia.org/wiki/Publisher) to be identified using a 
unique number. This unique number can then be linked to any of the 
numerous other identifiers that are used across the media industries to 
identify names and other forms of identity.

https://en.wikipedia.org/wiki/International_Standard_Name_Identifier

The interesting thing about the numbers listed by Eric is that although 
there is overlap, each has some identities that the others don't. AFAIK, 
ORCID will not include people who died before the system was developed, 
and ISNI may not include folks whose works are long out of print. Also, 
the maintenance of them differs: ISNI is assigned by an agency (as is 
LCNA), while ORCIDs are considered to be under the control of the 
identified person.


kc

On 6/4/14, 9:20 PM, David Lowe wrote:

Been tracking the ORCiD/ISNI thing for a few months now.  Here was a nice 
graphic that I think this group will appreciate of how CU-Boulder sees ORCiDs 
across their systems from a recent ORCiD event:

[cid:image001.jpg@01CF79CD.64C36FA0]



Individual academics (especially grant-related folks) should get and use ORCiDs, which 
either the individual can register for or their affiliate institutions may assign in 
batches with an institutional membership.  An ISNI, in contrast, may be assigned to an 
individual, often for rights-tracking purposes, as in the case of recorded performances.  
Also ISNIs can be given to organizations (corporate bodies, in catalogerese) 
while ORCiDs are only for individuals.

--DBL



-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Eric 
Lease Morgan
Sent: Wednesday, June 04, 2014 3:55 PM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] orcid and researcherid and scopus, oh my



On Jun 4, 2014, at 3:32 PM, Jodi Schneider 
jschnei...@pobox.commailto:jschnei...@pobox.com wrote:




However, I believe that ISNI is bridging between these various sources
-- certainly including LC and VIAF [1], and also ORCID [2].
[1] http://www.isni.org/content/data-contributors
[2] From
http://orcid.org/content/what-relationship-between-isni-and-orcid
ORCID identifiers utilize a format compliant with the ISNI ISO standard.
ISNI has reserved a block of identifiers for use by ORCID, so there
will be no overlaps in assignments.





I'm glad ISNI's have been mentioned. Thank you.



I had not heard about ISNIs until very recently, and I was both embarrassed as 
well as surprised. Sure, I've read the Web pages, [1] but what are ISNIs? How 
are they different from other identifiers, and is it just my imagination or 
have they simply not become as popular here in the 'States as they are in 
Europe, and why?



[1] ISNI - http://isni.org/



-

Eric M.



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] orcid and researcherid and scopus, oh my

2014-06-06 Thread Karen Coyle

Eric,

I do think that there is a difference between identification and the 
uses that libraries have made of authorities. Part of it is historical 
-- that libraries began using authority control before the use of 
computers. Library authority control controls the form of the display of 
the name as its way to achieve uniqueness. Some of the identifiers 
listed do not provide a preferred form of the name. Instead, they 
provide a machine-readable identifier that brings together the variant 
forms of the name that have been used in publications.


Among these is VIAF, btw. VIAF identifies a person by clustering the 
authority records for that person. These authority files have preferred 
name forms, but there is no preferred display form of the name 
associated with the VIAF identifier.


This brings up the question of whether/when identifiers can perform the 
functions that are performed today by library authority control; is a 
single form of the name needed if the person can be identified in 
another way? Catalogers, in my experience, answer Yes, users need to 
see a single form of the name. However, the world of journal articles 
simply does not provide this, and it appears that ambiguity in names 
will be resolved through identifiers rather than name forms.


kc


On 6/4/14, 8:27 PM, Eric Lease Morgan wrote:

  * ORCID - http://orcid.org/-0002-9952-7800
  * ResearcherID - http://www.researcherid.com/rid/F-2062-2014
  * Scopus - http://www.scopus.com/authid/detail.url?authorId=25944695600
  * VIAF - http://viaf.org/viaf/26290254
  * LC - http://id.loc.gov/authorities/names/n94036700
  * ISNI - http://isni.org/isni/35290715

How have any of y'all used theses sorts of identifiers, and what problems do 
you think you will be able to solve by doing so?


Each of these identifiers are essentially keys in a table, and the table often 
points to written works. To what degree are these sorts of things intended to 
be “authority records” and to what degree are they simply expected to be 
identifiers? What’s the difference? I do know that things like ORCIDs are 
intended to be included in grant and journal submissions — so they are keys 
also pointing to things like names, addresses, affiliations, etc. If there is 
computer-readable data/information at the other end of the identifiers, then, 
the data/information could be collected to create reports, such as on-the-fly 
curriculum vitas or departmental publication reports. —ELM


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] separate list for jobs

2014-05-07 Thread Karen Coyle

On 5/7/14, 5:37 AM, Sarah Shealy wrote:

you're creating an artificial divide between us and them. No one is 
'intruding' on anyone - it's a public forum for people who work in LIBRARIES (aka - places where 
inclusion is supposed to rule) without any filtering of participants, so technically anyone with an 
email address is part of the 'ingroup'.
Yeah, Sarah! I think this is a good reminder that the attitude of if 
you don't know, you don't belong here is NOT what c4l is about. 
Instead, it should be if you don't know, let me help you learn what you 
need to know. At the same time, if someone asks a question that you 
don't have the time or patience to answer, just let it flow by -- 
someone else will catch it. You don't have to read every post to c4l, 
and you definitely are not obligated to answer questions that come to 
the list.


Delete key. The all-important delete key.

kc

--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Italian lectures on semantic web and Linked Data

2014-05-07 Thread Karen Coyle

Mille grazie, Stefano! - kc

On 5/7/14, 8:32 AM, Stefano Bargioni wrote:

Hi, Karen:
the recorderd video is available at https://www.youtube.com/watch?v=ZN9gEM0RmqE.
Eric Lease Morgan starts at 36min
Giovanni Bergamin starts at 1h 30min
Stefano Bargioni starts at 2h 03min
Koha Gruppo Italiano https://www.facebook.com/KohaGruppoItaliano will add 
links to ppt asap.
Thanks. Stefano

On 06/mag/2014, at 16.29, Karen Coyle wrote:


Thanks, Stefano. The time zone difference makes this pretty much impossible for 
some of us. Will the talks be recorded for later viewing?

kc

On 5/6/14, 2:09 AM, Stefano Bargioni wrote:

Koha Gruppo Italiano is pleased to announce that you can follow the conference 
Italian lectures on semantic web and Linked Data: practical examples for 
libraries via a streaming on
https://www.youtube.com/watch?v=ZN9gEM0RmqE
starting from May 7th, 2014 9:00 AM (CET). Mostly in Italian.
The full program is available at
www.pusc.it/sites/default/files/bib/7maggio2014.pdf
You are warmly welcome to participate.
Stefano

PS : apologies for cross posting.

--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Italian lectures on semantic web and Linked Data

2014-05-06 Thread Karen Coyle
Thanks, Stefano. The time zone difference makes this pretty much 
impossible for some of us. Will the talks be recorded for later viewing?


kc

On 5/6/14, 2:09 AM, Stefano Bargioni wrote:

Koha Gruppo Italiano is pleased to announce that you can follow the conference 
Italian lectures on semantic web and Linked Data: practical examples for 
libraries via a streaming on
https://www.youtube.com/watch?v=ZN9gEM0RmqE
starting from May 7th, 2014 9:00 AM (CET). Mostly in Italian.
The full program is available at
www.pusc.it/sites/default/files/bib/7maggio2014.pdf
You are warmly welcome to participate.
Stefano

PS : apologies for cross posting.


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] barriers to open metadata?

2014-04-30 Thread Karen Coyle
My question has to do with discoverability. Let's say that I have a 
bibliographic database and I want to add the OCLC work identifiers to 
it. Obviously I don't want to do it by hand. I might have ISBNs, but in 
some cases I will have a regular author/title-type citation.


and let's say that I am asking this for two different institutions:
a) is an OCLC member institution
b) is not

Thanks,
kc



On 4/30/14, 8:47 AM, Dan Scott wrote:

On Tue, Apr 29, 2014 at 11:37 PM, Roy Tennant roytenn...@gmail.com wrote:

This has now instead become a reasonable recommendation
concerning ODC-BY licensing [3] but the confusion and uncertainty
about which records an OCLC member may redistribute remains.

[3] http://www.oclc.org/news/releases/2012/201248.en.html

Allow me to try to put this confusion and uncertainty to rest once and for
all:

ALL THE THINGS. ALL.

At least as far as we are concerned. I think it's well past time to put the
past in the past.

That's great, Roy. That's a *lot* simpler than parsing the
recommendations, WCRR, community norms, and such at [A, B] :)


Meanwhile, we have just put nearly 200 million works records up as linked
open data. [1], [2], [3]. If that doesn't rock the library open linked data
world, then no one is paying attention.
Roy

[1] http://oclc.org/en-US/news/releases/2014/201414dublin.html
[2]
http://dataliberate.com/2014/04/worldcat-works-197-million-nuggets-of-linked-data/
[3] http://hangingtogether.org/?p=3811

Yes, that is really awesome. But Laura was asking about barriers to
open metadata, so damn you for going off-topic with PR around a lack
of barriers to some metadata (which, for those who have not looked
yet, have a nice ODC-BY licensing statement at the bottom of a given
Works page) :)

A. http://oclc.org/worldcat/community/record-use.en.html
B. 
http://oclc.org/worldcat/community/record-use/data-licensing/questions.en.html


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] getting URIs, was: [CODE4LIB] barriers to open metadata?

2014-04-30 Thread Karen Coyle
Thanks, Richard. I ask because it's one of the most common questions 
that I get -- often about WorldCat, but in general about any source of 
URIs -- How do I connect my data (text forms) to their URIs? And these 
questions usually come from library or archive projects with little or 
no programming staff. So it seems like we need to be able to answer that 
question so that people can get linked up. In fact, it seems to me that 
the most pressing need right now is an easy way (or one that someone 
else can do for you at a reasonable cost) to connect the text string 
identifiers that we have to URIs. I envision something like what we 
went through when we moved from AACR name forms to AACR2 name forms, and 
libraries were able to send their MARC records to a service that 
returned the records with the new name form. In this case, though, such 
a service would return the data with the appropriate URIs added. (In the 
case of MARC, in the $0 subfield.)


It's great that the big guys like LC and OCLC are providing URIs for 
resources. But at the moment I feel like it's grapes dangling just 
beyond the reach of the folks we want to connect to. Any ideas on how to 
make this easy are welcome. And I do think that there's great potential 
for an enterprising start-up to provide an affordable service for 
libraries and archives. Of course, an open source pass in your data in 
x or y format and we'll return it with URIs embedded would be great, 
but I think it would be reasonable to charge for such a service.


kc


On 4/30/14, 9:59 AM, Richard Wallis wrote:

To unpack the several questions lurking in Karen’s question.

As to being able to use the WorldCat Works data/identifiers there is no
difference between a or b - it is ODC-BY licensed data.

Getting a Work URI may be easier for a) as they should be able to identify
the OCLC Number and hence use the linked data from it’s URI 
http://worldcat.org/oclc/{ocn} to pick up the link to it’s work.

Tools such as xISBN http://xisbn.worldcat.org/xisbnadmin/doc/api.htm can
step you towards identifier lookups and are openly available for low volume
usage.

Citation lookup is more a bib lookup feature, that you could get an OCLC
Number from. One of colleagues may be helpful on the particulars of this.

Apologies for being WorldCat specific, but Karen did ask.

~Richard.


On 30 April 2014 17:15, Karen Coyle li...@kcoyle.net wrote:


My question has to do with discoverability. Let's say that I have a
bibliographic database and I want to add the OCLC work identifiers to it.
Obviously I don't want to do it by hand. I might have ISBNs, but in some
cases I will have a regular author/title-type citation.

and let's say that I am asking this for two different institutions:
a) is an OCLC member institution
b) is not

Thanks,
kc




On 4/30/14, 8:47 AM, Dan Scott wrote:


On Tue, Apr 29, 2014 at 11:37 PM, Roy Tennant roytenn...@gmail.com
wrote:


This has now instead become a reasonable recommendation

concerning ODC-BY licensing [3] but the confusion and uncertainty
about which records an OCLC member may redistribute remains.

[3] http://www.oclc.org/news/releases/2012/201248.en.html


Allow me to try to put this confusion and uncertainty to rest once and
for
all:

ALL THE THINGS. ALL.

At least as far as we are concerned. I think it's well past time to put
the
past in the past.


That's great, Roy. That's a *lot* simpler than parsing the
recommendations, WCRR, community norms, and such at [A, B] :)

  Meanwhile, we have just put nearly 200 million works records up as linked

open data. [1], [2], [3]. If that doesn't rock the library open linked
data
world, then no one is paying attention.
Roy

[1] http://oclc.org/en-US/news/releases/2014/201414dublin.html
[2]
http://dataliberate.com/2014/04/worldcat-works-197-million-
nuggets-of-linked-data/
[3] http://hangingtogether.org/?p=3811


Yes, that is really awesome. But Laura was asking about barriers to
open metadata, so damn you for going off-topic with PR around a lack
of barriers to some metadata (which, for those who have not looked
yet, have a nice ODC-BY licensing statement at the bottom of a given
Works page) :)

A. http://oclc.org/worldcat/community/record-use.en.html
B. http://oclc.org/worldcat/community/record-use/data-
licensing/questions.en.html


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet






--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] getting URIs, was: [CODE4LIB] barriers to open metadata?

2014-04-30 Thread Karen Coyle
Jonathan, I think we can point to some interesting benefits. If you take 
a look at what the BBC has done with their Wildlife site [1] and then 
look at the new FAO catalog [2] you can see how a page can be enhanced 
with useful data based on URIs in the bibliographic records. Imagine 
being able to add the short author bio from Wikipedia to a record 
display. etc. etc. [3] Or linking from a person as subject to the New 
York times data page for that person. [4]


Now, I know that your reply will be: but only if the vendors do it. 
Well, godammnit, we sure as hell can't wait for them - they are 
followers, not leaders. (And maybe this will give a boost to OS catalogs 
that don't have to wait for the unwieldy barge of library systems to 
make its change of direction.)


Note also that linked data is already happening in libraries in Europe, 
and the entire Europeana and DPLA are being developed as LD databases. 
This isn't some far out future nuttiness. We're actually running behind.


kc

[1] http://www.w3.org/2001/sw/sweo/public/UseCases/BBC/
[2] Info: http://aims.fao.org/agris; search interface: 
http://agris.fao.org/agris-search/index.do

[3] try this out in: https://apps.facebook.com/WorldCat/
[4] http://data.nytimes.com/N20483401082089183163 (R. Nixon) which links 
to page with a huge list of articles 
http://topics.nytimes.com/top/reference/timestopics/people/n/richard_milhous_nixon/index.html


On 4/30/14, 11:13 AM, Jonathan Rochkind wrote:
If you want libraries to spend money on adding URI's to their data, 
there is going to need to be some clear benefit they get from doing it 
-- and it needs to be a pretty near-term benefit, not Well, some day 
all these awesome things might happen, because linked data.



On 4/30/14 1:34 PM, Karen Coyle wrote:

Thanks, Richard. I ask because it's one of the most common questions
that I get -- often about WorldCat, but in general about any source of
URIs -- How do I connect my data (text forms) to their URIs? And these
questions usually come from library or archive projects with little or
no programming staff. So it seems like we need to be able to answer that
question so that people can get linked up. In fact, it seems to me that
the most pressing need right now is an easy way (or one that someone
else can do for you at a reasonable cost) to connect the text string
identifiers that we have to URIs. I envision something like what we
went through when we moved from AACR name forms to AACR2 name forms, and
libraries were able to send their MARC records to a service that
returned the records with the new name form. In this case, though, such
a service would return the data with the appropriate URIs added. (In the
case of MARC, in the $0 subfield.)

It's great that the big guys like LC and OCLC are providing URIs for
resources. But at the moment I feel like it's grapes dangling just
beyond the reach of the folks we want to connect to. Any ideas on how to
make this easy are welcome. And I do think that there's great potential
for an enterprising start-up to provide an affordable service for
libraries and archives. Of course, an open source pass in your data in
x or y format and we'll return it with URIs embedded would be great,
but I think it would be reasonable to charge for such a service.

kc


On 4/30/14, 9:59 AM, Richard Wallis wrote:

To unpack the several questions lurking in Karen’s question.

As to being able to use the WorldCat Works data/identifiers there is no
difference between a or b - it is ODC-BY licensed data.

Getting a Work URI may be easier for a) as they should be able to
identify
the OCLC Number and hence use the linked data from it’s URI 
http://worldcat.org/oclc/{ocn} to pick up the link to it’s work.

Tools such as xISBN http://xisbn.worldcat.org/xisbnadmin/doc/api.htm
can
step you towards identifier lookups and are openly available for low
volume
usage.

Citation lookup is more a bib lookup feature, that you could get an 
OCLC
Number from. One of colleagues may be helpful on the particulars of 
this.


Apologies for being WorldCat specific, but Karen did ask.

~Richard.


On 30 April 2014 17:15, Karen Coyle li...@kcoyle.net wrote:


My question has to do with discoverability. Let's say that I have a
bibliographic database and I want to add the OCLC work identifiers to
it.
Obviously I don't want to do it by hand. I might have ISBNs, but in 
some

cases I will have a regular author/title-type citation.

and let's say that I am asking this for two different institutions:
a) is an OCLC member institution
b) is not

Thanks,
kc




On 4/30/14, 8:47 AM, Dan Scott wrote:


On Tue, Apr 29, 2014 at 11:37 PM, Roy Tennant roytenn...@gmail.com
wrote:


This has now instead become a reasonable recommendation

concerning ODC-BY licensing [3] but the confusion and uncertainty
about which records an OCLC member may redistribute remains.

[3] http://www.oclc.org/news/releases/2012/201248.en.html

Allow me to try to put this confusion and uncertainty to rest

Re: [CODE4LIB] getting URIs, was: [CODE4LIB] barriers to open metadata?

2014-04-30 Thread Karen Coyle
Obviously openRefine will be used in many applications, but you've got 
to get your data TO openrefine, and you've got to do some programming to 
do that, and then to return the data to however you store it. OpenRefine 
is a great tool, but not a complete solution, IMO.


kc

On 4/30/14, 10:47 AM, Simon Brown wrote:

What about OpenRefine?


On Wed, Apr 30, 2014 at 10:34 AM, Karen Coyle li...@kcoyle.net wrote:


Thanks, Richard. I ask because it's one of the most common questions that
I get -- often about WorldCat, but in general about any source of URIs --
How do I connect my data (text forms) to their URIs? And these questions
usually come from library or archive projects with little or no programming
staff. So it seems like we need to be able to answer that question so that
people can get linked up. In fact, it seems to me that the most pressing
need right now is an easy way (or one that someone else can do for you at a
reasonable cost) to connect the text string identifiers that we have to
URIs. I envision something like what we went through when we moved from
AACR name forms to AACR2 name forms, and libraries were able to send their
MARC records to a service that returned the records with the new name form.
In this case, though, such a service would return the data with the
appropriate URIs added. (In the case of MARC, in the $0 subfield.)

It's great that the big guys like LC and OCLC are providing URIs for
resources. But at the moment I feel like it's grapes dangling just beyond
the reach of the folks we want to connect to. Any ideas on how to make this
easy are welcome. And I do think that there's great potential for an
enterprising start-up to provide an affordable service for libraries and
archives. Of course, an open source pass in your data in x or y format and
we'll return it with URIs embedded would be great, but I think it would be
reasonable to charge for such a service.

kc


On 4/30/14, 9:59 AM, Richard Wallis wrote:


To unpack the several questions lurking in Karen’s question.

As to being able to use the WorldCat Works data/identifiers there is no
difference between a or b - it is ODC-BY licensed data.

Getting a Work URI may be easier for a) as they should be able to identify
the OCLC Number and hence use the linked data from it’s URI 
http://worldcat.org/oclc/{ocn} to pick up the link to it’s work.

Tools such as xISBN http://xisbn.worldcat.org/xisbnadmin/doc/api.htm
can
step you towards identifier lookups and are openly available for low
volume
usage.

Citation lookup is more a bib lookup feature, that you could get an OCLC
Number from. One of colleagues may be helpful on the particulars of this.

Apologies for being WorldCat specific, but Karen did ask.

~Richard.


On 30 April 2014 17:15, Karen Coyle li...@kcoyle.net wrote:

  My question has to do with discoverability. Let's say that I have a

bibliographic database and I want to add the OCLC work identifiers to it.
Obviously I don't want to do it by hand. I might have ISBNs, but in some
cases I will have a regular author/title-type citation.

and let's say that I am asking this for two different institutions:
a) is an OCLC member institution
b) is not

Thanks,
kc




On 4/30/14, 8:47 AM, Dan Scott wrote:

  On Tue, Apr 29, 2014 at 11:37 PM, Roy Tennant roytenn...@gmail.com

wrote:

  This has now instead become a reasonable recommendation

concerning ODC-BY licensing [3] but the confusion and uncertainty
about which records an OCLC member may redistribute remains.

[3] http://www.oclc.org/news/releases/2012/201248.en.html

  Allow me to try to put this confusion and uncertainty to rest once

and
for
all:

ALL THE THINGS. ALL.

At least as far as we are concerned. I think it's well past time to put
the
past in the past.

  That's great, Roy. That's a *lot* simpler than parsing the

recommendations, WCRR, community norms, and such at [A, B] :)

   Meanwhile, we have just put nearly 200 million works records up as
linked


open data. [1], [2], [3]. If that doesn't rock the library open linked
data
world, then no one is paying attention.
Roy

[1] http://oclc.org/en-US/news/releases/2014/201414dublin.html
[2]
http://dataliberate.com/2014/04/worldcat-works-197-million-
nuggets-of-linked-data/
[3] http://hangingtogether.org/?p=3811

  Yes, that is really awesome. But Laura was asking about barriers to

open metadata, so damn you for going off-topic with PR around a lack
of barriers to some metadata (which, for those who have not looked
yet, have a nice ODC-BY licensing statement at the bottom of a given
Works page) :)

A. http://oclc.org/worldcat/community/record-use.en.html
B. http://oclc.org/worldcat/community/record-use/data-
licensing/questions.en.html

  --

Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet





--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet






--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] barriers to open metadata?

2014-04-30 Thread Karen Coyle
Roy, the question that I have is, as I say below, about DISCOVERABILITY 
of URIs, not intellectual property issues. It's great that there are 
lots of URIs for useful things out in the world, but they don't jump 
into your data store on their own through some kind of magic. To me, the 
big problem today is that of populating legacy data with useful 
identifiers. I know that some folks have worked at making connections 
between subject headings in their catalog and the URIs available through 
id.loc.gov - and as I recall, it turns out to be fairly frustrating. It 
seems to be that the solution to this is that providers of URIs and 
users of URIs have to both make an effort to meet half-way, or at a 
mutally convenient location. It simply is not enough to say: Hey, look! 
I've got all of these URIs. Good luck! So let's talk about how we make 
that connection.


kc

On 4/30/14, 1:17 PM, Roy Tennant wrote:

Also, this:

OCLC identifiers, and Linked Data URIs, are always in the public domain.
Independent of the data and/or information content (which may be subject to
individual licensing terms open or otherwise) that they identify, or link
to, OCLC identifiers (e.g. OCLC Numbers, VIAF IDs, or WorldCat Work URIs)
can be treated as if they are in the public domain and can be included in
any data exposure mechanism or activity as public domain data.

http://www.oclc.org/developer/develop/linked-data.en.html

Roy


On Wed, Apr 30, 2014 at 9:59 AM, Richard Wallis 
richard.wal...@dataliberate.com wrote:


To unpack the several questions lurking in Karen’s question.

As to being able to use the WorldCat Works data/identifiers there is no
difference between a or b - it is ODC-BY licensed data.

Getting a Work URI may be easier for a) as they should be able to identify
the OCLC Number and hence use the linked data from it’s URI 
http://worldcat.org/oclc/{ocn} to pick up the link to it’s work.

Tools such as xISBN http://xisbn.worldcat.org/xisbnadmin/doc/api.htm can
step you towards identifier lookups and are openly available for low volume
usage.

Citation lookup is more a bib lookup feature, that you could get an OCLC
Number from. One of colleagues may be helpful on the particulars of this.

Apologies for being WorldCat specific, but Karen did ask.

~Richard.


On 30 April 2014 17:15, Karen Coyle li...@kcoyle.net wrote:


My question has to do with discoverability. Let's say that I have a
bibliographic database and I want to add the OCLC work identifiers to it.
Obviously I don't want to do it by hand. I might have ISBNs, but in some
cases I will have a regular author/title-type citation.

and let's say that I am asking this for two different institutions:
a) is an OCLC member institution
b) is not

Thanks,
kc




On 4/30/14, 8:47 AM, Dan Scott wrote:


On Tue, Apr 29, 2014 at 11:37 PM, Roy Tennant roytenn...@gmail.com
wrote:


This has now instead become a reasonable recommendation

concerning ODC-BY licensing [3] but the confusion and uncertainty
about which records an OCLC member may redistribute remains.

[3] http://www.oclc.org/news/releases/2012/201248.en.html


Allow me to try to put this confusion and uncertainty to rest once and
for
all:

ALL THE THINGS. ALL.

At least as far as we are concerned. I think it's well past time to put
the
past in the past.


That's great, Roy. That's a *lot* simpler than parsing the
recommendations, WCRR, community norms, and such at [A, B] :)

  Meanwhile, we have just put nearly 200 million works records up as

linked

open data. [1], [2], [3]. If that doesn't rock the library open linked
data
world, then no one is paying attention.
Roy

[1] http://oclc.org/en-US/news/releases/2014/201414dublin.html
[2]
http://dataliberate.com/2014/04/worldcat-works-197-million-
nuggets-of-linked-data/
[3] http://hangingtogether.org/?p=3811


Yes, that is really awesome. But Laura was asking about barriers to
open metadata, so damn you for going off-topic with PR around a lack
of barriers to some metadata (which, for those who have not looked
yet, have a nice ODC-BY licensing statement at the bottom of a given
Works page) :)

A. http://oclc.org/worldcat/community/record-use.en.html
B. http://oclc.org/worldcat/community/record-use/data-
licensing/questions.en.html


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet




--
Richard Wallis
Founder, Data Liberate
http://dataliberate.com
Tel: +44 (0)7767 886 005

Linkedin: http://www.linkedin.com/in/richardwallis
Skype: richard.wallis1
Twitter: @rjw



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] barriers to open metadata?

2014-04-30 Thread Karen Coyle

On 4/30/14, 6:37 PM, Roy Tennant wrote:

In the end there may need to be reconciliation services just like we had
similar services in the card-catalog-to-digital years.
Roy
Roy, yes, that's what I'm assuming. I think we are indeed in the same 
leaky boat we were in in the 1970's when all of a sudden we realized 
that in the future we wanted our data to be digital but most of what 
we had was definitely analog. In the early days, we thought it was an 
impossible task to convert our cards to MARC, but it turned out to be 
possible.


I believe that linking our heading strings (the ones that hopefully 
resemble the prefLabel on someone's authority file) to identifiers is 
not as hard as people assume, especially if we have systems that can 
learn -- that is, that can build up cases of synonyms (e.g. Smith, 
John with title Here's my book == Smith, John J. with title Here's 
my book). This is what the AACR-AACR2 services did. OCLC surely does a 
lot of this when merging manifestations, and undoubtedly did so when 
determining what are works, and when bringing authority entries together 
for VIAF. No, you don't get 100% perfection, but we don't get that now 
with any of our services.


And for all of those who keep suggesting Open Refine -- it's like you 
walk into bakery to buy a cake and they hand you flour, eggs, milk and 
show you where the oven is. Yes, it can be done. But you want the cake 
-- if you could do and wanted to *make* a cake you wouldn't be in the 
bakery, you'd be home in your kitchen. So in case it isn't clear, I'm 
talking cake, not cake making. How are we going to provide cake to the 
library and archives masses? And, if you are feeling entrepreneurial, 
wouldn't this be a good time to open a bakery?


kc




On Wed, Apr 30, 2014 at 5:53 PM, Karen Coyle li...@kcoyle.net wrote:


Roy, the question that I have is, as I say below, about DISCOVERABILITY of
URIs, not intellectual property issues. It's great that there are lots of
URIs for useful things out in the world, but they don't jump into your data
store on their own through some kind of magic. To me, the big problem today
is that of populating legacy data with useful identifiers. I know that some
folks have worked at making connections between subject headings in their
catalog and the URIs available through id.loc.gov - and as I recall, it
turns out to be fairly frustrating. It seems to be that the solution to
this is that providers of URIs and users of URIs have to both make an
effort to meet half-way, or at a mutally convenient location. It simply is
not enough to say: Hey, look! I've got all of these URIs. Good luck! So
let's talk about how we make that connection.

kc

On 4/30/14, 1:17 PM, Roy Tennant wrote:


Also, this:

OCLC identifiers, and Linked Data URIs, are always in the public domain.
Independent of the data and/or information content (which may be subject
to
individual licensing terms open or otherwise) that they identify, or link
to, OCLC identifiers (e.g. OCLC Numbers, VIAF IDs, or WorldCat Work URIs)
can be treated as if they are in the public domain and can be included in
any data exposure mechanism or activity as public domain data.

http://www.oclc.org/developer/develop/linked-data.en.html

Roy


On Wed, Apr 30, 2014 at 9:59 AM, Richard Wallis 
richard.wal...@dataliberate.com wrote:

  To unpack the several questions lurking in Karen’s question.

As to being able to use the WorldCat Works data/identifiers there is no
difference between a or b - it is ODC-BY licensed data.

Getting a Work URI may be easier for a) as they should be able to
identify
the OCLC Number and hence use the linked data from it’s URI 
http://worldcat.org/oclc/{ocn} to pick up the link to it’s work.

Tools such as xISBN http://xisbn.worldcat.org/xisbnadmin/doc/api.htm
can
step you towards identifier lookups and are openly available for low
volume
usage.

Citation lookup is more a bib lookup feature, that you could get an OCLC
Number from. One of colleagues may be helpful on the particulars of this.

Apologies for being WorldCat specific, but Karen did ask.

~Richard.


On 30 April 2014 17:15, Karen Coyle li...@kcoyle.net wrote:

  My question has to do with discoverability. Let's say that I have a

bibliographic database and I want to add the OCLC work identifiers to
it.
Obviously I don't want to do it by hand. I might have ISBNs, but in some
cases I will have a regular author/title-type citation.

and let's say that I am asking this for two different institutions:
a) is an OCLC member institution
b) is not

Thanks,
kc




On 4/30/14, 8:47 AM, Dan Scott wrote:

  On Tue, Apr 29, 2014 at 11:37 PM, Roy Tennant roytenn...@gmail.com

wrote:

  This has now instead become a reasonable recommendation

concerning ODC-BY licensing [3] but the confusion and uncertainty
about which records an OCLC member may redistribute remains.

[3] http://www.oclc.org/news/releases/2012/201248.en.html

  Allow me to try to put this confusion and uncertainty to rest

Re: [CODE4LIB] Enough PHP To Be Dangerous At ALA Annual

2014-04-11 Thread Karen Coyle

Tim,

Good for you! I would find this more useful than about 99.2% of the 
offerings at ALA. If I were to attend the conference, I would definitely 
be there. In fact, Las Vegas is close enough that I will think about 
coming just for this.


kc

On 4/11/14, 5:00 PM, Tim Spalding wrote:

For fun I'm throwing a day-long, almost-free introduction to PHP session
alongside the pre-conference day at ALA Annual.

It's called Enough PHP To Be Dangerous, and will be held Friday, June 27,
2014 a few blocks from the Convention Center.

Blog post is here:
http://blog.librarything.com/thingology/2014/04/come-learn-php-at-ala-2014/

It's for people who don't know much about programming. So, it's NOT FOR
YOU. You'd be bored! But you might want to tell a colleague. And if you're
interested in helping out, it would be great having another person or two
helping attendees figure out why their script won't run.

I'm doing it for fun and because I can. Conferences are ridiculously
expensive to exhibit at, so why not do a little more with it? For
simplicity's sake, I didn't hook up with LITA or anyone else; I just rented
a room and contracted for coffee and sandwiches.

If you're interested in going, or helping out, read the blog post and email
me.

Best,
Tim



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] New Zealand Chapter

2014-04-09 Thread Karen Coyle
Really, Jay, you've got the t-shirt and the logo - what else could be 
needed? :-)


My take: reach out to a small but vibrant group; make sure it is diverse 
(type of library, type of person), and have a first meeting about how to 
have a second meeting. The first meeting may need to be folks in a 
nearby locale; the second would reach out to institutions that can add 
some support (space, time, motion).


kc

On 4/9/14, 9:17 PM, Jay Gattuso wrote:

Hi all,

Long time listener, first time caller.

We don't have a C4L chapter over here in New Zealand, and I wondered what we 
would need to do to align the small group of Lib  / GLAM coders with the 
broader C4L group.

One of my colleagues did make this: http://i.imgur.com/XgGP9vX.jpg

We are  also setting up a two day code/hack fest, focusing on our Digital 
Preservation concerns, in June.

I'd also really like to run the hackfest under a C4L banner.

Any thoughts?

J

Jay Gattuso | Digital Preservation Analyst | Preservation, Research and 
Consultancy
National Library of New Zealand | Te Puna Mātauranga o Aotearoa
PO Box 1467 Wellington 6140 New Zealand | +64 (0)4 474 3064
jay.gatt...@dia.govt.nzmailto:jay.gatt...@natlib.govt.nz


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] code4lib italy

2014-03-11 Thread Karen Coyle
I happen to be in Italy at the moment, having just attended a conference 
on digital libraries here. [1] My general impression is that actual 
coding in libraries is rare, but I could ask around among librarians I 
know here and see what they say. There are some libraries that have had 
to create their own systems, and this might be a good outlet for those 
folks. It could also bring out projects that aren't well known.


kc
[1] http://www.aib.it/attivita/congressi/c2014/fsr2014/

On 3/10/14, 7:54 PM, Eric Lease Morgan wrote:

I wonder whether there are enough people and enough interest to organize a 
Code4Lib Italy event. Hmmm... —Eric Lease Morgan


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] code4lib italy

2014-03-11 Thread Karen Coyle
This is great. As soon as there is a group convened, we can try to find 
some way to coordinate for a meeting between US and Italy. Then, of 
course, there is the rest of the globe to cover... ;-)


kc

On 3/11/14, 9:22 AM, Nicola Carboni wrote:

Hi,
I can surely say that the interest on the topic is very high, and there is a 
community (maybe not so big as in U.S.) that would definitively participate.
I would definitively going on with this idea, also as a starting point to 
create a group of librarian/developer in Italy
I will try to forward this to some people that are interested on the topic.

Nicola




--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Fwd: RFC 7154 - IETF Guidelines for Conduct

2014-03-03 Thread Karen Coyle
Wow. I really like the IETF policy. It's not so much a code of conduct, 
per se, but it's a great statement of who we are. I think a discussion 
along those lines for c4l would be very interesting. Where IETF has a 
statement like:


 3. IETF participants devise solutions for the global Internet that 
meet the needs of diverse technical and operational environments.


...c4l could say something about promoting solutions for all types and 
all sizes of libraries. A statement about sharing would also be nice. 
And maybe something about welcoming folks with various levels of 
technical expertise.


kc

On 3/3/14, 10:36 PM, Peter Murray wrote:

Code4LibCon is coming up in a few weeks.  I’m sure there will be reminders 
about the community-generated code of conduct 
(https://github.com/code4lib/antiharassment-policy/blob/master/code_of_conduct.md).
  I think it also useful to take a look at what just passed the IETF that 
governs their meetings.  In particular, I like how it gets past the “Thou shalt 
not” and gets to the “Thou shall”.

http://tools.ietf.org/html/rfc7154


Peter
--
Peter Murray
Assistant Director, Technology Services Development
LYRASIS
peter.mur...@lyrasis.org
+1 678-235-2955
800.999.8558 x2955


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Welcome to Roy4Lib

2014-02-21 Thread Karen Coyle
Remember, folks, this is the Roy whose office number at the CDL was 
404. And it was true.


kc

On 2/21/14, 3:13 PM, Roy Tennant wrote:

roy4lib.org is ALWAYS down. I mean, it just makes too much sense for it to
be in any other state.
Roy


On Fri, Feb 21, 2014 at 2:31 PM, Rosalyn Metz rosalynm...@gmail.com wrote:


it appears that roy4lib.org is also down


On Fri, Feb 21, 2014 at 1:59 PM, Frumkin, Jeremy 
frumk...@u.library.arizona.edu wrote:


Welcome to the Roy4Lib discussion list. This list is intended to
facilitate discussion on Roy Tennant's new world library order, the role
of bacon (including kosher and vegetarian based varieties) in this
context, and the long, long, long, long, long drawn out death of MARC.

If you believe you have subscribed to this list in error, please email

the

admin at r...@roy4lib.org.




Jeremy Frumkin
Assistant Dean / Chief Technology Strategist
University of Arizona Libraries

+1 520.626.7296
frumk...@u.library.arizona.edu

Any intelligent fool can make things bigger and more complex... It takes
a touch of genius - and a lot of courage to move in the opposite
direction. - Albert Einstein



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Fwd: [rules] Publication of the RDA Element Vocabularies

2014-01-24 Thread Karen Coyle

On 1/24/14, 6:56 AM, Jon Phipps wrote:


Thanks for reminding me that this is an academic panel discussion in front
of an audience, rather than a conversation.



Not entirely clear what you meant by that, but I do think that we have a 
very practical issue in front of us, and it's one of the things that, 
IMO, is holding back the adoption of linked data: the limitations of the 
tools in this area. As I said above, there is no reason why we should be 
working with raw URIs in our work, but few tools present the 
human-readable labels to the developer. So we are unfortunately forced 
to work directly with rdaa:P50209 even though we would prefer to be 
working with addressee of (the rdfs:label). Although we shouldn't be 
designing vocabularies to make up for the limitations of the tools, it's 
basically inevitable if we want to get things done. (There are, BTW, 
enterprise-level tools, but they are beyond the $$ of most folks on this 
list.)


I also think that rdfs:label presents us with the same problem that we 
found with SKOS that led to SKOS-XL and content as text -- there are 
times when you need to say something more about the label; more than 
what language it is in. It seems quite logical to me that you would have 
one label for experts, another for the general public; one label for 
those doing input, another for your basic UI; one label for children, 
another for adults; etc. You could do that in your application software, 
but then you aren't sharing it. That you found the need for a local 
reg:name is evidence of this, but it, too, will prove to be inadequate 
for some needs.


kc

--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Fwd: [rules] Publication of the RDA Element Vocabularies

2014-01-23 Thread Karen Coyle

On 1/23/14, 4:01 PM, Robert Sanderson wrote:
So in my opinion, as is everything in the mail of course, this is even 
worse. Now instead of 1600 properties, you have 1600 * (number of 
languages +1) properties. And you're going to see them appearing in 
uses of the ontology. Either stick with your opaque identifiers or 
pick a language for the readable ones, and best practice would be 
English, but doing both is a disaster in the making.


Actually, it's more than that. Because, as you see below, for each 
property there are two URIs, a reg:name, and an rdfs:label. The lexical 
URI is based on the reg:name not the rdfs:label. So that makes one 
opaque identifier, one lexical identifer, and two display forms. The 
reg:name is camel case and condensed, as compared to the label, which is 
space-delimited words. Both have language designations.


The rdfs:label is written as a verb phrase, with either has or is -- 
a form that I generally find annoying in practice -- while the reg:name 
is noun-ish. I'm sympathetic to the has form in some situations, 
such as when reading a triple as a statement, but the verb-enhanced form 
would seem odd on an input form, for example. Having both could be 
useful, but since RDF doesn't recognize label types it only works in a 
closed world.


!--Property: has address of the corporate body--
rdf:Description rdf:about=http://rdaregistry.info/Elements/a/P50036;
  rdfs:isDefinedBy rdf:resource=http://rdaregistry.info/Elements/a/; /
  reg:status rdf:resource=http://metadataregistry.org/uri/RegStatus/1001; /
  reg:name xml:lang=enaddressOfTheCorporateBody/reg:name
  rdfs:label xml:lang=enhas address of the corporate body/rdfs:label
  skos:definition xml:lang=enRelates a corporate body to the address of a 
corporate body's headquarters or offices, or an e-mail or internet address for the 
body./skos:definition
  rdf:type rdf:resource=http://www.w3.org/1999/02/22-rdf-syntax-ns#Property; 
/
  rdfs:domain rdf:resource=http://rdaregistry.info/Elements/c/C10005; /
  rdfs:subPropertyOf rdf:resource=http://rdaregistry.info/Elements/u/P60512; 
/
  owl:sameAs 
rdf:resource=http://rdaregistry.info/Elements/a/addressOfTheCorporateBody; /
/rdf:Description


kc

--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Fwd: [rules] Publication of the RDA Element Vocabularies

2014-01-22 Thread Karen Coyle
 are distributed in sets
(the number of elements in each set is given in brackets):
Agent properties
[http://metadataregistry.org/schema/show/id/81.html]http://metadataregistry.org/schema/show/id/81.html(226)
Expression properties
[http://metadataregistry.org/schema/show/id/78.html]http://metadataregistry.org/schema/show/id/78.html(236)
Item properties
[http://metadataregistry.org/schema/show/id/80.html]http://metadataregistry.org/schema/show/id/80.html(54)
Manifestation properties
[http://metadataregistry.org/schema/show/id/79.html]http://metadataregistry.org/schema/show/id/79.html(213)
Work properties
[http://metadataregistry.org/schema/show/id/77.html]http://metadataregistry.org/schema/show/id/77.html(232)
Unconstrained properties
[http://metadataregistry.org/schema/show/id/82.html]http://metadataregistry.org/schema/show/id/82.html(698)
Classes 
[http://metadataregistry.org/schema/show/id/83.html]http://metadataregistry.org/schema/show/id/83.html(8)

Follow the links to see details of each element set.

Questions or comments on the content of the element sets may be addressed
to the Chair of the JSC, Gordon Dunsire [jsch...@rdatoolkit.org]. Questions
and comments on the encoding of the vocabularies or on the Open Metadata
Registry may be addressed to Diane Hillmann [metadata.ma...@gmail.com].


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Fwd: [rules] Publication of the RDA Element Vocabularies

2014-01-22 Thread Karen Coyle
On 1/22/14, 3:17 PM, Robert Sanderson wrote:

 The we might change what this means argument is also problematic -- if
 you change what it means, then you should change the URI! Otherwise people
 will continue to use them incorrectly, plus the legacy data generated with
 the previous definition will suddenly change what it's saying.
Rob, absolutely right. If the semantics change, then you need a new
property. But labels can change (or more can be added). However, the
library world still equates labels with data -- that is, that our data
is one-to-one with what we display. That's a huge problem, and it's very
hard getting people to think differently about that.

I've been looking at the output of the RDA vocabularies over the last
couple of days and it IS quite difficult to do so with properties named
something like P3058. There is a strong case to be made for mnemonics,
although I also take Jon's point that when a property has a name like
dc:title it is easy for folks to assume they really know what it
means. I would still prefer something memorable at this stage.


 Finally, 1600 properties... good luck with that.
Yes. And remember, RDA was designed to be a *simpler* cataloging code.
Can you imagine if it weren't?!

kc



 Rob



 On Wed, Jan 22, 2014 at 3:03 PM, Hamilton, Gill g.hamil...@nls.uk wrote:

 Je ne comprends pas l'anglais.
 Je ne comprends pas l'URI otherDesignationAssociatedWithTheCorporateBody

 私は日本人です。私は理解していない、そのURI

 Opaque URIs with human readable labels helps in an international context.

 Just my two yens worth :)
 G

 -
 Gill Hamilton
 Digital Access Manager
 National Library of Scotland
 George IV Bridge
 Edinburgh EH1 1EW, Scotland
 e: g.hamil...@nls.uk
 t: +44 (0)131 623 3770
 Skype: gill.hamilton.nls

 
 From: Code for Libraries [CODE4LIB@LISTSERV.ND.EDU] on behalf of Dan
 Scott [deni...@gmail.com]
 Sent: 22 January 2014 21:10
 To: CODE4LIB@LISTSERV.ND.EDU
 Subject: Re: [CODE4LIB] Fwd: [rules] Publication of the RDA Element
 Vocabularies

 Hi Karen:

 On Wed, Jan 22, 2014 at 3:16 PM, Karen Coyle li...@kcoyle.net wrote:
 I can't address the first points, but I can speak a bit to the question
 of
 meaningful URIs. In the original creation of the RDA elements,
 meaningful
 URIs were used based on the actual RDA terminology. This resulted in URIs
 like:


 http://rdvocab.info/Elements/alternativeChronologicalDesignationOfLastIssueOrPartOfSequence
 and...


 http://rdvocab.info/Elements/alternativeChronologicalDesignationOfLastIssueOrPartOfSequenceManifestation
 Not only that, the terminology for some elements changed over time,
 which in
 some cases meant deprecating a property that was then overly confusing
 based
 on its name.

 Now, I agree that one possibility would have been for the JSC to develop
 meaningful but reasonably short property names. Another possibility is
 that
 we cease looking at URIs and begin to work with labels, since URIs are
 for
 machines and labels are for humans. Unfortunately, much RDF software
 still
 expects you to work with the underlying URI rather than the human-facing
 label. We need to get through that stage as quickly as possible, because
 it's causing us to put effort into URI naming that would be best used
 for
 other analysis activities.
 Thanks for responding on this front. I understand that, while the
 vocabulary was in heavy active development it might have been painful
 to adjust as elements changed, but given that this marks the actual
 publication of the vocabulary, that churn should have settled down,
 and then this part of the JSC's contribution to semantic web could
 have semantics applied at both the micro and macro level.

 I guess I see URIs as roughly parallel to API names; as long as humans
 are assembling programs, we're likely to benefit from having
 meaningful (no air quotes required) names... even if sometimes the
 meaning drifts over time and the code  APIs need to be refactored.
 Dealing with sequentially numbered alphanumeric identifiers reminds me
 rather painfully of MARC.

 For what it's worth (and it might not be worth much) curl
 http://rdaregistry.info/Elements/a/P50101 | grep reg:name | sort |
 uniq -c shows that the reg:name property is unique across all of the
 agent properties, at least. Remnants of the earlier naming effort? If
 that pattern holds, those could have been simply used for the
 identifiers in place of P#. The most unwieldy of those appears
 to be otherDesignationAssociatedWithTheCorporateBody (which _is_
 unwieldy, certainly, but still more meaningful than
 http://rdaregistry.info/Elements/a/P50033).

 Perhaps it's not too late?
 Follow us on Twitter and Facebook

 National Library of Scotland, Scottish Charity, No: SCO11086

 This communication is intended for the addressee(s) only. If you are not
 the addressee please inform the sender and delete the email from your
 system. The statements and opinions expressed

Re: [CODE4LIB] rdf ontologies for archival descriptions

2014-01-19 Thread Karen Coyle
There is an ontology for archival description developed by the Europeana 
project. The web site for the ontology is here:

http://pro.europeana.eu/edm-documentation
and there are some good articles about it here and here:
http://ontogenealogy.com/europeana-data-model-edm/
http://semanticweb.com/tag/europeana-data-model

Note that Europeana provides an XSD version of the ontology because that 
is what most of the contributors could handle. (The list of contributors 
is here: http://www.europeana.eu/portal/europeana-providers.html).


Europeana shows on the linked data cloud as a major content provider.

When looking for vocabularies, you might want to have a look at the 
Linked Open Vocabularies site:

   http://lov.okfn.org/dataset/lov/
This lists all of the vocabularies in use in the LOD cloud.

You can search for properties, and a search on time or place gives 
you many options.


Note that these are properties, and there is still the question of 
values or objects. For example, if you are going to be including 
geographic places in your ontology, then you probably want to use URIs 
from the geoNames database. (http://www.geonames.org).


kc

On 1/18/14, 6:39 PM, Eric Lease Morgan wrote:

If you were to select a set of RDF ontologies intended to be used in the linked 
data of archival descriptions, then what ontologies would you select?

For simplicity's sake, RDF ontologies are akin to the fields in MARC records or the 
entities in EAD/XML files. Articulated more accurately, they are the things denoting 
relationships between subjects and objects in RDF triples. In this light, they are akin 
to the verbs in all but the most simplistic of sentences. But if they are akin to verbs, 
then they bring with them all of the nuance and subtlety of human written language. And 
human written language, in order to be an effective human communications device, comes 
with two equally important prerequisites: 1) a writer who can speak to an intended 
audience, and 2) a reader with a certain level of intelligence. A writer who does not use 
the language of the intended audience speaks to few, and a reader who does not 
bring something to the party goes away with little understanding. Because the 
effectiveness of every writer is not perfect, and because not every reader comes to the 
party with a certain level of intelligence, wr!

it!

  ten language is imperfect. Similarly, the ontologies of linked data are 
imperfect. There are no perfect ontologies nor absolutely correct uses of them. 
There are only best practices and common usages.

This being the case, ontologies still need to be selected in order for linked 
data to be manifested. What ontologies would you suggest be used when creating 
linked data for archival descriptions? Here are a few possibilities, listed in 
no priority order:

   * Dublin Core Terms - This ontology is rather bibliographic in
 nature, and provides a decent framework for describing much of
 the content of archival descriptions.

   * FOAF - Archival collections often originate from individual
 people. Such is the scope of FOAF, and FOAF is used by a number
 of other sets of linked data.

   * MODS - Because many archival descriptions are rooted in MARC
 records, and MODS is easily mapped from MARC.

   * Schema.org - This is an up-and-coming ontology heralded by the
 600-pound gorillas in the room -- Google, Microsoft, Yahoo, etc.
 While the ontology has not been put into practice for very long,
 it is growing and wide ranging.

   * RDF - This ontology is necessary because linked data is
 manifested as... RDF

   * RDFS - This ontology may be necessary because the archival
 community may be creating some of its own ontologies.

   * OWL and SKOS - Both of these ontologies seem to be used to
 denote relationships between terms in other ontologies. In this
 way they are used to create classification schemes and thesauri.
 For example, they allow the implementor to denote creator in one
 ontology is the same as author in another ontology. Or they
 allow country in one ontology to be denoted as a parent
 geographic term for city in another ontology.

While some or all of these ontologies may be useful for linked data of archival 
descriptions, what might some other ontologies include? (Remember, it is often 
better to select existing ontologies rather than inventing, unless there is 
something distinctly unique about a particular domain.) For example, how about an 
ontology denoting times? Or how about one for places? FOAF is good for people, but what 
about organizations or institutions?

Inquiring minds would like to know.

—
Eric Morgan


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] COinS metadata format support

2014-01-18 Thread Karen Coyle
Roy, I'm not sure what tips you over into sarcasm mode (unless it's 
anything I say), but 1) the answer is a few posts down, albeit not in 
any detail 2) as a member-based organization that exists to serve its 
members, I would think that OCLC would want to encourage  the gathering 
of information about members needs, rather than responding to them 
negatively.


The transition to a new data model, and in particular the use of 
identifiers for things, is (IMO) going to open up a lot of opportunities 
for services. This will be like the transition to MARC from cards 
(retrospective conversion) and the transition from AACR headings to 
AACR2 headings, on steroids. It's not too early to start, and one of the 
early steps is that of matching our string-based data with URIs for the 
same thing. I'd be interested in engaging in some brainstorming about 
how to make this happen as efficiently and cost effective as possible. 
(Maybe this would be a good BOF at c4l14?)


kc


On 1/17/14, 2:10 PM, Roy Tennant wrote:

I agree that most Schema.org markup would be generated on the fly from a
database, so I'm a bit unclear on what kind of tool we (OCLC) or someone
else is being asked to develop. Perhaps someone could spell it out for this
dense person? Thanks,
Roy


On Fri, Jan 17, 2014 at 12:02 PM, Karen Coyle li...@kcoyle.net wrote:


Dave, my impression is that most schema.org code is being generated
programmatically out of database-driven web sites. The list of tools (not
many) is here:

http://schema.rdfs.org/tools.html

As OCLC is very involved in schema.org, it would make sense to encourage
them to develop something. Looking at what they do today in WorldCat, the
big advantage that they seem to have is that they can locate URIs for names
(VIAF) and subjects (LCSH).

kc

On 1/17/14, 10:49 AM, Bigwood, David wrote:


Is there an easy-to-use tool for schema.org microdata? I was fond of the
OCLC NJ COinS tool, but it has gone. Anything of that ilk?

Thanks,
Dave

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of
Michael J. Giarlo
Sent: Friday, January 17, 2014 12:34 PM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] COinS metadata format support

Chad,

In that case, I wonder if you might get more mileage out of
schema.orgmicrodata instead of COinS.  There are undoubtedly more clients
out there that can make sense of HTML5 microdata than COinS, which is
really showing its age and is a niche technology.

-Mike



On Fri, Jan 17, 2014 at 10:25 AM, Chad Mills cmmi...@rci.rutgers.edu
wrote:

  Jodi,

No.  I am working with our repository resources which is an eclectic
mixture of resource types.  I just want to simply embed our metadata
in our search results and record displays for other tools to use.  It
seems cheap and reasonable to do I just didn't want to limit this
feature to only certain resource types.

Best,
Chad


- Original Message -
From: Jodi Schneider jschnei...@pobox.com
To: CODE4LIB@LISTSERV.ND.EDU
Sent: Friday, January 17, 2014 12:54:43 PM
Subject: Re: [CODE4LIB] COinS metadata format support

Hi Chad,

Are these subscription images/video/audio that libraries have? The
original purpose of COinS, as I understand it, was to get people to
subscription copies. Depending on what you're doing (i.e. the
purpose/intended use) there might be a better standard these days.

In case it helps there's more info here:
http://ocoins.info/
(though it looks like the generator isn't up any longer, maybe due to
OCLC New Jersey hosting?)

Hopefully you'll get some more helpful advice from others!

-Jodi




On Fri, Jan 17, 2014 at 5:36 PM, Chad Mills cmmi...@rci.rutgers.edu
wrote:

  I was able to easily find and create COinS for books and journals.

I started thinking about images, video, audio, etc.  I see
references to 'info:ofi/fmt:kev:mtx:unknown' as a rft_val_fmt value
some places.  I


would


assume if I went down that road the rft.genre would have a value of
'unknown' as well.  Is there some other alternative I am missing
when handling other formats?

Thanks!

--
Chad Mills
Digital Library Architect
Ph: 848.932.5924
Fax: 848.932.1386
Cell: 732.309.8538

Rutgers University Libraries
Scholarly Communication Center
Room 409D, Alexander Library
169 College Avenue, New Brunswick, NJ 08901

http://rucore.libraries.rutgers.edu/



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] COinS metadata format support

2014-01-17 Thread Karen Coyle
The missing formats are missing because they were never defined in the 
OpenURL standard. There is a registry of formats [1] that was designed 
to be updatable, although few updates have been done. I worked on the 
document type standards that are there today, and would happily help 
develop additional types if we can get some assurance that they could be 
added to the standard. I honestly do not know what mechanism exists for 
adding new types, but if anyone is interested we could ping the 
appropriate folks at OCLC and see.


kc
[1] 
http://alcme.oclc.org/openurl/servlet/OAIHandler?verb=ListRecordsmetadataPrefix=oai_dcset=Core:Metadata+Formats


On 1/17/14, 8:36 AM, Chad Mills wrote:

I was able to easily find and create COinS for books and journals.  I started 
thinking about images, video, audio, etc.  I see references to 
'info:ofi/fmt:kev:mtx:unknown' as a rft_val_fmt value some places.  I would 
assume if I went down that road the rft.genre would have a value of 'unknown' 
as well.  Is there some other alternative I am missing when handling other 
formats?

Thanks!

--
Chad Mills
Digital Library Architect
Ph: 848.932.5924
Fax: 848.932.1386
Cell: 732.309.8538

Rutgers University Libraries
Scholarly Communication Center
Room 409D, Alexander Library
169 College Avenue, New Brunswick, NJ 08901

http://rucore.libraries.rutgers.edu/


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] COinS metadata format support

2014-01-17 Thread Karen Coyle
Dave, my impression is that most schema.org code is being generated 
programmatically out of database-driven web sites. The list of tools 
(not many) is here:


http://schema.rdfs.org/tools.html

As OCLC is very involved in schema.org, it would make sense to encourage 
them to develop something. Looking at what they do today in WorldCat, 
the big advantage that they seem to have is that they can locate URIs 
for names (VIAF) and subjects (LCSH).


kc

On 1/17/14, 10:49 AM, Bigwood, David wrote:

Is there an easy-to-use tool for schema.org microdata? I was fond of the OCLC 
NJ COinS tool, but it has gone. Anything of that ilk?

Thanks,
Dave

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Michael 
J. Giarlo
Sent: Friday, January 17, 2014 12:34 PM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] COinS metadata format support

Chad,

In that case, I wonder if you might get more mileage out of schema.orgmicrodata 
instead of COinS.  There are undoubtedly more clients out there that can make 
sense of HTML5 microdata than COinS, which is really showing its age and is a 
niche technology.

-Mike



On Fri, Jan 17, 2014 at 10:25 AM, Chad Mills cmmi...@rci.rutgers.eduwrote:


Jodi,

No.  I am working with our repository resources which is an eclectic
mixture of resource types.  I just want to simply embed our metadata
in our search results and record displays for other tools to use.  It
seems cheap and reasonable to do I just didn't want to limit this
feature to only certain resource types.

Best,
Chad


- Original Message -
From: Jodi Schneider jschnei...@pobox.com
To: CODE4LIB@LISTSERV.ND.EDU
Sent: Friday, January 17, 2014 12:54:43 PM
Subject: Re: [CODE4LIB] COinS metadata format support

Hi Chad,

Are these subscription images/video/audio that libraries have? The
original purpose of COinS, as I understand it, was to get people to
subscription copies. Depending on what you're doing (i.e. the
purpose/intended use) there might be a better standard these days.

In case it helps there's more info here:
http://ocoins.info/
(though it looks like the generator isn't up any longer, maybe due to
OCLC New Jersey hosting?)

Hopefully you'll get some more helpful advice from others!

-Jodi




On Fri, Jan 17, 2014 at 5:36 PM, Chad Mills cmmi...@rci.rutgers.edu
wrote:


I was able to easily find and create COinS for books and journals.
I started thinking about images, video, audio, etc.  I see
references to 'info:ofi/fmt:kev:mtx:unknown' as a rft_val_fmt value
some places.  I

would

assume if I went down that road the rft.genre would have a value of
'unknown' as well.  Is there some other alternative I am missing
when handling other formats?

Thanks!

--
Chad Mills
Digital Library Architect
Ph: 848.932.5924
Fax: 848.932.1386
Cell: 732.309.8538

Rutgers University Libraries
Scholarly Communication Center
Room 409D, Alexander Library
169 College Avenue, New Brunswick, NJ 08901

http://rucore.libraries.rutgers.edu/



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Mapping LCSH to DDC

2013-12-10 Thread Karen Coyle
I've often thought that this would be an interesting exercise if someone 
would undertake it.


Just a reminder: in theory (IN THEORY) the first subject heading in an 
LC record is the one most semantically close to the assigned subject 
classification. So perhaps a first pass with the FIRST 6xx might give a 
more refined matching. And then it would be interesting to compare that 
with the results using all 600-651's.


kc

On 12/10/13, 1:18 PM, Edward Summers wrote:

Not a naive idea at all. If you have the stomach for it, you could extract the 
Subject Heading / Dewey combinations out of say the LC Catalog MARC data [1] to 
use as training data for some kind of clustering [2] algorithm. You might even 
be able to do something simple like keep a count of the Dewey ranges associated 
with each subject heading.

I’m kind of curious myself, so I could work on getting the subject heading / 
dewey combinations if you want?

//Ed

[1] https://archive.org/details/marc_records_scriblio_net
[2] https://en.wikipedia.org/wiki/Cluster_analysis

On Dec 10, 2013, at 8:18 AM, Irina Arndt ar...@mpdl.mpg.de wrote:


Hi CODE4LIB,

we would like to add DDC classes to a bunch of MARC records, which contains 
only LoC Subject Headings.
Does anybody know, if a mapping between LCSH and DDC is anywhere existent (and 
available)?

I understood, that WebDewey http://www.oclc.org/dewey/versions/webdewey.en.html 
 might provide such a service, but

· we are no OCLC customers or subscribers to WebDewey

· even if we were, I'm not sure, if the service matches our needs

I'm thinking of a tool, where I can upload my list of subject headings and get 
back a list, where the matching Dewey classes have been added (but a 'simple' 
csv file with LCSH terms and DDC classes would be helpful as well- I am fully 
aware, that neither LCSH nor DDC are simple at all...) . Naïve idea...?

Thanks for any clues,
Irina


---

Irina Arndt
Max Planck Digital Library (MPDL)
Library System Coordinator
Amalienstr. 33
D-80799 Muenchen, Germany

Tel. +49 89 38602-254
Fax +49 89 38602-290

Email: ar...@mpdl.mpg.demailto:ar...@mpdl.mpg.de
http://www.mpdl.mpg.de


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Mapping LCSH to DDC

2013-12-10 Thread Karen Coyle
Yeah, Ed! I'm totally looking forward to results. Unlikely as it is, if 
there's anything I can do


and I understand about limiting to 650, but ... well, let's see how it goes.

kc

On 12/10/13, 1:37 PM, Edward Summers wrote:

I was going to try to reduce the space a bit by focusing on 650 fields. Each 
record with a Dewey number will be a tab separated line, that will include each 
650 field in order. So something like:

305.42/0973 tab Women's rights -- United States -- History -- Sources. tab Women 
-- United States -- History — Sources tab Manuscripts, American -- Facsimiles.

I thought it might be a place to start at least … it’s running on an ec2 
instance right now :-)

//Ed

On Dec 10, 2013, at 4:26 PM, Karen Coyle li...@kcoyle.net wrote:


I've often thought that this would be an interesting exercise if someone would 
undertake it.

Just a reminder: in theory (IN THEORY) the first subject heading in an LC 
record is the one most semantically close to the assigned subject 
classification. So perhaps a first pass with the FIRST 6xx might give a more 
refined matching. And then it would be interesting to compare that with the 
results using all 600-651's.

kc

On 12/10/13, 1:18 PM, Edward Summers wrote:

Not a naive idea at all. If you have the stomach for it, you could extract the 
Subject Heading / Dewey combinations out of say the LC Catalog MARC data [1] to 
use as training data for some kind of clustering [2] algorithm. You might even 
be able to do something simple like keep a count of the Dewey ranges associated 
with each subject heading.

I’m kind of curious myself, so I could work on getting the subject heading / 
dewey combinations if you want?

//Ed

[1] https://archive.org/details/marc_records_scriblio_net
[2] https://en.wikipedia.org/wiki/Cluster_analysis

On Dec 10, 2013, at 8:18 AM, Irina Arndt ar...@mpdl.mpg.de wrote:


Hi CODE4LIB,

we would like to add DDC classes to a bunch of MARC records, which contains 
only LoC Subject Headings.
Does anybody know, if a mapping between LCSH and DDC is anywhere existent (and 
available)?

I understood, that WebDewey http://www.oclc.org/dewey/versions/webdewey.en.html 
 might provide such a service, but

· we are no OCLC customers or subscribers to WebDewey

· even if we were, I'm not sure, if the service matches our needs

I'm thinking of a tool, where I can upload my list of subject headings and get 
back a list, where the matching Dewey classes have been added (but a 'simple' 
csv file with LCSH terms and DDC classes would be helpful as well- I am fully 
aware, that neither LCSH nor DDC are simple at all...) . Naïve idea...?

Thanks for any clues,
Irina


---

Irina Arndt
Max Planck Digital Library (MPDL)
Library System Coordinator
Amalienstr. 33
D-80799 Muenchen, Germany

Tel. +49 89 38602-254
Fax +49 89 38602-290

Email: ar...@mpdl.mpg.demailto:ar...@mpdl.mpg.de
http://www.mpdl.mpg.de


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet





--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] transforming marc to rdf

2013-12-06 Thread Karen Coyle

On 12/5/13 8:11 AM, Eric Lease Morgan wrote:
Where will I get the URIs from? I will get them by combining some sort 
of unique code (like an OCLC symbol) or namespace with the value of 
the MARC records' 001 fields.


You actually need 3 URIs per triple:

subject URI (which is what I believe you are creating, above)
predicate URI (the data element URI, like 
http://purl.org/dc/terms/title) http://purl.org/dc/terms/title
object URI (the URI for the data you are providing, like 
http://id.loc.gov/authorities/names/n94036700)


The first two MUST be URIs. The third SHOULD be a URI but can also be a 
string. However, strings, in the linked data space, do NOT LINK. If you 
only have strings in the object/value space then you can run searches 
against your data, but your data cannot link to other data. Creating 
linked data that doesn't link isn't terribly useful.


(In case this doesn't make sense to anyone reading, I have a slide deck 
that illustrates this. I've uploaded it to: 
http://kcoyle.net/presentations/3webIntro.pptx )


A key first step for all of us is to start getting identifiers into our 
data, even before we start thinking about linked data. MARC records in 
systems that recognize authority control should be able to store or 
provide on output the URI of every authority-controlled entity. This 
should not be terribly difficult (ok, famous last words, I know). But if 
your vendor system can flip headings then it should also be able to 
provide a URI (especially since LC has conveniently made their URIs 
derivable from the LC record numbers).


With identifiers for things, THEN you are really linking.

kc


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Interested in enhancing call-number browse in Millennium catalog

2013-12-04 Thread Karen Coyle
cynthia, I can't help you but I can encourage you -- honestly, the call 
numbers alone (especially in LCC where the notation is not a strict 
hierarchy) are close to useless without the meanings. One might even 
make the meanings more visible than the call numbers themselves (bolder, 
placed first)... Go for it!


Next: a smart phone app that reads books on a shelf and reports what the 
call numbers mean. I'd call it Where the F am I, anyway?


kc

On 12/3/13 6:35 PM, Harper, Cynthia wrote:

I'm thinking of trying to enhance the call-number browse pages on a
Millennium catalog with meanings of the classification ranges taken from
the LoC Classification database.

http://id.loc.gov/authorities/classification.html

a typical call-number browse page might look like this:
http://librarycatalog.vts.edu/search~S1?/cBX100.7.B632+1999/cbx++100.7+b632+1999/-3,-1,,E/browsehttp://librarycatalog.vts.edu/search%7ES1?/cBX100.7.B632+1999/cbx++100.7+b632+1999/-3,-1,,E/browse

I'd like to intersperse the call-number listing with call-number range
meanings like

BX100 - Christian denominations - Eastern churches

Has anyone tried this?  Can you point me to the API documentation for the
LC Classification?

Cindy Harper


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] book cover api

2013-12-04 Thread Karen Coyle

Open Library book covers come in S, M and L -

https://openlibrary.org/dev/docs/api/covers

Of course, if what you want isn't exactly one of those...

kc

On 12/4/13 9:34 AM, Kaile Zhu wrote:

A while ago, we had a discussion about book cover APIs.  I tried some of those 
mentioned and found they are working to some degree, but none of them would 
offer the size I want.  The flexibility of the size is just not there.  The 
size I am looking for is like this:
http://img1.imagesbn.com/p/9780316227940_p0_v2_s114x166.JPG

Anybody has found a way of implementing book cover api to your specifications 
successfully and is willing to share that with me?  Off-line if you want.  Much 
appreciation.  Thanks.

Kelly Zhu
405-974-5957
kz...@uco.edu

**Bronze+Blue=Green** The University of Central Oklahoma is Bronze, Blue, and 
Green! Please print this e-mail only if absolutely necessary!

**CONFIDENTIALITY** This e-mail (including any attachments) may contain 
confidential, proprietary and privileged information. Any unauthorized 
disclosure or use of this information is prohibited.


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Code4lib 2014 Diversity Scholarships: Call for Applications

2013-11-25 Thread Karen Coyle
http://grcc.edu/library | http://grcc.edu/library/socialmedia


On 11/25/2013 at 10:43 AM, Dan Eveland devel...@gmail.com wrote:

So, by diversity you mean every single type of person except white

male

that believes they are actually male. Is that accurate? So... diverse
except for one category specifically excluded through these rules. Is

there

any other category other then this one, specific, group of people who

are

not qualified to receive one of these scholarships? Really, I'd like

to

know. Perhaps it would have been more efficient to list who cannot

get the

help they need.


On Mon, Nov 18, 2013 at 9:19 PM, Jason Ronallo jrona...@gmail.com

wrote:

For the Code4Lib 2014 Conference, 9 scholarships have been

sponsored

to promote diversity.

CLIR/DLF has sponsored 5 scholarships, EBSCO has sponsored 2
scholarships, ProQuest has sponsored 1 full scholarship, and Sumana
Harihareswara has sponsored half a scholarship which was matched by
ProQuest. All sponsors have left it up to the discretion of the
Code4Lib 2014 Scholarship Committee for how to award these

diversity

scholarships.

The Code4Lib Scholarship Committee will award 9 diversity

scholarships

based on merit and need. Each scholarship will provide up to $1,000

to

cover travel costs and conference fees for a qualified attendee to
attend the 2014 Code4Lib Conference, which will be held in Raleigh,
North Carolina, from March 24 - 27, 2014.

CONFERENCE INFO

For more information on the Code4Lib Conference, please see the
conference website:
http://code4lib.org/conference/2014

You can see write-ups of previous Code4Lib Conferences:
http://journal.code4lib.org/articles/6848
http://journal.code4lib.org/articles/2717
http://journal.code4lib.org/articles/998
http://journal.code4lib.org/articles/72

CODE4LIB 2014 DIVERSITY SCHOLARSHIPS ELIGIBILITY, CRITERIA, AND
REQUIREMENTS

To qualify for a scholarship, an applicant must be interested in
actively contributing to the mission and goals of the Code4Lib
Conference.

- Four scholarships will be awarded to any woman or transgendered

person.

- Four scholarships will be awarded to any person of Hispanic or
Latino, Black or African-American, Asian, Native Hawaiian or

Pacific

Islander, or American Indian or Alaskan Native descent.
- One scholarship will be awarded to the best remaining candidate

who

meets any of the previously mentioned eligibility requirements.

Eligible applicants may apply based on multiple criteria, but no
applicant will receive more than one scholarship. Past winners of

any

Code4Lib scholarship are not eligible for a scholarship.

The scholarship recipients will be selected based upon their merit

and

financial needs.

Scholarship recipients are required to write and submit a brief

trip

report to the Code4Lib 2014 Scholarships Committee by April 1, 2014

to

be posted to the Code4Lib wiki. The report should address: (a) what
kind of experience they had at the conference, (b) what they have
learned, (c) what suggestions they have for future attendees and
conference organizers.

All reimbursement forms and receipts must be received by May 26,

2014.

HOW TO APPLY

To apply, please send an email to Jason Ronallo

(jrona...@gmail.com)

with the subject heading “Code4Lib 2014 Diversity Scholarship
Application” containing the following (combined into a single

attached

PDF, if possible):

1. A brief letter of interest, which:
 - Identifies your eligibility for a diversity scholarship
 - Describes your interest in the conference and how you intend

to

participate
 - Discusses your merit and needs for the scholarship
2. A résumé or CV
3. Contact information for two professional or academic references
The application deadline is Dec. 13, 2013, 5pm EST. The scholarship
committee will notify successful candidates the week of Jan. 6,

2013.

SPONSORS

We would like to thank our sponsors for supporting the Code4Lib

2014

Diversity Scholarships.

Council on Library and Information Resources http://www.clir.org/
Digital Library Federation http://www.diglib.org/
EBSCO http://www.ebsco.com/
ProQuest http://www.proquest.com
Sumana Harihareswara http://www.harihareswara.net/



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] Code4lib 2014 Diversity Scholarships: Call for Applications

2013-11-25 Thread Karen Coyle

On 11/25/13 12:17 PM, Matthew Sherman wrote:

I am going to reiterate my push to turn this conversation to a discussion
for funding options for everyone who wants to attend Code4Lib 2014.  I
think that will be a much better use of our time.


Agreed. Someone mentioned state scholarships that many of us didn't know 
about. I think a page on the c4l wiki for scholarship opportunities 
could be a win. People could add any that they hear about.


And I wince a bit at mentioning this, but with actual non-profit status, 
c4l might be more able to solicit donations. Or perhaps the group could 
find a non-profit partner that could help out in that regard?


kc




On Mon, Nov 25, 2013 at 3:13 PM, Erik Hetzner erik.hetz...@ucop.edu wrote:


Hi all,

I can’t believe we are having this conversation again.

I have nothing to add except to say that rather than feed the troll,
you might do what I did, and turn your frustration at this thread
arising *once again* into a donation to the Ada Initiative or similar
organization. Sadly, it seems that one cannot contribute to the
diversity scholarships, as I would be happy to do so. If anybody knows
how, please let me know.

best, Erik



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Karen Coyle
Eric, I think this skips a step - which is the design step in which you 
create a domain model that uses linked data as its basis. RDF is not a 
serialization; it actually may require you to re-think the basic 
structure of your metadata. The reason for that is that it provides 
capabilities that record-based data models do not. Rather than starting 
with current metadata, you need to take a step back and ask: what does 
my information world look like as linked data?


I repeat: RDF is NOT A SERIALIZATION.

kc

On 11/19/13 5:04 AM, Eric Lease Morgan wrote:

I believe participating in the Semantic Web and providing content via the principles of 
linked data is not rocket surgery, especially for cultural heritage 
institutions -- libraries, archives, and museums. Here is a simple recipe for their 
participation:

   1. use existing metadata standards (MARC, EAD, etc.) to describe
  collections

   2. use any number of existing tools to convert the metadata to
  HTML, and save the HTML on a Web server

   3. use any number of existing tools to convert the metadata to
  RDF/XML (or some other serialization of RDF), and save the
  RDF/XML on a Web server

   4. rest, congratulate yourself, and share your experience with
  others in your domain

   5. after the first time though, go back to Step #1, but this time
  work with other people inside your domain making sure you use as
  many of the same URIs as possible

   6. after the second time through, go back to Step #1, but this
  time supplement access to your linked data with a triple store,
  thus supporting search

   7. after the third time through, go back to Step #1, but this
  time use any number of existing tools to expose the content in
  your other information systems (relational databases, OAI-PMH
  data repositories, etc.)

   8. for dessert, cogitate ways to exploit the linked data in your
  domain to discover new and additional relationships between URIs,
  and thus make the Semantic Web more of a reality

What do you think?

I am in the process of writing a guidebook on the topic of linked data and 
archives. In the guidebook I will elaborate on this recipe and provide 
instructions for its implementation. [1]

[1] guidebook - http://sites.tufts.edu/liam/

--
Eric Lease Morgan
University of Notre Dame


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Karen Coyle
Eric, if you want to leap into the linked data world in the fastest, 
easiest way possible, then I suggest looking at microdata markup, e.g. 
schema.org.[1] Schema.org does not require you to transform your data at 
all: it only requires mark-up of your online displays. This makes sense 
because as long as your data is in local databases, it's not visible to 
the linked data universe anyway; so why not take the easy way out and 
just add linked data to your public online displays? This doesn't 
require a transformation of your entire record (some of which may not be 
suitable as linked data in any case), only those things that are 
likely to link usefully. This latter generally means things for which 
you have an identifier. And you make no changes to your database, only 
to display.


OCLC is already producing this markup in WorldCat records [2]-- not 
perfectly, of course, lots of warts, but it is a first step. However, it 
is a first step that makes more sense to me than *transforming* or 
*cross-walking* current metadata. It also, I believe, will help us 
understand what bits of our current metadata will make the transition to 
linked data, and what bits should remain as accessible documents that 
users can reach through linked data.


kc
[1] http://schema.org, and look at the work going on to add 
bibliographic properties at 
http://www.w3.org/community/schemabibex/wiki/Main_Page
[2] look at the linked data section of any WorldCat page for a single 
item, such 
ashttp://www.worldcat.org/title/selection-of-early-statistical-papers-of-j-neyman/oclc/527725referer=brief_results




On 11/19/13 7:54 AM, Eric Lease Morgan wrote:

On Nov 19, 2013, at 9:41 AM, Karen Coyle li...@kcoyle.net wrote:


Eric, I think this skips a step - which is the design step in which you
create a domain model that uses linked data as its basis. RDF is not a
serialization; it actually may require you to re-think the basic
structure of your metadata. The reason for that is that it provides
capabilities that record-based data models do not. Rather than starting
with current metadata, you need to take a step back and ask: what does
my information world look like as linked data?


I respectfully disagree. I do not think it necessary to create a domain model 
ahead of time; I do not think it is necessary for us to re-think our metadata 
structures. There already exists tools enabling us — cultural heritage 
institutions — to manifest our metadata as RDF. The manifestations may not be 
perfect, but “we need to learn to walk before we run” and the metadata 
structures we have right now will work for right now. As we mature we can 
refine our processes. I do not advocate “stepping back and asking”. I advocate 
looking forward and doing. —Eric Morgan


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] mass convert jpeg to pdf

2013-11-08 Thread Karen Coyle
+1 for the viewer concept, and I'll add that viewing  downloading meet 
different needs and should both be offered if possible. (said because of 
recently having had to download huge PDFs just to glance at a few pages).


kc

On 11/8/13 11:10 AM, Edward Summers wrote:

It is sad to me that converting to PDF for viewing off the Web seems like the 
answer. Isn’t there a tiling viewer (like Leaflet) that could be used to render 
jpeg derivatives of the original tif files in Omeka?

For an example of using Leaflet (usually used for working with maps) in this 
way checkout NYTimes Machine Beta:

 http://apps.beta620.nytimes.com/timesmachine/1969/07/20/issue.html

//Ed

On Nov 8, 2013, at 2:00 PM, Kyle Banerjee kyle.baner...@gmail.com wrote:


We are in the process of migrating our digital collections from CONTENTdm
to Omeka and are trying to figure out what to do about the compound objects
-- the vast majority of which are digitized books.

The source files are actually hi res tiffs but since ginormous objects
broken into hundreds of pieces (each of which can be well over 100MB in
size) aren't exactly friendly to use, we'd like to stitch them into
individual pdf's that can be viewed more conveniently

My game plan is to simply have a script pull the files down as jpegs which
can be fed to imagemagick which can theoretically do everything I need.
However, I've never actually done anything like this before, so I wanted to
see if there's a method that people have used for combining lots of images
into pdfs that works particularly well. Thanks,

kyle


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] display book covers

2013-11-07 Thread Karen Coyle
I wanted to mention that I pointed to the Open Library covers API 
earlier, and it so happens that was on a day when OL covers was down 
(possibly due to the disruption of the fire at the Internet Archive). So 
if you've been having trouble with OL covers, it has been reported that 
they are now available again.


https://openlibrary.org/dev/docs/api/covers

kc


On 11/7/13 6:57 AM, Brent Ferguson wrote:

LibraryThing's API may have book cover art , can't remember...

We pay for Syndetics (cover art service) (owned by Bowker) and can use that for 
our cover art outside PAC


From: Code for Libraries [CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Chris 
Fitzpatrick [chrisfitz...@gmail.com]
Sent: Thursday, November 07, 2013 9:56 AM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] display book covers

Hi,


I think you can do this all with JS or Coffeescript.

Here's a fiddle :

http://jsfiddle.net/chrisfitzpat/t69Xs/




On Tue, Nov 5, 2013 at 10:12 PM, Daryl Grenz grenzda...@hotmail.com wrote:


Powell's Books provides an API (http://api.powells.com/stable) and direct
links to their book covers by ISBN13 only.
Regarding the limit on daily use of the Google Books API, I think from
when I used it before that if you access cover links through the Dynamic
Links API (https://developers.google.com/books/docs/dynamic-links) there
is no daily limit.
- Daryl


Date: Tue, 5 Nov 2013 15:13:35 +
From: aw...@rockhall.org
Subject: [CODE4LIB] display book covers
To: CODE4LIB@LISTSERV.ND.EDU

Hi all,

Anyone have some good resources about tools for gathering book cover

images?  I'm building that into our next catalog update, which uses
Blacklight, but I'm not necessarily looking for Rails-only approaches.  My
questions are more general:

What sources are out there?  (ex. Google Books, amazon)

Making it work?
I'm trying out Google Books at the moment, just making a call to their

API.  This can be asynchronously and loaded after the rest of the page, or
cached, perhaps even store the url in solr or a database table?

Tools?
I am trying out a Google Books gem[1], which is just a wrapper for the

api.

Other thoughts?

Thanks in advance,

…adam

__
Adam Wead
Systems and Digital Collections Librarian
Library + Archives
Rock and Roll Hall of Fame and Museum
216.515.1960
aw...@rockhall.org

[1] https://github.com/zeantsoi/GoogleBooks
This communication is a confidential and proprietary business

communication. It is intended solely for the use of the designated
recipient(s). If this communication is received in error, please contact
the sender and delete this communication.




--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] rdf serialization

2013-11-07 Thread Karen Coyle

Ethan, thanks, it's good to have examples.

I'd say that for simple linking SPARQL may not be necessary, perhaps 
should be avoided, but IF you need something ELSE, say a query WHERE you 
have conditions, THEN you may find that a query language is needed.


kc

On 11/6/13 9:14 AM, Ethan Gruber wrote:

I think that the answer to #1 is that if you want or expect people to use
your endpoint that you should document how it works: the ontologies, the
models, and a variety of example SPARQL queries, ranging from simple to
complex.  The British Museum's SPARQL endpoint (
http://collection.britishmuseum.org/sparql) is highly touted, but how many
people actually use it?  I understand your point about SPARQL being too
complicated for an API interface, but the best examples of services built
on SPARQL are probably the ones you don't even realize are built on SPARQL
(e.g., http://numismatics.org/ocre/id/ric.1%282%29.aug.4A#mapTab).  So on
one hand, perhaps only the most dedicated and hardcore researchers will
venture to construct SPARQL queries for your endpoint, but on the other,
you can build some pretty visualizations based on SPARQL queries conducted
in the background from the user's interaction with a simple html/javascript
based interface.

Ethan


On Wed, Nov 6, 2013 at 11:54 AM, Ross Singer rossfsin...@gmail.com wrote:


Hey Karen,

It's purely anecdotal (albeit anecdotes borne from working at a company
that offered, and has since abandoned, a sparql-based triple store
service), but I just don't see the interest in arbitrary SPARQL queries
against remote datasets that I do against linking to (and grabbing) known
items.  I think there are multiple reasons for this:

1) Unless you're already familiar with the dataset behind the SPARQL
endpoint, where do you even start with constructing useful queries?
2) SPARQL as a query language is a combination of being too powerful and
completely useless in practice: query timeouts are commonplace, endpoints
don't support all of 1.1, etc.  And, going back to point #1, it's hard to
know how to optimize your queries unless you are already pretty familiar
with the data
3) SPARQL is a flawed API interface from the get-go (IMHO) for the same
reason we don't offer a public SQL interface to our RDBMSes

Which isn't to say it doesn't have its uses or applications.

I just think that in most cases domain/service-specific APIs (be they
RESTful, based on the Linked Data API [0], whatever) will likely be favored
over generic SPARQL endpoints.  Are n+1 different APIs ideal?  I am pretty
sure the answer is no, but that's the future I foresee, personally.

-Ross.
0. https://code.google.com/p/linked-data-api/wiki/Specification


On Wed, Nov 6, 2013 at 11:28 AM, Karen Coyle li...@kcoyle.net wrote:


Ross, I agree with your statement that data doesn't have to be RDF all
the way down, etc. But I'd like to hear more about why you think SPARQL
availability has less value, and if you see an alternative to SPARQL for
querying.

kc



On 11/6/13 8:11 AM, Ross Singer wrote:


Hugh, I don't think you're in the weeds with your question (and, while I
think that named graphs can provide a solution to your particular

problem,

that doesn't necessarily mean that it doesn't raise more questions or
potentially more frustrations down the line - like any new power, it can
be
used for good or evil and the difference might not be obvious at first).

My question for you, however, is why are you using a triple store for
this?
   That is, why bother with the broad and general model in what I assume
is a
closed world assumption in your application?

We don't generally use XML databases (Marklogic being a notable
exception),
or MARC databases, or insert your transmission format of

choice-specific

databases because usually transmission formats are designed to account

for

lots and lots of variations and maximum flexibility, which generally is
the
opposite of the modeling that goes into a specific app.

I think there's a world of difference between modeling your data so it

can

be represented in RDF (and, possibly, available via SPARQL, but I think
there is *far* less value there) and committing to RDF all the way down.
   RDF is a generalization so multiple parties can agree on what data
means,
but I would have a hard time swallowing the argument that

domain-specific

data must be RDF-native.

-Ross.


On Wed, Nov 6, 2013 at 10:52 AM, Hugh Cayless philomou...@gmail.com
wrote:

  Does that work right down to the level of the individual triple though?

If
a large percentage of my triples are each in their own individual

graphs,

won't that be chaos? I really don't know the answer, it's not a
rhetorical
question!

Hugh

On Nov 6, 2013, at 10:40 , Robert Sanderson azarot...@gmail.com

wrote:

  Named Graphs are the way to solve the issue you bring up in that post,

in
my opinion.  You mint an identifier for the graph, and associate the
provenance and other information with that.  This then gets ingested

Re: [CODE4LIB] rdf serialization

2013-11-07 Thread Karen Coyle

Ross, I think you are not alone, as per this:

http://howfuckedismydatabase.com/nosql/

kc

On 11/6/13 8:54 AM, Ross Singer wrote:

Hey Karen,

It's purely anecdotal (albeit anecdotes borne from working at a company
that offered, and has since abandoned, a sparql-based triple store
service), but I just don't see the interest in arbitrary SPARQL queries
against remote datasets that I do against linking to (and grabbing) known
items.  I think there are multiple reasons for this:

1) Unless you're already familiar with the dataset behind the SPARQL
endpoint, where do you even start with constructing useful queries?
2) SPARQL as a query language is a combination of being too powerful and
completely useless in practice: query timeouts are commonplace, endpoints
don't support all of 1.1, etc.  And, going back to point #1, it's hard to
know how to optimize your queries unless you are already pretty familiar
with the data
3) SPARQL is a flawed API interface from the get-go (IMHO) for the same
reason we don't offer a public SQL interface to our RDBMSes

Which isn't to say it doesn't have its uses or applications.

I just think that in most cases domain/service-specific APIs (be they
RESTful, based on the Linked Data API [0], whatever) will likely be favored
over generic SPARQL endpoints.  Are n+1 different APIs ideal?  I am pretty
sure the answer is no, but that's the future I foresee, personally.

-Ross.
0. https://code.google.com/p/linked-data-api/wiki/Specification


On Wed, Nov 6, 2013 at 11:28 AM, Karen Coyle li...@kcoyle.net wrote:


Ross, I agree with your statement that data doesn't have to be RDF all
the way down, etc. But I'd like to hear more about why you think SPARQL
availability has less value, and if you see an alternative to SPARQL for
querying.

kc



On 11/6/13 8:11 AM, Ross Singer wrote:


Hugh, I don't think you're in the weeds with your question (and, while I
think that named graphs can provide a solution to your particular problem,
that doesn't necessarily mean that it doesn't raise more questions or
potentially more frustrations down the line - like any new power, it can
be
used for good or evil and the difference might not be obvious at first).

My question for you, however, is why are you using a triple store for
this?
   That is, why bother with the broad and general model in what I assume
is a
closed world assumption in your application?

We don't generally use XML databases (Marklogic being a notable
exception),
or MARC databases, or insert your transmission format of choice-specific
databases because usually transmission formats are designed to account for
lots and lots of variations and maximum flexibility, which generally is
the
opposite of the modeling that goes into a specific app.

I think there's a world of difference between modeling your data so it can
be represented in RDF (and, possibly, available via SPARQL, but I think
there is *far* less value there) and committing to RDF all the way down.
   RDF is a generalization so multiple parties can agree on what data
means,
but I would have a hard time swallowing the argument that domain-specific
data must be RDF-native.

-Ross.


On Wed, Nov 6, 2013 at 10:52 AM, Hugh Cayless philomou...@gmail.com
wrote:

  Does that work right down to the level of the individual triple though?

If
a large percentage of my triples are each in their own individual graphs,
won't that be chaos? I really don't know the answer, it's not a
rhetorical
question!

Hugh

On Nov 6, 2013, at 10:40 , Robert Sanderson azarot...@gmail.com wrote:

  Named Graphs are the way to solve the issue you bring up in that post,

in
my opinion.  You mint an identifier for the graph, and associate the
provenance and other information with that.  This then gets ingested as


the


4th URI into a quad store, so you don't lose the provenance information.

In JSON-LD:
{
   @id : uri-for-graph,
   dcterms:creator : uri-for-hugh,
   @graph : [
// ... triples go here ...
   ]
}

Rob



On Wed, Nov 6, 2013 at 7:42 AM, Hugh Cayless philomou...@gmail.com


wrote:


I wrote about this a few months back at

  http://blogs.library.duke.edu/dcthree/2013/07/27/the-

trouble-with-triples/


I'd be very interested to hear what the smart folks here think!

Hugh

On Nov 5, 2013, at 18:28 , Alexander Johannesen 
alexander.johanne...@gmail.com wrote:

  But the

question to every piece of meta data is *authority*, which is the part
of RDF that sucks.


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] rdf serialization

2013-11-06 Thread Karen Coyle
Ben, Yes, I copied from the browser URIs, and that was sloppy. However, 
it was the quickest thing to do, plus it was addressed to a human, not a 
machine. The URI for the LC entry is there on the page. Unfortunately, 
the VIAF URI is called Permalink -- which isn't obvious.


I guess if I want anyone to answer my emails, I need to post mistakes. 
When I post correct information, my mail goes unanswered (not even a 
thanks). So, thanks, guys.


kc

On 11/6/13 12:47 AM, Ben Companjen wrote:

Karen,

The URIs you gave get me to webpages *about* the Declaration of
Independence. I'm sure it's just a copy/paste mistake, but in this context
you want the exact right URIs of course. And by better I guess you meant
probably more widely used and probably longer lasting? :)

LOC URI for the DoI (the work) is without .html:
http://id.loc.gov/authorities/names/n79029194


VIAF URI for the DoI is without trailing /:
http://viaf.org/viaf/179420344

Ben
http://companjen.name/id/BC - me
http://companjen.name/id/BC.html - about me


On 05-11-13 19:03, Karen Coyle li...@kcoyle.net wrote:


Eric, I found an even better URI for you for the Declaration of
Independence:

http://id.loc.gov/authorities/names/n79029194.html

Now that could be seen as being representative of the name chosen by the
LC Name Authority, but the related VIAF record, as per the VIAF
definition of itself, represents the real world thing itself. That URI is:

http://viaf.org/viaf/179420344/

I noticed that this VIAF URI isn't linked from the Wikipedia page, so I
will add that.

kc


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] rdf serialization

2013-11-06 Thread Karen Coyle
Ross, I agree with your statement that data doesn't have to be RDF all 
the way down, etc. But I'd like to hear more about why you think SPARQL 
availability has less value, and if you see an alternative to SPARQL for 
querying.


kc


On 11/6/13 8:11 AM, Ross Singer wrote:

Hugh, I don't think you're in the weeds with your question (and, while I
think that named graphs can provide a solution to your particular problem,
that doesn't necessarily mean that it doesn't raise more questions or
potentially more frustrations down the line - like any new power, it can be
used for good or evil and the difference might not be obvious at first).

My question for you, however, is why are you using a triple store for this?
  That is, why bother with the broad and general model in what I assume is a
closed world assumption in your application?

We don't generally use XML databases (Marklogic being a notable exception),
or MARC databases, or insert your transmission format of choice-specific
databases because usually transmission formats are designed to account for
lots and lots of variations and maximum flexibility, which generally is the
opposite of the modeling that goes into a specific app.

I think there's a world of difference between modeling your data so it can
be represented in RDF (and, possibly, available via SPARQL, but I think
there is *far* less value there) and committing to RDF all the way down.
  RDF is a generalization so multiple parties can agree on what data means,
but I would have a hard time swallowing the argument that domain-specific
data must be RDF-native.

-Ross.


On Wed, Nov 6, 2013 at 10:52 AM, Hugh Cayless philomou...@gmail.com wrote:


Does that work right down to the level of the individual triple though? If
a large percentage of my triples are each in their own individual graphs,
won't that be chaos? I really don't know the answer, it's not a rhetorical
question!

Hugh

On Nov 6, 2013, at 10:40 , Robert Sanderson azarot...@gmail.com wrote:


Named Graphs are the way to solve the issue you bring up in that post, in
my opinion.  You mint an identifier for the graph, and associate the
provenance and other information with that.  This then gets ingested as

the

4th URI into a quad store, so you don't lose the provenance information.

In JSON-LD:
{
  @id : uri-for-graph,
  dcterms:creator : uri-for-hugh,
  @graph : [
   // ... triples go here ...
  ]
}

Rob



On Wed, Nov 6, 2013 at 7:42 AM, Hugh Cayless philomou...@gmail.com

wrote:

I wrote about this a few months back at


http://blogs.library.duke.edu/dcthree/2013/07/27/the-trouble-with-triples/

I'd be very interested to hear what the smart folks here think!

Hugh

On Nov 5, 2013, at 18:28 , Alexander Johannesen 
alexander.johanne...@gmail.com wrote:


But the
question to every piece of meta data is *authority*, which is the part
of RDF that sucks.


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] rdf serialization

2013-11-05 Thread Karen Coyle

On 11/5/13 6:45 AM, Ed Summers wrote:

I'm with Ross though:

... and Karen!


I find it much to read rdf as turtle or json-ld than it is rdf/xml.


It's easier to read, but it's also easier to create *correctly*, and 
that, to me, is the key point. Folks who are used to XML have a certain 
notion of data organization in mind. Working with RDF in XML one tends 
to fall into the XML data think rather than the RDF concepts.


I have suggested (repeatedly) to LC on the BIBFRAME list that they 
should use turtle rather than RDF/XML in their examples -- because I 
suspect that they may be doing some XML think in the background. This 
seems to be the case because in some of the BIBFRAME documents the 
examples are in XML but not RDF/XML. I find this rather ... disappointing.


I also find it useful to create pseudo-code triples using whatever 
notation I find handy, as in the example I provided earlier for Eric. 
Writing out actual valid triples is a pain, but seeing your data as 
triples is very useful.


kc

--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] display book covers

2013-11-05 Thread Karen Coyle

Open Library's CoverStore API information is here:
  https://openlibrary.org/dev/docs/api/covers

kc


On 11/5/13 7:13 AM, Adam Wead wrote:

Hi all,

Anyone have some good resources about tools for gathering book cover images?  
I'm building that into our next catalog update, which uses Blacklight, but I'm 
not necessarily looking for Rails-only approaches.  My questions are more 
general:

What sources are out there?  (ex. Google Books, amazon)

Making it work?
I'm trying out Google Books at the moment, just making a call to their API.  
This can be asynchronously and loaded after the rest of the page, or cached, 
perhaps even store the url in solr or a database table?

Tools?
I am trying out a Google Books gem[1], which is just a wrapper for the api.

Other thoughts?

Thanks in advance,

…adam

__
Adam Wead
Systems and Digital Collections Librarian
Library + Archives
Rock and Roll Hall of Fame and Museum
216.515.1960
aw...@rockhall.org

[1] https://github.com/zeantsoi/GoogleBooks
This communication is a confidential and proprietary business communication. It 
is intended solely for the use of the designated recipient(s). If this 
communication is received in error, please contact the sender and delete this 
communication.


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


  1   2   3   4   5   >