Re: [CODE4LIB] getting Worldcat records

2008-02-07 Thread Reese, Terry
Since these are your libraries' records, you can certainly download them
again from OCLC.  I've also known libraries in the past that have been
able to have oclc generate a subset of records from their database --
though in these cases, this always has involved a cost to purchase the
records.  In terms of how easy it is to do on your own -- if you don't
have OCLC do it, you would likely need a list of all the OCLC numbers
that you are interested in.  With that list, you could easily batch
export the data again from Worldcat using Connexion.

--TR



 -Original Message-
 From: Code for Libraries [mailto:[EMAIL PROTECTED] On Behalf
Of
 Alberto Accomazzi
 Sent: Thursday, February 07, 2008 10:35 AM
 To: CODE4LIB@listserv.nd.edu
 Subject: [CODE4LIB] getting Worldcat records

 Our project maintains a database of bibliographic metadata for all
 things in astronomy and most of physics.  We'd like to add records for
 books that have been recently added to our library and to correlate
 existing records with the library holdings.  Sounds easy enough, but
 because of the intricacies of Harvard libraries administration we
 haven't been able to get a dump of the records, much less a feed.

 The recent emails about OCLC worldcat records made me wonder if we
 could
 get the equivalent data from them (since our library subscribes to
 them).  Essentially what I'd like is a dump of all QB and QC records
in
 OCLC entered by Harvard, so we can index them and then point to the
 library record in OCLC.  Is this (a) legal, (b) feasible, (c) easy?  I
 assume the answer to (a) and (b) is yes, since we have our library's
 support.  If not, are there alternatives?  I learned about openlibrary
 only yesterday, so I haven't had a chance to explore what's in it
 yet...

 Thanks,

 -- Alberto


[CODE4LIB] Date and time for Evergreen Gathering at PLA

2008-02-07 Thread Jason Etheridge
Folks,

I apologize for the cross-posting (and re-iteration of some
information), but I want to spread this far and wide.

We're organizing a (currently informal) gathering of Evergreen(*)
users, enthusiasts, and interested individuals, during the Public
Library Association National Conference (March 25-29, 2008 in
Minneapolis, Minnesota).

This is what we have so far for a meeting date and location:  March
27, 2008 5:30pm-7:30pm in the Minneapolis Marriott City Center (exact
room to be determined)

Anyone is free to attend, but if you would, let me know if you plan to
do so we can be better prepared.  Also, if you would like to help
organize (or are simply interested), we have a Google Document for
collaborators that I've been adding folks to.  Just let me know if you
would like to be added to that as well.

Further information on this gathering will be disseminated via the
OPEN-ILS-GENERAL mailing list (http://open-ils.org/listserv.php) and
the project's blog (http://open-ils.org/blog/), so you all won't have
to get spammed by me again. :)

Thanks!

(*) Evergreen is an open-source library automation system originally
developed by the Georgia Public Library Service for its PINES program,
with the hope that others throughout the world would benefit from it
and help contribute to its development (which is indeed happening).

--
Jason Etheridge
 | VP, Community Support and Advocacy
 | Equinox Software, Inc. / The Evergreen Experts
 | phone:  1-877-OPEN-ILS (673-6457)
 | email:  [EMAIL PROTECTED]
 | web:  http://www.esilibrary.com


[CODE4LIB] Looking for Bib and LC subject authority data

2008-02-07 Thread Charles Antoine Julien, Mr
A kind fellow on NGC4Lib suggested I mention this here. 

 

I'm developing a 3D fly-through interface for an LCSH organized
collection but I'm having difficulty finding a library willing to give
me a subset of their data (i.e., subject headings (broad to narrow
terms) and the bib records to which they have been assigned).  They just
don't see why they should help me.  Their value added isn't clear to
them since this is experimental and I have no wish turn this into a
business (I like to build and test solutions...selling them isn't my
piece of pie).

 

I'm planning to import the data into Access or SQL Server (depending how
much I get) and partly normalize the bib records so subject terms for
each item are in a separate one-to-many table.  I also need the
authority data to establish where each subject term (and its associated
bib records) resides in the broad to narrow term hierarchy...this is
more useful in the sciences which seems to have 4-6 levels deep.  

 

Jonathan Rochkind (kind fellow in question) suggested the following

 

-I could access data directly through Z39.5...

-I could take LC subject authority data in MARC format from a certain
grey-area-legal source

-I could take bib records (and their associated LCSH terms) from 

http://simile.mit.edu/wiki/Dataset_Collection Particularly:
http://simile.mit.edu/rdf-test-data/barton/compressed/

In particular, the Barton collection. That will be in the MODS format,
which will actually be easier to work with than library standard MARC. 

Or http://www.archive.org/details/marc_records_scriblio_net

 

Obviously I'm not looking forward to parsing MARC data although I've
heard there are scripts for this.  

 

Additional suggestions and/or comments would be greatly appreciated.

 

Thanks a bunch,

 

Charles-Antoine Julien

Ph.D Candidate

School of Information Studies

McGill University

 


Re: [CODE4LIB] Records for Open Library

2008-02-07 Thread K.G. Schneider
I had another thought (ouch... hurts...) which is this: if OCLC had to
open up its data, then it would have to improve its services to survive.

K.G. Schneider

On Thu, 7 Feb 2008 10:28:14 -0600, Danielle Plumer
[EMAIL PROTECTED] said:
 Kevin Kelly had an interesting post on The Technium last week about these
 sorts of issues
 (http://www.kk.org/thetechnium/archives/2008/01/better_than_fre.php), and
 his conclusion is exactly along the lines of Karen's post.

 His assumptions are:

 When copies are super abundant, they become worthless.
 When copies are super abundant, stuff which can't be copied becomes
 scarce and valuable.

 So he concludes:

 When copies are free, you need to sell things which can not be copied.

 The things which cannot be copied are services -- he lists eight
 generatives that have value. These are immediacy, personalization,
 interpretation, authenticity, accessibility, embodiment, patronage, and
 findability. Trust is also mentioned as a intangible asset with
 significant value.

 I find that this is a compelling argument, and it seems to be in line
 with things I hear coming out of OCLC Research, at least, and from the
 folks at Open Library, too. It will take time for an organization with as
 much inertia as OCLC has to change its modus operandi, but I think it
 will come. However, unlike others, I tend to be an optimist in the
 morning and a cynic by nightfall, so we'll see...

 Danielle Cunniff Plumer, Coordinator
 Texas Heritage Digitization Initiative
 Texas State Library and Archives Commission
 512.463.5852 (phone) / 512.936.2306 (fax)
 [EMAIL PROTECTED]

 -Original Message-
 From: Code for Libraries [mailto:[EMAIL PROTECTED] Behalf Of
 K.G. Schneider
 Sent: Thursday, February 07, 2008 7:04 AM
 To: CODE4LIB@listserv.nd.edu
 Subject: Re: [CODE4LIB] Records for Open Library


  Maybe Roy will answer that one -- but I doubt its that difficult to guess.
  OCLC's primary value is its bibliographic database and the information
  about its member's holdings.  Nearly all of it's services are built around
  this.  If they gave that information up to the Open Library, it would most
  certainly undermine their ILL, Cataloging and Grid Services initiatives.
  However, if a handful of members in relation to their membership
  participate in the program -- its no skin off their noses.
 
  --TR
 

 You know, I realize that's the going-in thinking, and OCLC has shared
 that
 with me. I fully understand the need for OCLC to protect its services.
 But I
 remember with a previous job that people (even some very important
 people)
 thought our product was our data, but it really wasn't: it was the
 services
 we wrapped around the data, including maintenance, delivery, affiliated
 products, etc. It's true that the data had to be good, but that goodness
 didn't come with a core dump of one-time static data. Keeping our data
 closed ultimately harmed us, perhaps perniciously, and I wish I had done
 a
 better job of championing a different path. I didn't have the skills or
 vocabulary and to this day I regret that.

 Karen G. Been there, done that, got the teeshirt Schneider
 [EMAIL PROTECTED]


[CODE4LIB] getting Worldcat records

2008-02-07 Thread Alberto Accomazzi

Our project maintains a database of bibliographic metadata for all
things in astronomy and most of physics.  We'd like to add records for
books that have been recently added to our library and to correlate
existing records with the library holdings.  Sounds easy enough, but
because of the intricacies of Harvard libraries administration we
haven't been able to get a dump of the records, much less a feed.

The recent emails about OCLC worldcat records made me wonder if we could
get the equivalent data from them (since our library subscribes to
them).  Essentially what I'd like is a dump of all QB and QC records in
OCLC entered by Harvard, so we can index them and then point to the
library record in OCLC.  Is this (a) legal, (b) feasible, (c) easy?  I
assume the answer to (a) and (b) is yes, since we have our library's
support.  If not, are there alternatives?  I learned about openlibrary
only yesterday, so I haven't had a chance to explore what's in it yet...

Thanks,

-- Alberto


Re: [CODE4LIB] Records for Open Library

2008-02-07 Thread Danielle Plumer
Kevin Kelly had an interesting post on The Technium last week about these sorts 
of issues (http://www.kk.org/thetechnium/archives/2008/01/better_than_fre.php), 
and his conclusion is exactly along the lines of Karen's post.

His assumptions are:

When copies are super abundant, they become worthless.
When copies are super abundant, stuff which can't be copied becomes scarce and 
valuable.

So he concludes:

When copies are free, you need to sell things which can not be copied.

The things which cannot be copied are services -- he lists eight generatives 
that have value. These are immediacy, personalization, interpretation, 
authenticity, accessibility, embodiment, patronage, and findability. Trust is 
also mentioned as a intangible asset with significant value.

I find that this is a compelling argument, and it seems to be in line with 
things I hear coming out of OCLC Research, at least, and from the folks at Open 
Library, too. It will take time for an organization with as much inertia as 
OCLC has to change its modus operandi, but I think it will come. However, 
unlike others, I tend to be an optimist in the morning and a cynic by 
nightfall, so we'll see...

Danielle Cunniff Plumer, Coordinator
Texas Heritage Digitization Initiative
Texas State Library and Archives Commission
512.463.5852 (phone) / 512.936.2306 (fax)
[EMAIL PROTECTED]

-Original Message-
From: Code for Libraries [mailto:[EMAIL PROTECTED] Behalf Of
K.G. Schneider
Sent: Thursday, February 07, 2008 7:04 AM
To: CODE4LIB@listserv.nd.edu
Subject: Re: [CODE4LIB] Records for Open Library


 Maybe Roy will answer that one -- but I doubt its that difficult to guess.
 OCLC's primary value is its bibliographic database and the information
 about its member's holdings.  Nearly all of it's services are built around
 this.  If they gave that information up to the Open Library, it would most
 certainly undermine their ILL, Cataloging and Grid Services initiatives.
 However, if a handful of members in relation to their membership
 participate in the program -- its no skin off their noses.

 --TR


You know, I realize that's the going-in thinking, and OCLC has shared that
with me. I fully understand the need for OCLC to protect its services. But I
remember with a previous job that people (even some very important people)
thought our product was our data, but it really wasn't: it was the services
we wrapped around the data, including maintenance, delivery, affiliated
products, etc. It's true that the data had to be good, but that goodness
didn't come with a core dump of one-time static data. Keeping our data
closed ultimately harmed us, perhaps perniciously, and I wish I had done a
better job of championing a different path. I didn't have the skills or
vocabulary and to this day I regret that.

Karen G. Been there, done that, got the teeshirt Schneider
[EMAIL PROTECTED]