Re: [CODE4LIB] getting Worldcat records
Since these are your libraries' records, you can certainly download them again from OCLC. I've also known libraries in the past that have been able to have oclc generate a subset of records from their database -- though in these cases, this always has involved a cost to purchase the records. In terms of how easy it is to do on your own -- if you don't have OCLC do it, you would likely need a list of all the OCLC numbers that you are interested in. With that list, you could easily batch export the data again from Worldcat using Connexion. --TR -Original Message- From: Code for Libraries [mailto:[EMAIL PROTECTED] On Behalf Of Alberto Accomazzi Sent: Thursday, February 07, 2008 10:35 AM To: CODE4LIB@listserv.nd.edu Subject: [CODE4LIB] getting Worldcat records Our project maintains a database of bibliographic metadata for all things in astronomy and most of physics. We'd like to add records for books that have been recently added to our library and to correlate existing records with the library holdings. Sounds easy enough, but because of the intricacies of Harvard libraries administration we haven't been able to get a dump of the records, much less a feed. The recent emails about OCLC worldcat records made me wonder if we could get the equivalent data from them (since our library subscribes to them). Essentially what I'd like is a dump of all QB and QC records in OCLC entered by Harvard, so we can index them and then point to the library record in OCLC. Is this (a) legal, (b) feasible, (c) easy? I assume the answer to (a) and (b) is yes, since we have our library's support. If not, are there alternatives? I learned about openlibrary only yesterday, so I haven't had a chance to explore what's in it yet... Thanks, -- Alberto
[CODE4LIB] Date and time for Evergreen Gathering at PLA
Folks, I apologize for the cross-posting (and re-iteration of some information), but I want to spread this far and wide. We're organizing a (currently informal) gathering of Evergreen(*) users, enthusiasts, and interested individuals, during the Public Library Association National Conference (March 25-29, 2008 in Minneapolis, Minnesota). This is what we have so far for a meeting date and location: March 27, 2008 5:30pm-7:30pm in the Minneapolis Marriott City Center (exact room to be determined) Anyone is free to attend, but if you would, let me know if you plan to do so we can be better prepared. Also, if you would like to help organize (or are simply interested), we have a Google Document for collaborators that I've been adding folks to. Just let me know if you would like to be added to that as well. Further information on this gathering will be disseminated via the OPEN-ILS-GENERAL mailing list (http://open-ils.org/listserv.php) and the project's blog (http://open-ils.org/blog/), so you all won't have to get spammed by me again. :) Thanks! (*) Evergreen is an open-source library automation system originally developed by the Georgia Public Library Service for its PINES program, with the hope that others throughout the world would benefit from it and help contribute to its development (which is indeed happening). -- Jason Etheridge | VP, Community Support and Advocacy | Equinox Software, Inc. / The Evergreen Experts | phone: 1-877-OPEN-ILS (673-6457) | email: [EMAIL PROTECTED] | web: http://www.esilibrary.com
[CODE4LIB] Looking for Bib and LC subject authority data
A kind fellow on NGC4Lib suggested I mention this here. I'm developing a 3D fly-through interface for an LCSH organized collection but I'm having difficulty finding a library willing to give me a subset of their data (i.e., subject headings (broad to narrow terms) and the bib records to which they have been assigned). They just don't see why they should help me. Their value added isn't clear to them since this is experimental and I have no wish turn this into a business (I like to build and test solutions...selling them isn't my piece of pie). I'm planning to import the data into Access or SQL Server (depending how much I get) and partly normalize the bib records so subject terms for each item are in a separate one-to-many table. I also need the authority data to establish where each subject term (and its associated bib records) resides in the broad to narrow term hierarchy...this is more useful in the sciences which seems to have 4-6 levels deep. Jonathan Rochkind (kind fellow in question) suggested the following -I could access data directly through Z39.5... -I could take LC subject authority data in MARC format from a certain grey-area-legal source -I could take bib records (and their associated LCSH terms) from http://simile.mit.edu/wiki/Dataset_Collection Particularly: http://simile.mit.edu/rdf-test-data/barton/compressed/ In particular, the Barton collection. That will be in the MODS format, which will actually be easier to work with than library standard MARC. Or http://www.archive.org/details/marc_records_scriblio_net Obviously I'm not looking forward to parsing MARC data although I've heard there are scripts for this. Additional suggestions and/or comments would be greatly appreciated. Thanks a bunch, Charles-Antoine Julien Ph.D Candidate School of Information Studies McGill University
Re: [CODE4LIB] Records for Open Library
I had another thought (ouch... hurts...) which is this: if OCLC had to open up its data, then it would have to improve its services to survive. K.G. Schneider On Thu, 7 Feb 2008 10:28:14 -0600, Danielle Plumer [EMAIL PROTECTED] said: Kevin Kelly had an interesting post on The Technium last week about these sorts of issues (http://www.kk.org/thetechnium/archives/2008/01/better_than_fre.php), and his conclusion is exactly along the lines of Karen's post. His assumptions are: When copies are super abundant, they become worthless. When copies are super abundant, stuff which can't be copied becomes scarce and valuable. So he concludes: When copies are free, you need to sell things which can not be copied. The things which cannot be copied are services -- he lists eight generatives that have value. These are immediacy, personalization, interpretation, authenticity, accessibility, embodiment, patronage, and findability. Trust is also mentioned as a intangible asset with significant value. I find that this is a compelling argument, and it seems to be in line with things I hear coming out of OCLC Research, at least, and from the folks at Open Library, too. It will take time for an organization with as much inertia as OCLC has to change its modus operandi, but I think it will come. However, unlike others, I tend to be an optimist in the morning and a cynic by nightfall, so we'll see... Danielle Cunniff Plumer, Coordinator Texas Heritage Digitization Initiative Texas State Library and Archives Commission 512.463.5852 (phone) / 512.936.2306 (fax) [EMAIL PROTECTED] -Original Message- From: Code for Libraries [mailto:[EMAIL PROTECTED] Behalf Of K.G. Schneider Sent: Thursday, February 07, 2008 7:04 AM To: CODE4LIB@listserv.nd.edu Subject: Re: [CODE4LIB] Records for Open Library Maybe Roy will answer that one -- but I doubt its that difficult to guess. OCLC's primary value is its bibliographic database and the information about its member's holdings. Nearly all of it's services are built around this. If they gave that information up to the Open Library, it would most certainly undermine their ILL, Cataloging and Grid Services initiatives. However, if a handful of members in relation to their membership participate in the program -- its no skin off their noses. --TR You know, I realize that's the going-in thinking, and OCLC has shared that with me. I fully understand the need for OCLC to protect its services. But I remember with a previous job that people (even some very important people) thought our product was our data, but it really wasn't: it was the services we wrapped around the data, including maintenance, delivery, affiliated products, etc. It's true that the data had to be good, but that goodness didn't come with a core dump of one-time static data. Keeping our data closed ultimately harmed us, perhaps perniciously, and I wish I had done a better job of championing a different path. I didn't have the skills or vocabulary and to this day I regret that. Karen G. Been there, done that, got the teeshirt Schneider [EMAIL PROTECTED]
[CODE4LIB] getting Worldcat records
Our project maintains a database of bibliographic metadata for all things in astronomy and most of physics. We'd like to add records for books that have been recently added to our library and to correlate existing records with the library holdings. Sounds easy enough, but because of the intricacies of Harvard libraries administration we haven't been able to get a dump of the records, much less a feed. The recent emails about OCLC worldcat records made me wonder if we could get the equivalent data from them (since our library subscribes to them). Essentially what I'd like is a dump of all QB and QC records in OCLC entered by Harvard, so we can index them and then point to the library record in OCLC. Is this (a) legal, (b) feasible, (c) easy? I assume the answer to (a) and (b) is yes, since we have our library's support. If not, are there alternatives? I learned about openlibrary only yesterday, so I haven't had a chance to explore what's in it yet... Thanks, -- Alberto
Re: [CODE4LIB] Records for Open Library
Kevin Kelly had an interesting post on The Technium last week about these sorts of issues (http://www.kk.org/thetechnium/archives/2008/01/better_than_fre.php), and his conclusion is exactly along the lines of Karen's post. His assumptions are: When copies are super abundant, they become worthless. When copies are super abundant, stuff which can't be copied becomes scarce and valuable. So he concludes: When copies are free, you need to sell things which can not be copied. The things which cannot be copied are services -- he lists eight generatives that have value. These are immediacy, personalization, interpretation, authenticity, accessibility, embodiment, patronage, and findability. Trust is also mentioned as a intangible asset with significant value. I find that this is a compelling argument, and it seems to be in line with things I hear coming out of OCLC Research, at least, and from the folks at Open Library, too. It will take time for an organization with as much inertia as OCLC has to change its modus operandi, but I think it will come. However, unlike others, I tend to be an optimist in the morning and a cynic by nightfall, so we'll see... Danielle Cunniff Plumer, Coordinator Texas Heritage Digitization Initiative Texas State Library and Archives Commission 512.463.5852 (phone) / 512.936.2306 (fax) [EMAIL PROTECTED] -Original Message- From: Code for Libraries [mailto:[EMAIL PROTECTED] Behalf Of K.G. Schneider Sent: Thursday, February 07, 2008 7:04 AM To: CODE4LIB@listserv.nd.edu Subject: Re: [CODE4LIB] Records for Open Library Maybe Roy will answer that one -- but I doubt its that difficult to guess. OCLC's primary value is its bibliographic database and the information about its member's holdings. Nearly all of it's services are built around this. If they gave that information up to the Open Library, it would most certainly undermine their ILL, Cataloging and Grid Services initiatives. However, if a handful of members in relation to their membership participate in the program -- its no skin off their noses. --TR You know, I realize that's the going-in thinking, and OCLC has shared that with me. I fully understand the need for OCLC to protect its services. But I remember with a previous job that people (even some very important people) thought our product was our data, but it really wasn't: it was the services we wrapped around the data, including maintenance, delivery, affiliated products, etc. It's true that the data had to be good, but that goodness didn't come with a core dump of one-time static data. Keeping our data closed ultimately harmed us, perhaps perniciously, and I wish I had done a better job of championing a different path. I didn't have the skills or vocabulary and to this day I regret that. Karen G. Been there, done that, got the teeshirt Schneider [EMAIL PROTECTED]