On 6/15/2011 9:31 AM, Eric Hellman wrote:
Clearly, Jonathan has gone through the process of getting his library to think 
through the integration, and it seems to work.

Thank you!

Has there been any opposition?

Not opposition exactly, but it doesn't work perfectly, and people are unhappy when it doesn't work. It can sometimes find the _wrong_ match on a 'foreign' site like Amazon etc. Or avoid finding a right one of course.

Or the definition of right/wrong can be not entirely clear too -- on a bib record for a video of an opera performed, is it right or wrong to supply a link to the print version of the opera? What if the software isn't smart enough to _tell_ you it's an alternate format (it's not), and the link is just in the single flat list of links?

Also issues with avoiding duplicate double URLs when things are in bib records AND in SFX kb AND maybe looked for otherwise by Umlaut. (we have _some_ HathiTrust URLs in our bib records, that came that way from OCLC, who knew?)

These things get really complicated, quickly. I am constantly finding time to do more tweaking, but it'll never be perfect, so people have to get used to lack of perfection. Still when I ask, okay, this HathiTrust/Amazon/Google linking feature is not going to be perfect, would you rather keep it with imperfections we may not be able to fix, or eliminate it -- nobody says eliminate.

What are the reasons that this sort of integration not more widespread? Are they 
technical or institutional? What can be done by producers of open access content to make 
this work better and easier? Are "unified" approaches being touted by vendors 
delivering something really different?

I think they are mostly technical. This stuff is _hard_, because of the (lack of) quality of our own metadata, the lack of quality of third party metadata, the lack of sufficient APIs and Services, and the lack of a local technical infrastructure to support tying everythign together.

So on the one hand, I'm trying to find time for an overhaul of Umlaut to make it easier for people to install and maintain, and I'm hoping I can get some more adoption at that point. To at least provide some open source "local technical infrastructure". Umlaut is intentionally designed to be as easy as possible to integrate with your existing catalog or other service points, as well as to provide 'just in time' services from third party external searches -- that's it's mission, this kind of just-in-time service. ("easy as possible" -- or as easy as I can make it, which sometimes still isn't easy enough, especially if you don't have local technical resources).

But still, it's metadata, metadata, metadata. So what can producers of open access content do to make this work better and easier?

1) Have good metadata for their content, especially including as many identifiers as possible -- ISBN, OCLCnum, LCCN. Even if you aren't an OCLC member and don't have an "OCLC record", if you can figure out what OCLC record represents this thing you've got, list it in the metadata. Even if the ISBN/OCLCnum/LCCN doesn't represent the _exact_ same thing, list it -- ideally somehow identified as 'an alternate manifestation'. Also have author, title, publisher, publication year metadata. If you can have author metadata as an NAF/VIAF controlled form or identifier, even better. Metadata is expensive, but metadata is valuable, the better it is, the better Umlaut's approach can work.

Share the metadata publically, in case someone wants to do something with it.

2) Provide an API that allows lookup of your open access content, searching against the good metadata from #1. Including identifier searches. The thing is, each of (dozens, hundreds, thousands) of open access content providers having such an API --- it's a burdensome expense for each of them, but it's also unrealistic for client software to talk to dozens/hundreds/thousands of APIs.

So this stuff needs to be aggregated in fewer major service points. It could be an aggregator of just metadata that links to content hosted on individual hosts, or it could be an aggregator of content itself. Either way, it needs a good API based on good metadata. "Google" doesn't work as such an aggregator, the APIs it has are too limited functionally and by ToS, and the results do not have sufficient metadata. Maybe the Internet Archive does -- although IA's API's and metadata are sometimes a bit sketchy (If you do put it in IA, make sure it somehow shows in the "Open Library" section and it's APIs -- the OL API's IA has are sufficient for Umlaut's use, but general Internet Archive APIs are not). Or maybe a new aggregator(s) have to be collectively created.

Reply via email to