[CODE4LIB] Unicode persistence (was: it's cool to hate on OpenURL)

2010-04-30 Thread Jakob Voss

Eric Hellman wrote:


May I just add here that of all the things we've talked about in
these threads, perhaps the only thing that will still be in use a
hundred years from now will be Unicode. إن شاء الله


Stuart Yeates wrote:

 Sadly, yes, I agree with you on this.

 Do you have any idea how demotivating that is for those of us
 maintaining collections with works containing characters that don't
 qualify for inclusion?

May I just add there that Unicode is evolving too and you can help to 
get missing characters included. One of the next updates will even 
include hundreds of icons such as a slice of pizza, a kissing couple, 
and the mount Fuji (See this zipped PDF: http://is.gd/bABl9 and 
http://en.wikipedia.org/wiki/Emoji).


I also bet that Unicode will be there in hundred years from now (and 
probably URIs) while things like XML and RDF may be little used then. 
But I fear that the Unicode may be used in a different way just like 
words in natural language change their meanings over the centuries.


And that's why wee need libraries (phew, at least one positive claim 
about these institutions we all are bound to ;-)


Jakob

--
Jakob Voß jakob.v...@gbv.de, skype: nichtich
Verbundzentrale des GBV (VZG) / Common Library Network
Platz der Goettinger Sieben 1, 37073 Göttingen, Germany
+49 (0)551 39-10242, http://www.gbv.de


Re: [CODE4LIB] it's cool to hate on OpenURL

2010-04-30 Thread Jakob Voss

Stuart Yeates wrote:

A great deal of heat has been vented in this thread, and at least a 
little light.


I'd like to invite everyone to contribute to the wikipedia page at 
http://en.wikipedia.org/wiki/OpenURL in the hopes that it evolves into a 
better overview of the protocol, the ecosystem and their place on th web.


[Hint: the best heading for a rant wikipedia is 'criticisms' but you'll 
still need to reference the key points. Links into this thread count as 
references, if you can't find anything else.]


Good point - but writing Wikipedia articles is more work than discussing 
on mailing lists ;-) Instead of improving the OpenURL article I started 
to add to the more relevant[1] COinS article:


http://en.wikipedia.org/wiki/COinS

Maybe some of you (Eric Hellman, Richard Cameron, Daniel Chudnov, Ross 
Singer, Herbert Van de Sompel ...) could fix the history section which I 
tried to reconstruct from historic sources[2] from the Internet without 
violating the Wikipedia NPOV which is hard if you write about things you 
were involved at.


Am I right that neither OpenURL nor COinS strictly defines a metadata 
model with a set of entities/attributes/fields/you-name-it and their 
definition? Apparently all ContextObjects metadata formats are based on 
non-normative implementation guidelines only ??


Cheers
Jakob

[1] My bet: What will remain from OpenURL will be a link server base 
URL that you attach a COinS to


[2] about five years ago, so its historic in terms of internet ;-) By 
the way does anyone have a copy of

http://dbk.ch.umist.ac.uk/wiki/index.php?title=Metadata_in_HTML ?

--
Jakob Voß jakob.v...@gbv.de, skype: nichtich
Verbundzentrale des GBV (VZG) / Common Library Network
Platz der Goettinger Sieben 1, 37073 Göttingen, Germany
+49 (0)551 39-10242, http://www.gbv.de


Re: [CODE4LIB] it's cool to hate on OpenURL (was: Twitter annotations...)

2010-04-30 Thread Owen Stephens
Dead ends from OpenURL enabled hyperlinks aren't a result of the standard
though, but rather an aspect of both the problem they are trying to solve,
and the conceptual way they try to do this.

I'd content these dead ends are an implementation issue - and despite this I
have to say that my experience on the ground is that feedback from library
users on the use of link resolvers is positive - much more so than many of
the other library systems I've been involved with.

What I do see as a problem is that this market seems to have essentially
stagnated, at least as far as I can see. I suspect the reasons for this are
complex, but it would be nice to see some more innovation in this area.

Owen

On Thu, Apr 29, 2010 at 6:14 PM, Ed Summers e...@pobox.com wrote:

 On Thu, Apr 29, 2010 at 12:08 PM, Eric Hellman e...@hellman.net wrote:
  Since this thread has turned into a discussion on OpenURL...
 
  I have to say that during the OpenURL 1.0 standardization process, we
 definitely had moments of despair. Today, I'm willing to derive satisfaction
 from it works and overlook shortcomings. It might have been otherwise.

 Personally, I've followed enough OpenURL enabled hyperlink dead ends
 to contest it works.

 //Ed




-- 
Owen Stephens
Owen Stephens Consulting
Web: http://www.ostephens.com
Email: o...@ostephens.com


Re: [CODE4LIB] Twitter annotations and library software

2010-04-30 Thread Owen Stephens
Alex,

Could you expand on how you think the problem that OpenURL tackles would
have been better approached with existing mechanisms? I'm not debating this
necessarily, but from my perspective when OpenURL was first introduced it
solved a real problem that I hadn't seen solved before.

Owen

On Thu, Apr 29, 2010 at 11:55 PM, Alexander Johannesen 
alexander.johanne...@gmail.com wrote:

 Hi,

 On Thu, Apr 29, 2010 at 22:47, Walker, David dwal...@calstate.edu wrote:
  I would suggest it's more because, once you step outside of the
  primary use case for OpenURL, you end-up bumping into *other* standards.

 These issues were raised all the back when it was created, as well. I
 guess it's easy to be clever in hindsight. :) Here's what I wrote
 about it 5 years ago (http://shelter.nu/blog-159.html) ;

 So let's talk about 'Not invented here' first, because surely, we're
 all guilty of this one from time to time. For example, lately I dug
 into the ANSI/NISO Z39.88 -2004 standard, better known as OpenURL. I
 was looking at it critically, I have to admit, comparing it to what I
 already knew about Web Services, SOA, http,
 Google/Amazon/Flickr/Del.icio.us API's, and various Topic Maps and
 semantic web technologies (I was the technical editor of Explorers
 Guide to the Semantic Web)

 I think I can sum up my experiences with OpenURL as such; why? Why
 have the library world invented a new way of doing things that already
 can be done quite well already? Now, there is absolutely nothing wrong
 with the standard per se (except a pretty darn awful choice of
 name!!), so I'm not here criticising the technical merits and the work
 put into it. No, it's a simple 'why' that I have yet to get a decent
 answer to, even after talking to the OpenURL bigwigs about it. I mean,
 come on; convince me! I'm not unreasonable, no truly, really, I just
 want to be convinced that we need this over anything else.


 Regards,

 Alex
 --
  Project Wrangler, SOA, Information Alchemist, UX, RESTafarian, Topic Maps
 --- http://shelter.nu/blog/ --
 -- http://www.google.com/profiles/alexander.johannesen ---




-- 
Owen Stephens
Owen Stephens Consulting
Web: http://www.ostephens.com
Email: o...@ostephens.com


Re: [CODE4LIB] it's cool to hate on OpenURL

2010-04-30 Thread Thomas Berger
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

Hi,

Jakob Voss schrieb:
...

 Am I right that neither OpenURL nor COinS strictly defines a metadata
 model with a set of entities/attributes/fields/you-name-it and their
 definition? Apparently all ContextObjects metadata formats are based on
 non-normative implementation guidelines only ??

You are right, the 129 page spec of Z39.88 only deals with these in
examples, and IIRC does not even decent pointers to the following:

There are some Core Metadata formats registered under
http://alcme.oclc.org/openurl/servlet/OAIHandler?verb=ListRecordsmetadataPrefix=oai_dcset=Core:Metadata+Formats

notably info:ofi/fmt:kev:mtx:book and info:ofi/fmt:kev:mtx:journal :

 http://www.openurl.info/registry/docs/mtx/info:ofi/fmt:kev:mtx:book ,
 http://www.openurl.info/registry/docs/mtx/info:ofi/fmt:kev:mtx:journal .

As for the semantics of the fields defined there, the description is
certainly not strictly defined in any sense but you can already see
that there is potential for confusion and especially no canonical way
(no way at all?) to achieve satisfying descriptions for several kinds of
non-mainstream objects:

* scholarly articles originally published online (when a journal
  context cannot be constructed)

* articles in books which are not by coincidence conference proceedings

...

ciao
Thomas Berger

-BEGIN PGP SIGNATURE-
Version: GnuPG v1.4.9 (Cygwin)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org

iJwEAQECAAYFAkvamhkACgkQYhMlmJ6W47P/vQP/XfExoDcfEuMGIgFUw4z4GOiM
LhFsKPkVR7wHKDgB+mEF+xnDJoP549Y31YEYKc5lMMz5fVMONUDwwiavVL6IdDvL
EU2jDOQO+WnLiw1XENYoTiuEP6bxyF+gojlBlEN7De1OlCJ97gqitwz6owmTxvp0
BfjBcj4X8//5KVMD0rw=
=1a4i
-END PGP SIGNATURE-


Re: [CODE4LIB] Twitter annotations and library software

2010-04-30 Thread Owen Stephens
Tim,

I'd vote for adopting the same approach as COinS on the basis it already has
some level of adoption, and we know covers at least some of the stuff
libraries and academic users (as used by both libraries and consumer tools
such as Zotero) might want to do. We are talking Books (from what you've
said), so we don't have to worry about other formats. (although it does mean
we can do journal articles and some other stuff as well for no effort)

Mendeley and Zotero already speak COinS, it is pretty simple, and there are
already several code libraries to deal with it.

It isn't where I hope we end up in the longterm but if we talk about this
happening tomorrow, why not use something that is relatively simple, already
has a good set of implementations, and we know works for several cases of
embedding book metadata in a web environment

Owen

On Thu, Apr 29, 2010 at 7:01 PM, Jakob Voss jakob.v...@gbv.de wrote:

 Dear Tim,


 you wrote:

 So this is my recommended framework for proceeding. Tim, I'm afraid
 you'll actually have to do the hard work yourself.


 No, I don't. Because the work isn't fundamentally that hard. A
 complex standard might be, but I never for a moment considered
 anything like that. We have *512 bytes*, and it needs to be usable by
 anyone. Library technology is usually fatally over-engineered, but
 this is a case where that approach isn't even possible.


 Jonathan did a very well summary - you just have to pick what you main
 focus of embedding bibliographic data is.


 A) I favour using the CSL-Record format which I summarized at

 http://wiki.code4lib.org/index.php/Citation_Style_Language

 because I had in mind that people want to have a nice looking citation of
 the publication that someone tweeted about. The drawback is that CSL is less
 adopted and will not always fit in 512 bytes


 B) If you main focus is to link Tweets about the same publication (and
 other stuff about this publication) than you must embed identifiers.
 LibraryThing is mainly based on two identifiers

 1) ISBN to identify editions
 2) LT work ids to identify works

 I wonder why LT work ids have not picked up more although you thankfully
 provide a full mapping to ISBN at
 http://www.librarything.com/feeds/thingISBN.xml.gz but nevermind. I
 thought that some LT records also contain other identifiers such as OCLC
 number, LOC number etc. but maybe I am wrong. The best way to specify
 identifiers is to use an URI (all relevant identifiers that I know have an
 URI form). For ISBN it is

 uri:isbn:{ISBN13}

 For LT Work-ID you can use the URL with your .com top level domain:

 http://www.librarything.com/work/{LTWORKID}http://www.librarything.com/work/%7BLTWORKID%7D

 That would fit for tweets about books with an ISBN and for tweets about a
 work which will make 99.9% of tweets from LT about single publications
 anyway.


 C) If your focus is to let people search for a publication in libraries
 than and to copy bibliographic data in reference management software then
 COinS is a way to go. COinS is based on OpenURL which I and others ranted
 about because it is a crapy library standard like MARC. But unlike other
 metadata formats COinS usually fits in less then 512 bytes. Furthermore you
 may have to deal with it for LibraryThing for libraries anyway.


 Although I strongly favour CSL as a practising library scientist and
 developer I must admit that for LibraryThing the best way is to embed
 identifiers (ISBN and LT Work-ID) and maybe COinS. As long as LibraryThing
 does not open up to more complex publications like preprints of
 proceeding-articles in series etc. but mainly deals with books and works
 this will make LibraryThing users happy.


  Then, three years from now, we can all conference-tweet about a CIL talk,
 about all the cool ways libraries are using Twitter, and how it's such a
 shame that the annotations standard wasn't designed with libraries in mind.


 How about a bet instead of voting. In three years will there be:

 a) No relevant Twitter annotations anyway
 b) Twitter annotations but not used much for bibliographic data
 c) A rich variety of incompatible bibliographic annotation standards
 d) Semantic Web will have solved every problem anyway
 ..

 Cheers
 Jakob

 --
 Jakob Voß jakob.v...@gbv.de, skype: nichtich
 Verbundzentrale des GBV (VZG) / Common Library Network
 Platz der Goettinger Sieben 1, 37073 Göttingen, Germany
 +49 (0)551 39-10242, http://www.gbv.de




-- 
Owen Stephens
Owen Stephens Consulting
Web: http://www.ostephens.com
Email: o...@ostephens.com


Re: [CODE4LIB] Twitter annotations and library software

2010-04-30 Thread Alexander Johannesen
On Fri, Apr 30, 2010 at 18:47, Owen Stephens o...@ostephens.com wrote:
 Could you expand on how you think the problem that OpenURL tackles would
 have been better approached with existing mechanisms?

As we all know, it's pretty much a spec for a way to template incoming
and outgoing URLs, defining some functionality along the way. As such,
URLs with basic URI templates and rewriting have been around for a
long time. Even longer than that is just the basics of HTTP which have
status codes and functionality to do exactly the same. We've been
doing link resolving since mid 90's, either as CGI scripts, or as
Apache modules, so none of this were new. URI comes in, you look it up
in a database, you cross-check with other REQUEST parameters (or
sessions, if you must, as well as IP addresses) and pop out a 303
(with some possible rewriting of the outgoing URL) (with the hack we
needed at the time to also create dummy pages with META tags
*shudder*).

So the idea was to standardize on a way to do this, and it was a good
idea as such. OpenURL *could* have had a great potential if it
actually defined something tangible, something concrete like a model
of interaction or basic rules for fishing and catching tokens and the
like, and as someone else mentioned, the 0.1 version was quite a good
start. But by the time when 1.0 came out, all the goodness had turned
so generic and flexible in such a complex way that handling it turned
you right off it. The standard also had a very difficult language, and
more specifically didn't use enough of the normal geeky language used
by sysadmins around. The more I tried to wrap my head around it, the
more I felt like just going back to CGI scripts that looked stuff up
in a database. It was easier to hack legacy code, which, well, defeats
the purpose, no?

Also, forgive me if I've forgotten important details; I've suppressed
this part of my life. :)


Kind regards,

Alex
-- 
 Project Wrangler, SOA, Information Alchemist, UX, RESTafarian, Topic Maps
--- http://shelter.nu/blog/ --
-- http://www.google.com/profiles/alexander.johannesen ---


Re: [CODE4LIB] Twitter annotations and library software

2010-04-30 Thread Owen Stephens
Thanks Alex,

This makes sense, and yes I see what your saying - and yes, if you end up
going back to custom coding because it's easier it does seem to defeat the
purpose.

However I'd argue that actually OpenURL 'succeeded' because it did manage to
get some level of acceptance (ignoring the question of whether it is v0.1 or
v1.0) - the cost of developing 'link resolvers' would have been much higher
if we'd been doing something different for each publisher/platform. In this
sense (I'd argue) sometimes crappy standards are better than none.

We've used OpenURL v1.0 in a recent project and because we were able to
simply pick up code already done for Zotero, and  we already had an OpenURL
resolver, the amount of new code we needed for this was minimal.

I think the point about Link Resolvers doing stuff that Apache and CGI
scripts were already doing is a good one - and I've argued before that what
we actually should do is separate some of this out (a bit like Johnathan did
with Umlaut) into an application that can answer questions about location
(what is generally called the KnowledgeBase in link resolvers) and the
applications that deal with analysing the context and the redirection

(To introduce another tangent in a tangential thread, interestingly (I
think!) I'm having a not dissimilar debate about Linked Data at the moment -
there are many who argue that it is too complex and that as long as you have
a nice RESTful interface you don't need to get bogged down in ontologies and
RDF etc. I'm still struggling with this one - my instinct is that it will
pay to standardise but so far I've not managed to convince even myself this
is more than wishful thinking at the moment)

Owen

On Fri, Apr 30, 2010 at 10:33 AM, Alexander Johannesen 
alexander.johanne...@gmail.com wrote:

 On Fri, Apr 30, 2010 at 18:47, Owen Stephens o...@ostephens.com wrote:
  Could you expand on how you think the problem that OpenURL tackles would
  have been better approached with existing mechanisms?

 As we all know, it's pretty much a spec for a way to template incoming
 and outgoing URLs, defining some functionality along the way. As such,
 URLs with basic URI templates and rewriting have been around for a
 long time. Even longer than that is just the basics of HTTP which have
 status codes and functionality to do exactly the same. We've been
 doing link resolving since mid 90's, either as CGI scripts, or as
 Apache modules, so none of this were new. URI comes in, you look it up
 in a database, you cross-check with other REQUEST parameters (or
 sessions, if you must, as well as IP addresses) and pop out a 303
 (with some possible rewriting of the outgoing URL) (with the hack we
 needed at the time to also create dummy pages with META tags
 *shudder*).

 So the idea was to standardize on a way to do this, and it was a good
 idea as such. OpenURL *could* have had a great potential if it
 actually defined something tangible, something concrete like a model
 of interaction or basic rules for fishing and catching tokens and the
 like, and as someone else mentioned, the 0.1 version was quite a good
 start. But by the time when 1.0 came out, all the goodness had turned
 so generic and flexible in such a complex way that handling it turned
 you right off it. The standard also had a very difficult language, and
 more specifically didn't use enough of the normal geeky language used
 by sysadmins around. The more I tried to wrap my head around it, the
 more I felt like just going back to CGI scripts that looked stuff up
 in a database. It was easier to hack legacy code, which, well, defeats
 the purpose, no?

 Also, forgive me if I've forgotten important details; I've suppressed
 this part of my life. :)


 Kind regards,

 Alex
 --
  Project Wrangler, SOA, Information Alchemist, UX, RESTafarian, Topic Maps
 --- http://shelter.nu/blog/ --
 -- http://www.google.com/profiles/alexander.johannesen ---




-- 
Owen Stephens
Owen Stephens Consulting
Web: http://www.ostephens.com
Email: o...@ostephens.com


Re: [CODE4LIB] Twitter annotations and library software

2010-04-30 Thread Alexander Johannesen
On Fri, Apr 30, 2010 at 20:29, Owen Stephens o...@ostephens.com wrote:
 However I'd argue that actually OpenURL 'succeeded' because it did manage to
 get some level of acceptance (ignoring the question of whether it is v0.1 or
 v1.0) - the cost of developing 'link resolvers' would have been much higher
 if we'd been doing something different for each publisher/platform. In this
 sense (I'd argue) sometimes crappy standards are better than none.

Well, perhaps. I see OpenURL as the natural progression from PURL, in
which both have their degree of success, however I'm careful using
that word as I live on the outside of the library world. It may well
be a success on the inside. :)

 I think the point about Link Resolvers doing stuff that Apache and CGI
 scripts were already doing is a good one - and I've argued before that what
 we actually should do is separate some of this out (a bit like Johnathan did
 with Umlaut) into an application that can answer questions about location
 (what is generally called the KnowledgeBase in link resolvers) and the
 applications that deal with analysing the context and the redirection

Yes, split it into smaller chunks is always smart, especially with
complex issues. For example, in the Topic Maps world, the who standard
(reference model, data model, query language, constraint language, XML
exchange language, various notational languages) is wrapped up with a
guide in the middle. Make them into smaller parcels, and make your
flexible point there. If you pop it all into one, no one will read it
and fully understand it. (And don't get me started on the WS-* set of
standards on the same issues ...)

 (To introduce another tangent in a tangential thread, interestingly (I
 think!) I'm having a not dissimilar debate about Linked Data at the moment -
 there are many who argue that it is too complex and that as long as you have
 a nice RESTful interface you don't need to get bogged down in ontologies and
 RDF etc. I'm still struggling with this one - my instinct is that it will
 pay to standardise but so far I've not managed to convince even myself this
 is more than wishful thinking at the moment)

Ah, now this is certainly up my alley. As you might have seen, I'm a
Topic Maps guy, and we have in our model a distinction between three
different kinds of identities; internal, external indicators and
published subject identifiers. The RDF world only had rdf:about, so
when you used www.somewhere.org, are you talking about that thing,
or does that thing represent something you're talking about? Tricky
stuff which has these days become a *huge* problem with Linked Data.
And yes, they're trying to solve that by issuing a HTTP 303 status
code as a means of declaring the identifiers imperative, which is a
*lot* of resolving to do on any substantial set of data, and in my
eyes a huge ugly hack. (And what if your Internet falls down? Tough.)

Anyway, here's more on these identity problems ;
   http://www.ontopia.net/topicmaps/materials/identitycrisis.html

As to the RESTful notions, they only take you as far as content-types
can take you. Sure, you can gleam semantics from it, but I reckon
there's an impedance mismatch between just the things librarians how
got down pat ; meta data vs. data. CRUD or, in this example, GPPD
(get/post/put/delete), who aren't in a dichotomy btw, can only
determine behavior that enables certain semantic paradigms, but cannot
speak about more complex relationships or even modest models. (Very
often models aren't actionable :)

The funny thing is that after all these years of working with Topic
Maps I find that these hard issues have been solved years ago, and the
rest of the world is slowly catching up to it. I blame the lame
DAML+OIL background of RDF and OWL, to be honest; a model too simple
to be elegantly advanced and too complex to be easily useful.


Kind regards,

Alex
-- 
 Project Wrangler, SOA, Information Alchemist, UX, RESTafarian, Topic Maps
--- http://shelter.nu/blog/ --
-- http://www.google.com/profiles/alexander.johannesen ---


Re: [CODE4LIB] it's cool to hate on OpenURL (was: Twitter annotations...)

2010-04-30 Thread Kyle Banerjee
 Dead ends from OpenURL enabled hyperlinks aren't a result of the standard
 though, but rather an aspect of both the problem they are trying to solve,
 and the conceptual way they try to do this.

 I'd content these dead ends are an implementation issue.


Absolutely. There is no inherent reason that an OpenURL has to be a plain
text link that must be manually clicked nor is there any requirement that
the resolver simply return an HTML page that may or may not contain plain
text links the user can click on.

An obvious thing for a resolver to be able to do is return results in JSON
so the OpenURL can be more than a static link. But since the standard
defines no such response, the site generating the OpenURL would have to know
something about the resolver.


 What I do see as a problem is that this market seems to have essentially
 stagnated, at least as far as I can see. I suspect the reasons for this are
 complex


I suspect they are simple. The easy part of OpenURL addresses a compelling
use case, but there just isn't that much pressure to take things to the next
level. If people really got that upset about dead links -- and in many
systems this is impossible because you'll be offered ILL fulfillment if no
electronic copy is available -- there would be more incentive to handle
resolution before the user clicks on the link.

In short, you reach a point of diminishing returns with OpenURL very
quickly.

kyle


[CODE4LIB] Job Posting: Database Specialist @ Penn State University Libraries

2010-04-30 Thread Janis Mathewson

Hi all,

We have an application developer job opportunity in my department at the 
Penn State Libraries.


Thanks,
Janis

Database Specialist
Level: 03
Work Unit: University Libraries
Department: Department Of Information Technologies
Job Number: 32060

Penn State University Libraries is seeking a database/applications 
specialist to support Library initiatives in Web application 
development, data analysis and database design. Design and development 
of Web/data applications incorporating the complete software development 
cycle including needs assessment, definition of requirements, prototype 
development, application design, testing, coding, documentation and 
integration of Web application services into other enterprise platforms 
as appropriate. Typically requires Bachelor’s degree plus four years of 
related experience or an equivalent combination of education and 
experience. Experience should include database design, data 
manipulation, developing and deploying software or an equivalent 
combination of education and experience. The candidate must be able to: 
keep informed of new technologies for use with application development 
initiatives; create complex queries and generate custom reports from 
Oracle tables and other data sources; assist in defining data needs and 
in interpretation of existing data for library administration; provide 
advanced SQL code to aid colleagues in database planning, querying, data 
transformation and analyzing; design database architecture for new 
Web/data applications; lead and participate on project teams; work 
closely with other developers in the unit and serve on cross functional 
teams; disseminate project information in written and oral formats to a 
technical and non-specialized audience in a variety of settings (public, 
small group, individual, etc.). The ideal candidate must have experience 
with: a server side scripting language such as ColdFusion, PHP, JSP, 
.NET etc., advanced SQL and database design skills, XHTML, CSS, 
demonstrated commitment to diversity; ability to effectively communicate 
technical information to a non-technical audience and excellent 
organizational, interpersonal and communication skills; and demonstrated 
ability to work in a team environment. Preferred skills include 
proficiency with Oracle databases and ColdFusion, strong understanding 
of XML/XSLT, AJAX, JavaScript, PL/SQL and familiarity with a web content 
management system. A commitment to providing outstanding customer 
service and a driven passion for technology are essential. The 
successful candidate must also possess the ability to interact 
effectively with individuals from a variety of cultures and backgrounds. 
The University Libraries is a multicultural environment that embraces 
respect and diversity.



To apply please visit: http://www.psu.jobs/Search/Opportunities.html

--
Janis Mathewson
Department for Information Technologies (I-Tech)
Penn State University Libraries
University Park, PA 16802-1812
(814) 865-4867


Re: [CODE4LIB] it's cool to hate on OpenURL

2010-04-30 Thread Ross Singer
On Fri, Apr 30, 2010 at 4:09 AM, Jakob Voss jakob.v...@gbv.de wrote:

 Am I right that neither OpenURL nor COinS strictly defines a metadata model
 with a set of entities/attributes/fields/you-name-it and their definition?
 Apparently all ContextObjects metadata formats are based on non-normative
 implementation guidelines only ??

You are right.  Z39.88 and (by extension) COinS really only defines
the ContextObject itself.  So it defines the carrier package, it's
administrative elements, referents, referrers, referringentities,
services, requester and resolver and their transports.

It doesn't really specify what should actually go into any of those
slots.  The idea is that it defers to the community profiles for that.

In the XML context object, you can send more than one metadata-by-val
element (or metadata-by-ref) per entity (ref, rfr, rfe, svc, req, res)
- I'm not sure what is supposed to happen, for example, if you send a
referent that has multiple MBV elements that don't actually describe
the same thing.

-Ross.


Re: [CODE4LIB] it's cool to hate on OpenURL (was: Twitter annotations...)

2010-04-30 Thread Ross Singer
On Fri, Apr 30, 2010 at 7:59 AM, Kyle Banerjee kyle.baner...@gmail.com wrote:

 An obvious thing for a resolver to be able to do is return results in JSON
 so the OpenURL can be more than a static link. But since the standard
 defines no such response, the site generating the OpenURL would have to know
 something about the resolver.

I actually think this lack of any specified response format is a large
factor in the stagnation of OpenURL as a technology.  Since a resolver
is under no obligation to do anything but present a web page it's
difficult for local entrepreneurial types to build upon the
infrastructure simply because there are no guarantees that it will
work anywhere else (or even locally, depending on your vendor, I
suppose), much less contribute back to the ecosystem.

Umlaut was able to exist because (for better or worse) SFX has an XML
output.  It has never been able to scale horizontally, however,
because to work with another vendor's link resolver (which should
actually be quite straightforward) it requires a connector to whatever
*their* proprietary API needs.

I could definitely see a project like Umlaut providing a 'de facto'
machine readable response for SAP 1/2 requests that content providers
could then use to start offering better integration at *their* end.

This assumes that more than 5 libraries would actually be using it, however.

-Ross.


Re: [CODE4LIB] it's cool to hate on OpenURL (was: Twitter annotations...)

2010-04-30 Thread Ed Summers
On Fri, Apr 30, 2010 at 9:09 AM, Ross Singer rossfsin...@gmail.com wrote:
 I actually think this lack of any specified response format is a large
 factor in the stagnation of OpenURL as a technology.  Since a resolver
 is under no obligation to do anything but present a web page it's
 difficult for local entrepreneurial types to build upon the
 infrastructure simply because there are no guarantees that it will
 work anywhere else (or even locally, depending on your vendor, I
 suppose), much less contribute back to the ecosystem.

I agree. And that's an issue with the standard, not the implementations.

//Ed


[CODE4LIB] Approaches to Did You Mean Query Spelling Suggestions

2010-04-30 Thread Cory Lown
I'm exploring options for implementing a spelling suggestion or basic query 
reformulation service in our home grown search application (searches library 
website, catalog, summon, and a few other bins). Right now, my thought is to 
provide results for whatever was searched for 'as is' and generate a link for 
an alternate search -- sort of like what The Google does.  I am concerned only 
with correcting spelling errors, not so much with topically related search 
suggestions.

The 3 options I've found that seem worth further investigation are:

- Yahoo Search spellingSuggestion service: 
http://developer.yahoo.com/search/web/V1/spellingSuggestion.html 

- GNU Aspell: http://aspell.net

- Ockham Spell service: http://spell.ockham.org/about/index.html . There is a 
thread on Code4Lib back in 2005 about this: 
http://serials.infomotions.com/code4lib/sru/?operation=searchRetrieveversion=1.1stylesheet=/code4lib/sru/style.xslquery=spelling+server

Anyone doing something like this? What tools are you using? What have you 
tried? What worked well? Have I overlooked an option that I should consider?

Thanks,

Cory


Cory Lown
NCSU Libraries
Raleigh, NC
cory_l...@ncsu.edu


Re: [CODE4LIB] it's cool to hate on OpenURL (was: Twitter annotations...)

2010-04-30 Thread Corey A Harper

Hi All,

Though hesitant to jump in here, I agree with Owen that the dead ends 
aren't a standards issue. The bloat of the standard is, as is the lack 
of a standardized response format, but the dead ends have to do with bad 
metadata being coded into open-URLs and with breakdowns in the 
connection between content aggregators/providers and knowledge base 
maintainers.


Work in this area isn't completely stagnant, though. The joint NISO/UK 
Serials Group's Knowledge Bases And Related Tools working group is 
looking towards solutions to exactly these problems.

http://www.uksg.org/kbart

Their initial report on best practice for content providers and KB 
maintainers is worth a look.


-Corey

Owen Stephens wrote:

Dead ends from OpenURL enabled hyperlinks aren't a result of the standard
though, but rather an aspect of both the problem they are trying to solve,
and the conceptual way they try to do this.

I'd content these dead ends are an implementation issue - and despite this I
have to say that my experience on the ground is that feedback from library
users on the use of link resolvers is positive - much more so than many of
the other library systems I've been involved with.

What I do see as a problem is that this market seems to have essentially
stagnated, at least as far as I can see. I suspect the reasons for this are
complex, but it would be nice to see some more innovation in this area.

Owen

On Thu, Apr 29, 2010 at 6:14 PM, Ed Summers e...@pobox.com wrote:


On Thu, Apr 29, 2010 at 12:08 PM, Eric Hellman e...@hellman.net wrote:

Since this thread has turned into a discussion on OpenURL...

I have to say that during the OpenURL 1.0 standardization process, we

definitely had moments of despair. Today, I'm willing to derive satisfaction
from it works and overlook shortcomings. It might have been otherwise.

Personally, I've followed enough OpenURL enabled hyperlink dead ends
to contest it works.

//Ed







--
Corey A Harper
Metadata Services Librarian
New York University Libraries
20 Cooper Square, 3rd Floor
New York, NY 10003-7112
212.998.2479
corey.har...@nyu.edu


Re: [CODE4LIB] it's cool to hate on OpenURL (was: Twitter annotations...)

2010-04-30 Thread Eric Hellman
Eek. I was hoping for something much simpler. Do you realize that you're asking 
for service taxonomy?

On Apr 30, 2010, at 10:22 AM, Ross Singer wrote:

 I think the basis of a response could actually be another context
 object with the 'services' entity containing a list of
 services/targets that are formatted in some way that is appropriate
 for the context and the referent entity enhanced with whatever the
 resolver can add to the puzzle.


[CODE4LIB] NYTSL Spring Meeting and Program: May 19, 2010

2010-04-30 Thread Lisa Genoese
Cross-posted; apologies for duplication.

*

Dear colleagues and friends,

Please join us for:



New York Technical Services Librarians

Spring Meeting  Program

Wednesday, May 19, 2010



Online registration is now open!

http://www.nytsl.orghttp://www.nytsl.org/



Space is limited so please register early.  Registration deadline: Friday, May 
14, 2010.



TOPIC: Communities of Interest : A New Model for Institutional Repositories



SPEAKER: Kate Wittenberg



As Project Director, Client and Partnership Development at Ithaka, Kate focuses 
on building partnerships among scholars, publishers, libraries, technology 
providers, and societies with an interest in promoting the development of 
digital scholarship and building and sustaining innovative initiatives. Before 
coming to Ithaka, Kate was the Director of EPIC (the Electronic Publishing 
Initiative at Columbia) a pioneering initiative in digital publishing, and a 
model partnership for libraries, presses, and academic IT departments. Some of 
the ventures produced by EPIC include CIAO (Columbia International Affairs 
Online), Gutenberg-E (a reinvention of the monograph as an electronic work), 
and Jazz Studies Online.



WHERE and WHEN:

South Court Auditorium

NYPL Humanities  Social Sciences Library

Fifth Avenue and 42nd Street

New York, N.Y. 10018



Wednesday, May 19, 2010

Refreshments, 5:00-6:00 PM

Meeting  Program, 6:00-8:00 PM



REGISTRATION and PRICING:

NYTSL now offers PayPal as a preferred payment method. Please go to 
http://www.nytsl.orghttp://www.nytsl.org/  for more information.  Mail-in 
registration forms are also available on the website.



Program (Members): $15.00

Program + Membership (Non-members and renewals), Sept. 2009-Aug. 2010 Academic 
Year: $20.00

Program only (Non-members): $25.00


For questions about membership status, please contact Lisa Genoese, 
lgeno...@nyam.org mailto:lgeno...@nyam.org


[CODE4LIB] New open-source inventory system

2010-04-30 Thread Yitzchak Schaffer

Hello all, apologies for cross-posting.

We have just released a PHP/symfony application for processing the 
inventory of ILS item, under the new BSD license.  This is great for 
those of us with an ILS that can do data export, but lacks an inventory 
module.  Code is included for importing items from III export files, as 
well as checking LC shelf-order.


http://bitbucket.org/yitznewton/ils-inventory

Have fun; let me know if you want to contribute.

--
Yitzchak Schaffer
Systems Manager
Touro College Libraries
33 West 23rd Street
New York, NY 10010
Tel (212) 463-0400 x5230
Fax (212) 627-3197
Email yitzchak.schaf...@tourolib.org

Access Problems? Contact systems.libr...@touro.edu


[CODE4LIB] SRU/ZeeRex explain question : record schemas

2010-04-30 Thread Jonathan Rochkind
This page:
http://www.loc.gov/standards/sru/resources/schemas.html

says:

The Explain document lists the XML schemas for a given database in which 
records may be transferred. Every schemas is unambiguously identified by a URI 
and a server may assign a short name, which may or may not be the same as the 
short name listed in the table below (and may differ from the short name that 
another server assigns).


But perusing the SRU/ZeeRex Explain documentation I've been able to find, I've 
been unable to find WHERE in the Explain document this information is 
listed/advertised. 

Can anyone clue me in?


Re: [CODE4LIB] SRU/ZeeRex explain question : record schemas

2010-04-30 Thread Ray Denenberg, Library of Congress

schemaInfo is what you're looking for I think.

Look at http://z3950.loc.gov:7090/voyager.

Line 74, for example,
schemaInfo
schema identifier=info:srw/schema/1/marcxml-v1.1 sort=false 
name=marcxml

titleMARCXML/title
/schema


Is this what you're looking for?

--Ray


- Original Message - 
From: Jonathan Rochkind rochk...@jhu.edu

To: CODE4LIB@LISTSERV.ND.EDU
Sent: Friday, April 30, 2010 3:57 PM
Subject: [CODE4LIB] SRU/ZeeRex explain question : record schemas



This page:
http://www.loc.gov/standards/sru/resources/schemas.html

says:

The Explain document lists the XML schemas for a given database in which 
records may be transferred. Every schemas is unambiguously identified by a 
URI and a server may assign a short name, which may or may not be the same 
as the short name listed in the table below (and may differ from the short 
name that another server assigns).



But perusing the SRU/ZeeRex Explain documentation I've been able to find, 
I've been unable to find WHERE in the Explain document this information is 
listed/advertised.


Can anyone clue me in? 


Re: [CODE4LIB] SRU/ZeeRex explain question : record schemas

2010-04-30 Thread LeVan,Ralph
There's a schemaInfo element right under the explain element that
carries that data.

Here's a pointer to the Explain record for my LCNAF database.

http://alcme.oclc.org/srw/search/lcnaf

Don't let the browser fool you!  View the source and you'll see the
actual XML that was returned.  The schemaInfo element is towards the
bottom.

Ralph

p.s. For my LCNAF friends on the list, note the change to the database
element in the serverInfo section.  It now include the update date and
number of records in the database!  That's being automatically generated
by the SRW server.


 -Original Message-
 From: Code for Libraries [mailto:code4...@listserv.nd.edu] On Behalf
Of
 Jonathan Rochkind
 Sent: Friday, April 30, 2010 3:58 PM
 To: CODE4LIB@LISTSERV.ND.EDU
 Subject: [CODE4LIB] SRU/ZeeRex explain question : record schemas
 
 This page:
 http://www.loc.gov/standards/sru/resources/schemas.html
 
 says:
 
 The Explain document lists the XML schemas for a given database in
which
 records may be transferred. Every schemas is unambiguously identified
by a
 URI and a server may assign a short name, which may or may not be the
same
 as the short name listed in the table below (and may differ from the
short name
 that another server assigns).
 
 
 But perusing the SRU/ZeeRex Explain documentation I've been able to
find, I've
 been unable to find WHERE in the Explain document this information is
 listed/advertised.
 
 Can anyone clue me in?


Re: [CODE4LIB] Approaches to Did You Mean Query Spelling Suggestions

2010-04-30 Thread Genny Engel
I am not a fan of services that give spelling suggestions based on their own 
web-wide universe of terms.  It's better to suggest only terms that are 
actually found within the smaller universe of your own materials.  That way the 
user isn't offered a link that's guaranteed to get them zero results.  However, 
this only works if you're actually indexing the contents of all your sources 
into a local index -- not if you're dynamically retrieving the results from 
different sources.

I don't have personal experience with any of the options you list, but from 
briefly looking at them, I would be inclined toward Aspell since you'd control 
the dictionary.  

Ideally the dictionary would auto-populate from the index the search engine 
builds.  We use Thunderstone Webinator http://www.thunderstone.com for our 
website search and it uses its own index for the spelling suggestions.  It also 
lists in parentheses the number of results that match each suggestion.   
http://www.sonomalibrary.org/search 


Genny Engel
Sonoma County Library
gen...@sonoma.lib.ca.us
707 545-0831 x581
www.sonomalibrary.org



 cory_l...@ncsu.edu 04/30/10 06:55AM 
I'm exploring options for implementing a spelling suggestion or basic query 
reformulation service in our home grown search application (searches library 
website, catalog, summon, and a few other bins). Right now, my thought is to 
provide results for whatever was searched for 'as is' and generate a link for 
an alternate search -- sort of like what The Google does.  I am concerned only 
with correcting spelling errors, not so much with topically related search 
suggestions.

The 3 options I've found that seem worth further investigation are:

- Yahoo Search spellingSuggestion service: 
http://developer.yahoo.com/search/web/V1/spellingSuggestion.html 

- GNU Aspell: http://aspell.net 

- Ockham Spell service: http://spell.ockham.org/about/index.html . There is a 
thread on Code4Lib back in 2005 about this: 
http://serials.infomotions..com/code4lib/sru/?operation=searchRetrieveversion=1.1stylesheet=/code4lib/sru/style.xslquery=spelling+server
 

Anyone doing something like this? What tools are you using? What have you 
tried? What worked well? Have I overlooked an option that I should consider?

Thanks,

Cory


Cory Lown
NCSU Libraries
Raleigh, NC
cory_l...@ncsu.edu


Re: [CODE4LIB] Approaches to Did You Mean Query Spelling Suggestions

2010-04-30 Thread Brad Dewar
Seconded.  We use Solr's SpellCheckComponent to accomplish exactly this.

Brad Dewar
bde...@stfx.ca



-Original Message-
From: Code for Libraries [mailto:code4...@listserv.nd.edu] On Behalf Of Genny 
Engel
Sent: April-30-10 6:00 PM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Approaches to Did You Mean Query Spelling Suggestions

I am not a fan of services that give spelling suggestions based on their own 
web-wide universe of terms.  It's better to suggest only terms that are 
actually found within the smaller universe of your own materials.  That way the 
user isn't offered a link that's guaranteed to get them zero results.  However, 
this only works if you're actually indexing the contents of all your sources 
into a local index -- not if you're dynamically retrieving the results from 
different sources.

I don't have personal experience with any of the options you list, but from 
briefly looking at them, I would be inclined toward Aspell since you'd control 
the dictionary.  

Ideally the dictionary would auto-populate from the index the search engine 
builds.  We use Thunderstone Webinator http://www.thunderstone.com for our 
website search and it uses its own index for the spelling suggestions.  It also 
lists in parentheses the number of results that match each suggestion.   
http://www.sonomalibrary.org/search 


Genny Engel
Sonoma County Library
gen...@sonoma.lib.ca.us
707 545-0831 x581
www.sonomalibrary.org



 cory_l...@ncsu.edu 04/30/10 06:55AM 
I'm exploring options for implementing a spelling suggestion or basic query 
reformulation service in our home grown search application (searches library 
website, catalog, summon, and a few other bins). Right now, my thought is to 
provide results for whatever was searched for 'as is' and generate a link for 
an alternate search -- sort of like what The Google does.  I am concerned only 
with correcting spelling errors, not so much with topically related search 
suggestions.

The 3 options I've found that seem worth further investigation are:

- Yahoo Search spellingSuggestion service: 
http://developer.yahoo.com/search/web/V1/spellingSuggestion.html 

- GNU Aspell: http://aspell.net 

- Ockham Spell service: http://spell.ockham.org/about/index.html . There is a 
thread on Code4Lib back in 2005 about this: 
http://serials.infomotions..com/code4lib/sru/?operation=searchRetrieveversion=1.1stylesheet=/code4lib/sru/style.xslquery=spelling+server
 

Anyone doing something like this? What tools are you using? What have you 
tried? What worked well? Have I overlooked an option that I should consider?

Thanks,

Cory


Cory Lown
NCSU Libraries
Raleigh, NC
cory_l...@ncsu.edu


Re: [CODE4LIB] Approaches to Did You Mean Query Spelling Suggestions

2010-04-30 Thread Chad Fennell
 Seconded.  We use Solr's SpellCheckComponent to accomplish exactly this.

+1


[CODE4LIB] It's cool to love milk and cookies

2010-04-30 Thread Simon Spero

I like chocolate milk.