Re: [CODE4LIB] Metadata

2012-02-14 Thread Mike Taylor
Reminds me of this article, which frighteningly is now eight years old:
 http://www.miketaylor.org.uk/tech/metadata.html

-- Mike.


On 14 February 2012 06:25, Kåre Fiedler Christiansen
k...@statsbiblioteket.dk wrote:
 You realize, of course, that discussing the use of the word metametadata 
 could be described as metametametadata? Which would make my post 
 metametametametadata. At which point it all turns into silliness (which it 
 certainly wasn't before... right? :-) ).

 Best,
  Kåre

 From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Kurt 
 Nordstrom

 I got such dirty looks when I used the term metametadata to
 describe
 something. ;)

 -Kurt

 On 02/13/2012 02:39 PM, Becky Yoose wrote:
  Could this conversation be described as metametadata?
 
  *runs, hides*
 
  Thanks,
  Becky
 
 



Re: [CODE4LIB] Metadata

2012-02-14 Thread Michael Hopwood
 The actual data (the novel) is not in the catalog (which is composed only of 
 metadata).

 That's a technical limitation.

It's also a legal/commercial limitation, as well as a question of provenance.

To summarise a lot of good points made already:

 An item of metadata is a relationship that someone claims to exist between 
two entities.
 - source: http://www.doi.org/topics/indecs/indecs_framework_2000.pdf

Cheers,

M

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Nate 
Vack
Sent: 13 February 2012 22:39
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Metadata

On Mon, Feb 13, 2012 at 4:25 PM, Genny Engel gen...@sonoma.lib.ca.us wrote:

 You simply can't use the average library catalog to look up Author X's novel 
 that starts with the sentence So a string walks into a bar.  The actual 
 data (the novel) is not in the catalog (which is composed only of metadata).

That's a technical limitation. If you're Google Books (or any other fulltext 
index), the actual data *is* in the catalog, and data and metadata are again 
functionally identical.

The best working definition of metadata I've come up with is something I have 
a field for in my data cataloging program.

I think it's kind of a circular issue: We know metadata and data are separate 
because our software and workflow require it. Software and workflows are 
designed to separate metadata and data because we know they're separate.

-n


Re: [CODE4LIB] Metadata

2012-02-14 Thread Nate Vack
On Tue, Feb 14, 2012 at 1:22 AM, Graham Triggs grahamtri...@gmail.com wrote:

 That's an interesting distinction though. Do you need all that data in
 order to make sense of the results? You don't [necessarily] need to
 know who conducted some research, or when they conducted it in order
 to analyse and make sense of the data. In the context of having the
 data, this other information becomes irrelevant in terms of
 understanding what that data says.

It is *essential* to understanding what the data says. Perhaps you
find out your sensor was on the fritz during a time period -- you need
to be able to know what datasets are suspect. Maybe the blood pressure
effect you're looking at is mediated by circadian rhythms, and hence,
times of day.

Not all of your data is necessary in every analysis, but a bunch of
blood pressure measurements in the absence of contextual information
is universally useless.

The metadata is part of the data.

-n


Re: [CODE4LIB] Metadata

2012-02-14 Thread Michael Hopwood
Having done research, and now working in a very varied metadata role, I don't 
quite understand this discussion about data that is or isn't metadata. 
Scientific data is a great example of structured data, but it's not impossible 
to distinguish it from metadata purely describing a dataset.

However, if you have scientific research data created during the experiments, 
even if it's operational, it's clearly part of the data. This doesn't mean 
there can't be metadata describing *that data*. Just because it's not glamorous 
data doesn't mean it's not essential to the scientific process. Similarly, just 
being about mundane or procedural things doesn't make data into metadata...!

You're absolutely right, the contextual information is certainly part of the 
experimental outcome in this example; otherwise it would be abstract data such 
as one might use in a textbook example.

Metadata would describe the dataset itself, not the scientific research. 
There's always a certain ambiguity involved in identifying the data as 
distinct from the metadata, and it's a false dichotomy to suggest metadata is 
not useful at all for the domain expert. It's contextual, and the definition is 
always at least partly based on your use case for the data and its description.

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Nate 
Vack
Sent: 14 February 2012 14:45
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Metadata

On Tue, Feb 14, 2012 at 1:22 AM, Graham Triggs grahamtri...@gmail.com wrote:

 That's an interesting distinction though. Do you need all that data in 
 order to make sense of the results? You don't [necessarily] need to 
 know who conducted some research, or when they conducted it in order 
 to analyse and make sense of the data. In the context of having the 
 data, this other information becomes irrelevant in terms of 
 understanding what that data says.

It is *essential* to understanding what the data says. Perhaps you find out 
your sensor was on the fritz during a time period -- you need to be able to 
know what datasets are suspect. Maybe the blood pressure effect you're looking 
at is mediated by circadian rhythms, and hence, times of day.

Not all of your data is necessary in every analysis, but a bunch of blood 
pressure measurements in the absence of contextual information is universally 
useless.

The metadata is part of the data.

-n


[CODE4LIB] New England Code4Lib Regional Conference

2012-02-14 Thread Greenspun, Cindy
Hello -

The planning process has begun for a New England regional gathering and we 
would like to try and get some potential attendance numbers together.  We put 
together a brief poll which should only take a minute or less to complete:
https://yalesurvey.qualtrics.com/SE/?SID=SV_24OogxklhikrpzK.

Dates: Friday, October 26 and Saturday, October 27
Location: Yale University, New Haven, CT

This will be a great opportunity to meet your peers at local institutions and 
converse on code4lib related topics!  You could plan to come for a day at a low 
cost day trip or stay for 2 days and maximize your interaction time.

We hope to see you there!  This email is sent on behalf of the NECode4lib 
planning committee.

Matthew Beacom
Michael Friscia
Cindy Greenspun
Michelle Hudson
James Luker
Mark Matienzo
Joe Montibello
Tito Sierra
Kalee Sprague
Randy Stern

Cindy Greenspun
IT and Project Manager
Database Administrator
cindy.greens...@yale.edu


Re: [CODE4LIB] RDF advice

2012-02-14 Thread Ethan Gruber
Hi Karen,

Thanks.  Would it be odd to use foaf:primaryTopic when FOAF isn't used to
describe other attributes of a concept?

Ethan

On Mon, Feb 13, 2012 at 5:59 PM, Karen Coyle li...@kcoyle.net wrote:

 On 2/13/12 1:43 PM, Ethan Gruber wrote:

 Hi Patrick,

 Thanks.  That does make sense.  Hopefully others will weigh in with
 agreement (or disagreement).  Sometimes these semantic languages are so
 flexible that it's unsettling.  There are a million ways to do something
 with only de facto standards rather than restricted schemas.  For what
 it's
 worth, the metadata files describe coin-types, an intellectual concept in
 numismatics succinctly described at
 http://coins.about.com/od/**coinsglossary/g/coin_type.htmhttp://coins.about.com/od/coinsglossary/g/coin_type.htm,
 not physical
 objects in a collection.


 I believe this is similar to what FOAF does with primary topic:
 http://xmlns.com/foaf/spec/#**term_primaryTopichttp://xmlns.com/foaf/spec/#term_primaryTopic

 In FOAF that usually points to a web page ABOUT the subject of the FOAF
 data, so a wikipedia web page about Stephen King would get this primary
 topic property. Presuming that your XML is http:// accessible, it might
 fit into this model.

 kc


 Ethan

 On Mon, Feb 13, 2012 at 4:28 PM, Patrick Murray-John
 patrickmjc...@gmail.com  wrote:

  Ethan,

 The semantics do seem odd there. It doesn't seem like a skos:Concept
 would
 typically link to a metadata record about -- if I'm following you right
 --
 a specific coin. Is this sort of a FRBRish approach, where your
 skos:Concept is similar to the abstraction of a frbr:Work (that is, the
 idea of a particular coin), where your metadata records are really
 describing the common features of a particular coin?

 If that's close, it seems like the richer metadata is really a sort of
 definition of the skos:Concept, so maybe skos:definition would do the
 trick? Something like this:

 ex:wheatPenny a skos:Concept ;
skos:prefLabel Wheat Penny ;
skos:definition Your richer, non RDF metadata document describing the
 front and back, years minted, etc.

 In XML that might be like:

 skos:Concept 
 about=http://example.org/wheatPennyhttp://example.org/**wheatPenny
 http://example.org/**wheatPenny http://example.org/wheatPenny

 
  skos:prefLabelWheat Penny/skos:prefLabel
  skos:definition
 Your richer, non RDF metadata document describing the front and back,
 years minted, etc.
  /skos:definition
  /skos:Concept


 It might raise an eyebrow to have, instead of a literal value for
 skos:definition, another set of structured, non RDF metadata. Better in
 that case to go with a document reference, and make your richer metadata
 a
 standalone document with its own URI:

 ex:wheatPenny skos:definition ex:wheatPennyDefinition**.xml

 skos:Concept 
 about=http://example.org/wheatPennyhttp://example.org/**wheatPenny
 http://example.org/**wheatPenny http://example.org/wheatPenny
 
 skos:definition 
 resource=http://example.org/wheatPenny.xmlhttp://example.org/**wheatPenny.xml
 http://**example.org/wheatPenny.xml http://example.org/wheatPenny.xml
 

 /
 /skos:Concept

 I'm looking at the Documentation as a Document Reference section in SKOS
 Primer : 
 http://www.w3.org/TR/2009/NOTE-skos-primer-20090818/http://www.w3.org/TR/2009/**NOTE-skos-primer-20090818/
 htt**p://www.w3.org/TR/2009/NOTE-**skos-primer-20090818/http://www.w3.org/TR/2009/NOTE-skos-primer-20090818/
 


 Again, if I'm following, that might be the closest approach.

 Hope that helps,
 Patrick



 On 02/11/2012 09:53 PM, Ethan Gruber wrote:

  Hi Patrick,

 The richer metadata model is an ontology for describing coins.  It is
 more
 complex than, say, VRA Core or MODS, but not as hierarchically
 complicated
 as an EAD finding aid.  I'd like to link a skos:Concept to one of these
 related metadata records.  It doesn't matter if I use  skos, owl, etc.
 to
 describe this relationship, so long as it is a semantically appropriate
 choice.

 Ethan

 On Sat, Feb 11, 2012 at 2:32 PM, Patrick Murray-John
 patrickmjc...@gmail.com   wrote:

  Ethan,


 Maybe I'm being daft in missing it, but could I ask about more details
 in
 the richer metadata model? My hunch is that, depending on the details
 of
 the information you want to bring in, there might be more precise
 alternatives to what's in SKOS. Are you aiming to have a link between a
 skos:Concept and texts/documents related to that concept?

 Patrick


 On 02/11/2012 03:14 PM, Ethan Gruber wrote:

  Hi Ross,


 Thanks for the input.  My main objective is to make the richer
 metadata
 available one way or another to people using our web services.  Do you
 think it makes more sense to link to a URI of the richer metadata
 document
 as skos:related (or similar)?  I've seen two uses for
 skos:related--one
 to
 point to related skos:concepts, the other to point to web resources
 associated with that concept, e.g., a wikipedia article.  I have a
 feeling
 the latter is incorrect, at least 

Re: [CODE4LIB] CODE4LIB Digest - 12 Feb 2012 to 13 Feb 2012 (#2012-42)

2012-02-14 Thread David Talley
When I read Nate's response, I thought that the distinction is the endpoint 
of the process: The data is what the user goes looking for, the stuff that 
satisfies the desire that started their search. The metadata is the path to 
get there. Then I remembered the old example of a student consulting an 
author catalog to grab the person's birth  death dates for a school report 
rather than to find a work produced by that author. Then Joel added a whole 
new layer with that imagined hide  seek process built around the metadata 
(almost gamefication, really), and again the metadata becomes the 
destination not the path. 

Is it a useful distinction to say the data's the *reason* for collecting 
the metadata in the first place? Without the need to give access to that 
copy of _A Tale of Two Cities_, either in a physical library or Google 
Books, the descriptive metadata never would be created. I'd agree with Nate 
that it doesn't matter much to the computer's processing routines, but to 
make the computer serve its user, those goals are paramount.

Apologies if that's overly conceptual for a list with 'code' in the name. 

David

--

Date:Mon, 13 Feb 2012 16:39:14 -0600
From:Nate Vack njv...@wisc.edu
Subject: Re: Metadata

[. . . snip]I think it's kind of a circular issue: We know metadata and 
data are
separate because our software and workflow require it. Software and
workflows are designed to separate metadata and data because we know
they're separate.

--

Date:Mon, 13 Feb 2012 21:09:30 -0500
From:Richard, Joel M richar...@si.edu
Subject: Re: Metadata

[. . . snip]The contents of _A Tale of Two Cities_ can now be seen in so 
many different ways: a histogram of word frequency, a chart of which 
characters have the most dialogue, locations in the novel can be mapped 
geographically over the course of the story. (I only wish I had an 
interactive map when reading A Game of Thrones to tell me who was where at 
which part of the novel!)

And you can then search for books that take place in certain cities, or in 
a time period, or have people who wear beige top hats in victorian England. 
The possibilities are endless! [snip . . . ]


Re: [CODE4LIB] CODE4LIB Digest - 12 Feb 2012 to 13 Feb 2012 (#2012-42)

2012-02-14 Thread David Talley
From the article Tod helpfully links: One of our implementation goals was 
to build a touch interface that
appeared to be completely dedicated and self-contained: we did not want
it to be apparent to the user that the interface had been created with
and was being driven by commodity components. 

I'm stuck by the self-contained nature of this project design, and 
similarly with the iPad catalog look-up tools. Are such implementations 
most successful with separate, narrowly defined goals? Or would a library 
want to keep a consistent interaction experience across the website, 
kiosks, and even physical space (signage, displays, functional process 
terminology, etc.)? I tend to think that even if specific interaction 
methods are tailored to provide particular information in specific 
contexts, they all need to be designed as components of the user's overall 
interaction experience.

David

--

Date:Mon, 13 Feb 2012 09:55:09 -0600
From:Tod Olson t...@uchicago.edu
Subject: Re: Touch Screens in the Library

NCSU has done some work you might be interested in.  See this article:

Lessons in Public Touchscreen Development
by Andreas K. Orphanides

http://journal.code4lib.org/articles/5832

-Tod

Tod Olson t...@uchicago.edu
Systems Librarian 
University of Chicago Library


Re: [CODE4LIB] RDF advice

2012-02-14 Thread Ross Singer
It should actually be foaf:isPrimaryTopicOf (the inverse of foaf:primaryTopic).

It's fine to use it with your skos:Concept because the domain is an
owl:Thing (that is, any RDF resource) and the range is a foaf:Document
(which can be any document of any kind), again that's the advantage of
RDF.

That said, it's a little hard to figure out if that's an ideal
property for your use case.

It might make more sense to mint your own property and embed your nuds
document in there (as an XMLLiteral type), since this seems like it's
only going to be specialized usage, anyway.

-Ross.

On Tue, Feb 14, 2012 at 11:04 AM, Ethan Gruber ewg4x...@gmail.com wrote:
 Hi Karen,

 Thanks.  Would it be odd to use foaf:primaryTopic when FOAF isn't used to
 describe other attributes of a concept?

 Ethan

 On Mon, Feb 13, 2012 at 5:59 PM, Karen Coyle li...@kcoyle.net wrote:

 On 2/13/12 1:43 PM, Ethan Gruber wrote:

 Hi Patrick,

 Thanks.  That does make sense.  Hopefully others will weigh in with
 agreement (or disagreement).  Sometimes these semantic languages are so
 flexible that it's unsettling.  There are a million ways to do something
 with only de facto standards rather than restricted schemas.  For what
 it's
 worth, the metadata files describe coin-types, an intellectual concept in
 numismatics succinctly described at
 http://coins.about.com/od/**coinsglossary/g/coin_type.htmhttp://coins.about.com/od/coinsglossary/g/coin_type.htm,
 not physical
 objects in a collection.


 I believe this is similar to what FOAF does with primary topic:
 http://xmlns.com/foaf/spec/#**term_primaryTopichttp://xmlns.com/foaf/spec/#term_primaryTopic

 In FOAF that usually points to a web page ABOUT the subject of the FOAF
 data, so a wikipedia web page about Stephen King would get this primary
 topic property. Presuming that your XML is http:// accessible, it might
 fit into this model.

 kc


 Ethan

 On Mon, Feb 13, 2012 at 4:28 PM, Patrick Murray-John
 patrickmjc...@gmail.com  wrote:

  Ethan,

 The semantics do seem odd there. It doesn't seem like a skos:Concept
 would
 typically link to a metadata record about -- if I'm following you right
 --
 a specific coin. Is this sort of a FRBRish approach, where your
 skos:Concept is similar to the abstraction of a frbr:Work (that is, the
 idea of a particular coin), where your metadata records are really
 describing the common features of a particular coin?

 If that's close, it seems like the richer metadata is really a sort of
 definition of the skos:Concept, so maybe skos:definition would do the
 trick? Something like this:

 ex:wheatPenny a skos:Concept ;
    skos:prefLabel Wheat Penny ;
    skos:definition Your richer, non RDF metadata document describing the
 front and back, years minted, etc.

 In XML that might be like:

 skos:Concept 
 about=http://example.org/wheatPennyhttp://example.org/**wheatPenny
 http://example.org/**wheatPenny http://example.org/wheatPenny

 
  skos:prefLabelWheat Penny/skos:prefLabel
  skos:definition
 Your richer, non RDF metadata document describing the front and back,
 years minted, etc.
  /skos:definition
  /skos:Concept


 It might raise an eyebrow to have, instead of a literal value for
 skos:definition, another set of structured, non RDF metadata. Better in
 that case to go with a document reference, and make your richer metadata
 a
 standalone document with its own URI:

 ex:wheatPenny skos:definition ex:wheatPennyDefinition**.xml

 skos:Concept 
 about=http://example.org/wheatPennyhttp://example.org/**wheatPenny
 http://example.org/**wheatPenny http://example.org/wheatPenny
 
 skos:definition 
 resource=http://example.org/wheatPenny.xmlhttp://example.org/**wheatPenny.xml
 http://**example.org/wheatPenny.xml http://example.org/wheatPenny.xml
 

 /
 /skos:Concept

 I'm looking at the Documentation as a Document Reference section in SKOS
 Primer : 
 http://www.w3.org/TR/2009/NOTE-skos-primer-20090818/http://www.w3.org/TR/2009/**NOTE-skos-primer-20090818/
 htt**p://www.w3.org/TR/2009/NOTE-**skos-primer-20090818/http://www.w3.org/TR/2009/NOTE-skos-primer-20090818/
 


 Again, if I'm following, that might be the closest approach.

 Hope that helps,
 Patrick



 On 02/11/2012 09:53 PM, Ethan Gruber wrote:

  Hi Patrick,

 The richer metadata model is an ontology for describing coins.  It is
 more
 complex than, say, VRA Core or MODS, but not as hierarchically
 complicated
 as an EAD finding aid.  I'd like to link a skos:Concept to one of these
 related metadata records.  It doesn't matter if I use  skos, owl, etc.
 to
 describe this relationship, so long as it is a semantically appropriate
 choice.

 Ethan

 On Sat, Feb 11, 2012 at 2:32 PM, Patrick Murray-John
 patrickmjc...@gmail.com   wrote:

  Ethan,


 Maybe I'm being daft in missing it, but could I ask about more details
 in
 the richer metadata model? My hunch is that, depending on the details
 of
 the information you want to bring in, there might be more precise
 alternatives to what's 

Re: [CODE4LIB] Touch Screens in the Library

2012-02-14 Thread Cynthia Ng
Thanks everyone for the links and ideas. We'll be looking into
implementing the touch screen once we finish our wayfinding tool, so
it's great to see how others have done it.

Please post if you encounter any more.

Thanks!


[CODE4LIB] Job: Workflow Consultant ReCAP Discovery to Delivery Project Workflow Consultant ReCap Discovery to Delivery Project at New York Public Library

2012-02-14 Thread jobs4lib
NYPL seeks a Workflow Consultant to support the work of the Workflow and
Technology Subcommittee (WTS). The Workflow Consultant will
work closely with the Planning Consultant and the technology team but will
provide deliverables directly to the WTS.

The Workflow Consultant will be engaged to perform the following activities:

1.
Evaluate current workflows associated with transferring materials to and
retrieving materials from the ReCAP facility

The Workflow Consultant will document existing workflows at the partner
libraries necessary to perform these activities:

• Select and prepare materials for transfer
to ReCAP

• Request, receive, and return materials
needed from ReCAP

The Workflow Consultant will also document workflows at the ReCAP facility
necessary to perform these activities:

• Ingest and shelve new materials from each
partner (including use of inventory control system)

• Respond to request, retrieve, deliver,
return and reshelve requested materials.

Deliverables:

2.
Recommend appropriate modifications to these workflows and related technology
services necessary to support the shared
collectionDeliverables: An initial report to the WTS due by July
31, 2012 describing current workflows and the Workflow Consultant's
recommendations for workflow changes and system changes necessary to support
ReCAP shared collections. A final report due September 30, 2012
based on continued participation in planning discussions with the WTS and the
project team, to incorporate modifications based on subsequent decisions about
services, organization, or collection management.



For complete RFP please contact Joseph Brucia at josephbru...@nypl.org



Brought to you by code4lib jobs: http://jobs.code4lib.org/job/790/


[CODE4LIB] Job: Technology Consultant ReCap Discovery to Delivery Project at New York Public Library

2012-02-14 Thread jobs4lib
NYPL seeks a Technology Consultant to support the work of the Workflow and
Technology Subcommittee (WTS). The Technology Consultant
will work closely with the Technology Architect and the Planning Consultant
but will also provide deliverables directly to the WTS.

The Technology Consultant will be engaged to perform the following activities:

1.
Prepare a technology environmental scan:

a.
Identify current technologies available in the ReCAP partner libraries and the
ReCAP facility to support discovery, request and delivery, collections
management, and collections development

Describe the software systems currently in use to support these library
functions at each of the three ReCAP partners and at the ReCAP
facility. Include local and external
systems that are used by these institutions.

b.
In conjunction with the Technology Architect, identify options for software
systems to support those four required elements of the shared collection
(discovery, request and delivery, collections management, and collections
development).

Based on research and domain knowledge, describe software systems (and related
services) that could be used to support those functions in the context of a
shared collection. Provide a general description of their
functionality and their attributes e.g. propriety or open-source, perpetual
license or subscription, and/or other attributes as advised by the Technology
Consultant. Provide a preliminary assessment of the pros and cons of each
system for use by ReCAP shared collections. Note: this activity is intended to
describe potential technologies in general, not to recommend specific choices
at this time.

 A written report to the Technology Architect
and the WTS due by April 31, 2012.


Make recommendations to the Technology Architect and the WTS about the design
of systems infrastructure to support both user-facing and staff-facing
services at each member institution to support shared collections.

Working as part of a team including the Technology Architect, the Workflow
Consultant, and the Planning Consultant, incorporate information about new
services, revised workflows, and system options to produce recommendations
about the system infrastructure needed to support ReCAP shared
collections. Specific tasks include:

• Review findings and recommendations made
by the Workflow Consultant, who will be engaged to evaluate current workflows
associated with selection from the partner libraries for ingest into the ReCAP
facility, and to recommend appropriate modifications to these workflows and
related technology services necessary to support the shared
collection. The Workflow Consultant's initial report is
expected to be available in summer 2012.

• Consider findings by the Planning
Consultant regarding desired organizational structure, governance, and cost-
sharing (as available).

• Assess existing software products
applicable to shared collections (as identified in Activity #1) and recommend
appropriate solutions for the ReCAP partners' shared collections system
architecture.

  * Identify any needed systems or functionality not currently available in 
existing software products, with recommendations about how best to fill gaps.

• Assist the Technology Architect in
gathering cost information for implementing or developing the recommended
system solutions.

Deliverable: A report to the Technology Architect and the
WTS due by September 30, 2012 describing the Technology
Consultant's recommendations for system solutions to support ReCAP shared
collections. The report will include any recommendations
with respect to priorities, development, or implementation scheduling.



For the complete RFP please contact Joseph Brucia at josephbru...@nypl.org



Brought to you by code4lib jobs: http://jobs.code4lib.org/job/791/


Re: [CODE4LIB] Berkeley DB and NOID

2012-02-14 Thread Joshua Gomez
Thanks John!  I tried running db_upgrade, but apparently the utilities were
not included in my setup. So I ran the following:

gomez@gwnma:/var/www/nd/t1$ sudo apt-get install db4.8-util

That installed the utilities, but I'm still having some trouble:

gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v -h NOID NOID/noid.bdb
db4.8_upgrade: Program version 4.8 doesn't match environment version 4.7
db4.8_upgrade: DB_ENV-open: DB_VERSION_MISMATCH: Database environment
version mismatch

I tried leaving off the environment home flag

gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v NOID/noid.bdb
db4.8_upgrade: NOID/noid.bdb upgraded successfully

That looks like it worked. However, when I try making a call to the service
I get the old error:

gomez@gwnma:/var/www/nd/t1$ curl localhost/nd/noidu_t1?mint+1
no Env object (DB_VERSION_MISMATCH: Database environment version mismatch)

-Josh


On Mon, Feb 13, 2012 at 5:42 PM, John A. Kunze j...@ucop.edu wrote:

 The standard BerkeleyDB library probably changed when you upgraded
 Ubuntu, and it complains that the NOID database (written with the old
 library) is incompatible.

 You should be able to use db_upgrade to convert the NOID database
 (NOID/noid.bdb).  db_upgrade is a command line utility that comes with
 BerkeleyDB.

 -John


 --- On Mon, 13 Feb 2012, Joshua Gomez wrote:

 Does anyone here have expertise with Berkeley DB?

 I was running an instance of NOID (which uses Berkeley DB) to mint and
 resolve ARKs.  I updated the OS for the server it was running on from
 Ubuntu 9 to Ubuntu 10.  Now NOID has stopped working and complains that
 the
 db version doesn't match: Program version 4.8 doesn't match environment
 version 4.7

 I have no experience at all with Berkeley DB and could use some advice.

 Thanks,
 Josh

 --
 Joshua Gomez
 Digital Library Programmer Analyst
 George Washington University Libraries
 2130 H St, NW Washington, DC 20052
 (202) 994-8267




-- 
Joshua Gomez
Digital Library Programmer Analyst
George Washington University Libraries
2130 H St, NW Washington, DC 20052
(202) 994-8267


Re: [CODE4LIB] Berkeley DB and NOID

2012-02-14 Thread John A. Kunze

Try removing the environment files before you upgrade (or after if
you didn't save the old file).  When I released that version of NOID
I have since regretted leaving in code that created a stub environment
(that actually isn't used by NOID) because it just creates upgrade
problems that I never figured out properly.

If all else fails and you're just using NOID to mint (eg, no binding),
it's quite easy to start a new minter from scratch and mint until you see
the last id you minted prior to conversion.  The order of minting is
deterministic, so in the end you'll have a new minter that's in the same
state as the old minter (again, provided you haven't been doing holds
and binds -- that's more complicated).

-John

--- On Tue, 14 Feb 2012, Joshua Gomez wrote:

Thanks John!  I tried running db_upgrade, but apparently the utilities were
not included in my setup. So I ran the following:

gomez@gwnma:/var/www/nd/t1$ sudo apt-get install db4.8-util

That installed the utilities, but I'm still having some trouble:

gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v -h NOID NOID/noid.bdb
db4.8_upgrade: Program version 4.8 doesn't match environment version 4.7
db4.8_upgrade: DB_ENV-open: DB_VERSION_MISMATCH: Database environment
version mismatch

I tried leaving off the environment home flag

gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v NOID/noid.bdb
db4.8_upgrade: NOID/noid.bdb upgraded successfully

That looks like it worked. However, when I try making a call to the service
I get the old error:

gomez@gwnma:/var/www/nd/t1$ curl localhost/nd/noidu_t1?mint+1
no Env object (DB_VERSION_MISMATCH: Database environment version mismatch)

-Josh


On Mon, Feb 13, 2012 at 5:42 PM, John A. Kunze j...@ucop.edu wrote:


The standard BerkeleyDB library probably changed when you upgraded
Ubuntu, and it complains that the NOID database (written with the old
library) is incompatible.

You should be able to use db_upgrade to convert the NOID database
(NOID/noid.bdb).  db_upgrade is a command line utility that comes with
BerkeleyDB.

-John


--- On Mon, 13 Feb 2012, Joshua Gomez wrote:


Does anyone here have expertise with Berkeley DB?

I was running an instance of NOID (which uses Berkeley DB) to mint and
resolve ARKs.  I updated the OS for the server it was running on from
Ubuntu 9 to Ubuntu 10.  Now NOID has stopped working and complains that
the
db version doesn't match: Program version 4.8 doesn't match environment
version 4.7

I have no experience at all with Berkeley DB and could use some advice.

Thanks,
Josh

--
Joshua Gomez
Digital Library Programmer Analyst
George Washington University Libraries
2130 H St, NW Washington, DC 20052
(202) 994-8267





--
Joshua Gomez
Digital Library Programmer Analyst
George Washington University Libraries
2130 H St, NW Washington, DC 20052
(202) 994-8267



[CODE4LIB] Job Posting: Data Processing Automation Programmer, University of Michigan, Digital Library Production Service (DLPS)

2012-02-14 Thread John Weise
http://umjobs.org/job_detail/66448/data_processing_automation_programmerUniversity
of Michigan, Digital Library Production Service (DLPS)Data Processing
Automation ProgrammerJob Summary

The Library Information Technology (LIT) division provides comprehensive
technology support and guidance for the University of Michigan Library
system, including hosting digital library collections, coordinating
electronic publishing initiatives, and supporting traditional library
services (circulation of materials and management of metadata).
The Digital Library Production Service (DLPS), a part of the LIT, is one of
the nation's premier organizations for the creation and support of digital
library resources and infrastructure, with production level support for
electronic library collections. Staff are responsible for creating online
access mechanisms, significant digital conversion activities, and they play
a role in the University of Michigan Library's work on the HathiTrust
digital library. Additionally, DLPS staff support large image and finding
aid collections, and contribute to the Library's development efforts with
many other formats.

This is a 3-year term position with the possibility for renewal.

This position has a target salary range of $40,000-$60,000 annually,
dependent on qualifications and experience of the selected candidate.
Responsibilities*

DLPS is looking for a talented, resourceful programmer to develop,
maintain, document, and monitor software systems. Primary focus will be
placed on developing highly reliable software tools for routine data
processing on a large scale. Specific processing tasks include file format
conversion, optical character recognition (OCR), metadata insertion,
transformation, validation, and transfer. Work includes assessing needs and
specifying software requirements. Development of web interfaces for process
management may be needed as well. Other tasks will vary but include, for
example, preparing documentation and the development of digital library
access systems.
Required Qualifications*

Bachelor's degree in computer science or an equivalent combination of
education and experience.
Facility with Linux or similar *nix operating systems.
Demonstrated experience programming with Perl, or, experience with Ruby on
Rails, Python, PHP plus willingness to learn and use Perl.
Strong analytical and troubleshooting skills.
Excellent verbal and written communication skills.
Ability to creatively improve workflows and processes.
Desired Qualifications*

Demonstrated experience building dynamic web interfaces with HTML, CSS, and
JavaScript.
Experience using version control systems in software development.
Demonstrated experience with SGML/XML and related technologies and
standards.
Demonstrated experience with database technology such as MySQL, including
database design and implementation.
Demonstrated experience developing APIs such as RESTful web services.
Familiarity with batch image processing techniques in applications such as
Photoshop and on the command line, ImageMagick.
U-M EEO/AA Statement

The University of Michigan is an equal opportunity/affirmative action
employer.


Re: [CODE4LIB] Berkeley DB and NOID

2012-02-14 Thread Joshua Gomez
That didn't work either. I started with a fresh copy of the NOID directory
from before I tried the upgrade command:

gomez@gwnma:/var/www/nd/t1$ ls NOID
__db.001  __db.003  __db.005  lock  log.01  noid.bdb
__db.002  __db.004  __db.006  log   logbdb  README

gomez@gwnma:/var/www/nd/t1$ sudo rm NOID/__*

gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v -h NOID noid.bdb
db4.8_upgrade: noid.bdb upgraded successfully

gomez@gwnma:/var/www/nd/t1$ curl localhost/nd/noidu_t1?mint+1
no Env object (No such file or directory)

I have been binding the IDs for the production NOID db (t1 is just the test
NOID), but I have also been storing them in a mysql DB on our dspace server
that I use to keep track of where things are in the workflow. So I suppose
I could try what you suggested: setup a new production NOID, mint as many
IDs as we have used so far, then rebind them all using the data in the
mysql DB. Redundant data is a good thing.

-Joshua

On Tue, Feb 14, 2012 at 1:07 PM, John A. Kunze j...@ucop.edu wrote:

 Try removing the environment files before you upgrade (or after if
 you didn't save the old file).  When I released that version of NOID
 I have since regretted leaving in code that created a stub environment
 (that actually isn't used by NOID) because it just creates upgrade
 problems that I never figured out properly.

 If all else fails and you're just using NOID to mint (eg, no binding),
 it's quite easy to start a new minter from scratch and mint until you see
 the last id you minted prior to conversion.  The order of minting is
 deterministic, so in the end you'll have a new minter that's in the same
 state as the old minter (again, provided you haven't been doing holds
 and binds -- that's more complicated).

 -John


 --- On Tue, 14 Feb 2012, Joshua Gomez wrote:

 Thanks John!  I tried running db_upgrade, but apparently the utilities
 were
 not included in my setup. So I ran the following:

 gomez@gwnma:/var/www/nd/t1$ sudo apt-get install db4.8-util

 That installed the utilities, but I'm still having some trouble:

 gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v -h NOID NOID/noid.bdb
 db4.8_upgrade: Program version 4.8 doesn't match environment version 4.7
 db4.8_upgrade: DB_ENV-open: DB_VERSION_MISMATCH: Database environment
 version mismatch

 I tried leaving off the environment home flag

 gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v NOID/noid.bdb
 db4.8_upgrade: NOID/noid.bdb upgraded successfully

 That looks like it worked. However, when I try making a call to the
 service
 I get the old error:

 gomez@gwnma:/var/www/nd/t1$ curl localhost/nd/noidu_t1?mint+1
 no Env object (DB_VERSION_MISMATCH: Database environment version
 mismatch)

 -Josh


 On Mon, Feb 13, 2012 at 5:42 PM, John A. Kunze j...@ucop.edu wrote:

  The standard BerkeleyDB library probably changed when you upgraded
 Ubuntu, and it complains that the NOID database (written with the old
 library) is incompatible.

 You should be able to use db_upgrade to convert the NOID database
 (NOID/noid.bdb).  db_upgrade is a command line utility that comes with
 BerkeleyDB.

 -John


 --- On Mon, 13 Feb 2012, Joshua Gomez wrote:

  Does anyone here have expertise with Berkeley DB?

 I was running an instance of NOID (which uses Berkeley DB) to mint and
 resolve ARKs.  I updated the OS for the server it was running on from
 Ubuntu 9 to Ubuntu 10.  Now NOID has stopped working and complains that
 the
 db version doesn't match: Program version 4.8 doesn't match environment
 version 4.7

 I have no experience at all with Berkeley DB and could use some advice.

 Thanks,
 Josh

 --
 Joshua Gomez
 Digital Library Programmer Analyst
 George Washington University Libraries
 2130 H St, NW Washington, DC 20052
 (202) 994-8267




 --
 Joshua Gomez
 Digital Library Programmer Analyst
 George Washington University Libraries
 2130 H St, NW Washington, DC 20052
 (202) 994-8267




-- 
Joshua Gomez
Digital Library Programmer Analyst
George Washington University Libraries
2130 H St, NW Washington, DC 20052
(202) 994-8267


Re: [CODE4LIB] Berkeley DB and NOID

2012-02-14 Thread John A. Kunze

Sounds good.  At this distance and given my focus on a new NOID release
(that separates minting from binding in separate databases and simplifies
installation by relying only on the standard DB_File), I don't think I
can help you devise a better strategy than the one you've come up with.

-John

--- On Tue, 14 Feb 2012, Joshua Gomez wrote:

That didn't work either. I started with a fresh copy of the NOID directory
from before I tried the upgrade command:

gomez@gwnma:/var/www/nd/t1$ ls NOID
__db.001  __db.003  __db.005  lock  log.01  noid.bdb
__db.002  __db.004  __db.006  log   logbdb  README

gomez@gwnma:/var/www/nd/t1$ sudo rm NOID/__*

gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v -h NOID noid.bdb
db4.8_upgrade: noid.bdb upgraded successfully

gomez@gwnma:/var/www/nd/t1$ curl localhost/nd/noidu_t1?mint+1
no Env object (No such file or directory)

I have been binding the IDs for the production NOID db (t1 is just the test
NOID), but I have also been storing them in a mysql DB on our dspace server
that I use to keep track of where things are in the workflow. So I suppose
I could try what you suggested: setup a new production NOID, mint as many
IDs as we have used so far, then rebind them all using the data in the
mysql DB. Redundant data is a good thing.

-Joshua

On Tue, Feb 14, 2012 at 1:07 PM, John A. Kunze j...@ucop.edu wrote:


Try removing the environment files before you upgrade (or after if
you didn't save the old file).  When I released that version of NOID
I have since regretted leaving in code that created a stub environment
(that actually isn't used by NOID) because it just creates upgrade
problems that I never figured out properly.

If all else fails and you're just using NOID to mint (eg, no binding),
it's quite easy to start a new minter from scratch and mint until you see
the last id you minted prior to conversion.  The order of minting is
deterministic, so in the end you'll have a new minter that's in the same
state as the old minter (again, provided you haven't been doing holds
and binds -- that's more complicated).

-John


--- On Tue, 14 Feb 2012, Joshua Gomez wrote:


Thanks John!  I tried running db_upgrade, but apparently the utilities
were
not included in my setup. So I ran the following:

gomez@gwnma:/var/www/nd/t1$ sudo apt-get install db4.8-util

That installed the utilities, but I'm still having some trouble:

gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v -h NOID NOID/noid.bdb
db4.8_upgrade: Program version 4.8 doesn't match environment version 4.7
db4.8_upgrade: DB_ENV-open: DB_VERSION_MISMATCH: Database environment
version mismatch

I tried leaving off the environment home flag

gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v NOID/noid.bdb
db4.8_upgrade: NOID/noid.bdb upgraded successfully

That looks like it worked. However, when I try making a call to the
service
I get the old error:

gomez@gwnma:/var/www/nd/t1$ curl localhost/nd/noidu_t1?mint+1
no Env object (DB_VERSION_MISMATCH: Database environment version
mismatch)

-Josh


On Mon, Feb 13, 2012 at 5:42 PM, John A. Kunze j...@ucop.edu wrote:

 The standard BerkeleyDB library probably changed when you upgraded

Ubuntu, and it complains that the NOID database (written with the old
library) is incompatible.

You should be able to use db_upgrade to convert the NOID database
(NOID/noid.bdb).  db_upgrade is a command line utility that comes with
BerkeleyDB.

-John


--- On Mon, 13 Feb 2012, Joshua Gomez wrote:

 Does anyone here have expertise with Berkeley DB?


I was running an instance of NOID (which uses Berkeley DB) to mint and
resolve ARKs.  I updated the OS for the server it was running on from
Ubuntu 9 to Ubuntu 10.  Now NOID has stopped working and complains that
the
db version doesn't match: Program version 4.8 doesn't match environment
version 4.7

I have no experience at all with Berkeley DB and could use some advice.

Thanks,
Josh

--
Joshua Gomez
Digital Library Programmer Analyst
George Washington University Libraries
2130 H St, NW Washington, DC 20052
(202) 994-8267





--
Joshua Gomez
Digital Library Programmer Analyst
George Washington University Libraries
2130 H St, NW Washington, DC 20052
(202) 994-8267





--
Joshua Gomez
Digital Library Programmer Analyst
George Washington University Libraries
2130 H St, NW Washington, DC 20052
(202) 994-8267



[CODE4LIB] Lift the Flap books

2012-02-14 Thread Sara Amato
If you were to have a 'lift the flap' type book that you wanted to digitize, 
for web display and use, what technology would you use for markup and display?

Visually I like the Internet Archive BookReader ( 
http://openlibrary.org/dev/docs/bookreader ), which says it can do 'foldouts', 
thought I haven't found an example of HOW to do that ... nor exactly what the 
metadata schema is.  


Re: [CODE4LIB] Lift the Flap books

2012-02-14 Thread stuart yeates

On 15/02/12 13:43, Sara Amato wrote:

If you were to have a 'lift the flap' type book that you wanted to digitize, 
for web display and use, what technology would you use for markup and display?

Visually I like the Internet Archive BookReader ( 
http://openlibrary.org/dev/docs/bookreader ), which says it can do 'foldouts', 
thought I haven't found an example of HOW to do that ... nor exactly what the 
metadata schema is.


Sounds like an ideal use for HTML, javascript and image transparency 
using OnMouseOver as trigger.


cheers
stuart

--
Stuart Yeates
Library Technology Services http://www.victoria.ac.nz/library/


Re: [CODE4LIB] Lift the Flap books

2012-02-14 Thread Avram Lyon
On Tue, Feb 14, 2012 at 4:43 PM, Sara Amato sam...@willamette.edu wrote:
 If you were to have a 'lift the flap' type book that you wanted to digitize, 
 for web display and use, what technology would you use for markup and display?

I don't have an answer, but Whitney Trettien at Duke has written about
this and other limitations of mainstream digitization from the
theoretical perspective:
http://blog.whitneyannetrettien.com/2011/05/digitizing-dead-end-branches.html
(and other posts in the blog)

- Avram

UCLA Slavic