Re: [CODE4LIB] Metadata
Reminds me of this article, which frighteningly is now eight years old: http://www.miketaylor.org.uk/tech/metadata.html -- Mike. On 14 February 2012 06:25, Kåre Fiedler Christiansen k...@statsbiblioteket.dk wrote: You realize, of course, that discussing the use of the word metametadata could be described as metametametadata? Which would make my post metametametametadata. At which point it all turns into silliness (which it certainly wasn't before... right? :-) ). Best, Kåre From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Kurt Nordstrom I got such dirty looks when I used the term metametadata to describe something. ;) -Kurt On 02/13/2012 02:39 PM, Becky Yoose wrote: Could this conversation be described as metametadata? *runs, hides* Thanks, Becky
Re: [CODE4LIB] Metadata
The actual data (the novel) is not in the catalog (which is composed only of metadata). That's a technical limitation. It's also a legal/commercial limitation, as well as a question of provenance. To summarise a lot of good points made already: An item of metadata is a relationship that someone claims to exist between two entities. - source: http://www.doi.org/topics/indecs/indecs_framework_2000.pdf Cheers, M -Original Message- From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Nate Vack Sent: 13 February 2012 22:39 To: CODE4LIB@LISTSERV.ND.EDU Subject: Re: [CODE4LIB] Metadata On Mon, Feb 13, 2012 at 4:25 PM, Genny Engel gen...@sonoma.lib.ca.us wrote: You simply can't use the average library catalog to look up Author X's novel that starts with the sentence So a string walks into a bar. The actual data (the novel) is not in the catalog (which is composed only of metadata). That's a technical limitation. If you're Google Books (or any other fulltext index), the actual data *is* in the catalog, and data and metadata are again functionally identical. The best working definition of metadata I've come up with is something I have a field for in my data cataloging program. I think it's kind of a circular issue: We know metadata and data are separate because our software and workflow require it. Software and workflows are designed to separate metadata and data because we know they're separate. -n
Re: [CODE4LIB] Metadata
On Tue, Feb 14, 2012 at 1:22 AM, Graham Triggs grahamtri...@gmail.com wrote: That's an interesting distinction though. Do you need all that data in order to make sense of the results? You don't [necessarily] need to know who conducted some research, or when they conducted it in order to analyse and make sense of the data. In the context of having the data, this other information becomes irrelevant in terms of understanding what that data says. It is *essential* to understanding what the data says. Perhaps you find out your sensor was on the fritz during a time period -- you need to be able to know what datasets are suspect. Maybe the blood pressure effect you're looking at is mediated by circadian rhythms, and hence, times of day. Not all of your data is necessary in every analysis, but a bunch of blood pressure measurements in the absence of contextual information is universally useless. The metadata is part of the data. -n
Re: [CODE4LIB] Metadata
Having done research, and now working in a very varied metadata role, I don't quite understand this discussion about data that is or isn't metadata. Scientific data is a great example of structured data, but it's not impossible to distinguish it from metadata purely describing a dataset. However, if you have scientific research data created during the experiments, even if it's operational, it's clearly part of the data. This doesn't mean there can't be metadata describing *that data*. Just because it's not glamorous data doesn't mean it's not essential to the scientific process. Similarly, just being about mundane or procedural things doesn't make data into metadata...! You're absolutely right, the contextual information is certainly part of the experimental outcome in this example; otherwise it would be abstract data such as one might use in a textbook example. Metadata would describe the dataset itself, not the scientific research. There's always a certain ambiguity involved in identifying the data as distinct from the metadata, and it's a false dichotomy to suggest metadata is not useful at all for the domain expert. It's contextual, and the definition is always at least partly based on your use case for the data and its description. -Original Message- From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Nate Vack Sent: 14 February 2012 14:45 To: CODE4LIB@LISTSERV.ND.EDU Subject: Re: [CODE4LIB] Metadata On Tue, Feb 14, 2012 at 1:22 AM, Graham Triggs grahamtri...@gmail.com wrote: That's an interesting distinction though. Do you need all that data in order to make sense of the results? You don't [necessarily] need to know who conducted some research, or when they conducted it in order to analyse and make sense of the data. In the context of having the data, this other information becomes irrelevant in terms of understanding what that data says. It is *essential* to understanding what the data says. Perhaps you find out your sensor was on the fritz during a time period -- you need to be able to know what datasets are suspect. Maybe the blood pressure effect you're looking at is mediated by circadian rhythms, and hence, times of day. Not all of your data is necessary in every analysis, but a bunch of blood pressure measurements in the absence of contextual information is universally useless. The metadata is part of the data. -n
[CODE4LIB] New England Code4Lib Regional Conference
Hello - The planning process has begun for a New England regional gathering and we would like to try and get some potential attendance numbers together. We put together a brief poll which should only take a minute or less to complete: https://yalesurvey.qualtrics.com/SE/?SID=SV_24OogxklhikrpzK. Dates: Friday, October 26 and Saturday, October 27 Location: Yale University, New Haven, CT This will be a great opportunity to meet your peers at local institutions and converse on code4lib related topics! You could plan to come for a day at a low cost day trip or stay for 2 days and maximize your interaction time. We hope to see you there! This email is sent on behalf of the NECode4lib planning committee. Matthew Beacom Michael Friscia Cindy Greenspun Michelle Hudson James Luker Mark Matienzo Joe Montibello Tito Sierra Kalee Sprague Randy Stern Cindy Greenspun IT and Project Manager Database Administrator cindy.greens...@yale.edu
Re: [CODE4LIB] RDF advice
Hi Karen, Thanks. Would it be odd to use foaf:primaryTopic when FOAF isn't used to describe other attributes of a concept? Ethan On Mon, Feb 13, 2012 at 5:59 PM, Karen Coyle li...@kcoyle.net wrote: On 2/13/12 1:43 PM, Ethan Gruber wrote: Hi Patrick, Thanks. That does make sense. Hopefully others will weigh in with agreement (or disagreement). Sometimes these semantic languages are so flexible that it's unsettling. There are a million ways to do something with only de facto standards rather than restricted schemas. For what it's worth, the metadata files describe coin-types, an intellectual concept in numismatics succinctly described at http://coins.about.com/od/**coinsglossary/g/coin_type.htmhttp://coins.about.com/od/coinsglossary/g/coin_type.htm, not physical objects in a collection. I believe this is similar to what FOAF does with primary topic: http://xmlns.com/foaf/spec/#**term_primaryTopichttp://xmlns.com/foaf/spec/#term_primaryTopic In FOAF that usually points to a web page ABOUT the subject of the FOAF data, so a wikipedia web page about Stephen King would get this primary topic property. Presuming that your XML is http:// accessible, it might fit into this model. kc Ethan On Mon, Feb 13, 2012 at 4:28 PM, Patrick Murray-John patrickmjc...@gmail.com wrote: Ethan, The semantics do seem odd there. It doesn't seem like a skos:Concept would typically link to a metadata record about -- if I'm following you right -- a specific coin. Is this sort of a FRBRish approach, where your skos:Concept is similar to the abstraction of a frbr:Work (that is, the idea of a particular coin), where your metadata records are really describing the common features of a particular coin? If that's close, it seems like the richer metadata is really a sort of definition of the skos:Concept, so maybe skos:definition would do the trick? Something like this: ex:wheatPenny a skos:Concept ; skos:prefLabel Wheat Penny ; skos:definition Your richer, non RDF metadata document describing the front and back, years minted, etc. In XML that might be like: skos:Concept about=http://example.org/wheatPennyhttp://example.org/**wheatPenny http://example.org/**wheatPenny http://example.org/wheatPenny skos:prefLabelWheat Penny/skos:prefLabel skos:definition Your richer, non RDF metadata document describing the front and back, years minted, etc. /skos:definition /skos:Concept It might raise an eyebrow to have, instead of a literal value for skos:definition, another set of structured, non RDF metadata. Better in that case to go with a document reference, and make your richer metadata a standalone document with its own URI: ex:wheatPenny skos:definition ex:wheatPennyDefinition**.xml skos:Concept about=http://example.org/wheatPennyhttp://example.org/**wheatPenny http://example.org/**wheatPenny http://example.org/wheatPenny skos:definition resource=http://example.org/wheatPenny.xmlhttp://example.org/**wheatPenny.xml http://**example.org/wheatPenny.xml http://example.org/wheatPenny.xml / /skos:Concept I'm looking at the Documentation as a Document Reference section in SKOS Primer : http://www.w3.org/TR/2009/NOTE-skos-primer-20090818/http://www.w3.org/TR/2009/**NOTE-skos-primer-20090818/ htt**p://www.w3.org/TR/2009/NOTE-**skos-primer-20090818/http://www.w3.org/TR/2009/NOTE-skos-primer-20090818/ Again, if I'm following, that might be the closest approach. Hope that helps, Patrick On 02/11/2012 09:53 PM, Ethan Gruber wrote: Hi Patrick, The richer metadata model is an ontology for describing coins. It is more complex than, say, VRA Core or MODS, but not as hierarchically complicated as an EAD finding aid. I'd like to link a skos:Concept to one of these related metadata records. It doesn't matter if I use skos, owl, etc. to describe this relationship, so long as it is a semantically appropriate choice. Ethan On Sat, Feb 11, 2012 at 2:32 PM, Patrick Murray-John patrickmjc...@gmail.com wrote: Ethan, Maybe I'm being daft in missing it, but could I ask about more details in the richer metadata model? My hunch is that, depending on the details of the information you want to bring in, there might be more precise alternatives to what's in SKOS. Are you aiming to have a link between a skos:Concept and texts/documents related to that concept? Patrick On 02/11/2012 03:14 PM, Ethan Gruber wrote: Hi Ross, Thanks for the input. My main objective is to make the richer metadata available one way or another to people using our web services. Do you think it makes more sense to link to a URI of the richer metadata document as skos:related (or similar)? I've seen two uses for skos:related--one to point to related skos:concepts, the other to point to web resources associated with that concept, e.g., a wikipedia article. I have a feeling the latter is incorrect, at least
Re: [CODE4LIB] CODE4LIB Digest - 12 Feb 2012 to 13 Feb 2012 (#2012-42)
When I read Nate's response, I thought that the distinction is the endpoint of the process: The data is what the user goes looking for, the stuff that satisfies the desire that started their search. The metadata is the path to get there. Then I remembered the old example of a student consulting an author catalog to grab the person's birth death dates for a school report rather than to find a work produced by that author. Then Joel added a whole new layer with that imagined hide seek process built around the metadata (almost gamefication, really), and again the metadata becomes the destination not the path. Is it a useful distinction to say the data's the *reason* for collecting the metadata in the first place? Without the need to give access to that copy of _A Tale of Two Cities_, either in a physical library or Google Books, the descriptive metadata never would be created. I'd agree with Nate that it doesn't matter much to the computer's processing routines, but to make the computer serve its user, those goals are paramount. Apologies if that's overly conceptual for a list with 'code' in the name. David -- Date:Mon, 13 Feb 2012 16:39:14 -0600 From:Nate Vack njv...@wisc.edu Subject: Re: Metadata [. . . snip]I think it's kind of a circular issue: We know metadata and data are separate because our software and workflow require it. Software and workflows are designed to separate metadata and data because we know they're separate. -- Date:Mon, 13 Feb 2012 21:09:30 -0500 From:Richard, Joel M richar...@si.edu Subject: Re: Metadata [. . . snip]The contents of _A Tale of Two Cities_ can now be seen in so many different ways: a histogram of word frequency, a chart of which characters have the most dialogue, locations in the novel can be mapped geographically over the course of the story. (I only wish I had an interactive map when reading A Game of Thrones to tell me who was where at which part of the novel!) And you can then search for books that take place in certain cities, or in a time period, or have people who wear beige top hats in victorian England. The possibilities are endless! [snip . . . ]
Re: [CODE4LIB] CODE4LIB Digest - 12 Feb 2012 to 13 Feb 2012 (#2012-42)
From the article Tod helpfully links: One of our implementation goals was to build a touch interface that appeared to be completely dedicated and self-contained: we did not want it to be apparent to the user that the interface had been created with and was being driven by commodity components. I'm stuck by the self-contained nature of this project design, and similarly with the iPad catalog look-up tools. Are such implementations most successful with separate, narrowly defined goals? Or would a library want to keep a consistent interaction experience across the website, kiosks, and even physical space (signage, displays, functional process terminology, etc.)? I tend to think that even if specific interaction methods are tailored to provide particular information in specific contexts, they all need to be designed as components of the user's overall interaction experience. David -- Date:Mon, 13 Feb 2012 09:55:09 -0600 From:Tod Olson t...@uchicago.edu Subject: Re: Touch Screens in the Library NCSU has done some work you might be interested in. See this article: Lessons in Public Touchscreen Development by Andreas K. Orphanides http://journal.code4lib.org/articles/5832 -Tod Tod Olson t...@uchicago.edu Systems Librarian University of Chicago Library
Re: [CODE4LIB] RDF advice
It should actually be foaf:isPrimaryTopicOf (the inverse of foaf:primaryTopic). It's fine to use it with your skos:Concept because the domain is an owl:Thing (that is, any RDF resource) and the range is a foaf:Document (which can be any document of any kind), again that's the advantage of RDF. That said, it's a little hard to figure out if that's an ideal property for your use case. It might make more sense to mint your own property and embed your nuds document in there (as an XMLLiteral type), since this seems like it's only going to be specialized usage, anyway. -Ross. On Tue, Feb 14, 2012 at 11:04 AM, Ethan Gruber ewg4x...@gmail.com wrote: Hi Karen, Thanks. Would it be odd to use foaf:primaryTopic when FOAF isn't used to describe other attributes of a concept? Ethan On Mon, Feb 13, 2012 at 5:59 PM, Karen Coyle li...@kcoyle.net wrote: On 2/13/12 1:43 PM, Ethan Gruber wrote: Hi Patrick, Thanks. That does make sense. Hopefully others will weigh in with agreement (or disagreement). Sometimes these semantic languages are so flexible that it's unsettling. There are a million ways to do something with only de facto standards rather than restricted schemas. For what it's worth, the metadata files describe coin-types, an intellectual concept in numismatics succinctly described at http://coins.about.com/od/**coinsglossary/g/coin_type.htmhttp://coins.about.com/od/coinsglossary/g/coin_type.htm, not physical objects in a collection. I believe this is similar to what FOAF does with primary topic: http://xmlns.com/foaf/spec/#**term_primaryTopichttp://xmlns.com/foaf/spec/#term_primaryTopic In FOAF that usually points to a web page ABOUT the subject of the FOAF data, so a wikipedia web page about Stephen King would get this primary topic property. Presuming that your XML is http:// accessible, it might fit into this model. kc Ethan On Mon, Feb 13, 2012 at 4:28 PM, Patrick Murray-John patrickmjc...@gmail.com wrote: Ethan, The semantics do seem odd there. It doesn't seem like a skos:Concept would typically link to a metadata record about -- if I'm following you right -- a specific coin. Is this sort of a FRBRish approach, where your skos:Concept is similar to the abstraction of a frbr:Work (that is, the idea of a particular coin), where your metadata records are really describing the common features of a particular coin? If that's close, it seems like the richer metadata is really a sort of definition of the skos:Concept, so maybe skos:definition would do the trick? Something like this: ex:wheatPenny a skos:Concept ; skos:prefLabel Wheat Penny ; skos:definition Your richer, non RDF metadata document describing the front and back, years minted, etc. In XML that might be like: skos:Concept about=http://example.org/wheatPennyhttp://example.org/**wheatPenny http://example.org/**wheatPenny http://example.org/wheatPenny skos:prefLabelWheat Penny/skos:prefLabel skos:definition Your richer, non RDF metadata document describing the front and back, years minted, etc. /skos:definition /skos:Concept It might raise an eyebrow to have, instead of a literal value for skos:definition, another set of structured, non RDF metadata. Better in that case to go with a document reference, and make your richer metadata a standalone document with its own URI: ex:wheatPenny skos:definition ex:wheatPennyDefinition**.xml skos:Concept about=http://example.org/wheatPennyhttp://example.org/**wheatPenny http://example.org/**wheatPenny http://example.org/wheatPenny skos:definition resource=http://example.org/wheatPenny.xmlhttp://example.org/**wheatPenny.xml http://**example.org/wheatPenny.xml http://example.org/wheatPenny.xml / /skos:Concept I'm looking at the Documentation as a Document Reference section in SKOS Primer : http://www.w3.org/TR/2009/NOTE-skos-primer-20090818/http://www.w3.org/TR/2009/**NOTE-skos-primer-20090818/ htt**p://www.w3.org/TR/2009/NOTE-**skos-primer-20090818/http://www.w3.org/TR/2009/NOTE-skos-primer-20090818/ Again, if I'm following, that might be the closest approach. Hope that helps, Patrick On 02/11/2012 09:53 PM, Ethan Gruber wrote: Hi Patrick, The richer metadata model is an ontology for describing coins. It is more complex than, say, VRA Core or MODS, but not as hierarchically complicated as an EAD finding aid. I'd like to link a skos:Concept to one of these related metadata records. It doesn't matter if I use skos, owl, etc. to describe this relationship, so long as it is a semantically appropriate choice. Ethan On Sat, Feb 11, 2012 at 2:32 PM, Patrick Murray-John patrickmjc...@gmail.com wrote: Ethan, Maybe I'm being daft in missing it, but could I ask about more details in the richer metadata model? My hunch is that, depending on the details of the information you want to bring in, there might be more precise alternatives to what's
Re: [CODE4LIB] Touch Screens in the Library
Thanks everyone for the links and ideas. We'll be looking into implementing the touch screen once we finish our wayfinding tool, so it's great to see how others have done it. Please post if you encounter any more. Thanks!
[CODE4LIB] Job: Workflow Consultant ReCAP Discovery to Delivery Project Workflow Consultant ReCap Discovery to Delivery Project at New York Public Library
NYPL seeks a Workflow Consultant to support the work of the Workflow and Technology Subcommittee (WTS). The Workflow Consultant will work closely with the Planning Consultant and the technology team but will provide deliverables directly to the WTS. The Workflow Consultant will be engaged to perform the following activities: 1. Evaluate current workflows associated with transferring materials to and retrieving materials from the ReCAP facility The Workflow Consultant will document existing workflows at the partner libraries necessary to perform these activities: • Select and prepare materials for transfer to ReCAP • Request, receive, and return materials needed from ReCAP The Workflow Consultant will also document workflows at the ReCAP facility necessary to perform these activities: • Ingest and shelve new materials from each partner (including use of inventory control system) • Respond to request, retrieve, deliver, return and reshelve requested materials. Deliverables: 2. Recommend appropriate modifications to these workflows and related technology services necessary to support the shared collectionDeliverables: An initial report to the WTS due by July 31, 2012 describing current workflows and the Workflow Consultant's recommendations for workflow changes and system changes necessary to support ReCAP shared collections. A final report due September 30, 2012 based on continued participation in planning discussions with the WTS and the project team, to incorporate modifications based on subsequent decisions about services, organization, or collection management. For complete RFP please contact Joseph Brucia at josephbru...@nypl.org Brought to you by code4lib jobs: http://jobs.code4lib.org/job/790/
[CODE4LIB] Job: Technology Consultant ReCap Discovery to Delivery Project at New York Public Library
NYPL seeks a Technology Consultant to support the work of the Workflow and Technology Subcommittee (WTS). The Technology Consultant will work closely with the Technology Architect and the Planning Consultant but will also provide deliverables directly to the WTS. The Technology Consultant will be engaged to perform the following activities: 1. Prepare a technology environmental scan: a. Identify current technologies available in the ReCAP partner libraries and the ReCAP facility to support discovery, request and delivery, collections management, and collections development Describe the software systems currently in use to support these library functions at each of the three ReCAP partners and at the ReCAP facility. Include local and external systems that are used by these institutions. b. In conjunction with the Technology Architect, identify options for software systems to support those four required elements of the shared collection (discovery, request and delivery, collections management, and collections development). Based on research and domain knowledge, describe software systems (and related services) that could be used to support those functions in the context of a shared collection. Provide a general description of their functionality and their attributes e.g. propriety or open-source, perpetual license or subscription, and/or other attributes as advised by the Technology Consultant. Provide a preliminary assessment of the pros and cons of each system for use by ReCAP shared collections. Note: this activity is intended to describe potential technologies in general, not to recommend specific choices at this time. A written report to the Technology Architect and the WTS due by April 31, 2012. Make recommendations to the Technology Architect and the WTS about the design of systems infrastructure to support both user-facing and staff-facing services at each member institution to support shared collections. Working as part of a team including the Technology Architect, the Workflow Consultant, and the Planning Consultant, incorporate information about new services, revised workflows, and system options to produce recommendations about the system infrastructure needed to support ReCAP shared collections. Specific tasks include: • Review findings and recommendations made by the Workflow Consultant, who will be engaged to evaluate current workflows associated with selection from the partner libraries for ingest into the ReCAP facility, and to recommend appropriate modifications to these workflows and related technology services necessary to support the shared collection. The Workflow Consultant's initial report is expected to be available in summer 2012. • Consider findings by the Planning Consultant regarding desired organizational structure, governance, and cost- sharing (as available). • Assess existing software products applicable to shared collections (as identified in Activity #1) and recommend appropriate solutions for the ReCAP partners' shared collections system architecture. * Identify any needed systems or functionality not currently available in existing software products, with recommendations about how best to fill gaps. • Assist the Technology Architect in gathering cost information for implementing or developing the recommended system solutions. Deliverable: A report to the Technology Architect and the WTS due by September 30, 2012 describing the Technology Consultant's recommendations for system solutions to support ReCAP shared collections. The report will include any recommendations with respect to priorities, development, or implementation scheduling. For the complete RFP please contact Joseph Brucia at josephbru...@nypl.org Brought to you by code4lib jobs: http://jobs.code4lib.org/job/791/
Re: [CODE4LIB] Berkeley DB and NOID
Thanks John! I tried running db_upgrade, but apparently the utilities were not included in my setup. So I ran the following: gomez@gwnma:/var/www/nd/t1$ sudo apt-get install db4.8-util That installed the utilities, but I'm still having some trouble: gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v -h NOID NOID/noid.bdb db4.8_upgrade: Program version 4.8 doesn't match environment version 4.7 db4.8_upgrade: DB_ENV-open: DB_VERSION_MISMATCH: Database environment version mismatch I tried leaving off the environment home flag gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v NOID/noid.bdb db4.8_upgrade: NOID/noid.bdb upgraded successfully That looks like it worked. However, when I try making a call to the service I get the old error: gomez@gwnma:/var/www/nd/t1$ curl localhost/nd/noidu_t1?mint+1 no Env object (DB_VERSION_MISMATCH: Database environment version mismatch) -Josh On Mon, Feb 13, 2012 at 5:42 PM, John A. Kunze j...@ucop.edu wrote: The standard BerkeleyDB library probably changed when you upgraded Ubuntu, and it complains that the NOID database (written with the old library) is incompatible. You should be able to use db_upgrade to convert the NOID database (NOID/noid.bdb). db_upgrade is a command line utility that comes with BerkeleyDB. -John --- On Mon, 13 Feb 2012, Joshua Gomez wrote: Does anyone here have expertise with Berkeley DB? I was running an instance of NOID (which uses Berkeley DB) to mint and resolve ARKs. I updated the OS for the server it was running on from Ubuntu 9 to Ubuntu 10. Now NOID has stopped working and complains that the db version doesn't match: Program version 4.8 doesn't match environment version 4.7 I have no experience at all with Berkeley DB and could use some advice. Thanks, Josh -- Joshua Gomez Digital Library Programmer Analyst George Washington University Libraries 2130 H St, NW Washington, DC 20052 (202) 994-8267 -- Joshua Gomez Digital Library Programmer Analyst George Washington University Libraries 2130 H St, NW Washington, DC 20052 (202) 994-8267
Re: [CODE4LIB] Berkeley DB and NOID
Try removing the environment files before you upgrade (or after if you didn't save the old file). When I released that version of NOID I have since regretted leaving in code that created a stub environment (that actually isn't used by NOID) because it just creates upgrade problems that I never figured out properly. If all else fails and you're just using NOID to mint (eg, no binding), it's quite easy to start a new minter from scratch and mint until you see the last id you minted prior to conversion. The order of minting is deterministic, so in the end you'll have a new minter that's in the same state as the old minter (again, provided you haven't been doing holds and binds -- that's more complicated). -John --- On Tue, 14 Feb 2012, Joshua Gomez wrote: Thanks John! I tried running db_upgrade, but apparently the utilities were not included in my setup. So I ran the following: gomez@gwnma:/var/www/nd/t1$ sudo apt-get install db4.8-util That installed the utilities, but I'm still having some trouble: gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v -h NOID NOID/noid.bdb db4.8_upgrade: Program version 4.8 doesn't match environment version 4.7 db4.8_upgrade: DB_ENV-open: DB_VERSION_MISMATCH: Database environment version mismatch I tried leaving off the environment home flag gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v NOID/noid.bdb db4.8_upgrade: NOID/noid.bdb upgraded successfully That looks like it worked. However, when I try making a call to the service I get the old error: gomez@gwnma:/var/www/nd/t1$ curl localhost/nd/noidu_t1?mint+1 no Env object (DB_VERSION_MISMATCH: Database environment version mismatch) -Josh On Mon, Feb 13, 2012 at 5:42 PM, John A. Kunze j...@ucop.edu wrote: The standard BerkeleyDB library probably changed when you upgraded Ubuntu, and it complains that the NOID database (written with the old library) is incompatible. You should be able to use db_upgrade to convert the NOID database (NOID/noid.bdb). db_upgrade is a command line utility that comes with BerkeleyDB. -John --- On Mon, 13 Feb 2012, Joshua Gomez wrote: Does anyone here have expertise with Berkeley DB? I was running an instance of NOID (which uses Berkeley DB) to mint and resolve ARKs. I updated the OS for the server it was running on from Ubuntu 9 to Ubuntu 10. Now NOID has stopped working and complains that the db version doesn't match: Program version 4.8 doesn't match environment version 4.7 I have no experience at all with Berkeley DB and could use some advice. Thanks, Josh -- Joshua Gomez Digital Library Programmer Analyst George Washington University Libraries 2130 H St, NW Washington, DC 20052 (202) 994-8267 -- Joshua Gomez Digital Library Programmer Analyst George Washington University Libraries 2130 H St, NW Washington, DC 20052 (202) 994-8267
[CODE4LIB] Job Posting: Data Processing Automation Programmer, University of Michigan, Digital Library Production Service (DLPS)
http://umjobs.org/job_detail/66448/data_processing_automation_programmerUniversity of Michigan, Digital Library Production Service (DLPS)Data Processing Automation ProgrammerJob Summary The Library Information Technology (LIT) division provides comprehensive technology support and guidance for the University of Michigan Library system, including hosting digital library collections, coordinating electronic publishing initiatives, and supporting traditional library services (circulation of materials and management of metadata). The Digital Library Production Service (DLPS), a part of the LIT, is one of the nation's premier organizations for the creation and support of digital library resources and infrastructure, with production level support for electronic library collections. Staff are responsible for creating online access mechanisms, significant digital conversion activities, and they play a role in the University of Michigan Library's work on the HathiTrust digital library. Additionally, DLPS staff support large image and finding aid collections, and contribute to the Library's development efforts with many other formats. This is a 3-year term position with the possibility for renewal. This position has a target salary range of $40,000-$60,000 annually, dependent on qualifications and experience of the selected candidate. Responsibilities* DLPS is looking for a talented, resourceful programmer to develop, maintain, document, and monitor software systems. Primary focus will be placed on developing highly reliable software tools for routine data processing on a large scale. Specific processing tasks include file format conversion, optical character recognition (OCR), metadata insertion, transformation, validation, and transfer. Work includes assessing needs and specifying software requirements. Development of web interfaces for process management may be needed as well. Other tasks will vary but include, for example, preparing documentation and the development of digital library access systems. Required Qualifications* Bachelor's degree in computer science or an equivalent combination of education and experience. Facility with Linux or similar *nix operating systems. Demonstrated experience programming with Perl, or, experience with Ruby on Rails, Python, PHP plus willingness to learn and use Perl. Strong analytical and troubleshooting skills. Excellent verbal and written communication skills. Ability to creatively improve workflows and processes. Desired Qualifications* Demonstrated experience building dynamic web interfaces with HTML, CSS, and JavaScript. Experience using version control systems in software development. Demonstrated experience with SGML/XML and related technologies and standards. Demonstrated experience with database technology such as MySQL, including database design and implementation. Demonstrated experience developing APIs such as RESTful web services. Familiarity with batch image processing techniques in applications such as Photoshop and on the command line, ImageMagick. U-M EEO/AA Statement The University of Michigan is an equal opportunity/affirmative action employer.
Re: [CODE4LIB] Berkeley DB and NOID
That didn't work either. I started with a fresh copy of the NOID directory from before I tried the upgrade command: gomez@gwnma:/var/www/nd/t1$ ls NOID __db.001 __db.003 __db.005 lock log.01 noid.bdb __db.002 __db.004 __db.006 log logbdb README gomez@gwnma:/var/www/nd/t1$ sudo rm NOID/__* gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v -h NOID noid.bdb db4.8_upgrade: noid.bdb upgraded successfully gomez@gwnma:/var/www/nd/t1$ curl localhost/nd/noidu_t1?mint+1 no Env object (No such file or directory) I have been binding the IDs for the production NOID db (t1 is just the test NOID), but I have also been storing them in a mysql DB on our dspace server that I use to keep track of where things are in the workflow. So I suppose I could try what you suggested: setup a new production NOID, mint as many IDs as we have used so far, then rebind them all using the data in the mysql DB. Redundant data is a good thing. -Joshua On Tue, Feb 14, 2012 at 1:07 PM, John A. Kunze j...@ucop.edu wrote: Try removing the environment files before you upgrade (or after if you didn't save the old file). When I released that version of NOID I have since regretted leaving in code that created a stub environment (that actually isn't used by NOID) because it just creates upgrade problems that I never figured out properly. If all else fails and you're just using NOID to mint (eg, no binding), it's quite easy to start a new minter from scratch and mint until you see the last id you minted prior to conversion. The order of minting is deterministic, so in the end you'll have a new minter that's in the same state as the old minter (again, provided you haven't been doing holds and binds -- that's more complicated). -John --- On Tue, 14 Feb 2012, Joshua Gomez wrote: Thanks John! I tried running db_upgrade, but apparently the utilities were not included in my setup. So I ran the following: gomez@gwnma:/var/www/nd/t1$ sudo apt-get install db4.8-util That installed the utilities, but I'm still having some trouble: gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v -h NOID NOID/noid.bdb db4.8_upgrade: Program version 4.8 doesn't match environment version 4.7 db4.8_upgrade: DB_ENV-open: DB_VERSION_MISMATCH: Database environment version mismatch I tried leaving off the environment home flag gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v NOID/noid.bdb db4.8_upgrade: NOID/noid.bdb upgraded successfully That looks like it worked. However, when I try making a call to the service I get the old error: gomez@gwnma:/var/www/nd/t1$ curl localhost/nd/noidu_t1?mint+1 no Env object (DB_VERSION_MISMATCH: Database environment version mismatch) -Josh On Mon, Feb 13, 2012 at 5:42 PM, John A. Kunze j...@ucop.edu wrote: The standard BerkeleyDB library probably changed when you upgraded Ubuntu, and it complains that the NOID database (written with the old library) is incompatible. You should be able to use db_upgrade to convert the NOID database (NOID/noid.bdb). db_upgrade is a command line utility that comes with BerkeleyDB. -John --- On Mon, 13 Feb 2012, Joshua Gomez wrote: Does anyone here have expertise with Berkeley DB? I was running an instance of NOID (which uses Berkeley DB) to mint and resolve ARKs. I updated the OS for the server it was running on from Ubuntu 9 to Ubuntu 10. Now NOID has stopped working and complains that the db version doesn't match: Program version 4.8 doesn't match environment version 4.7 I have no experience at all with Berkeley DB and could use some advice. Thanks, Josh -- Joshua Gomez Digital Library Programmer Analyst George Washington University Libraries 2130 H St, NW Washington, DC 20052 (202) 994-8267 -- Joshua Gomez Digital Library Programmer Analyst George Washington University Libraries 2130 H St, NW Washington, DC 20052 (202) 994-8267 -- Joshua Gomez Digital Library Programmer Analyst George Washington University Libraries 2130 H St, NW Washington, DC 20052 (202) 994-8267
Re: [CODE4LIB] Berkeley DB and NOID
Sounds good. At this distance and given my focus on a new NOID release (that separates minting from binding in separate databases and simplifies installation by relying only on the standard DB_File), I don't think I can help you devise a better strategy than the one you've come up with. -John --- On Tue, 14 Feb 2012, Joshua Gomez wrote: That didn't work either. I started with a fresh copy of the NOID directory from before I tried the upgrade command: gomez@gwnma:/var/www/nd/t1$ ls NOID __db.001 __db.003 __db.005 lock log.01 noid.bdb __db.002 __db.004 __db.006 log logbdb README gomez@gwnma:/var/www/nd/t1$ sudo rm NOID/__* gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v -h NOID noid.bdb db4.8_upgrade: noid.bdb upgraded successfully gomez@gwnma:/var/www/nd/t1$ curl localhost/nd/noidu_t1?mint+1 no Env object (No such file or directory) I have been binding the IDs for the production NOID db (t1 is just the test NOID), but I have also been storing them in a mysql DB on our dspace server that I use to keep track of where things are in the workflow. So I suppose I could try what you suggested: setup a new production NOID, mint as many IDs as we have used so far, then rebind them all using the data in the mysql DB. Redundant data is a good thing. -Joshua On Tue, Feb 14, 2012 at 1:07 PM, John A. Kunze j...@ucop.edu wrote: Try removing the environment files before you upgrade (or after if you didn't save the old file). When I released that version of NOID I have since regretted leaving in code that created a stub environment (that actually isn't used by NOID) because it just creates upgrade problems that I never figured out properly. If all else fails and you're just using NOID to mint (eg, no binding), it's quite easy to start a new minter from scratch and mint until you see the last id you minted prior to conversion. The order of minting is deterministic, so in the end you'll have a new minter that's in the same state as the old minter (again, provided you haven't been doing holds and binds -- that's more complicated). -John --- On Tue, 14 Feb 2012, Joshua Gomez wrote: Thanks John! I tried running db_upgrade, but apparently the utilities were not included in my setup. So I ran the following: gomez@gwnma:/var/www/nd/t1$ sudo apt-get install db4.8-util That installed the utilities, but I'm still having some trouble: gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v -h NOID NOID/noid.bdb db4.8_upgrade: Program version 4.8 doesn't match environment version 4.7 db4.8_upgrade: DB_ENV-open: DB_VERSION_MISMATCH: Database environment version mismatch I tried leaving off the environment home flag gomez@gwnma:/var/www/nd/t1$ sudo db4.8_upgrade -v NOID/noid.bdb db4.8_upgrade: NOID/noid.bdb upgraded successfully That looks like it worked. However, when I try making a call to the service I get the old error: gomez@gwnma:/var/www/nd/t1$ curl localhost/nd/noidu_t1?mint+1 no Env object (DB_VERSION_MISMATCH: Database environment version mismatch) -Josh On Mon, Feb 13, 2012 at 5:42 PM, John A. Kunze j...@ucop.edu wrote: The standard BerkeleyDB library probably changed when you upgraded Ubuntu, and it complains that the NOID database (written with the old library) is incompatible. You should be able to use db_upgrade to convert the NOID database (NOID/noid.bdb). db_upgrade is a command line utility that comes with BerkeleyDB. -John --- On Mon, 13 Feb 2012, Joshua Gomez wrote: Does anyone here have expertise with Berkeley DB? I was running an instance of NOID (which uses Berkeley DB) to mint and resolve ARKs. I updated the OS for the server it was running on from Ubuntu 9 to Ubuntu 10. Now NOID has stopped working and complains that the db version doesn't match: Program version 4.8 doesn't match environment version 4.7 I have no experience at all with Berkeley DB and could use some advice. Thanks, Josh -- Joshua Gomez Digital Library Programmer Analyst George Washington University Libraries 2130 H St, NW Washington, DC 20052 (202) 994-8267 -- Joshua Gomez Digital Library Programmer Analyst George Washington University Libraries 2130 H St, NW Washington, DC 20052 (202) 994-8267 -- Joshua Gomez Digital Library Programmer Analyst George Washington University Libraries 2130 H St, NW Washington, DC 20052 (202) 994-8267
[CODE4LIB] Lift the Flap books
If you were to have a 'lift the flap' type book that you wanted to digitize, for web display and use, what technology would you use for markup and display? Visually I like the Internet Archive BookReader ( http://openlibrary.org/dev/docs/bookreader ), which says it can do 'foldouts', thought I haven't found an example of HOW to do that ... nor exactly what the metadata schema is.
Re: [CODE4LIB] Lift the Flap books
On 15/02/12 13:43, Sara Amato wrote: If you were to have a 'lift the flap' type book that you wanted to digitize, for web display and use, what technology would you use for markup and display? Visually I like the Internet Archive BookReader ( http://openlibrary.org/dev/docs/bookreader ), which says it can do 'foldouts', thought I haven't found an example of HOW to do that ... nor exactly what the metadata schema is. Sounds like an ideal use for HTML, javascript and image transparency using OnMouseOver as trigger. cheers stuart -- Stuart Yeates Library Technology Services http://www.victoria.ac.nz/library/
Re: [CODE4LIB] Lift the Flap books
On Tue, Feb 14, 2012 at 4:43 PM, Sara Amato sam...@willamette.edu wrote: If you were to have a 'lift the flap' type book that you wanted to digitize, for web display and use, what technology would you use for markup and display? I don't have an answer, but Whitney Trettien at Duke has written about this and other limitations of mainstream digitization from the theoretical perspective: http://blog.whitneyannetrettien.com/2011/05/digitizing-dead-end-branches.html (and other posts in the blog) - Avram UCLA Slavic