The 40th European Library Automation Group Systems Seminar will be in the Royal library, Copenhagen (Denmark), June 6-9 2016 ELAG is Europe’s premier conference on the application of information technology in libraries and documentation centres. For over thirty years, the ELAG (European Library Automation Group) Conference has provided library and IT professionals with the opportunity to discuss new technologies, to review on-going developments and to exchange best practices. The theme that was chosen for this year conference is ‘EXIT’. We talk a lot about future developments, but not so much about what we leave behind. We transition data and applications. We stop doing some things and start doing new things. What effort do we put in transitioning data? How do we implement new services ? How do we implement new data models and architecture and what problems do we encounter in that process ? How do we evaluate the effect of transitions ? Are your proud of the work you have been doing ? Do you have a great idea ? Are you able to teach your colleagues something ? Make yourself heard by doing a presentation at Elag 2016 ! Head over to our conference website and submit a proposal for a presentation, workshop or bootcamp! http://elag.org Patrick Hochstenbach - digital architect University Library Ghent Rozier 9 - 9000 Ghent - Belgium patrick.hochstenb...@ugent.be +32 (0)9 264 7980
Dear All, You are warmly invited to a one day event in Ghent, Belgium, hosted by the International Image Interoperability Framework community (http://iiif.io/) and Ghent University Library (http://lib.ugent.be/), describing the power and potential of interoperable image delivery over the Web. The day will showcase how institutions are leveraging IIIF to reduce total cost and time to deploy image delivery solutions, while simultaneously improving end user experience with a new host of rich and dynamic features. It will also highlight how to participate in this growing movement to take advantage of the common framework. This event will be valuable for organizational decision makers, repository and collection managers, software engineers; for cultural heritage or STEM (science / technology / engineering / medicine) institutions; or for anyone engaged with image-based resources on the Web. The event will be held at the beautiful Ghent Opera House on Tuesday December 8th, 2015. There is no cost to attend, but please register on EventBrite: http://iiif-ghent-2015.eventbrite.com/ A detailed program and further logistical information is available at: http://iiif.io/event/2015/ghent.html There will be many opportunities for discussion, questions and networking throughout the day with new and existing partners including national libraries, top tier research institutions, commercial providers and major aggregators. Please register now on EventBrite (http://iiif-ghent-2015.eventbrite.com/), and join iiif-disc...@googlegroups.com
for announcements and discussion regarding the event.
SAFE-PLN partners are glad to announce their data archiving grid is open to international partners. Using a data grid, the five institutions agree to archive each others’ open access collections across two continents, two timezones, four languages and in seven copies to guarantee perpetual access to their scientific heritage. The head of libraries of five universities have signed a collaboration agreement on deploying a preservation solution for open access collections of dissertations, publications and research data: Ghent University, Memorial University of Newfoundland, Universität Bielefeld, Université Catholique de Louvain and Université Libre de Bruxelles Their solution, named SAFE-PLN (SAFE Archiving Federation Private LOCKSS Network), consists in deploying an internationally distributed network to safeguard core collections. It is based on the open-source preservation software LOCKSS (http://www.lockss.org/community/networks/) developed by Stanford University. Worldwide, SAFE-PLN is the third fully operational PLN with an international scope. More information is available on the project website: www.safepln.org.
Hi, I need some advise on creating MODS records for our institutional repository. In particular I wonder how best to express the different access restrictions on digital files when a record contains more than one full-text file. E.g. what we do now is write something like: location url displayLabel=ruimtelijk_bestuursrecht_Geert_13-12-10.pdfhttps://biblio.ugent.be/publication/1927382/file/1927384/url /location physicalDescription internetMediaTypeapplication/pdf/internetMediaType /physicalDescription accessCondition type=restrictionOnAccessrestricted (changes to open on 2016-01-01)/accessCondition and this repeated for every full-text file in the record I don't like this solution because: 1. This make the MODS context-sensitive: the order of local, physical, accessCondition has a meaning (the first accessCondition is for the first location, the second accessCondition ois for the second loaction etc etc). As I understand the order of elementents in MODS shouldn't matter. 2. Access conditions and embargo's are free-text! Are there best practices we should use? Greetings from Belgium Patrick Ghent University Library
[CODE4LIB] PubLister/LibreCat Software Developer Workshop on 29 and 30 November 2012 | Bielefeld Germany
***Apologies for cross-posting*** We would like to invite you to the joint PubLister/LibreCat Software Developer Workshop on 29 and 30 November 2012 (1 - 7pm ; 9am - 2pm CEST) at Bielefeld University. Building off last year's PubLister Symposium and Workshop http://pub.uni-bielefeld.de/workshop/ , the University Libraries of Lund, Gent and Bielefeld share the common vision of 1) Creating a high-level system of building blocks that can be reused when creating repository-like applications: Project Catmandu 2) Creating a next-generation repository service based on these building blocks: Project LibreCat. http://librecat.org/ After showcasing exemplary implementations at the three Universities, the workshop will get you started building repository applications with Catmandu. The sources needed are distributed both via CPAN http://search.cpan.org/search?m=allq=catmandu and GitHub https://github.com/LibreCat . A brief tutorial can be found here http://librecat.org/tutorial/index.html The workshop takes place at the Center of Excellence - Cognitive Interaction Technology (CITEC) at Bielefeld University. We acknowledge the support of the German Research Foundation (DFG). For free registration please contact publikationsdienste...@uni-bielefeld.de Please be aware that the number of participants is limited. Registration deadline is November 15. With kind regards, Najko Jahn --- Najko Jahn Universität Bielefeld - UniversitÃ¤tsbibliothek Universitätsstr. 25 - 33615 Bielefeld - www.ub.uni-bielefeld.de Office A2-139 Tel. +49 (0) 521/106-2546 Fax +49 (0) 521/106-4052 najko.j...@uni-bielefeld.de skype: najko.jahn
Or in case you like Java ..I've started working last year on a Java application to automatically print documents from a hot folder. Just pushed the code: https://github.com/phochste/PrintApp Patrick From: Code for Libraries [CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Kozlowski,Brendon [bkozlow...@sals.edu] Sent: Tuesday, April 03, 2012 6:15 PM To: CODE4LIB@LISTSERV.ND.EDU Subject: [CODE4LIB] Silently print (no GUI) in Windows I'm curious to know if anyone has discovered ways of silently printing documents from such Windows applications as: - Acrobat Reader (current version) - Microsoft Office 2007 (Word, Excel, Powerpoint, Visio, etc...) - Windows Picture and Fax Viewer I unfortunately haven't had much luck finding any resources on this. I'd like to be able to receive documents in a queue like fashion to a single PC and simply print them off as they arrive. However, automating the loading/exiting of the full-blown application each time, and on-demand, seems a little too cumbersome and unnecessary. I have not yet decided on whether I'd be scripting it (PHP, AutoIT, batch files, VBS, Powershell, etc...) or learning and then writing a .NET application. If .NET solutions use the COM object, the scripting becomes a potential candidate. Unfortunately I need to know how, or even if, it's even possible to do first. Thank you for any and all feedback or assistance. Brendon Kozlowski Web Administrator Saratoga Springs Public Library 49 Henry Street Saratoga Springs, NY, 12866  584-7860 x217 Please consider the environment before printing this message. To report this message as spam, offensive, or if you feel you have received this in error, please send e-mail to ab...@sals.edu including the entire contents and subject of the message. It will be reviewed by staff and acted upon appropriately.
Dear All At Ghent Universty Library in Belgium we are re-evaluating our web strategy in several internal workshops with faculty librarians. The theme of our next meeting will be about discovery, and one-search-box integrated search solutions. We see these days a lot of vendors and cloud providing solutions that provide these features which addresses strategic goals of libraries. We are also interested in a methodological question: what are the underlying use-cases libraries are trying to solve for where a new suite of services provide a solution. Does you have insights you are willing to share. Searching for methodologies, principles of heuristics, statistics, research and teaching practices. Best from Belgium Patrick Skype: patrick.hochstenbach Patrick Hochstenbach Digital Architect University Library+32(0)92647980 Ghent University * Rozier 9 * 9000 * Gent
Dear Nate, There is a trade-off: do you want very fast processing of data - go for binary data. do you want to share your data globally easily in many (not per se library related) environments - go for XML/RDF. Open your data and do both :-) Pat Sent from my iPhone On 25 Oct 2010, at 20:39, Nate Vack njv...@wisc.edu wrote: Hi all, I've just spent the last couple of weeks delving into and decoding a binary file format. This, in turn, got me thinking about MARCXML. In a nutshell, it looks like it's supposed to contain the exact same data as a normal MARC record, except in XML form. As in, it should be round-trippable. What's the advantage to this? I can see using a human-readable format for poorly-documented file formats -- they're relatively easy to read and understand. But MARC is well, well-documented, with more than one free implementation in cursory searching. And once you know a binary file's format, it's no harder to parse than XML, and the data's smaller and processing faster. So... why the XML? Curious, -Nate
Dear all, At Ghent University the department of telecommunications and information processing is brainstorming on a project on citation linking. They have quite some expertise in flexible querying and information retrieval. They would like to try out their algorithms on public training sets of references and bibliographic data. The task is to train their algorithms to find all matches between citations and a corpus of publications. The challenge (as we all know from related projects/products) is to match the 'bad' citation data with 'good' publication data. Are there some public datasets available which were human tested examined to really get good precision/recall numbers for the proposed algorithms? Datasets which are/can be used in current/future shootouts between citation matching algorithms? Thanks Patrick Skype: patrick.hochstenbach Patrick Hochstenbach Digital Architect University Library+32(0)92647980 Ghent University * Rozier 9 * 9000 * Gent
Hi, We have a interesting job opening for a developer on site at Ghent University Library for 13 months. You'll join our team to participate in creating a image search engine for high resolution scans of old manuscripts. Experience with Java is very welcome. Like to learn djatoka,imageio, clojure, solr, couchdb? Please contact me. Job description (in Dutch): http://www.ugent.be/nl/nieuwsagenda/vacatures/atp/contract-tijdelijk/ivh Best, Patrick
Nothing beats E- http://en.wikipedia.org/wiki/E_%28programming_language%29 sexy e - 924,000 hits But oh poor Erlang sexy erlang - 2 hits (both of them telling me: erlang isn't sexy) P@ -Original Message- From: Code for Libraries on behalf of Tim Spalding Sent: Fri 26-3-2010 4:21 To: CODE4LIB@LISTSERV.ND.EDU Subject: Re: [CODE4LIB] newbie Ruby may be sexy but sexy ruby on rails gets only four hits. As for sexy python, well, no comment. T On Thu, Mar 25, 2010 at 10:36 PM, Andrew Hankinson andrew.hankin...@gmail.com wrote: Just out of curiosity I tried them in quotes: sexy ruby - 72,200 sexy python - 37,900 sexy php - 25,100 sexy java - 16,100 sexy asp - 14,800 sexy perl - 8,080 sexy C++ - 177 sexy FORTRAN - 67 sexy COBOL - 8 I tried sexy lisp but the results were skewed by speech impediment fetishes. Which I'd say is even less strange than 8 people thinking you can write sexy COBOL.
Depends on your datamodel, Godar. You could also consider databases like CouchDB. Not XML ..but if your datamodel can fit into JSON. Efficient serving of docs over HTTP is their trademark, like scaling through replication. Lucene. CouchDB has Lucene integration..but I find it somewhat flaky. In my case I did batch index jobs of the database. In another project we could (I don't say easily) fit the datamodel into MySQL. Our developers could then reuse all the MySQL tools, scripts. The sysadmin was happy. So first consider if XML is really needed throughout the whole codebase. Are you working with textual documents in XML, or database dumps in XML? Best, P@ Skype: patrick.hochstenbach Patrick Hochstenbach Software Architect University Library +32(0)92647980 Ghent University * Rozier 9 * 9000 * Gent -Original Message- From: Code for Libraries on behalf of Andrew Nagy Sent: Mon 18-1-2010 1:28 To: CODE4LIB@LISTSERV.ND.EDU Subject: Re: [CODE4LIB] Q: what is the best open source native XML database I've had the best luck with eXist and BerkeleyDB XML. Both support XQuery and have indexing features based on any XML structure. Andrew On 1/16/10, Godmar Back god...@gmail.com wrote: Hi, we're currently looking for an XML database to store a variety of small-to-medium sized XML documents. The XML documents are unstructured in the sense that they do not follow a schema or DTD, and that their structure will be changing over time. We'll need to do efficient searching based on elements, attributes, and full text within text content. More importantly, the documents are mutable. We'll like to bring documents or fragments into memory in a DOM representation, manipulate them, then put them back into the database. Ideally, this should be done in a transaction-like manner. We need to efficiently serve document fragments over HTTP, ideally in a manner that allows for scaling through replication. We would prefer strong support for Java integration, but it's not a must. Have other encountered similar problems, and what have you been using? So far, we're researching: eXist-DB (http://exist.sourceforge.net/ ), Base-X (http://www.basex.org/ ), MonetDB/XQuery (http://www.monetdb.nl/XQuery/ ), Sedna (http://modis.ispras.ru/sedna/index.html ). Wikipedia lists a few others here: http://en.wikipedia.org/wiki/XML_database I'm wondering to what extent systems such as Lucene, or even digital object repositories such as Fedora could be coaxed into this usage scenario. Thanks for any insight you have or experience you can share. - Godmar -- Sent from my mobile device