[CODE4LIB] Job: Developer Needed for Omeka/Scripto + Wordpress Newman website archive at The National Institute for Newman Studies

2015-08-05 Thread jobs
Developer Needed for Omeka/Scripto + Wordpress Newman website archive The National Institute for Newman Studies Pittsburgh Please follow the link below to see all the project details. https://docs.google.com/document/d/1poSC- A7V_TPhlx2uXKSx2zUdXJl_i4Dtz3q0jiIBuVc/edit Please contact me

[CODE4LIB] Job: RESEARCH DATA MANAGEMENT LIBRARIAN at Indiana University Bloomington

2015-08-05 Thread jobs
RESEARCH DATA MANAGEMENT LIBRARIAN Indiana University Bloomington Bloomington **RESEARCH DATA MANAGEMENT LIBRARIAN ASSISTANT LIBRARIAN OR ASSOCIATE LIBRARIAN INDIANA UNIVERSITY BLOOMINGTON LIBRARIES** Founded in 1820, Indiana University Bloomington has grown from a small state seminary

[CODE4LIB] Processing Circ data

2015-08-05 Thread Harper, Cynthia
Hi all. What are you using to process circ data for ad-hoc queries. I usually extract csv or tab-delimited files - one row per item record, with identifying bib record data, then total checkouts over the given time period(s). I have been importing these into Access then grouping them by bib

Re: [CODE4LIB] Processing Circ data

2015-08-05 Thread Owen Stephens
Another option might be to use OpenRefine http://openrefine.org - this should easily handle 250,000 rows. I find it good for basic data analysis, and there are extensions which offer some visualisations (e.g. the VIB BITs extension which will plot simple data using d3

Re: [CODE4LIB] Processing Circ data

2015-08-05 Thread Kevin Ford
Hi Cindy, This doesn't quite address your issue, but, unless you've hit the 2 GB Access size limit [1], Access can handle a good deal more than 250,000 item records (rows, yes?) you cited. What makes you think you've hit the limit? Slowness, something else? All the best, Kevin [1]

Re: [CODE4LIB] Processing Circ data

2015-08-05 Thread Harper, Cynthia
Well, I guess it could be bad data, but I don't know how to tell. I think I've done more than this before. I have a Find duplicates query that groups by bib record number. That query seemed to take about 40 minutes to process. Then I added a criterion to limit to only records that had 0 circs

Re: [CODE4LIB] Processing Circ data

2015-08-05 Thread Kevin Ford
On the surface, your difficulties suggest you may need look at a few optimization tactics. Apologies if these are things you've already considered and addressed - just offering a suggestion. This page [1] is for Access 2003 but the items under Improve query performance should apply - I think

[CODE4LIB] Fwd: Survey on embedded metadata in digital objects

2015-08-05 Thread Edward M. Corrado
Dear Colleagues, You are invited to participate in a survey designed to collect information on the practice of embedding metadata into digital objects. The purpose of the survey is to explore the cost and benefit of embedding additional (i.e. LAM-generated) metadata into digital objects, to