Re: [CODE4LIB] Rdio playlist

2013-02-04 Thread Matt Schultz
Last time I took a road trip/weekend trip to Chicago I spun up a small
Spotify playlist of Chicago-based bands that included:

Telefon Tel Aviv
Smashing Pumpkins
Ministry
Pigface
Disappears
Secret Colours
Smith Westerns
Screaching Weasel
Wilco

Most, but not all, are in-line with my tastes, so I know there are lots
more Chicago-based artists that could be added. I'm not on Rdio or going to
Code4Lib, but may get on and listen to your playlist. If nothing else I'll
fire up my old playlist and be there in spirit.


On Sun, Feb 3, 2013 at 10:44 PM, William Denton w...@pobox.com wrote:

 Do you ever put together a playlist or some situationally appropriate
 music when you're taking a trip?  Do you ever do this when you're going to
 a conference?  If you answered yes to either, and you're on Rdio, pitch in
 and add some stuff to a Code4Lib Chicago playlist:

 http://www.rdio.com/people/**wdenton/playlists/2229053/**
 Code4Lib_2013_in_Chicago/http://www.rdio.com/people/wdenton/playlists/2229053/Code4Lib_2013_in_Chicago/

 I reckon we should be able to collaboratively edit the perfect playlist
 for use by hundreds of librarians and coders descending on Chicago in
 February, useful when either a) drinking local beers, b) hacking, c)
 walking through the Art Institute of Chicago, or d) all of the above.

 Bill
 --
 William Denton
 Toronto, Canada
 http://www.miskatonic.org/




-- 
Matt Schultz
Program Manager
Educopia Institute, MetaArchive Cooperative
http://www.metaarchive.org
matt.schu...@metaarchive.org
616-566-3204


Re: [CODE4LIB] Why we need multiple discovery services engine?

2013-02-04 Thread Emily Lynema
Here at NCSU, we use our locally-hosted Endeca service for our catalog
and Serials
Solutions Summon as an article search solution. Why do this?

1. Our next-gen library catalog (Endeca) came first. This was before Solr hit
the library world, and before library vendors started working on
improving their bundled catalog apps. Our bundled catalog was terrible, and we
wanted something better. This was back in the day when everyone was doing
federated search for articles (think MetaLib).

2. 4-5 years down the road, a number of vendors (Ebsco, Serials
Solutions, etc.)
were getting into the web scale discovery business. Aka, one big index that
includes everything, in particular the citation content that libraries have
historically not had local access to index / search. We bought Summon to
solve the article search problem that federated searching never resolved
for us. We wanted one access point for less experienced users who needed to
find articles. Since we had backed away from federated search for articles,
this was our big pain point; we already had a catalog we liked.

We've actually loaded our catalog content into Summon, as well. So why
keep both?
We've done a LOT of work adding functionality into our local catalog, including
enhanced title searching,lots of supplemental content, a quite complex
local requesting system. So we can't just switch to the Summon interface
without some effort.

In addition, we have found that we prefer the bento box approach to
searching across formats, as opposed to the integrated index approach
of Summon.
At least at this moment. We use this in the search across our library
website [1]. It's just really, really hard to always surface the
right kind of thing the user is looking for when the things you're indexing are
different in nature (ex: bibliographic record vs. full-text of
newspaper article). With the bento box approach, you have better
opportunities to surface the different types of content available, while
still having local systems optimized for specific content types.

Maybe that's a long-winded excuse for not pushing to break down silos
more. Time
will probably tell.

-emily

[1] http://www.lib.ncsu.edu/search/?q=java


-

Date:Fri, 1 Feb 2013 04:21:01 +
From:Jonathan Rochkind rochk...@jhu.edu
Subject: Re: Why we need multiple discovery services engine?

So, there are two categories of solutions here -- 1) local indexes, where
you create the index yourself, like blacklight or vufind (both based on a
local Solr).  2) vendor-hosted indexes, where the vendor includes all sorts
of things in their index that you the customer don't have local metadata
for, mostly including lots and lots of scholarly article citations.

If you want to include scholarly article citations, you probably can't do
that with a local index solution. Although some consortiums have done some
interesting stuff in that area, let's just say it takes a lot of resources
to do. For most people, if you want to include article search in your
index, it's not feasilbe to do so with a local index. So only
VuFind/Blacklight with a local Solr is out, if you want article search.

You _can_ load local content in a vendor-hosted index like
EDS/Primo/Summon. So plenty of people do choose a vendor-hosted index
product as their only discovery tool, including both local metadata and
vendor-provided metadata. As you suggest.

But some people want the increased control that a locally controlled Solr
index gives you, for the local metadata where it's feasible. So use a local
index product. But still want the article search you can get with a
vendor-hosted index product. So they use both.

There is also at least some reasons to believe that our users don't mind
and may even prefer having local results and hosted metadata results
presented seperately (although probably preferably in a consistent UI),
rather than merged.

A bunch more discussion of these issues is included in my blog post at:
http://bibwild.wordpress.com/2012/10/02/article-search-improvement-strategy/

From: Code for Libraries [CODE4LIB@LISTSERV.ND.EDU] on behalf of Wayne Lam [
wing...@gmail.com]
Sent: Thursday, January 31, 2013 9:31 PM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: [CODE4LIB] Why we need multiple discovery services engine?

Hi all,

I saw in numerous of library website, many of them would have their own
based discovery services (e.g. blacklight / vufind) and at the same time
they will have vendor based discovery services (e.g. EDS / Primo / Summon).
Instead of having to maintain 2 separate system, why not put everything
into just one? Any special reason or concern?

Best

Wayne

--
Emily Lynema
Associate Department Head
Information Technology, NCSU Libraries
919-513-8031
emily_lyn...@ncsu.edu


Re: [CODE4LIB] Anyone have a SUSHI client?

2013-02-04 Thread Joshua Welker
Tom,

I am the guy who wrote sushi.py around this time last year. My apologies for 
the shabbiness of the code. It was meant to be primarily a proof of concept. 
It's definitely incomplete. I only completed the DB3 and JR1 report logic up to 
this point, but it would be easy enough to add other report types. You're also 
right that sushi.py doesn't do anything to dedupe data, but it would be very 
simple to write a script that reads through the SQL records and deletes dupes. 
You could also use the built-in UNIQUE flag in MySQL when creating your table 
so that duplicate records just don't get saved. If you use the CSV export 
functionality of sushi.py, Excel has some built-in dedupe features that would 
help as well.

Let me know if you'd like some help modifying sushi.py. I sort of gave up on it 
last spring. SUSHI implementation among vendors is still pretty shabby, and 
there are still some weaknesses in the SUSHI standard (I wrote about them in 
the Nov 2012 issue of Computers in Libraries). The productivity gains I was 
seeing from using SUSHI ended up being pretty low.

Josh Welker


-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Tom 
Keays
Sent: Friday, January 25, 2013 8:40 AM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Anyone have a SUSHI client?

I've been looking briefly at sushi.py, as a way to orient myself to collecting 
stats this way. I'm not intending to single out sushi.py, but looking at it 
(mainly the data structure at this point, and not the code itself), raises some 
questions about the best approach for collecting SUSHI data.

sushi.py seems to have a small number of routines; mainly to retrieve the XML 
file from a vendor and ingest the data in that file into a MySQL database. 
There are only MySQL tables for COUNTER JR1, DR1, DR2, and DR2 reports and they 
mirror, to a degree, the structure of the item records returned in the SUSHI 
xml. Here are the skeletons of 2 of the sushi.py SQL
tables:

counter_jr1
  id int,
  print_issn varchar,
  online_issn varchar,
  platform varchar,
  item_name text,
  data_type varchar,
  date_begin datetime,
  date_end datetime,
  ft_pdf int,
  ft_html int,
  ft_total varchar

counter_db3
  id int,
  platform varchar,
  item_name text,
  data_type varchar,
  date_begin datetime,
  date_end datetime,
  searches int,
  sessions int

On the face of it, this seems like a pretty good data structure (although I 
have a couple of concerns, that I will get to) but my main question is whether 
there is any agreement about how to collect this data? If I were to dig into 
some of the other SUSHI packages mentioned in this thread, what would I find 
there? Excel-formatted COUNTER reports are simply a table of columns, 
representing various fields, such as title (for JR1), platform, publisher (for 
JR1), ISSN (for JR1), etc., followed by columns for up to 12 months of the 
collected year, and then summary data.  JR1 reports have fulltext HTML, PDF, 
and Total columns. DR1 has two rows, one for searches and one for sesssions, 
with YTD totals in the final column. Similar data structures exist for other 
COUNTER reports. They rely on the user to interpret them and probably ought not 
to inform a decision for structuring the data in a database. Is there been any 
best practice for how COUNTER data is modeled in a database?

There are other COUNTER reports besides those four. For instance, some journal 
vendors do indeed report searches and sessions using the DR3 report, but others 
use the equivalent JR4 report, so I would have expected sushi.py to have a 
mechanism to collect these. Does SUSHI only deliver JR1, DR1, DR2, and DR2 
reports, or is this a problem with sushi.py?

Now, one of the selling points for SUSHI is that if a vendor ever advises that 
you should re-collect data for a given time period, the xml you receive is 
structured such that the act of collecting OUGHT TO update, rather than 
duplicate, data previously collected. However in sushi.py's SQL structure, 
which gives every row a unique (auto-incremented) ID number, there would have 
to be logic applied during the ingest to prevent multiple instances of data 
collected from the same vendor for the same time period.
So, that's a concern.

I'm also concerned about what is represented in the ft_pdf, ft_html, and 
ft_total fields. In the Excel COUNTER reports, the ft_pdf, ft_html, and 
ft_total columns simply tabulate the YTD totals and the only way you would be 
able to derive a monthly breakdown would be to collect 12 monthly reports and 
analyze the differences from month to month -- something that most libraries 
don't do. I have to go back and confirm this, but I don't think the SUSHI 
reports are giving a month-only breakdown for those fields, so I wonder about 
their inclusion in that table. I guess my question is what is returned in the 
SUSHI xml report: monthly or yearly figures for the ft_pdf, ft_html, and 
ft_total fields?


Re: [CODE4LIB] Rdio playlist

2013-02-04 Thread Keith Jenkins
Can't believe no one has yet mentioned Chicago band I Fight Dragons
-- they mix NES-controller chiptunes with electric guitars, playing
rockin' covers of Mario Bros., Legend of Zelda, Contra, etc. in
addition to some originals that have been remixed by other Chicago
musicians.  They just got back from their War of Cyborg Liberation
Tour.


[CODE4LIB] Linked data [was: Why we need multiple discovery services engine?]

2013-02-04 Thread Donna Campbell
In mentioning pushing to break down silos more, it brings to mind a
question I've had about linked data.

From what I've read thus far, the idea of breaking down silos of
information seems like a good one in that it makes finding information
easier but doesn't it also remove some of the markers of finding credible
sources? Doesn't it blend accurate sources and inaccurate sources?


Donna R. Campbell
Technical Services  Systems Librarian
(215) 935-3872 (phone)
(267) 295-3641 (fax)
Mailing Address (via USPS):
Westminster Theological Seminary Library
P.O. Box 27009
Philadelphia, PA 19118  USA
Shipping Address (via UPS or FedEx):
Westminster Theological Seminary Library
2960 W. Church Rd.
Glenside, PA 19038  USA

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of
Emily Lynema
Sent: Monday, February 04, 2013 9:56 AM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Why we need multiple discovery services engine?

Here at NCSU, we use our locally-hosted Endeca service for our catalog
and Serials
Solutions Summon as an article search solution. Why do this?

1. Our next-gen library catalog (Endeca) came first. This was before Solr
hit
the library world, and before library vendors started working on
improving their bundled catalog apps. Our bundled catalog was terrible,
and we
wanted something better. This was back in the day when everyone was doing
federated search for articles (think MetaLib).

2. 4-5 years down the road, a number of vendors (Ebsco, Serials
Solutions, etc.)
were getting into the web scale discovery business. Aka, one big index
that
includes everything, in particular the citation content that libraries
have
historically not had local access to index / search. We bought Summon to
solve the article search problem that federated searching never resolved
for us. We wanted one access point for less experienced users who needed
to
find articles. Since we had backed away from federated search for
articles,
this was our big pain point; we already had a catalog we liked.

We've actually loaded our catalog content into Summon, as well. So why
keep both?
We've done a LOT of work adding functionality into our local catalog,
including
enhanced title searching,lots of supplemental content, a quite complex
local requesting system. So we can't just switch to the Summon interface
without some effort.

In addition, we have found that we prefer the bento box approach to
searching across formats, as opposed to the integrated index approach
of Summon.
At least at this moment. We use this in the search across our library
website [1]. It's just really, really hard to always surface the
right kind of thing the user is looking for when the things you're
indexing are
different in nature (ex: bibliographic record vs. full-text of
newspaper article). With the bento box approach, you have better
opportunities to surface the different types of content available, while
still having local systems optimized for specific content types.

Maybe that's a long-winded excuse for not pushing to break down silos
more. Time
will probably tell.

-emily

[1] http://www.lib.ncsu.edu/search/?q=java


Re: [CODE4LIB] Anyone have a SUSHI client?

2013-02-04 Thread Tom Keays
Hi Joshua,

I was mainly looking at your program, not for the code, but as a way to
bring myself up to speed about current practices in modeling the COUNTER
data. I'm trying to avoid reinventing something that has already been well
thought through. I apologize for calling out your model. You have gotten
much further than I have. Some of the other respondents in this thread have
set me straight on some things I was very fuzzy on going in.

How go about I collecting and storing the data is still something I haven't
resolved yet. I personally would prefer a Python solution, but there forces
here at MPOW that suggest I should build a data repository in SharePoint.
Assuming that is the case, Serial Solution's open source SUSHI harvester
written in .NET might actually be the way for me to go. So, my next step is
to look at their data model and see what reports they collect and store.

As an aside, I'm also now wondering if de-duping is strictly necessary as
long as there is a field to record the date the report was generated.
 De-duping (or maybe just deprecating duplicate data) could be separate
from the collection process.

Best,
Tom

On Mon, Feb 4, 2013 at 10:07 AM, Joshua Welker jwel...@sbuniv.edu wrote:

 Tom,

 I am the guy who wrote sushi.py around this time last year. My apologies
 for the shabbiness of the code. It was meant to be primarily a proof of
 concept. It's definitely incomplete. I only completed the DB3 and JR1
 report logic up to this point, but it would be easy enough to add other
 report types. You're also right that sushi.py doesn't do anything to dedupe
 data, but it would be very simple to write a script that reads through the
 SQL records and deletes dupes. You could also use the built-in UNIQUE flag
 in MySQL when creating your table so that duplicate records just don't get
 saved. If you use the CSV export functionality of sushi.py, Excel has some
 built-in dedupe features that would help as well.

 Let me know if you'd like some help modifying sushi.py. I sort of gave up
 on it last spring. SUSHI implementation among vendors is still pretty
 shabby, and there are still some weaknesses in the SUSHI standard (I wrote
 about them in the Nov 2012 issue of Computers in Libraries). The
 productivity gains I was seeing from using SUSHI ended up being pretty low.

 Josh Welker


 -Original Message-
 From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of
 Tom Keays
 Sent: Friday, January 25, 2013 8:40 AM
 To: CODE4LIB@LISTSERV.ND.EDU
 Subject: Re: [CODE4LIB] Anyone have a SUSHI client?

 I've been looking briefly at sushi.py, as a way to orient myself to
 collecting stats this way. I'm not intending to single out sushi.py, but
 looking at it (mainly the data structure at this point, and not the code
 itself), raises some questions about the best approach for collecting SUSHI
 data.

 sushi.py seems to have a small number of routines; mainly to retrieve the
 XML file from a vendor and ingest the data in that file into a MySQL
 database. There are only MySQL tables for COUNTER JR1, DR1, DR2, and DR2
 reports and they mirror, to a degree, the structure of the item records
 returned in the SUSHI xml. Here are the skeletons of 2 of the sushi.py SQL
 tables:

 counter_jr1
   id int,
   print_issn varchar,
   online_issn varchar,
   platform varchar,
   item_name text,
   data_type varchar,
   date_begin datetime,
   date_end datetime,
   ft_pdf int,
   ft_html int,
   ft_total varchar

 counter_db3
   id int,
   platform varchar,
   item_name text,
   data_type varchar,
   date_begin datetime,
   date_end datetime,
   searches int,
   sessions int

 On the face of it, this seems like a pretty good data structure (although
 I have a couple of concerns, that I will get to) but my main question is
 whether there is any agreement about how to collect this data? If I were to
 dig into some of the other SUSHI packages mentioned in this thread, what
 would I find there? Excel-formatted COUNTER reports are simply a table of
 columns, representing various fields, such as title (for JR1), platform,
 publisher (for JR1), ISSN (for JR1), etc., followed by columns for up to 12
 months of the collected year, and then summary data.  JR1 reports have
 fulltext HTML, PDF, and Total columns. DR1 has two rows, one for searches
 and one for sesssions, with YTD totals in the final column. Similar data
 structures exist for other COUNTER reports. They rely on the user to
 interpret them and probably ought not to inform a decision for structuring
 the data in a database. Is there been any best practice for how COUNTER
 data is modeled in a database?

 There are other COUNTER reports besides those four. For instance, some
 journal vendors do indeed report searches and sessions using the DR3
 report, but others use the equivalent JR4 report, so I would have expected
 sushi.py to have a mechanism to collect these. Does SUSHI only deliver JR1,
 DR1, DR2, and DR2 reports, or is this a problem with 

[CODE4LIB] conf presenters: a kind request

2013-02-04 Thread Jonathan Rochkind
We are all very excited about the conference next week, to speak to our 
peers and to hear what our peers have to say!


I would like to suggest that those presenting be considerate to your 
audience, and actually prepare your talk in advance!


You may think you can get away with making some slides that morning 
during someone elses talk and winging it; nobody will notice right? Or 
they wont' care if they do?


From past years, I can say that for me at least, yeah, I can often tell 
who hasn't actually prepared their talk. And I'll consider it 
disrespectful to the time of the audience, who voted for your talk and 
then got on airplanes to come see it, and you didn't spend the time to 
plan it advance and make it as high quality for them as you could.


I don't mean to make people nervous about public speaking. The code4lib 
audience is a very kind and generous audience, they are a good audience. 
It'll go great! Just maybe repay their generosity by actually preparing 
your talk in advance, you know?  Do your best, it'll go great!


If you aren't sure how to do this, the one thing you can probably do to 
prepare (maybe this is obvious) is practice your presentation in 
advance, with a timer, just once.  In front of a friend or just by 
yourself. Did you finish on time, and get at least half of what was 
important in? Then you're done preparing, that was it!  Yes, if you're 
going to have slides, this means making your slides or notes/outline in 
advance so you can practice your delivery just once!


Just practice it once in advance (even the night before, as a last 
resort!), and it'll go great!


Jonathan


Re: [CODE4LIB] Why we need multiple discovery services engine?

2013-02-04 Thread Owen, Will
I think this pretty accurately reflects the sentiments at the University
of North Carolina, one of NSCU's partners in developing the Endeca
interface to our catalog.  It's as much a mechanism for delivering
services as it is for discovery, and as such reflects the physical nature
of the collections it includes.  It's not possible at this point for us to
build those services into the Summon environment.  +1 all around to
Emily's comments.

Will Owen

On 2/4/13 9:56 AM, Emily Lynema emilylyn...@gmail.com wrote:

Here at NCSU, we use our locally-hosted Endeca service for our catalog
and Serials
Solutions Summon as an article search solution. Why do this?

1. Our next-gen library catalog (Endeca) came first. This was before Solr
hit
the library world, and before library vendors started working on
improving their bundled catalog apps. Our bundled catalog was terrible,
and we
wanted something better. This was back in the day when everyone was doing
federated search for articles (think MetaLib).

2. 4-5 years down the road, a number of vendors (Ebsco, Serials
Solutions, etc.)
were getting into the web scale discovery business. Aka, one big index
that
includes everything, in particular the citation content that libraries
have
historically not had local access to index / search. We bought Summon to
solve the article search problem that federated searching never resolved
for us. We wanted one access point for less experienced users who needed
to
find articles. Since we had backed away from federated search for
articles,
this was our big pain point; we already had a catalog we liked.

We've actually loaded our catalog content into Summon, as well. So why
keep both?
We've done a LOT of work adding functionality into our local catalog,
including
enhanced title searching,lots of supplemental content, a quite complex
local requesting system. So we can't just switch to the Summon interface
without some effort.

In addition, we have found that we prefer the bento box approach to
searching across formats, as opposed to the integrated index approach
of Summon.
At least at this moment. We use this in the search across our library
website [1]. It's just really, really hard to always surface the
right kind of thing the user is looking for when the things you're
indexing are
different in nature (ex: bibliographic record vs. full-text of
newspaper article). With the bento box approach, you have better
opportunities to surface the different types of content available, while
still having local systems optimized for specific content types.

Maybe that's a long-winded excuse for not pushing to break down silos
more. Time
will probably tell.

-emily

[1] http://www.lib.ncsu.edu/search/?q=java


-

Date:Fri, 1 Feb 2013 04:21:01 +
From:Jonathan Rochkind rochk...@jhu.edu
Subject: Re: Why we need multiple discovery services engine?

So, there are two categories of solutions here -- 1) local indexes, where
you create the index yourself, like blacklight or vufind (both based on a
local Solr).  2) vendor-hosted indexes, where the vendor includes all
sorts
of things in their index that you the customer don't have local metadata
for, mostly including lots and lots of scholarly article citations.

If you want to include scholarly article citations, you probably can't do
that with a local index solution. Although some consortiums have done some
interesting stuff in that area, let's just say it takes a lot of resources
to do. For most people, if you want to include article search in your
index, it's not feasilbe to do so with a local index. So only
VuFind/Blacklight with a local Solr is out, if you want article search.

You _can_ load local content in a vendor-hosted index like
EDS/Primo/Summon. So plenty of people do choose a vendor-hosted index
product as their only discovery tool, including both local metadata and
vendor-provided metadata. As you suggest.

But some people want the increased control that a locally controlled Solr
index gives you, for the local metadata where it's feasible. So use a
local
index product. But still want the article search you can get with a
vendor-hosted index product. So they use both.

There is also at least some reasons to believe that our users don't mind
and may even prefer having local results and hosted metadata results
presented seperately (although probably preferably in a consistent UI),
rather than merged.

A bunch more discussion of these issues is included in my blog post at:
http://bibwild.wordpress.com/2012/10/02/article-search-improvement-strateg
y/

From: Code for Libraries [CODE4LIB@LISTSERV.ND.EDU] on behalf of Wayne
Lam [
wing...@gmail.com]
Sent: Thursday, January 31, 2013 9:31 PM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: [CODE4LIB] Why we need multiple discovery services engine?

Hi all,

I saw in numerous of library website, many of them would have their own
based discovery 

Re: [CODE4LIB] Linked data [was: Why we need multiple discovery services engine?]

2013-02-04 Thread Ross Singer
On Feb 4, 2013, at 10:34 AM, Donna Campbell dcampb...@wts.edu wrote:

 In mentioning pushing to break down silos more, it brings to mind a
 question I've had about linked data.
 
 From what I've read thus far, the idea of breaking down silos of
 information seems like a good one in that it makes finding information
 easier but doesn't it also remove some of the markers of finding credible
 sources? Doesn't it blend accurate sources and inaccurate sources?

Provenance is especially important in this context, which I think is a crucial 
role that libraries can play.

-Ross.

 
 
 Donna R. Campbell
 Technical Services  Systems Librarian
 (215) 935-3872 (phone)
 (267) 295-3641 (fax)
 Mailing Address (via USPS):
 Westminster Theological Seminary Library
 P.O. Box 27009
 Philadelphia, PA 19118  USA
 Shipping Address (via UPS or FedEx):
 Westminster Theological Seminary Library
 2960 W. Church Rd.
 Glenside, PA 19038  USA
 
 -Original Message-
 From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of
 Emily Lynema
 Sent: Monday, February 04, 2013 9:56 AM
 To: CODE4LIB@LISTSERV.ND.EDU
 Subject: Re: [CODE4LIB] Why we need multiple discovery services engine?
 
 Here at NCSU, we use our locally-hosted Endeca service for our catalog
 and Serials
 Solutions Summon as an article search solution. Why do this?
 
 1. Our next-gen library catalog (Endeca) came first. This was before Solr
 hit
 the library world, and before library vendors started working on
 improving their bundled catalog apps. Our bundled catalog was terrible,
 and we
 wanted something better. This was back in the day when everyone was doing
 federated search for articles (think MetaLib).
 
 2. 4-5 years down the road, a number of vendors (Ebsco, Serials
 Solutions, etc.)
 were getting into the web scale discovery business. Aka, one big index
 that
 includes everything, in particular the citation content that libraries
 have
 historically not had local access to index / search. We bought Summon to
 solve the article search problem that federated searching never resolved
 for us. We wanted one access point for less experienced users who needed
 to
 find articles. Since we had backed away from federated search for
 articles,
 this was our big pain point; we already had a catalog we liked.
 
 We've actually loaded our catalog content into Summon, as well. So why
 keep both?
 We've done a LOT of work adding functionality into our local catalog,
 including
 enhanced title searching,lots of supplemental content, a quite complex
 local requesting system. So we can't just switch to the Summon interface
 without some effort.
 
 In addition, we have found that we prefer the bento box approach to
 searching across formats, as opposed to the integrated index approach
 of Summon.
 At least at this moment. We use this in the search across our library
 website [1]. It's just really, really hard to always surface the
 right kind of thing the user is looking for when the things you're
 indexing are
 different in nature (ex: bibliographic record vs. full-text of
 newspaper article). With the bento box approach, you have better
 opportunities to surface the different types of content available, while
 still having local systems optimized for specific content types.
 
 Maybe that's a long-winded excuse for not pushing to break down silos
 more. Time
 will probably tell.
 
 -emily
 
 [1] http://www.lib.ncsu.edu/search/?q=java


Re: [CODE4LIB] usability testing software

2013-02-04 Thread Nate Hill
Thanks everyone for your responses!
Lots to play with.
N

On Mon, Feb 4, 2013 at 9:41 AM, Rogers, Nathan roger...@indiana.edu wrote:

 Depending what your budget there are a number of good web based services
 out there. This breakdown is pretty good although I only have firsthand
 experience with WebSort (free version) and Crazyegg. If all you need to
 capture is audio/video I know some people that have made due with
 Camtasia, which is available for OS X.

 On 1/31/13 10:35 AM, Nate Hill nathanielh...@gmail.com wrote:

 Hi all,
 Years ago I had the opportunity to use Morae to do some usability testing.
 http://www.techsmith.com/morae.html
 I may have an opportunity to put together a little bit of a usability
 testing lab at my library, and I wonder if anyone can suggest a similar
 product but...
 I'd like it to run on Macs.
 Suggestions?
 thanks
 
 --
 Nate Hill
 nathanielh...@gmail.com
 http://4thfloor.chattlibrary.org/
 http://www.natehill.net




-- 
Nate Hill
nathanielh...@gmail.com
http://4thfloor.chattlibrary.org/
http://www.natehill.net


Re: [CODE4LIB] conf presenters: a kind request

2013-02-04 Thread Bill Dueber
I'm gonna add to this briefly, and probably a bit less tactfully than
Jonathan :-)

   - My number-one complaint about past presentations: Don't have slides we
   can't read. You probably can't read this, but... isn't a helpful thing to
   hear during a presentation. Make it legible, or figure out a different way
   to present the information. A kick-ass poster or UML diagram or flowchart
   or whatever isn't kick-ass when we can't read it. It's just an
   uninformative blur.  [Note: this doesn't mean you shouldn't include the
   kick-ass poster when you upload your slides. Please do!]
   - Make sure your content fits well in the time allotted. You're not
   there to get through as much as possible. You're there to best use our
   collective time to make the argument that what you're doing is
   important/impressive/worth knowing, and to convey *as much of the
   interesting bits as you can without rushing*. The goal isn't for you to
   get lots of words out of your mouth; the goal is for us to understand them.
   If you absolutely can't cut it down to a point where you're not rushing,
   then you haven't done the hard work of distilling out the interesting bits,
   and you should get on that right away.
   - On the flip side, don't present for 8mn and leave plenty of time for
   questions. Odds are your'e not saying anything interesting enough to
   elicit questions in those 8 minutes. If you really only have 8mn of
   content, well, you shouldn't have proposed a talk. But odds are you *do*
   have interesting things to say, and may want to chat with your colleagues
   to figure out exactly what that is.
   - Don't make the 3.38 million messages on creating a non-threatening
   environment be for naught. Please.

As Jonathan said: this is a great, great audience. We're all forgiving,
we're all interested, we're all eager to lean new things and figure out how
to apply them to our own situations. We love to hear about your successes.
We *love* to hear about failures that include a way for us to avoid them,
and you're going to be well-received no matter what because a bunch of
people voted to hear you!





On Mon, Feb 4, 2013 at 10:47 AM, Jonathan Rochkind rochk...@jhu.edu wrote:

 We are all very excited about the conference next week, to speak to our
 peers and to hear what our peers have to say!

 I would like to suggest that those presenting be considerate to your
 audience, and actually prepare your talk in advance!

 You may think you can get away with making some slides that morning during
 someone elses talk and winging it; nobody will notice right? Or they wont'
 care if they do?

 From past years, I can say that for me at least, yeah, I can often tell
 who hasn't actually prepared their talk. And I'll consider it disrespectful
 to the time of the audience, who voted for your talk and then got on
 airplanes to come see it, and you didn't spend the time to plan it advance
 and make it as high quality for them as you could.

 I don't mean to make people nervous about public speaking. The code4lib
 audience is a very kind and generous audience, they are a good audience.
 It'll go great! Just maybe repay their generosity by actually preparing
 your talk in advance, you know?  Do your best, it'll go great!

 If you aren't sure how to do this, the one thing you can probably do to
 prepare (maybe this is obvious) is practice your presentation in advance,
 with a timer, just once.  In front of a friend or just by yourself. Did you
 finish on time, and get at least half of what was important in? Then you're
 done preparing, that was it!  Yes, if you're going to have slides, this
 means making your slides or notes/outline in advance so you can practice
 your delivery just once!

 Just practice it once in advance (even the night before, as a last
 resort!), and it'll go great!

 Jonathan




-- 
Bill Dueber
Library Systems Programmer
University of Michigan Library


[CODE4LIB] Job: Librarian for Assessment at Yale University Library

2013-02-04 Thread jobs
**Librarian for Assessment  
Program Development  Research

Yale University Library

Librarian 1-3**

  
  
Schedule: Full-time (37.5 hours per week); Standard Work Week (M-F, 8:30-5:00)

  
  
Yale University offers exciting opportunities for achievement and growth in
New Haven, Connecticut. Conveniently located between Boston
and New York, New Haven is the creative capital of Connecticut with cultural
resources that include two major art museums, a critically acclaimed repertory
theater, state-of-the-art concert hall, and world-renowned schools of
Architecture, Art, Drama, and Music.

  
  
  
**The University and the Library**  
  
The Yale University Library, as one of the world's leading research libraries,
collects, organizes, preserves, and provides access to and services for a rich
and unique record of human thought and creativity. It fosters intellectual
growth and supports the teaching and research missions of Yale University and
scholarly communities worldwide. A distinctive strength is its rich spectrum
of resources, including around 12.8 million volumes and information in all
media, ranging from ancient papyri to early printed books to electronic
databases. The Library is engaging in numerous projects to expand access to
its physical and digital collections. Housed in eighteen buildings including
the Sterling Memorial Library, the Beinecke Rare Book and Manuscript Library,
and the Bass Library, it employs a dynamic and diverse staff of approximately
five hundred who offer innovative and flexible services to library
readers. For additional information on the Yale University
Library, please visit the Library's web site at www.library.yale.edu.

  
  
  
**Position Focus**  
  
The Librarian for Assessment provides leadership and vision for assessment,
measurement, planning and analysis throughout the Yale University Library
(YUL) and strives to ensure that these activities are integral parts of the
Library's programs, services, and collections. The Librarian for Assessment
oversees and participates in assessment efforts throughout YUL; serves as an
internal consultant for data-gathering and assessment activities conducted by
other Library staff; works with Library colleagues to analyze and report
assessment data; represents the Library in campus, regional and national
assessment efforts; evaluates the effectiveness of Library assessment efforts
and how they support the mission and strategic goals of the Library and the
University; and recommends ways to strengthen the Library's assessment and
measurement programs. The successful applicant for this position must possess
strong analytical skills, a firm understanding of Library services and work
processes, and strong interpersonal and listening skills. This position
reports to the Associate University Librarian for Program Development and
Research.

  
  
  
**Principal Responsibilities**  
  
1. Directs assessment efforts within the Yale University
Library. Initiates assessment activities and provides
consultation for assessment work done by other Library staff and
departments. Promotes awareness and communication of other
related assessment efforts within the Library, University, and externally.

  
2. Works with Library IT and other organizations within the
University and external to it to gather and create tools that enable managers
and staff to make data-driven decisions.

  
3. Works with the Director of Collection Development to
research, design, and test methodologies for collection assessment appropriate
to the needs of the Yale University Library.

  
4. Analyzes assessment-related data and communicates
assessment activities and results to appropriate individuals and groups,
including Library staff and the Yale campus community.

  
5. Develops and maintains expertise in assessment methods,
techniques and best practices.

  
6. Establishes training and documentation programs for
Library staff on use of appropriate assessment tools and methods.

  
7. Evaluates effectiveness of Library assessment activities
on a regular basis and makes recommendations on ways to strengthen assessment
work, including support needed. Develops and fosters a
culture of assessment within the Library.

  
8. Contributes to the profession and represents the Library
and the University in the academic, scholarly, and professional community.
Responds to, collaborates with and participates as appropriate in other
campus, regional and national assessment-related efforts.

  
  
  
**Required Education and Experience**  
  
Master's degree in Library Science from an American Library Association
accredited Library school. In selected instances, a post-graduate degree in a
related discipline may be required or substituted for an MLS. Appointment to
this rank is limited to two years at which time it is expected that the
individual will develop necessary requirements to meet expectations of
performance at the Librarian 2 level.

  
  
  
**Required Skills and Abilities** 

Re: [CODE4LIB] Why we need multiple discovery services engine?

2013-02-04 Thread Bill Dueber
Here at UMich, we do *not* yet load our catalog data into Summon. Partially
because it's not at all clear that the patrons want books and articles
mixed in that way, but mostly because we've worked really hard to
customize the way relevancy and indexing works in our catalog, and we don't
have that level of customization with Summon yet.

Believe me, if i thought I could get what the institution wants, I'd give
up writing/maintaining the catalog in a heartbeat.


On Mon, Feb 4, 2013 at 10:48 AM, Owen, Will o...@email.unc.edu wrote:

 I think this pretty accurately reflects the sentiments at the University
 of North Carolina, one of NSCU's partners in developing the Endeca
 interface to our catalog.  It's as much a mechanism for delivering
 services as it is for discovery, and as such reflects the physical nature
 of the collections it includes.  It's not possible at this point for us to
 build those services into the Summon environment.  +1 all around to
 Emily's comments.

 Will Owen

 On 2/4/13 9:56 AM, Emily Lynema emilylyn...@gmail.com wrote:

 Here at NCSU, we use our locally-hosted Endeca service for our catalog
 and Serials
 Solutions Summon as an article search solution. Why do this?
 
 1. Our next-gen library catalog (Endeca) came first. This was before Solr
 hit
 the library world, and before library vendors started working on
 improving their bundled catalog apps. Our bundled catalog was terrible,
 and we
 wanted something better. This was back in the day when everyone was doing
 federated search for articles (think MetaLib).
 
 2. 4-5 years down the road, a number of vendors (Ebsco, Serials
 Solutions, etc.)
 were getting into the web scale discovery business. Aka, one big index
 that
 includes everything, in particular the citation content that libraries
 have
 historically not had local access to index / search. We bought Summon to
 solve the article search problem that federated searching never resolved
 for us. We wanted one access point for less experienced users who needed
 to
 find articles. Since we had backed away from federated search for
 articles,
 this was our big pain point; we already had a catalog we liked.
 
 We've actually loaded our catalog content into Summon, as well. So why
 keep both?
 We've done a LOT of work adding functionality into our local catalog,
 including
 enhanced title searching,lots of supplemental content, a quite complex
 local requesting system. So we can't just switch to the Summon interface
 without some effort.
 
 In addition, we have found that we prefer the bento box approach to
 searching across formats, as opposed to the integrated index approach
 of Summon.
 At least at this moment. We use this in the search across our library
 website [1]. It's just really, really hard to always surface the
 right kind of thing the user is looking for when the things you're
 indexing are
 different in nature (ex: bibliographic record vs. full-text of
 newspaper article). With the bento box approach, you have better
 opportunities to surface the different types of content available, while
 still having local systems optimized for specific content types.
 
 Maybe that's a long-winded excuse for not pushing to break down silos
 more. Time
 will probably tell.
 
 -emily
 
 [1] http://www.lib.ncsu.edu/search/?q=java
 
 
 -
 
 Date:Fri, 1 Feb 2013 04:21:01 +
 From:Jonathan Rochkind rochk...@jhu.edu
 Subject: Re: Why we need multiple discovery services engine?
 
 So, there are two categories of solutions here -- 1) local indexes, where
 you create the index yourself, like blacklight or vufind (both based on a
 local Solr).  2) vendor-hosted indexes, where the vendor includes all
 sorts
 of things in their index that you the customer don't have local metadata
 for, mostly including lots and lots of scholarly article citations.
 
 If you want to include scholarly article citations, you probably can't do
 that with a local index solution. Although some consortiums have done some
 interesting stuff in that area, let's just say it takes a lot of resources
 to do. For most people, if you want to include article search in your
 index, it's not feasilbe to do so with a local index. So only
 VuFind/Blacklight with a local Solr is out, if you want article search.
 
 You _can_ load local content in a vendor-hosted index like
 EDS/Primo/Summon. So plenty of people do choose a vendor-hosted index
 product as their only discovery tool, including both local metadata and
 vendor-provided metadata. As you suggest.
 
 But some people want the increased control that a locally controlled Solr
 index gives you, for the local metadata where it's feasible. So use a
 local
 index product. But still want the article search you can get with a
 vendor-hosted index product. So they use both.
 
 There is also at least some reasons to believe that our users don't mind
 and may even prefer having local results and 

Re: [CODE4LIB] Jan. 9th Moving Up to the Cloud

2013-02-04 Thread Jimmy Ghaphery
Belated followup, we have posted a YouTube video of this event.

http://www.library.vcu.edu/blog/news/2013/01/video-library-technology-leade.html

thanks,

Jimmy


On Thu, Jan 3, 2013 at 1:17 PM, Jimmy Ghaphery jghap...@vcu.edu wrote:

 Please see the announcement below for the event on January 9th which will
 also be streamed.

 thanks,

 Jimmy

 *
 Library technology-science leaders to gather at VCU on Jan. 9 for Moving
 Up to the Cloud event*

 As one of the first major research universities to launch a pioneering new
 cloud-resident enterprise library platform, Virginia Commonwealth
 University is hosting an educational session Jan. 9, designed to help other
 libraries around the world prepare for the next generation of library
 technologies.

 VCU Libraries went live in the fall semester with Alma, cloud-resident
 enterprise library software from Ex Libris Inc. Alma is a unified
 resource-management system that replaced the system VCU had used for the
 past decade, which had also been developed by Ex Libris. VCU is the third
 library in North America to have launched Alma.

 Speakers at the Jan. 9 event:


- Marshall Breeding, the national authority on enterprise library
systems, is editor of Library Technology Guides and author of a monthly
column in Computers in Libraries.
- Mark Ryland, chief solutions architect for Amazon Web Service’s
Worldwide Public Sector team, began working with Microsoft Federal Systems
as a senior architectural engineer and has seen the industry grow up.
- Mark Triest, president of Ex Libris North America since 2010,
previously was president of Inigral, a social-media start-up delivering
applications for the education market. He also served as senior vice
president of global sales at Sunguard Higher Education.
- John Ulmschneider, university librarian at VCU, is a past president
and board member of the Association of Southeastern Research Libraries and
leads VCU’s growing and innovative research-library system.
- Gene Damon, panel moderator, is director of library automation for
the Virginia Community College System, one of the nation’s largest
community-college systems. He is chair of the steering committee for the
Virtual Library of Virginia (VIVA), and brings a distinguished background
of leadership in enterprise library systems in both corporate and academic
environments.

 An audience of Virginia and Washington-area academic library leaders will
 participate in two panel sessions, which will be streamed live on the Web
 and posted for viewing by interested parties worldwide.

 Get details, register to attend, or watch 
 livehttp://www.library.vcu.edu/events/cloud/


 --
 Jimmy Ghaphery
 Head, Library Information Systems
 VCU Libraries
 804-827-3551




-- 
Jimmy Ghaphery
Head, Library Information Systems
VCU Libraries
804-827-3551


Re: [CODE4LIB] conf presenters: a kind request

2013-02-04 Thread Joe Hourcle
On Feb 4, 2013, at 11:25 AM, Bill Dueber wrote:

[trimmed (and agreed with all of that)]

 As Jonathan said: this is a great, great audience. We're all forgiving,
 we're all interested, we're all eager to lean new things and figure out how
 to apply them to our own situations. We love to hear about your successes.
 We *love* to hear about failures that include a way for us to avoid them,
 and you're going to be well-received no matter what because a bunch of
 people voted to hear you!

I'd actually be interested in people's complaints about bad presentations;
I've been keeping notes for years, with the intention of making a
presentation on giving better presentations.  (but it's much harder than
it sounds, as I plan on making all of the mistakes during the presentation)


 On Mon, Feb 4, 2013 at 10:47 AM, Jonathan Rochkind rochk...@jhu.edu wrote:
 
 We are all very excited about the conference next week, to speak to our
 peers and to hear what our peers have to say!
 
 I would like to suggest that those presenting be considerate to your
 audience, and actually prepare your talk in advance!

[trimmed]

 Just practice it once in advance (even the night before, as a last
 resort!), and it'll go great!


I did one of those 'Ignite' talks this year; because it's auto-
advancing slides, I went over it multiple times.  My recommendation
is that you try to get various co-workers as guinea pigs.  I even
subjected one of my neighbors to it, even though he wasn't necessarily
part of the intended audience.

They gave me a lot of feed back -- asking for clarification on bits,
we realized I could trim down a couple of slides, giving me more
slides to expand other bits.  I still screwed up the presentation,
but it would have been much worse if I hadn't practiced.

My local ASIST chapter used to run 'preview' events before the
annual meeting, where the local folks presenting at annual were
invited to give their talks.  If nothing else, it forced you to
have it done a couple of weeks early, but more importantly, it
gave me a chance to have a similar audience to what would be
at the main meeting ... one of my talks bombed hard;  it was on
standards  protocols for scientific data, and I hadn't considered
just how bad a talk that's 50% acronyms would go over.  I was
able to change how I presented the material so it wasn't quite
so painful the second time around.

There's only been once when practicing in advanced made for a worse
presentation ... and that's because when I finished, PowerPoint asked
me if I wanted to save the timings ... what ever you do, do *not*
tell it yes.  Because then it'll auto-advance your slides, so when
you skip over one slide during the practice, it'll not let you
have it up during the real talk.

(There's a setting to turn off use of timings ... and the audience
laughed when I kept scolding the computer, but it still felt
horrible when I was up there)

And it's important that you *must* practice in front of other
people.  How fast you think it's going to take you, or how fast
it takes you talking to yourself is nothing like talking in
front of other people.

...

So, all of that being said, some of the things I've made a note
of over the years.  (it's incomplete, as I've still take notes
by hand, and there are more items on the back pages of the 
various memo books I've had over the years)

* Get there before the session, and test your presentation on the
  same hardware as it's going to be presented from.  This is
  especially important if you're a Mac user, and presenting from
  a PC, or visa-versa.  Look for odd fonts, images that didn't
  load, videos, abnormal gamma, bad font sizes (may result in
  missing test), missing characters, incorrect justification, etc.

* If you're going to be presenting from your own machine, still
  test it out, to make sure that you have all of the necessary
  adaptors, that you know what needs to be done to switch the
  monitor, that the machine detects the projector at a reasonable
  size and the gamma's adjusted correctly.  (and have it loaded
  in advance; you're wasting enough time switching machines).
  And start switching machines while the last presenter's doing
  QA ... and if you lose 5 min because of switching, prepare
  to cut your talk short, force the following presenters to lose
  time)

* Have a backup plan, with the presentation stashed on a website
  that you've memorized the URL to, *and* on a USB stick.
  (website is safer vs. virus transfer, only use the USB stick
  if there's no internet)  And put the file at the top level of
  the USB stick, not buried 12 folders deep.

* If they have those clip on microphones, put it on your label
  on the same side as the screen is to you.  (so whenever you
  turn to look at the screen, it still picks up your voice)

* If you have a stationary mic, you have to actually stay near
  it or it doesn't work.

* Hand-held mics suck unless you're used to them, as most of us
  aren't used to holding our hand up 

Re: [CODE4LIB] Anyone have a SUSHI client?

2013-02-04 Thread Joshua Welker
Tom,

I'm glad you are getting some help from looking at my code. That was the reason 
I put it up. When I was designing a SUSHI client last year, there was virtually 
nothing available to show how to do it, so I was hoping sushi.py would prevent 
someone from having to stumble in the dark quite as much.

Funny you mention your organization wanting you to use Sharepoint. We also 
originally used Sharepoint for storing some of our operational data, but in the 
end, Sharepoint is such a closed environment that it wasn't able to meet our 
needs. However, even if you are 100% locked into Sharepoint, you should still 
be able to use Python code to access the SUSHI web services and then store it 
in Sharepoint using some of its APIs (if you can wade through Sharepoint's docs 
enough to figure out those APIs). 

Are you familiar with XAMPP? It's a light-weight Windows package that lets you 
run MySQL and some other stuff on a desktop computer very easily. Even if you 
don't have access to an official web server from your institution, you could 
still run sushi.py just on your local computer and have it store the data in 
your local MySQL database. Then you could just create some backups of that 
database and store them as files on Sharepoint or a shared drive or whatever 
method you have available. Just a thought.

Josh Welker

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Tom 
Keays
Sent: Monday, February 04, 2013 9:39 AM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Anyone have a SUSHI client?

Hi Joshua,

I was mainly looking at your program, not for the code, but as a way to bring 
myself up to speed about current practices in modeling the COUNTER data. I'm 
trying to avoid reinventing something that has already been well thought 
through. I apologize for calling out your model. You have gotten much further 
than I have. Some of the other respondents in this thread have set me straight 
on some things I was very fuzzy on going in.

How go about I collecting and storing the data is still something I haven't 
resolved yet. I personally would prefer a Python solution, but there forces 
here at MPOW that suggest I should build a data repository in SharePoint.
Assuming that is the case, Serial Solution's open source SUSHI harvester 
written in .NET might actually be the way for me to go. So, my next step is to 
look at their data model and see what reports they collect and store.

As an aside, I'm also now wondering if de-duping is strictly necessary as long 
as there is a field to record the date the report was generated.
 De-duping (or maybe just deprecating duplicate data) could be separate from 
the collection process.

Best,
Tom

On Mon, Feb 4, 2013 at 10:07 AM, Joshua Welker jwel...@sbuniv.edu wrote:

 Tom,

 I am the guy who wrote sushi.py around this time last year. My 
 apologies for the shabbiness of the code. It was meant to be primarily 
 a proof of concept. It's definitely incomplete. I only completed the 
 DB3 and JR1 report logic up to this point, but it would be easy enough 
 to add other report types. You're also right that sushi.py doesn't do 
 anything to dedupe data, but it would be very simple to write a script 
 that reads through the SQL records and deletes dupes. You could also 
 use the built-in UNIQUE flag in MySQL when creating your table so that 
 duplicate records just don't get saved. If you use the CSV export 
 functionality of sushi.py, Excel has some built-in dedupe features that would 
 help as well.

 Let me know if you'd like some help modifying sushi.py. I sort of gave 
 up on it last spring. SUSHI implementation among vendors is still 
 pretty shabby, and there are still some weaknesses in the SUSHI 
 standard (I wrote about them in the Nov 2012 issue of Computers in 
 Libraries). The productivity gains I was seeing from using SUSHI ended up 
 being pretty low.

 Josh Welker


 -Original Message-
 From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf 
 Of Tom Keays
 Sent: Friday, January 25, 2013 8:40 AM
 To: CODE4LIB@LISTSERV.ND.EDU
 Subject: Re: [CODE4LIB] Anyone have a SUSHI client?

 I've been looking briefly at sushi.py, as a way to orient myself to 
 collecting stats this way. I'm not intending to single out sushi.py, 
 but looking at it (mainly the data structure at this point, and not 
 the code itself), raises some questions about the best approach for 
 collecting SUSHI data.

 sushi.py seems to have a small number of routines; mainly to retrieve 
 the XML file from a vendor and ingest the data in that file into a 
 MySQL database. There are only MySQL tables for COUNTER JR1, DR1, DR2, 
 and DR2 reports and they mirror, to a degree, the structure of the 
 item records returned in the SUSHI xml. Here are the skeletons of 2 of 
 the sushi.py SQL
 tables:

 counter_jr1
   id int,
   print_issn varchar,
   online_issn varchar,
   platform varchar,
   item_name text,
   data_type varchar,
 

Re: [CODE4LIB] Linked data [was: Why we need multiple discovery services engine?]

2013-02-04 Thread Joe Hourcle
On Feb 4, 2013, at 10:34 AM, Donna Campbell wrote:

 In mentioning pushing to break down silos more, it brings to mind a
 question I've had about linked data.
 
 From what I've read thus far, the idea of breaking down silos of
 information seems like a good one in that it makes finding information
 easier but doesn't it also remove some of the markers of finding credible
 sources? Doesn't it blend accurate sources and inaccurate sources?

Yes, yes it does.

The 'intelligence' community has actually been talking about this
problem with RDF for years.  My understanding is that they use
RDF quads (not triples) so that they have an extra parameter to
track the source.  (it might be that they use something larger
than a quad).

From what I remember (the conversation was years ago), they
have to be able to mark information as suspect (eg, they find
that one of the sources is unreliable, then re-run all all of
the analysis without that source's contribution to determine
if they came to the same result).

I don't know enough about the implementation of linked data
systems, so if there's some way to filter which sources are
considered for input, or if there's any tracking of the RDF
triples once they're parsed out.


-Joe


[CODE4LIB] kickstarter for a homebrow project, from mistym

2013-02-04 Thread Jonathan Rochkind

homebrew is what many of us use to install code packages on OSX.

mistym (Misty De Meo) is a frequent hanger outer in #code4lib, and has 
helped many of us with stuff (including homebrew). She's also one of the 
homebrew maintainers.


Here's a kickstarter from mistym, to help homebrew with some automated 
testing infrastructure, to better handle user code submissions.


It's notable that they're asking for money for actual equipment, not 
developer time reimbursement or anything. They're not asking for much.


Consider kicking in some bucks if you feel like it!

http://www.kickstarter.com/projects/homebrew/brew-test-bot


[CODE4LIB] Job: Electronic Resources / Reference Librarian at Boston Athenæum

2013-02-04 Thread jobs
The Boston Athenaeum seeks a librarian fluent in emerging technologies to
manage electronic resources and serve as part of team that covers the
reference desk. This full-time exempt position reports to the Head of Reader
Services and supports the work of the Boston Athenaeum across all departments.

  
Responsibilities:

  * Manage electronic resources; including acquisition, licensing, processing, 
and maintenance 
* Coordinate with serials acquisition in Technical Services Department.
* Oversee collection development (working with subject bibliographers) and 
negotiate purchase with vendor or consortium liaison
* Maintain electronic resources webpage
* Participate in project to digitize finding aids
* Serve on committee that maintains website
* Work with all departments' collection management systems
  * Provide general and specialized reference assistance and occasionally work 
at circulation
  * Work some evenings and Saturdays in scheduled rotation among reference staff
  * Offer public presentations of reference services to members
  * Serve on a weekly basis in special collections room
  * Must be able to lift 40 pounds
  * Other duties as assigned
Required Qualifications:

  
Education

  * MLS from an ALA-accredited institution
Experience

  * Experience offering public services in a library, using cataloging software 
(preferably OCLC), content management systems (e.g., Drupal), digital 
collection management software (e.g., ContentDM and The Museum System), email 
marketing software (e.g., Constant Contact)
  * Experience editing and maintaining webpages and websites, proficiency with 
HTML and CSS
  * Proficiency with Microsoft Office; experience with image editing
Preferred Qualifications:

  * Familiarity with visual resource librarianship, Drupal, EAD, Raiser's Edge, 
proxy servers
  * Experience dealing with vendors
  * Experience editing text, knowledge of Adobe Photoshop
  * Familiarity with PHP, XSL and XML
Position is available immediately.

  
For further information see [www.bostonathenaeum.org/node/39](http://www.bosto
nathenaeum.org/node/39). E-mail resume and cover letter to
h...@bostonathenaeum.org.



Brought to you by code4lib jobs: http://jobs.code4lib.org/job/5987/


[CODE4LIB] Job: E-Resources Librarian and Web and Digital Services Librarian (2 positions) (CUNY Queens College, New York) at CUNY Queens College

2013-02-04 Thread jobs
The Queens College Libraries (City University of New York) are seeking
candidates for two positions: E-Resources Librarian and Web and Digital
Services Librarian. E-Resources Librarian sought by the Benjamin S. Rosenthal
Library at Queens College, The City University of New York
([http://qcpages.qc.cuny.edu/library](http://qcpages.qc.cuny.edu/library)).
Anticipated start date is July 1, 2013. Reporting to the Associate Librarian
for Public Services, the E-Resources Librarian effectively coordinates and
leads e-resources activities which include management, budgeting, policy
development, statistical and qualitative reporting and improvement in a high-
tech multi-disciplinary library.

  
For further information on the position (See Position 6560) and how to apply: 
[http://www.qc.cuny.edu/HR/Pages/JobListings.aspx](http://www.qc.cuny.edu/HR/P
ages/JobListings.aspx) Web  Digital Services Librarian sought by the Benjamin
S. Rosenthal Library at Queens College, the City University of New York
([http://qcpages.qc.cuny.edu/library](http://qcpages.qc.cuny.edu/library)).
Anticipated start date is July 1, 2013. The Web  Digital Services Librarian
is responsible as point person and coordinator for leading the ongoing
activities required to maintain useful, accessible and user-friendly delivery
of the Queens College Libraries digital library services. The Web  Digital
Services Librarian is responsible for design and delivery of a suite of
library web sites and augmentative digital services which are reflective of
current standards and best practices in academic library web site design,
current pedagogical technologies and teaching and learning trends. The
successful candidate reports to the Associate Librarian for Public Services
and will be a member of a digital library team, working closely with the
E-Resources Librarian, Library Systems Office staff and the campus and CUNY
technology units' members. The position may encompass supervision. For further
information on the position (See Position 6559) and how to apply: [http://www.
qc.cuny.edu/HR/Pages/JobListings.aspx](http://www.qc.cuny.edu/HR/Pages/JobList
ings.aspx)

  
EMPLOYMENT OPPORTUNITY We are committed to enhancing our diverse academic
community by actively encouraging people with disabilities, minorities,
veterans, and women to apply. We take pride in our pluralistic community and
continue to seek excellence through diversity and inclusion. EO/AA Employer



Brought to you by code4lib jobs: http://jobs.code4lib.org/job/6021/