Re: [CODE4LIB] Rdio playlist
Last time I took a road trip/weekend trip to Chicago I spun up a small Spotify playlist of Chicago-based bands that included: Telefon Tel Aviv Smashing Pumpkins Ministry Pigface Disappears Secret Colours Smith Westerns Screaching Weasel Wilco Most, but not all, are in-line with my tastes, so I know there are lots more Chicago-based artists that could be added. I'm not on Rdio or going to Code4Lib, but may get on and listen to your playlist. If nothing else I'll fire up my old playlist and be there in spirit. On Sun, Feb 3, 2013 at 10:44 PM, William Denton w...@pobox.com wrote: Do you ever put together a playlist or some situationally appropriate music when you're taking a trip? Do you ever do this when you're going to a conference? If you answered yes to either, and you're on Rdio, pitch in and add some stuff to a Code4Lib Chicago playlist: http://www.rdio.com/people/**wdenton/playlists/2229053/** Code4Lib_2013_in_Chicago/http://www.rdio.com/people/wdenton/playlists/2229053/Code4Lib_2013_in_Chicago/ I reckon we should be able to collaboratively edit the perfect playlist for use by hundreds of librarians and coders descending on Chicago in February, useful when either a) drinking local beers, b) hacking, c) walking through the Art Institute of Chicago, or d) all of the above. Bill -- William Denton Toronto, Canada http://www.miskatonic.org/ -- Matt Schultz Program Manager Educopia Institute, MetaArchive Cooperative http://www.metaarchive.org matt.schu...@metaarchive.org 616-566-3204
Re: [CODE4LIB] Why we need multiple discovery services engine?
Here at NCSU, we use our locally-hosted Endeca service for our catalog and Serials Solutions Summon as an article search solution. Why do this? 1. Our next-gen library catalog (Endeca) came first. This was before Solr hit the library world, and before library vendors started working on improving their bundled catalog apps. Our bundled catalog was terrible, and we wanted something better. This was back in the day when everyone was doing federated search for articles (think MetaLib). 2. 4-5 years down the road, a number of vendors (Ebsco, Serials Solutions, etc.) were getting into the web scale discovery business. Aka, one big index that includes everything, in particular the citation content that libraries have historically not had local access to index / search. We bought Summon to solve the article search problem that federated searching never resolved for us. We wanted one access point for less experienced users who needed to find articles. Since we had backed away from federated search for articles, this was our big pain point; we already had a catalog we liked. We've actually loaded our catalog content into Summon, as well. So why keep both? We've done a LOT of work adding functionality into our local catalog, including enhanced title searching,lots of supplemental content, a quite complex local requesting system. So we can't just switch to the Summon interface without some effort. In addition, we have found that we prefer the bento box approach to searching across formats, as opposed to the integrated index approach of Summon. At least at this moment. We use this in the search across our library website [1]. It's just really, really hard to always surface the right kind of thing the user is looking for when the things you're indexing are different in nature (ex: bibliographic record vs. full-text of newspaper article). With the bento box approach, you have better opportunities to surface the different types of content available, while still having local systems optimized for specific content types. Maybe that's a long-winded excuse for not pushing to break down silos more. Time will probably tell. -emily [1] http://www.lib.ncsu.edu/search/?q=java - Date:Fri, 1 Feb 2013 04:21:01 + From:Jonathan Rochkind rochk...@jhu.edu Subject: Re: Why we need multiple discovery services engine? So, there are two categories of solutions here -- 1) local indexes, where you create the index yourself, like blacklight or vufind (both based on a local Solr). 2) vendor-hosted indexes, where the vendor includes all sorts of things in their index that you the customer don't have local metadata for, mostly including lots and lots of scholarly article citations. If you want to include scholarly article citations, you probably can't do that with a local index solution. Although some consortiums have done some interesting stuff in that area, let's just say it takes a lot of resources to do. For most people, if you want to include article search in your index, it's not feasilbe to do so with a local index. So only VuFind/Blacklight with a local Solr is out, if you want article search. You _can_ load local content in a vendor-hosted index like EDS/Primo/Summon. So plenty of people do choose a vendor-hosted index product as their only discovery tool, including both local metadata and vendor-provided metadata. As you suggest. But some people want the increased control that a locally controlled Solr index gives you, for the local metadata where it's feasible. So use a local index product. But still want the article search you can get with a vendor-hosted index product. So they use both. There is also at least some reasons to believe that our users don't mind and may even prefer having local results and hosted metadata results presented seperately (although probably preferably in a consistent UI), rather than merged. A bunch more discussion of these issues is included in my blog post at: http://bibwild.wordpress.com/2012/10/02/article-search-improvement-strategy/ From: Code for Libraries [CODE4LIB@LISTSERV.ND.EDU] on behalf of Wayne Lam [ wing...@gmail.com] Sent: Thursday, January 31, 2013 9:31 PM To: CODE4LIB@LISTSERV.ND.EDU Subject: [CODE4LIB] Why we need multiple discovery services engine? Hi all, I saw in numerous of library website, many of them would have their own based discovery services (e.g. blacklight / vufind) and at the same time they will have vendor based discovery services (e.g. EDS / Primo / Summon). Instead of having to maintain 2 separate system, why not put everything into just one? Any special reason or concern? Best Wayne -- Emily Lynema Associate Department Head Information Technology, NCSU Libraries 919-513-8031 emily_lyn...@ncsu.edu
Re: [CODE4LIB] Anyone have a SUSHI client?
Tom, I am the guy who wrote sushi.py around this time last year. My apologies for the shabbiness of the code. It was meant to be primarily a proof of concept. It's definitely incomplete. I only completed the DB3 and JR1 report logic up to this point, but it would be easy enough to add other report types. You're also right that sushi.py doesn't do anything to dedupe data, but it would be very simple to write a script that reads through the SQL records and deletes dupes. You could also use the built-in UNIQUE flag in MySQL when creating your table so that duplicate records just don't get saved. If you use the CSV export functionality of sushi.py, Excel has some built-in dedupe features that would help as well. Let me know if you'd like some help modifying sushi.py. I sort of gave up on it last spring. SUSHI implementation among vendors is still pretty shabby, and there are still some weaknesses in the SUSHI standard (I wrote about them in the Nov 2012 issue of Computers in Libraries). The productivity gains I was seeing from using SUSHI ended up being pretty low. Josh Welker -Original Message- From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Tom Keays Sent: Friday, January 25, 2013 8:40 AM To: CODE4LIB@LISTSERV.ND.EDU Subject: Re: [CODE4LIB] Anyone have a SUSHI client? I've been looking briefly at sushi.py, as a way to orient myself to collecting stats this way. I'm not intending to single out sushi.py, but looking at it (mainly the data structure at this point, and not the code itself), raises some questions about the best approach for collecting SUSHI data. sushi.py seems to have a small number of routines; mainly to retrieve the XML file from a vendor and ingest the data in that file into a MySQL database. There are only MySQL tables for COUNTER JR1, DR1, DR2, and DR2 reports and they mirror, to a degree, the structure of the item records returned in the SUSHI xml. Here are the skeletons of 2 of the sushi.py SQL tables: counter_jr1 id int, print_issn varchar, online_issn varchar, platform varchar, item_name text, data_type varchar, date_begin datetime, date_end datetime, ft_pdf int, ft_html int, ft_total varchar counter_db3 id int, platform varchar, item_name text, data_type varchar, date_begin datetime, date_end datetime, searches int, sessions int On the face of it, this seems like a pretty good data structure (although I have a couple of concerns, that I will get to) but my main question is whether there is any agreement about how to collect this data? If I were to dig into some of the other SUSHI packages mentioned in this thread, what would I find there? Excel-formatted COUNTER reports are simply a table of columns, representing various fields, such as title (for JR1), platform, publisher (for JR1), ISSN (for JR1), etc., followed by columns for up to 12 months of the collected year, and then summary data. JR1 reports have fulltext HTML, PDF, and Total columns. DR1 has two rows, one for searches and one for sesssions, with YTD totals in the final column. Similar data structures exist for other COUNTER reports. They rely on the user to interpret them and probably ought not to inform a decision for structuring the data in a database. Is there been any best practice for how COUNTER data is modeled in a database? There are other COUNTER reports besides those four. For instance, some journal vendors do indeed report searches and sessions using the DR3 report, but others use the equivalent JR4 report, so I would have expected sushi.py to have a mechanism to collect these. Does SUSHI only deliver JR1, DR1, DR2, and DR2 reports, or is this a problem with sushi.py? Now, one of the selling points for SUSHI is that if a vendor ever advises that you should re-collect data for a given time period, the xml you receive is structured such that the act of collecting OUGHT TO update, rather than duplicate, data previously collected. However in sushi.py's SQL structure, which gives every row a unique (auto-incremented) ID number, there would have to be logic applied during the ingest to prevent multiple instances of data collected from the same vendor for the same time period. So, that's a concern. I'm also concerned about what is represented in the ft_pdf, ft_html, and ft_total fields. In the Excel COUNTER reports, the ft_pdf, ft_html, and ft_total columns simply tabulate the YTD totals and the only way you would be able to derive a monthly breakdown would be to collect 12 monthly reports and analyze the differences from month to month -- something that most libraries don't do. I have to go back and confirm this, but I don't think the SUSHI reports are giving a month-only breakdown for those fields, so I wonder about their inclusion in that table. I guess my question is what is returned in the SUSHI xml report: monthly or yearly figures for the ft_pdf, ft_html, and ft_total fields?
Re: [CODE4LIB] Rdio playlist
Can't believe no one has yet mentioned Chicago band I Fight Dragons -- they mix NES-controller chiptunes with electric guitars, playing rockin' covers of Mario Bros., Legend of Zelda, Contra, etc. in addition to some originals that have been remixed by other Chicago musicians. They just got back from their War of Cyborg Liberation Tour.
[CODE4LIB] Linked data [was: Why we need multiple discovery services engine?]
In mentioning pushing to break down silos more, it brings to mind a question I've had about linked data. From what I've read thus far, the idea of breaking down silos of information seems like a good one in that it makes finding information easier but doesn't it also remove some of the markers of finding credible sources? Doesn't it blend accurate sources and inaccurate sources? Donna R. Campbell Technical Services Systems Librarian (215) 935-3872 (phone) (267) 295-3641 (fax) Mailing Address (via USPS): Westminster Theological Seminary Library P.O. Box 27009 Philadelphia, PA 19118 USA Shipping Address (via UPS or FedEx): Westminster Theological Seminary Library 2960 W. Church Rd. Glenside, PA 19038 USA -Original Message- From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Emily Lynema Sent: Monday, February 04, 2013 9:56 AM To: CODE4LIB@LISTSERV.ND.EDU Subject: Re: [CODE4LIB] Why we need multiple discovery services engine? Here at NCSU, we use our locally-hosted Endeca service for our catalog and Serials Solutions Summon as an article search solution. Why do this? 1. Our next-gen library catalog (Endeca) came first. This was before Solr hit the library world, and before library vendors started working on improving their bundled catalog apps. Our bundled catalog was terrible, and we wanted something better. This was back in the day when everyone was doing federated search for articles (think MetaLib). 2. 4-5 years down the road, a number of vendors (Ebsco, Serials Solutions, etc.) were getting into the web scale discovery business. Aka, one big index that includes everything, in particular the citation content that libraries have historically not had local access to index / search. We bought Summon to solve the article search problem that federated searching never resolved for us. We wanted one access point for less experienced users who needed to find articles. Since we had backed away from federated search for articles, this was our big pain point; we already had a catalog we liked. We've actually loaded our catalog content into Summon, as well. So why keep both? We've done a LOT of work adding functionality into our local catalog, including enhanced title searching,lots of supplemental content, a quite complex local requesting system. So we can't just switch to the Summon interface without some effort. In addition, we have found that we prefer the bento box approach to searching across formats, as opposed to the integrated index approach of Summon. At least at this moment. We use this in the search across our library website [1]. It's just really, really hard to always surface the right kind of thing the user is looking for when the things you're indexing are different in nature (ex: bibliographic record vs. full-text of newspaper article). With the bento box approach, you have better opportunities to surface the different types of content available, while still having local systems optimized for specific content types. Maybe that's a long-winded excuse for not pushing to break down silos more. Time will probably tell. -emily [1] http://www.lib.ncsu.edu/search/?q=java
Re: [CODE4LIB] Anyone have a SUSHI client?
Hi Joshua, I was mainly looking at your program, not for the code, but as a way to bring myself up to speed about current practices in modeling the COUNTER data. I'm trying to avoid reinventing something that has already been well thought through. I apologize for calling out your model. You have gotten much further than I have. Some of the other respondents in this thread have set me straight on some things I was very fuzzy on going in. How go about I collecting and storing the data is still something I haven't resolved yet. I personally would prefer a Python solution, but there forces here at MPOW that suggest I should build a data repository in SharePoint. Assuming that is the case, Serial Solution's open source SUSHI harvester written in .NET might actually be the way for me to go. So, my next step is to look at their data model and see what reports they collect and store. As an aside, I'm also now wondering if de-duping is strictly necessary as long as there is a field to record the date the report was generated. De-duping (or maybe just deprecating duplicate data) could be separate from the collection process. Best, Tom On Mon, Feb 4, 2013 at 10:07 AM, Joshua Welker jwel...@sbuniv.edu wrote: Tom, I am the guy who wrote sushi.py around this time last year. My apologies for the shabbiness of the code. It was meant to be primarily a proof of concept. It's definitely incomplete. I only completed the DB3 and JR1 report logic up to this point, but it would be easy enough to add other report types. You're also right that sushi.py doesn't do anything to dedupe data, but it would be very simple to write a script that reads through the SQL records and deletes dupes. You could also use the built-in UNIQUE flag in MySQL when creating your table so that duplicate records just don't get saved. If you use the CSV export functionality of sushi.py, Excel has some built-in dedupe features that would help as well. Let me know if you'd like some help modifying sushi.py. I sort of gave up on it last spring. SUSHI implementation among vendors is still pretty shabby, and there are still some weaknesses in the SUSHI standard (I wrote about them in the Nov 2012 issue of Computers in Libraries). The productivity gains I was seeing from using SUSHI ended up being pretty low. Josh Welker -Original Message- From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Tom Keays Sent: Friday, January 25, 2013 8:40 AM To: CODE4LIB@LISTSERV.ND.EDU Subject: Re: [CODE4LIB] Anyone have a SUSHI client? I've been looking briefly at sushi.py, as a way to orient myself to collecting stats this way. I'm not intending to single out sushi.py, but looking at it (mainly the data structure at this point, and not the code itself), raises some questions about the best approach for collecting SUSHI data. sushi.py seems to have a small number of routines; mainly to retrieve the XML file from a vendor and ingest the data in that file into a MySQL database. There are only MySQL tables for COUNTER JR1, DR1, DR2, and DR2 reports and they mirror, to a degree, the structure of the item records returned in the SUSHI xml. Here are the skeletons of 2 of the sushi.py SQL tables: counter_jr1 id int, print_issn varchar, online_issn varchar, platform varchar, item_name text, data_type varchar, date_begin datetime, date_end datetime, ft_pdf int, ft_html int, ft_total varchar counter_db3 id int, platform varchar, item_name text, data_type varchar, date_begin datetime, date_end datetime, searches int, sessions int On the face of it, this seems like a pretty good data structure (although I have a couple of concerns, that I will get to) but my main question is whether there is any agreement about how to collect this data? If I were to dig into some of the other SUSHI packages mentioned in this thread, what would I find there? Excel-formatted COUNTER reports are simply a table of columns, representing various fields, such as title (for JR1), platform, publisher (for JR1), ISSN (for JR1), etc., followed by columns for up to 12 months of the collected year, and then summary data. JR1 reports have fulltext HTML, PDF, and Total columns. DR1 has two rows, one for searches and one for sesssions, with YTD totals in the final column. Similar data structures exist for other COUNTER reports. They rely on the user to interpret them and probably ought not to inform a decision for structuring the data in a database. Is there been any best practice for how COUNTER data is modeled in a database? There are other COUNTER reports besides those four. For instance, some journal vendors do indeed report searches and sessions using the DR3 report, but others use the equivalent JR4 report, so I would have expected sushi.py to have a mechanism to collect these. Does SUSHI only deliver JR1, DR1, DR2, and DR2 reports, or is this a problem with
[CODE4LIB] conf presenters: a kind request
We are all very excited about the conference next week, to speak to our peers and to hear what our peers have to say! I would like to suggest that those presenting be considerate to your audience, and actually prepare your talk in advance! You may think you can get away with making some slides that morning during someone elses talk and winging it; nobody will notice right? Or they wont' care if they do? From past years, I can say that for me at least, yeah, I can often tell who hasn't actually prepared their talk. And I'll consider it disrespectful to the time of the audience, who voted for your talk and then got on airplanes to come see it, and you didn't spend the time to plan it advance and make it as high quality for them as you could. I don't mean to make people nervous about public speaking. The code4lib audience is a very kind and generous audience, they are a good audience. It'll go great! Just maybe repay their generosity by actually preparing your talk in advance, you know? Do your best, it'll go great! If you aren't sure how to do this, the one thing you can probably do to prepare (maybe this is obvious) is practice your presentation in advance, with a timer, just once. In front of a friend or just by yourself. Did you finish on time, and get at least half of what was important in? Then you're done preparing, that was it! Yes, if you're going to have slides, this means making your slides or notes/outline in advance so you can practice your delivery just once! Just practice it once in advance (even the night before, as a last resort!), and it'll go great! Jonathan
Re: [CODE4LIB] Why we need multiple discovery services engine?
I think this pretty accurately reflects the sentiments at the University of North Carolina, one of NSCU's partners in developing the Endeca interface to our catalog. It's as much a mechanism for delivering services as it is for discovery, and as such reflects the physical nature of the collections it includes. It's not possible at this point for us to build those services into the Summon environment. +1 all around to Emily's comments. Will Owen On 2/4/13 9:56 AM, Emily Lynema emilylyn...@gmail.com wrote: Here at NCSU, we use our locally-hosted Endeca service for our catalog and Serials Solutions Summon as an article search solution. Why do this? 1. Our next-gen library catalog (Endeca) came first. This was before Solr hit the library world, and before library vendors started working on improving their bundled catalog apps. Our bundled catalog was terrible, and we wanted something better. This was back in the day when everyone was doing federated search for articles (think MetaLib). 2. 4-5 years down the road, a number of vendors (Ebsco, Serials Solutions, etc.) were getting into the web scale discovery business. Aka, one big index that includes everything, in particular the citation content that libraries have historically not had local access to index / search. We bought Summon to solve the article search problem that federated searching never resolved for us. We wanted one access point for less experienced users who needed to find articles. Since we had backed away from federated search for articles, this was our big pain point; we already had a catalog we liked. We've actually loaded our catalog content into Summon, as well. So why keep both? We've done a LOT of work adding functionality into our local catalog, including enhanced title searching,lots of supplemental content, a quite complex local requesting system. So we can't just switch to the Summon interface without some effort. In addition, we have found that we prefer the bento box approach to searching across formats, as opposed to the integrated index approach of Summon. At least at this moment. We use this in the search across our library website [1]. It's just really, really hard to always surface the right kind of thing the user is looking for when the things you're indexing are different in nature (ex: bibliographic record vs. full-text of newspaper article). With the bento box approach, you have better opportunities to surface the different types of content available, while still having local systems optimized for specific content types. Maybe that's a long-winded excuse for not pushing to break down silos more. Time will probably tell. -emily [1] http://www.lib.ncsu.edu/search/?q=java - Date:Fri, 1 Feb 2013 04:21:01 + From:Jonathan Rochkind rochk...@jhu.edu Subject: Re: Why we need multiple discovery services engine? So, there are two categories of solutions here -- 1) local indexes, where you create the index yourself, like blacklight or vufind (both based on a local Solr). 2) vendor-hosted indexes, where the vendor includes all sorts of things in their index that you the customer don't have local metadata for, mostly including lots and lots of scholarly article citations. If you want to include scholarly article citations, you probably can't do that with a local index solution. Although some consortiums have done some interesting stuff in that area, let's just say it takes a lot of resources to do. For most people, if you want to include article search in your index, it's not feasilbe to do so with a local index. So only VuFind/Blacklight with a local Solr is out, if you want article search. You _can_ load local content in a vendor-hosted index like EDS/Primo/Summon. So plenty of people do choose a vendor-hosted index product as their only discovery tool, including both local metadata and vendor-provided metadata. As you suggest. But some people want the increased control that a locally controlled Solr index gives you, for the local metadata where it's feasible. So use a local index product. But still want the article search you can get with a vendor-hosted index product. So they use both. There is also at least some reasons to believe that our users don't mind and may even prefer having local results and hosted metadata results presented seperately (although probably preferably in a consistent UI), rather than merged. A bunch more discussion of these issues is included in my blog post at: http://bibwild.wordpress.com/2012/10/02/article-search-improvement-strateg y/ From: Code for Libraries [CODE4LIB@LISTSERV.ND.EDU] on behalf of Wayne Lam [ wing...@gmail.com] Sent: Thursday, January 31, 2013 9:31 PM To: CODE4LIB@LISTSERV.ND.EDU Subject: [CODE4LIB] Why we need multiple discovery services engine? Hi all, I saw in numerous of library website, many of them would have their own based discovery
Re: [CODE4LIB] Linked data [was: Why we need multiple discovery services engine?]
On Feb 4, 2013, at 10:34 AM, Donna Campbell dcampb...@wts.edu wrote: In mentioning pushing to break down silos more, it brings to mind a question I've had about linked data. From what I've read thus far, the idea of breaking down silos of information seems like a good one in that it makes finding information easier but doesn't it also remove some of the markers of finding credible sources? Doesn't it blend accurate sources and inaccurate sources? Provenance is especially important in this context, which I think is a crucial role that libraries can play. -Ross. Donna R. Campbell Technical Services Systems Librarian (215) 935-3872 (phone) (267) 295-3641 (fax) Mailing Address (via USPS): Westminster Theological Seminary Library P.O. Box 27009 Philadelphia, PA 19118 USA Shipping Address (via UPS or FedEx): Westminster Theological Seminary Library 2960 W. Church Rd. Glenside, PA 19038 USA -Original Message- From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Emily Lynema Sent: Monday, February 04, 2013 9:56 AM To: CODE4LIB@LISTSERV.ND.EDU Subject: Re: [CODE4LIB] Why we need multiple discovery services engine? Here at NCSU, we use our locally-hosted Endeca service for our catalog and Serials Solutions Summon as an article search solution. Why do this? 1. Our next-gen library catalog (Endeca) came first. This was before Solr hit the library world, and before library vendors started working on improving their bundled catalog apps. Our bundled catalog was terrible, and we wanted something better. This was back in the day when everyone was doing federated search for articles (think MetaLib). 2. 4-5 years down the road, a number of vendors (Ebsco, Serials Solutions, etc.) were getting into the web scale discovery business. Aka, one big index that includes everything, in particular the citation content that libraries have historically not had local access to index / search. We bought Summon to solve the article search problem that federated searching never resolved for us. We wanted one access point for less experienced users who needed to find articles. Since we had backed away from federated search for articles, this was our big pain point; we already had a catalog we liked. We've actually loaded our catalog content into Summon, as well. So why keep both? We've done a LOT of work adding functionality into our local catalog, including enhanced title searching,lots of supplemental content, a quite complex local requesting system. So we can't just switch to the Summon interface without some effort. In addition, we have found that we prefer the bento box approach to searching across formats, as opposed to the integrated index approach of Summon. At least at this moment. We use this in the search across our library website [1]. It's just really, really hard to always surface the right kind of thing the user is looking for when the things you're indexing are different in nature (ex: bibliographic record vs. full-text of newspaper article). With the bento box approach, you have better opportunities to surface the different types of content available, while still having local systems optimized for specific content types. Maybe that's a long-winded excuse for not pushing to break down silos more. Time will probably tell. -emily [1] http://www.lib.ncsu.edu/search/?q=java
Re: [CODE4LIB] usability testing software
Thanks everyone for your responses! Lots to play with. N On Mon, Feb 4, 2013 at 9:41 AM, Rogers, Nathan roger...@indiana.edu wrote: Depending what your budget there are a number of good web based services out there. This breakdown is pretty good although I only have firsthand experience with WebSort (free version) and Crazyegg. If all you need to capture is audio/video I know some people that have made due with Camtasia, which is available for OS X. On 1/31/13 10:35 AM, Nate Hill nathanielh...@gmail.com wrote: Hi all, Years ago I had the opportunity to use Morae to do some usability testing. http://www.techsmith.com/morae.html I may have an opportunity to put together a little bit of a usability testing lab at my library, and I wonder if anyone can suggest a similar product but... I'd like it to run on Macs. Suggestions? thanks -- Nate Hill nathanielh...@gmail.com http://4thfloor.chattlibrary.org/ http://www.natehill.net -- Nate Hill nathanielh...@gmail.com http://4thfloor.chattlibrary.org/ http://www.natehill.net
Re: [CODE4LIB] conf presenters: a kind request
I'm gonna add to this briefly, and probably a bit less tactfully than Jonathan :-) - My number-one complaint about past presentations: Don't have slides we can't read. You probably can't read this, but... isn't a helpful thing to hear during a presentation. Make it legible, or figure out a different way to present the information. A kick-ass poster or UML diagram or flowchart or whatever isn't kick-ass when we can't read it. It's just an uninformative blur. [Note: this doesn't mean you shouldn't include the kick-ass poster when you upload your slides. Please do!] - Make sure your content fits well in the time allotted. You're not there to get through as much as possible. You're there to best use our collective time to make the argument that what you're doing is important/impressive/worth knowing, and to convey *as much of the interesting bits as you can without rushing*. The goal isn't for you to get lots of words out of your mouth; the goal is for us to understand them. If you absolutely can't cut it down to a point where you're not rushing, then you haven't done the hard work of distilling out the interesting bits, and you should get on that right away. - On the flip side, don't present for 8mn and leave plenty of time for questions. Odds are your'e not saying anything interesting enough to elicit questions in those 8 minutes. If you really only have 8mn of content, well, you shouldn't have proposed a talk. But odds are you *do* have interesting things to say, and may want to chat with your colleagues to figure out exactly what that is. - Don't make the 3.38 million messages on creating a non-threatening environment be for naught. Please. As Jonathan said: this is a great, great audience. We're all forgiving, we're all interested, we're all eager to lean new things and figure out how to apply them to our own situations. We love to hear about your successes. We *love* to hear about failures that include a way for us to avoid them, and you're going to be well-received no matter what because a bunch of people voted to hear you! On Mon, Feb 4, 2013 at 10:47 AM, Jonathan Rochkind rochk...@jhu.edu wrote: We are all very excited about the conference next week, to speak to our peers and to hear what our peers have to say! I would like to suggest that those presenting be considerate to your audience, and actually prepare your talk in advance! You may think you can get away with making some slides that morning during someone elses talk and winging it; nobody will notice right? Or they wont' care if they do? From past years, I can say that for me at least, yeah, I can often tell who hasn't actually prepared their talk. And I'll consider it disrespectful to the time of the audience, who voted for your talk and then got on airplanes to come see it, and you didn't spend the time to plan it advance and make it as high quality for them as you could. I don't mean to make people nervous about public speaking. The code4lib audience is a very kind and generous audience, they are a good audience. It'll go great! Just maybe repay their generosity by actually preparing your talk in advance, you know? Do your best, it'll go great! If you aren't sure how to do this, the one thing you can probably do to prepare (maybe this is obvious) is practice your presentation in advance, with a timer, just once. In front of a friend or just by yourself. Did you finish on time, and get at least half of what was important in? Then you're done preparing, that was it! Yes, if you're going to have slides, this means making your slides or notes/outline in advance so you can practice your delivery just once! Just practice it once in advance (even the night before, as a last resort!), and it'll go great! Jonathan -- Bill Dueber Library Systems Programmer University of Michigan Library
[CODE4LIB] Job: Librarian for Assessment at Yale University Library
**Librarian for Assessment Program Development Research Yale University Library Librarian 1-3** Schedule: Full-time (37.5 hours per week); Standard Work Week (M-F, 8:30-5:00) Yale University offers exciting opportunities for achievement and growth in New Haven, Connecticut. Conveniently located between Boston and New York, New Haven is the creative capital of Connecticut with cultural resources that include two major art museums, a critically acclaimed repertory theater, state-of-the-art concert hall, and world-renowned schools of Architecture, Art, Drama, and Music. **The University and the Library** The Yale University Library, as one of the world's leading research libraries, collects, organizes, preserves, and provides access to and services for a rich and unique record of human thought and creativity. It fosters intellectual growth and supports the teaching and research missions of Yale University and scholarly communities worldwide. A distinctive strength is its rich spectrum of resources, including around 12.8 million volumes and information in all media, ranging from ancient papyri to early printed books to electronic databases. The Library is engaging in numerous projects to expand access to its physical and digital collections. Housed in eighteen buildings including the Sterling Memorial Library, the Beinecke Rare Book and Manuscript Library, and the Bass Library, it employs a dynamic and diverse staff of approximately five hundred who offer innovative and flexible services to library readers. For additional information on the Yale University Library, please visit the Library's web site at www.library.yale.edu. **Position Focus** The Librarian for Assessment provides leadership and vision for assessment, measurement, planning and analysis throughout the Yale University Library (YUL) and strives to ensure that these activities are integral parts of the Library's programs, services, and collections. The Librarian for Assessment oversees and participates in assessment efforts throughout YUL; serves as an internal consultant for data-gathering and assessment activities conducted by other Library staff; works with Library colleagues to analyze and report assessment data; represents the Library in campus, regional and national assessment efforts; evaluates the effectiveness of Library assessment efforts and how they support the mission and strategic goals of the Library and the University; and recommends ways to strengthen the Library's assessment and measurement programs. The successful applicant for this position must possess strong analytical skills, a firm understanding of Library services and work processes, and strong interpersonal and listening skills. This position reports to the Associate University Librarian for Program Development and Research. **Principal Responsibilities** 1. Directs assessment efforts within the Yale University Library. Initiates assessment activities and provides consultation for assessment work done by other Library staff and departments. Promotes awareness and communication of other related assessment efforts within the Library, University, and externally. 2. Works with Library IT and other organizations within the University and external to it to gather and create tools that enable managers and staff to make data-driven decisions. 3. Works with the Director of Collection Development to research, design, and test methodologies for collection assessment appropriate to the needs of the Yale University Library. 4. Analyzes assessment-related data and communicates assessment activities and results to appropriate individuals and groups, including Library staff and the Yale campus community. 5. Develops and maintains expertise in assessment methods, techniques and best practices. 6. Establishes training and documentation programs for Library staff on use of appropriate assessment tools and methods. 7. Evaluates effectiveness of Library assessment activities on a regular basis and makes recommendations on ways to strengthen assessment work, including support needed. Develops and fosters a culture of assessment within the Library. 8. Contributes to the profession and represents the Library and the University in the academic, scholarly, and professional community. Responds to, collaborates with and participates as appropriate in other campus, regional and national assessment-related efforts. **Required Education and Experience** Master's degree in Library Science from an American Library Association accredited Library school. In selected instances, a post-graduate degree in a related discipline may be required or substituted for an MLS. Appointment to this rank is limited to two years at which time it is expected that the individual will develop necessary requirements to meet expectations of performance at the Librarian 2 level. **Required Skills and Abilities**
Re: [CODE4LIB] Why we need multiple discovery services engine?
Here at UMich, we do *not* yet load our catalog data into Summon. Partially because it's not at all clear that the patrons want books and articles mixed in that way, but mostly because we've worked really hard to customize the way relevancy and indexing works in our catalog, and we don't have that level of customization with Summon yet. Believe me, if i thought I could get what the institution wants, I'd give up writing/maintaining the catalog in a heartbeat. On Mon, Feb 4, 2013 at 10:48 AM, Owen, Will o...@email.unc.edu wrote: I think this pretty accurately reflects the sentiments at the University of North Carolina, one of NSCU's partners in developing the Endeca interface to our catalog. It's as much a mechanism for delivering services as it is for discovery, and as such reflects the physical nature of the collections it includes. It's not possible at this point for us to build those services into the Summon environment. +1 all around to Emily's comments. Will Owen On 2/4/13 9:56 AM, Emily Lynema emilylyn...@gmail.com wrote: Here at NCSU, we use our locally-hosted Endeca service for our catalog and Serials Solutions Summon as an article search solution. Why do this? 1. Our next-gen library catalog (Endeca) came first. This was before Solr hit the library world, and before library vendors started working on improving their bundled catalog apps. Our bundled catalog was terrible, and we wanted something better. This was back in the day when everyone was doing federated search for articles (think MetaLib). 2. 4-5 years down the road, a number of vendors (Ebsco, Serials Solutions, etc.) were getting into the web scale discovery business. Aka, one big index that includes everything, in particular the citation content that libraries have historically not had local access to index / search. We bought Summon to solve the article search problem that federated searching never resolved for us. We wanted one access point for less experienced users who needed to find articles. Since we had backed away from federated search for articles, this was our big pain point; we already had a catalog we liked. We've actually loaded our catalog content into Summon, as well. So why keep both? We've done a LOT of work adding functionality into our local catalog, including enhanced title searching,lots of supplemental content, a quite complex local requesting system. So we can't just switch to the Summon interface without some effort. In addition, we have found that we prefer the bento box approach to searching across formats, as opposed to the integrated index approach of Summon. At least at this moment. We use this in the search across our library website [1]. It's just really, really hard to always surface the right kind of thing the user is looking for when the things you're indexing are different in nature (ex: bibliographic record vs. full-text of newspaper article). With the bento box approach, you have better opportunities to surface the different types of content available, while still having local systems optimized for specific content types. Maybe that's a long-winded excuse for not pushing to break down silos more. Time will probably tell. -emily [1] http://www.lib.ncsu.edu/search/?q=java - Date:Fri, 1 Feb 2013 04:21:01 + From:Jonathan Rochkind rochk...@jhu.edu Subject: Re: Why we need multiple discovery services engine? So, there are two categories of solutions here -- 1) local indexes, where you create the index yourself, like blacklight or vufind (both based on a local Solr). 2) vendor-hosted indexes, where the vendor includes all sorts of things in their index that you the customer don't have local metadata for, mostly including lots and lots of scholarly article citations. If you want to include scholarly article citations, you probably can't do that with a local index solution. Although some consortiums have done some interesting stuff in that area, let's just say it takes a lot of resources to do. For most people, if you want to include article search in your index, it's not feasilbe to do so with a local index. So only VuFind/Blacklight with a local Solr is out, if you want article search. You _can_ load local content in a vendor-hosted index like EDS/Primo/Summon. So plenty of people do choose a vendor-hosted index product as their only discovery tool, including both local metadata and vendor-provided metadata. As you suggest. But some people want the increased control that a locally controlled Solr index gives you, for the local metadata where it's feasible. So use a local index product. But still want the article search you can get with a vendor-hosted index product. So they use both. There is also at least some reasons to believe that our users don't mind and may even prefer having local results and
Re: [CODE4LIB] Jan. 9th Moving Up to the Cloud
Belated followup, we have posted a YouTube video of this event. http://www.library.vcu.edu/blog/news/2013/01/video-library-technology-leade.html thanks, Jimmy On Thu, Jan 3, 2013 at 1:17 PM, Jimmy Ghaphery jghap...@vcu.edu wrote: Please see the announcement below for the event on January 9th which will also be streamed. thanks, Jimmy * Library technology-science leaders to gather at VCU on Jan. 9 for Moving Up to the Cloud event* As one of the first major research universities to launch a pioneering new cloud-resident enterprise library platform, Virginia Commonwealth University is hosting an educational session Jan. 9, designed to help other libraries around the world prepare for the next generation of library technologies. VCU Libraries went live in the fall semester with Alma, cloud-resident enterprise library software from Ex Libris Inc. Alma is a unified resource-management system that replaced the system VCU had used for the past decade, which had also been developed by Ex Libris. VCU is the third library in North America to have launched Alma. Speakers at the Jan. 9 event: - Marshall Breeding, the national authority on enterprise library systems, is editor of Library Technology Guides and author of a monthly column in Computers in Libraries. - Mark Ryland, chief solutions architect for Amazon Web Service’s Worldwide Public Sector team, began working with Microsoft Federal Systems as a senior architectural engineer and has seen the industry grow up. - Mark Triest, president of Ex Libris North America since 2010, previously was president of Inigral, a social-media start-up delivering applications for the education market. He also served as senior vice president of global sales at Sunguard Higher Education. - John Ulmschneider, university librarian at VCU, is a past president and board member of the Association of Southeastern Research Libraries and leads VCU’s growing and innovative research-library system. - Gene Damon, panel moderator, is director of library automation for the Virginia Community College System, one of the nation’s largest community-college systems. He is chair of the steering committee for the Virtual Library of Virginia (VIVA), and brings a distinguished background of leadership in enterprise library systems in both corporate and academic environments. An audience of Virginia and Washington-area academic library leaders will participate in two panel sessions, which will be streamed live on the Web and posted for viewing by interested parties worldwide. Get details, register to attend, or watch livehttp://www.library.vcu.edu/events/cloud/ -- Jimmy Ghaphery Head, Library Information Systems VCU Libraries 804-827-3551 -- Jimmy Ghaphery Head, Library Information Systems VCU Libraries 804-827-3551
Re: [CODE4LIB] conf presenters: a kind request
On Feb 4, 2013, at 11:25 AM, Bill Dueber wrote: [trimmed (and agreed with all of that)] As Jonathan said: this is a great, great audience. We're all forgiving, we're all interested, we're all eager to lean new things and figure out how to apply them to our own situations. We love to hear about your successes. We *love* to hear about failures that include a way for us to avoid them, and you're going to be well-received no matter what because a bunch of people voted to hear you! I'd actually be interested in people's complaints about bad presentations; I've been keeping notes for years, with the intention of making a presentation on giving better presentations. (but it's much harder than it sounds, as I plan on making all of the mistakes during the presentation) On Mon, Feb 4, 2013 at 10:47 AM, Jonathan Rochkind rochk...@jhu.edu wrote: We are all very excited about the conference next week, to speak to our peers and to hear what our peers have to say! I would like to suggest that those presenting be considerate to your audience, and actually prepare your talk in advance! [trimmed] Just practice it once in advance (even the night before, as a last resort!), and it'll go great! I did one of those 'Ignite' talks this year; because it's auto- advancing slides, I went over it multiple times. My recommendation is that you try to get various co-workers as guinea pigs. I even subjected one of my neighbors to it, even though he wasn't necessarily part of the intended audience. They gave me a lot of feed back -- asking for clarification on bits, we realized I could trim down a couple of slides, giving me more slides to expand other bits. I still screwed up the presentation, but it would have been much worse if I hadn't practiced. My local ASIST chapter used to run 'preview' events before the annual meeting, where the local folks presenting at annual were invited to give their talks. If nothing else, it forced you to have it done a couple of weeks early, but more importantly, it gave me a chance to have a similar audience to what would be at the main meeting ... one of my talks bombed hard; it was on standards protocols for scientific data, and I hadn't considered just how bad a talk that's 50% acronyms would go over. I was able to change how I presented the material so it wasn't quite so painful the second time around. There's only been once when practicing in advanced made for a worse presentation ... and that's because when I finished, PowerPoint asked me if I wanted to save the timings ... what ever you do, do *not* tell it yes. Because then it'll auto-advance your slides, so when you skip over one slide during the practice, it'll not let you have it up during the real talk. (There's a setting to turn off use of timings ... and the audience laughed when I kept scolding the computer, but it still felt horrible when I was up there) And it's important that you *must* practice in front of other people. How fast you think it's going to take you, or how fast it takes you talking to yourself is nothing like talking in front of other people. ... So, all of that being said, some of the things I've made a note of over the years. (it's incomplete, as I've still take notes by hand, and there are more items on the back pages of the various memo books I've had over the years) * Get there before the session, and test your presentation on the same hardware as it's going to be presented from. This is especially important if you're a Mac user, and presenting from a PC, or visa-versa. Look for odd fonts, images that didn't load, videos, abnormal gamma, bad font sizes (may result in missing test), missing characters, incorrect justification, etc. * If you're going to be presenting from your own machine, still test it out, to make sure that you have all of the necessary adaptors, that you know what needs to be done to switch the monitor, that the machine detects the projector at a reasonable size and the gamma's adjusted correctly. (and have it loaded in advance; you're wasting enough time switching machines). And start switching machines while the last presenter's doing QA ... and if you lose 5 min because of switching, prepare to cut your talk short, force the following presenters to lose time) * Have a backup plan, with the presentation stashed on a website that you've memorized the URL to, *and* on a USB stick. (website is safer vs. virus transfer, only use the USB stick if there's no internet) And put the file at the top level of the USB stick, not buried 12 folders deep. * If they have those clip on microphones, put it on your label on the same side as the screen is to you. (so whenever you turn to look at the screen, it still picks up your voice) * If you have a stationary mic, you have to actually stay near it or it doesn't work. * Hand-held mics suck unless you're used to them, as most of us aren't used to holding our hand up
Re: [CODE4LIB] Anyone have a SUSHI client?
Tom, I'm glad you are getting some help from looking at my code. That was the reason I put it up. When I was designing a SUSHI client last year, there was virtually nothing available to show how to do it, so I was hoping sushi.py would prevent someone from having to stumble in the dark quite as much. Funny you mention your organization wanting you to use Sharepoint. We also originally used Sharepoint for storing some of our operational data, but in the end, Sharepoint is such a closed environment that it wasn't able to meet our needs. However, even if you are 100% locked into Sharepoint, you should still be able to use Python code to access the SUSHI web services and then store it in Sharepoint using some of its APIs (if you can wade through Sharepoint's docs enough to figure out those APIs). Are you familiar with XAMPP? It's a light-weight Windows package that lets you run MySQL and some other stuff on a desktop computer very easily. Even if you don't have access to an official web server from your institution, you could still run sushi.py just on your local computer and have it store the data in your local MySQL database. Then you could just create some backups of that database and store them as files on Sharepoint or a shared drive or whatever method you have available. Just a thought. Josh Welker -Original Message- From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Tom Keays Sent: Monday, February 04, 2013 9:39 AM To: CODE4LIB@LISTSERV.ND.EDU Subject: Re: [CODE4LIB] Anyone have a SUSHI client? Hi Joshua, I was mainly looking at your program, not for the code, but as a way to bring myself up to speed about current practices in modeling the COUNTER data. I'm trying to avoid reinventing something that has already been well thought through. I apologize for calling out your model. You have gotten much further than I have. Some of the other respondents in this thread have set me straight on some things I was very fuzzy on going in. How go about I collecting and storing the data is still something I haven't resolved yet. I personally would prefer a Python solution, but there forces here at MPOW that suggest I should build a data repository in SharePoint. Assuming that is the case, Serial Solution's open source SUSHI harvester written in .NET might actually be the way for me to go. So, my next step is to look at their data model and see what reports they collect and store. As an aside, I'm also now wondering if de-duping is strictly necessary as long as there is a field to record the date the report was generated. De-duping (or maybe just deprecating duplicate data) could be separate from the collection process. Best, Tom On Mon, Feb 4, 2013 at 10:07 AM, Joshua Welker jwel...@sbuniv.edu wrote: Tom, I am the guy who wrote sushi.py around this time last year. My apologies for the shabbiness of the code. It was meant to be primarily a proof of concept. It's definitely incomplete. I only completed the DB3 and JR1 report logic up to this point, but it would be easy enough to add other report types. You're also right that sushi.py doesn't do anything to dedupe data, but it would be very simple to write a script that reads through the SQL records and deletes dupes. You could also use the built-in UNIQUE flag in MySQL when creating your table so that duplicate records just don't get saved. If you use the CSV export functionality of sushi.py, Excel has some built-in dedupe features that would help as well. Let me know if you'd like some help modifying sushi.py. I sort of gave up on it last spring. SUSHI implementation among vendors is still pretty shabby, and there are still some weaknesses in the SUSHI standard (I wrote about them in the Nov 2012 issue of Computers in Libraries). The productivity gains I was seeing from using SUSHI ended up being pretty low. Josh Welker -Original Message- From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Tom Keays Sent: Friday, January 25, 2013 8:40 AM To: CODE4LIB@LISTSERV.ND.EDU Subject: Re: [CODE4LIB] Anyone have a SUSHI client? I've been looking briefly at sushi.py, as a way to orient myself to collecting stats this way. I'm not intending to single out sushi.py, but looking at it (mainly the data structure at this point, and not the code itself), raises some questions about the best approach for collecting SUSHI data. sushi.py seems to have a small number of routines; mainly to retrieve the XML file from a vendor and ingest the data in that file into a MySQL database. There are only MySQL tables for COUNTER JR1, DR1, DR2, and DR2 reports and they mirror, to a degree, the structure of the item records returned in the SUSHI xml. Here are the skeletons of 2 of the sushi.py SQL tables: counter_jr1 id int, print_issn varchar, online_issn varchar, platform varchar, item_name text, data_type varchar,
Re: [CODE4LIB] Linked data [was: Why we need multiple discovery services engine?]
On Feb 4, 2013, at 10:34 AM, Donna Campbell wrote: In mentioning pushing to break down silos more, it brings to mind a question I've had about linked data. From what I've read thus far, the idea of breaking down silos of information seems like a good one in that it makes finding information easier but doesn't it also remove some of the markers of finding credible sources? Doesn't it blend accurate sources and inaccurate sources? Yes, yes it does. The 'intelligence' community has actually been talking about this problem with RDF for years. My understanding is that they use RDF quads (not triples) so that they have an extra parameter to track the source. (it might be that they use something larger than a quad). From what I remember (the conversation was years ago), they have to be able to mark information as suspect (eg, they find that one of the sources is unreliable, then re-run all all of the analysis without that source's contribution to determine if they came to the same result). I don't know enough about the implementation of linked data systems, so if there's some way to filter which sources are considered for input, or if there's any tracking of the RDF triples once they're parsed out. -Joe
[CODE4LIB] kickstarter for a homebrow project, from mistym
homebrew is what many of us use to install code packages on OSX. mistym (Misty De Meo) is a frequent hanger outer in #code4lib, and has helped many of us with stuff (including homebrew). She's also one of the homebrew maintainers. Here's a kickstarter from mistym, to help homebrew with some automated testing infrastructure, to better handle user code submissions. It's notable that they're asking for money for actual equipment, not developer time reimbursement or anything. They're not asking for much. Consider kicking in some bucks if you feel like it! http://www.kickstarter.com/projects/homebrew/brew-test-bot
[CODE4LIB] Job: Electronic Resources / Reference Librarian at Boston Athenæum
The Boston Athenaeum seeks a librarian fluent in emerging technologies to manage electronic resources and serve as part of team that covers the reference desk. This full-time exempt position reports to the Head of Reader Services and supports the work of the Boston Athenaeum across all departments. Responsibilities: * Manage electronic resources; including acquisition, licensing, processing, and maintenance * Coordinate with serials acquisition in Technical Services Department. * Oversee collection development (working with subject bibliographers) and negotiate purchase with vendor or consortium liaison * Maintain electronic resources webpage * Participate in project to digitize finding aids * Serve on committee that maintains website * Work with all departments' collection management systems * Provide general and specialized reference assistance and occasionally work at circulation * Work some evenings and Saturdays in scheduled rotation among reference staff * Offer public presentations of reference services to members * Serve on a weekly basis in special collections room * Must be able to lift 40 pounds * Other duties as assigned Required Qualifications: Education * MLS from an ALA-accredited institution Experience * Experience offering public services in a library, using cataloging software (preferably OCLC), content management systems (e.g., Drupal), digital collection management software (e.g., ContentDM and The Museum System), email marketing software (e.g., Constant Contact) * Experience editing and maintaining webpages and websites, proficiency with HTML and CSS * Proficiency with Microsoft Office; experience with image editing Preferred Qualifications: * Familiarity with visual resource librarianship, Drupal, EAD, Raiser's Edge, proxy servers * Experience dealing with vendors * Experience editing text, knowledge of Adobe Photoshop * Familiarity with PHP, XSL and XML Position is available immediately. For further information see [www.bostonathenaeum.org/node/39](http://www.bosto nathenaeum.org/node/39). E-mail resume and cover letter to h...@bostonathenaeum.org. Brought to you by code4lib jobs: http://jobs.code4lib.org/job/5987/
[CODE4LIB] Job: E-Resources Librarian and Web and Digital Services Librarian (2 positions) (CUNY Queens College, New York) at CUNY Queens College
The Queens College Libraries (City University of New York) are seeking candidates for two positions: E-Resources Librarian and Web and Digital Services Librarian. E-Resources Librarian sought by the Benjamin S. Rosenthal Library at Queens College, The City University of New York ([http://qcpages.qc.cuny.edu/library](http://qcpages.qc.cuny.edu/library)). Anticipated start date is July 1, 2013. Reporting to the Associate Librarian for Public Services, the E-Resources Librarian effectively coordinates and leads e-resources activities which include management, budgeting, policy development, statistical and qualitative reporting and improvement in a high- tech multi-disciplinary library. For further information on the position (See Position 6560) and how to apply: [http://www.qc.cuny.edu/HR/Pages/JobListings.aspx](http://www.qc.cuny.edu/HR/P ages/JobListings.aspx) Web Digital Services Librarian sought by the Benjamin S. Rosenthal Library at Queens College, the City University of New York ([http://qcpages.qc.cuny.edu/library](http://qcpages.qc.cuny.edu/library)). Anticipated start date is July 1, 2013. The Web Digital Services Librarian is responsible as point person and coordinator for leading the ongoing activities required to maintain useful, accessible and user-friendly delivery of the Queens College Libraries digital library services. The Web Digital Services Librarian is responsible for design and delivery of a suite of library web sites and augmentative digital services which are reflective of current standards and best practices in academic library web site design, current pedagogical technologies and teaching and learning trends. The successful candidate reports to the Associate Librarian for Public Services and will be a member of a digital library team, working closely with the E-Resources Librarian, Library Systems Office staff and the campus and CUNY technology units' members. The position may encompass supervision. For further information on the position (See Position 6559) and how to apply: [http://www. qc.cuny.edu/HR/Pages/JobListings.aspx](http://www.qc.cuny.edu/HR/Pages/JobList ings.aspx) EMPLOYMENT OPPORTUNITY We are committed to enhancing our diverse academic community by actively encouraging people with disabilities, minorities, veterans, and women to apply. We take pride in our pluralistic community and continue to seek excellence through diversity and inclusion. EO/AA Employer Brought to you by code4lib jobs: http://jobs.code4lib.org/job/6021/