Re: [CODE4LIB] Google can give you answers, but librarians give you the right answers

2016-04-01 Thread Andrew Anderson
On Apr 1, 2016, at 0:31, Cornel Darden Jr. <corneldarde...@gmail.com> wrote:

> "Google can give you answers, but librarians give you the right answers."
> 
> Library: "because not everything on the internet is true"
> 
> Some people applauded the statement and were like: "yay librarians!"
> 
> Others thought it was a very ignorant statement. And many patrons caused a 
> huge backlash. It was interesting as the library responded to the irritated 
> patrons. 

While I understand the motivation behind these statements, it also presents as 
“You’re doing it wrong!”, which is likely part of the reason for the backlash.  

Some of the more effective materials that I’ve seen created to communicate this 
concept effectively show sample search engine results with millions of hits of 
varying quality juxtaposed against commercial databases with dozens of high 
quality hits, letting the user draw their own conclusion that they would rather 
look through a few dozen relevant items than all the chaff from the search 
engine results.

Don’t tell them they’re doing it wrong, let them see that there’s a better way 
and let them chose the better option willingly.

-- 
Andrew Anderson, President & CEO, Library and Information Resources Network, 
Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes


Re: [CODE4LIB] Internet of Things

2016-03-31 Thread Andrew Anderson
For those who were not previously aware of IoT, here’s a primer focused 
specifically on the library space:

https://www.oclc.org/publications/nextspace/articles/issue24/librariesandtheinternetofthings.en.html

IMHO this is still a very young concept, and not even fully imagined yet, so 
there is no reason to feel like you’ve missed the boat, when the ship hasn’t 
even reached the dock yet.

-- 
Andrew Anderson, President & CEO, Library and Information Resources Network, 
Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Mar 30, 2016, at 22:16, Lesli M <les...@gmail.com> wrote:

> I feel compelled to pipe up about the comment "Very sad that a librarian 
> didn't know what it was."
> 
> Librarians come in all flavors and varieties. Until I worked in a medical 
> library, I had no idea what a systematic review was. I had no idea there was 
> a variety of librarian called "clinical librarian."
> 
> Do you know the hot new interest for law libraries? Medical libraries? 
> Science libraries?
> 
> The IoT is a specific area of interest. Just like every other special 
> interest out there.
> 
> Is it really justified to expect all librarians of all flavors and varieties 
> to know this very tech-ish thing called IoT?
> 
> Lesli


Re: [CODE4LIB] [patronprivacy] Let's Encrypt and EZProxy

2016-01-16 Thread Andrew Anderson
On Jan 15, 2016, at 13:20, Salazar, Christina <christina.sala...@csuci.edu> 
wrote:

> Something that I also see implied here is why aren’t vendors doing a better 
> job collaborating with the developers of EZProxy, instead of only putting the 
> pressure on Let’s Encrypt to support wildcard certs (although I kind of think 
> that’s the better way to go).


Because it’s easier than actually taking the time to fully understand the 
platforms and how all the pieces fit together.  

I’ve lost track of how many discussions I have had with various vendors 
recently over:

* Why they need to encode URLs before trying to pass them to another service 
like EZproxy's login handler
* Why they really do need to pay attention to what RFC 2616 Section 3.2.2 and 
RFC 2396 Section 2.2 have to say regarding the use of the reserved character in 
URLs
* Why it’s a bad idea to add “DJ google.com” in the EZproxy stanza
* Why it’s a bad idea to add “DJ ” in the EZproxy stanza
* Why it’s a bad idea to add “DJ ” in the EZproxy 
stanza

Instead of trying to understand how proxied access works, someone just keeps 
slapping “DJ ” or “HJ ” into the service stanza 
until the service starts working, and then never revisits the final product to 
see if those additions were really necessary.  Do this for a few platform 
iterations, and the resulting stanza can become insane.

The conversations typically go something like this:

Me: “Why are you trying to proxy google.com services?” 
Vendor: “Because we’re loading the jQuery JavaScript library from their CDN."
Me: “And how are you handling registering all your customer’s IP addresses with 
Google?” 
…  … 
Vendor: “We don’t”.
Me: “Then why do you think you need that in your proxy stanza?”. 
…  …
Vendor: “We . . . don’t?”
Me: “Exactly. And how are you reaping the performance benefits of a CDN service 
if you’re funneling all of the unauthenticated web traffic through a proxy 
server instead of allowing the CDN to do what it does best and keeping the 
proxy server out of the middle of that transaction?"
Vendor: “We . . . aren’t?”
Me: “That’s right, by adding ‘DJ ’ to your stanza, you have 
successfully negated the performance benefits of using a CDN service.”

-- 
Andrew Anderson, President & CEO, Library and Information Resources Network, 
Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes


Re: [CODE4LIB] Let's Encrypt and EZProxy

2016-01-14 Thread Andrew Anderson
Eric,

Check out Startcom’s StartSSL service (https://www.startssl.com), for $120 you 
have the ability to generate 3-year wildcard certificates with their 
Organizational Validation level of service.

Andrew

-- 
Andrew Anderson, President & CEO, Library and Information Resources Network, 
Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Jan 14, 2016, at 21:33, Eric Hellman <e...@hellman.net> wrote:

> I would also go with the $120 3 year wildcard cert for ezproxy. What vendor 
> are you using?
>> On Jan 14, 2016, at 7:23 PM, Cary Gordon <listu...@chillco.com> wrote:
>> 
>> I love the idea of Let’s Encrypt, but I recently bought a three year 
>> wildcard cert subscription for about $120. I would need to fall firmly into 
>> the true believer category to go the route you suggest.
>> 
>> Cary
>> 
>>> On Jan 14, 2016, at 11:20 AM, Eric Hellman <e...@hellman.net> wrote:
>>> 
>>> A while back, the issue of needing a wildcard certificate (not supported by 
>>> Lets Encrypt) for EZProxy was discussed.
>>> 
>>> In my discussions with publishers about switching to HTTPS, EZProxy 
>>> compatibility has been the most frequently mentioned stumbling block 
>>> preventing a complete switch to HTTPS for some HTTPS-ready  publishers. In 
>>> two cases that I know of, a publisher which has been HTTPS-only was asked 
>>> by a library customer to provide insecure service (oh the horror!) for this 
>>> reason.
>>> 
>>> It's been pointed out to me that while Lets Encrypt is not supporting 
>>> wildcard certificates, up to 100 hostnames can be supported on a single LE 
>>> certificate. A further limit on certificates issued per week per domain 
>>> would mean that up to 500 hostnames can be registered with LE in a week.
>>> 
>>> Are there EZProxy instances out there that need more than 500 hostnames, 
>>> assuming that all services are switched to HTTPS?
>>> 
>>> Also, I blogged my experience talking to people about privacy at #ALAMW16.
>>> http://go-to-hellman.blogspot.com/2016/01/not-using-https-on-your-website-is-like.html
>>>  
>>> <http://go-to-hellman.blogspot.com/2016/01/not-using-https-on-your-website-is-like.html>
>>> 
>>> Eric
>>> 
>>> 
>>> Eric Hellman
>>> President, Free Ebook Foundation
>>> Founder, Unglue.it https://unglue.it/
>>> https://go-to-hellman.blogspot.com/
>>> twitter: @gluejar
>>> 
> 


Re: [CODE4LIB] FOSS recommendations for online-only library

2015-08-23 Thread Andrew Anderson
I would recommend Apache’s mod_proxy over Squid for a library setting, as it 
can be morphed into a general rewriting proxy easier than Squid can for 
off-site access.

It’s true that both can be made to perform the rewriting function, but the bar 
for entry is lower for Apache and it supports a broader set of authentication 
options than Squid does.

-- 
Andrew Anderson, President  CEO, Library and Information Resources Network, 
Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Aug 23, 2015, at 0:45, Cornel Darden Jr. corneldarde...@gmail.com wrote:

 Hello,
 
 There are open-source proxies available. I would give squid a try. 
 http://wiki.squid-cache.org/Features/Authentication
 
 At such a library, public domain materials are awesome! I would look into 
 calibre as an ebook server and mamager. http://calibre-ebook.com
 
 Of course, project Gutenberg and the internet archive will supply calibre 
 with thousands of free books. Also, look into drm free publishers. With squid 
 active, many non-drm options can be realized for eBooks too. Do not allow 
 access to databases without authentication. 
 
 Sent from my iPhone
 
 On Aug 22, 2015, at 11:06 PM, Nicole Askin nask...@alumni.ubc.ca wrote:
 
 1. We don't currently have such technology, though we are definitely
 looking at it beyond this project as well
 2. Either. From my understanding there aren't many/any comprehensive free
 discovery products. We're currently making do with a Google custom search
 engine, which is a very suboptimal solution
 3. Yes. I'm working on learning what I can, and we're working on tech
 support options.
 Thanks,
 Nicole
 
 On Fri, Aug 21, 2015 at 2:11 PM, Kevin Hawkins 
 kevin.s.hawk...@ultraslavonic.info wrote:
 
 We should probably clarify you're needs a bit.
 
 Will you need technology that manages authentication of authorized users,
 or does your non-profit already have some tool (like a user login or proxy
 server) that can decide which users should be able to get access to your
 resources?
 
 You mention discovery options ... are you thinking of a discovery
 product or old-fashioned federated search that provides a single user
 search interface that searches across many or all of your licensed
 products?  And a link resolver?
 
 As a general rule of thumb, you can either have limited tech support or
 use open-source software but not both.  :(
 
 Kevin
 
 
 On 8/20/15 5:04 PM, Nicole Askin wrote:
 
 Hello all,
 I'm working with a non-profit that is offering access to research
 databases
 for patrons that do not otherwise have it. We are hoping to develop a
 library portal to support users, ideally including both article- and
 journal-level search. We'd like to do this as much as possible using
 *only*
 free and open source software, so I'm looking for recommendations on what
 to use and, crucially, what works well together.
 Some parameters:
 -We have no physical location or physical holdings - don't need
 circulation
 or anything in that category, although access stats would be nice
 -We do not have our own hosted materials - no need for a CMS
 -We have very limited tech support
 
 Any thoughts? I've been playing around with VuFind and reSearcher so far
 but am definitely open to other possibilities, particularly if there are
 good discovery options available.
 
 Thanks,
 Nicole
 
 


Re: [CODE4LIB] Protocol-relative URLs in MARC

2015-08-17 Thread Andrew Anderson
There are multiple questions embedded in this:

1) What does the MARC standard have to say about 856$u?

$u - Uniform Resource Identifier

Uniform Resource Identifier (URI), which provides standard syntax for locating 
an object using existing Internet protocols. Field 856 is structured to allow 
for the creation of a URL from the concatenation of other separate 856 
subfields. Subfield $u may be used instead of those separate subfields or in 
addition to them.

Subfield $u may be repeated only if both a URN or a URL or more than one URN 
are recorded.

Used for automated access to an electronic item using one of the Internet 
protocols or by resolution of a URN. Subfield $u may be repeated only if both a 
URN and a URL or more than one URN are recorded. Field 856 is repeated if more 
than one URL needs to be recorded.

Here, it is established that $u uses a URI, which leads to….

2) What do the RFCs say about protocol-relative URIs?

http://tools.ietf.org/html/rfc3986#section-4.1

  URI-reference is used to denote the most common usage of a resource
   identifier.

  URI-reference = URI / relative-ref

   A URI-reference is either a URI or a relative reference.  If the
   URI-reference's prefix does not match the syntax of a scheme followed
   by its colon separator, then the URI-reference is a relative
   reference.

So by the stated use of URIs in the MARC standard, and the RFC definition of 
the URI relative reference, there should be no standards basis by which 
protocol relative URLs should not be valid for use in 856.

Expanding out to the software support, most tools that I have used with general 
URL manipulation in general have no problems with this format, but I have only 
used PyMARC for manipulating MARC records, not any of the other MARC editors. 
If they try to be too clever about data validation and not quite clever enough 
about standards and patterns, there could be issues at this level.

As for browser support, IE7  IE8 have issues with double-loading some 
resources when used in this manner, but those browsers are becoming nearly 
extinct, so I would not anticipate client-side issues as long as the 
intermediate system that consumed the 856 record and render it for display can 
handle this.  Our web properties switched to using this pattern several years 
ago to avoid the “insecure content” warnings and we have had no issues on the 
client side.  

Then the other consumers of MARC data come into play — title lists, link 
resolvers, proxy servers, etc.  A lot of what I’ve seen in this space are 
lipstick wearing dinosaurs of a code base, so unless the vendor is particularly 
good about keeping up with current web patterns, this is where I would expect 
the most challenges.  There may be implicit or explicit assumptions built into 
systems that would break with protocol-relative URLs, e.g. if the value is 
passed directly to a proxy server, it may not know what to do without a scheme 
prefixed to the URI, and attempt to serve local content instead.

That said, there is a big push recently for dropping non-SSL connections in 
general (going so far as to call the protocol relative URIs an anti-pattern), 
so is it really worth all the potential pain and suffering to make your links 
scheme-agnostic, when maybe it would be a better investment in time to switch 
them all to SSL instead?  This dovetails nicely with some of the discussions I 
have had recently with electronic services librarians about how to protect 
patron privacy in an online world by using SSL as an arrow in that quiver.

Andrew

-- 
Andrew Anderson, President  CEO, Library and Information Resources Network, 
Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Aug 17, 2015, at 16:41, Stuart A. Yeates syea...@gmail.com wrote:

 I'm in the middle of some work which includes touching the 856s in lots of
 MARC records pointing to websites we control. The websites are available on
 both https://example.org/ and http://example.org/
 
 Can I put //example.org/ in the MARC or is this contrary to the standard?
 
 Note that there is a separate question about whether various software
 systems support this, but that's entirely secondary to the question of the
 standard.
 
 cheers
 stuart
 --
 ...let us be heard from red core to black sky


Re: [CODE4LIB] quick question: CloudFlare

2015-06-19 Thread Andrew Anderson
We have had good experience with it so far, yes.  Do you have a specific use 
case that you’re concerned about?

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Jun 19, 2015, at 12:58, Kun Lin l...@whitman.edu wrote:

 Quick question:
 
 
 
 Who is using CloudFlare for their library website? Are they very
 accommodating in using CNAME?
 
 
 
 Thanks
 
 Kun Lin


Re: [CODE4LIB] quick question: CloudFlare

2015-06-19 Thread Andrew Anderson
That’s a bit sub-optimal regarding how they handle domain setup, I agree.  You 
can get partial functionality by adding a NS record in your existing DNS 
servers for pointing specific records to their DNS servers even without going 
through the full domain delegation process.  After some testing, we were 
sufficiently happy with their service to move forward with the full delegation, 
but this technique worked well for kicking the tires without making the full 
commitment to their DNS service.

The down side to using the NS trick is that their SSL handling will not be 
fully active unless you do the whole domain.  Depending on what you hope to 
accomplish, that may be the make-or-break decision for using their service or 
not.  You can still do SSL on the host under some circumstances, but I believe 
all entries in the top level domain must use their certificates when 
acceleration is active.  Subdomains can still use the SSL certificate on the 
host even without full delegation.

Another reason to consider letting them handle your DNS (if you can) is that 
they have some pretty interesting plans for adding DNSSEC support for later 
this year.

At any rate, what I would suggest you consider is something like this:

testIN  NS  ns1.ns.cloudflare.com
IN  NS  ns2.ns.cloudflare.com

and replace ns1 and ns2 with the name servers assigned to your account.

Of course, you need a “test” record created on the CloudFlare end to serve the 
appropriate DNS entries.  This configuration will send all DNS queries for the 
test host to CloudFlare’s servers and through their acceleration infrastructure.

Hope this helps,
Andrew

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Jun 19, 2015, at 18:29, Kun Lin l...@whitman.edu wrote:

 In most case, Cloudflare will want you to delete the whole domain to their
 DNS server. This is impossible for us to do. Therefore, I am trying to
 figure out CNAME option.
 
 Thanks
 Kun
 
 -Original Message-
 From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of
 Andrew Anderson
 Sent: Friday, June 19, 2015 3:24 PM
 To: CODE4LIB@LISTSERV.ND.EDU
 Subject: Re: [CODE4LIB] quick question: CloudFlare
 
 We have had good experience with it so far, yes.  Do you have a specific
 use case that you're concerned about?
 
 --
 Andrew Anderson, Director of Development, Library and Information
 Resources Network, Inc.
 http://www.lirn.net/ | http://www.twitter.com/LIRNnotes |
 http://www.facebook.com/LIRNnotes
 
 On Jun 19, 2015, at 12:58, Kun Lin l...@whitman.edu wrote:
 
 Quick question:
 
 
 
 Who is using CloudFlare for their library website? Are they very
 accommodating in using CNAME?
 
 
 
 Thanks
 
 Kun Lin


Re: [CODE4LIB] Let's implement the referrer meta tag

2015-06-12 Thread Andrew Anderson
Or just SSL enable your library web site.  Few vendors support SSL today, so 
crossing the HTTP/HTTPS barrier is supposed to automatically disable referring 
URL passing.

http://www.w3.org/Protocols/rfc2616/rfc2616-sec15.html#sec15.1.3

15.1.3 Encoding Sensitive Information in URI's

Because the source of a link might be private information or might reveal an 
otherwise private information source, it is strongly recommended that the user 
be able to select whether or not the Referer field is sent. For example, a 
browser client could have a toggle switch for browsing openly/anonymously, 
which would respectively enable/disable the sending of Referer and From 
information.

Clients SHOULD NOT include a Referer header field in a (non-secure) HTTP 
request if the referring page was transferred with a secure protocol.

Authors of services which use the HTTP protocol SHOULD NOT use GET based forms 
for the submission of sensitive data, because this will cause this data to be 
encoded in the Request-URI. Many existing servers, proxies, and user agents 
will log the request URI in some place where it might be visible to third 
parties. Servers can use POST-based form submission instead

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Jun 12, 2015, at 0:24, Conal Tuohy conal.tu...@gmail.com wrote:

 Assuming your library web server has a front-end proxy (I guess this is
 pretty common) or at least runs inside Apache httpd or something, then
 rather than use the HTML meta tag, it might be easier to set the referer
 policy via the Content-Security-Policy HTTP header field.
 
 https://w3c.github.io/webappsec/specs/content-security-policy/#content-security-policy-header-field
 
 e.g. in Apache httpd with mod_headers:
 
 Header set Content-Security-Policy referrer 'no-referrer'
 
 
 
 On 12 June 2015 at 13:55, Frumkin, Jeremy A - (frumkinj) 
 frumk...@email.arizona.edu wrote:
 
 Eric -
 
 Many thanks for raising awareness of this. It does feel like encouraging
 good practice re: referrer meta tag would be a good thing, but I would not
 know where to start to make something like this required practice. Did you
 have some thoughts on that?
 
 — jaf
 
 ---
 Jeremy Frumkin
 Associate Dean / Chief Technology Strategist
 University of Arizona Libraries
 
 +1 520.626.7296
 j...@arizona.edu
 ——
 A person who never made a mistake never tried anything new. - Albert
 Einstein
 
 
 
 
 
 
 
 
 
 On 6/11/15, 8:25 AM, Eric Hellman e...@hellman.net wrote:
 
 
 http://go-to-hellman.blogspot.com/2015/06/protect-reader-privacy-with-referrer.html
 
 http://go-to-hellman.blogspot.com/2015/06/protect-reader-privacy-with-referrer.html
 
 
 I hope this is easy to deploy on library websites, because the privacy
 enhancement is significant.
 
 I'd be very interested to know of sites that are using it; I know Thomas
 Dowling implemented a referrer policy on http://oatd.org/ 
 http://oatd.org/
 
 Would it be a good idea to make it a required practice for libraries?
 
 
 Eric Hellman
 President, Gluejar.Inc.
 Founder, Unglue.it https://unglue.it/
 http://go-to-hellman.blogspot.com/
 twitter: @gluejar
 


Re: [CODE4LIB] making EZproxy http/https transparent

2015-03-03 Thread Andrew Anderson
https://pluto.potsdam.edu/ezproxywiki/index.php/SSL#Wildcard_certificate

(You can safely ignore the SSL warning, pluto uses self-signed certificates)

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Mar 3, 2015, at 11:46, Karl Holten khol...@switchinc.org wrote:

 If you're using proxy by hostname, it's my understanding that you need to 
 purchase a SSL certificate for each secure domain, otherwise you get security 
 errors. Depending on how many domains you have, the cost of this can add up. 
 Maintaining it is a headache too because it seems like vendors often don't 
 bother to notify you they're making a switch.
 
 If there's some way to avoid doing this, I would love to know!
 
 Karl Holten
 Systems Integration Specialist
 SWITCH Inc
 414-382-6711
 
 -Original Message-
 From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of 
 Stuart A. Yeates
 Sent: Monday, March 2, 2015 5:27 PM
 To: CODE4LIB@LISTSERV.ND.EDU
 Subject: [CODE4LIB] making EZproxy http/https transparent
 
 In the last couple of months we've had to update a number of EZproxy stanzas 
 as either tools migrate to HTTPS-only or people try to access HTTP/HTTPS 
 parallel resources using browsers that automatically detect HTTP/HTTPS 
 parallel resources and switch users to the HTTPS version (think current 
 Chrome, anything with the HTTPSeverywhere plugin).
 
 We'd like to avoid updating our config.txt piecemeal on the basis of 
 user-gernated error-reports
 
 We're thinking of going through our EZproxy config.txt and adding an H 
 https:// for every H or URL entry. (Domain and DomainJavascript already work 
 for both HTTP and HTTPS).
 
 Has anyone tried anything like this? Are there pitfalls?
 
 cheers
 stuart
 --
 ...let us be heard from red core to black sky


Re: [CODE4LIB] [RESOLVED] Re: HTTPS EZproxy question / RFC 6125

2014-12-24 Thread Andrew Anderson
There are 3 basic approaches to rewriting proxy servers that I have seen in the 
wild, each with their own strengths and weaknesses:

1) Proxy by port

This is the original EZproxy model, where each proxied resource gets its own 
port number.  This runs afoul of firewall rules to non port 80/443 resources, 
and it creates a problem for SSL access, as clients try both HTTP and HTTPS to 
the same port number, and EZproxy is not setup to differentiate both protocols 
accessing the same port.  With more and more resources moving to HTTPS, the end 
of this solution as a viable option is in sight.

2) Proxy by hostname

This is the current preferred EZproxy model, as it addresses the HTTP(S) port 
issue, but as you have identified, it instead creates a hostname mangling 
issue, and now I’m curious myself about how EZproxy will handle a hyphenated 
SSL site as well with HttpsHyphens enabled.  I /think/ it does the right thing 
by mapping the hostname back to the original internally, as a “-“ in hostnames 
for release versioning is how the Google App Engine platform works, but I have 
not explicitly investigated that.

3) Proxy by path

A different proxy product that we use, Muse Proxy from Edulib, leverages proxy 
by path, where the original website URL is deconstructed and passed to the 
proxy server as query arguments.  This approach has worked fairly well as it 
cleanly avoids the hostname mangling issues, though some of the new “single 
page web apps” that use JavaScript routing patterns can be interesting, so the 
vendor has added proxy by hostname support as an option for those sites as a 
fallback.

So there is no perfect solution, but some work better than others.  I’m looking 
forward to expanding our use of the proxy by path approach, as that is a very 
clean approach to this problem, and it seems to have fewer caveats than the 
other two approaches.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Dec 18, 2014, at 17:04, Stuart A. Yeates syea...@gmail.com wrote:

 It appears that the core of my problem was that I was unaware of
 
 Option HttpsHyphens / NoHttpsHyphens
 
 which toggle between proxying on
 
 https://www.somedb.com.ezproxy.yourlib.org
 
 and
 
 https://www-somedb-com.ezproxy.yourlib.org
 
 and allows infinitely nested domains to be proxied using a simple
 wildcard cert by compressing things.
 
 The paranoid in me is screaming that there's an interesting brokenness
 in here when a separate hosted resource is at https://www-somedb.com/,
 but I'm trying to overlook that.
 
 cheers
 stuart
 --
 ...let us be heard from red core to black sky
 
 
 On Mon, Dec 15, 2014 at 9:24 AM, Stuart A. Yeates syea...@gmail.com wrote:
 Some resources are only available only via HTTPS. Previously we used a
 wildcard certificate, I can't swear that it was ever tested as
 working, but we weren't getting any complaints.
 
 Recently browser security has been tightened and RFC 6125 has appeared
 and been implemented and proxing of https resources with a naive
 wildcard cert no longer works (we're getting complaints and are able
 to duplicate the issues).
 
 At 
 https://security.stackexchange.com/questions/10538/what-certificates-are-needed-for-multi-level-subdomains
 there is an interesting solution with multiple wildcards in the same
 cert:
 
 foo.com
 *.foo.com
 *.*.foo.com
 ...
 
 There is also the possibility that we can just grep the logs for every
 machine name ever accessed and generate a huge list.
 
 Has anyone tried these options? Successes? Failures? Thoughts?
 
 cheers
 stuart
 
 
 --
 ...let us be heard from red core to black sky


Re: [CODE4LIB] Functional Archival Resource Keys

2014-12-11 Thread Andrew Anderson
I’m not commenting on whether inflections are good, bad, or ugly, but simply 
looking at this from the perspective of real-world hurdles, unexpected 
interactions, and implementation challenges that are going to be run into by 
the selection of an existing reserved character as an inflection indicator.  It 
looks like we disagree on the concept that “no one is using it” as it has a 
clearly defined role in the URI specification, and it is not uncommon to use 
“?”’s as a cache-busting mechanism when clearly no one intends to fetch an 
object’s metadata when they do so.

Taking a step back, this seems like a false economy vs a more expressive and 
human-friendly mechanism for defining access to metadata and policy for the 
object in question.

There are a number of different approaches that could be taken to achieve the 
stated goals of ARK without overloading the purpose of an existing defined 
reserved character, and I think that the project would be doing itself a favor 
by exploring the alternatives to find an approach that does not have the 
potential to slow adoption due to technical and political reasons.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Dec 10, 2014, at 14:28, John Kunze j...@ucop.edu wrote:

 I don't know the precise constraints you're working under, but Henry
 Thompson of the W3C TAG (Technical Architecture Group) has advocated for
 consideration of the ARK approach to the TAG's meetings.
 
 The terminal '?' is sort of a no-brainer, but clearly it stretches the URI
 spec; on the plus side, it's ripe for definition since no one else is using
 it.  It was Jonathan Rees (also of the W3C TAG) who pointed out the need
 for an additional response header, just in case some service actually was
 responding query strings that overlapped with inflections.
 
 Just to be clear, the ARKs don't own the inflections concept (in fact the
 ARK scheme is unusual in not owning things, such as a resolver).  If you
 think inflections are a good/bad idea for ARKs, chances are you'll think
 the same for other kinds of identifiers.  As Clifford Lynch once said, the
 '?' inflection should work for all URLs.
 
 On Tue, Dec 9, 2014 at 10:09 PM, Andrew Anderson and...@lirn.net wrote:
 
 RFC and expectation violations make my brain hurt.
 
 Overloading an operator that has a clearly defined role in HTTP URIs (
 https://tools.ietf.org/html/rfc7230#section-2.7.1) creates the potential
 for /so/ many unexpected interactions between browsers (
 https://code.google.com/p/chromium/issues/detail?id=108690), HTTP caches,
 URL rewriting servers, etc. that implementations, adopters, and users are
 going to be playing a long game of whack-a-mole working around them.
 
 The proposal is already carving out a URI namespace in the form of “ark:”:
 
  http://ark.cdlib.org/ark:/13030/tf5p30086k?
 
 So why not take advantage of the fact that any system processing the
 “ark:” namespace is already going to have to be a custom application and
 adopt a RESTful path to communicate the service requested instead?
 
  http://ark.cdlib.org/ark:metadata/13030/tf5p30086k
  http://ark.cdlib.org/ark:policy/13030/tf5p30086k
 
 If a web services style implementation is undesired, what about creating
 another reserved character or overload a character that is already used in
 URIs but not part of the HTTP URI specification, “!?
 
 Or, if a standard approach for HTTP header implementation were proposed
 and adopted, it is not unreasonable to imagine that browsers might adopt
 methods that would allow the average user access to the inflections without
 jumping through hoops once adoption reaches critical mass.
 
 There are many approaches and techniques that could be employed here that
 would not require overloading “?” in HTTP URIs that there really is no
 excuse for trying to do so.
 
 --
 Andrew Anderson, Director of Development, Library and Information
 Resources Network, Inc.
 http://www.lirn.net/ | http://www.twitter.com/LIRNnotes |
 http://www.facebook.com/LIRNnotes
 
 On Dec 9, 2014, at 9:25, Ethan Gruber ewg4x...@gmail.com wrote:
 
 I'm using a few applications in Tomcat, so inflections are much more
 difficult to implement than content negotiation. I can probably tweak the
 Apache settings to do a proxypass for inflections by modifying the
 examples
 above.
 
 I agree with Conal, though. Inflections are puzzling at best and bad
 architecture at worst, and the sooner the community puts forward a more
 standard solution, the better.
 
 On Mon, Dec 8, 2014 at 7:21 PM, John Kunze j...@ucop.edu wrote:
 
 Just as a URL permits an ordinary user with a web browser to get to an
 object, inflections permit an ordinary user to see metadata (without
 curl
 or code).
 
 There's nothing to prevent a server from supporting both the HTTP Accept
 header (content negotiation) and inflections.  If you can do

Re: [CODE4LIB] Stack Overflow

2014-11-04 Thread Andrew Anderson
On Nov 4, 2014, at 9:42, Joshua Welker wel...@ucmo.edu wrote:

 3. Libraries have a culture of
 protecting vendors from criticism. Sure, we do lots of criticism behind
 closed doors, but nowhere that leaves an online footprint.

Oops.  Someone should have told me that rule before I openly and repeatedly 
criticized EBSCO for having a broken DNS configuration that is celebrating the 
2-year anniversary of my in-depth bug report to them, along with a specific 
resolution path that their IT department has demonstrated an amazing resolve to 
ignore despite repeated pings to their customer service representatives to keep 
the issue active over the past 2 years.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes


Re: [CODE4LIB] Why learn Unix?

2014-10-27 Thread Andrew Anderson
There is something of a natural symbiosis between *NIX and libraries.  If you 
have not already found it, read Unix as Literature for some background on why 
those who like the written word are drawn to *NIX naturally.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Oct 27, 2014, at 10:02, Siobhain Rivera siori...@indiana.edu wrote:

 Hi everyone,
 
 I'm part of the ASIST Student Chapter and Indiana University, and we're
 putting together a series of workshops on Unix. We've noticed that a lot of
 people don't seem to have a good idea of why they should learn Unix,
 particularly the reference/non technology types. We're going to do some
 more research to make a fact sheet about the uses of Unix, but I thought
 I'd pose the question to the list - what do you think are reasons
 librarians need to know Unix, even if they aren't in particularly tech
 heavy jobs?
 
 I'd appreciate any input. Have a great week!
 
 Siobhain Rivera
 Indiana University Bloomington
 Library Science, Digital Libraries Specialization
 ASIST-SC, Webmaster


Re: [CODE4LIB] Requesting a Little IE Assistance

2014-10-13 Thread Andrew Anderson
I’ve never attempted this, but instead of linking to the text files directly, 
can you include the text files in an iframe and leverage that to apply 
sizing/styling information to the iframe content?

Something like:

html
body
iframe src=“/path/to/file.txt”/iframe
/body
/html

That structure, combined with some javascript tricks might get you where you 
need to be:

http://stackoverflow.com/questions/4612374/iframe-inherit-from-parent

Of course, if you’re already going that far, you’re not too far removed from 
just pulling the text file into a nicely formatted container via AJAX, and 
styling that container as needed, without the iframe hackery.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Oct 13, 2014, at 9:59, Matthew Sherman matt.r.sher...@gmail.com wrote:

 For anyone who knows Internet Explore, is there a way to tell it to use
 word wrap when it displays txt files?  This is an odd question but one of
 my supervisors exclusively uses IE and is going to try to force me to
 reupload hundreds of archived permissions e-mails as text files to a
 repository in a different, less preservable, file format if I cannot tell
 them how to turn on word wrap.  Yes it is as crazy as it sounds.  Any
 assistance is welcome.
 
 Matt Sherman


Re: [CODE4LIB] Forwarding blog post: Apple, Android and NFC – how should libraries prepare? (RFID stuffs)

2014-10-08 Thread Andrew Anderson
On Oct 8, 2014, at 4:54, Ross Singer rossfsin...@gmail.com wrote:

 We’re generally in need of a spec, not a standard, I’ve found (although 
 they’re definitely not mutually exclusive!).


The wonderful thing about standards, is that there are so many to choose from.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes


Re: [CODE4LIB] Library app basics

2014-10-07 Thread Andrew Anderson
Before launching into a native app, start with the functional requirements to 
see if what you want to accomplish could be done in a well designed mobile web 
site, or if you actually need the advanced features that native development 
would make available.

For example, there is a _lot_ that you can do in jQuery Mobile backed by a 
strong AJAX backend that looks like a native app, yet does not subject you to 
the stringent requirements of having to do multi-platform development and worry 
about submitting to multiple vendors for approval.  

There is already some support for media capture for photos/video/sound in HTML5 
on some devices that you can use for interactive experiences like snapping a 
photo, sending it to the server for processing, and having the server send back 
something relevant.  See 
http://www.html5rocks.com/en/tutorials/getusermedia/intro/ for some information 
on what is possible currently, and then imagine what you could do with book 
covers, bar codes, maybe even tapping into the NFC chips in smartphones to 
tickle those RFID chips everyone is talking about this week.

As a data point, I have seen estimates that put mobile app development costs 
between $5,000 and $50,000, depending on their complexity, amount of UI/UX 
design and testing, graphics development, etc, so if you are operating without 
a budget and are having to scrounge for devices just to test with, a smart 
mobile web site may be a better starting point anyway.  It’s less of an 
unknown, using familiar tools, doesn’t require testing hardware, and doesn’t 
have an onerous vendor approval step to deal with.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Oct 7, 2014, at 14:51, Will Martin w...@will-martin.net wrote:

 My boss has directed me to start looking into producing a phone app for the 
 library, or better yet finding a way to integrate with the existing 
 campus-wide app.  Could I pick the list's brains?
 
 1) Is there some tolerably decent cross-platform app language, or am I going 
 to be learning 3 different languages for iOS, Android, and Windows phone?  
 I've dabbled in all kinds of things, but my bread-and-butter work has been 
 PHP on a LAMP stack.  Apps aren't written in that, so new language time.
 
 2) The library's selection of mobile devices consists of 2 iPads and a Galaxy 
 tablet.  We don't have phones for testing.  My personal phone is a 
 12-year-old flip phone which doesn't run apps.  Can I get by with emulators?  
 What are some good ones?  The budget for the project is zero, so I don't 
 think dedicated testing devices are in the cards unless I upgrade my own 
 phone, which I probably ought to anyway.
 
 3) What are some best practices for library app design?  We were thinking the 
 key functionality would be personal account management (what have I got 
 checked out, renew my stuff, etc), hours, lab availability, search the 
 catalog, and ask a librarian.  Anything missing?  Too much stuff?
 
 Will Martin
 
 Web Services Librarian
 Chester Fritz Library
 
 P.S.  I sent this a couple days ago and wondered why it hadn't shown up -- 
 only to realize I accidently sent it to j...@code4lib.org rather than the 
 actual list serv address.  Whoops, embarrassing!


Re: [CODE4LIB] What is the real impact of SHA-256? - Updated

2014-10-06 Thread Andrew Anderson
My concern would be more that given proven weaknesses in MD5, do I want to risk 
that 1 in a billion chance that the “right” bit error creeps into an archive 
that manages to not impact the checksum, thus creating the illusion that the 
archive integrity has not been violated?

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Oct 2, 2014, at 18:34, Jonathan Rochkind rochk...@jhu.edu wrote:

 For checksums for ensuring archival integrity, are cryptographic flaws 
 relavent? I'm not sure, is part of the point of a checksum to ensure against 
 _malicious_ changes to files?  I honestly don't know. (But in most systems, 
 I'd guess anyone who had access to maliciously change the file would also 
 have access to maliciously change the checksum!)
 
 Rot13 is not suitable as a checksum for ensuring archival integrity however, 
 because it's output is no smaller than it's input, which is kind of what 
 you're looking for. 
 
 
 From: Code for Libraries [CODE4LIB@LISTSERV.ND.EDU] on behalf of Cary Gordon 
 [listu...@chillco.com]
 Sent: Thursday, October 02, 2014 5:51 PM
 To: CODE4LIB@LISTSERV.ND.EDU
 Subject: Re: [CODE4LIB] What is the real impact of SHA-256? - Updated
 
 +1
 
 MD5 is little better than ROT13. At least with ROT13, you have no illusions.
 
 We use SHA 512 for most work. We don't do finance or national security, so it 
 is a good fit for us.
 
 Cary
 
 On Oct 2, 2014, at 12:30 PM, Simon Spero sesunc...@gmail.com wrote:
 
 Intel skylake processors have dedicated sha instructions.
 See: https://software.intel.com/en-us/articles/intel-sha-extensions
 
 Using a tree hash approach (which is inherently embarrassingly parallel)
 will leave io time dominant. This approach is used by Amazon glacier - see
 http://docs.aws.amazon.com/amazonglacier/latest/dev/checksum-calculations.html
 
 MD5 is broken, and cannot be used for any security purposes. It cannot be
 used for deduplication if any of the files are in the directories of
 security researchers!
 
 If security is not a concern then there are many faster hashing algorithms
 that avoid the costs imposed by the need to defend against adversaries.
 See siphash, murmur, cityhash, etc.
 
 Simon
 On Oct 2, 2014 11:18 AM, Alex Duryee a...@avpreserve.com wrote:
 
 Despite some of its relative flaws, MD5 is frequently selected over SHA-256
 in archives as the checksum algorithm of choice. One of the primary factors
 here is the longer processing time required for SHA-256, though there have
 been no empirical studies calculating that time difference and its overall
 impact on checksum generation and verification in a preservation
 environment.
 
 AVPreserve Consultant Alex Duryee recently ran a series of tests comparing
 the real time and cpu time used by each algorithm. His newly updated white
 paper What Is the Real Impact of SHA-256? presents the results and comes
 to some interesting conclusions regarding the actual time difference
 between the two and what other factors may have a greater impact on your
 selection decision and file monitoring workflow. The paper can be
 downloaded for free at
 
 http://www.avpreserve.com/papers-and-presentations/whats-the-real-impact-of-sha-256/
 .
 __
 
 Alex Duryee
 *AVPreserve*
 350 7th Ave., Suite 1605
 New York, NY 10001
 
 office: 917-475-9630
 
 http://www.avpreserve.com
 Facebook.com/AVPreserve http://facebook.com/AVPreserve
 twitter.com/AVPreserve
 


Re: [CODE4LIB] LibGuides v2 - Templates and Nav

2014-09-18 Thread Andrew Anderson
There are ways around this, e.g. http://api.jquerymobile.com/taphold/

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Sep 17, 2014, at 21:17, Jonathan Rochkind rochk...@jhu.edu wrote:

 Mouse hover is not available to anyone using a touch device rather than a 
 mouse, as well as being problematic for keyboard access.
 
 While there might be ways to make the on-hover UI style keyboard accessible 
 (perhaps in some cases activating on element focus in addition toon hover), 
 there aren't really any good ones I can think for purely touch devices (which 
 don't really trigger focus state either).
 
 An increasing amount of web use, of course, is mobile touch devices, and 
 probably will continue to be and to increase for some time, including on 
 library properties.
 
 So I think probably on-hover UI should simply be abandoned at this point, 
 even if some people love it, it will be inaccessible to an increasing portion 
 of our users with no good accomodations.
 
 Jonathan
 
 On 9/17/14 4:25 PM, Jesse Martinez wrote:
 On the same token, we're making it a policy to not use mouse hover over
 effects to display database/asset descriptions in LG2 until this can become
 keyboard accessible. This is a beloved feature from LG1 so I'm hoping
 SpringShare read my pestering emails about this...
 
 Jesse
 
 On Wed, Sep 17, 2014 at 3:38 PM, Brad Coffield bcoffield.libr...@gmail.com
 wrote:
 
 Johnathan,
 
 That point is well taken. Accessibility, to me, shouldn't be a tacked-on
 we'll do the best we can sort of thing. It's an essential part of being a
 library being open to all users. Unfortunately I know our site has a lot of
 work to be done regarding accessibility. I'll also pay attention to that
 when/if I make mods to the v2 templates.
 
 On Wed, Sep 17, 2014 at 1:49 PM, Jonathan LeBreton lebre...@temple.edu
 wrote:
 
 I might mention here that we (Temple University)  found LibGuides 2.0  to
 offer some noteworthy improvements in section 508 accessibility
 when compared with version 1.0.   Accessibility is a particular point of
 concern for the whole institution as we look across the city, state, and
 country at other institutions that have been called out and settled with
 various disability advocacy groups.
 So we moved to v. 2.0 during the summer in order to have those
 improvements in place for the fall semester, as well as to get the value
 from some other developments in v. 2.0 that benefit all customers.
 
 When I see email on list about making  modifications to templates and
 such, it gives me a bit of concern on this score that by doing so,  one
 might easily begin to make the CMS framework for content less accessible.
   I thought I should voice that.This is not to say that one shouldn't
 customize and explore enhancements etc.,  but one should do so with some
 care if you are operating with similar mandates or concerns.Unless I
 am
 mistaken, several of the examples noted are now throwing 508 errors that
 are not in the out-of-the box  LibGuide templates and which are not the
 result of an individual content contributor/author inserting bad stuff
 like images without alt tags.
 
 
 
 
 Jonathan LeBreton
 Senior Associate University Librarian
 Editor:  Library  Archival Security
 Temple University Libraries
 Paley M138,  1210 Polett Walk, Philadelphia PA 19122
 voice: 215.204.8231
 fax: 215.204.5201
 mobile: 215.284.5070
 email:  lebre...@temple.edu
 email:  jonat...@temple.edu
 
 -Original Message-
 From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of
 Cindi Blyberg
 Sent: Wednesday, September 17, 2014 12:03 PM
 To: CODE4LIB@LISTSERV.ND.EDU
 Subject: Re: [CODE4LIB] LibGuides v2 - Templates and Nav
 
 Hey everyone!
 
 Not to turn C4L into Support4LibGuides, but... :)
 
 The infrastructure for all the APIs is in place; currently, the Guides
 API
 and the Subjects API are functioning.  Go to Tools  API  Get Guides to
 see the general structure of the URL.  Replace guides with subjects
 to
 retrieve your subjects.  You will need your LibGuides site ID, which you
 can get from the LibApps Dashboard screen.
 
 Word is that it will not take long to add other API calls on the back
 end;
 if you need these now, please do email supp...@springshare.com and
 reference this conversation.
 
 As for v1, we are planning on supporting it for 2 more years--that said,
 we would never leave anyone hanging, so if it takes longer than that to
 get
 everyone moved over, we're ready for that.
 
 Best,
  -Cindi
 
 On Wed, Sep 17, 2014 at 10:46 AM, Nadaleen F Tempelman-Kluit 
 n...@nyu.edu
 
 wrote:
 
 Hi all-
 While we're on the topic of LibGuides V2, when will the GET subjects
 API (and other API details) be in place? We're in a holding pattern
 until we get those details and we've not been able to get any timeline
 as to when those assets

Re: [CODE4LIB] Anybody know a way to add a MARC tag on-mass to a file of MARC records

2014-08-28 Thread Andrew Anderson
I’ve had a lot of success with pymarc for this.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Aug 28, 2014, at 14:37, Schwartz, Raymond schwart...@wpunj.edu wrote:

 I need to automate this in a script.  As far as I can tell.  You cannot do 
 this with MarcEdit.
 
 -Original Message-
 From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Jane 
 Costanza
 Sent: Thursday, August 28, 2014 2:33 PM
 To: CODE4LIB@LISTSERV.ND.EDU
 Subject: Re: [CODE4LIB] Anybody know a way to add a MARC tag on-mass to a 
 file of MARC records
 
 MarcEdit is a free MARC editing utility.
 
 http://marcedit.reeset.net/
 
 Jane Costanza
 Associate Professor/Head of Discovery Services Trinity University San 
 Antonio, Texas
 210-999-7612
 jcost...@trinity.edu
 http://digitalcommons.trinity.edu/
 http://lib.trinity.edu/
 
 
 On Thu, Aug 28, 2014 at 1:26 PM, Schwartz, Raymond schwart...@wpunj.edu
 wrote:
 
 Anybody know a way to add a MARC tag on-mass to a file of MARC 
 records.  I need to add the tag 918 $a with the contents DELETE to 
 each of the records.
 
 Thanks in advance. /Ray
 
 Ray Schwartz
 Systems Specialist Librarian schwart...@wpunj.edu
 blocked::mailto:schwart...@wpunj.edu
 David and Lorraine Cheng Library   Tel: +1 973 720-3192
 William Paterson University Fax: +1 973 720-2585
 300 Pompton RoadMobile: +1 201
 424-4491
 Wayne, NJ 07470-2103 USA
 http://nova.wpunj.edu/schwartzr2/
 http://euphrates.wpunj.edu/faculty/schwartzr2/
 


Re: [CODE4LIB] Does 'Freedom to Read' require us to systematically privilege HTTPS over HTTP?

2014-06-18 Thread Andrew Anderson
On Jun 17, 2014, at 17:09, Stuart Yeates stuart.yea...@vuw.ac.nz wrote:

 On 06/17/2014 08:49 AM, Galen Charlton wrote:
 On Sun, Jun 15, 2014 at 4:03 PM, Stuart Yeates stuart.yea...@vuw.ac.nz 
 wrote:
 As I read it, 'Freedom to Read' means that we have to take active steps to
 protect that rights of our readers to read what they want and  in private.
 [snip]
 * building HTTPS Everywhere-like functionality into LMSs (such functionality
 may already exist, I'm not sure)
 
 Many ILSs can be configured to require SSL to access their public
 interfaces, and I think it would be worthwhile to encourage that as a
 default expectation for discovery interfaces.
 
 However, I think that's only part of the picture for ILSs.  Other
 parts would include:
 
 * staff training on handling patron and circulation data
 * ensuring that the ILS has the ability to control (and let users
 control) how much circulation and search history data gets retained
 * ensuring that the ILS backup policy strikes the correct balance
 between having enough for disaster recovery while not keeping
 individually identifiable circ history forever
 * ensuring that contracts with ILS hosting providers and services that
 access patron data from the ILS have appropriate language concerning
 data retention and notification of subpoenas.
 
 Compared to other contributors to this thread, I appear to be (a) less 
 worried about state actors than our commercial partners and (b) keener to see 
 relatively straight forward technical fixes that just work 'for free' across 
 large classes of library systems. Things like:
 
 * An ILS module that pulls the HTTPS Everywhere ruleset from 
 https://gitweb.torproject.org/https-everywhere.git/tree/HEAD:/src/chrome/content/rules
  and applies those rules as a standard data-cleanup step on all imported data 
 (MARC, etc).
 
 * A plugin to the CMS that drives the library's websites / blogs / whatever 
 and uses the same rulesets to default all links to HTTPS.
 
 * An EzProxy plugin (or howto) on silently redirectly users to HTTPS over 
 HTTP sites.
 
 cheers
 stuart

This is something that I have been interested in as well, and I have been 
asking our content providers when they will make their content available via 
HTTPS, but so far with very little uptake.  Perhaps if enough customers start 
asking, it will get enough exposure internally to drive adoption of HTTPS for 
the content side.

I looked into what EZproxy offers for the user side, and that product does not 
currently have the ability to do HTTPS to HTTP proxying, even though there is 
no technical reason why it could not be done (look at how many HTTPS sites run 
Apache in a reverse proxy to HTTP servers internally for load balancing, etc.)  

EZproxy makes the assumption that a HTTP resource will always be accessed over 
HTTP, and you cannot configure a HTTPS entry point to HTTP services to at least 
secure the side of the communication channel that is going to contain more 
identifiable information about the user, before it becomes aggregated into the 
general proxy stream.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes


Re: [CODE4LIB] Does 'Freedom to Read' require us to systematically privilege HTTPS over HTTP?

2014-06-18 Thread Andrew Anderson
EZproxy already handles HTTPS connections for HTTPS enabled services today, and 
on modern hardware (i.e. since circa 2005), cryptographic processing far 
surpasses the speed of most network connections, so I do not accept the “it’s 
too heavy” argument against it supporting the HTTPS to HTTP functionality.  
Even embedded systems with 500MHz CPUs can terminate SSL VPNs at over 100Mb/s 
these days.

All I am saying is that the model where you expose HTTPS to the patron and 
still continue to use HTTP for the vendor is not possible with EZproxy today, 
and there is no technical reason why it could not do so, but rather a policy 
decision.  While HTTPS to HTTP translation would not completely solve the 
entire point of the original posting, it would be a step in the right direction 
until the rest of the world caught up.

As an aside, the lightweight nature of EZproxy seems to be becoming its 
Achilles Heel these days, as modern web development methods seem to be pushing 
the boundaries of its capabilities pretty hard.  The stance that EZproxy only 
supports what it understands is going to be a problem when vendors adopt 
HTTP/2.0, SDCH encoding, web sockets, etc., just as AJAX caused issues 
previously.  Most vendor platforms are Java based, and once Jetty starts 
supporting these features, the performance chasm between dumbed-down proxy 
connections and direct connections is going to become even more significant 
than it is today.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Jun 18, 2014, at 11:20, Cary Gordon listu...@chillco.com wrote:

 One of the reasons that EZProxy is so fast and resource-efficient is that
 it is very lightweight. HTTPS to HTTP processing would require that
 EZProzy, or another proxy layer behind it, provide an HTTPS endpoint.
 Building this into EZProxy, I think, would not be a good fit for
 their model.
 
 I think that it would be simpler to just do everything in nginx, or
 possibly node.
 
 Cary
 
 On Wednesday, June 18, 2014, Andrew Anderson and...@lirn.net wrote:
 
 On Jun 17, 2014, at 17:09, Stuart Yeates stuart.yea...@vuw.ac.nz
 javascript:; wrote:
 
 On 06/17/2014 08:49 AM, Galen Charlton wrote:
 On Sun, Jun 15, 2014 at 4:03 PM, Stuart Yeates stuart.yea...@vuw.ac.nz
 javascript:; wrote:
 As I read it, 'Freedom to Read' means that we have to take active
 steps to
 protect that rights of our readers to read what they want and  in
 private.
 [snip]
 * building HTTPS Everywhere-like functionality into LMSs (such
 functionality
 may already exist, I'm not sure)
 
 Many ILSs can be configured to require SSL to access their public
 interfaces, and I think it would be worthwhile to encourage that as a
 default expectation for discovery interfaces.
 
 However, I think that's only part of the picture for ILSs.  Other
 parts would include:
 
 * staff training on handling patron and circulation data
 * ensuring that the ILS has the ability to control (and let users
 control) how much circulation and search history data gets retained
 * ensuring that the ILS backup policy strikes the correct balance
 between having enough for disaster recovery while not keeping
 individually identifiable circ history forever
 * ensuring that contracts with ILS hosting providers and services that
 access patron data from the ILS have appropriate language concerning
 data retention and notification of subpoenas.
 
 Compared to other contributors to this thread, I appear to be (a) less
 worried about state actors than our commercial partners and (b) keener to
 see relatively straight forward technical fixes that just work 'for free'
 across large classes of library systems. Things like:
 
 * An ILS module that pulls the HTTPS Everywhere ruleset from
 https://gitweb.torproject.org/https-everywhere.git/tree/HEAD:/src/chrome/content/rules
 and applies those rules as a standard data-cleanup step on all imported
 data (MARC, etc).
 
 * A plugin to the CMS that drives the library's websites / blogs /
 whatever and uses the same rulesets to default all links to HTTPS.
 
 * An EzProxy plugin (or howto) on silently redirectly users to HTTPS
 over HTTP sites.
 
 cheers
 stuart
 
 This is something that I have been interested in as well, and I have been
 asking our content providers when they will make their content available
 via HTTPS, but so far with very little uptake.  Perhaps if enough customers
 start asking, it will get enough exposure internally to drive adoption of
 HTTPS for the content side.
 
 I looked into what EZproxy offers for the user side, and that product does
 not currently have the ability to do HTTPS to HTTP proxying, even though
 there is no technical reason why it could not be done (look at how many
 HTTPS sites run Apache in a reverse proxy to HTTP servers internally for
 load balancing, etc.)
 
 EZproxy makes the assumption that a HTTP resource

Re: [CODE4LIB] Windows XP EOL

2014-03-05 Thread Andrew Anderson
You’d be amazed at what you can do with port 80/443 access, so while that is a 
deterrent, it is not a solution that will make any guarantees that the machines 
cannot do anything nefarious.

Adding a proxy server in front of the machines with a whitelist of allowed web 
sites instead of NAT would go further, but at the end of that day you’re still 
talking about taking a 14 year old operating system that is no longer supported 
and connecting it to the internet.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Mar 5, 2014, at 7:20, Michael Bond mb...@the-forgotten.org wrote:

 Why not setup your XP boxes to use a private network (10.x.x.x or 
 192.168.x.x) and put them behind a heavily fire walled NAT solution. Could be 
 setup on the network level or with a router and a linux box running IP 
 tables. Lots of ways to do it. 
 
 Install and keep updated Firefox or Chrome, lock down the machines so that 
 users don’t have permissions to install anything, and setup a whitelist of 
 programs that are allowed to be run (takes a little bit of work, but its very 
 doable. We did this in WVU Libraries on all our machines [500 or so], public 
 and staff, until we got our virtualized desktops in place). 
 
 You can’t disallow Internet Explorer from running, but you can limit the 
 websites that it is allowed to visit. You could even go as far as only 
 allowing it to connect to the local host, but likely anything ‘on campus’ 
 would be fine.
 
 I’m assuming you are using some sort of image management solution (Ghost, at 
 the very least). So once you get an image setup it shouldn’t be that bad to 
 maintain and deploy. And if something does become exploited, you can can 
 re-image the machine. 
 
 Configure the NAT to not allow any traffic to come from that private network 
 other than ports 80 and 443 (and any other legitimate port that you need). 
 that way if a machine does become compromised it can’t do (much) harm outside 
 of your private XP network. 
 
 If you need AD authentication you can set that all up in the ACLs for the 
 network as well so that they can only contact a specific authentication 
 server. If you absolutely needed to you could even put an auth server on the 
 same private network that has a trust back to your main auth servers. Put 2 
 network interfaces in it and it can live on 2 networks so you don’t have to 
 poke a hole through your private networks ACLs to get back to the main auth 
 servers. 
 
 Its not an ideal situation, but if you can’t afford new machines and you 
 absolutely need to keep your XP machines running there are ways of doing it. 
 But at what point does it become cost prohibitive with your time compared to 
 investing in new hardware?
 
 If you don’t do something though, you’ll be spending all your time rebuilding 
 compromised XP boxes eventually. 
 
 Michael Bond
 mb...@the-forgotten.org
 
 
 
 On Mar 4, 2014, at 4:55 PM, Riley Childs rchi...@cucawarriors.com wrote:
 
 Not to stomp around, but 1 hour is a LONG time for an unpatched computer, 
 especially when in close proximity to other unpatched computers! DeepFreeze 
 is great, but it is not a long term solution, also starting next week you 
 will get a nag screen every time you login telling you about the EOL.
 
 Riley Childs
 Student
 Asst. Head of IT Services
 Charlotte United Christian Academy
 (704) 497-2086
 RileyChilds.net
 Sent from my Windows Phone, please excuse mistakes
 
 From: Benjamin Stewartmailto:benjamin.stew...@unbc.ca
 Sent: ‎3/‎4/‎2014 4:46 PM
 To: CODE4LIB@LISTSERV.ND.EDUmailto:CODE4LIB@LISTSERV.ND.EDU
 Subject: Re: [CODE4LIB] Windows XP EOL
 
 Hello everyone
 
 (I have been in IT for 25+ years, k-7 for 15 years and now 10 months UNBC
 Library)
 
 
 If I worked for an organization that did not have the money to go either
 replacement Win7 or Linux desktop for usability issues.
 
 I would contact Faronics and get a deal for educational licenses to
 install Deepfreeze.
 Then setup all workstation basic accounts and to reboot if idle for 1
 hour. (and shut down, startup between set times)
 Deepfreeze also has a remote console to unfreeze and refreeze for
 maintenance to the workstation. (e.g. browser updates flash adobe)
 This in hand with PDQ deploy/inventory works very nice. (Basic version
 free)
 
 
 Last option would (no possible for most places) contact the Dell official
 lease site via direct or eBay. (there is a Canada and US supplier)
 
 You can by nice 780 Dell with win7 pro for about $140 with shipping.
 Some companies like Dell of HP have be know to also donate to non-profit.
 
 ~Ben
 
 System Administrator
 Geoffrey R. Weller library
 UNBC, BC Canada
 PH (250) 960-6605
 benjamin.stew...@unbc.ca
 
 
 
 
 
 
 
 On 2014-03-04, 11:12 AM, Ingraham Dwyer, Andy adw...@library.ohio.gov
 wrote:
 
 I would

Re: [CODE4LIB] Windows XP EOL

2014-03-05 Thread Andrew Anderson
On Mar 5, 2014, at 15:37, Marc Truitt mtru...@mta.ca wrote:

 Perhaps that's why several contributors to this thread have suggested
 that M$' EOL declaration aside, why give it up?  XP, I'll miss ya...

XP: The new DOS 3.3?

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes


Re: [CODE4LIB] Proquest search api?

2014-02-17 Thread Andrew Anderson
The document you want to request from ProQuest support was called 
Federated-Search.docx when they sent it to me.  This will address many of your 
documentation needs.

ProQuest used to have an excel spreadsheet with all of the product codes for 
the databases available for download from 
http://support.proquest.com/kb/article?ArticleId=3698source=articlec=12cid=26,
 but it appears to no longer be available from that source.  ProQuest support 
should be able to answer where it went when you request the federated search 
document.

You may receive multiple 856 fields for Citation/Abstract, Full Text, and 
Scanned PDF:

=856  41$3Citation/Abstract$uhttp://search.proquest.com/docview/...
=856  40$3Full Text$uhttp://search.proquest.com/docview/...
=856  40$3Scanned PDF$uhttp://search.proquest.com/docview/...

I would suggest that rather than relying on the 2nd indicator, you should parse 
subfield 3 instead to find the format that you prefer.  You see the multiple 
856 fields in the MARC records for ProQuest holdings as well, as that is how 
ProQuest handles coverage gaps in titles, so if you have ever processed 
ProQuest MARC records before, you should be already prepared for this.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Feb 17, 2014, at 10:28, Jonathan Rochkind rochk...@jhu.edu wrote:

 I still haven't managed to get info from Proquest support, but thanks to off 
 list hints from another coder, I have discovered the Proquest SRU endpoint, 
 which I think is the thing they call the XML gateway.
 
 Here's an example query:
 
 http://fedsearch.proquest.com/search/sru/pqdtft?operation=searchRetrieveversion=1.2maximumRecords=30startRecord=1query=title%3D%22global%20warming%22%20AND%20author%3DCastet
 
 For me, coming from an IP address recognized as 'on campus' for our general 
 Proquest access, no additional authentication is required to use this API. 
 I'm not sure if we at some point prior had them activate the XML Gateway 
 for us, likely for a federated search product, or if it's just this way for 
 everyone.
 
 The path component after /sru, pqdtft is the database code for Proquest 
 Dissertations and Theses. I'm not sure where you find a list of these 
 database codes in general; if you've made a succesful API request to that 
 endpoint, there will be a diagnosticMessage element near the end of the 
 response listing all database codes you have access to (but without 
 corresponding full English names, you kind of have to guess).
 
 The value of the 'query' parameter is a valid CQL query, as usual for SRU. 
 Unfortunately, there seems to be no SRU explain response to tell you what 
 fields/operators are available. But guessing often works, title, author, 
 and date are all available -- I'm not sure exactly how 'date' works, need 
 to experiment more. The CQL query param above un-escaped is:
 
 title=global warming AND author=Castet
 
 Responses seem to be in MARCXML, and that seems to be the only option.
 
 It looks like you can tell if a full text is available (on Proquest platform) 
 for a given item, based on whether there's an 856 field with second indicator 
 set to 0 -- that will be a URL to full text. I think. It looks like. Did I 
 mention if there are docs for any of this, I haven't found them?
 
 So, there you go, a Proquest search API!
 
 Jonathan
 
 
 
 On 2/12/14 3:44 PM, Jonathan Rochkind wrote:
 Aha, thinking to google search for proquest z3950 actually got me some
 additional clues!
 
 Sites that are currently using Z39.50 to search ProQuest are advised to
 consider moving to the XML gateway.
 
 in Google snippets for:
 
 http://www.proquest.com/assets/downloads/products/techrequirements_np.pdf
 
 Also If you are using the previous XML
 gateway for access other than with a federated search vendor, please
 contact our support center at
 www.proquest.com/go/migrate and we can get you the new XML gateway
 implementation documentation.
 
 Okay, so now I at least know that something called the XML Gateway
 exists, and that's what I want info on or ask about!  (Why are our
 vendors so reluctant to put info on their services online?)
 
 I am not a huge fan of z3950, and am not ordinarily optimistic about
 it's ability to actually do what I need, but I'd use it if it was all
 that was available; in this case, it seems like Proquest is recommending
 you do NOT use it, but use this mysterious 'XML gateway'.
 
 
 
 On 2/12/14 3:29 PM, Eric Lease Morgan wrote:
 On Feb 12, 2014, at 3:22 PM, Jonathan Rochkind rochk...@jhu.edu wrote:
 
 I feel like at some point I heard there was a search API for the
 Proquest content/database platform.
 
 
 While it may not be the coolest, I’d be willing to bet Proquest
 supports Z39.50. I used it lately to do some interesting queries
 against the New York Times Historical Newspapers Database (index). [1]
 Okay. I know

Re: [CODE4LIB] EZProxy changes / alternatives ?

2014-02-03 Thread Andrew Anderson
For me it’s a little more concrete, and a little less abstract when it comes to 
why a viable alternative to EZproxy is necessary.  It has very little to do 
with the cost of EZproxy itself, and much more to do with support, features, 
and functionality.

There exists a trivial DoS attack against EZproxy that I reported to OCLC about 
2 years ago, and has not been addressed yet.

Native IPv6 support by EZproxy has slipped by years now.  I have patrons using 
IPv6 for access today that I want to provide a better experience than forcing 
them to use a 6to4 gateway at their ISP.

You cannot proxy https to http with EZproxy to secure the patron to proxy side 
of the proxy communication, increasing your patron’s privacy.

I have requested that OCLC make a minor change to their existing AD 
authentication support to enable generic LDAP/Kerberos authentication that was 
denied because “no one wants it”.  Since they support AD, 95% of the code 
required already exists, and would make a lot more sense than some of the other 
authentication schemes that EZproxy already supports.  This closes the door on 
integration with eDirectory, IPA, SUN Directory Server, OpenLDAP, etc. for no 
good reason.

OCLC has been the steward of EZproxy for over 5 years now, and in that time, 
they are yet to fully document the software.  Every few months some new obscure 
configuration option gets discussed on the EZproxy list that I’ve never seen 
before, and I have been working with this software for over a decade now.  This 
is not only limited to existing configuration options, either — there was no 
documentation on the new MimeFilter option when it was first introduced.  I 
would have expected that the IT staff at OCLC that is managing the EZproxy 
service would have demanded full documentation by now, and that documentation 
would have been released to customers as well.

EZproxy does not cluster well.  The peering support is functional, but not 
seamless when there is a failure.  When a proxy in the server pool goes down, 
the patron is prompted for authentication again when they land on a new proxy 
server, since EZproxy does not share session state.  External load balancers 
cannot fix this problem, either, for the same reason.

EZproxy does not support gzip compression, causing library access use an 
additional 80-90% bandwidth for textual content (HTML, CSS, JS, etc).

EZproxy does not support caching, causing library access to use an additional 
30-50% additional bandwidth for cacheable web assets. (And yes, you can park a 
cache in front of EZproxy to offset this, which is how I collected the 30-50% 
numbers, but doing so breaks the “it’s easy and just works” model that EZproxy 
promises.)

Combine the lack of gzip support with the lack of caching support, and you are 
looking at around a 60-80% overall increase in bandwidth consumption.  When you 
have a user community measured in hundreds of users, things like gzip 
compression and caching may not matter as much, but when your user community is 
measured in the hundreds of thousands of patrons, these things really do 
matter, and mean the difference between doubling your bandwidth costs this 
year, or deferring that expense 5-7 years down the road.

So it’s not _just_ $500 per year when you take a step back and look at the 
bigger picture.  It’s $500 per year, plus the per Mb cost of your internet 
connection — both inbound and outbound — which can be measured in hundreds of 
dollars per month for larger sites.  If you could could cut that by 2/3 just by 
switching to a different proxy solution, that might get your attention, even if 
you shifted the $500/yr support costs to a different entity.  

Imagine never hearing “wow this library network is slow” again because a web 
page that used to load 1MB of content was able to gzip that down to 600KB, and 
300KB of that content was served off the local proxy server, leaving just 300KB 
to pull off the remote server.  How much is a better user experience worth to 
you?

Bottom line: competition is good.  Just look at how Internet Explorer is almost 
a sane browser now, thanks largely to competition from Firefox and Chrome.  If 
coming up with a viable alternative to EZproxy using open source tools causes a 
security, features, and functionality arms race, then everyone wins.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Jan 31, 2014, at 18:43, Kyle Banerjee kyle.baner...@gmail.com wrote:

 On Fri, Jan 31, 2014 at 3:10 PM, Salazar, Christina 
 christina.sala...@csuci.edu wrote:
 
 I think though that razor thin budgets aside, the EZProxy using community
 is vulnerable to what amounts to a monopoly. Don't get any ideas, OCLC
 peeps (just kiddin') but now we're so captive to EZProxy, what are our
 options if OCLC wants to gradually (or not so gradually) jack up the price?
 
 Does

Re: [CODE4LIB] EZProxy changes / alternatives ?

2014-01-31 Thread Andrew Anderson
EZproxy is a self-installing statically compiled single binary download, with a 
built-in administrative interface that makes most common administrative tasks 
point-and-click, that works on Linux and Windows systems, and requires very 
little in the way of resources to run.  It also has a library of a few hundred 
vendor stanzas that can be copied and pasted and work the majority of the time.

To successfully replace EZproxy in this setting, it would need to be packaged 
in such a way that it is equally easy to install and maintain, and the library 
of vendor stanzas would need to be developed as apache conf.d files.

Re: nginx from another reply in this thread, I am keeping my eye on it for 
future projects, but one thing it does not have currently is the wealth of 
Apache modules.  Some of the authentication that is commonly used in a library 
setting are supported by existing Apache modules, while nginx does not support 
them. Since it was developed with a different set of priorities, supporting 
things like Athens/CAS/SAML were not the main focus of nginx historically.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Jan 31, 2014, at 12:43, Timothy Cornwell tc...@cornell.edu wrote:

 I have an IT background and some apache proxy experience, and it seems fairly 
 easy - for me.  I understand it may not be for libraries with limited IT 
 resources.  I am not at all familiar with EZProxy, so I have to ask:
 
 What is it about EZProxy that makes it attractive for those libraries with 
 limited IT resources?
 
 -T
 
 
 
 -Original Message-
 From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Kyle 
 Banerjee
 Sent: Friday, January 31, 2014 12:14 PM
 To: CODE4LIB@LISTSERV.ND.EDU
 Subject: Re: [CODE4LIB] EZProxy changes / alternatives ?
 
 Many good ideas in this thread.
 
 One thing I'd just like to throw out there is that there are some ideas that 
 may be good to distribute in the form of virtual machines and this might be 
 one of them.
 
 Proxying is needed by practically all libraries and takes little in terms of 
 systems resources. But many libraries with limited IT resources would have 
 trouble implementing alternatives to ezproxy -- especially if they have to 
 use authentication features not supported by Apache HTTPD. Even for those who 
 do have enough staff time, it seems kind of nuts to have everyone spending 
 time solving the same problems.
 
 kyle
 
 
 On Fri, Jan 31, 2014 at 5:43 AM, Ryan Eby ryan...@gmail.com wrote:
 
 There was actually a breakout in 2011? Code4lib discussing Apache and 
 using it as a proxy. I believe Terry Reese and Jeremy Frumkin, then 
 from Oregon?, were the ones leading it. There was lots of interest but 
 I'm not sure if anything took off or if they have documentation 
 somewhere of how far they got. I remember it being about getting 
 something a consortia of libraries could use together so may have been 
 more complex requirements than what is looked for here.
 
 
 http://wiki.code4lib.org/index.php/Can_we_hack_on_this:_Open_Extensibl
 e_Proxy:_going_beyond_EZProxy%3F
 
 --
 Ryan Eby
 


Re: [CODE4LIB] EZProxy changes / alternatives ?

2014-01-29 Thread Andrew Anderson
 for testing:

Location “/badpath”
ProxyHTMLEnable Off
SetOutputFilter INFLATE;dummy-html-to-plain
ExtFilterOptions LogStdErr Onfail=remove
/Location
ExtFilterDefine dummy-html-to-plain mode=output intype=text/html 
outtype=text/plain cmd=“/bin/cat -“

So what’s currently missing in the Apache HTTPd solution?

- Services that use an authentication token (predominantly ebook vendors) need 
special support written.  I have been entertaining using mod_lua for this to 
make this support relatively easy for someone who is not hard-core technical to 
maintain.

- Services that are not IP authenticated, but use one of the Form-based 
authentication variants.  I suspect that an approach that injects a script tag 
into the page pointing to javascript that handles the form fill/submission 
might be a sane approach here.  This should also cleanly deal with the ASP.net 
abominations that use __PAGESTATE to store sessions client-side instead of 
server-side.

- EZproxy’s built-in DNS server (enabled with the “DNS” directive) would need 
to be handled using a separate DNS server (there are several options to choose 
from).

- In this setup, standard systems-level management and reporting tools would be 
used instead of the /admin interface in EZproxy

- In this setup, the functionality of the EZproxy /menu URL would need to be 
handled externally.  This may not be a real issue, as many academic sites 
already use LMS or portal systems instead of the EZproxy to direct students to 
resources, so this feature may not be as critical to replicate.

- And of course, extensive testing.  While the above ProQuest stanza works for 
the main ProQuest search interface, it won’t work for everyone, everywhere just 
yet.

Bottom line: Yes, Apache HTTPd is a viable EZproxy alternative if you have a 
system administrator who knows their way around Apache HTTPd, and are willing 
to spend some time getting to know your vendor services intimately.

All of this testing was done on Fedora 19 for the 2.4 version of HTTPd, which 
should be available in RHEL7/CentOS7 soon, so about the time that hard 
decisions are to be made regarding EZproxy vs something else, that something 
else may very well be Apache HTTPd with vendor-specific configuration files.

-- 
Andrew Anderson, Director of Development, Library and Information Resources 
Network, Inc.
http://www.lirn.net/ | http://www.twitter.com/LIRNnotes | 
http://www.facebook.com/LIRNnotes

On Jan 29, 2014, at 14:42, Margo Duncan mdun...@uttyler.edu wrote:

 Would you *have* to be hosted? We're in a rural part of the USA and network 
 connections from here to anywhere aren't great, so we try to host most 
 everything we can.  EZProxy really is EZ to host yourself.
 
 Margo
 
 -Original Message-
 From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of 
 stuart yeates
 Sent: Wednesday, January 29, 2014 1:40 PM
 To: CODE4LIB@LISTSERV.ND.EDU
 Subject: Re: [CODE4LIB] EZProxy changes / alternatives ?
 
 The text I've seen talks about [e]xpanded reporting capabilities to support 
 management decisions in forthcoming versions and encourages towards the 
 hosted solution.
 
 Since we're in .nz, they'd put our hosted proxy server in .au, but the 
 network connection between .nz and .au is via the continental .us, which puts 
 an extra trans-pacific network loop in 99% of our proxied network connections.
 
 cheers
 stuart
 
 On 30/01/14 03:14, Ingraham Dwyer, Andy wrote:
 OCLC announced in April 2013 the changes in their license model for North 
 America.  EZProxy's license moves from requiring a one-time purchase of 
 US$495 to a *annual* fee of $495, or through their hosted service, with the 
 fee depending on scale of service.  The old one-time purchase license is no 
 longer offered for sale as of July 1, 2013.  I don't have any details about 
 pricing for other parts of the world.
 
 An important thing to recognize here, is that they cannot legally change the 
 terms of a license that is already in effect.  The software you have 
 purchased under the old license is still yours to use, indefinitely.  OCLC 
 has even released several maintenance updates during 2013 that are available 
 to current license-holders.  In fact, they released V5.7 in early January 
 2014, and made that available to all license-holders.  However, all updates 
 after that version are only available to holders of the yearly subscription. 
  The hosted product is updated to the most current version automatically.
 
 My recommendation is:  If your installation of EZProxy works, don't change 
 it.  Yet.  Upgrade your installation to the last version available under the 
 old license, and use that for as long as you can.  At this point, there are 
 no world-changing new features that have been added to the product.  There 
 is speculation that IPv6 support will be the next big feature-add, but I 
 haven't heard anything official.  Start