Re: [CODE4LIB] rdf triplestores

2013-11-11 Thread Ethan Gruber
I've been using Apache Fuseki (
http://jena.apache.org/documentation/serving_data/) for almost a year, in
production since the spring.  It's a SPARQL server with a built in TBD.
It's easy to use, and takes about 5 minutes to get working on your desktop
or server.

Ethan


On Mon, Nov 11, 2013 at 1:17 AM, Richard Wallis 
richard.wal...@dataliberate.com wrote:

 I've had some success with 4Store: http://4store.org

 Used it on mac laptop to load the WorldCat most highly held resources:
 http://dataliberate.com/2012/08/putting-worldcat-data-into-a-triple-store/

 As to the point about loading RDF/XML, especially if you have a large
 amount of data.

- Triplestores much prefer raw triples for large amounts of data
- Chopping up files of triples into smaller chunks is also often
beneficial as it reduces memory footprints and can take advantage of
multithreading.  It is also far easier to recover from errors such as
 bad
data etc.
- A bit of unix command line wizardry (split followed a simple for-loop)
is fairly standard practice

 Also raw triples are often easier to produce - none of that mucking about
 producing correctly formatted XML - and you can chop, sort, and play about
 with them using powerful unix command line tools.

 ~Richard.


 On 11 November 2013 18:19, Scott Turnbull scott.turnb...@aptrust.org
 wrote:

  I've primarily used Sesame myself.  The http based queries made it pretty
  easy to script against.
 
  http://www.openrdf.org/
 
 
  On Mon, Nov 11, 2013 at 12:12 AM, Eric Lease Morgan emor...@nd.edu
  wrote:
 
   What is your favorite RDF triplestore?
  
   I am able to convert numerous library-related metadata formats into
   RDF/XML. In a minimal way, I can then contribute to the Semantic Web by
   simply putting the resulting files on an HTTP file system. But if I
 were
  to
   import my RDF/XML into a triplestore, then I could do a lot more. Jena
   seems like a good option. So does Openlink Virtuoso.
  
   What experience do y'all have with these tools, and do you know how to
   import RDF/XML into them?
  
   --
   Eric Lease Morgan
  
 
 
 
  --
  *Scott Turnbull*
  APTrust Technical Lead
  scott.turnb...@aptrust.org
  www.aptrust.org
  678-379-9488
 



 --
 Richard Wallis
 Founder, Data Liberate
 http://dataliberate.com
 Tel: +44 (0)7767 886 005

 Linkedin: http://www.linkedin.com/in/richardwallis
 Skype: richard.wallis1
 Twitter: @rjw



Re: [CODE4LIB] rdf triplestores

2013-11-11 Thread Stefano Bargioni
My +1 for Joseki.
sb

On 11/nov/2013, at 06.12, Eric Lease Morgan wrote:

 What is your favorite RDF triplestore?
 
 I am able to convert numerous library-related metadata formats into RDF/XML. 
 In a minimal way, I can then contribute to the Semantic Web by simply putting 
 the resulting files on an HTTP file system. But if I were to import my 
 RDF/XML into a triplestore, then I could do a lot more. Jena seems like a 
 good option. So does Openlink Virtuoso. 
 
 What experience do y'all have with these tools, and do you know how to import 
 RDF/XML into them?
 
 -- 
 Eric Lease Morgan
 


Re: [CODE4LIB] rdf triplestores

2013-11-11 Thread Ross Singer
I've used Fuseki a lot and really like it, although configuration for
things like LARQ (full text indexing) historically has been a little
underdocumented (and it can be a little difficult to understand what
component is in charge of what task).

4-Store is super simple to get up and running with, as well, but I haven't
used it in production for anything.

-Ross.


On Mon, Nov 11, 2013 at 8:52 AM, Stefano Bargioni bargi...@pusc.it wrote:

 My +1 for Joseki.
 sb

 On 11/nov/2013, at 06.12, Eric Lease Morgan wrote:

  What is your favorite RDF triplestore?
 
  I am able to convert numerous library-related metadata formats into
 RDF/XML. In a minimal way, I can then contribute to the Semantic Web by
 simply putting the resulting files on an HTTP file system. But if I were to
 import my RDF/XML into a triplestore, then I could do a lot more. Jena
 seems like a good option. So does Openlink Virtuoso.
 
  What experience do y'all have with these tools, and do you know how to
 import RDF/XML into them?
 
  --
  Eric Lease Morgan
 



Re: [CODE4LIB] rdf triplestores

2013-11-11 Thread Kevin Ford
I'll second Richard on this.   4store is fairly quick to set up and get 
going.  It comes with command-line tools and an HTTP option.


FWIW, ID.LOC.GOV uses 4store in its stack.

Yours,
Kevin


On 11/11/2013 01:17 AM, Richard Wallis wrote:

I've had some success with 4Store: http://4store.org

Used it on mac laptop to load the WorldCat most highly held resources:
http://dataliberate.com/2012/08/putting-worldcat-data-into-a-triple-store/

As to the point about loading RDF/XML, especially if you have a large
amount of data.

- Triplestores much prefer raw triples for large amounts of data
- Chopping up files of triples into smaller chunks is also often
beneficial as it reduces memory footprints and can take advantage of
multithreading.  It is also far easier to recover from errors such as bad
data etc.
- A bit of unix command line wizardry (split followed a simple for-loop)
is fairly standard practice

Also raw triples are often easier to produce - none of that mucking about
producing correctly formatted XML - and you can chop, sort, and play about
with them using powerful unix command line tools.

~Richard.


On 11 November 2013 18:19, Scott Turnbull scott.turnb...@aptrust.orgwrote:


I've primarily used Sesame myself.  The http based queries made it pretty
easy to script against.

http://www.openrdf.org/


On Mon, Nov 11, 2013 at 12:12 AM, Eric Lease Morgan emor...@nd.edu
wrote:


What is your favorite RDF triplestore?

I am able to convert numerous library-related metadata formats into
RDF/XML. In a minimal way, I can then contribute to the Semantic Web by
simply putting the resulting files on an HTTP file system. But if I were

to

import my RDF/XML into a triplestore, then I could do a lot more. Jena
seems like a good option. So does Openlink Virtuoso.

What experience do y'all have with these tools, and do you know how to
import RDF/XML into them?

--
Eric Lease Morgan





--
*Scott Turnbull*
APTrust Technical Lead
scott.turnb...@aptrust.org
www.aptrust.org
678-379-9488







Re: [CODE4LIB] rdf triplestores

2013-11-11 Thread Jason Stirnaman
Eric,
We just did a workshop at C4LMidwest on getting up and running with Fuseki and 
RDF/XML. Here's the 3-part tutorial (for OS X, but translates easily to Linux):
http://jstirnaman.wordpress.com/2013/10/11/installing-fuseki-with-jena-and-tdb-on-os-x/

Jason

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Eric 
Lease Morgan
Sent: Sunday, November 10, 2013 11:12 PM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: [CODE4LIB] rdf triplestores

What is your favorite RDF triplestore?

I am able to convert numerous library-related metadata formats into RDF/XML. In 
a minimal way, I can then contribute to the Semantic Web by simply putting the 
resulting files on an HTTP file system. But if I were to import my RDF/XML into 
a triplestore, then I could do a lot more. Jena seems like a good option. So 
does Openlink Virtuoso. 

What experience do y'all have with these tools, and do you know how to import 
RDF/XML into them?

-- 
Eric Lease Morgan


Re: [CODE4LIB] We should use HTTPS on code4lib.org

2013-11-11 Thread Ordway, Ryan
HTTPS prevents passive monitoring at the application level, but there is
still nothing stopping the government from issuing a subpoena for the
webserver log files. They can still see what you're doing at the network
level, granted they can only see source and destination IPs and ports. With
enough analysis, some level of usage can be inferred.

Unless we want to turn off the access logs and remove any Google Analytics
tags, there are still ways for usage to be monitored and tracked. Unless
you always delete your browser history or use something like Chrome's
Incognito mode, your browser is tracking you too.

If Internet Archive has issues with HTTPS, it's easy enough to exclude them
from any URL rewriting so they can continue to harvest via HTTP.





On Sun, Nov 10, 2013 at 2:45 AM, Mitar mmi...@gmail.com wrote:

 Hi!

 On Wed, Nov 6, 2013 at 5:37 PM, Riley Childs ri...@tfsgeo.com wrote:
  Why? HTTPS is used when there is sensitive data involved, code4lib.org(at 
  least to my knowledge) does not have sensitive data?

 It is not just about the security of the users but privacy of the
 users as well. Internet Archive moved to HTTPS so that nobody could
 monitor what their users are accessing.


 Mitar

 --
 http://mitar.tnode.com/
 https://twitter.com/mitar_m



Re: [CODE4LIB] rdf triplestores

2013-11-11 Thread Tom Johnson
We use 4Store at Oregon State University. I recommend it as very easy to
put up.

I've gone so far as to launch it live in a 20 minute talk.

- Tom


On Mon, Nov 11, 2013 at 8:52 AM, Kevin Ford k...@3windmills.com wrote:

 I'll second Richard on this.   4store is fairly quick to set up and get
 going.  It comes with command-line tools and an HTTP option.

 FWIW, ID.LOC.GOV uses 4store in its stack.

 Yours,
 Kevin



 On 11/11/2013 01:17 AM, Richard Wallis wrote:

 I've had some success with 4Store: http://4store.org

 Used it on mac laptop to load the WorldCat most highly held resources:
 http://dataliberate.com/2012/08/putting-worldcat-data-into-
 a-triple-store/

 As to the point about loading RDF/XML, especially if you have a large
 amount of data.

 - Triplestores much prefer raw triples for large amounts of data
 - Chopping up files of triples into smaller chunks is also often
 beneficial as it reduces memory footprints and can take advantage of
 multithreading.  It is also far easier to recover from errors such as
 bad
 data etc.
 - A bit of unix command line wizardry (split followed a simple
 for-loop)
 is fairly standard practice

 Also raw triples are often easier to produce - none of that mucking about
 producing correctly formatted XML - and you can chop, sort, and play about
 with them using powerful unix command line tools.

 ~Richard.


 On 11 November 2013 18:19, Scott Turnbull scott.turnb...@aptrust.org
 wrote:

  I've primarily used Sesame myself.  The http based queries made it pretty
 easy to script against.

 http://www.openrdf.org/


 On Mon, Nov 11, 2013 at 12:12 AM, Eric Lease Morgan emor...@nd.edu
 wrote:

  What is your favorite RDF triplestore?

 I am able to convert numerous library-related metadata formats into
 RDF/XML. In a minimal way, I can then contribute to the Semantic Web by
 simply putting the resulting files on an HTTP file system. But if I were

 to

 import my RDF/XML into a triplestore, then I could do a lot more. Jena
 seems like a good option. So does Openlink Virtuoso.

 What experience do y'all have with these tools, and do you know how to
 import RDF/XML into them?

 --
 Eric Lease Morgan




 --
 *Scott Turnbull*
 APTrust Technical Lead
 scott.turnb...@aptrust.org
 www.aptrust.org
 678-379-9488







Re: [CODE4LIB] mass convert jpeg to pdf

2013-11-11 Thread Stuart Snydman
Annotorius has been integrated with OpenLayers [1] to support annotation of 
maps on zoomable images.  Quite excellent work indeed, thanks to Rainer.  

As part of IIIF [2] and Shared Canvas [3] we have been targeting a similar 
OpenSeadragon integration with Annotorius and then making this a 
feature/modality in the Mirador image comparison environment [4]. Part of the 
roadmap for Mirador is to have annotation viewing and making integrated with 
OpenSeadragon (or similarly tiled) zooming.

-Stu

[1] http://annotorious.github.io/demos/openlayers-annotation.html
[2] http://iiif.io/
[3] http://www.shared-canvas.org/
[4] http://iiif.io/mirador/

On Nov 10, 2013, at 12:34 PM, Edward Summers wrote:

 Annotorious [1] is a neat little JavaScript library for adding annotations to 
 an image, and displaying them later. I might be wrong, but it doesn’t appear 
 to support zoomable images at the moment. I do see there was some 
 cross-project activity with OpenSeaDragon [2] so maybe asking over there will 
 yield some leads? Ranier Simon gave a excellent, brief presentation about 
 Annotorious at iAnnotate earlier this year. [3]
 
 Leaflet [4] is widely known as a JavaScript library for doing maps; but the 
 tiling that goes on when displaying maps is very similar to zooming on other 
 images like in OpenSeaDragon. Because it is oriented around maps, it 
 definitely supports drawing paths, polygons, other shapes, and there are lots 
 of plugins [5] for various things, including overlaying stuff over the image 
 with Raphael.
 
 Another thing to look at from the digital library research angle might be the 
 SharedCanvas work [6,7]. I’m similarly curious to hear if other people have 
 done annotation with zoomable interfaces before. 
 
 Wondering out loud a bit: don’t your archivists need to make the annotations 
 on a zoomable interface, even if your end-users don’t?
 
 //Ed
 
 [1] http://annotorious.github.io/
 [2] https://github.com/openseadragon/openseadragon/issues/14
 [3] http://www.youtube.com/watch?v=-HgWIkBeQNM
 [4] http://leafletjs.com/
 [5] http://leafletjs.com/plugins.html
 [6] http://www.shared-canvas.org/
 [7] http://link.springer.com/article/10.1007%2Fs00799-012-0098-8
 
 On Nov 10, 2013, at 9:41 AM, Ethan Gruber ewg4x...@gmail.com wrote:
 Does anyone have experience with an image zooming engine in conjunction
 with image annotation? I don't want end users to annotate things
 themselves, but allow them to click on annotations added by an archivist.