[CODE4LIB] RHEV-M

2013-11-19 Thread Lynne E. Grigsby
Is anyone using RHEV-M to manager their Linux virtual server environment?
 We are interested in talking to someone about their experience.  We have
found plenty of people to talk to about VMware, but would like to
investigate an alternative.

Thanks for your help.
Lynne


[CODE4LIB] ruby-marc api design feedback wanted

2013-11-19 Thread Jonathan Rochkind

ruby-marc users, a question.

I am working on some Marc8 to UTF-8 conversion for ruby-marc.

Sometimes, what appears to be an illegal byte will appear in the Marc8 
input, and it can not be converted to UTF8.


The software will support two alternatives when this happens: 1) Raising 
an exception. 2) Replacing the illegal byte with a replacement char 
and/or omitting it.


I feel like most of the time, users are going to want #2.  I know that's 
what I'm going to want nearly all the time.


Yet, still, I am feeling uncertain whether that should be the default. 
Which should be the default behavior, #1 or #2?  If most people most of 
the time are going to want #2 (is this true?), then should that be the 
default behavior?   Or should #1 still be the default behavior, because 
by default bad input should raise, not be silently recovered from, even 
though most people most of the time won't want that, heh.


Jonathan


[CODE4LIB] Programmatic retrieval of book series info?

2013-11-19 Thread Cab Vinton
Our circ staff would like to retrospectively add labels to our fiction
series so patrons can suss out series information at a glance (which
series, numbering).

I can pull out the titles, authors, ISBNs, etc. from our catalog
easily enough, but I'm wondering how best to approach matching each
title with the appropriate series info.

GoodReads has an API, for example, but wondering if anyone on the list
has already come up with a good method or has suggestions.

Many thanks in advance,

Cab Vinton, Director
Plaistow Public Library
Plaistow, NH


[CODE4LIB] Job: Associate Developer (Ruby on Rails) at WGBH Educational Foundation

2013-11-19 Thread jobs
Associate Developer (Ruby on Rails)
WGBH Educational Foundation
Boston

**Position Title:** Associate Developer (Ruby on Rails)  
  
**Position Type:** Project Contract 12/02/13 to 12/31/14+  
  
**Company:** WGBH Educational Foundation  
  
**Department:** Media Library & Archives  
  
**Department Overview:**  
WGBH produces the best and most well-known television, radio and online
programs for public media. The WGBH Media Library and Archives preserves and
helps re-purpose WGBH creations into the future. The MLA establishes the
policies and procedures for the access, acquisition, intellectual control, and
preservation of WGBH's physical media and digital production and
administrative assets. The MLA also offers production organization of archival
materials from projects start up to shut down, research services, rights
clearances, and licenses WGBH stock footage.

  
**Position Overview:**  
The WGBH Media Library and Archives system will be based on the Hydra Project

technology stack, which includes Ruby on Rails, Blacklight, Apache Solr, and
the

Fedora Commons repository. Working closely with the Media Library and
Archive's

Director, Project Manager, Developer and Systems Analyst, as well as a WGBH

Interactive Designer, the web developer will continue to develop the Open
Vault

website: http://openvault.wgbh.org and ongoing work to improve the digital
asset

management system.

  
**Ideal candidates should be:**  
• comfortable working in teams of 2 to 6

• able to communicate clearly and respectfully to all team members, both
technical and non-technical

• willing to explore new technologies

  
**Duties** will depend on individual strengths, but may include any of:  
• general Rails development

• streaming video integration and presentation

• organizing and writing documentation

• usage stats and analytics

• DevOps and deployment

• performance stats, analysis and optimization

  
**Required skills** for all tasks include having working knowledge of:  
• Ruby >= 1.9.3

• Rails >= 3.2.0, common conventions, patterns, and best practices, TDD with
Rspec and Capybara (or equivalent)

• Github

• CSS3 + HTML5

• XML basics

• working from command line (OS X or Linux)

  
**Other skills** that will come in handy for other project tasks include having 
a experience with:  
• SCSS

• jQuery

• Twitter Bootstrap

• building REST apis

• Rails gem patterns

• HTML 5 video and various javascript players

• Rails deployment w/ Capistrano

• SQL basics

• Fedora Commons repository

• working with XML, e.g. translating using XSL, parsing with Nokogiri,

Education Requirements:

Bachelor's Degree in Computer Science required.

  
**Compensation:**  
Compensation for this position will be determined by the skills, background,
education and availability of the candidate for the Contract period.

  
**Applying for the Position:**  
Candidates should apply at www.wgbh.org/careers. Reference Job REQ# P-1075

  
**Further questions** can be addressed to Dani Baptista 
(dani_bapti...@wgbh.org). Please put "REQ# P-1075 Associate Developer" in the 
subject line. 



Brought to you by code4lib jobs: http://jobs.code4lib.org/job/10762/


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Bigwood, David
+1 for schema.org as one of the first steps. COinS are another useful simple 
mark-up if the data is already there.

I'm looking forward to the book.

Sincerely,
David Bigwood
Lunar and Planetary Institute


-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Karen 
Coyle
Sent: Tuesday, November 19, 2013 10:10 AM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] linked data recipe

Eric, if you want to leap into the linked data world in the fastest, easiest 
way possible, then I suggest looking at microdata markup, e.g. 
schema.org.[1] Schema.org does not require you to transform your data at
all: it only requires mark-up of your online displays. This makes sense because 
as long as your data is in local databases, it's not visible to the linked data 
universe anyway; so why not take the easy way out and just add linked data to 
your public online displays? This doesn't require a transformation of your 
entire record (some of which may not be suitable as linked data in any case), 
only those "things" that are likely to link usefully. This latter generally 
means "things for which you have an identifier." And you make no changes to 
your database, only to display.

OCLC is already producing this markup in WorldCat records [2]-- not perfectly, 
of course, lots of warts, but it is a first step. However, it is a first step 
that makes more sense to me than *transforming* or
*cross-walking* current metadata. It also, I believe, will help us understand 
what bits of our current metadata will make the transition to linked data, and 
what bits should remain as accessible documents that users can reach through 
linked data.

kc
[1] http://schema.org, and look at the work going on to add bibliographic 
properties at http://www.w3.org/community/schemabibex/wiki/Main_Page
[2] look at the "linked data" section of any WorldCat page for a single item, 
such 
ashttp://www.worldcat.org/title/selection-of-early-statistical-papers-of-j-neyman/oclc/527725&referer=brief_results


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Karen Coyle
Ethan, it looks to me like it depends on who you are and who is your 
target. In the schema.org clan there is still a majority using 
microdata, but my impression is that these are the online sales sites 
whose primary interest is SEO. RDFa lite is moving up generally [0], yet 
I haven't seen a clear statement that the search engines consider it = 
microdata (even though the two are very close). Perhaps they do? 
Recently it was announced that JSON-LD is now an "official" schema.org 
markup. The advantage of JSON-LD is that it separates the display from 
the mark-up so there is less of a formatting issue. However, it also 
opens it all up to scamming - well, to easier scamming than with the 
other two formats.


Meanwhile, as more and more folks discover schema.org there is more and 
more demand for additions to what was originally an extremely simple set 
of properties. Some predict that it will crumble under its own 
disorderliness, a metadata tower of Babel.


Regardless of that, I still think that the web is the place for linked 
data, even though there are quite a few enterprise implementations of ld 
that do not present a public face. I'd prefer to have some idea of what 
we want to link to, why, and how it will help users. There are some 
examples, like FAO's Open Agris [1], but I'd like to see more. (And I'm 
not sure what LIBRIS [2] is doing with their catalog, which is reported 
to be a triple-store.)


kc
[0] http://webdatacommons.org/
[1] http://agris.fao.org/openagris/
[2] http://libris.kb.se/?language=en
On 11/19/13 8:28 AM, Ethan Gruber wrote:

Hasn't the pendulum swung back toward RDFa Lite (
http://www.w3.org/TR/rdfa-lite/) recently?  They are fairly equivalent, but
I'm not sure about all the politics involved.


On Tue, Nov 19, 2013 at 11:09 AM, Karen Coyle  wrote:


Eric, if you want to leap into the linked data world in the fastest,
easiest way possible, then I suggest looking at microdata markup, e.g.
schema.org.[1] Schema.org does not require you to transform your data at
all: it only requires mark-up of your online displays. This makes sense
because as long as your data is in local databases, it's not visible to the
linked data universe anyway; so why not take the easy way out and just add
linked data to your public online displays? This doesn't require a
transformation of your entire record (some of which may not be suitable as
linked data in any case), only those "things" that are likely to link
usefully. This latter generally means "things for which you have an
identifier." And you make no changes to your database, only to display.

OCLC is already producing this markup in WorldCat records [2]-- not
perfectly, of course, lots of warts, but it is a first step. However, it is
a first step that makes more sense to me than *transforming* or
*cross-walking* current metadata. It also, I believe, will help us
understand what bits of our current metadata will make the transition to
linked data, and what bits should remain as accessible documents that users
can reach through linked data.

kc
[1] http://schema.org, and look at the work going on to add bibliographic
properties at http://www.w3.org/community/schemabibex/wiki/Main_Page
[2] look at the "linked data" section of any WorldCat page for a single
item, such ashttp://www.worldcat.org/title/selection-of-early-
statistical-papers-of-j-neyman/oclc/527725&referer=brief_results




On 11/19/13 7:54 AM, Eric Lease Morgan wrote:


On Nov 19, 2013, at 9:41 AM, Karen Coyle  wrote:

  Eric, I think this skips a step - which is the design step in which you

create a domain model that uses linked data as its basis. RDF is not a
serialization; it actually may require you to re-think the basic
structure of your metadata. The reason for that is that it provides
capabilities that record-based data models do not. Rather than starting
with current metadata, you need to take a step back and ask: what does
my information world look like as linked data?


I respectfully disagree. I do not think it necessary to create a domain
model ahead of time; I do not think it is necessary for us to re-think our
metadata structures. There already exists tools enabling us — cultural
heritage institutions — to manifest our metadata as RDF. The manifestations
may not be perfect, but “we need to learn to walk before we run” and the
metadata structures we have right now will work for right now. As we mature
we can refine our processes. I do not advocate “stepping back and asking”.
I advocate looking forward and doing. —Eric Morgan


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet



--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Edward Summers
I think this is a nice list Eric. I particularly like the iterative approach. 
I’m not a huge fan of #6, and #7 seems like it might be challenging from a data 
synchronization perspective.  But it’s still a nice list.

While I think it’s right that you don’t want to let the perfect (a complete and 
perfect domain model) be the enemy of the good (iterative data publishing on 
the Web), it definitely helps if a Linked Data project has an idea of what 
types of resources it is putting on the Web, and how they potentially fit in 
with other stuff that’s already there.

I honestly think the hardest thing is to establish *why* you want to publish 
data on the Web: who is it for, how will they use it, etc. If the honest answer 
is simply “it is the right thing to do”, “we want to get a grant” or “we want 
to build the semweb” that’s fine, but it’s not ideal. Ideally there’s an actual 
use case where exposing structured data on the Web yields potential benefits, 
that can be realized with Linked Data. 

//Ed

On Nov 19, 2013, at 11:55 AM, Eric Lease Morgan  wrote:

> On Nov 19, 2013, at 11:09 AM, Karen Coyle  wrote:
> 
>> Eric, if you want to leap into the linked data world in the fastest, 
>> easiest way possible, then I suggest looking at microdata markup, e.g. 
>> schema.org. [1] …
>> 
>> [1] http://schema.org
> 
> 
> I don’t advocate this as the fastest, easiest way possible because it forces 
> RDF “aggregators” to parse HTML, and thus passes a level of complexity down 
> the processing chain. Expose RDF as RDF, not embedded in another format. I do 
> advocate the inclusion of schema.org mark-up, RDFa, etc. into HTML but rather 
> as a level of refinement. —Eric Morgan


[CODE4LIB] Job: Metadata Integration and Delivery Specialist at University of Virginia Library

2013-11-19 Thread jobs
Metadata Integration and Delivery Specialist
University of Virginia Library
Charlottesville, VA

The University of Virginia Library seeks a Metadata Integration and Delivery
Specialist for Metadata Management Services. The University
of Virginia Library is place of infinite possibility! We
are looking for high energy, innovative professionals with integrity and a
strong work ethic. We seek individuals who are not afraid of taking risk and
have the proven ability and/or potential to get positive results, manage
change, and collaborate with others effectively. We want creative
professionals who possess a keen and deep understanding of what it takes to
continuously improve and maintain a major academic research library.

  
Metadata Management Services facilitates access to library managed content,
participates in partnerships to share expertise and provide innovation in data
management, and engages fully in the mission of the University of
Virginia. Reporting to the Metadata Librarian, the Metadata
Integration and Delivery Specialist supports that mission by collaborating
with colleagues within and outside of the department and Library to determine
and document best practices and workflows for non-MARC metadata creation. This
position is charged with finding creative, collaborative and sustainable
solutions for providing and managing metadata for a variety of physical and
digital resources. This position will also have responsibility for describing
resources in a variety of formats, following local and national standards. The
employee in this position is encouraged to be current with the community of
practice for non-MARC metadata and stay abreast of developments within the
broader field of information organization.

  
Qualifications

  
Required: Bachelor's degree. Experience with library
metadata standards and application, and demonstrated ability to work
independently and collaboratively across groups to achieve
objectives. Demonstrated ability to create metadata
according to established rules and standards. Ability to
communicate effectively orally and in writing. Excellent
interpersonal skills.

  
Preferred: Master of Science in Library Science or related
field.

  
To Apply: Complete a Candidate Profile, attach a cover letter, cv, and contact
information for three professional references through Jobs@UVA (Posting
#0613248). For assistance with this process contact
Charlotte Albright (cd...@virginia.edu), Library Human Resources Generalist at
(434) 243-3509. Review of applications will begin December
6, 2013.

  
The University of Virginia is an affirmative action/equal opportunity employer
committed to diversity, equity, and inclusiveness.



Brought to you by code4lib jobs: http://jobs.code4lib.org/job/10761/


[CODE4LIB] Job: Business Analyst at Harvard University

2013-11-19 Thread jobs
Business Analyst
Harvard University
Cambridge, MA

The Business Analyst will support leadership of the Harvard
Library by identifying operational data needs, mining information systems, and
providing analysis to inform business performance assessment and
decisions. Reporting to the Director, Financial Planning
and Analysis, the Business Analyst will:

  

  * Maintain and develop the library's information and organizational 
performance systems
  * Provide reporting and supporting assessment advice and analysis to inform 
strategic decisions
  * Enhance the library's information and assessment systems and ability to 
collect and report useful data
  * Support librarians in finding and using data in their daily work
  
This role will collaborate with staff across a wide range of functions and
disciplines in order to better coordinate the collection and creation of data
as it relates to the performance goals of the Harvard Library.

  
  
**Duties and Responsibilities:**  
  
The Business Analyst works closely with the Harvard Library leadership and
with librarians across the broader library system. Specific
components of the role include:

  
_Maintain and develop the library's information and organizational performance
systems_

  * In consultation with Harvard Library leadership, design, develop and 
maintain systems to measure library progress as it relates to library strategic 
initiatives and operational efficiency
  * Improve methods used to measure key metrics
  * Develop and monitor annual baselines for key metrics and monitor progress
  * Lead the preparation of organizational performance updates and associated 
reporting
  * Communicate and explain findings to library senior leadership
  
_Provide reporting and supporting assessment advice and analysis to inform
strategic decisions_

  * Work collaboratively with Harvard Library leadership to discover 
information needs and to identify, prioritize, design and implement new 
reporting tools and solutions
  * Analyze information (e.g., bibliographic, financial, usage) in support of 
evolving needs
  * Provide consultation and technical assistance to Harvard Library leadership 
with use and interpretation of data
  * Respond to requests for data and/or reports on a variety of issues related 
to the library collections reporting and statistics, including ad hoc analyses 
and special requests
  * Lead in data collection and aggregation across partnerships with other 
institutions (including Borrow Direct Partners) to aggregate data and business 
intelligence to help improve and expand alliances
  * Lead the data collection efforts for national surveys (ARL statistics) and 
the preparation of data for annual HL reports
  
_Enhance the Library's assessment systems and ability to collect and report
useful data_

  * Develop structures for collecting, storing and presenting or sharing data 
in consultation with appropriate IT and Library colleagues
  * Design, develop, troubleshoot, document, and analyze reports for Harvard 
Library
  * Maintain internal sites to house data, distribute reports, associated 
documentation and explanatory materials
  * Provide consultation to Harvard Library leadership on developing assessment 
surveys and user studies
  
_Support librarians in finding and using data in their daily work_

  * Facilitate data-driven decision making throughout the Harvard Library by 
committing to make data as accessible and transparent as possible; 
teaching/mentoring librarians how to access, query, manipulate, and visualize 
data
  * Develop reporting mechanisms to enable managers and staff to run and 
analyze data in support of data driven decision making
**Basic Qualifications**

  * Bachelor's degree or equivalent education or work experience required
  * Seven plus years of experience working in IT, business intelligence, 
library assessment, financial or statistical analysis, financial reporting 
and/or libraries
  * Strong knowledge of Excel and of database tools (e.g., Microsoft Access)
**Additional Qualifications**  

  * MLS, MBA, and/or advanced degree in statistics, information sciences, or 
engineering
  * Experience building reports and queries with Business 
Intelligence/Analytics tools (Cognos BI, Microsoft)
  * Working knowledge of libraries, both within Harvard and more general 
industry trends
  * Knowledge of Integrated Library Systems, such as Aleph
  * Knowledge of statistical software packages (e.g. SPSS, R, STATA)
  * Knowledge of database tools such as Quickbase and content management systems
  * Proven ability to deliver quality analysis in a fast-paced environment
  * Excellent analytic, written and verbal communication skills
  * Ability to process and distill complex information, present complicated 
information in easily comprehensible formats
  * Ability to work in ambiguous environment: to define objectives, set tasks, 
build relationships, and achieve outcomes
  * Exceptional numeric skills: able to use number

Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Eric Lease Morgan
On Nov 19, 2013, at 11:09 AM, Karen Coyle  wrote:

> Eric, if you want to leap into the linked data world in the fastest, 
> easiest way possible, then I suggest looking at microdata markup, e.g. 
> schema.org. [1] …
> 
> [1] http://schema.org


I don’t advocate this as the fastest, easiest way possible because it forces 
RDF “aggregators” to parse HTML, and thus passes a level of complexity down the 
processing chain. Expose RDF as RDF, not embedded in another format. I do 
advocate the inclusion of schema.org mark-up, RDFa, etc. into HTML but rather 
as a level of refinement. —Eric Morgan


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Eric Lease Morgan
On Nov 19, 2013, at 9:54 AM, Aaron Rubinstein  
wrote:

> I think you’ve hit the nail on the head here, Karen. I would just
> add, or maybe reassure, that this does not necessarily require
> rethinking your existing metadata but how to translate that
>   ^
> existing metadata into a linked data environment. Though this
> 
> might seem like a pain, in many cases it will actually inspire
> you to go back and improve/increase the value of that existing
> metadata...


There are tools allowing people to translate existing metadata into a linked 
data environment, and for right now, I advocate that they are good enough. I 
will provide simplistic examples.

For people who maintain MARC records:

  1. convert the MARC records to MARCXML with the MARCXML Toolkit [1]
  2. convert the MARCXML to RDF/XML in the manner of BIBFRAME’s transformation 
service [2]
  3. save the resulting RDF/XML on a Web server
  4. convert the MARC (or MARCXML) into (valid) HTML
  5. save the resulting HTML on a Web server
  6. for extra credit, implement a content negotiation service for the HTML and 
RDF/XML
  7. for extra extra credit, implement a SPARQL endpoint for your RDF

If one does Steps #1 through #5, then they are doing linked data and 
participating in the Semantic Web. That is the goal.

For people who maintain EAD files:

  1. transform the EAD files into RDF/XML with a stylesheet created by the 
Archives Hub [3]
  2. save the resulting RDF/XML on a Web server
  3. transform the EAD into HTML, using your favorite EAD to HTML stylesheet [4]
  4. save the resulting HTML on a Web server
  5. for extra credit, implement a content negotiation service for the HTML and 
RDF/XML
  6. for extra extra credit, implement a SPARQL endpoint for your RDF

If one does Steps #1 through #4 of this example, then they are doing linked 
data and participating in the Semantic Web. That is the goal.

In both examples the end result will be a valid linked data implementation. Not 
complete. Not necessarily as thorough as desired. Not necessarily as accurate 
as desired. But valid. Such a process will not expose false, incorrect 
data/information, but rather data/information that is intended to be 
maintained, improved, and updated on a continual basis.

Finally, I want to highlight a distinction between well-formed, valid, and 
accurate information — linked data. I will use XML as an example. XML can be 
“well-formed”. This means it is syntactically correct. Specific characters are 
represented by entities. Elements are correctly opened and closed. The whole 
structure has a single root. Etc. The next level up is “valid”. Valid XML is 
XML that conforms to a DTD or schema; it is semantically correct. It means that 
required elements exist, and are presented in a particular order. Specific 
attributes used in elements are denoted. And in the case of schemas, values in 
elements and attributes take on particular shapes beyond simple character data. 
Finally XML can be “accurate” (my term). This means the assertions in the XML 
are true. For example, there is nothing stopping me from putting the title of a 
work in an author element. How is the computer expected to know the difference? 
It can’t. Alternatively, the title could be presente!
 d as “Thee Adventrs Av Tom Sawher”, when the more accurate title may be “The 
Adventures of Tom Sawyer”. Well-formedness and validity is the domain of 
computers. Accuracy is the domain of humans. In the world of linked data, you 
are not participating if your published data is not “well-formed”. (Go back to 
start.) You are participating if it is “valid”. But you are really doing really 
well if the data is “accurate”. 

Let’s not make this more difficult than it really is.

[1] MARCXML Toolkit - linked at http://www.loc.gov/standards/marcxml/
[2] BIBFRAME’s transformation service - 
http://bibframe.org/tools/transform/start
[3] Archives Hub stylesheet - http://data.archiveshub.ac.uk/xslt/ead2rdf.xsl
[4] EAD to HTML - for example, 
http://www.catholicresearch.net/data/ead/ead2html.xsl

— 
Eric Morgan


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Ethan Gruber
Hasn't the pendulum swung back toward RDFa Lite (
http://www.w3.org/TR/rdfa-lite/) recently?  They are fairly equivalent, but
I'm not sure about all the politics involved.


On Tue, Nov 19, 2013 at 11:09 AM, Karen Coyle  wrote:

> Eric, if you want to leap into the linked data world in the fastest,
> easiest way possible, then I suggest looking at microdata markup, e.g.
> schema.org.[1] Schema.org does not require you to transform your data at
> all: it only requires mark-up of your online displays. This makes sense
> because as long as your data is in local databases, it's not visible to the
> linked data universe anyway; so why not take the easy way out and just add
> linked data to your public online displays? This doesn't require a
> transformation of your entire record (some of which may not be suitable as
> linked data in any case), only those "things" that are likely to link
> usefully. This latter generally means "things for which you have an
> identifier." And you make no changes to your database, only to display.
>
> OCLC is already producing this markup in WorldCat records [2]-- not
> perfectly, of course, lots of warts, but it is a first step. However, it is
> a first step that makes more sense to me than *transforming* or
> *cross-walking* current metadata. It also, I believe, will help us
> understand what bits of our current metadata will make the transition to
> linked data, and what bits should remain as accessible documents that users
> can reach through linked data.
>
> kc
> [1] http://schema.org, and look at the work going on to add bibliographic
> properties at http://www.w3.org/community/schemabibex/wiki/Main_Page
> [2] look at the "linked data" section of any WorldCat page for a single
> item, such ashttp://www.worldcat.org/title/selection-of-early-
> statistical-papers-of-j-neyman/oclc/527725&referer=brief_results
>
>
>
>
> On 11/19/13 7:54 AM, Eric Lease Morgan wrote:
>
>> On Nov 19, 2013, at 9:41 AM, Karen Coyle  wrote:
>>
>>  Eric, I think this skips a step - which is the design step in which you
>>> create a domain model that uses linked data as its basis. RDF is not a
>>> serialization; it actually may require you to re-think the basic
>>> structure of your metadata. The reason for that is that it provides
>>> capabilities that record-based data models do not. Rather than starting
>>> with current metadata, you need to take a step back and ask: what does
>>> my information world look like as linked data?
>>>
>>
>> I respectfully disagree. I do not think it necessary to create a domain
>> model ahead of time; I do not think it is necessary for us to re-think our
>> metadata structures. There already exists tools enabling us — cultural
>> heritage institutions — to manifest our metadata as RDF. The manifestations
>> may not be perfect, but “we need to learn to walk before we run” and the
>> metadata structures we have right now will work for right now. As we mature
>> we can refine our processes. I do not advocate “stepping back and asking”.
>> I advocate looking forward and doing. —Eric Morgan
>>
>
> --
> Karen Coyle
> kco...@kcoyle.net http://kcoyle.net
> m: 1-510-435-8234
> skype: kcoylenet
>


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Karen Coyle
Eric, if you want to leap into the linked data world in the fastest, 
easiest way possible, then I suggest looking at microdata markup, e.g. 
schema.org.[1] Schema.org does not require you to transform your data at 
all: it only requires mark-up of your online displays. This makes sense 
because as long as your data is in local databases, it's not visible to 
the linked data universe anyway; so why not take the easy way out and 
just add linked data to your public online displays? This doesn't 
require a transformation of your entire record (some of which may not be 
suitable as linked data in any case), only those "things" that are 
likely to link usefully. This latter generally means "things for which 
you have an identifier." And you make no changes to your database, only 
to display.


OCLC is already producing this markup in WorldCat records [2]-- not 
perfectly, of course, lots of warts, but it is a first step. However, it 
is a first step that makes more sense to me than *transforming* or 
*cross-walking* current metadata. It also, I believe, will help us 
understand what bits of our current metadata will make the transition to 
linked data, and what bits should remain as accessible documents that 
users can reach through linked data.


kc
[1] http://schema.org, and look at the work going on to add 
bibliographic properties at 
http://www.w3.org/community/schemabibex/wiki/Main_Page
[2] look at the "linked data" section of any WorldCat page for a single 
item, such 
ashttp://www.worldcat.org/title/selection-of-early-statistical-papers-of-j-neyman/oclc/527725&referer=brief_results




On 11/19/13 7:54 AM, Eric Lease Morgan wrote:

On Nov 19, 2013, at 9:41 AM, Karen Coyle  wrote:


Eric, I think this skips a step - which is the design step in which you
create a domain model that uses linked data as its basis. RDF is not a
serialization; it actually may require you to re-think the basic
structure of your metadata. The reason for that is that it provides
capabilities that record-based data models do not. Rather than starting
with current metadata, you need to take a step back and ask: what does
my information world look like as linked data?


I respectfully disagree. I do not think it necessary to create a domain model 
ahead of time; I do not think it is necessary for us to re-think our metadata 
structures. There already exists tools enabling us — cultural heritage 
institutions — to manifest our metadata as RDF. The manifestations may not be 
perfect, but “we need to learn to walk before we run” and the metadata 
structures we have right now will work for right now. As we mature we can 
refine our processes. I do not advocate “stepping back and asking”. I advocate 
looking forward and doing. —Eric Morgan


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Ethan Gruber
yo, i get it


On Tue, Nov 19, 2013 at 10:54 AM, Ross Singer  wrote:

> I don't know what your definition of "serialization" is, but I don't know
> of any where "data model" and "formatted output of a data model" are
> synonymous.
>
> RDF is a data model *not* a serialization.
>
> -Ross.
>
>
> On Tue, Nov 19, 2013 at 10:45 AM, Ethan Gruber  wrote:
>
> > I see that serialization has a different definition in computer science
> > than I thought it did.
> >
> >
> > On Tue, Nov 19, 2013 at 10:36 AM, Ross Singer 
> > wrote:
> >
> > > That's still not a "serialization".  It's just a similar data model.
> > >  Pretty huge difference.
> > >
> > > -Ross.
> > >
> > >
> > > On Tue, Nov 19, 2013 at 10:31 AM, Ethan Gruber 
> > wrote:
> > >
> > > > I'm not sure that I agree that RDF is not a serialization.  It really
> > > > depends on the context of the system and intended use of the linked
> > data.
> > > > For example, TEI is designed with a specific purpose which cannot be
> > > > replicated in RDF (at least, not very easily at all), but deriving
> RDF
> > > from
> > > > highly-linked TEI to put into an endpoint can open doors to queries
> > which
> > > > are otherwise impossible to make on the data.  This certainly
> requires
> > > some
> > > > rethinking of the way texts interact.  But perhaps it may be best to
> > say
> > > > that RDF *can* (but not necessarily) be a derivation, rather than
> > > > serialization, of some larger, more complex canonical data model.
> > > >
> > > > Ethan
> > > >
> > > >
> > > > On Tue, Nov 19, 2013 at 9:54 AM, Aaron Rubinstein <
> > > > arubi...@library.umass.edu> wrote:
> > > >
> > > > > I think you’ve hit the nail on the head here, Karen. I would just
> > add,
> > > or
> > > > > maybe reassure, that this does not necessarily require rethinking
> > your
> > > > > existing metadata but how to translate that existing metadata into
> a
> > > > linked
> > > > > data environment. Though this might seem like a pain, in many cases
> > it
> > > > will
> > > > > actually inspire you to go back and improve/increase the value of
> > that
> > > > > existing metadata.
> > > > >
> > > > > This is definitely looking awesome, Eric!
> > > > >
> > > > > Aaron
> > > > >
> > > > > On Nov 19, 2013, at 9:41 AM, Karen Coyle  wrote:
> > > > >
> > > > > > Eric, I think this skips a step - which is the design step in
> which
> > > you
> > > > > create a domain model that uses linked data as its basis. RDF is
> not
> > a
> > > > > serialization; it actually may require you to re-think the basic
> > > > structure
> > > > > of your metadata. The reason for that is that it provides
> > capabilities
> > > > that
> > > > > record-based data models do not. Rather than starting with current
> > > > > metadata, you need to take a step back and ask: what does my
> > > information
> > > > > world look like as linked data?
> > > > > >
> > > > > > I repeat: RDF is NOT A SERIALIZATION.
> > > > > >
> > > > > > kc
> > > > > >
> > > > > > On 11/19/13 5:04 AM, Eric Lease Morgan wrote:
> > > > > >> I believe participating in the Semantic Web and providing
> content
> > > via
> > > > > the principles of linked data is not "rocket surgery", especially
> for
> > > > > cultural heritage institutions -- libraries, archives, and museums.
> > > Here
> > > > is
> > > > > a simple recipe for their participation:
> > > > > >>
> > > > > >>   1. use existing metadata standards (MARC, EAD, etc.) to
> describe
> > > > > >>  collections
> > > > > >>
> > > > > >>   2. use any number of existing tools to convert the metadata to
> > > > > >>  HTML, and save the HTML on a Web server
> > > > > >>
> > > > > >>   3. use any number of existing tools to convert the metadata to
> > > > > >>  RDF/XML (or some other "serialization" of RDF), and save
> the
> > > > > >>  RDF/XML on a Web server
> > > > > >>
> > > > > >>   4. rest, congratulate yourself, and share your experience with
> > > > > >>  others in your domain
> > > > > >>
> > > > > >>   5. after the first time though, go back to Step #1, but this
> > time
> > > > > >>  work with other people inside your domain making sure you
> use
> > > as
> > > > > >>  many of the same URIs as possible
> > > > > >>
> > > > > >>   6. after the second time through, go back to Step #1, but this
> > > > > >>  time supplement access to your linked data with a triple
> > store,
> > > > > >>  thus supporting search
> > > > > >>
> > > > > >>   7. after the third time through, go back to Step #1, but this
> > > > > >>  time use any number of existing tools to expose the content
> > in
> > > > > >>  your other information systems (relational databases,
> OAI-PMH
> > > > > >>  data repositories, etc.)
> > > > > >>
> > > > > >>   8. for dessert, cogitate ways to exploit the linked data in
> your
> > > > > >>  domain to discover new and additional relationships between
> > > URIs,
> > > > > >>  and thus make the Semantic Web more of a reality
> > > > > >>
> > > > > >

Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Ross Singer
I don't know what your definition of "serialization" is, but I don't know
of any where "data model" and "formatted output of a data model" are
synonymous.

RDF is a data model *not* a serialization.

-Ross.


On Tue, Nov 19, 2013 at 10:45 AM, Ethan Gruber  wrote:

> I see that serialization has a different definition in computer science
> than I thought it did.
>
>
> On Tue, Nov 19, 2013 at 10:36 AM, Ross Singer 
> wrote:
>
> > That's still not a "serialization".  It's just a similar data model.
> >  Pretty huge difference.
> >
> > -Ross.
> >
> >
> > On Tue, Nov 19, 2013 at 10:31 AM, Ethan Gruber 
> wrote:
> >
> > > I'm not sure that I agree that RDF is not a serialization.  It really
> > > depends on the context of the system and intended use of the linked
> data.
> > > For example, TEI is designed with a specific purpose which cannot be
> > > replicated in RDF (at least, not very easily at all), but deriving RDF
> > from
> > > highly-linked TEI to put into an endpoint can open doors to queries
> which
> > > are otherwise impossible to make on the data.  This certainly requires
> > some
> > > rethinking of the way texts interact.  But perhaps it may be best to
> say
> > > that RDF *can* (but not necessarily) be a derivation, rather than
> > > serialization, of some larger, more complex canonical data model.
> > >
> > > Ethan
> > >
> > >
> > > On Tue, Nov 19, 2013 at 9:54 AM, Aaron Rubinstein <
> > > arubi...@library.umass.edu> wrote:
> > >
> > > > I think you’ve hit the nail on the head here, Karen. I would just
> add,
> > or
> > > > maybe reassure, that this does not necessarily require rethinking
> your
> > > > existing metadata but how to translate that existing metadata into a
> > > linked
> > > > data environment. Though this might seem like a pain, in many cases
> it
> > > will
> > > > actually inspire you to go back and improve/increase the value of
> that
> > > > existing metadata.
> > > >
> > > > This is definitely looking awesome, Eric!
> > > >
> > > > Aaron
> > > >
> > > > On Nov 19, 2013, at 9:41 AM, Karen Coyle  wrote:
> > > >
> > > > > Eric, I think this skips a step - which is the design step in which
> > you
> > > > create a domain model that uses linked data as its basis. RDF is not
> a
> > > > serialization; it actually may require you to re-think the basic
> > > structure
> > > > of your metadata. The reason for that is that it provides
> capabilities
> > > that
> > > > record-based data models do not. Rather than starting with current
> > > > metadata, you need to take a step back and ask: what does my
> > information
> > > > world look like as linked data?
> > > > >
> > > > > I repeat: RDF is NOT A SERIALIZATION.
> > > > >
> > > > > kc
> > > > >
> > > > > On 11/19/13 5:04 AM, Eric Lease Morgan wrote:
> > > > >> I believe participating in the Semantic Web and providing content
> > via
> > > > the principles of linked data is not "rocket surgery", especially for
> > > > cultural heritage institutions -- libraries, archives, and museums.
> > Here
> > > is
> > > > a simple recipe for their participation:
> > > > >>
> > > > >>   1. use existing metadata standards (MARC, EAD, etc.) to describe
> > > > >>  collections
> > > > >>
> > > > >>   2. use any number of existing tools to convert the metadata to
> > > > >>  HTML, and save the HTML on a Web server
> > > > >>
> > > > >>   3. use any number of existing tools to convert the metadata to
> > > > >>  RDF/XML (or some other "serialization" of RDF), and save the
> > > > >>  RDF/XML on a Web server
> > > > >>
> > > > >>   4. rest, congratulate yourself, and share your experience with
> > > > >>  others in your domain
> > > > >>
> > > > >>   5. after the first time though, go back to Step #1, but this
> time
> > > > >>  work with other people inside your domain making sure you use
> > as
> > > > >>  many of the same URIs as possible
> > > > >>
> > > > >>   6. after the second time through, go back to Step #1, but this
> > > > >>  time supplement access to your linked data with a triple
> store,
> > > > >>  thus supporting search
> > > > >>
> > > > >>   7. after the third time through, go back to Step #1, but this
> > > > >>  time use any number of existing tools to expose the content
> in
> > > > >>  your other information systems (relational databases, OAI-PMH
> > > > >>  data repositories, etc.)
> > > > >>
> > > > >>   8. for dessert, cogitate ways to exploit the linked data in your
> > > > >>  domain to discover new and additional relationships between
> > URIs,
> > > > >>  and thus make the Semantic Web more of a reality
> > > > >>
> > > > >> What do you think?
> > > > >>
> > > > >> I am in the process of writing a guidebook on the topic of linked
> > data
> > > > and archives. In the guidebook I will elaborate on this recipe and
> > > provide
> > > > instructions for its implementation. [1]
> > > > >>
> > > > >> [1] guidebook - http://sites.tufts.edu/liam/
> > > > >>
> > > > >> --
> >

Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Eric Lease Morgan
On Nov 19, 2013, at 9:41 AM, Karen Coyle  wrote:

> Eric, I think this skips a step - which is the design step in which you 
> create a domain model that uses linked data as its basis. RDF is not a 
> serialization; it actually may require you to re-think the basic 
> structure of your metadata. The reason for that is that it provides 
> capabilities that record-based data models do not. Rather than starting 
> with current metadata, you need to take a step back and ask: what does 
> my information world look like as linked data?


I respectfully disagree. I do not think it necessary to create a domain model 
ahead of time; I do not think it is necessary for us to re-think our metadata 
structures. There already exists tools enabling us — cultural heritage 
institutions — to manifest our metadata as RDF. The manifestations may not be 
perfect, but “we need to learn to walk before we run” and the metadata 
structures we have right now will work for right now. As we mature we can 
refine our processes. I do not advocate “stepping back and asking”. I advocate 
looking forward and doing. —Eric Morgan


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Eric Lease Morgan
On Nov 19, 2013, at 8:48 AM, Robert Forkel  wrote:

> while I also think this is not rocket surgery, I'd like to point out that
> trial (and potentially error) as suggested by your "go back to step #1"
> instructions is not a good solution to coming up with URIs. I think once
> published - i.e. put on a webserver - you should be able to keep the URIs
> in your RDF persistent. Otherwise you are polluting the Semantic Web with
> dead links and make it hard for aggregators to find out whether the data
> they harvested is still valid.
> 
> So while iterative approaches are pragmatic and often work out well, for
> the particular issue of coming up with URIs I'd recommend spending as much
> thought before publishing as you can spend.


Intellectually, I completely understand.

Practically, I still advocate putting publishing the linked data as soon as 
possible. Knowledge is refined over time. The data being published is not 
incorrect nor invalid, just not as good as it could be. Data aggregators will 
refresh their stores and old information will go to "Big Byte Heaven”. It is 
just like a library collection. The “best” books are collected. The good ones 
get used. The old ones get weeded or relegated to off-site storage. What 
remains is a current perception of truth. Building library collections is a 
process that is never done nor never perfect. Linked data is a literal 
reflection of library collections, therefore linked data is never done nor 
never perfect either. URIs will break. Books will be removed from the 
collection. URIs will go stale. 

The process of providing linked data is a lot like painting a painting. The 
painting is painted as a whole, from start to finish. One does not get one 
corner of the canvass perfect and move on from there. An idea is articulated. 
An outlined is drawn. The outline is refined, and the painting gradually comes 
to life. Many times paintings are never finished but worked, reworked, and 
worked some more. 

If the profession looks to make perfect its list of URIs, then it will never 
leave the starting gate. I know that is not being advocated, but since one can 
not measure the timeless validity of a URI, I advocate that the current URIs 
are good enough. There is an understanding of a commitment to updating them and 
refining them in the future.

— 
Eric Morgan


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Ethan Gruber
I see that serialization has a different definition in computer science
than I thought it did.


On Tue, Nov 19, 2013 at 10:36 AM, Ross Singer  wrote:

> That's still not a "serialization".  It's just a similar data model.
>  Pretty huge difference.
>
> -Ross.
>
>
> On Tue, Nov 19, 2013 at 10:31 AM, Ethan Gruber  wrote:
>
> > I'm not sure that I agree that RDF is not a serialization.  It really
> > depends on the context of the system and intended use of the linked data.
> > For example, TEI is designed with a specific purpose which cannot be
> > replicated in RDF (at least, not very easily at all), but deriving RDF
> from
> > highly-linked TEI to put into an endpoint can open doors to queries which
> > are otherwise impossible to make on the data.  This certainly requires
> some
> > rethinking of the way texts interact.  But perhaps it may be best to say
> > that RDF *can* (but not necessarily) be a derivation, rather than
> > serialization, of some larger, more complex canonical data model.
> >
> > Ethan
> >
> >
> > On Tue, Nov 19, 2013 at 9:54 AM, Aaron Rubinstein <
> > arubi...@library.umass.edu> wrote:
> >
> > > I think you’ve hit the nail on the head here, Karen. I would just add,
> or
> > > maybe reassure, that this does not necessarily require rethinking your
> > > existing metadata but how to translate that existing metadata into a
> > linked
> > > data environment. Though this might seem like a pain, in many cases it
> > will
> > > actually inspire you to go back and improve/increase the value of that
> > > existing metadata.
> > >
> > > This is definitely looking awesome, Eric!
> > >
> > > Aaron
> > >
> > > On Nov 19, 2013, at 9:41 AM, Karen Coyle  wrote:
> > >
> > > > Eric, I think this skips a step - which is the design step in which
> you
> > > create a domain model that uses linked data as its basis. RDF is not a
> > > serialization; it actually may require you to re-think the basic
> > structure
> > > of your metadata. The reason for that is that it provides capabilities
> > that
> > > record-based data models do not. Rather than starting with current
> > > metadata, you need to take a step back and ask: what does my
> information
> > > world look like as linked data?
> > > >
> > > > I repeat: RDF is NOT A SERIALIZATION.
> > > >
> > > > kc
> > > >
> > > > On 11/19/13 5:04 AM, Eric Lease Morgan wrote:
> > > >> I believe participating in the Semantic Web and providing content
> via
> > > the principles of linked data is not "rocket surgery", especially for
> > > cultural heritage institutions -- libraries, archives, and museums.
> Here
> > is
> > > a simple recipe for their participation:
> > > >>
> > > >>   1. use existing metadata standards (MARC, EAD, etc.) to describe
> > > >>  collections
> > > >>
> > > >>   2. use any number of existing tools to convert the metadata to
> > > >>  HTML, and save the HTML on a Web server
> > > >>
> > > >>   3. use any number of existing tools to convert the metadata to
> > > >>  RDF/XML (or some other "serialization" of RDF), and save the
> > > >>  RDF/XML on a Web server
> > > >>
> > > >>   4. rest, congratulate yourself, and share your experience with
> > > >>  others in your domain
> > > >>
> > > >>   5. after the first time though, go back to Step #1, but this time
> > > >>  work with other people inside your domain making sure you use
> as
> > > >>  many of the same URIs as possible
> > > >>
> > > >>   6. after the second time through, go back to Step #1, but this
> > > >>  time supplement access to your linked data with a triple store,
> > > >>  thus supporting search
> > > >>
> > > >>   7. after the third time through, go back to Step #1, but this
> > > >>  time use any number of existing tools to expose the content in
> > > >>  your other information systems (relational databases, OAI-PMH
> > > >>  data repositories, etc.)
> > > >>
> > > >>   8. for dessert, cogitate ways to exploit the linked data in your
> > > >>  domain to discover new and additional relationships between
> URIs,
> > > >>  and thus make the Semantic Web more of a reality
> > > >>
> > > >> What do you think?
> > > >>
> > > >> I am in the process of writing a guidebook on the topic of linked
> data
> > > and archives. In the guidebook I will elaborate on this recipe and
> > provide
> > > instructions for its implementation. [1]
> > > >>
> > > >> [1] guidebook - http://sites.tufts.edu/liam/
> > > >>
> > > >> --
> > > >> Eric Lease Morgan
> > > >> University of Notre Dame
> > > >
> > > > --
> > > > Karen Coyle
> > > > kco...@kcoyle.net http://kcoyle.net
> > > > m: 1-510-435-8234
> > > > skype: kcoylenet
> > >
> >
>


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Ross Singer
That's still not a "serialization".  It's just a similar data model.
 Pretty huge difference.

-Ross.


On Tue, Nov 19, 2013 at 10:31 AM, Ethan Gruber  wrote:

> I'm not sure that I agree that RDF is not a serialization.  It really
> depends on the context of the system and intended use of the linked data.
> For example, TEI is designed with a specific purpose which cannot be
> replicated in RDF (at least, not very easily at all), but deriving RDF from
> highly-linked TEI to put into an endpoint can open doors to queries which
> are otherwise impossible to make on the data.  This certainly requires some
> rethinking of the way texts interact.  But perhaps it may be best to say
> that RDF *can* (but not necessarily) be a derivation, rather than
> serialization, of some larger, more complex canonical data model.
>
> Ethan
>
>
> On Tue, Nov 19, 2013 at 9:54 AM, Aaron Rubinstein <
> arubi...@library.umass.edu> wrote:
>
> > I think you’ve hit the nail on the head here, Karen. I would just add, or
> > maybe reassure, that this does not necessarily require rethinking your
> > existing metadata but how to translate that existing metadata into a
> linked
> > data environment. Though this might seem like a pain, in many cases it
> will
> > actually inspire you to go back and improve/increase the value of that
> > existing metadata.
> >
> > This is definitely looking awesome, Eric!
> >
> > Aaron
> >
> > On Nov 19, 2013, at 9:41 AM, Karen Coyle  wrote:
> >
> > > Eric, I think this skips a step - which is the design step in which you
> > create a domain model that uses linked data as its basis. RDF is not a
> > serialization; it actually may require you to re-think the basic
> structure
> > of your metadata. The reason for that is that it provides capabilities
> that
> > record-based data models do not. Rather than starting with current
> > metadata, you need to take a step back and ask: what does my information
> > world look like as linked data?
> > >
> > > I repeat: RDF is NOT A SERIALIZATION.
> > >
> > > kc
> > >
> > > On 11/19/13 5:04 AM, Eric Lease Morgan wrote:
> > >> I believe participating in the Semantic Web and providing content via
> > the principles of linked data is not "rocket surgery", especially for
> > cultural heritage institutions -- libraries, archives, and museums. Here
> is
> > a simple recipe for their participation:
> > >>
> > >>   1. use existing metadata standards (MARC, EAD, etc.) to describe
> > >>  collections
> > >>
> > >>   2. use any number of existing tools to convert the metadata to
> > >>  HTML, and save the HTML on a Web server
> > >>
> > >>   3. use any number of existing tools to convert the metadata to
> > >>  RDF/XML (or some other "serialization" of RDF), and save the
> > >>  RDF/XML on a Web server
> > >>
> > >>   4. rest, congratulate yourself, and share your experience with
> > >>  others in your domain
> > >>
> > >>   5. after the first time though, go back to Step #1, but this time
> > >>  work with other people inside your domain making sure you use as
> > >>  many of the same URIs as possible
> > >>
> > >>   6. after the second time through, go back to Step #1, but this
> > >>  time supplement access to your linked data with a triple store,
> > >>  thus supporting search
> > >>
> > >>   7. after the third time through, go back to Step #1, but this
> > >>  time use any number of existing tools to expose the content in
> > >>  your other information systems (relational databases, OAI-PMH
> > >>  data repositories, etc.)
> > >>
> > >>   8. for dessert, cogitate ways to exploit the linked data in your
> > >>  domain to discover new and additional relationships between URIs,
> > >>  and thus make the Semantic Web more of a reality
> > >>
> > >> What do you think?
> > >>
> > >> I am in the process of writing a guidebook on the topic of linked data
> > and archives. In the guidebook I will elaborate on this recipe and
> provide
> > instructions for its implementation. [1]
> > >>
> > >> [1] guidebook - http://sites.tufts.edu/liam/
> > >>
> > >> --
> > >> Eric Lease Morgan
> > >> University of Notre Dame
> > >
> > > --
> > > Karen Coyle
> > > kco...@kcoyle.net http://kcoyle.net
> > > m: 1-510-435-8234
> > > skype: kcoylenet
> >
>


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Ethan Gruber
I'm not sure that I agree that RDF is not a serialization.  It really
depends on the context of the system and intended use of the linked data.
For example, TEI is designed with a specific purpose which cannot be
replicated in RDF (at least, not very easily at all), but deriving RDF from
highly-linked TEI to put into an endpoint can open doors to queries which
are otherwise impossible to make on the data.  This certainly requires some
rethinking of the way texts interact.  But perhaps it may be best to say
that RDF *can* (but not necessarily) be a derivation, rather than
serialization, of some larger, more complex canonical data model.

Ethan


On Tue, Nov 19, 2013 at 9:54 AM, Aaron Rubinstein <
arubi...@library.umass.edu> wrote:

> I think you’ve hit the nail on the head here, Karen. I would just add, or
> maybe reassure, that this does not necessarily require rethinking your
> existing metadata but how to translate that existing metadata into a linked
> data environment. Though this might seem like a pain, in many cases it will
> actually inspire you to go back and improve/increase the value of that
> existing metadata.
>
> This is definitely looking awesome, Eric!
>
> Aaron
>
> On Nov 19, 2013, at 9:41 AM, Karen Coyle  wrote:
>
> > Eric, I think this skips a step - which is the design step in which you
> create a domain model that uses linked data as its basis. RDF is not a
> serialization; it actually may require you to re-think the basic structure
> of your metadata. The reason for that is that it provides capabilities that
> record-based data models do not. Rather than starting with current
> metadata, you need to take a step back and ask: what does my information
> world look like as linked data?
> >
> > I repeat: RDF is NOT A SERIALIZATION.
> >
> > kc
> >
> > On 11/19/13 5:04 AM, Eric Lease Morgan wrote:
> >> I believe participating in the Semantic Web and providing content via
> the principles of linked data is not "rocket surgery", especially for
> cultural heritage institutions -- libraries, archives, and museums. Here is
> a simple recipe for their participation:
> >>
> >>   1. use existing metadata standards (MARC, EAD, etc.) to describe
> >>  collections
> >>
> >>   2. use any number of existing tools to convert the metadata to
> >>  HTML, and save the HTML on a Web server
> >>
> >>   3. use any number of existing tools to convert the metadata to
> >>  RDF/XML (or some other "serialization" of RDF), and save the
> >>  RDF/XML on a Web server
> >>
> >>   4. rest, congratulate yourself, and share your experience with
> >>  others in your domain
> >>
> >>   5. after the first time though, go back to Step #1, but this time
> >>  work with other people inside your domain making sure you use as
> >>  many of the same URIs as possible
> >>
> >>   6. after the second time through, go back to Step #1, but this
> >>  time supplement access to your linked data with a triple store,
> >>  thus supporting search
> >>
> >>   7. after the third time through, go back to Step #1, but this
> >>  time use any number of existing tools to expose the content in
> >>  your other information systems (relational databases, OAI-PMH
> >>  data repositories, etc.)
> >>
> >>   8. for dessert, cogitate ways to exploit the linked data in your
> >>  domain to discover new and additional relationships between URIs,
> >>  and thus make the Semantic Web more of a reality
> >>
> >> What do you think?
> >>
> >> I am in the process of writing a guidebook on the topic of linked data
> and archives. In the guidebook I will elaborate on this recipe and provide
> instructions for its implementation. [1]
> >>
> >> [1] guidebook - http://sites.tufts.edu/liam/
> >>
> >> --
> >> Eric Lease Morgan
> >> University of Notre Dame
> >
> > --
> > Karen Coyle
> > kco...@kcoyle.net http://kcoyle.net
> > m: 1-510-435-8234
> > skype: kcoylenet
>


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Aaron Rubinstein
I think you’ve hit the nail on the head here, Karen. I would just add, or maybe 
reassure, that this does not necessarily require rethinking your existing 
metadata but how to translate that existing metadata into a linked data 
environment. Though this might seem like a pain, in many cases it will actually 
inspire you to go back and improve/increase the value of that existing metadata.

This is definitely looking awesome, Eric!

Aaron

On Nov 19, 2013, at 9:41 AM, Karen Coyle  wrote:

> Eric, I think this skips a step - which is the design step in which you 
> create a domain model that uses linked data as its basis. RDF is not a 
> serialization; it actually may require you to re-think the basic structure of 
> your metadata. The reason for that is that it provides capabilities that 
> record-based data models do not. Rather than starting with current metadata, 
> you need to take a step back and ask: what does my information world look 
> like as linked data?
> 
> I repeat: RDF is NOT A SERIALIZATION.
> 
> kc
> 
> On 11/19/13 5:04 AM, Eric Lease Morgan wrote:
>> I believe participating in the Semantic Web and providing content via the 
>> principles of linked data is not "rocket surgery", especially for cultural 
>> heritage institutions -- libraries, archives, and museums. Here is a simple 
>> recipe for their participation:
>> 
>>   1. use existing metadata standards (MARC, EAD, etc.) to describe
>>  collections
>> 
>>   2. use any number of existing tools to convert the metadata to
>>  HTML, and save the HTML on a Web server
>> 
>>   3. use any number of existing tools to convert the metadata to
>>  RDF/XML (or some other "serialization" of RDF), and save the
>>  RDF/XML on a Web server
>> 
>>   4. rest, congratulate yourself, and share your experience with
>>  others in your domain
>> 
>>   5. after the first time though, go back to Step #1, but this time
>>  work with other people inside your domain making sure you use as
>>  many of the same URIs as possible
>> 
>>   6. after the second time through, go back to Step #1, but this
>>  time supplement access to your linked data with a triple store,
>>  thus supporting search
>> 
>>   7. after the third time through, go back to Step #1, but this
>>  time use any number of existing tools to expose the content in
>>  your other information systems (relational databases, OAI-PMH
>>  data repositories, etc.)
>> 
>>   8. for dessert, cogitate ways to exploit the linked data in your
>>  domain to discover new and additional relationships between URIs,
>>  and thus make the Semantic Web more of a reality
>> 
>> What do you think?
>> 
>> I am in the process of writing a guidebook on the topic of linked data and 
>> archives. In the guidebook I will elaborate on this recipe and provide 
>> instructions for its implementation. [1]
>> 
>> [1] guidebook - http://sites.tufts.edu/liam/
>> 
>> --
>> Eric Lease Morgan
>> University of Notre Dame
> 
> -- 
> Karen Coyle
> kco...@kcoyle.net http://kcoyle.net
> m: 1-510-435-8234
> skype: kcoylenet


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Karen Coyle
Eric, I think this skips a step - which is the design step in which you 
create a domain model that uses linked data as its basis. RDF is not a 
serialization; it actually may require you to re-think the basic 
structure of your metadata. The reason for that is that it provides 
capabilities that record-based data models do not. Rather than starting 
with current metadata, you need to take a step back and ask: what does 
my information world look like as linked data?


I repeat: RDF is NOT A SERIALIZATION.

kc

On 11/19/13 5:04 AM, Eric Lease Morgan wrote:

I believe participating in the Semantic Web and providing content via the principles of 
linked data is not "rocket surgery", especially for cultural heritage 
institutions -- libraries, archives, and museums. Here is a simple recipe for their 
participation:

   1. use existing metadata standards (MARC, EAD, etc.) to describe
  collections

   2. use any number of existing tools to convert the metadata to
  HTML, and save the HTML on a Web server

   3. use any number of existing tools to convert the metadata to
  RDF/XML (or some other "serialization" of RDF), and save the
  RDF/XML on a Web server

   4. rest, congratulate yourself, and share your experience with
  others in your domain

   5. after the first time though, go back to Step #1, but this time
  work with other people inside your domain making sure you use as
  many of the same URIs as possible

   6. after the second time through, go back to Step #1, but this
  time supplement access to your linked data with a triple store,
  thus supporting search

   7. after the third time through, go back to Step #1, but this
  time use any number of existing tools to expose the content in
  your other information systems (relational databases, OAI-PMH
  data repositories, etc.)

   8. for dessert, cogitate ways to exploit the linked data in your
  domain to discover new and additional relationships between URIs,
  and thus make the Semantic Web more of a reality

What do you think?

I am in the process of writing a guidebook on the topic of linked data and 
archives. In the guidebook I will elaborate on this recipe and provide 
instructions for its implementation. [1]

[1] guidebook - http://sites.tufts.edu/liam/

--
Eric Lease Morgan
University of Notre Dame


--
Karen Coyle
kco...@kcoyle.net http://kcoyle.net
m: 1-510-435-8234
skype: kcoylenet


[CODE4LIB] ALA's Carroll Preston Baber Research Grant--Call for Proposals

2013-11-19 Thread Mary Popp
*Carroll Preston Baber research grant call for proposals*


 Do you have a project that is just waiting for the right funding?  Are you
thinking about ways that libraries can improve services to users?


The American Library Association (ALA) gives an annual grant for those
conducting research that will lead to the improvement of services to users.
The Carroll Preston Baber Research Grant is given to one or more librarians
or library educators who will conduct innovative research that could lead
to an improvement in services to any specified group of people.


The grant, up to $3,000, will be given to a proposed project that aims to
answer a question of vital importance to the library community that is
national in scope. Among the review panel criteria are:

   - The research problem is clearly defined, with a specific question or
   questions that can be answered by collecting data. The applicant(s) clearly
   describe a strategy for data collection whose methods are appropriate to
   the research question(s). A review of the literature, methodologies, etc.
   is not considered research (e.g., methodology review rather than
   application of a methodology) for purposes of the award, except where the
   literature review is the primary method of collecting data.
   - The research question focuses on benefits to library users and should
   be applied and have practical value as opposed to theoretical.
   - The applicant(s) demonstrate ability to undertake and successfully
   complete the project. The application provides evidence that sufficient
   time and resources have been allocated to the effort. Appropriate
   institutional commitment to the project has been secured.

Any ALA member may apply, and the Jury would welcome projects that involve
both a practicing librarian and a researcher.


Deadline is *January 8, 2014*.

Check out this web site to find procedures and an application form:
http://www.ala.org/ala/aboutala/offices/ors/orsawards/baberresearchgrant/babercarroll.cfm
 See the section on *How to Apply*.  Also see related documents linked near
the bottom of the page for:

Schedule and Procedures
http://www.ala.org/offices/ors/orsawards/baberresearchgrant/schedandprocedures

Proposal Requirements and Application Cover Sheet:
http://www.ala.org/offices/ors/orsawards/baberresearchgrant/requirements

Full press release:
http://www.ala.org/news/press-releases/2013/11/baber-research-grant-proposals-due-january-8

Questions?   Contact: Mary Pagliero Popp at p...@indiana.edu.


-- 
*Mary*
*--*

*Mary Pagliero Popp, Chair, Baber Award Jury, American Library Association*


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Robert Forkel
Hi Eric,
while I also think this is not rocket surgery, I'd like to point out that
trial (and potentially error) as suggested by your "go back to step #1"
instructions is not a good solution to coming up with URIs. I think once
published - i.e. put on a webserver - you should be able to keep the URIs
in your RDF persistent. Otherwise you are polluting the Semantic Web with
dead links and make it hard for aggregators to find out whether the data
they harvested is still valid.
So while iterative approaches are pragmatic and often work out well, for
the particular issue of coming up with URIs I'd recommend spending as much
thought before publishing as you can spend.
best
robert



On Tue, Nov 19, 2013 at 2:04 PM, Eric Lease Morgan  wrote:

> I believe participating in the Semantic Web and providing content via the
> principles of linked data is not "rocket surgery", especially for cultural
> heritage institutions -- libraries, archives, and museums. Here is a simple
> recipe for their participation:
>
>   1. use existing metadata standards (MARC, EAD, etc.) to describe
>  collections
>
>   2. use any number of existing tools to convert the metadata to
>  HTML, and save the HTML on a Web server
>
>   3. use any number of existing tools to convert the metadata to
>  RDF/XML (or some other "serialization" of RDF), and save the
>  RDF/XML on a Web server
>
>   4. rest, congratulate yourself, and share your experience with
>  others in your domain
>
>   5. after the first time though, go back to Step #1, but this time
>  work with other people inside your domain making sure you use as
>  many of the same URIs as possible
>
>   6. after the second time through, go back to Step #1, but this
>  time supplement access to your linked data with a triple store,
>  thus supporting search
>
>   7. after the third time through, go back to Step #1, but this
>  time use any number of existing tools to expose the content in
>  your other information systems (relational databases, OAI-PMH
>  data repositories, etc.)
>
>   8. for dessert, cogitate ways to exploit the linked data in your
>  domain to discover new and additional relationships between URIs,
>  and thus make the Semantic Web more of a reality
>
> What do you think?
>
> I am in the process of writing a guidebook on the topic of linked data and
> archives. In the guidebook I will elaborate on this recipe and provide
> instructions for its implementation. [1]
>
> [1] guidebook - http://sites.tufts.edu/liam/
>
> --
> Eric Lease Morgan
> University of Notre Dame
>


Re: [CODE4LIB] linked data recipe

2013-11-19 Thread Brian Zelip
It's a great start Eric. It helps me think that I can do it.  Looking
forward to more.

Brian Zelip
UIUC


On Tue, Nov 19, 2013 at 7:04 AM, Eric Lease Morgan  wrote:

> I believe participating in the Semantic Web and providing content via the
> principles of linked data is not "rocket surgery", especially for cultural
> heritage institutions -- libraries, archives, and museums. Here is a simple
> recipe for their participation:
>
>   1. use existing metadata standards (MARC, EAD, etc.) to describe
>  collections
>
>   2. use any number of existing tools to convert the metadata to
>  HTML, and save the HTML on a Web server
>
>   3. use any number of existing tools to convert the metadata to
>  RDF/XML (or some other "serialization" of RDF), and save the
>  RDF/XML on a Web server
>
>   4. rest, congratulate yourself, and share your experience with
>  others in your domain
>
>   5. after the first time though, go back to Step #1, but this time
>  work with other people inside your domain making sure you use as
>  many of the same URIs as possible
>
>   6. after the second time through, go back to Step #1, but this
>  time supplement access to your linked data with a triple store,
>  thus supporting search
>
>   7. after the third time through, go back to Step #1, but this
>  time use any number of existing tools to expose the content in
>  your other information systems (relational databases, OAI-PMH
>  data repositories, etc.)
>
>   8. for dessert, cogitate ways to exploit the linked data in your
>  domain to discover new and additional relationships between URIs,
>  and thus make the Semantic Web more of a reality
>
> What do you think?
>
> I am in the process of writing a guidebook on the topic of linked data and
> archives. In the guidebook I will elaborate on this recipe and provide
> instructions for its implementation. [1]
>
> [1] guidebook - http://sites.tufts.edu/liam/
>
> --
> Eric Lease Morgan
> University of Notre Dame
>


[CODE4LIB] linked data recipe

2013-11-19 Thread Eric Lease Morgan
I believe participating in the Semantic Web and providing content via the 
principles of linked data is not "rocket surgery", especially for cultural 
heritage institutions -- libraries, archives, and museums. Here is a simple 
recipe for their participation:

  1. use existing metadata standards (MARC, EAD, etc.) to describe
 collections

  2. use any number of existing tools to convert the metadata to
 HTML, and save the HTML on a Web server

  3. use any number of existing tools to convert the metadata to
 RDF/XML (or some other "serialization" of RDF), and save the
 RDF/XML on a Web server

  4. rest, congratulate yourself, and share your experience with
 others in your domain

  5. after the first time though, go back to Step #1, but this time
 work with other people inside your domain making sure you use as
 many of the same URIs as possible

  6. after the second time through, go back to Step #1, but this
 time supplement access to your linked data with a triple store,
 thus supporting search

  7. after the third time through, go back to Step #1, but this
 time use any number of existing tools to expose the content in
 your other information systems (relational databases, OAI-PMH
 data repositories, etc.)

  8. for dessert, cogitate ways to exploit the linked data in your
 domain to discover new and additional relationships between URIs,
 and thus make the Semantic Web more of a reality 

What do you think?

I am in the process of writing a guidebook on the topic of linked data and 
archives. In the guidebook I will elaborate on this recipe and provide 
instructions for its implementation. [1]

[1] guidebook - http://sites.tufts.edu/liam/

--
Eric Lease Morgan
University of Notre Dame


Re: [CODE4LIB] Tool for managing subscription content metadata

2013-11-19 Thread raffaele messuti
Barnes, Hugh wrote:
> I have been thinking along the lines of Mediawiki, maybe with a good

the mediawiki wikibase[1] extension (that runs wikidata.org)
should be a good solution for your needs.

otherwise i've always heard good things about topincs[2]
(but i confess to have some limitations understanding
data modeling using topicmaps)

bye.


[1] http://www.mediawiki.org/wiki/Wikibase
[2] http://www.cerny-online.com/topincs/


--
raffaele, @atomotic