[CODE4LIB] NEWS RELEASE: VIVO 1.8.1 Now Available–Improved Performance, New Visualizations

2015-11-10 Thread Graham Triggs
FOR IMMEDIATE RELEASE

November 10, 2015

Read it online: 
http://vivoweb.org/blog/2015/11/now-available-vivo-181-—-improved-performance-and-new-visualizations
 
<http://vivoweb.org/blog/2015/11/now-available-vivo-181-%E2%80%94-improved-performance-and-new-visualizations>
Contact: Graham Triggs <gtri...@duraspace.org <mailto:gtri...@duraspace.org>>

NOW AVAILABLE: VIVO 1.8.1 — Improved Performance and New Visualizations

Winchester, MA  On November 10, 2015 VIVO 1.8.1 was released by the VIVO team.  
This new release offers users vastly improved performance, new and better 
visualizations, as well as bug fixes.

• Download VIVO 1.8.1: http://bit.ly/vivo-1_8_1-release 
<http://bit.ly/vivo-1_8_1-release>

Full release notes are available on the VIVO wiki: 
https://wiki.duraspace.org/display/VIVO/VIVO+v1.8.1+Release+Notes 
<https://wiki.duraspace.org/display/VIVO/VIVO+v1.8.1+Release+Notes>.

Performance improvements

Users should see up to 75% reduction in time to display profiles compared to 
VIVO 1.8, and a 30% reduction compared to VIVO 1.7. These findings have been 
observed on both small data sets (200 people, 3,500 articles), and large data 
sets (4,500 people, 40,000 articles).  Most profiles - even large/complex 
profiles - display within approximately two seconds.

Visualizations

The ability to examine the VIVO network from many points-of-view is at the 
heart of connecting researchers, ideas and resources. VIVO 1.8.1 offers users 
new and improved visualizations that make it easier to “see” formative research 
as it emerges at individual, institutional and topic levels. Click here to see 
an example of the new Visualizations in 1.8.1

New Javascript based versions of Co-Author and Co-Investigator  are networks 
available.  These new visualizations do not require Flash and display on mobile 
devices such as phones and tablets.

All visualizations have significant performance improvements. Maps of Science, 
Temporal Graphs and Co-author and  Co-Investigator networks now all complete in 
just a few seconds.  The very largest Maps of Science may require up to two 
minutes to complete.

Additional Improvements

New AltMetric badges are enabled on publications by default, providing direct 
link to AltMetric information regarding the publication.

Additional improvements reduce time and resource usage for indexing and 
inferencing. More than a dozen have been made to improve user experience.

We look forward to hearing from you about VIVO 1.8.1. Please contact VIVO Tech 
Lead Graham Triggs with questions, suggestions and feedback: 
<gtri...@duraspace.org <mailto:gtri...@duraspace.org>>.

About VIVO

VIVO (http://vivoweb.org <http://vivoweb.org/>) is an open source, open 
ontology, open process platform for hosting information about the interests, 
activities and accomplishments of scientists and scholars. VIVO supports open 
development and integration of science and scholarship through simple, standard 
semantic web! technologies. VIVO was originally funded by Cornell University 
and the National Institutes of Health (U24 RR029822) and is currently a 
community-supported project under the DuraSpace umbrella.

How Does DuraSpace Help?

The DuraSpace (http://duraspace.org <http://duraspace.org/>) organization is an 
independent 501(c)(3) not-for-profit providing leadership and innovation for 
open technologies that promote durable, persistent access and discovery of 
digital data. Our values are expressed in our organizational byline, "Committed 
to our digital future."

DuraSpace works collaboratively with organizations that use VIVO to advance the 
design, development and sustainability of the project. As a non-profit, 
DuraSpace provides technical leadership, sustainability planning, fundraising, 
community development, marketing and communications, collaborations and 
strategic partnerships, and administration.

Re: [CODE4LIB] separate list for Jobs

2014-05-09 Thread Graham Triggs
On 8 May 2014 18:08, Scott Fisher scott.fis...@ucop.edu wrote:

 4.  I find arguments to the effect of I love looking at jobs² orthoginal
 to the discussion since weıre not talking about disallowing job postings,
 but just moving them to a separate list.  Anyone who is interested enough
 in jobs could also add a separate jobs list to go to their daily email
 inbox, so Iım not sure how it would be a loss of all jobs emails they
 like.  I suppose this argument essentially comes down to the same kind of
 argument as the pro-email-filter one.  Basically that argument is, ³just
 do something different to receive the emails you want the way you want
 them.²  But in this case the argument is coming right back at you from the
 other direction of suggting a separate email list.


I have filters set up, and find they just don't work reliably. OK, they
work 9 times out of 10, but things always slip through.

Imho, there are more people inconvenienced by having jobs on the list
(setting up filters, filters not working, unable to filter digests, etc.)
than there are people inconvenienced by having a separate list for jobs (is
there really anyone that can't sign up for a separate list? is it
impossible to add a subscribe link to the jobs list in a footer?)

(actually, I note there isn't a footer on this mailing list at all - must
be one of the only mailing lists that doesn't include a footer with
unsubscribe and general information. Technically, doesn't this mean it
falls foul of anti-spam legislation in some territories?)

G


Re: [CODE4LIB] Anyone have access to well-disambiguated sets of publication data?

2013-07-10 Thread Graham Triggs
Hi Paul,

I guess this rather depends on your purposes. From the way you've asked the
question, it sounds like you are looking for a control set of data to
compare your own efforts of author disambiguation to (rather than simply
having good sources of disambiguated data - presumably for feeding into a
VIVO instance)?

In case you haven't looked at it already, you might find the Profiles RNS
Disambiguation Engine useful:

http://profiles.catalyst.harvard.edu/docs/ProfilesRNS_DisambiguationEngine.pdf

Although this would just cover Medline/PubMed data. This could work just as
a source of disambiguated data for you, not just as a control set for your
own implementations.

Additionally, if you are not adverse to using other software to provide
disambiguated data, rather than implementing your own solution, then you
might want to look at research information management software (e.g.
Symplectic Elements). These specialise in acquiring data from a number of
data sources (including PubMed), and helping you create a clean,
disambiguated set of publication data - and typically provide APIs that
allow you to interact and/or extract that information for re-use in other
systems.

Regards,
G




On 9 July 2013 16:32, Paul Albert paa2...@med.cornell.edu wrote:

 I am exploring methods for author disambiguation, and I would like to have
 access to one or more set of well-disambiguated data set containing:
 – a unique author identifier (email address, institutional identifier)
 – a unique article identifier (PMID, DOI, etc.)
 – a unique journal identifier (ISSN)

 Definition for well-disambiguated – for a given set of authors, you know
 the identity of their journal articles to a precision and recall of greater
 than 90-95%.

 Any ideas?

 thanks,
 Paul


 Paul Albert
 Project Manager, VIVO
 Weill Cornell Medical Library
 646.962.2551



Re: [CODE4LIB] It's all job postings!

2012-08-15 Thread Graham Triggs
On 6 August 2012 13:19, Ed Summers e...@pobox.com wrote:
 150 people responded about whether jobs.code4lib.org posting should
 come to the discussion list:

 yes: 132
 no: 10
 who cares: 8

 93% in support or agnostic seems to be a good indicator that the
 postings should continue to come to the list for now.

I'm not entirely convinced about that assessment. I quite readily
agree that the jobs should be posted to *a* mailing list, I'm not so
sure that it should be this mailing list.

It's been discussed about filtering the jobs sent to the list, but I
already filter the code4lib mailing list into a tag. It's been a bit
of a faff, but I've subdivided the filtering so that I can get the
messages sent from jobs@... to go to a different tag. But then Ed
replied to one, so now it appears in both tags, and because I'm using
Gmail, it takes the whole thread with it.

So filtering really isn't a solution.

Rather than just asking whether jobs should come to this mailing list,
maybe we can ask whether a separate mailing list should be set up,
specifically for jobs. The two mailing lists could be cross promoted
(e.g. a standard footer), and people can choose whether they want or
don't want to receive them. And we can still have
discussions/follow-ups about those jobs on that mailing list.

Even though the vast majority of the postings aren't applicable to me,
I would probably still sign up to a separate jobs mailing list as it
is of interest - but I would at least then be able to keep that
separate from the main discussions, which is something I can't
effectively do right now.

G


Re: [CODE4LIB] Metadata

2012-02-16 Thread Graham Triggs
On 16 February 2012 12:13, suzanne.pilsk suzanne.pi...@gmail.com wrote:
 (Pssst: Does it matter if you call it data and I call it metadata?)

Yes. Yes it does. Although possibly not for the reasons you might
think I'll give.

I think it's really important that developers (and researchers) get
out of the mindset that:

a) metadata isn't data
b) metadata isn't important
c) data can't be both metadata and data

Because the chances are that it is all important, but in different
ways, and has different needs in frequency of use, indexing, etc.

Treating everything the same, and in one big bucket, is likely to slow
down the processes of indexing, locating, and scanning over possible
resources.

So there is a need to think about things in terms of what is required
for different purposes - what is needed for locating and describing
resources, and everything that is required to use those resources. And
that may well mean having bits of information that are duplicated in
more than one bucket.

But what we can't be is scared of identifying that there are different
needs and requirements, different scopes of relevant data. Or that
users actually have names for these scopes.

G


Re: [CODE4LIB] Metadata

2012-02-13 Thread Graham Triggs
On 13 February 2012 15:57, Nate Vack njv...@wisc.edu wrote:
 My take on this discussion, coming from a research lab: Metadata isn't meta.

Well, coming from a publishing and repositories world, my take is
slightly different.

 For example, in recordings of, say, blood pressure over time, it's
 common to think about things such as participant identifiers,
 acquisition dates, event markers, and sampling rates as metadata,
 and the actual measurements as data.

 But really: those meta things aren't ancillary to data analysis;
 they're essential in keeping analyses organized, and often important
 parameters in running an analysis at all.

That's an interesting distinction though. Do you need all that data in
order to make sense of the results? You don't [necessarily] need to
know who conducted some research, or when they conducted it in order
to analyse and make sense of the data. In the context of having the
data, this other information becomes irrelevant in terms of
understanding what that data says.

But in a wider context, you do need such additional information in
order to be able to use it. If you don't know who conducted the
research, when it was conducted, etc. then you can't reference it.You
can't place it into another context (a follow up study to validate the
findings, or see if something has changed over time). And it's not a
case of saying something has to fall into one category or another, it
may be necessary / useful in both.

 Breaking things down into data versus metadata I think, encourages a
 false (and not very interesting) dichotomy. If information has a use,
 call it what it is: data. Store everything that's useful.

The problem isn't that we have labels for data to be used in different
contexts. It's that just because something does have a label [and may
not necessarily be important to you in your context], that doesn't
mean that it's any less important than something else with another
label.

G


Re: [CODE4LIB] Google Analytics w/ Sub-sub-domains

2012-02-07 Thread Graham Triggs
Having a quick look at your site(s), it appears that those links through
resources... are redirecting through other pages.

If so, it may have more to do with not being able to follow the flow of
page accesses, than the ability to track a sub domain.

G
On Feb 7, 2012 12:30 PM, Predmore, Andrew andrew.predm...@yale.edu
wrote:

 Yes, the code is not consistent across our pages.  That is the problem I
 am trying to solve.  The pages are served by dozens of different systems.
 Therefore, I had held off on changing all of them until I knew I had
 working code.

 At this point, based on my research and the feedback here I will be going
 with this:

 _gaq.push(['_setDomainName', '.yale.edu']);

 After I get that code on the majority of pages, I can start looking into
 the separate problems I am having with drop-offs and setting up a profile
 that will show me the sub-domains.

 Thank you to everyone for the help.


 --
 Clayton Andrew Predmore
 Manager, Web Operations
 Yale University Library
 andrew.predm...@yale.edu






 On 2/6/12 4:53 PM, Brian Tingle brian.tingle.cdlib@gmail.com
 wrote:

 Henry, that is what you need to do if you want to track the same page to
 two different google analytics properties and you are using the
 legacy synchronous code.  It sounds like yale wants to collect all this
 use
 under one UA- google analytics property (it is just that the property
 spans
 multiple subdomains).
 
 I think the link I sent to
 
 http://code.google.com/apis/analytics/docs/tracking/gaTrackingSite.html
 ht
 tp://%22
 
 addresses the yale case; and I the way I read it adding this:
 
  _gaq.push(['_setDomainName', '.yale.edu']);
 
 Or  _gaq.push(['_setDomainName', 'library.yale.edu']);
 
 on _every_ page should  work.
 
 Right now, I only see _setDomainName on the home page.  If this is not the
 _same_ on all the pages, the cookies won't be shared as users move between
 the sites.
 
 For example;
 
 view-source:http://www.library.yale.edu/researcheducation/
 
 This page is missing
 
 _gaq.push(['_setDomainName', '.yale.edu']);
 
  It will only work prospectively (it won't change the past) but when all
 pages are sharing the same _setDomainName then they should all share the
 same cookies and the links between pages should be counted correctly.
 
 But google analytics can get tricky, just when I think I understand
 something it changes.  I find I have to double check things a lot with the
 debug toolbar to make sure the right stuff is getting sent to google (esp.
 when setting up custom events or setting up multiple trackers on the same
 page).  You should be able to use it to verify that the same session and
 cookies are being used as you go from page to page.  In the chrome debug
 nowadays you can right click on the console log and select Preserve Log
 upon navigation which makes this a lot easier.