I'd like to restate the initial question.
Why did wikidata choose shex instead of other approaches?
>From this very detailed comparison
http://book.validatingrdf.com/bookHtml013.html (thank you Andra!) I could
see arguments in both directions. I'm curious to know what swayed the
wikidata
Stas,
One thing that would be extremely useful right away would be an integration
of the free text search from MediaWiki API. That is one area SPARQL does
not handle well, Mediawiki does well, and its pretty important for many
applications. If there were some clever way of mixing (fast!) free
-- Forwarded message --
Reminder: Deadline May 4th
Full CFP: http://www.humancomputation.com/2017/submit.html
The Fifth AAAI Conference on Human Computation and Crowdsourcing (HCOMP
2017) will be held in Quebec City, Canada, Oct. 24-26, 2017. It will be
sponsored by the
The 'all cities' part of your query seems strange. If you replace it with
a simple 'instance of city' - ?x wdt:P31 wd:Q515 . it returns 145 results
in 341ms. http://tinyurl.com/m4ffatg
Why would you ask for things that have 'occupation' or 'position' with
value city?
Also, I'm not familiar with
Denny,
This is a really good question: "How bad / good does Wikidata as a whole
fit the role of an open vocabulary for content tagging?"
but I suspect that it needs some qualifiers on it.
For what domains?
For what purposes?
I think the answer will vary by context and by timestamp. (But I
Dario,
One message you can send is that they can and should use existing
controlled vocabularies and ontologies to construct the metadata they want
to share. For example, MeSH descriptors would be a good way for them to
organize the 'primary topic' assertions for their articles and would make
it
Slava,
This sounds interesting I would be happy to be a test subject in the bio or
semantic web domain.
(There is no reason this project needs to be reviewed by Wikimedia Research
or anyone else. The data is open, please go ahead and use it to do useful
research like this.) Whether or not any
That would be the ideal spot - if it said what portion of the data was CC0
or is specifically declared as such for the purposes of allowing it into
wikidata.
On Wed, Oct 5, 2016 at 10:40 PM, Federico Leva (Nemo) <nemow...@gmail.com>
wrote:
> Benjamin Good, 05/10/2016 23:33:
>
wrote:
> Benjamin Good, 05/10/2016 19:44:
>
>> As a specific example, we have informal (e.g. an email to us) permission
>> to import data from the Disease Ontology [2] and UniProt [3] but would
>> like to make those informal agreements 'official' and public.
>>
>
&
Thanks Egon, thats a reasonable approach to keep the ball rolling. How do
you plan to handle updates ?
On Wed, Oct 5, 2016 at 12:15 PM, Egon Willighagen <
egon.willigha...@gmail.com> wrote:
> On Wed, Oct 5, 2016 at 7:44 PM, Benjamin Good <ben.mcgee.g...@gmail.com>
> wrote:
&g
On Wed, Aug 17, 2016 at 11:44 AM, Andy Mabbett <a...@pigsonthewing.org.uk>
wrote:
> On 17 August 2016 at 00:43, Benjamin Good <ben.mcgee.g...@gmail.com>
> wrote:
>
> > I am about to propose much more widespread use of Property:P2888 "exact
> > match"
Perhaps it would be more productive if I give a very specific example.
(I'd prefer a general, wikidata-wide policy but it sounds like that isn't
going to happen.)
We are working on integrating wikidata with many of the ontologies that are
part of the OBO Foundry [1]. These include, for example,
I am about to propose much more widespread use of Property:P2888 "exact
match" for linking from a wikidata item to a URI that should resolve to
linked data about the same concept from another semantic web resource.
(amongst the biomedical items that our team works with)
Is there any other pattern
Perhaps the new 'exact match'?
https://m.wikidata.org/wiki/Property:P2888
> On Jun 17, 2016, at 8:26 PM, Thad Guidry wrote:
>
> Oh boy. I thought I had a few things figured out with Wikidata...until
> I read through the Property Talk discussions for
>
>
Hi Tom,
I think the example you have there is actually linked up properly at the
moment?
https://en.wikipedia.org/wiki/SPATA5 is about both the gene and the protein
as are most Wikipedia articles of this nature. And it is linked to the
gene the way we encourage modeling
Hi Julie,
We've thought a lot about this, but not done anything formally yet. There
is an example of this happening to improve the disease ontology presented
in this paper [1].
Mechanically, parties interested in a particular swath of data linked to
their resource could set up repeated SPARQL
+1000 to Markus K's main arguments above. Yes noise will be introduced as
different people come in to make edits. Administrators locking them out
isn't the best way to solve that problem in an open system. There are
other options, as he raised - both technical and social.
Our group maintains
awesome, thank you!
In case its useful for anyone, I was using wikidata to teach biologists and
chemists about knowledge graphs (many tinyurls toward the end)
http://www.slideshare.net/goodb/computing-on-the-shoulders-of-giants
On the second part of the course students were provided with the
I don't want to suggest a technology, but the URL shortener on wdqs has
been EXTREMELY USEFUL to me and it would be major bummer to lose it. I
recently taught a class and used a variety of examples of SPARQL queries
over wikidata. Having that shortener made it much faster for me to
assemble the
I think no one would disagree that the current viewing experience on the
main wikidata.org interface is not ideal. Keep in mind though that this is
fundamentally an impossible UI task. There are many many different ways
that the data about an item in wikidata might be best used/viewed depending
If a group (lets say either mine or Lydia's team) was interested in doing
what I propose, could we request and sign such an NDA? This doesn't seem
insurmountable right ?
On Tue, Apr 26, 2016 at 10:01 AM, Dan Garry <dga...@wikimedia.org> wrote:
> On 26 April 2016 at 08:41, Benj
I'll start with the simple question than give the longer context. Is there
any way to know how many times an item or a claim appears in the results of
a query to query.wikidata.org ? Are there any other ways to quantify
query/application usage of specific wikidata content?
Background. The
James,
This is cool. The integration between Wikipedia and Wikidata is really
smooth and useful. I wonder if you could let users do Wikidata editing
right in context rather than having them have to find the link on the top
there and jump out of the app and into the editor ? There might be
ionality" (though
>>>> "people born since 1600 with no nationality" would have worked
>>>> better), "people with no occupation", "buildings that don't have a
>>>> 'located in' value", etc.
>>>>
>>>> Th
mage than good. Several P1889s would
> be nice.
>
> I think, as someone suggested, that adding P1889 and having Wikibase
> merging looking at P1889 would be a solution.
>
>
> /Finn
>
>
> On 11/10/2015 12:34 AM, Benjamin Good wrote:
>
>> Magnus,
>>
>>
In another thread, we are discussing the preponderance of problematic
merges of gene/protein items. One of the hypotheses raised to explain the
volume and nature of these merges (which are often by fairly inexperienced
editors and/or people that seem to only do merges) was that they were
coming
unmerging
> game. :-)
>
> /Finn
>
>
>
> On 11/10/2015 05:54 PM, Benjamin Good wrote:
>
>> In another thread, we are discussing the preponderance of problematic
>> merges of gene/protein items. One of the hypotheses raised to explain
>> the volume and nature of th
Magnus,
We are seeing more and more of these problematic merges. See:
http://tinyurl.com/ovutz5x for the current list of (today 61) problems.
Are these coming from the wikidata game?
All of the editors performing the merges seem to be new and the edit
patterns seem to match the game. I thought
A section in the talk page associated with the article in question would
seem to solve this (definitely real) problem? - assuming that a would-be
editor was aware of the talk page.
Alternatively, you could propose a generic property with a text field that
could be added to items on an as-needed
I am having the same kinds of 500 problems. Bitbucket is generally
suffering today: http://status.bitbucket.org
On Thu, Oct 22, 2015 at 12:27 PM, Markus Kroetzsch <
markus.kroetz...@tu-dresden.de> wrote:
> On 22.10.2015 19:29, Dario Taraborelli wrote:
>
>> I’m constantly getting 500 errors.
>>
Hi everyone, thanks for the encouragement!
The next technical step is actually a reworking of the template/lua code
for producing that infobox. The current one works for the few genes I've
manually tested it on, but the way it uses arbitrary access (via nesting
structures in the template code)
I've noticed that a number of EN Wikipedia articles are starting to use
wikidata e.g. all of these templates:
https://en.wikipedia.org/wiki/Category:Templates_using_data_from_Wikidata
This is great, and I have already started trying to encourage more of this
in the areas we are working on (genes,
They wrote a really insightful paper about how their processes for
large-scale data curation worked. Among may other things, they
investigated mechanical turk 'micro tasks' versus hourly workers and
generally found the latter to be more cost effective.
The Anatomy of a Large-Scale Human
keep them around,
then just give users a way to filter them out in your API's is all that I
ask. :)
Thad
+ThadGuidry https://www.google.com/+ThadGuidry
On Mon, Jun 15, 2015 at 10:53 AM, Benjamin Good ben.mcgee.g...@gmail.com
wrote:
This is an important question. There are apparently
I recently introduced wikidata to a (very computationally savvy) colleague
by sending him this link:
https://www.wikidata.org/wiki/Q423111
His response is indicative of an interface problem that I think is actually
very important:
Is there a simple way to get the RDF for a given concept? The
We have a hackathon starting tomorrow morning (California time). It would
be fantastic if we could hack on adding our gene wikidata content to a
Wikipedia instance using this new ability. We too have been anxiously
awaiting this development.
Is there a sandbox environment somewhere that we
On Thu, May 7, 2015 at 12:32 PM, Lydia Pintscher
lydia.pintsc...@wikimedia.de wrote:
On Thu, May 7, 2015 at 9:28 PM, Benjamin Good ben.mcgee.g...@gmail.com
wrote:
We have a hackathon starting tomorrow morning (California time). It
would
be fantastic if we could hack on adding our gene
could be generated as a
whole.
-Data could be queried/exported from WD based on an ontology by simply
selecting the whole or parts of an ontology.
This approach has been suggested and discussed by Benjamin Good, Elvira
Mitraka, Andra Wagmeester, Andrew Su and me. As an example, we put
together
,
Have you checked tool-labs already?
http://tools.wmflabs.org/hay/directory/#/keyword/wikidata
Maybe you find something useful...
Best,
Claudia
On Dec 15, 2014, at 12:42 PM, Lydia Pintscher
lydia.pintsc...@wikimedia.de wrote:
On Wed, Dec 10, 2014 at 8:28 PM, Benjamin Good ben.mcgee.g
Are there are any tools currently that could do something like this:
Input: A category/class - something that hangs off subclassOf P279
relationships. For example: gene Q7187
Output: an interactive visual representation of the properties that are
being used by the entities connected to this
Sorry, not sure if this is the right place to post this bug report?
https://www.wikidata.org/wiki/Wikidata:List_of_properties/all#Medicine
reports quite a few messages like:
*The time allocated for running scripts has expired.**The time allocated
for running scripts has expired.**The time
a string, a
coordinate, etc.
I *think* the reason you're getting some results for
CLAIM[486:string] is because the system is somehow matching these up
with the items that have null/unknown values for 486 and returning
those.
Andrew.
On 27 October 2014 03:52, Benjamin Good ben.mcgee.g
.
On 24 October 2014 18:13, Benjamin Good ben.mcgee.g...@gmail.com wrote:
Using the autolist 2 tool (http://tools.wmflabs.org/autolist/index.php),
I
enter the WDQ query:
CLAIM[486:D008180]
which I think means Give me the items that have MeSH id = D008180
. It
does indeed return
Using the autolist 2 tool (http://tools.wmflabs.org/autolist/index.php), I
enter the WDQ query:
CLAIM[486:D008180]
which I think means Give me the items that have MeSH id = D008180 . It
does indeed return an item with that claim Q1495661
but also returns items that have an *no value* as the
, 2014 at 10:54 PM, Benjamin Good ben.mcgee.g...@gmail.com
wrote:
Very cool project! It really shows some great potential for wikidata.
In quickly playing with it I found it difficult to find properties that
matched my intention most of the time. I think this is not the fault of
the tool
Very cool project! It really shows some great potential for wikidata. In
quickly playing with it I found it difficult to find properties that
matched my intention most of the time. I think this is not the fault of
the tool but an indication of an area where wikidata could be improved
(e.g.
Magnus is right on here. We are hoping to begin things by establishing
wikidata as a central hub for gene identifier lookup and matching. (This
is a big a problem in bioinformatics). With this (hopefully fairly bit
stable) starting point of wikidata entities, we will be expanding the
statements
potential
for structuring large sets of biological data. Thanks for your excellent
work!
Cheers,
Eric
https://www.wikidata.org/wiki/User:Emw
On Mon, Oct 6, 2014 at 4:21 PM, Benjamin Good ben.mcgee.g...@gmail.com
wrote:
I thought folks might like to know that every human gene (according
I thought folks might like to know that every human gene (according to the
United States National Center for Biotechnology Information) now has a
representative entity on wikidata. I hope that these are the seeds for
some amazing applications in biology and medicine.
Well done Andra and
wrote:
Wow! That's pretty cool work!
Do you have any plans to keep the data fresh?
On Mon Oct 06 2014 at 1:22:12 PM Benjamin Good ben.mcgee.g...@gmail.com
wrote:
I thought folks might like to know that every human gene (according to
the United States National Center for Biotechnology
Would be happy to. Let me know suggested size and how to get it over to
you.
thanks
-Ben
On Mon, Oct 6, 2014 at 1:40 PM, Lydia Pintscher
lydia.pintsc...@wikimedia.de wrote:
Hey Ben and Andra,
On Mon, Oct 6, 2014 at 10:30 PM, Benjamin Good ben.mcgee.g...@gmail.com
wrote:
Yes indeed we do
Sounds good. Will do.
On Mon, Oct 6, 2014 at 1:53 PM, Lydia Pintscher
lydia.pintsc...@wikimedia.de wrote:
On Mon, Oct 6, 2014 at 10:48 PM, Benjamin Good ben.mcgee.g...@gmail.com
wrote:
Would be happy to. Let me know suggested size and how to get it over to
you.
Sweet. As long as you
Based on the CHEBI ontology perspective, alcohol is a class with subclasses
like 'aromatic alcohol' which has subclasses like 'benzyl alcohols' which
has subclasses like 'methylbenzyl alcohol' and so on.
These relationships seem worth capturing and subclass seems like a
reasonable way to do it.
for the disease identifiers
though of course we would love to have some feedback about that!
-Ben
On Fri, Sep 12, 2014 at 1:28 AM, Joe Filceolaire filceola...@gmail.com
wrote:
Is this being inported into wikidata?
Joe
On 11 Sep 2014 23:16, Benjamin Good ben.mcgee.g...@gmail.com wrote:
FYI we are already
biology.
On 11 September 2014 17:06, Benjamin Good ben.mcgee.g...@gmail.com
wrote:
Thats great Andy! Hope to see you connecting with
https://www.wikidata.org/wiki/Wikidata:WikiProject_Molecular_biology -
specifically with regard to drug information.
cheers
-Ben
On Wed, Sep 10
How are related properties calculated?
Is the definition of a Class something that has a subclass relationship?
Or?
Very cool...
-Ben
On Mon, Sep 8, 2014 at 9:24 AM, Markus Krötzsch
mar...@semantic-mediawiki.org wrote:
On 08.09.2014 14:53, Markus Krötzsch wrote:
...
://www.mediawiki.org/wiki/PWB
Both branches support Wikidata
Best
On 8/29/14, Benjamin Good ben.mcgee.g...@gmail.com wrote:
Which python framework should a new developer use to make a wikidata
editing bot?
thanks
-Ben
It does, but only on the very bottom with a see also.
Somehow I ended up on
https://github.com/jcreus/pywikidata
first.
which is two years out of date and very similarly named..
-ben
On Fri, Aug 29, 2014 at 10:17 AM, Derric Atzrott
datzr...@alizeepathology.com wrote:
There is
Which python framework should a new developer use to make a wikidata
editing bot?
thanks
-Ben
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l
Tom,
I totally agree with your sentiments here. Two questions.
Do you believe there is any valuable use for upper ontologies in the
wikidata system at all at this stage?
Could you describe how you see a bubble-up classification scheme working in
this context in a little detail? I can imagine
OK, thanks for your reply. We will watch for new developments and
incorporate them into our work as they are ready.
Keep up the good work on this important project!
-Ben
On Fri, Sep 13, 2013 at 1:20 PM, Daniel Kinzler daniel.kinz...@wikimedia.de
wrote:
Am 13.09.2013 18:24, schrieb Benjamin
Thanks Lydia, here is a direct link to our wikidata project idea.
http://sulab.org/gsoc13/#idea1
There is clearly already a lot of interest in wikidata among the gsoc
student community. If students (or potential mentors) have additional
ideas for biology-related wikidata projects, please do get
Thanks for the feedback. I think we will push forward and work directly on
wikidata. This is a conversion from database(s)-wikipedia to
database(s)-wikidata-wikipedia
All of our code will be and is open source. We'd be happy to share and to
build on what other bot developers are doing. Is
I am considering the task of converting the templates from the gene
articles in Wikipedia (http://en.wikipedia.org/wiki/Portal:Gene_Wiki) to
use/create wikidata assertions. This involves an extensive update of the
template structure as well as the code for the bot that keeps them in sync
with
64 matches
Mail list logo