Re: [Wikidata-l] weekly summary #156

2015-05-09 Thread Ricordisamoa

Il 03/05/2015 16:13, Lydia Pintscher ha scritto:

Hey folks :)

Here's your summary of what happened around Wikidata over the last week.


  Events https://www.wikidata.org/wiki/Wikidata:Events/Press/Blogs
  https://www.wikidata.org/wiki/Wikidata:Press_coverage

  * Wikidata was presented at a Swedish Linked Data Network Meet-up

https://www.eventbrite.com/e/lankade-data-natverkstraff-swedish-linked-data-network-meet-up-tickets-15689895901
in Gothenburg.
  * The most important Wikipedia pages

http://www.gizmodo.in/news/The-Most-Important-Wikipedia-Pages-From-the-First-Open-Wiki-Ranking/articleshow/47082229.cms


  Other Noteworthy Stuff

  * P107 (main type)

https://www.wikidata.org/w/index.php?title=Property:P107action=editredlink=1
is finally orphaned and deleted, 20 months after deprecation. This
was the first property used 100 times.
  * The royal baby https://www.wikidata.org/wiki/Q18002970 was
quickly updated on Wikidata. (her family tree
https://tools.wmflabs.org/family/ancestors.php?q=Q18002970lang=en)
  * New tool by Magnus: Find pictures on geography.co.uk, upload them
to Commons and add them to Wikidata
https://tools.wmflabs.org/fist/geograph_candidates.html
  * All Van Gogh Museum paintings are now on Commons and Wikidata. Go
go SumOfAllPaintings peeps!
  * List of topics with links in at least X Wikipedias
https://tools.wmflabs.org/wikidata-todo/static/multi_wp/
  * All French Senators matched using MixNMatch
https://tools.wmflabs.org/mix-n-match/?mode=catalog_detailscatalog=44
\o/


  Did you know?

  * Newest properties: represented by
https://www.wikidata.org/wiki/Property:P1875, Netflix Identifier
https://www.wikidata.org/wiki/Property:P1874, maximum number of
players https://www.wikidata.org/wiki/Property:P1873, minimum
number of players https://www.wikidata.org/wiki/Property:P1872,
CERL ID https://www.wikidata.org/wiki/Property:P1871, Name
Assigning Authority Number
https://www.wikidata.org/wiki/Property:P1870, Hall of Valor ID
https://www.wikidata.org/wiki/Property:P1869, ballots cast
https://www.wikidata.org/wiki/Property:P1868, eligible voters
https://www.wikidata.org/wiki/Property:P1867, catholic-hierarchy
diocese ID https://www.wikidata.org/wiki/Property:P1866,
Wikidata example geographic coordinates
https://www.wikidata.org/wiki/Property:P1865, Wikidata example
monolingual text https://www.wikidata.org/wiki/Property:P1864,
Wikidata example property
https://www.wikidata.org/wiki/Property:P1863, Wikidata example
quantity https://www.wikidata.org/wiki/Property:P1862, Wikidata
example time https://www.wikidata.org/wiki/Property:P1861,
Wikidata example URL
https://www.wikidata.org/wiki/Property:P1860, Wikidata example
item value https://www.wikidata.org/wiki/Property:P1859,
Wikidata example string
https://www.wikidata.org/wiki/Property:P1858, Wikidata example
media file https://www.wikidata.org/wiki/Property:P1856,
Wikidata property example
https://www.wikidata.org/wiki/Property:P1855


  Development

  * We rolled out usage tracking on the first two wikis (French
Wikisource and Dutch Wikipedia). Users should not notice anything.
More wikis will follow in the next weeks. This is the remaining
step for enabling arbitrary access on wikis other than Commons.
  * The students team is working hard to get a first release of the
improved constraint reports and checks against 3rd party databases
out.
  * Ricordisamoa fixed the issue with long descriptions being cut off.



Actually Thiemo did the most of it ;)
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Wikidata for Wiktionary

2015-05-08 Thread Ricordisamoa

Il 07/05/2015 14:08, Andy Mabbett ha scritto:

On 7 May 2015 at 11:57, Ricordisamoa ricordisa...@openmailbox.org wrote:


Let's focus on Commons, OpenStreetMap, queries, arbitrary access, new
datatypes?

OSM in what context?


Adding mutual links, keeping them up to date, building applications that 
use both databases, etc.

https://wiki.openstreetmap.org/wiki/Wikidata



Also, we should throw WikiSpecies into the mix.



This reminds me of some old discussions... [1] 
https://www.wikidata.org/wiki/Wikidata:Project_chat/Archive/2013/02#Include_Wikispecies_into_Wikidata 
[2] 
https://www.wikidata.org/wiki/Wikidata:Project_chat/Archive/2013/04#Wikispecies 
[3] 
https://www.wikidata.org/wiki/Wikidata:Project_chat/Archive/2014/02#Wikispecies 
etc.
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Wikidata for Wiktionary

2015-05-07 Thread Ricordisamoa

Hi Denny,
I would strongly advise against connecting Wiktionary to Wikidata in the 
status quo, mainly for the reasons Gerard summarized.
While wikt's 'data model' probably makes sense for a spelling-based 
dictionary, it does not for a concept-based knowledge base like ours.
Even turning Wiktionary into an OmegaWiki 
https://meta.wikimedia.org/wiki/OmegaWiki-like project seems unlikely 
feasible without an intermediate step.
Let's focus on Commons, OpenStreetMap, queries, arbitrary access, new 
datatypes?


Il 07/05/2015 04:54, Denny Vrandečić ha scritto:
It is rather clear that everyone wants Wikidata to also support 
Wiktionary, and there have been plenty of proposals in the last few 
years. I think that the latest proposals are sufficiently similar to 
go for the next step: a break down of the tasks needed to get this done.


Currently, the idea of having Wikidata supporting Wiktionary is 
stalled because it is regarded as a large monolithic task, and as such 
it is hard to plan and commit to. I tried to come up with a task 
break-down, and discussed it with Lydia and Daniel, and now, as said 
in the last office hour, here it is for discussion and community input.


https://www.wikidata.org/wiki/Wikidata:Wiktionary/Development/Proposals/2015-05 



I think it would be really awesome if we would start moving in this 
direction. Wiktionary supported by Wikidata could quickly become one 
of the crucial pieces of infrastructure for the Web as a whole, but in 
particular for Wikipedia and its future development.


Cheers,
Denny
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Wikidata for Wiktionary

2015-05-07 Thread Ricordisamoa

Il 07/05/2015 16:03, Daniel Kinzler ha scritto:

Am 07.05.2015 um 14:56 schrieb Yair Rand:

Task 1 as described on the proposal page isn't completely clear on how it would
work. Would the generated items have Q-ids? Would it be possible to link
Wiktionary entries to non-Wiktionary pages in the very rare situations that make
sense (articles on particular series of (not-language-associated)
symbols/characters)?

Task 1 (Interlanguage-Links for Wiktionary) would not involve Wikidata or
Wikibase at all. It would be a standalone extension linking pages with identical
names between wikis.


It's ok then! I have been thinking about something like that for some 
time...


___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Making queries on Wikibase

2015-04-29 Thread Ricordisamoa

Il 29/04/2015 20:56, Luca Martinelli ha scritto:

Dear all,

I need to know about the possibility of making queries on a Wikibase
instance. I think it is possible to make queries on data on a
particular instance only with external tools at the moment, right?

Thanks for the answer. :)



Wikidata-Query-Service 
https://phabricator.wikimedia.org/project/profile/891/
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Data templates

2015-04-18 Thread Ricordisamoa

Il 11/04/2015 13:29, Antoine Isaac ha scritto:

Hi,

Is the 'template' word so bad? Paraphrasing Daniel's definition of the 
MediaWiki template, one could see a 'WikiData template' as
a set of of properties that can be re-used, e.g. to make create 
statements about a certain class. (the 'parameter' bit could be 
understood as adding or removing properties from the templates, e.g. 
using twice a property or adding a new one when it's needed).


What we're after seems to exist already, described as 'item structure':
http://www.wikidata.org/wiki/Wikidata:WikiProject_Visual_arts/Item_structure 


Or 'list of properties':
https://www.wikidata.org/wiki/Wikidata:List_of_properties/Works

'schema convention' matches the idea, but the wording may be too 
abstract. I come from a community that calls such things 'description 
set profiles'; such expressions have a hard time being adopted in less 
technical communities...



About the text values. A big +1 to Daniel at not trying to represent 
semi-structured text, which is meant to piggyback structured data in 
legacy systems that can't handle it. The matter is rather the 
availability in Wikidata of text-like summaries like the 
dbpedia-owl:abstract at http://dbpedia.org/page/Castle . Having things 
like this together with the Wikidata data would be great for 
data-reusers like us, instead of having to fetch it from elsewhere!


There's TextExtracts 
https://www.mediawiki.org/wiki/Extension:TextExtracts for that: 
example 
https://en.wikipedia.org/w/api.php?action=queryprop=extractsexsentences=1explaintext=1titles=Castle




Antoine  ---
Antoine Isaac
RD Manager, Europeana.eu
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Data templates

2015-04-08 Thread Ricordisamoa

Il 04/04/2015 23:45, Stas Malyshev ha scritto:

Hi!


For things that actually *are* free text, and not terribly long, a monolongual
(or, in the future, multilingual) text property could be used. quote already
exists, abstract could be added, pending community discussion. Length
limitations can be adjusted if need be.

Maybe if the need of bigger texts arises we can have separate field
type? Right now the storage model is not very good for storing texts of
non-negligible sizes, especially multilingual ones (x800 languages).
OTOH, we have a type that allows us to use multimedia by integrating
with Commons. So maybe the same idea with using some other wiki -
quotes? sources? for bigger text snippets would work too? Just
brainstorming here :)


spamStructured Wikiquote 
https://meta.wikimedia.org/wiki/Structured_Wikiquote/spam





What I was warning against is continuing the misuse of text fields for
semi-structured or even fully structured data that I have often seen in GLAM
meta-data. That kind of thing should not be copied to Wikidata.

Right. I think it may be useful here to understand which kinds of text
we're talking about which can't be structured but are big enough to
cause concern. I.e. if it's quotes - we already have wikiquote, right? Etc.



___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Wikidata periodic table

2015-04-07 Thread Ricordisamoa
Forgot: the code is formally under review here 
https://gerrit.wikimedia.org/r/202610/. GPL-3.0+ as the old WikiPeriod.


Il 08/04/2015 01:18, Ricordisamoa ha scritto:
I'd like to announce a new Labs tool to show a periodic table 
https://tools.wmflabs.org/ptable/.
It is based on WikiPeriod's PHP code (in turn ported from JavaScript) 
and features several improvements:


  * 'tiles' are wider and taller;
  * most of them are now provided with a background color (the same as
Wikipedia's

https://en.wikipedia.org/wiki/Periodic_table#Periodic_table_legend_for_category)
based on the elements' subclass of property
https://www.wikidata.org/wiki/Property:P279 (the same that
powers period/group detection);
  * for labels, Wikidata's built-in language fallback is used instead
of just falling back to English;
  * a public JSON API https://tools.wmflabs.org/ptable/api is
available for everyone!

And some more under the hood:

  * rewritten in Python with Jinja2:
  o more object-oriented
  o presentation is split from actual logic
  o less vulnerable to XSS attacks
  * a LRU (least recently used) cache with a maximum TTL (per-item
time-to-live) value of 6 hours is used to avoid hitting data
sources on every request;
  * both the Wikidata API and Wikidata Query can be used
interchangeably as sources.

I had to create some items such as Q19753344 
https://www.wikidata.org/wiki/Q19753344 and Q19753345 
https://www.wikidata.org/wiki/Q19753345 to properly categorize 
elements. My knowledge of chemistry is limited, so please report/fix 
every mistake you can find ;-)


Future plans include:

  * oxidation state
  * images
  * responsive design
  * alternative table structures

___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Weekly Summary #152

2015-04-07 Thread Ricordisamoa

Il 05/04/2015 17:18, John Lewis ha scritto:

Hi everyone,

Here is the latest update of everything going on around Wikidata!


  Discussions

  * Closed RfC: Reforming administrator inactivity criteria

https://www.wikidata.org/wiki/Wikidata:Requests_for_comment/Reforming_administrator_inactivity_criteria


  Events https://www.wikidata.org/wiki/Wikidata:Events/Press/Blogs
  https://www.wikidata.org/wiki/Wikidata:Press_coverage

  * WikiArabia
http://arabia.wikimedia.tn/index.php?title=Main_Page takes place
in Monastir, Tunisia, 3-5 April
  * The GLAM-WIKI 2015 http://www.glamwiki.nl/ conference in The
Hague (10-12 April) features several presentations and tutorials
about Wikidata for/with cultural institutions.
  * The Library world will use Wikidata

http://ultimategerardm.blogspot.nl/2015/04/viaf-move-over-wikipedia-for-wikidata.html
 to
link its information to any and all Wikipedias. No longer English
only, but every Wikipedia will be exposed in this way.
  * Freebase, SEO and Wikidata
http://infobib.de/2015/04/01/seo-freebase-und-wikidata/



Hmm... German?


  * Office hour on IRC covering overall status/development, Freebase
and admin inactivity criteria RfC. You can read the log

https://tools.wmflabs.org/meetbot/wikimedia-office/2015/wikimedia-office.2015-03-31-16.00.log.html.


  Other Noteworthy Stuff

  * Magnus wrote a short tour of Wikidata's tool ecosystem
https://tools.wmflabs.org/wikidata-todo/tour.html#slide.3D0.
  * A first version of the Primary Sources Tool has been released
https://lists.wikimedia.org/pipermail/wikidata-l/2015-April/005724.html.
It'll help with migrating Freebase data and more.
  * Italian Wikipedia's quality festival

https://it.wikipedia.org/wiki/Wikipedia:Festival_della_qualit%C3%A0/Aprile_2015
 is
focusing on interwiki links and Wikidata this month. Help them out?
  * Lots of new databases have been added to Mix n Match
https://tools.wmflabs.org/mix-n-match/.
  * Screenshots of the current state of new constraint reports and
checks against 3rd party databases have been posted.
https://twitter.com/wikidata/status/582545229884063744


  Did you know?

  * Newest properties: choreographer
https://www.wikidata.org/wiki/Property:P1809, senat.fr ID
https://www.wikidata.org/wiki/Property:P1808, Great Aragonese
Encyclopedia ID https://www.wikidata.org/wiki/Property:P1807


  Development

  * Wikidata development started 3 years ago
https://twitter.com/wikidata/status/583220841711845376. 3 to
everyone who is a part of it.
  * Went through all the feedback we got for improving watchlist
integration on Wikipedia and co and posted our assesment

https://www.wikidata.org/wiki/Wikidata:Watchlist_integration_improvement_input#First_iteration_on_feedback_by_dev_team
  * Put the infrastructure for creating Turtle
https://en.wikipedia.org/wiki/Turtle_%28syntax%29-Beta dumps in
place. All new Wikidata dumps will be in
https://dumps.wikimedia.org/wikidatawiki/entities/ from Monday on
(the old * directory will be kept around and receive new json
dumps for backwards compatibility).
  * Reduced size of entities pages by removing no longer needed data
(to make the UI faster).
  * Fixed bug that sometimes caused dates and other types of values to
be cut short when quickly saving. (phabricator:T92831
https://phabricator.wikimedia.org/T92831)
  * Fixed issues with setting focus after clicking edit.

You can see all open bugs related to Wikidata here 
https://phabricator.wikimedia.org/maniphest/query/4RotIcw5oINo/#R.



  Monthly Tasks

  * Hack on one of these
https://phabricator.wikimedia.org/maniphest/query/R8GRzX1eH0tb/#R.
  * Help fix these items
https://www.wikidata.org/wiki/Wikidata:The_Game/Flagged_items which
have been flagged using Wikidata - The Game.
  * Help develop the next summary here!
https://www.wikidata.org/wiki/Wikidata:Status_updates/Next
  * Contribute to a Showcase item
https://www.wikidata.org/wiki/Wikidata:Showcase_items
  * Help translate
https://www.wikidata.org/wiki/Special:LanguageStats or proofread
pages in your own language!



___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Data templates

2015-04-04 Thread Ricordisamoa

Il 02/04/2015 13:51, Daniel Kinzler ha scritto:

Am 02.04.2015 um 09:03 schrieb Valentine Charles:

Regarding the dimensions, it is great to know that it is on your plate. I was
wondering is there a place where we can see the classes/properties that are in
the pipeline and participate to discussions around them?

Note however that for dimensions, the issue is not creating the property, but
teaching the software about units, so that such a property would make sense. A
*lot* of properties are waiting for unit support: length, height, speed,
distance, and many more are blocked on units.



See Wikidata:Property proposal/Pending/2 
https://www.wikidata.org/wiki/Wikidata:Property_proposal/Pending/2 for 
a list.
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] 3 years!

2015-04-01 Thread Ricordisamoa

Il 01/04/2015 12:46, Lydia Pintscher ha scritto:

Hey folks :)

3 years ago we started the development on Wikidata. I'm amazed at how
far we've come together in those 3 years. 3 to everyone who's a part
of this and helped us get to where we are. Looking forward to what we
can achieve in the next 3 ;-)


Cheers
Lydia

    {{}}      
    {{}}      
    {{}}      
    {{}}      
    {{}}      
    {{}}      
    {{}}      
    {{}}      
    {{}}      
    {{}}      
    {{}}      
    {{}}      
    {{}}      
    {{}}      
__  __ _  _  __ _  _   ___
\ \/ /|_   _|| |/ /|_   _||  __ \ /\  |__ __|  /\
 \ \  /\  / /   | |  | ' /   | |  | |  | |   /  \| | /  \
  \ \/  \/ /| |  |  | |  | |  | |  / /\ \   | |   / /\ \
   \  /\  /_| |_ | . \  _| |_ | |__| | /  \  | |  /  \
\/  \/|_||_|\_\|_||_/ /_/\_\ |_| /_/\_\
  \\   // \\   // \\   // \\   //
   \\   @@@   //   \\  //   \\   ééé //   \\   %%%   //
\\ @ // \\  // \\ è // \\ % //
 \\ @@@ //   \\  //   \\ ééé //   \\ %%% //

___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Initial release of the primary sources tool

2015-04-01 Thread Ricordisamoa

Il 01/04/2015 21:30, Denny Vrandečić ha scritto:

Even better are pull requests!


Unfortunately, Google requires anyone to sign a Contributor License 
Agreement.


___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Kian: The first neural network to serve Wikidata

2015-03-19 Thread Ricordisamoa

Il 20/03/2015 01:11, Amir Ladsgroup ha scritto:

OK, I have some news:
1- Today I rewrote some parts of Kian and now it automatically chooses 
regulation parameter (lambda), thus predictions are more accurate. I 
wanted to push changes to the github but It seems my ssh has issues. 
It'll be there soon
2- (Important) I wrote a code that can find possible mistakes in 
Wikidata based on Kian. The code will be in github soon. Check out 
this link http://tools.wmflabs.org/dexbot/possible_mistakes_fr.txt. 
It's result from comparing French Wikipedia against Wikidata e.g. this 
line:

Q2994923: 1 (d), 0.257480420229 (w) [0, 0, 1, 2, 0]
1 (d) means Wikidata thinks it's a human

0.25... (w) means French Wikipedia thinks it's not a human (with 74.3% 
certainty)


And if you check the link you can see it's a mistake in Wikidata. 
Please check other results and fix them.


Tell me if you want this test to be ran from another language too.

3- I used Kian to import unconnected pages from French Wikipedia and 
created about 1900 items. The result is here 
http://tools.wmflabs.org/dexbot/kian_res_fr.txt and please check if 
anything in this list is not human and tell me and I run some error 
analysis.



Best

Great!
Unfortunately, some files seem to have been published with the wrong 
character encoding.

E.g. the first name shows up as Échécrate in my browsers.
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Kian: The first neural network to serve Wikidata

2015-03-19 Thread Ricordisamoa

Il 20/03/2015 01:11, Amir Ladsgroup ha scritto:

OK, I have some news:
1- Today I rewrote some parts of Kian and now it automatically chooses 
regulation parameter (lambda), thus predictions are more accurate. I 
wanted to push changes to the github but It seems my ssh has issues. 
It'll be there soon
2- (Important) I wrote a code that can find possible mistakes in 
Wikidata based on Kian. The code will be in github soon. Check out 
this link http://tools.wmflabs.org/dexbot/possible_mistakes_fr.txt. 
It's result from comparing French Wikipedia against Wikidata e.g. this 
line:

Q2994923: 1 (d), 0.257480420229 (w) [0, 0, 1, 2, 0]
1 (d) means Wikidata thinks it's a human

0.25... (w) means French Wikipedia thinks it's not a human (with 74.3% 
certainty)


And if you check the link you can see it's a mistake in Wikidata. 
Please check other results and fix them.


Tell me if you want this test to be ran from another language too.

3- I used Kian to import unconnected pages from French Wikipedia and 
created about 1900 items. The result is here 
http://tools.wmflabs.org/dexbot/kian_res_fr.txt and please check if 
anything in this list is not human and tell me and I run some error 
analysis.



Best
The data are based on dumps, aren't they? Wikidata hasn't been thinking 
Q73823 https://www.wikidata.org/wiki/Q73823 is a human since 21 Feb.
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] mapping template parameters using Wikidata?

2015-03-18 Thread Ricordisamoa

Il 04/03/2015 18:00, Ricordisamoa ha scritto:
That's what Translatemplate 
https://tools.wmflabs.org/translatemplate/ is for! (thanks to you 
and Daniel for the idea :-)
It uses mwparserfromhell to parse DBpedia mappings and 'translate' 
templates in-place.

I've added you to the service group so you can fiddle with it.
The code got a bit hackish to work around a mwparserfromhell/Labs bug. 
If you're happy with the result we can publish it in Gerrit. 

And the new version is live! You can test it right now in your browser.
While it is not meant for production yet, I am very happy with the 
result. Amir, if you can come up with a better name I'll go straight to 
the Gerrit repository requests :-)
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Names, Aliases, Copyright (and a little OpenStreetMap)

2015-03-12 Thread Ricordisamoa
I completely forgot we already had the excellent transliteration gadget 
https://www.wikidata.org/wiki/MediaWiki:Gadget-SimpleTransliterate.js 
by Ebraminio https://www.wikidata.org/wiki/User:Ebraminio.
Just made a rough patch 
https://www.wikidata.org/w/index.php?title=MediaWiki:Gadget-SimpleTransliterate.jsdiff=20370oldid=155995810 
to make it work with the new UI. Enjoy!


Il 12/03/2015 11:24, Daniel Kinzler ha scritto:

Am 12.03.2015 um 10:03 schrieb Gerard Meijssen:

Hoi,
What would you do with the many, many Chinese place names in Wikidata where we
have nothing but Chinese ? It is completely useless to me in this way. A good
transliterations works for me. Like most people beyond that I do not care much
about it being official or sourced.

Decent automatic translitteration is fine I think. Automatic word-for-word
*translation* however seems rather problematic.


___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Kian: The first neural network to serve Wikidata

2015-03-08 Thread Ricordisamoa

Sounds promising! It'd be good to have the code publicly viewable.

Il 07/03/2015 13:24, Amir Ladsgroup ha scritto:

Hey,
I spent last few weeks working on this lights off [1] and now it's 
ready to work!


Kian is a three-layered neural network with flexible number of inputs 
and outputs. So if we can parametrize a job, we can teach him easily 
and get the job done.


For example and as the first job. We want to add P31:5 (human) to 
items of Wikidata based on categories of articles in Wikipedia. The 
only thing we need to is get list of items with P31:5 and list of 
items of not-humans (P31 exists but not 5 in it). then get list of 
category links in any wiki we want[2] and at last we feed these files 
to Kian and let him learn. Afterwards if we give Kian other articles 
and their categories, he classifies them as human, not human, or 
failed to determine. As test I gave him categories of ckb wiki (a 
small wiki) and worked pretty well and now I'm creating the training 
set from German Wikipedia and the next step will be English Wikipedia. 
Number of P31:5 will drastically increase this week.


I would love comments or ideas for tasks that Kian can do.


[1]: Because I love surprises
[2]: select pp_value, cl_to from page_props join categorylinks on 
pp_page = cl_from where pp_propname = 'wikibase_item';

Best
--
Amir



___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] mapping template parameters using Wikidata?

2015-03-04 Thread Ricordisamoa

Il 03/03/2015 21:40, Amir E. Aharoni ha scritto:
Thanks, that's a step forward. Now the question is how to bring this 
all together.


The context that interests me the most is translating an article in 
ContentTranslation. Let's go with an architect.[1] I am translating an 
article about an architect from English to Dutch, and it has {{Infobox 
architect}} at the top. How would ContentTranslation, a MediaWiki 
extension installed on the Wikimedia cluster, know that the name 
parameter is naam in Dutch?


Currently, in theory, it would:
1. Find that there's a corresponding infobox in Dutch using the 
interlanguage link: 
https://nl.wikipedia.org/wiki/Sjabloon:Infobox_architect
2. Go to dbpedia and find the English template: 
http://mappings.dbpedia.org/index.php/Mapping_en:Infobox_architect

3. Find that name is foaf:name
4. Go to dbpedia and find the Dutch template: 
http://mappings.dbpedia.org/index.php/Mapping_nl:Infobox_architect

5. Find that foaf:name is naam

... And then repeat steps 1 to 5 for each parameter.

Is this something that is possible to query now? (I'm not even talking 
about performance.)


That's what Translatemplate https://tools.wmflabs.org/translatemplate/ 
is for! (thanks to you and Daniel for the idea :-)
It uses mwparserfromhell to parse DBpedia mappings and 'translate' 
templates in-place.

I've added you to the service group so you can fiddle with it.
The code got a bit hackish to work around a mwparserfromhell/Labs bug. 
If you're happy with the result we can publish it in Gerrit.




Even if it is possible to query it, is it good to be dependent on an 
external website for this? Maybe it makes sense to import the data 
from dbpedia to Wikidata? It's absolutely not a rhetorical question - 
maybe it is OK to use dbpedia.


[1] {{Infobox cricketer}} exists in the Dutch Wikipedia, but doesn' 
appear in the Dutch mappings in dbpedia.



--
Amir Elisha Aharoni · אָמִיר אֱלִישָׁע אַהֲרוֹנִי
http://aharoni.wordpress.com
‪“We're living in pieces,
I want to live in peace.” – T. Moore‬

2015-03-03 20:39 GMT+02:00 Daniel Kinzler daniel.kinz...@wikimedia.de 
mailto:daniel.kinz...@wikimedia.de:


Am 03.03.2015 um 18:48 schrieb Amir E. Aharoni:
 Trying again... It's a really important topic for me.

 How do I go about proposing storing information about templates
parameters
 mapping to the community? I kinda understand how Wikidata works,
and it sounds
 like something that could be implemented using the current
properties, but
 thoughts about moving this forward would be very welcome.

Hi Amir!

We had a call today with the dbPedia folks, about exactly this topic!

The dbPedia mapping wiki[1] has this information, at least to some
extent. Let's
say you are looking at {{Cricketer Infobox}} on en. You can look
out the DBPedia
mappings for the template parameters on their mapping page[2].
There you can see
that the country parameter maps to the country proeprty in the
dbpedia
ontology[2], which in turn uses owl:equivalentProperty to
cross-link P17[4].

I assume this info is also available in machine readable form
somewhere, but I
don't know where offhand.

Today we discussed that this mapping should also be available in
the opposite
direction: on Wikidata, you can use P1628 (equivalent property) to
cross-reference the dbPedia ontology. I just added this info to
the country
property.

let me know if this helps :)
-- daniel

[1] http://mappings.dbpedia.org/index.php/
[2] http://mappings.dbpedia.org/index.php/Mapping_en:Cricketer_Infobox
[3] http://mappings.dbpedia.org/index.php/OntologyProperty:Country
[4] https://www.wikidata.org/wiki/Property:P17

--
Daniel Kinzler
Senior Software Developer

Wikimedia Deutschland
Gesellschaft zur Förderung Freien Wissens e.V.

___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org mailto:Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l




___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


[Wikidata-l] Call for development openness

2015-02-17 Thread Ricordisamoa

Hi.
I recently started following mediawiki/extensions/Wikibase on Gerrit, 
and quite astonishingly found that nearly all of the 100 most recently 
updated changes appear to be owned by WMDE employees (exceptions being 
one change by Legoktm and some from L10n-bot). This is not the case, for 
example, with mediawiki/core.
While this may be desired by the Wikidata team for corporate reasons, I 
feel that encouraging code review by volunteers would empower both 
Wikidata and third-party communities with new ways of contributing to 
the project and raise awareness of the development team's goals in the 
long term.
The messy naming conventions play a role too, i.e. Extension:Wikibase 
https://www.mediawiki.org/w/index.php?title=Extension:Wikibaseredirect=no 
is supposed to host technical documentation but instead redirects to the 
Wikibase https://www.mediawiki.org/wiki/Wikibase portal, with actual 
documentation split into Extension:Wikibase Repository 
https://www.mediawiki.org/wiki/Extension:Wikibase_Repository and 
Extension:Wikibase Client 
https://www.mediawiki.org/wiki/Extension:Wikibase_Client, apparently 
ignoring the fact that the code is actually developed in a single 
repository (correct me if I'm wrong). Just to add some more confusion, 
there's also Extension:Wikidata build 
https://www.mediawiki.org/wiki/Extension:Wikidata_build with no 
documentation.
And what about wmde on GitHub https://github.com/wmde with countless 
creatively-named repos? They make life even harder for potential 
contributors.
Finally, the ever-changing client-side APIs make gadgets development a 
pain in the ass.

Sorry if this sounds like a slap in the face, but it had to be said.
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Call for development openness

2015-02-17 Thread Ricordisamoa

Il 17/02/2015 12:53, Lydia Pintscher ha scritto:

On Tue, Feb 17, 2015 at 12:43 PM, Ricordisamoa
ricordisa...@openmailbox.org wrote:

Hi.
I recently started following mediawiki/extensions/Wikibase on Gerrit, and
quite astonishingly found that nearly all of the 100 most recently updated
changes appear to be owned by WMDE employees (exceptions being one change by
Legoktm and some from L10n-bot). This is not the case, for example, with
mediawiki/core.
While this may be desired by the Wikidata team for corporate reasons, I feel
that encouraging code review by volunteers would empower both Wikidata and
third-party communities with new ways of contributing to the project and
raise awareness of the development team's goals in the long term.

How would you like to see us encourage this more? It is nothing we
actively do not want of course.
Using a single code review system and a simpler repository structure 
will indirectly encourage them.
I'm now seeing Wikibase/Programmer's guide to Wikibase 
https://www.mediawiki.org/wiki/Wikibase/Programmer%27s_guide_to_Wikibase, 
which seems fairly detailed but partly duplicates the Gerrit help pages.



The messy naming conventions play a role too, i.e. Extension:Wikibase is
supposed to host technical documentation but instead redirects to the
Wikibase portal, with actual documentation split into Extension:Wikibase
Repository and Extension:Wikibase Client, apparently ignoring the fact that
the code is actually developed in a single repository (correct me if I'm
wrong). Just to add some more confusion, there's also Extension:Wikidata
build with no documentation.

There are different repositories. They just get merged into one for deployment.
Really? AFAICS development occurs on mediawiki/extensions/Wikibase 
https://git.wikimedia.org/summary/?r=mediawiki/extensions/Wikibase.git 
and on GitHub.
mediawiki/extensions/WikibaseRepository 
https://git.wikimedia.org/summary/?r=mediawiki/extensions/WikibaseRepository.git 
and mediawiki/extensions/WikibaseClient 
https://git.wikimedia.org/summary/?r=mediawiki/extensions/WikibaseClient.git 
also exist but have always been empty.
Even mediawiki/extensions/WikibaseRepo 
https://git.wikimedia.org/summary/?r=mediawiki/extensions/WikibaseRepo.git 
appears to exist according to Gitblit, but not according to Gerrit nor 
GitHub...



And what about wmde on GitHub with countless creatively-named repos? They
make life even harder for potential contributors.

Agreed. Something we want to tackle.
Out of curiosity, was GitHub chosen because it fitted with your 
workflow? Will you embrace Differential when it comes?



Finally, the ever-changing client-side APIs make gadgets development a pain
in the ass.

Agreed but as I said this is going to be painful for a little longer
until we have done the UI redesign. After that I want it to be more
stable again obviously.

Thanks. Is there a task/page where progress is tracked?



Cheers
Lydia

___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Call for development openness

2015-02-17 Thread Ricordisamoa

Il 17/02/2015 13:33, Ricordisamoa ha scritto:

Il 17/02/2015 12:53, Lydia Pintscher ha scritto:

On Tue, Feb 17, 2015 at 12:43 PM, Ricordisamoa
ricordisa...@openmailbox.org  wrote:

Hi.
I recently started following mediawiki/extensions/Wikibase on Gerrit, and
quite astonishingly found that nearly all of the 100 most recently updated
changes appear to be owned by WMDE employees (exceptions being one change by
Legoktm and some from L10n-bot). This is not the case, for example, with
mediawiki/core.
While this may be desired by the Wikidata team for corporate reasons, I feel
that encouraging code review by volunteers would empower both Wikidata and
third-party communities with new ways of contributing to the project and
raise awareness of the development team's goals in the long term.

How would you like to see us encourage this more? It is nothing we
actively do not want of course.
Using a single code review system and a simpler repository structure 
will indirectly encourage them.
I'm now seeing Wikibase/Programmer's guide to Wikibase 
https://www.mediawiki.org/wiki/Wikibase/Programmer%27s_guide_to_Wikibase, 
which seems fairly detailed but partly duplicates the Gerrit help pages.

The messy naming conventions play a role too, i.e. Extension:Wikibase is
supposed to host technical documentation but instead redirects to the
Wikibase portal, with actual documentation split into Extension:Wikibase
Repository and Extension:Wikibase Client, apparently ignoring the fact that
the code is actually developed in a single repository (correct me if I'm
wrong). Just to add some more confusion, there's also Extension:Wikidata
build with no documentation.

There are different repositories. They just get merged into one for deployment.
Really? AFAICS development occurs on mediawiki/extensions/Wikibase 
https://git.wikimedia.org/summary/?r=mediawiki/extensions/Wikibase.git 
and on GitHub.
mediawiki/extensions/WikibaseRepository 
https://git.wikimedia.org/summary/?r=mediawiki/extensions/WikibaseRepository.git 
and mediawiki/extensions/WikibaseClient 
https://git.wikimedia.org/summary/?r=mediawiki/extensions/WikibaseClient.git 
also exist but have always been empty.
Even mediawiki/extensions/WikibaseRepo 
https://git.wikimedia.org/summary/?r=mediawiki/extensions/WikibaseRepo.git 
appears to exist according to Gitblit, but not according to Gerrit nor 
GitHub...
Upd: I found https://github.com/wmde/WikibaseRepository, 
https://github.com/wmde/WikibaseClient and 
https://github.com/wmde/WikibaseLib, but they're marked as experimental 
splits and have no commits since Oct 2014, so I suppose they're dead.

And what about wmde on GitHub with countless creatively-named repos? They
make life even harder for potential contributors.

Agreed. Something we want to tackle.
Out of curiosity, was GitHub chosen because it fitted with your 
workflow? Will you embrace Differential when it comes?

Finally, the ever-changing client-side APIs make gadgets development a pain
in the ass.

Agreed but as I said this is going to be painful for a little longer
until we have done the UI redesign. After that I want it to be more
stable again obviously.

Thanks. Is there a task/page where progress is tracked?


Cheers
Lydia

___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] annotating red links

2015-02-12 Thread Ricordisamoa

Il 11/02/2015 22:08, Amir E. Aharoni ha scritto:
2015-02-11 22:14 GMT+02:00 Ricordisamoa ricordisa...@openmailbox.org 
mailto:ricordisa...@openmailbox.org:

 Adding non-existing pages to Wikidata items?
 Using a syntax like [Q42[notexistingpagetitle]]?

Is this a suggestion for possible syntax or something that actually 
works somewhere?
It is only a humble attempt at designing a syntax that should be 
familiar with most wikitexters but shouldn't break the existing content 
corpus :-)


But yeah, something like this - something that includes the title of a 
page that doesn't exist in this wiki, but may some to exist some day, 
and the Q number.


--
Amir Elisha Aharoni · אָמִיר אֱלִישָׁע אַהֲרוֹנִי
http://aharoni.wordpress.com
‪“We're living in pieces,
I want to live in peace.” – T. Moore‬

___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] annotating red links

2015-02-12 Thread Ricordisamoa

Il 12/02/2015 11:23, Amir E. Aharoni ha scritto:
 The other is to extend the link syntax similar to image syntax, for 
example
 with [[Article Name|Alternate Text|wd=Q1234]]. This should be 
minimally disruptive

 to the editors.

Yes - this would be more or less perfect, but it would require changes 
in core MediaWiki. If nothing else works, then it's possible, but 
seems harder to get through in practice.



Wouldn't it be feasible by means of a hook?


--
Amir Elisha Aharoni · אָמִיר אֱלִישָׁע אַהֲרוֹנִי
http://aharoni.wordpress.com
‪“We're living in pieces,
I want to live in peace.” – T. Moore‬



___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] annotating red links

2015-02-11 Thread Ricordisamoa

Il 11/02/2015 20:26, Amir E. Aharoni ha scritto:

Hi,

TL;DR: How can a red link be annotated in a semantic way with a 
foreign article title or a Wikidata Q item number?


Imagine: I'm writing a Wikipedia article in Russian. There's a red 
link in it. I don't have time to write the target article for that 
link now, but I'm sure that it should exist. In fact, that article 
does exist in the English Wikipedia.


I want the link to be red (fr the usual wiki reasons), but until the 
Russian article is written, I want to give the software a hint about 
which topic it is supposed to be about. Telling it the English article 
name would be one way to do it. Giving it the Wikidata Q item number 
would be an even better way to do it.


Unfortunately, MediaWiki does not currently have true syntax to do 
either. (Correct me if I'm wrong.)


Some Wikipedias may have templates that do something like this (e.g. 
Russian: https://ru.wikipedia.org/wiki/Template:En ). But there's 
nothing that is uniform to all projects.


*Why* is it useful to give the software this hint in the first place? 
Most simplistically, it's useful to the reader - in case that reader 
knows English, she can at least read something.


But there's something bigger. When the ContentTranslation extension 
translates links, it automatically adapts links that can be found. 
What to do about those that can't be auto-adapted? It frequently 
happens when Wikipedians translate articles that many links in the 
created articles turn out to be red. We'd love to get 
ContentTranslation to help the translators make those articles by 
writing relevant articles with as few clicks as possible, and that is 
only possible by annotating the red links with the topics to which 
they belong.


So, any ideas?
What do other Wikipedias for such annotation?
Is it imaginable to add wiki syntax for such a thing?
Can anybody think of a hack that reuses the current [[link]] syntax to 
add such annotation?


--
Amir Elisha Aharoni · אָמִיר אֱלִישָׁע אַהֲרוֹנִי
http://aharoni.wordpress.com
‪“We're living in pieces,
I want to live in peace.” – T. Moore‬


Adding non-existing pages to Wikidata items?
Using a syntax like [Q42[notexistingpagetitle]]?

___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Freebase's incompatible types and Property description permissions

2015-01-08 Thread Ricordisamoa

Il 08/01/2015 20:37, Thad Guidry ha scritto:


On Thu, Jan 8, 2015 at 12:17 PM, Federico Leva (Nemo) 
nemow...@gmail.com mailto:nemow...@gmail.com wrote:


Thad Guidry, 08/01/2015 18:58:

Unless the P17 Country property had an expanded definition of
sovereign
state (or originating sovereign state) of this item


That's more like P27. Both are rather flexible though, see their
talk pages.
https://www.wikidata.org/wiki/Property_talk:P17


1. How does Wikidata want to handle Property / Statement rule
enforcement and Freebase's incompatible types ?


I'm not sure how this is an example of incompatible type, it
sounds more like a type Freebase didn't have. Handling such
differences is possible by tweaking property descriptions and
adding constraints. P17 is already declared incompatible with
instance of: human. If you make music band a subclass of
human, then this statement about U2 will be reported by bots as
a constraint violation.


Right, Freebase would not stick a Property called Country right on 
an instance of a Music Band.  We would put Country under the Musical 
Group type, and give it a better definition like The nation or 
territory that this item originated from.  Freebase's Properties 
always live under a Freebase Type, like Musical Group.  Which is why 
on Wikidata, even seeing P17 on the U2 topic page makes me wonder what 
kind of schema Wikidata is trying to pull off.  But it appears that 
someone did not really read the description page of P17, like I just 
did, then they would see it just is not allowed like that, but instead 
should have used P27, but then you can't have a date of birth for a 
Musical Group (band), which voids using even P27 on an instance of band.


I understand, there are many holes in Wikidata's schema currently.  I 
am one of several Freebase experts coming over that can help Wikidata 
identify those problematic Schema. :-)



2. How does Wikidata want to handle locking down Property
descriptions
(Freebase uses Permissions and Owners), where the complete
meaning of
something being changed might cause severe wrongful polluted
data ?


There is no such thing in wikis.
http://c2.com/cgi/wiki?WikiDesignPrinciples
https://meta.wikimedia.org/wiki/The_wiki_way


But Wikidata is not a wiki in the true sense, or should not be 
purported as one Because it is not Schema-less, but in fact, 
prescribes to a publicly editable and agreed upon Schema model.


One thing I did notice is that the Wikidata Schema model is actually 
composed of both agreement on the 2 tabs of 
https://www.wikidata.org/wiki/Property_talk:P17  both the Property 
tab, AND the Discussions tabcombined...give the effective model of 
the Property...whereas in Freebase, we would just have the Property, 
where all rules and definitions about it are stored (Discussions about 
a Property were stored on our wiki and also our mailing list).  I 
enjoy the Wikidata way a bit more compared to Freebase, the benefit 
being a primary place to see the defines of the Property as well as 
the Discussion and questions about it in the past.


The errors are corrected after the fact; the central control
system is not made of permissions, but of checks like the
constraint violations bots mentioned above. What other pollutions
of the data you have in mind?


And that is my worry.  That the Schema model is publicly editable at 
any time.  And constraint violations are only effective against a 
Well Defined Property.  But what if I do not Well Define that 
property, or worst, I completely change the meaning of that Property.  
Imagine if I suddenly change the meaning of one of your MySQL table 
columns... like, PERSON suddenly becomes FURNITURE. That can happen 
with Wikidata's publicly editable Schema modelif someone 
maliciously changes the description of that P17 Country to something 
very generic like a state oh really ?  What kind of state ?  
Nations only ? Or territories considered as an organized political 
community under one government.? or both ?  it appears that P17's 
Discussion clarifies this a bit, and defines it a bit more narrowly 
and would not allow just any territory with a political community.


We have the same problem in Freebase, where if by public agreement, we 
change the meaning of a Property so much that it might cause erroneous 
data statements, then we deprecate that Property and create a new one, 
splitting off the various statements into their proper form and 
letting the Community know, and also performing the data tasks to 
subscribe the old data to the new Schema.


The pollution of data would happen if by agreement P17's Discussion 
page drastically changed the intended meaning of it, then all the data 
that used P17 would need to be cleaned up.


How does Wikidata intend to deal with those kinds of changes to 
Property 

Re: [Wikidata-l] Integration of ProteinBoxBot to the Mediawiki Gerrit

2014-12-22 Thread Ricordisamoa

Il 11/11/2014 17:18, Sebastian Burgstaller ha scritto:

Hi everyone!

I would like to quickly introduce myself, my name is Sebastian 
(https://www.wikidata.org/wiki/User:Sebotic) and I will join the lab 
of Andrew Su (sulab.org) towards the end of this year. As you probably 
know, our main aim regarding WikiData is to integrate human genomics 
and medical data, this will also be my primary project.


We therefore thought that it might be interesting to the WikiData 
community and also helpful for us, if we could integrate our sources 
for the ProteinBoxBot into the Mediawiki Gerrit code review system and 
receive  contributions from the large community.


The projects on Mediawiki Gerrit seems to be primarily concerning 
Mediawiki itself or addons to it. Would it be possible, to get our 
ProteinBoxBot project integrated on Gerrit? It is currently hosted on 
https://bitbucket.org/sulab/wikidatagenebot and is under strong 
development by Andra Wagmeester and I am about to join him in this 
effort.


Thank you!

Best regards,
Sebastian

Yes, please!
Per my previous proposal 
https://lists.wikimedia.org/pipermail/wikidata-l/2014-October/004804.html, 
I think the project would benefit from being hosted on the same platform 
as Pywikibot.

And what about a 'ProteinBoxBot' Phabricator project?
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Sitelinks to Redirects

2014-12-09 Thread Ricordisamoa

Il 16/10/2014 18:50, Jane Darnell ha scritto:

Purodha,
Redirects are cheap - so cheap in fact, that they take up more space 
when you delete them
Every deletion of any page (as almost every action in MediaWiki) 
increases the size of the database.

That doesn't mean the wiki is more cluttered.
, so even if they are misspelled or whatever, they are mostly left to 
rot unless they break something (for example when someone wants to use 
a redlink like [[redlink]] and someone else makes a redirect for 
redlink). I don't think there is any Wikimedia project that actively 
deletes redirects.


In general, redirects are supposed to be used as alternate names for 
the same thing, and in Wikidata, this is done by typing in alternate 
labels. Of course people also use redirects as a way of bundling 
concepts - just take a look at all the redirects to the article for 
insurance for all the types of insurance that don't yet have their 
own article.


Before Wikidata there were lots of interwiki links to redirects, and 
this caused multiple issues with unresolvable interwikilinks. Wikidata 
was invented to be able to use persistent identifiers for Wikipedia 
articles. Now everyone is surprised that now the interwikilinks work 
differently from before. The fact that redirects are not supported is 
by design and not a bug. Going forward, instead of making redirects, 
Wikidatans should just keep creating items in Wikidata and let the 
Wikipedias take care of themselves by letting them create articles and 
redirects in the normal wiki way. It should not be a goal for Wikidata 
to sitelink to every redirect in every Wikipedia, just as it is not a 
goal to sitelink to every image on Wikimedia Commons.


The subject at hand in this email thread is that instead of creating 
an article, the user ThurnerRupert made a redirect in the German 
Wikipedia called afrikanische Pflaume that links to Prunus and 
expected to be able to interwikilink this redirect via the Wikidata 
item for African Plum to the French Wikipedia's article for safou. 
I would say that Wikidata should not support this workflow and it is 
incorrect editing behavior. This has nothing to do with the numbers of 
redirects or whether or not they need to be deleted by anybody.


Jane
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


[Wikidata-l] standardize Wikidata WikiProject logos

2014-10-21 Thread Ricordisamoa
Currently, we have different styles: c:Category:Wikidata task forces 
https://commons.wikimedia.org/wiki/Category:Wikidata_task_forces.

What could a guideline be?
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] How can I increase the throughput of ProteinBoxBot?

2014-10-18 Thread Ricordisamoa

Il 03/10/2014 22:31, Legoktm ha scritto:

On 9/30/14 10:00 AM, Andra Waagmeester wrote:

Would it be accepted if we increase the throughput by running multiple
instances of ProteinBoxBot in parallel. If so, what would be an accepted
number of parallel instances of a bot to run? We can run multiple
instances from different geographical locations if necessary.

https://www.mediawiki.org/wiki/API:Etiquette recommends that you don't
make parallel requests. But if you're just going to run another instance
of your bot I doubt it would cause any problems.

Could you provide a link to your source code (I didn't see a link on the
bot's userpage)? IIRC, you might be using pywikibot, but I don't
remember exactly. Myself and other pywikibot developers can help you
optimize your code :)

-- Legoktm
I suppose the latest version is 
https://bitbucket.org/chinmay26/proteinboxbot/.
Pywikibot recently introduced a very simple way of manipulating items 
and sending back the changes via wbeditentity.
If the whole Gene Wiki Project codebase were hosted on 
gerrit.wikimedia.org, more people could help ;-)


___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l