Hi y'all, Good ideas.
Queries for such a big number of items are often timing out. Here is a working QLever query for items with more than 10 sitelinks : https://qlever.cs.uni-freiburg.de/wikidata/VdiLsm. There is only one result, you can decrease the value for more results. Reminder, QLever results are not updated in real time, it's based on dumps (who are late because right now, results are from 29.01.2025). Cheers, Nicolas Le ven. 28 févr. 2025 à 18:05, Amir E. Aharoni <[email protected]> a écrit : > I tried some queries, and they all timed out :( > > I'm not very good at SPARQL. > > But I agree with Andy: Dividing hundreds of thousands of items into small > groups that can be processed by people who are likely to know something > relevant about those items, is probably a better way to try to handle it > than just looking at a huge pile of items. > > Some ways to divide them that I can think of immediately: > 1. Having a sitelink to particular languages. > 2. Having a label or a description in a particular languages. > 3. Having certain characteristics in the label, like length, or presence > of certain characters (even a mostly arbitrary characteristic, like "label > starts with the letters 'Mi' " or "has digits in label", is better than > nothing). > > If someone can make a bunch of queries that do something like this and > actually work (and don't time out), this can be a nice beginning. > > בתאריך יום ו׳, 28 בפבר׳ 2025, 11:42, מאת Andy Mabbett < > [email protected]>: > >> On Fri, 28 Feb 2025 at 16:06, Romaine Wiki <[email protected]> >> wrote: >> >> > There are another 493k items with only one identifier and no other >> statement. >> > https://qlever.cs.uni-freiburg.de/wikidata/Z8OkZi?exec=true >> > Often that single identifier is just the Google Knowledge Graph ID >> (P2671). >> >> The first half-dozen or so I checked all also have a Wikipedia link in >> one or more languages. >> >> Maybe it would be worth making a query for each of the top, say twenty >> languages and posting on the relevant Village Pump? >> >> Or having an article or talk page template added by a bot, to each >> affected article? >> >> -- >> Andy Mabbett >> https://pigsonthewing.org.uk >> _______________________________________________ >> Wikidata mailing list -- [email protected] >> Public archives at >> https://lists.wikimedia.org/hyperkitty/list/[email protected]/message/4QUE62RXUBELM5HGQRNSZDKWR2QKSH4D/ >> To unsubscribe send an email to [email protected] >> > _______________________________________________ > Wikidata mailing list -- [email protected] > Public archives at > https://lists.wikimedia.org/hyperkitty/list/[email protected]/message/OBQ3R7C4SAHDAYD5O2D7LMCWTVJ3LX5K/ > To unsubscribe send an email to [email protected] >
_______________________________________________ Wikidata mailing list -- [email protected] Public archives at https://lists.wikimedia.org/hyperkitty/list/[email protected]/message/VIXWQD2A4OF5JSXS7ZKTRGWYHZ6Q7WC7/ To unsubscribe send an email to [email protected]
