[Wikidata-l] Tables
Hi there, can tables be stored within wikidata database? I mean simple 2D tables like excel spreadsheets... Jan ___ Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l
Re: [Wikidata-l] Tables
Am 27.05.2014 13:15, schrieb Jan Kučera: Hi there, can tables be stored within wikidata database? I mean simple 2D tables like excel spreadsheets... No. The data structure does not lend itself to representing tabular data easily. I think a separate extension for handling tabular data (probably as CSV) would be quite nice. It wouldn't be hard to do either. But I'm not sure how that would tie into wikidata as a project. -- daniel -- Daniel Kinzler Senior Software Developer Wikimedia Deutschland Gesellschaft zur Förderung Freien Wissens e.V. ___ Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l
Re: [Wikidata-l] Tables
I am not an expert of Wikidata but I work a lot in integration of databases with several middlewares or tools. I think that the best way is to ask a solution to store and to download data in a format compatible with datasheets like CSV. Excel is considered also a tool to do some basic analysis, but it can be connected easily to a data source (if well structured). Excel itself is not a good approach to store data, so it's not a good solution to keep the data in excel format in a database. Doesn't make sense to store a 2D tables in a database in my opinion because the data have no sense and they are not helpful to anyone. They can be stored like a text file, but I would not imagine the series of errors that can be generated importing these data again. Regards On Tue, May 27, 2014 at 1:15 PM, Jan Kučera kozuc...@gmail.com wrote: Hi there, can tables be stored within wikidata database? I mean simple 2D tables like excel spreadsheets... Jan ___ Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l -- Ilario Valdelli Wikimedia CH Verein zur Förderung Freien Wissens Association pour l’avancement des connaissances libre Associazione per il sostegno alla conoscenza libera Switzerland - 8008 Zürich Wikipedia: Ilario https://meta.wikimedia.org/wiki/User:Ilario Facebook: Ilario Valdelli https://www.facebook.com/ivaldelli Twitter: Ilario Valdelli https://twitter.com/ilariovaldelli Linkedin: Ilario Valdelli http://www.linkedin.com/profile/view?id=6724469 Tel: +41764821371 http://www.wikimedia.ch ___ Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l
Re: [Wikidata-l] Tables
Please, leave your comments here too: https://meta.wikimedia.org/wiki/Requests_for_comment/How_to_deal_with_open_datasets I've been gathering comments from several people, and in the next days I will try to summarize these suggestions to be discussed on irc. Thanks, Micru On Tue, May 27, 2014 at 2:12 PM, Ilario Valdelli valde...@gmail.com wrote: I am not an expert of Wikidata but I work a lot in integration of databases with several middlewares or tools. I think that the best way is to ask a solution to store and to download data in a format compatible with datasheets like CSV. Excel is considered also a tool to do some basic analysis, but it can be connected easily to a data source (if well structured). Excel itself is not a good approach to store data, so it's not a good solution to keep the data in excel format in a database. Doesn't make sense to store a 2D tables in a database in my opinion because the data have no sense and they are not helpful to anyone. They can be stored like a text file, but I would not imagine the series of errors that can be generated importing these data again. Regards On Tue, May 27, 2014 at 1:15 PM, Jan Kučera kozuc...@gmail.com wrote: Hi there, can tables be stored within wikidata database? I mean simple 2D tables like excel spreadsheets... Jan ___ Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l -- Ilario Valdelli Wikimedia CH Verein zur Förderung Freien Wissens Association pour l’avancement des connaissances libre Associazione per il sostegno alla conoscenza libera Switzerland - 8008 Zürich Wikipedia: Ilario https://meta.wikimedia.org/wiki/User:Ilario Facebook: Ilario Valdelli https://www.facebook.com/ivaldelli Twitter: Ilario Valdelli https://twitter.com/ilariovaldelli Linkedin: Ilario Valdellihttp://www.linkedin.com/profile/view?id=6724469 Tel: +41764821371 http://www.wikimedia.ch ___ Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l -- Etiamsi omnes, ego non ___ Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l
Re: [Wikidata-l] Subclass of/instance of
On Thu, May 15, 2014 at 5:03 PM, Markus Krötzsch mar...@semantic-mediawiki.org wrote: I applaud your comparison of inferencing with a form of decompression. I think this is a nice intuition (in fact, some people have researched semantic compression where one tries to reduce the size of a knowledge base by eliminating things that follow from the rest anyway). Markus, sorry the delay answering to this, I had to let the ideas grow for a while. I also like the idea of decompression, that is what makes your database of inferred data even more useful. There is a lot of data that can be inferred, and not just from following the relationships, but by computing. For instance population density which can be calculated from area and population, or aggregates of the population of each town in a district. Another source of inferred statements are wp categories. Most of them are very easily translatable into statements, and the other way round too. A place where to store and process these inferences would be most useful if WD is not the right place. You also say: Constraints are a great start. We should now ask how we could improve the management of constraints in the future, and which constraints we will have then. The first step will be having them as statements, then having them as queries, and finally automating their correction, either by semi-automatic tools, or with gamification. How to automatically transform a constraint into a game to solve the outliers it might be also an interesting topic. And of course, more far fetched, but nevertheless relevant is how to connect the property to a perceptual mechanism. About improving the reliability: yes, as wikidata grows bigger some statements become more important. There is something to be learnt about how neural nets work, specially strengthening most-used (or traveled, or accepted, or viewed) connections. Another process little understood now is the need to forget, or in wikidata terms auto-deprecate information that is no longer current. Not very relevant now, but something to keep in mind for the next years. Cheers, Micru ___ Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l