On AcaWiki:

This sounds in line with  AcaWiki's larger goal, and the small
community there is generally open to new ideas about how to structure
pages and data.  I also think the project would be appropriate as a
Wikimedia project, which would address many of the self-hosting issues
and tie into similar work on a WikiScholar project.  No need to have
multiple tiny projects when a single one would do.

> I think we want to specifically target
> our annotated bibliography to researchers, but AcaWiki appears to be
> targeting laypeople as well as researchers (and IMO it would be very
> tricky to do both well).

You could allow each biblio page to decide who its audience is.  If
there is ever a conflict between a lay and a specialist audience, you
can have two sets of annotations.  I'd like to see this happen in
practice before optimizing against it.

> * I don't think the focus on "summaries" is right. I think we need a
> structured infobox plus semi-structured text (e.g. sections for
> contributions, evidence, weaknesses, questions).

Again, I think both could be appropriate for a stub bibliography page;
and that a great one would include both a summary and structured
sections and infobox data.  [acawiki does like infobox-style
structure]

> * It doesn't look like a MediaWiki. Since the MW software is so

This is easy to fix -- people who like the current acawiki look can
use their own skin.


On Data-scraping and WikiScholar parallels:

>> My only experience with "scraping" pages is with Zotero, and it does it
>> beautifully. I assume (but don't know) that the current generation of
>> other bibliography software would also do a good job. Anyway, Zotero has
>> a huge support community, and scrapers for major sources (including
>> Google Scholar for articles and Amazon for books) are kept very well up
>> to date for the most part.
>
> Perhaps I'm just unlucky, then - I've only ever tried it on ACM papers
> (which it failed to do well, so I stopped).

Brian Mingus, who is working on WikiScholar (another related project
which may be suitable) has a great deal of exprience with scraping,
both using APIs and otherwise, and that is the foundation of his
effort.

> I don't know anything about how article IDs works in Zotero, but how to
> build a unique ID for each is an interesting, subtle, and important
> problem.

This is important, and has also been discussed elsewhere.  Some of
this discussion would be appropriate here:
http://meta.wikimedia.org/wiki/Talk:WikiScholar

-- 
Samuel Klein          identi.ca:sj           w:user:sj          +1 617 529 4266

_______________________________________________
Wiki-research-l mailing list
Wiki-research-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wiki-research-l

Reply via email to