On Nov 19, 2007, at 3:41 PM, Chris Hostetter wrote:


: info, etc. could be stripped fairly easily. So, we wouldn't necessarily know : who is searching for "Yonik Seeley" when we see that query term, just that it : was searched for. Maybe we can inquire to infrastructure what is even

It's a largely theoretical arguement (particularly relating to a subset of results on a specific domain as opposed to a subset from a specific search
engine) but the nutshell is: there may in fact be identifiable info in
the query string itself, so it's good to have some sanity checking before
exposing the queries to the world.

Agreed.



: At any rate, I think the bigger issue is finding a good set of data and query : logs that we can use. An alternate way is to just start creating a query set : based on the Wikipedia data, but that isn't as "real world" as query logs are.

I think looking at refer URLs containing query strings grouped by TLP site would give us lots of useful "small" collections of docs and query strings that are considered "relevent" (albeit: not by a human judgement, but by
some other search engine -- it's a start)

if you take something like the online HTTPD manual, each URL can be easily mapped to a machine parsable XML version, and i'm sure we can find plenty
of good query strings in the refer logs for httpd.apache.org.


+1


: Here's another possible thought: What if we took our own java- user mailing : list for a time period and we used the subject line or some other piece of : info in the text (maybe we can automatically identify questions (not hard to : do for simple cases (just identify sentences ending in ?), which would give us
: enough, methinks) and treat them as queries?  This may be a decent

two concerns i would have:
1) the person asking the question doesn't always know what to ask about
    (the X/Y problem) which could lead to missleading query/result
    matches.
 2) people aren't always "on topic" ... discussions can branch/evolve
without subjects changing (formatl documentation doesn't really have
    this problem)

Both true, but as with the other scenarios (except TREC) there is a human in the loop and we don't have to take every question available, just 100 or so good ones. Maybe we could even use the FAQs applied against the archive.

The other hard part about the mail archive is that you are likely to have matches against emails which are asking the question and not just those emails answering the question. Not sure if those are relevant or not. Sometimes, for me, just reading how someone else phrased the problem is enough to spur an answer.



: Of course, we could see if there is a way to purchase the TREC data
: (donations, anyone?) and make it available to committers on zones. This is

if spending money is an option, but spending enough money for TREC isn't an option, something i've been considering is using Amazon's mechanical turk to generate judgements ... take some seed data (ie: refer log query
strings and the title/summary/url of the top 5 URLs for each) and give
mturk users $0.05 to rank those 5 in order of how well they match.



I believe the TREC collection costs somewhere around $300, so it isn't going to break the bank. Perhaps we could ask the board to pay it or maybe we could arrange for donations. I'd be willing to kick in up to $50 to have it available, but I still don't like this route since only committers can have access b/c it is on zones and I don't know that this is that high of a priority for committers. Instead, I want something researchers and upstart grad students can easily download and try out and that we can all then discuss b/c we all have the data. Furthermore, by having multiple data sets, we can hopefully avoid the overtuning problem.

-Grant

---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to