I would also point you at many of Mr. Underwood's blog posts, as they have
helped me quite a bit :)

http://techblog.chegg.com/2012/12/12/measuring-search-relevance-with-mrr/

On Wed, Feb 24, 2016 at 11:37 AM, Doug Turnbull <
dturnb...@opensourceconnections.com> wrote:

> For relevance, I would also look at retention metrics. Harder to tie back
> to a specific search. But what happens after the conversion? Did they
> purchase the product and hate it? Or did they come back for more? Retention
> metrics say a lot about the whole experience. But for many search-heavy
> applications search is 90% of the user experience. Was it really relevant
> if they purchased a product, but were dissatisfied? Did search make a
> promise that wasn't delivered on? This is something I personally noodle
> about and not something I have a canned solution for.
>
> There's an obsession with what I think of as "engagement metrics" or
> "session metrics". Engagement metrics like CTR and are handy because
> they're easy to tie to a search. Search, search, search <click>, search,
> search, <click> <purchase>.
>
> I'm always cautious of click-thru metrics. Beware of the biases in your
> clickthru metrics
>
> http://opensourceconnections.com/blog/2014/10/08/when-click-scoring-can-hurt-search-relevance-a-roadmap-to-better-signals-processing-in-search/
>
> Another reason to be cautious is user behavioral data can require
> domain-specific interpretation. A good book on recommendor can talk more
> about interpreting user behavior to see if an item was relevant. For
> examples Pratical Recommender Systems by Kim Falk (a Manning MEAP) spends a
> great deal of time talking through gathering evidence whether a user liked
> the thing they clicked on or not. For example, did the user click a movie
> and go back immediately? Start watching a movie and go back in 5 minutes
> indicating they hated it? Or watch a movie all the way through?
>
> Related to interpreting behavior -- understand the kinds of searchers out
> there. Understand what sort of user experience you've built. Informational
> searchers doing research will look at every item and evaluate them. For
> example, a paralegal searching a legal application may need to examine
> every result carefully. Navigational searchers want to hunt for one thing.
> Everyday e-commerce searchers clicking on every result is probably
> disastrous. However, the purchasing dept of an organization MIGHT look at
> every result and that might be ok.
>
> Beware of search's long tail. You can gather metrics on all your searches
> and where users are clicking, but search has a notorious long tail. Many of
> my clients have meaningful metrics over perhaps the top 50 results before
> quickly going off into obscurity of statistical insignificance per search.
> This depends entirely on the type of search application you're developing.
> Some kind of niche product with a handful of searches per day? Or giant
> e-commerce site?
>
> Sometimes what's simpler is to do usability testing or to sit with an
> expert user and gather relevance judgments--grades on what's relevant and
> what's not. (this is what we do with Quepid). This works particularly well
> for these niche, expert search subjects
>
> Anyway, there's still quite a bit of art to interpreting search metrics. I
> would argue to keep the human and domain expert in the loop understanding
> and interpreting metrics. But its a yin-and-yang. You also need to be able
> to tell that supposed domain expert when they're wrong.
>
> Sorry for long winded email, but these topics dominate my
> dreams/nightmares these days :)
>
> Best
> -Doug
>
>
>
>
> On Wed, Feb 24, 2016 at 11:20 AM, Walter Underwood <wun...@wunderwood.org>
> wrote:
>
>> Click through rate (CTR) is fundamental. That is easy to understand and
>> integrates well with other business metrics like conversion. CTR is at
>> least one click anywhere in the result set (first page, second page, …).
>> Count multiple clicks as a single success. The metric is, “at least one
>> click”.
>>
>> No hit rate is sort of useful, but you need to know which queries are
>> getting no hits, so you can fix it.
>>
>> For latency metrics, look at 90th percentile or 95th percentile. Average
>> is useless because response time is a one-sided distribution, so it will be
>> thrown off by outliers. Percentiles have a direct customer satisfaction
>> interpretation. 90% of searches were under one second, for example. Median
>> response time should be very, very fast because of caching in Solr. During
>> busy periods, our median response time is about 1.5 ms.
>>
>> Number of different queries per conversion is a good way to look how
>> query assistance is working. Things like autosuggest, fuzzy, etc.
>>
>> About 10% of queries will be misspelled, so you do need to deal with that.
>>
>> Finding underperforming queries is trickier. I really need to write an
>> article on that.
>>
>> “Search Analytics for Your Site” by Lou Rosenfeld is a good introduction.
>>
>> http://rosenfeldmedia.com/books/search-analytics-for-your-site/ <
>> http://rosenfeldmedia.com/books/search-analytics-for-your-site/>
>>
>> Sea Urchin is doing some good work in search metrics:
>> https://seaurchin.io/ <https://seaurchin.io/>
>>
>> wunder
>> Walter Underwood
>> wun...@wunderwood.org
>> http://observer.wunderwood.org/  (my blog)
>> Search Guy, Chegg
>>
>> > On Feb 24, 2016, at 2:38 AM, Emir Arnautovic <
>> emir.arnauto...@sematext.com> wrote:
>> >
>> > Hi Bill,
>> > You can take a look at Sematext's search analytics (
>> https://sematext.com/search-analytics). It provides some of metrics you
>> mentioned, plus some additional (top queries, CTR, click stats, paging
>> stats etc.). In combination with Sematext's performance metrics (
>> https://sematext.com/spm) you can have full picture of your search
>> infrastructure.
>> >
>> > Regards,
>> > Emir
>> >
>> > --
>> > Monitoring * Alerting * Anomaly Detection * Centralized Log Management
>> > Solr & Elasticsearch Support * http://sematext.com/
>> >
>> >
>> > On 24.02.2016 04:07, William Bell wrote:
>> >> How do others look at search metrics?
>> >>
>> >> 1. Search conversion? Do you look at searches and if the user does not
>> >> click on a result, and reruns the search that would be a failure?
>> >>
>> >> 2. How to measure auto complete success metrics?
>> >>
>> >> 3. Facets/filters could be considered negative, since we did not find
>> the
>> >> results that the user wanted, and now they are filtering - who to
>> measure?
>> >>
>> >> 4. One easy metric is searches with 0 results. We could auto expand
>> the geo
>> >> distance or ask the user "did you mean" ?
>> >>
>> >> 5. Another easy one would be tech performance: "time it takes in
>> seconds to
>> >> get a result".
>> >>
>> >> 6. How to measure fuzzy? How do you know you need more synonyms? How to
>> >> measure?
>> >>
>> >> 7. How many searches it takes before the user clicks on a result?
>> >>
>> >> Other ideas? Is there a video or presentation on search metrics that
>> would
>> >> be useful?
>> >>
>> >
>>
>>
>
>
> --
> *Doug Turnbull **| *Search Relevance Consultant | OpenSource Connections
> <http://opensourceconnections.com>, LLC | 240.476.9983
> Author: Relevant Search <http://manning.com/turnbull>
> This e-mail and all contents, including attachments, is considered to be
> Company Confidential unless explicitly stated otherwise, regardless
> of whether attachments are marked as such.
>



-- 
*Doug Turnbull **| *Search Relevance Consultant | OpenSource Connections
<http://opensourceconnections.com>, LLC | 240.476.9983
Author: Relevant Search <http://manning.com/turnbull>
This e-mail and all contents, including attachments, is considered to be
Company Confidential unless explicitly stated otherwise, regardless
of whether attachments are marked as such.

Reply via email to