On Wed, Sep 11, 2013 at 12:07 AM, Sean Owen <[email protected]> wrote:
> > 2. Do we have to tune the "similarityclass" parameter in item-based CF? > If > > so, do we compare the mean average precision values based on validation > > data, and then report the same for the test set? > > > > > Yes you are conceptually looking over the entire hyper-parameter space. If > the similarity metric is one of those, you are trying different metrics. > Grid search, just brute-force trying combinations, works for 1-2 > hyper-parameters. Otherwise I'd try randomly choosing parameters, really, > or else it will take way too long to explore. You try to pick > hyper-parameters 'nearer' to those that have yielded better scores. > Or use a real exploration algorithm. For my favorite (hear that horn blowing?) see this article on recorded step meta-mutation.<http://arxiv.org/abs/0803.3838> The idea is a randomized search, but with something akin to momentum. This lets you search nasty landscapes with pretty pretty good robustness and smooth ones with fast convergence. The code and theory are simple and there is an implementation in Mahout.
