Thank you very much. That makes sense. -- Young
At 2010-08-30 14:59:14,"Sebastian Schelter" <[email protected]> wrote: >Hi Young, > >there are several tweaks available that can help you reduce response time. > >* you can precompute the item-similarities offline with >ItemSimilarityJob and load the results into memory in the online >recommender, so it can just look them up from RAM >* in ItemSimilarityJob you can impose a limit on the number of similar >items per single item in the results, this way you can limit the number >of similarities taken into consideration while computing recommendations >* at the start of the recommendation process, a CandidateItemsStrategy >implementation is used to identify the initial set of items that might >be worth recommending. There are several implementations available that >might be worth trying and you can also create your own version optimized >for your usecase > >Furthermore one of the key benefits of item-based recommendation is that >the item relations tend to be very static. Thus it might be sufficient >to precompute the item similarities in intervals (like once a day for >example depending on your usecase). So if you only update your data once >a day for example, you can consider the data readonly between those >updates which makes it an ideal caching candidate. If you manage to >cache your computed recommendations answering a request from an >in-memory cache should be accomplished in less than 1ms. > >--sebastian > >Am 29.08.2010 21:47, schrieb Young: >> Hi all, >> >> Based on 1M dataset, about how many requests could be expected to be handled >> at a time when using item-based recommender if the engine runs on a Core2 >> 2.4G CPU and 4G meomory. >> >> Thank you very much. >> >> -- Young >> >> >
