Item-Based Collaborative Filtering Recommendation Algorithms - Highlighted Paper
Item-Based Collaborative Filtering Recommendation Algorithms - Highlighted Paper
Algorithms
285
item-based algorithms may be able to provide the same qual- In one study using synthetic data, Horting produced better
ity as the user-based algorithms with less online computa- predictions than a nearest neighbor algorithm [1].
tion. Schafer et al., [26] present a detailed taxonomy and exam-
ples of recommender systems used in E-commerce and how
1.1 Related Work they can provide one-to-one personalization and at the same
can capture customer loyalty. Although these systems have
In this section we brie
y present some of the research lit- been successful in the past, their widespread use has exposed
erature related to collaborative ltering, recommender sys- some of their limitations such as the problems of sparsity in
tems, data mining and personalization. the data set, problems associated with high dimensionality
Tapestry [10] is one of the earliest implementations of col- and so on. Sparsity problem in recommender system has
laborative ltering-based recommender systems. This sys- been addressed in [23, 11]. The problems associated with
tem relied on the explicit opinions of people from a close-knit high dimensionality in recommender systems have been dis-
community, such as an oÆce workgroup. However, recom- cussed in [4], and application of dimensionality reduction
mender system for large communities cannot depend on each techniques to address these issues has been investigated in
person knowing the others. Later, several ratings-based au- [24].
tomated recommender systems were developed. The Grou- Our work explores the extent to which item-based recom-
pLens research system [19, 16] provides a pseudonymous menders, a new class of recommender algorithms, are able
collaborative ltering solution for Usenet news and movies. to solve these problems.
Ringo [27] and Video Recommender [14] are email and web-
based systems that generate recommendations on music and 1.2 Contributions
movies, respectively. A special issue of Communications of This paper has three primary research contributions:
the ACM [20] presents a number of dierent recommender
systems. 1. Analysis of the item-based prediction algorithms and
Other technologies have also been applied to recommender identication of dierent ways to implement its sub-
systems, including Bayesian networks, clustering, and Hort- tasks.
ing. Bayesian networks create a model based on a training
set with a decision tree at each node and edges represent- 2. Formulation of a precomputed model of item similarity
ing user information. The model can be built o-line over a to increase the online scalability of item-based recom-
matter of hours or days. The resulting model is very small, mendations.
very fast, and essentially as accurate as nearest neighbor 3. An experimental comparison of the quality of several
methods [6]. Bayesian networks may prove practical for en- dierent item-based algorithms to the classic user-based
vironments in which knowledge of user preferences changes (nearest neighbor) algorithms.
slowly with respect to the time needed to build the model
but are not suitable for environments in which user prefer- 1.3 Organization
ence models must be updated rapidly or frequently.
Clustering techniques work by identifying groups of users The rest of the paper is organized as follows. The next
who appear to have similar preferences. Once the clusters section provides a brief background in collaborative lter-
are created, predictions for an individual can be made by av- ing algorithms. We rst formally describe the collaborative
eraging the opinions of the other users in that cluster. Some ltering process and then discuss its two variants memory-
clustering techniques represent each user with partial par- based and model-based approaches. We then present some
ticipation in several clusters. The prediction is then an aver- challenges associated with the memory-based approach. In
age across the clusters, weighted by degree of participation. section 3, we present the item-based approach and describe
Clustering techniques usually produce less-personal recom- dierent sub-tasks of the algorithm in detail. Section 4 de-
mendations than other methods, and in some cases, the clus- scribes our experimental work. It provides details of our
ters have worse accuracy than nearest neighbor algorithms data sets, evaluation metrics, methodology and results of
[6]. Once the clustering is complete, however, performance dierent experiments and discussion of the results. The -
can be very good, since the size of the group that must be nal section provides some concluding remarks and directions
analyzed is much smaller. Clustering techniques can also for future research.
be applied as a "rst step" for shrinking the candidate set
in a nearest neighbor algorithm or for distributing nearest- 2. COLLABORATIVE FILTERING BASED
neighbor computation across several recommender engines. RECOMMENDER SYSTEMS
While dividing the population into clusters may hurt the
accuracy or recommendations to users near the fringes of Recommender systems systems apply data analysis tech-
their assigned cluster, pre-clustering may be a worthwhile niques to the problem of helping users nd the items they
trade-o between accuracy and throughput. would like to purchase at E-Commerce sites by producing
a predicted likeliness score or a list of top{N recommended
Horting is a graph-based technique in which nodes are items for a given user. Item recommendations can be made
users, and edges between nodes indicate degree of similarity using dierent methods. Recommendations can be based
between two users [1]. Predictions are produced by walking on demographics of the users, overall top selling items, or
the graph to nearby nodes and combining the opinions of past buying habit of users as a predictor of future items.
the nearby users. Horting diers from nearest neighbor as Collaborative Filtering (CF) [19, 27] is the most success-
the graph may be walked through other users who have not ful recommendation technique to date. The basic idea of
rated the item in question, thus exploring transitive rela- CF-based algorithms is to provide item recommendations
tionships that nearest neighbor algorithms do not consider. or predictions based on the opinions of other like-minded
286
users. The opinions of users can be obtained explicitly from on other items. The model building process is performed
the users or by using some implicit measures. by dierent machine learning algorithms such as Bayesian
network, clustering, and rule-based approaches. The
2.0.1 Overview of the Collaborative Filtering Pro- Bayesian network model [6] formulates a probabilistic model
cess for collaborative ltering problem. Clustering model treats
The goal of a collaborative ltering algorithm is to sug- collaborative ltering as a classication problem [2, 6, 29]
gest new items or to predict the utility of a certain item for and works by clustering similar users in same class and esti-
a particular user based on the user's previous likings and mating the probability that a particular user is in a partic-
the opinions of other like-minded users. In a typical CF sce- ular class C , and from there computes the conditional prob-
nario, there is a list of m users U = fu1 ; u2 ; : : : ; u g and a
m ability of ratings. The rule-based approach applies associ-
list of n items I = fi1 ; i2 ; : : : ; i g. Each user u has a list
n i ation rule discovery algorithms to nd association between
of items I i , which the user has expressed his/her opinions
u co-purchased items and then generates item recommenda-
about. Opinions can be explicitly given by the user as a rat- tion based on the strength of the association between items
ing score, generally within a certain numerical scale, or can [25].
be implicitly derived from purchase records, by analyzing
timing logs, by mining web hyperlinks and so on [28, 16]. 2.0.2 Challenges of User-based Collaborative Filter-
Note that I i I and it is possible for I i to be a null-set.
u u
ing Algorithms
There exists a distinguished user u 2 U called the active
a User-based collaborative ltering systems have been very
user for whom the task of a collaborative ltering algorithm successful in past, but their widespread use has revealed
is to nd an item likeliness that can be of two forms. some potential challenges such as:
Prediction is a numerical value, P , expressing the Sparsity. In practice, many commercial recommender
a;j
predicted likeliness of item i 62 I a for the active user systems are used to evaluate large item sets (e.g., Ama-
j u
u . This predicted value is within the same scale (e.g., [Link] recommends books and [Link] recom-
a
from 1 to 5) as the opinion values provided by u . mends music albums). In these systems, even active
a
users may have purchased well under 1% of the items
Recommendation is a list of N items, I I , that r (1% of 2 million books is 20; 000 books). Accordingly,
the active user will like the most. Note that the recom- a recommender system based on nearest neighbor al-
mended list must be on items not already purchased by gorithms may be unable to make any item recommen-
the active user, i.e., I \ I a = . This interface of CF
r u dations for a particular user. As a result the accuracy
algorithms is also known as Top-N recommendation. of recommendations may be poor.
Figure 1 shows the schematic diagram of the collaborative Scalability. Nearest neighbor algorithms require com-
ltering process. CF algorithms represent the entire m n putation that grows with both the number of users
user-item data as a ratings matrix, A. Each entry a in A and the number of items. With millions of users and
items, a typical web-based recommender system run-
i;j
287
Item for which prediction
i1 i2 .. ij .. in is sought
u1
u2
. Pa,j (prediction on
. Prediction item j for the active
user)
ua
. Recommendation
. {T i1 , T i2 , ..., T iN } Top-N
um list of items for the
active user
Active user
Input (ratings table) CF-Algorithm Output interface
schemes have been proposed to compute the association be- sim(i; j ) is given by
tween items ranging from probabilistic approach [6] to more
sim(i; j ) = cos(~i; ~j ) = ~ i j ~
traditional item-item correlations [15, 13]. We present a de- ~ ~
tailed analysis of our approach in the next section. kik2 kj k2
where \" denotes the dot-product of the two vectors.
3. ITEM-BASED COLLABORATIVE FILT-
ERING ALGORITHM 3.1.2 Correlation-based Similarity
In this section we study a class of item-based recommen- In this case, similarity between two items i and j is mea-
dation algorithms for producing predictions to users. Unlike sured by computing the Pearson-r correlation corr . To i;j
the user-based collaborative ltering algorithm discussed in make the correlation computation accurate we must rst
Section 2, the item-based approach looks into the set of isolate the co-rated cases (i.e., cases where the users rated
items the target user has rated and computes how simi- both i and j ) as shown in Figure 2. Let the set of users who
lar they are to the target item i and then selects k most both rated i and j are denoted by U then the correlation
similar items fi1 ; i2 ; : : : ; i g. At the same time their cor-
k similarity is given by
responding similarities fs 1 ; s 2 ; : : : ; s g are also computed.
i i ik
Once the most similar items are found, the prediction is P2 (R R )(R R )
then computed by taking a weighted average of the target
user's ratings on these similar items. We describe these two
sim(i; j ) = qP u U u;i
qPi u;j j
:
2 (R R )2 2 (R R )2
aspects, namely, the similarity computation and the predic- u U u;i i u U u;j j
tion generation in details here. Here R denotes the rating of user u on item i, R is the
u;i i
these two vectors. Formally, in the m n ratings matrix Here R is the average of the u-th user's ratings.
in Figure 2, similarity between items i and j , denoted by
u
288
si,j =?
1 2 3 i j n-1 n
1 R R
2 - R
3.2 Prediction Computation The regression model parameters and are determined
The most important step in a collaborative ltering sys- by going over both of the rating vectors. is the error of
tem is to generate the output interface in terms of prediction. the regression model.
Once we isolate the set of most similar items based on the
similarity measures, the next step is to look into the tar- 3.3 Performance Implications
get users ratings and use a technique to obtain predictions. The largest E-Commerce sites operate at a scale that
Here we consider two such techniques. stresses the direct implementation of collaborative ltering.
3.2.1 Weighted Sum In neighborhood-based CF systems, the neighborhood for-
As the name implies, this method computes the prediction mation process, especially the user-user similarity computa-
on an item i for a user u by computing the sum of the ratings tion step turns out to be the performance bottleneck, which
given by the user on the items similar to i. Each ratings is in turn can make the whole process unsuitable for real-time
weighted by the corresponding similarity s between items recommendation generation. One way of ensuring high scal-
i and j . Formally, using the notion shown in Figure 3 we
i;j
ability is to use a model-based approach. Model-based sys-
can denote the prediction P as tems have the potential to contribute to recommender sys-
u;i
tems to operate at a high scale. The main idea here to iso-
P (s R ) late the neighborhood generation and prediction generation
P u;i = P
all similar items, N
(js j)
i;N
u;N
steps.
In this paper, we present a model-based approach to pre-
Basically, this approach tries to capture how the active compute item-item similarity scores. The similarity compu-
user rates the similar items. The weighted sum is scaled by tation scheme is still correlation-based but the computation
the sum of the similarity terms to make sure the prediction is performed on the item space. In a typical E-Commerce
is within the predened range. scenario, we usually have a set of item that is static com-
pared to the number of users that changes most often. The
3.2.2 Regression static nature of items leads us to the idea of precomput-
This approach is similar to the weighted sum method but ing the item similarities. One possible way of precomputing
instead of directly using the ratings of similar items it uses the item similarities is to compute all-to-all similarity and
an approximation of the ratings based on regression model. then performing a quick table look-up to retrieve the re-
In practice, the similarities computed using cosine or cor- quired similarity values. This method, although saves time,
relation measures may be misleading in the sense that two requires an O(n2 ) space for n items.
rating vectors may be distant (in Euclidean sense) yet may The fact that we only need a small fraction of similar items
have very high similarity. In that case using the raw ratings to compute predictions leads us to an alternate model-based
of the \so-called" similar item may result in poor prediction. scheme. In this scheme, we retain only a small number of
The basic idea is to use the same formula as the weighted similar items. For each item j we compute the k most sim-
sum technique, but instead of using the similar item N 's ilar items, where k n and record these item numbers
\raw" ratings values R 's, this model uses their approx- and their similarities with j . We term k as the model size.
Based on this model building step, our prediction genera-
u;N
289
;;;
;;;;
;;;;;
;;
1 2 3 i-1 i i+1 n-1 n
1
2
;;;;;
;;
1 2 3 i-1 i m-1 m
u R R R - R
si,i-1
si,3
si,1 si,m
u R R R - R
;;;;;
;;
prediction
m-1
weighted sum regression-based
m
Figure 3: Item-based collaborative ltering algorithm. The prediction generation process is illustrated for 5
neighbors
prediction is computed using basic item-based collaborative rows (i.e., 943 users) and 1682 columns (i.e., 1682 movies
ltering algorithm. that were rated by at least one of the users). For our ex-
We observe a quality-performance trade-o here: to en- periments, we also take another factor into consideration,
sure good quality we must have a large model size, which sparsity level of data sets. For the data matrix R This is
leads to the performance problems discussed above. In one dened as 1 nonzero entries
total entries . The sparsity level of the
extreme, we can have a model size of n, which will ensure the Movie data set is, therefore, 1 9431682 , which is 0:9369.
100 000 ;
exact same quality as the original scheme but will have high Throughout the paper we term this data set as ML.
space complexity. However, our model building step ensures
that we retain the most similar items. While generating pre- 4.2 Evaluation Metrics
dictions, these items contribute the most to the prediction Recommender systems research has used several types of
scores. Accordingly, we hypothesize that this model-based
approach will provide reasonably good prediction quality measures for evaluating the quality of a recommender sys-
with even a small model size and hence provide a good per- tem. They can be mainly categorized into two classes:
formance. We experimentally validate our hypothesis later
in this paper. In particular, we experiment with the model Statistical accuracy metrics evaluate the accuracy of a
size by varying the number of similar items to be stored. system by comparing the numerical recommendation
Then we perform experiments to compute prediction and scores against the actual user ratings for the user-item
response-time to determine the impact of the model size on pairs in the test dataset. Mean Absolute Error (MAE)
quality and performance of the whole system. between ratings and predictions is a widely used met-
ric. MAE is a measure of the deviation of recommen-
4. EXPERIMENTAL EVALUATION dations from their true user-specied values. For each
ratings-prediction pair < p ; q > this metric treats the i i
4.1 Data set absolute error between them, i.e., jp q j equally. The i i
N
mender system. MovieLens is a web-based research recom-
mender system that debuted in Fall 1997. Each week hun- The lower the MAE, the more accurately the recom-
dreds of users visit MovieLens to rate and receive recom- mendation engine predicts user ratings. Root Mean
mendations for movies. The site now has over 43000 users Squared Error (RMSE), and Correlation are also used
who have expressed opinions on 3500+ dierent movies. We as statistical accuracy metric.
randomly selected enough users to obtain 100; 000 ratings
from the database (we only considered users that had rated Decision support accuracy metrics evaluate how eec-
20 or more movies). We divided the database into a train- tive a prediction engine is at helping a user select high-
ing set and a test set. For this purpose, we introduced a quality items from the set of all items. These metrics
variable that determines what percentage of data is used as assume the prediction process as a binary operation-
training and test sets; we call this variable x. A value of either items are predicted (good) or not (bad). With
x = 0:8 would indicate 80% of the data was used as train- this observation, whether a item has a prediction score
ing set and 20% of the data was used as test set. The data of 1:5 or 2:5 on a ve-point scale is irrelevant if the user
set was converted into a user-item matrix A that had 943 only chooses to consider predictions of 4 or higher. The
290
most commonly used decision support accuracy met-
rics are reversal rate, weighted errors and ROC sensi- Relative performance of different similarity
tivity [23].
measures
0.86
We used MAE as our choice of evaluation metric to re- 0.84
MAE
0.76
same ordering of dierent experimental schemes in terms of 0.74
prediction quality. 0.72
0.7
0.68
4.2.1 Experimental Procedure 0.66
Adjusted cosine Pure cosine Correlation
Experimental steps. We started our experiments by rst
dividing the data set into a training and a test portion. Be- Figure 4: Impact of the similarity computation mea-
fore starting full experimental evaluation of dierent algo- sure on item-based collaborative ltering algorithm.
rithms we determined the sensitivity of dierent parameters
to dierent algorithms and from the sensitivity plots we xed
the optimum values of these parameters and used them for larity algorithms, we implemented the algorithm to compute
the rest of the experiments. To determine the parameter the neighborhood and used weighted sum algorithm to gen-
sensitivity, we work only with the training data and further erate the prediction. We ran these experiments on our train-
subdivide it into a training and test portion and carried on ing data and used test set to compute Mean Absolute Error
our experiments on them. For conducted a 10-fold cross val- (MAE). Figure 4 shows the experimental results. It can be
idation of our experiments by randomly choosing dierent observed from the results that osetting the user-average for
training and test sets each time and taking the average of cosine similarity computation has a clear advantage, as the
the MAE values. MAE is signicantly lower in this case. Hence, we select the
adjusted cosine similarity for the rest of our experiments.
Benchmark user-based system. To compare the perfor-
mance of item-based prediction we also entered the training 4.3.2 Sensitivity of Training/Test Ratio
ratings set into a collaborative ltering recommendation en- To determine the sensitivity of density of the data set, we
gine that employs the Pearson nearest neighbor algorithm carried out an experiment where we varied the value of x
(user-user). For this purpose we implemented a
exible pre- from 0:2 to 0:9 in an increment of 0:1. For each of these
diction engine that implements user-based CF algorithms. training/test ratio values we ran our experiments using the
We tuned the algorithm to use the best published Pearson two prediction generation techniques{basic weighted sum
nearest neighbor algorithm and congured it to deliver the and regression based approach. Our results are shown in
highest quality prediction without concern for performance Figure 5. We observe that the quality of prediction increase
(i.e., it considered every possible neighbor to form optimal as we increase x. The regression-based approach shows bet-
neighborhoods). ter results than the basic scheme for low values of x but
as we increase x the quality tends to fall below the basic
Experimental platform. All our experiments were imple- scheme. From the curves, we select x = 0:8 as an optimum
mented using C and compiled using optimization
ag 06. value for our subsequent experiments.
We ran all our experiments on a linux based PC with Intel
Pentium III processor having a speed of 600 MHz and 2GB 4.3.3 Experiments with neighborhood size
of RAM. The size of the neighborhood has signicant impact on the
prediction quality [12]. To determine the sensitivity of this
4.3 Experimental Results parameter, we performed an experiment where we varied
In this section we present our experimental results of ap- the number of neighbors to be used and computed MAE.
plying item-based collaborative ltering techniques for gen- Our results are shown in Figure 5. We can observe that
erating predictions. Our results are mainly divided into two the size of neighborhood does aect the quality of predic-
parts|quality results and performance results. In assess- tion. But the two methods show dierent types of sensitiv-
ing the quality of recommendations, we rst determined the ity. The basic item-item algorithm improves as we increase
sensitivity of some parameters before running the main ex- the neighborhood size from 10 to 30, after that the rate of
periment. These parameters include the neighborhood size, increase diminishes and the curve tends to be
at. On the
the value of the training/test ratio x, and eects of dierent other hand, the regression-based algorithm shows decrease
similarity measures. For determining the sensitivity of var- in prediction quality with increased number of neighbors.
ious parameters, we focused only on the training data set Considering both trends we select 30 as our optimal choice
and further divided it into a training and a test portion and of neighborhood size.
used them to learn the parameters.
4.3.4 Quality Experiments
4.3.1 Effect of Similarity Algorithms Once we obtain the optimal values of the parameters, we
We implemented three dierent similarity algorithms ba- compare both of our item-based approaches with the bench-
sic cosine, adjusted cosine and correlation as described in mark user-based algorithm. We present the results in Fig-
Section 3.1 and tested them on our data sets. For each simi- ure 6. It can be observed from the charts that the basic
291
Sensitivity of the parameter x Sensitivity of the Neighborhood Size
0.87
0.85
0.751
0.83
0.81
MAE
MAE
0.746
0.79
0.77 0.741
0.75
0.73 0.736
10
20
30
40
50
60
70
80
90
100
125
150
175
200
0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
Train/test ratio, x No. of Neighbors
item-item algorithm out performs the user based algorithm Sensitivity of the model size
at all values of x (neighborhood size = 30) and all values of (at selected train/test ratio)
MAE
0:726 respectively. The regression-based algorithm, how- 0.78
pared to the user-based algorithm. We also compared our x=0.3 x=0.5 x=0.8
292
Item-item vs. User-user at Selected Item-item vs. User-user at Selected
Neighborhood Sizes (at x=0.8) Density Levels (at No. of Nbr = 30)
0.755 0.84
0.75 0.82
0.745 0.8
MAE
MAE
0.74 0.78
0.735 0.76
0.73 0.74
0.725 0.72
10 20 60 90 125 200 0.2 0.5 0.8 0.9
No. of neighbors Train/test ratio, x
user-user item-item user-user item-item
item-item-regression nonpers item-item-regression nonpers
Figure 6: Comparison of prediction quality of item-item and user-user collaborative ltering algorithms. We
compare prediction qualities at x = 0:2; 0:5; 0:8 and 0:9.
4.4.1 Impact of the model size on run-time and through- line performance. Furthermore, due to the model-based ap-
put proach, it is possible to retain only a small subset of items
Given the quality of prediction is reasonably good with and produce reasonably good prediction quality. Our ex-
small model size, we focus on the run-time and throughput perimental results support that claim. Therefore, the item-
of the system. We recorded the time required to generate item scheme is capable in addressing the two most important
predictions for the entire test set and plotted them in a challenges of recommender systems for E-Commerce{quality
chart with varying model size. We plotted the run time at of prediction and high performance.
dierent x values. Figure 8 shows the plot. Note here that
at x = 0:25 the whole system has to make prediction for
25; 000 test cases. From the plot we observe a substantial 5. CONCLUSION
dierence in the run-time between the small model size and Recommender systems are a powerful new technology for
the full item-item prediction case. For x = 0:25 the run-time extracting additional value for a business from its user data-
is 2:002 seconds for a model size of 200 as opposed to 14:11 bases. These systems help users nd items they want to buy
for the basic item-item case. This dierence is even more from a business. Recommender systems benet users by en-
prominent with x = 0:8 where a model size of 200 requires abling them to nd items they like. Conversely, they help
only 1:292 seconds and the basic item-item case requires the business by generating more sales. Recommender sys-
36:34 seconds. tems are rapidly becoming a crucial tool in E-commerce on
These run-time numbers may be misleading as we com- the Web. Recommender systems are being stressed by the
puted them for dierent training/test ratios where the work- huge volume of user data in existing corporate databases,
load size, i.e., number of predictions to be generated is dif- and will be stressed even more by the increasing volume
ferent (recall that at x = 0:3 our algorithm uses 30; 000 of user data available on the Web. New technologies are
ratings as training data and uses the rest of 70; 000 ratings needed that can dramatically improve the scalability of rec-
as test data to compare predictions generated by the system ommender systems.
to the actual ratings). To make the numbers comparable we In this paper we presented and experimentally evaluated
compute the throughput (predictions generated per second) a new algorithm for CF-based recommender systems. Our
for the model based and basic item-item schemes. Figure 8 results show that item-based techniques hold the promise of
charts these results. We see that for x = 0:3 and at a model allowing CF-based algorithms to scale to large data sets and
size of 100 the system generates 70; 000 ratings in 1:487 sec- at the same time produce high-quality recommendations.
onds producing a throughput rate of 47; 361 where as the
basic item-item scheme produced a throughput of 4961 only.
At x = 0:8 these two numbers are 21; 505 and 550 respec- 6. ACKNOWLEDGMENTS
tively. Funding for this research was provided in part by the
National Science Foundation under grants IIS 9613960, IIS
4.5 Discussion 9734442, and IIS 9978717 with additional funding by Net
From the experimental evaluation of the item-item collab- Perceptions Inc. This work was also supported by NSF
orative ltering scheme we make some important observa- CCR-9972519, EIA-9986042, ACI-9982274 by Army Research
tions. First, the item-item scheme provides better quality OÆce contract DA/DAAG55-98-1-0441, by the DOE ASCI
of predictions than the use-user (k-nearest neighbor) scheme. program and by Army High Performance Computing Re-
The improvement in quality is consistent over dierent neigh- search Center contract number DAAH04-95-C-0008. Access
borhood size and training/test ratio. However, the improve- to computing facilities was provided by AHPCRC, Min-
ment is not signicantly large. The second observation is nesota Supercomputer Institute. We like to thank anony-
that the item neighborhood is fairly static, which can be mous reviewers for their valuable comments.
potentially pre-computed, which results in very high on-
293
Recommendation time vs. Model size Throughput vs. Model size
(at selected train/test ratio) (at selected train/test ratio)
35.75 100000
Throughput (recs./sec)
Rec. time (sec.)
30.75 80000
25.75
20.75 60000
15.75 40000
10.75
20000
5.75
0.75 0
25 50 75 100 125 150 175 200 item- 25 50 75 100 125 150 175 200 item-
item item
Model size Model size
x=0.3 x=0.5 x=0.8 x=0.3 x=0.5 x=0.8
Figure 8: Recommendation time and throughput comparison between model-based scheme and full item-item
scheme. The comparisons are shown at three dierent x values.
294
Filtering Agents to Improve Prediction Quality in the [27] Shardanand, U., and Maes, P. (1995). Social
GroupLens Research Collaborative Filtering System. Information Filtering: Algorithms for Automating
In Proceedings of CSCW '98, Seattle, WA. 'Word of Mouth'. In Proceedings of CHI '95. Denver,
[24] Sarwar, B. M., Karypis, G., Konstan, J. A., and Riedl, CO.
J. (2000). Application of Dimensionality Reduction in [28] Terveen, L., Hill, W., Amento, B., McDonald, D., and
Recommender System{A Case Study. In ACM Creter, J. (1997). PHOAKS: A System for Sharing
WebKDD 2000 Workshop. Recommendations. Communications of the ACM,
[25] Sarwar, B. M., Karypis, G., Konstan, J. A., and Riedl, 40(3). pp. 59-62.
J. (2000). Analysis of Recommendation Algorithms for [29] Ungar, L. H., and Foster, D. P. (1998) Clustering
E-Commerce. In Proceedings of the ACM EC'00 Methods for Collaborative Filtering. In Workshop on
Conference. Minneapolis, MN. pp. 158-167 Recommender Systems at the 15th National
[26] Schafer, J. B., Konstan, J., and Riedl, J. (1999). Conference on Articial Intelligence.
Recommender Systems in E-Commerce. In
Proceedings of ACM E-Commerce 1999 conference.
295