dual hybrid

Upload: ezekill

Post on 03-Apr-2018

214 views

Category:

Documents


0 download

TRANSCRIPT

  • 7/29/2019 Dual Hybrid

    1/9

    30 1541-1672/07/$25.00 2007 IEEE IEEE INTELLIGENT SYSTEMSPublished by the IEEE Computer Society

    R e c o m m e n d e r S y s t e m s

    Collaborative FilteringUsing Dual InformationSourcesJinhyung Cho, Kwiseok Kwon, and Yongtae Park, Seoul National University

    With the proliferation of e-commerce, recommender systems have become an

    important component of personalized e-commerce services and are essential

    for e-commerce providers to remain competitive. One of the most successful recom-

    mendation techniques is collaborative filtering, whose performance has been proved

    in various e-commerce applications.1,2 CF auto-

    mates the word of mouth process.3 It forms a pre-

    dictor group that serves as an information source

    for recommendations.

    However, conventional CF methods suffer from a

    few fundamental limitations such as the cold-startproblem,data sparsity problem, and recommender reli-

    ability problem.4,5Thus, they have trouble dealing with

    high-involvement, knowledge-intensive domains such

    as e-learning video on demand. To overcome these

    problems, researchers have proposed recommendation

    techniques such as a hybrid approach combining CF

    with content-based filtering.4 Because e-commerce

    Web sites for e-learning often have various product

    categories, extracting the many attributes of these cat-

    egories for content-based filtering is extremely bur-

    densome. So, it might be practical to overcome these

    limitations by improving the CF method itself.

    Conventional CF methods base their recommen-

    dations on a single recommender group. Our CF

    method forms dual recommender groupsa similar-

    usersgroup and an expert-usersgroupas credible

    information sources. Then, it analyzes each groups

    influence on the target customers for the target prod-

    uct categories.

    Using this method, weve developed DISCORS

    (DualInformation Source Model-Based Collabora-

    tiveRecommender System) and applied it to a high-

    involvement product: e-learning VoD content. In

    experiments,DISCORS outperformed conventional CF

    methods in situations involving variations in the

    product domain and in data sparsity.

    CF from the consumerpsychology viewpoint

    When deciding what to purchase,consumers depend

    on a variety of information sources and have different

    acceptance levels for each source. Influencing factors

    can be the product domain characteristics, the con-sumers degree of involvement with the product, and

    the users level of knowledge about the product.6

    In the real world, a person making decisions about

    movies or daily necessities will seek the opinions of

    neighbors with similar preferences. On the other

    hand, when choosing expensive products or services

    for long-term use, such as a notebook computer or

    an educational program, an individuals decision is

    strongly influenced by people with professional

    expertise in that field. Customer preferences for rec-

    ommendation sources might also differ within a

    product domain. For example, when choosing a

    movie, some customers prefer neighbors opinions

    while others prefer expertsopinions.

    In this article, source diversity refers to the variety

    of information sources, and source receptivity refers

    to the level of a customers acceptance of a source.

    Source receptivity can differ across customers or the

    involvement level of product domainswe call this

    heterogeneous source receptivity. Product involve-

    mentrefers to the level of personal relevancethat is,

    the level of importance of a product or ones interest

    in it.7 High-involvement products are those for which

    the buyer is prepared to spend considerable time and

    effort in searching. Low-involvement products are

    bought frequently with a minimum of thought and

    Conventional

    collaborative-filtering

    methods use onlyone information

    source to provide

    recommendations.

    Using two sources

    similar users and

    expert users

    enables more effective,

    more adaptive

    recommendations.

  • 7/29/2019 Dual Hybrid

    2/9

    effort because they arent of vital concern and

    have no great impact on the consumers

    lifestyle. Unfortunately, existing CF methods

    dont consider source diversity, heterogeneous

    source receptivity, or product involvement.

    Similarity-based CFand its limitations

    In traditional CF methods, the single rec-

    ommender group comprises the nearest

    neighbors with preferences similar to those

    of a target user. So, these methods are called

    similarity-based CF.

    As we mentioned before,SCF methods suf-

    fer from the recommender reliability problem.

    That is, a recommender might not be reliable

    for a given item or set of items, even though the

    recommenders and target users preferences

    are similar.5 For example, when looking formovie recommendations, well often turn to

    our friends, on the basis that we have similar

    movie preferences overall. However, a partic-

    ular friend might not be reliable when it comes

    to recommending a particular type of movie.

    Trust-based CF and its limitationsTo solve the recommender reliability prob-

    lem, researchers have proposed trust-based

    CF.5,8,9 Such methods derive the neighbors

    trust explicitly or implicitly and use it as a

    supplementary criterion of similarity to select

    more credible neighbors.Each trust-based CF method employs a dif-

    ferent meaning of trust. For example, trust can

    imply trustworthinesshow much a user can

    trust other users in a trust network. Such trust-

    aware CF9 uses an explicitly rated trust value

    to select trustworthy users as a recommender

    group. In this way, it solves the recommender

    reliability problem. However, it doesnt

    account for source diversity or heterogeneous

    source receptivity.

    Second, trust can imply expertise or competencya users ability to

    make an accurate recommendation.5,8 In this case, CF can account for

    source diversity and recommender reliability by forming a recommender

    group based on both similarity and expertise. To do this, it uses the prod-

    uct or mean of the values for similarity and expertise. However, because

    this method equally weights similarity and expertise when combining

    them, without considering a variety of user or product domain charac-

    teristics, it doesnt account for heterogeneous source receptivity.

    Group influence and thedual-information-source model

    Most people belong to a number of different groups and perhaps

    would like to belong to several others. However,not all groups exert

    the same amount of influence on an individual. Sociologists use ref-

    erence group to refer to those groups that can modify or reinforce an

    individuals attitudes. Group behavior theory in consumer psychol-

    ogy has adopted the reference-group concept to consumer behavior.

    It holds that two reference groupssimilar users and experts

    strongly influence a consumers buying decision and that consumers

    perceive these groups as credible information sources.7

    Similarity- and trust-based methods view CF from a personal-influence

    perspective (see figure 1a). On the basis of group behavior theory, our

    method views CF from a group-influence perspective (see figure 1b).

    It builds information sources for the recommendations to a target user

    in accordance with two criteriasimilarity and expertise. As we men-

    tioned before, it employs dual recommender groupsa similar-user

    group and an expert-user groupas the information sources.

    This model overcomes the recommender reliability problem by

    using not just the similarity criterion but also the expertise criterion,

    and it accounts for source diversity by utilizing multiple recom-

    mender groups. In addition, it accounts for heterogeneous source

    receptivity by determining each groups level of influence on an indi-

    vidual user (source receptivity is the same concept as group influ-

    MAY/JUNE 2007 www.computer.org/intelligent 31

    Target user

    Target user

    Personal influence

    Similar users

    0.8

    1.2

    Our proposed CF (DISCORS)

    Group influence(source receptivity)

    with personal influence

    Similar users

    0.8

    0.7

    Similar & trustworthy users

    Similarity-based CF

    Trust-based CF

    0.1

    0.1

    0.2

    0.1

    0.9

    All users

    Personal influence

    Group influence

    0.90.1

    0.80.6 0.7

    0.5

    0.10.2

    0.10.9

    0.10.3

    0.20.2

    0.80.7

    0.70.3

    0.90.1

    0.3

    0.8

    0.10.9

    0.10.2

    0.20.2

    0.10.3

    Personal influence

    Similarity Trustworthiness

    Similarity Expertise Source receptivity

    All users

    All users

    Target user

    Expert users

    (a)

    (b)

    Figure 1. Two views of collaborative filtering: (a) Similarity-based CF and trust-basedCF take a personal-influence perspective. (b) DISCORS takes a group-influence

    perspective.

  • 7/29/2019 Dual Hybrid

    3/9

    ence, but from the users viewpoint). Consequently, a more person-

    alized recommendation is possible, taking into account variations in

    product domains and user tendencies.

    Redefining expertiseand trustworthiness

    According to the source credibility model, proposed in consumer

    psychology studies on word-of-mouth communication, an informationsources credibility comprises expertise,trustworthiness, similarity, and

    attraction.7 Here, expertise is the extent to which a source is perceived

    as being capable of providing correct information,while trustworthiness

    implies the degree to which a source is perceived as providing infor-

    mation reflecting that sources actual feelings or opinions.

    On the basis of this understanding of expertise, we define the

    expert-users group as users who have been carrying out a number

    of activities in a category that includes the target item so that they

    have a high probability of giving accurate recommendations to other

    users. We measure expertise by incorporating an appropriate mea-

    surement and various factors for the weights.

    DISCORSOwing to recent advances in multimedia and network technolo-

    gies, e-learning has become a promising alternative to traditional

    classroom learning. Web-based e-learning content services offer thou-

    sands of online courses. Currently,most e-learning content providers

    still offer all learners the same content, failing to satisfy individual

    learners. So, to provide more personalized content delivery, thereby

    increasing their competitiveness, they need to offer more relevant

    recommendation methods.

    Unfortunately, most recommendation methods focus on relatively

    low-involvement, entertainment product domains such as movies, cell

    phone wallpaper images, and music. So, we developed DISCORS as a

    viable alternative. Also, for high-involvement, knowledge-intensive

    product domains such as e-learning VoD, collecting sufficient explicit

    rating data from customers is difficult. To over-

    come this difficulty, DISCORS employs Web-

    usage mining to create users rating profiles

    from their implicit Web-usage behavior.

    The DISCORS recommendation process

    combines offline mining and online recom-mendation (see figure 2).

    Offline miningThis subprocess has three phases: create

    each users rating profile, form the dual infor-

    mation sources, and extract each users

    source receptivity.

    Creating user rating profiles. These profiles

    describe a users preference regarding each

    item by mining the Web-usage data collected

    in the e-learning VoD Web site. DISCORS con-

    structs each profile according to the three basicsteps of online VoD service use: click-through,

    preview, and payment. The relative frequency

    with which a user performs these steps for an

    item serves as an implicit preference rating;

    we assume that if the usage frequency for an

    item is relatively high, the user has a high preference for that item.

    We define the user rating profile,Ru,i, by modifying a previous

    approach10 to make it suitable for e-learning VoD content service:

    (1)

    whereRu,i is the rating profile of user u for item i, and m is the number of

    items. , ,and are the number of click-throughs, previews, and

    payments by a user for each item. The value ofRu,i is a sum of the nor-

    malized value of , , and . It ranges from 0 to 3, with a larger

    value indicating a stronger preference. It increases with the frequency

    of each stepclick-through,preview, and payment. Although each steps

    weights appear equal in equation 1, they arent, because customers who

    purchased specific content not only clicked the related Web pages but

    also previewed the content. So,Ru,i, which is used in the subsequent

    phases, is the normalized and weighted sum of , , and .

    Forming the dual information sources. In this phases first step,

    DISCORS selects users similar to a target user a and calculates their

    relative preference for a target item i.

    We define similar usersas a group of users with preference ratings

    similar to those the target user has had. To measure similarity, we

    employ Pearsons correlation coefficient, which is the most widely

    Ru ip

    ,Ru iv

    ,Ru ic

    ,

    Ru i

    p

    ,

    Ru i,

    Ru i

    c

    ,

    Ru ip

    ,Ru i,Ru i,

    R

    R R

    Ru i

    u ic

    i mu ic

    i mu ic,

    , ,

    ,

    min

    max=

    ( )

    ( )

    1

    1 1

    ( )

    + ( )

    i mu ic

    u iv

    i mu iv

    i m

    R

    R R

    min

    min

    m

    ,

    , ,1

    1aax min

    min

    , ,

    ,

    R R

    R R

    u iv

    i mu iv

    u ip

    i mu

    ( ) ( )

    +

    1

    1,,

    , ,max min

    ip

    i mu ip

    i mu ip

    R R

    ( )( ) ( )

    1 1

    R e c o m m e n d e r S y s t e m s

    32 www.computer.org/intelligent IEEE INTELLIGENT SYSTEMS

    Implicit

    Web usefeedback

    ExpertExpert users

    usersprediction

    predictionterm

    calculation

    Similarusersprediction

    termcalculation

    term

    Offline mining process Online recommendation process

    Phase 1 Phase 2 Phase 3

    Similarity-basedcollaborative

    filtering

    Recom-mendationgeneration

    Similarusers

    predictionterm

    Expertise-basedcollaborative

    filtering

    Dual-information-

    sourceformation

    Sourcereceptivityextraction

    Userratingprofile

    creation

    Userratingprofiledata-base

    Userratingprofile

    DB

    Webusagedata-base

    Target user

    Receptivitydatabase

    Figure 2. The DISCORS recommendation process combines offline mining and onlinerecommendation.

  • 7/29/2019 Dual Hybrid

    4/9

    used in conventional CF methods. We define the similarity between a

    target user a and another user u, s(u, a), as

    (2)

    whereRu,i is the rating of user u for item i, the item co-rated by two users.

    On the basis of previous research,1 we select the users that have a

    similarity threshold higher than 0.3 as the target users neighbors.

    We define theprediction term of similar users (neighbors), S(a, i),

    as the similarity-weighted sum of similar users relative preference

    compared to their average preference. We calculate this term as

    (3)

    whereRu,i is the rating of similar user u for item i, and n is the num-

    ber of users similar to target user a.

    In the second step, DISCORS selects the category experts and cal-

    culates their prediction term. As part of this step, we devised a

    measure of expertise reflecting the users activity and prediction

    competency. Expertise can be measured at the total-item level (a

    movie expert), category level (an action movie expert), or indi-

    vidual-item level (a Titanic expert).5 For recommendation,

    expertise measured in a more specific domain can be more pre-

    dictable. However, individual-item-level expertise isnt meaning-

    ful for real-world recommendations (Hes an expert on the movie

    Titanic. So what?).

    So, in this study, we measure expertise at the category level. Wedefine the expertise, e, of user u for category c as

    (4)

    where U(j) is the users who exhibited Web-usage behaviors for item

    j (except for target user u), C(i) is the item set that has the Web usages

    for the category of target item i,Nc(i) is the cardinality ofC(i) , and

    (u, c) is the activity weighting.

    We define (u, c) as 1 1/n (where n is the number of ratings in

    the category) to obtain a higher value of expertise for a user as that

    user rates more items in that category. We select the users who have

    expertise within the top 3 percent per category as expert users,because

    they showed the best recommendation results in terms of precision

    and coverage compared to other top-ranking user groups. We define

    the expert usersprediction term,E(i), as the expertise-weighted sum

    of expert users relative preference compared to their average prefer-

    ence. We calculate this term as

    (5)

    whereRu,i is the rating of expert user u for item i, and n is the num-

    ber of expert users in category c.

    Extracting source receptivity. The final phase extracts the hetero-

    geneous source receptivity for information sources with reference to

    the users and product domains. In other words, we build the follow-ing source receptivity model, under the assumption that each user

    demonstrates different susceptibility to the group influence:

    (6)

    Here, a and i denote a target customer and item number. ks andke are

    the importance weights ofS(a, i) andE(i)that is, a users receptiv-

    ity to the similar-user groups recommendation and the expert-user

    groups recommendation, respectively. We estimate these by multiple

    regression analysis using the least-squares method. If an overlap exists

    between similar and expert users, theres a high probability of multi-

    collinearity between S(a, i) andE(i). In that case, measuring each

    groups influence is difficult. So, when multicollinearity exists betweentwo variables (we consider it to exist if the variance inflation factor is

    greater than 10), we use ridge regression analysis to estimateks and ke.

    Online recommendationDISCORS generates personalized recommendations in real time by

    combining each users source receptivity values with each informa-

    tion sources prediction terms. Here we explain recommendation pro-

    cedures for an existing user, a new user, and a new item. Figure 3

    shows the pseudocode for our recommendation algorithm.

    Recommendation for an existing user. DISCORS gets the prediction

    terms after it forms each recommender group for the target user and

    item. Next, it finds the target users source receptivity, extracted dur-ing offline mining. Then, it determines the target users recommen-

    dation score by multiplying each sources prediction term and the

    users receptivity values:

    (7)

    If only one information source exists (for example, there are no

    similar users who rated the target item), DISCORS employs the single-

    source receptivity calculated with the existing information source

    only (see figure 3).

    Recommendation for a new user or a new item. Because finding sim-

    ilar users for a new user is impossible, DISCORS provides recommen-

    dations based on the expert users for the items category. For new

    that is, early-stageitems, we might not find similar users who rated

    the item, either. So, recommendations are once again based on those

    expert users. As time passes and Web use increases, DISCORS will apply

    the same recommendation procedure for existing users to these cases.

    Pilot system implementationOur pilot DISCORS system consists of six software agents and five data-

    bases (see figure 4a). Figure 4b shows the Web interface. The pilot sys-

    tem operates each agent independently so that the whole system remains

    stable during experimental substitutions or adjustment of an agent.

    Theuser profile creation agentcreates and manages user rating pro-

    files through offline Web-usage-mining tasks such as periodic collect-

    P a i R k S a i k E i Ca s eprediction , ,( ) = + ( ) + ( ) +

    R a i R k S a i k E i Ca s epast_rating , ,( ) = ( ) + ( ) +

    E iR R u c

    u c

    u i u eu

    n

    eu

    n( ) =

    ( ) ( )( )

    =

    =

    , ,

    ,

    1

    1

    e

    u j a ja U jj C i

    c i

    u c u c

    R R

    N, ,

    , ,

    ( ) = ( )

    ( ) ( )

    (

    1

    ))

    S a iR R u a

    u a

    u i u su

    n

    su

    n,

    ,

    ,

    ,

    ( ) =( ) ( )

    ( )=

    =

    1

    1

    su i u a i ai

    m

    u i u a

    u aR R R R

    R R R

    ,, ,

    ,

    ( ) =( ) ( )

    ( )

    = 12

    ,,i ai

    m

    i

    mR( )==

    2

    11

    MAY/JUNE 2007 www.computer.org/intelligent 33

  • 7/29/2019 Dual Hybrid

    5/9

    ing, parsing, and analyzing of real transaction data from the Web-usage

    database, customer database, and product database. It integrates the

    Web-usage data in a form suitable for the recommendation method.

    The SCF agentandECF agentactivate and manage the parts of our

    CF algorithm that calculate the similar-user groups and expert-user

    groups prediction terms, respectively. DISCORS uses these prediction

    terms to extract source receptivity and generate recommendations.

    The receptivity extraction agentextracts each users receptivity for

    the dual information sources. For this task, the agent analyzes users

    past rating profiles and the dual information sourcesprediction terms.

    The recommendation generation agent

    makes a personalized recommendation list

    for each target user according to the algo-

    rithm in figure 3. For each target user, it

    determines recommended products that

    reflect his or her source receptivity.Finally, the Web interface management

    agent provides user interfaces enabling

    Web-usage behaviors such as selecting a

    category or content, previewing e-learning

    content, and making electronic payments.

    Figure 4b shows the interface for present-

    ing recommendation lists.

    Evaluating DISCORS performanceWe wanted to answer these questions:

    How does DISCORS perform compared to

    CF methods based on a single-informa-tion-source model?

    How does the degree of product involve-

    ment affect the performance of DISCORS

    compared to that of CF methods based on

    a single-information-source model?

    How does data sparsity affect the perfor-

    mance of DISCORS compared to that of SCF

    methods?

    We compared DISCORS to three CF rec-

    ommender systems based on a single-infor-

    mation-source model. We used these bench-

    mark systems:

    SCF (similarity-based CF)a single-

    information-source model with one

    criterion,

    ECF (expertise-based CF)a single-

    information-source model with one

    criterion,

    HCF (hybrid CF with similarity weight-

    ing and expertise weighting)a single-

    information-source model with two criteria

    and homogeneous-source receptivity, and

    DISCORSa dual-information-source model

    with two criteria and heterogeneous-source

    receptivity.

    Evaluation metricsTo evaluate Discors performance,we employed two broad classes

    of recommendation accuracy metrics.

    The first ispredictive-accuracy metrics. Here, we use the mean

    absolute errorto compare each systems predictive accuracy. MAE

    is the absolute difference between a real and a predicted rating value.

    We use coverage, the number of items for which predictions can be

    formed as a percentage of the total number of items, to compare the

    range of recommendations for each system.

    The second class is classification accuracy metrics. To evaluate

    how well the recommendation lists match the users preferences, we

    R e c o m m e n d e r S y s t e m s

    34 www.computer.org/intelligent IEEE INTELLIGENT SYSTEMS

    Figure 3. Pseudocode for the recommendation algorithm.

    Algorithm : DISCORSInput :R: user rating matrix;k: source receptivity;

    OutputP(a, i) : Recommendation Score;

    Main()for all items i

    for all user uif (u= new user) then call NewUserRecomm();

    elseif (i= new item) then call NewItemRecomm();else call ExUserRecomm();

    endifendfor

    endfor

    end Main()

    ExUserRecomm()begincalculate prediction term of dual recommender groups S(a, i), E(i);if (both S(a, i) and E(i) exist) then select ks, ke, C;calculate P(a, i) = Avg(Ra) + ksS(a, i) + keE(i) + C;elseif (E(i) doesnt exist) then select kso, Cs; //kso is the single-source(similar-user group) receptivitycalculate P(a, i) = Avg(Ra) + ksoS(a, i) + Cs;elseif (S(a, i) doesnt exist) then select keo, Ce; //keois the single-source(expert-user group) receptivitycalculate P(a, i) = Avg(Ra) + keoE(i) + Ce;endifend ExUserRecomm( )

    NewUserRecomm( )begincalculate prediction term of expert-user group E(i);calculate P(a, i) = Avg(Ru) + E(i); //Avg(Ru) is the average of entire users ratingsend NewUserRecomm()

    NewItemRecomm()begincalculate prediction term of expert-user group E(i);select keo, Ce;calculate P(a, i) = Avg(Ra) + keoE(i) + Ce;end NewItemRecomm()

  • 7/29/2019 Dual Hybrid

    6/9

    MAY/JUNE 2007 www.computer.org/intelligent 35

    employ the widely usedprecision, recall,

    and F1 measures. If a user rates an item as

    being greater than 70 percent of the maxi-

    mum preference value (which is 3) or has

    purchased the item, we consider that the user

    prefers that item. DISCORS recommends anitem when that items predicted rating is

    greater than 70 percent of the maximum

    preference value. Precision refers to the

    number of recommended items that a user

    actually prefers, whereas recall refers to the

    number of preferred items that the system

    actually recommends. F1 is the harmonic

    mean of precision and recallthat is, (2 *

    precision * recall)/(precision + recall).

    A preliminary experimentwith research data

    Before we implemented DISCORS, we eval-uated its feasibility with a research data set

    that was open to the public. In the prelimi-

    nary experiment, we used the MovieLens

    data set consisting of approximately 1 mil-

    lion ratings involving 6,040 users and 3,900

    movies (www.grouplens.org/node/12). To

    evaluate each recommender system, we sep-

    arated this data set into two parts:

    a modeling set containing the 6,040 users

    ratings of 3,000 movies and

    a validation set containing those usersrat-

    ings of the remaining 900 movies.

    Table 1 shows the results. The MAE of

    DISCORS is approximately 4.5 percent lower

    than that of SCF, 8.7 percent lower than that

    of ECF, and 5.8 percent lower than that of

    HCF, at a significance level of 1 percent.

    Although the performance gain of DISCORS

    over SCF isnt high, it does indicate our sys-

    tems superiority, and we expect the gain to

    increase as product involvement or data spar-

    sity increases. The coverage of DISCORS

    exceeds that of SCF by 9.30 percent and that

    of HCF by 2.90 percent.

    Experiments with realWeb-usage data

    Paran.com (www.paran.com) is a Web

    portal operated by Korea Telecom Hitel, a

    subsidiary of Korea Telecom. This site,

    which has approximately 16 million sub-

    scribers and 8 million unique visitors per

    week, is a major Korean digital-content provider. Paran.com pro-

    vided us with the Web-usage data and purchasing data pertaining

    to e-learning content for foreign languages and to digital-comics

    content (digitalized comic books), logged from 1 January to 30

    June 2006. The e-learning content comprises various English,

    Japanese, and Chinese categories and provides 192 items. The dig-

    ital-comics content comprises 456 items in eight categories (such

    as action and drama).

    Through data preparation and Web-usage mining, we obtained

    14,731 ratings of 1,452 users for 126 items in the e-learning content

    Receptivityextraction

    agent

    User ratingprofile

    database

    User profile creation agent

    (b)

    (a)

    Web-usagedatabase

    Productdatabase

    Customerdatabase

    SCF agent(similarity-based CF)

    Similarity calculation

    Similar-user-group formation

    Similar usersprediction term calculation

    Recommendation generation agent

    Web interface management agent

    ECF agent(expertise-based CF)

    Expertise calculation

    Expert-user-group formation

    Expert usersprediction term calculationReceptivity

    database

    Figure 4. The implementation of the pilot DISCORS system: (a) the system architectureand (b) the Web interface for presenting recommendation lists.

  • 7/29/2019 Dual Hybrid

    7/9

    and 373,514 ratings of 11,245 users for 318 items in the digital-

    comics content, standardized from 0 to 3. We divided each data set

    into a modeling set and a validation set. The modeling set contained

    randomly selected items amounting to 80 percent of the total items;

    the validation set contained the remaining items.

    The e-learning content is more expensive (US$44 to $50 per lec-ture) than the digital-comics content ($0.1 to $0.3 per volume). Also,

    typical usage of e-learning content lasts more than one monthmuch

    longer than for digital comics. So, we classified the e-learning con-

    tent as a relatively high-involvement product and the digital-comics

    content as a relatively low-involvement product. We then compared

    the systems performance on the basis of product involvement.

    Classification accuracy with product involvement. The rating-pro-

    file data we used in this experiment didnt come directly from users;

    we inferred the data through Web-usage and purchasing results. So,

    we compared the systemsperformance by measuring classification

    accuracy. We expected that DISCORS would perform better than sin-

    gle-information-source CF because our system considers sourcediversity and heterogeneous source receptivity. Also, because con-

    sumers will more likely listen to experts opinions as their product

    involvement increases,we assumed that DISCORS would perform bet-

    ter for high-involvement products.

    Table 2 shows the results. DISCORS outperformed SCF by 26.0 per-

    cent for e-learning content and 10.34 percent for digital-comics con-

    tent, with F1 values at a significance level of 1 percent. Furthermore,the performance gain of DISCORS over SCF was significantly higher

    for e-learning than for digital comics. This supports our hypothesis

    that DISCORS performs even better as product involvement increases.

    We initially expected that DISCORS would perform worse than ECF

    for e-learning content because consumer reliance on experts tends

    to increase as product involvement increases. Contrary to our expec-

    tation, DISCORS outperformed ECF by 16.33 percent for e-learning

    content and 13.08 percent for digital-comics content. However, the

    difference isnt statistically significant.

    DISCORS also outperformed HCF by 7.18 percent for e-learning

    content and 9.83 percent for digital-comics content. This result sup-

    ports our assumption that a dual-information-source model can out-

    perform a single-information-source model. However, the differencein the performance gains across products isnt significant.

    The effects of sparsity. CF methodsperfor-

    mance depends on the availability of a criti-

    cal mass of ratings. Conventional CF meth-

    ods exhibit the data-sparsity problem; that is,

    the recommendation quality decreases sud-

    denly as data sparsity increases. We assumed

    that a CF method using a dual-information-

    source model will perform well even with

    data sparsity. To prove this assumption, we

    compared the performance of DISCORS to SCF

    for various levels of data density.The original data density levels were 1

    14,731/(1,452 * 126) = 0.9195 for e-learning

    R e c o m m e n d e r S y s t e m s

    36 www.computer.org/intelligent IEEE INTELLIGENT SYSTEMS

    Table 1. The predictive accuracy of DISCORS

    and three benchmark collaborative-filtering systems.

    System Mean absolute error Coverage (%) t-value (p < 0.01)

    DISCORS 0.6923 98.96

    SCF (similarity- 0.7250 90.52 6.952based CF) (4.5%)* (9.30%)*

    ECF (expertise- 0.7584 85.71 24.075based CF) (8.7%)* (15.50%)*

    HCF (hybrid CF) 0.7348 96.13 21.534

    (5.8%)* (2.90%)*

    *The figures in parentheses indicate the performance gain of DISCORS over that benchmark system.

    Table 2. The classification accuracy of DISCORS and three benchmark systems,for high (e-learning) and low (digital-comics) product involvement.

    Precision Recall F1

    System E-learning Comics E-learning Comics E-learning Comics

    DISCORS 0.3258 0.3363 0.2088 0.3187 0.2545 0.3272

    SCF 0.2257 0.2696 0.1828 0.3295 0.2020 0.2966(44.35%)* (24.74%)* (14.23%)* (3.30%)* (26.00%)* (10.34%)*

    t-value between DISCORS and SCF 82.563 63.452 8.875 4.512 8.924 4.572

    t-value between domains 8.276 3.548 4.853

    ECF 0.2972 0.2677 0.1731 0.3149 0.2188 0.2894(9.62%)* (25.63%)* (20.62%)* (1.20%)* (16.33%)* (13.08%)*

    t-value between DISCORS and ECF 2.872 3.548 15.602 4.669 12.245 9.642

    t-value between domains 3.300 4.887 0.085

    HCF 0.2804 0.2887 0.2059 0.3078 0.2374 0.2979(16.19%)* (16.49%)* (1.41%)* (3.53%)* (7.18%)* (9.83%)*

    t-value between DISCORS and HCF 12.533 13.234 0.072 8.187 10.669 6.715

    t-value between domains 0.069 0.877 0.819

    *The figures in parentheses indicate the performance gain of DISCORS over that benchmark system.p < 0.01

  • 7/29/2019 Dual Hybrid

    8/9

    content and 1 373,514/(11,245 * 318) = 0.8955 for digital comics.

    We obtained seven different density levels as follows. After dividing

    the data sets into modeling and validation portions, we retained 100

    percent, 87.5 percent,75 percent, 62.5 percent, 50 percent, 37.5 per-

    cent, and 25 percent of the nonzero entries in the modeling set, by ran-

    domly removing nonzero entries.Density affects the F1 values for DISCORS

    and SCF and the performance gain of DISCORS

    for both e-learning and digital-comics con-

    tent (see figure 5). For e-learning, as the den-

    sity decreases, the performance gain increases

    from 26.0 percent to 36.4 percent (F = 2.956,

    p < 0.01). For digital comics, as the density

    decreases, the performance gain increases

    from 10.3 percent to 32.0 percent (F = 4.275,

    p < 0.01). The F statistic provides a test for

    the statistical significance of the difference in

    the observed DISCORS performance gain over

    sparsity levels through ANOVA analysis.These results have three interesting impli-

    cations. First, the lower the data density (the

    higher the data sparsity), the better the perfor-

    mance gain of DISCORS relative to SCF. This

    implies that DISCORS helps mitigate the data-

    sparsity problem regardless of product domain.

    Second, for relatively low product involve-

    ment, the performance gain of DISCORS rela-

    tive to SCF is more sensitive to data density.

    This implies that for high-involvement prod-

    ucts, sparsity doesnt affect SCF; however,this

    isnt the case for low-involvement products.

    Finally, ECF, a component of DISCORS, is moreeffective for high-involvement products. This

    supports our assumption that consumers tend

    to be more receptive to experts opinions as

    product involvement increases.

    Visualizing source receptivity. We used a

    visualization to analyze the users source

    receptivity with variations in the product

    domain. In figure 6, thex-axis represents receptivity to similar

    users recommendations (ks), and they-axis represents receptivity

    to expert usersrecommendations (ke). For a low-involvement prod-

    uct (digital comics), most users have a low dependency on experts,

    as we expected. The centroid of the users segment for that product

    MAY/JUNE 2007 www.computer.org/intelligent 37

    (b)

    0.0000

    0.0500

    0.1000

    0.1500

    0.2000

    0.2500

    0.3000

    F1

    0.0000

    0.0500

    0.1000

    0.1500

    0.2000

    0.2500

    0.3000

    0.3500

    100 88 75 63 50 38 25

    100 88 75 63 50 38 25

    Data density (% of original)

    F1

    0.0

    5.0

    10.0

    15.0

    20.0

    25.0

    30.0

    35.0

    40.0

    DISCORS

    performancega

    in(%)

    DISCORS (D) SCF (S) (D S)/S

    DISCORS (D) SCF (S) (D S)/S

    (a) Data density (% of original)

    0.0

    5.0

    10.0

    15.0

    20.0

    25.0

    30.0

    35.0

    40.0

    DISCORSperformancega

    in(%)

    Figure 5. How data sparsity affects DISCORS and SCF for (a) a high-involvement product(e-learning) and (b) a low-involvement product (digital comics).

    (a) (b)

    Receptiv

    itytoexpertuse

    rs

    recommen

    dation

    (ke)

    Receptivity to similar users recommendation (ks) Receptivity to similar users recommendation (ks)

    3.02.52.01.51.00.50.5 0

    3.0

    2.5

    2.0

    1.5

    1.0

    0.5

    0

    0.5

    1.0

    1.5

    2.0

    Receptiv

    itytoexpertuse

    rs

    recommen

    dation

    (ke)

    3.0

    2.5

    2.0

    1.5

    1.0

    0.5

    0

    0.5

    1.0

    1.5

    2.02.52.01.51.00.50.5 0

    Figure 6. Users source receptivity for (a) a high-involvement product (e-learning, 1,452 users) and (b) a low-involvement product

    (digital comics, 11,245 users).

  • 7/29/2019 Dual Hybrid

    9/9

    R e c o m m e n d e r S y s t e m s

    38 www.computer.org/intelligent IEEE INTELLIGENT SYSTEMS

    is (1.126, 0.239). However, for a high-involvement product (e-learn-

    ing), many users exhibit expert-user dependencies; their centroid is

    (0.8034, 0.5337).

    We observed that you can classify users into expert-dependent

    users and neighbor-dependent users with respect to product

    domains. Accordingly, marketing staff in an e-commerce company

    can identify the most effective information source on the basis of

    the characteristics of individuals and product domains. Conse-

    quently, expert-dependent users could receive Web or mobile con-

    tent that reflects expert users recommendations, even for low-

    involvement products. Similarly, neighbor-dependent users could

    receive neighbors recommendations, even for high-involvement

    products. This strategy will enable more effective and more adap-

    tive personalized marketing.

    Because our results are based on data of a particular e-commercesite and a specific research data set, we need to evaluate DIS-CORS with data sets from various e-commerce product domains. In

    addition, we need to devise a more refined technique for analyzing

    Web usage that can automatically extract both user preference and

    user credibility. Also, it would be interesting to expand DISCORS to

    other challenging e-commerce domains or environments that require

    a recommendation method. Although we implemented DISCORS for

    providing e-learning services in this study, we believe its generallyapplicable to a variety of e-commerce recommender systems.

    AcknowledgmentsWe thank Korea Telecom Hitel Paran.com for providing us with the

    Web usage data used in this research, and Jeeyoung Yoon for his researchassistance.

    References1. J.L. Herlocker et al., Evaluating Collaborative Filtering Recommender

    Systems,ACM Trans. Information Systems, vol. 22,no. 1, 2004,pp. 553.

    2. B. Sarwar et al., Analysis of Recommendation Algorithms for E-com-merce, Proc. 2nd ACM Conf. Electronic Commerce (EC 00), ACMPress, 2000, pp. 158167.

    3. U. Shardanand and P. Maes, Social Information Filtering: Algorithmsfor Automating Word of Mouth, Proc. Human Factors in ComputingSystems Conf. (CHI 95), ACM Press, 1995, pp. 210217.

    4. G. Adomavicius and A. Tuzhilin, Toward the Next Generation of Rec-ommender Systems:A Survey of the State-of-the-Art and Possible Exten-sions,IEEE Trans. Knowledge and Data Eng., vol. 17, no. 6, 2005, pp.

    734749.

    5. J. ODonovan and B. Smyth, Trust in Recommender Systems, Proc.10th Intl Conf. Intelligent User Interfaces (IUI 05), ACM Press, 2005,pp. 167174.

    6. D.F. Duhan et al., Influences on Consumer Use of Word-of-Mouth Rec-ommendation Sources,J. Academy of Marketing Science, vol. 25, Fall1997, pp. 283295.

    7. T.S. Robertson, J. Zielinski, and S. Ward, Consumer Behavior, Scott,Foresman and Co., 1984.

    8. T. Riggs and R. Wilensky, An Algorithm for Automated Rating ofReviewers, Proc. 1st ACM/IEEE-CS Joint Conf. Digital Libraries(JCDL 01), ACM Press, 2001, pp. 381387.

    9. P. Massa and P. Avesani, Trust-Aware Collaborative Filtering for Rec-ommender Systems, On the Move to Meaningful Internet Systems 2004:CoopIS, DOA, and ODBASE, LNCS 3290, Springer, 2004, pp. 492508.

    10. Y.H. Cho, J.K. Kim, and S.H. Kim, A Personalized Recommender Sys-tem Based on Web Usage Mining and Decision Tree Induction,ExpertSystems with Applications, vol. 23, no. 3, 2002, pp. 329342.

    For more information on this or any other computing topic, please visit ourDigital Library at www.computer.org/publications/dlib.

    T h e A u t h o r s

    Jinhyung Cho is an assistant professor in theDongyang Technical Colleges Department of

    Computer and Information Engineering and aPhD candidate in the Seoul National UniversitysInterdisciplinary Graduate Program of Technol-ogy and Management. His research interests in-clude Web personalization, e-business, socialcomputing, and knowledge management systems.He received his MS in computer engineering from

    the Korea Advanced Institute of Science and Technology. Contact him atthe Dept. of Computer and Information Eng., Dongyang Technical Col-lege, 62-160 Kochuk-Dong, Kuro-Gu, Seoul, 152-714, Korea; [email protected].

    Kwiseok Kwon is an assistant professor in theAnyang Technical Colleges Department of E-business and a PhD candidate in the SeoulNational Universitys Interdisciplinary Graduate

    Program of Technology and Management. Hisresearch interests include Web personalization,new-service development, and the Semantic Web.He received his MS from the InterdisciplinaryGraduate Program of Technology and Manage-

    ment. Contact him at the Dept. of E-business, Anyang Technical College,San 39-1,Anyang 3-Dong, Manan-Gu,Anyang, Gyeonggi-Do, 430-749,Korea; [email protected].

    Yongtae Park is a professor in the Seoul NationalUniversitys Department of Industrial Engineeringand served as the director of SNUs Interdiscipli-nary Graduate Program of Technology and Man-agement. His research interests include knowl-edge network analysis and online-service creation.He received his PhD in operations management

    from the University of Wisconsin-Madison. Con-tact him at the Dept. of Industrial Eng., Seoul

    National Univ., San 56-1, Shillim-Dong, Kwanak-Gu, Seoul, 151-742,Korea; [email protected].