Sie sind auf Seite 1von 11

In data mining, association rule learning is a popular

and well researched method for discovering interesting relations between variables in large databases. Piatetsky-Shapiro [1] describes analyzing and presenting strong rules discovered in databases using different measures of interestingness. Based on the concept of strong rules, Agrawal et al. [2] introduced association rules for discovering regularities between products in large scale transaction data recorded by point-of-sale (POS) systems in supermarkets.

e.g., promotional pricing or product placements. In

addition to the above example from For example, the rule found in the sales data of a supermarket would indicate that if a customer buys onions and potatoes together, he or she is likely to also buy beef. Such information can be used as the basis for decisions about marketing activities such market basket analysis association rules are employed today in many application areas including Web usage mining, intrusion detection and bioinformatics.

Definition Following the original definition by Agrawal et al [2] the problem of association rule mining is defined as: Let

be a set of n binary attributes called items. Let be a set of transactions called the database. Each transaction in D has a unique transaction ID and contains a subset of the items in I. A rule is defined as an implication of the form where And

The sets of items (for short itemsets) X and Y are called

antecedent (left-hand-side or LHS) and consequent (right-hand-side or RHS) of the rule.

Example data base with 4 items and 5 transactions

Trans Id
1

Milk
1

Bread
1

Butter
0

Beer
0

To illustrate the concepts, we use a small example from

the supermarket domain. The set of items is I = {milk,bread,butter,beer} and a small database containing the items (1 codes presence and 0 absence of an item in a transaction) is shown in the table to the right. An example rule for the supermarket could be

meaning that if milk and bread is bought, customers also buy butter.

Note: this example is extremely small. In practical

applications, a rule needs a support of several hundred itemsets before it can be considered statistically significant, and datasets often contain thousands or millions of itemsets.
To select interesting rules from the set of all possible

rules, constraints on various measures of significance and interest can be used. The best-known constraints are minimum thresholds on support and confidence.

The support supp(X) of an itemset X is defined as the

proportion of transactions in the data set which contain the itemset. In the example database, the itemset {milk,bread} has a support of 2 / 5 = 0.4 since it occurs in 40% of all transactions (2 out of 5 transactions). The confidence of a rule is defined
For example, the rule has a confidence of 0.2 / 0.4 = 0.5 in the database, which means that for 50% of the transactions containing milk and bread the rule is correct.

Confidence can be interpreted as an estimate of the

probability P(Y | X), the probability of finding the RHS of the rule in transactions under the condition that these transactions also contain the LHS [3] The lift of a rule is defined as

or the ratio of the observed confidence to that expected by chance. The rule has a lift of

The conviction of a rule is defined as

The rule

has a conviction of

, and be interpreted as the ratio of the expected frequency that X occurs without Y if they were independent to the observed frequency.

Association rules are required to satisfy a user-specified

minimum support and a user-specified minimum confidence at the same time. To achieve this, association rule generation is a two-step process. First, minimum support is applied to find all frequent itemsets in a database. In a second step, these frequent itemsets and the minimum confidence constraint are used to form rules. While the second step is straight forward, the first step needs more attention.

Finding all frequent itemsets in a database is difficult

since it involves searching all possible itemsets (item combinations). The set of possible itemsets is the power set over I and has size 2n 1 (excluding the empty set which is not a valid itemset). Although the size of the powerset grows exponentially in the number of items n in I, efficient search is possible using the downward-closure property of support [2](also called anti-monotonicity[4]) which guarantees that for a frequent itemset also all its subsets are frequent and thus for an infrequent itemset, all its supersets must be infrequent. Exploiting this property, efficient algorithms (e.g., Apriori [5] and Eclat [6]) can find all frequent itemsets.

Das könnte Ihnen auch gefallen