Association rule mining Flashcards

(8 cards)

1
Q

Associative rule mining

A

is an unsupervised learning technique used to discover interesting relationships between items in large datasets, particularly in market basket analysis.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Mining Frequent Itemsets

A

Finding combinations of items that occur together frequently in transactions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

The Apriori Algorithm

A

Principle: If an itemset is frequent, all its subsets must also be frequent (anti-monotone property)

  • The support of an itemset never exceeds the support of its subsets
  • This is known as the anti-monotone property of support
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Apriori algorithm steps

A

How it works (Step-by-step):

Find frequent individual items

Count how often each item appears.

Keep only those that meet the minimum support threshold.

Generate larger itemsets

Combine frequent items to form 2-item sets, 3-item sets, and so on.

Keep only those that are still frequent (meet support threshold).

Stop when no larger frequent sets can be made.

Generate rules from frequent itemsets

For each frequent itemset, create rules like A → B.

Keep only the rules with high confidence (above the threshold).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Confidence

A

Confidence measures how often the rule is true. eg Given that someone buys beer, how often do they also buy wine?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Drawback of confidence

A

Confidence can be misleading because it doesn’t consider how common the right-hand side item already is. For example, even if P(Coffee | Tea) = 0.75, but P(Coffee) = 0.9, buying tea actually reduces the likelihood of buying coffee.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what are closed and maximal sets?

from these transactions which are closed and maximal

A

An itemset is closed if no proper superset has the same support.
An itemset is maximal if it is frequent and has no frequent superset.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

frequent?

A

Frequent = support ≥ min_sup

How well did you know this?
1
Not at all
2
3
4
5
Perfectly