Famous (and fake) example:
People who buy more diapers buy more beer
If person X buys diapers,
Person X buys beer
Conclusion: put expensive beer next to the diapers
Association rules imply causality by their if-then nature
But causality can go either direction
If person X buys diapers, and person X is male, and it is after 7pm, then person Y buys beer and tortilla chips and salsa
Can be harder to use, sometimes eliminated from consideration
Generating hypotheses to study further
Finding unexpected connections
Is there a surprisingly ineffective instructor or math problem?
Are there e-learning resources that tend to be selected together?
Find rules
Evaluate rules
Find rules
Evaluate rules
Support/Coverage
Confidence
“Interestingness”
Number of data points that fit the rule, divided by the total number of data points
(Variant: just the number of data points that fit the rule)
Rule:If a student took Advanced Data Mining, the student took Intro Statistics
Support/coverage?

Rule:If a student took Advanced Data Mining, the student took Intro Statistics
Support/coverage?
2/11= 0.1818

Number of data points that fit the rule, divided by the number of data points that fit the rule’s IF condition
Equivalent to precision in classification
Also referred to as accuracy, just to make things confusing
NOT equivalent to accuracy in classification
Rule:If a student took Advanced Data Mining, the student took Intro Statistics
Confidence?

Rule:If a student took Advanced Data Mining, the student took Intro Statistics
Confidence?
2/6 = 0.33

Implementations of Association Rule Mining sometimes differ based on whether the values for support and confidence (and other metrics)
Are calculated based on exact cases
Or some other grouping variable (sometimes called “customer” in specific packages)
Let’s say you are looking at whether boredom follows frustration
If Frustrated at time N,Then Bored at time N+1

If you just calculate it this way,
Confidence = 4/5

But if you treat student as your “customer” grouping variable
Then whole rule applies for A, C And IF applies for A, C
So confidence = 1

The association rule mining community differs from most other methodological communities by acknowledging that cut-offs for support and confidence are arbitrary
Researchers typically adjust them to find a desirable number of rules to investigate, ordering from best-to-worst…
Rather than arbitrarily saying that all rules over a certain cut-off are “good”
Support and confidence aren’t enough
Why not?
Possible to generate large numbers of trivial associations
Students who took a course took its prerequisites (AUTHORS REDACTED, 2009)
Students who do poorly on the exams fail the course (AUTHOR REDACTED, 2009)
Not quite what it sounds like
Typically defined as measures other than support and confidence
Rather than an actual measure of the novelty or usefulness of the discovery
Cosine
\[\frac{P(A^{\wedge}B)}{\sqrt{P(A)*P(B))}}\]
Measures co-occurrence
Merceron & Yacef (2008) note that it is easy to interpret (numbers closer to 1 than 0 are better; over 0.65 is desirable)
Rule:If a student took BDEMOOC, the student published EDM paper
Cosine? \[\frac{P(A,B)}{\sqrt{P(A)*P(B))}}\]
(0.4)/sqrt(0.5*0.5) = 0.8

Lift\[ \frac{Confidence(A\to B)}{P(B)} = \frac{P(A^{\wedge}B )}{P(A)*P(B)} \]
Measures whether data points that have both A and B are more common than would be expected from the base rate of each
Merceron & Yacef (2008) note that it is easy to interpret (lift over 1 indicates stronger association)
Rule:If a student took BDEMOOC, the student published EDM paper
Lift?
\[ \frac{P(A^{\wedge}B )}{P(A)*P(B)} \]
(0.4)/(0.5*0.5) = 1.6

(Tan, Kumar, & Srivastava, 2002)
Jaccard
\[ \frac{P(A^{\wedge}B)}{P(A)+P(B)- P(A^{\wedge}B)} \]
Measures the relative degree to which having A and B together is more likely than having either A or B but not both
Compared “interestingness” measures to human judgments about how interesting the rules were
They found that Jaccard and Cosine were the best single predictors
And that Lift had predictive power independent of them
But they also found that the correlations between [Jaccard and Cosine] and [human ratings of interestingness] were negative
Find rules
Evaluate rules
Generate frequent itemset
Generate rules from frequent itemset
Generate all single items, take those with support over threshold – {i1}
Generate all pairs of items from items in {i1}, take those with support over threshold – {i2}
Generate all triplets of items from items in {i2}, take those with support over threshold – {i3}
And so on…
Then form joint itemset of all itemsets
Given a frequent itemset, take all items with at least two components
Generate rules from these items
Eliminate rules with confidence below threshold
Negative association rules (Brin et al., 1997)
What doesn’t go together?(especially if probability suggests that two things should go together)
People who buy diapers don’t buy car wax, even though 30-year old males buy both?
People who take advanced data mining don’t take hierarchical linear models, even though everyone who takes either has advanced math?
Students who game the system don’t go off-task?
What might be some reasonable applications for Association Rule Mining in education?
Can you brainstorm some ways you might use them in your own work?