A Supervised Explainable Machine Learning Algorithm for Generating High-Precision Decision Rules
- ABC Rule Miner is a supervised rule-based learning algorithm designed to generate interpretable IF–THEN rules for predicting a target behavioral label.
- This algorithm was introduced in the paper : https://www.sciencedirect.com/science/article/pii/S1084804520302368
- Unlike Apriori (unsupervised), ABC Rule Miner uses a target column (e.g., Recommendation_Label) to grow branches in a tree-like structure and extract behavior-focused rules.
Major contributions of the paper :
- Identify redundancy in associations while producing user behavioral rules based on multi-dimensional contexts.
- A rule based ML method (Association Generation Tree) to discover non redundant behavioral associations.
- Conducting range of experiments on real smart phone dataset, comparing results of ABC rule miner approach with traditional methods ( Apriori).
- Understand the working of ABC Rule Miner.
- Implement ABC Rule Miner and Apriori on an E-learning dataset and understand and compare the results.
This project implements:
- ABC rule generation
- Rule pruning
- Confidence filtering
- Coverage evaluation
- Benchmarking against Apriori
- Visualizations (rule count vs. confidence threshold)
git clone https://github.com/Joyee2004/ABC-Rule-Miner.git
cd ABC-Rule-MinerNow in the experiments directory, run each file (.ipynb)/ code snippet.
- Contextual E-Learning Learner Interaction Dataset.
- It contains data from e-learning platforms, covering usage details.
- Dataset link : https://www.kaggle.com/datasets/zoya77/contextual-e-learning-learner-interaction-dataset
1. Input:
i. Training dataset
ii. Target behavior column
iii. Confidence threshold t
2. Recursive Branch Growth:
i. At each step, select an attribute that best separates data toward the target label.
ii. Expand branches similar to a decision tree.
3. Rule Extraction: Convert each path from root → leaf into an IF–THEN rule.
4. Pruning:
i. Remove rules below confidence threshold t.
ii. Remove redundant rule subsets.
5. Evaluation:
i. Apply rules on test data.
ii. Compute Accuracy, Precision, Recall, F1.
Observations:
From confidence 0.1 to 0.5 , there is no change in result
From 0.6 to 0.8 , performance increases.
From 0.8 to 0.9, performance decreases.
| t | Accuracy | Precision | Recall | F1 |
|---|---|---|---|---|
| 0.1 | 0.631 | 0.631 | 1 | 0.774 |
| 0.2 | 0.631 | 0.631 | 1 | 0.774 |
| 0.3 | 0.631 | 0.631 | 1 | 0.774 |
| 0.4 | 0.631 | 0.631 | 1 | 0.774 |
| 0.5 | 0.631 | 0.631 | 1 | 0.774 |
| 0.6 | 0.943 | 0.917 | 1 | 0.957 |
| 0.7 | 0.943 | 0.917 | 1 | 0.957 |
| 0.8 | 0.943 | 0.917 | 1 | 0.957 |
| 0.9 | 0.887 | 0.93 | 0.888 | 0.909 |
Observations:
1. Number of rules generated:
Apriori : increases with decrease in confidence.
ABC Rule Miner : increases with increase in confidence.
From 0.1 to 0.5, performance of both are similar, but ABC generates lesser rules.
2. Apriori generates redundant rules.
1. From confidence threshold 0.1 to 0.5, ABC Rule miner gives same result as that of Apriori but with lesser number of rules.
2. For ABC rule Miner, an optimal value of t is 0.6 where it outperforms apriori.
ABC Rule Miner eliminates redundant rules.
3. A tradeoff between number of rules and performance exists.