Finding Non-Coincidental Sporadic Rules Using Apriori-Inverse

Finding Non-Coincidental Sporadic Rules Using Apriori-Inverse

Yun Sing Koh (University of Otago, New Zealand), Nathan Rountree (University of Otago, New Zealand) and Richard O’Keefe (University of Otago, New Zealand)
Copyright: © 2006 |Pages: 17
DOI: 10.4018/jdwm.2006040102
OnDemand PDF Download:
$37.50

Abstract

Discovering association rules efficiently is an important data mining problem. We define sporadic rules as those with low support but high confidence; for example, a rare association of two symptoms indicating a rare disease. To find such rules using the well-known Apriori algorithm, minimum support has to be set very low, producing a large number of trivial frequent itemsets. To alleviate this problem, we propose a new method of discovering sporadic rules without having to produce all other rules above the minimum support threshold. The new method, called Apriori-Inverse, is a variation of the Apriori algorithm that uses the notion of maximum support instead of minimum support to generate candidate itemsets. Candidate itemsets of interest to us fall below a maximum support value but above a minimum absolute support value. Rules above maximum support are considered frequent rules, which are of no interest to us, whereas rules that occur by chance fall below the minimum absolute support value. We define two classes of sporadic rule: perfectly sporadic rules (those that consist only of items falling below maximum support) and imperfectly sporadic rules (those that may contain items over the maximum support threshold). This article is an expanded version of Koh and Rountree (2005).

Complete Article List

Search this Journal:
Reset
Open Access Articles: Forthcoming
Volume 13: 4 Issues (2017): Forthcoming, Available for Pre-Order
Volume 12: 4 Issues (2016)
Volume 11: 4 Issues (2015)
Volume 10: 4 Issues (2014)
Volume 9: 4 Issues (2013)
Volume 8: 4 Issues (2012)
Volume 7: 4 Issues (2011)
Volume 6: 4 Issues (2010)
Volume 5: 4 Issues (2009)
Volume 4: 4 Issues (2008)
Volume 3: 4 Issues (2007)
Volume 2: 4 Issues (2006)
Volume 1: 4 Issues (2005)
View Complete Journal Contents Listing