Performance Comparison of PSO and Hybrid PSO-GA in Hiding Fuzzy Sensitive Association Rules

Performance Comparison of PSO and Hybrid PSO-GA in Hiding Fuzzy Sensitive Association Rules

Sathiyapriya Krishnamoorthy (PSG College of Technology, India), Sudha Sadasivam G. (PSG College of Technology, India) and Rajalakshmi M. (Coimbatore Institute of Technology, India)
DOI: 10.4018/978-1-5225-5396-0.ch009
OnDemand PDF Download:
No Current Special Offers


Explosion of data analysis techniques facilitate organizations to publish microdata about individuals. While the released data sets provide valuable information to researchers, it is possible to infer sensitive information from the published non-sensitive data using association rule mining. An association rule is characterized as sensitive if its confidence is above disclosure threshold. These sensitive rules should be made uninteresting before releasing the dataset publicly. This is done by modifying the data that support the sensitive rules, so that the confidence of these sensitive rules is reduced below disclosure threshold. The main goal of the proposed system is to hide a set of sensitive association rules by perturbing the quantitative data that contains sensitive knowledge using PSO and hybrid PSO-GA with minimum side effects like lost rules, ghost rules. The performance of PSO and Hybrid PSO-GA approach in effectively hiding fuzzy association rule is also compared. Experimental results demonstrate that hybrid approach is efficient in terms of lost rules, number of modifications, hiding failure.
Chapter Preview

1. Introduction

Data or knowledge mining aims at discovering previously unknown knowledge from large collection of data items. Association rule mining is a popular technique in data mining to find frequent patterns, associations, correlations among set of items or objects in transactional databases. An association rule is an implication of the form X→Y, where both X and Y are set of attributes (items) from the database. X is called as the body (Left Hand Side (LHS)) of the rule and Y is called as the head (Right Hand Side (RHS)) of the rule. For example, an association rule in a market data may be defined as, Pen=> ink (support = 50%, confidence =65%). In 50% of the transactions, 65% of the people buying pen also buy ink in the same transaction; 50% and 65% represent the support and the confidence, respectively. Support is the percentage of number of transactions that contain both X and Y. Confidence is the ratio of the support of XUY to the support of X (Sathiyapriya Sudhasadasivam & Suganya 2014).

The Apriori algorithm was the common algorithm that uses the concept of frequent item sets to mine association rules in transaction data (Srikant & Agrawal,1995). The Apriori algorithm generates large number of candidate item set. The Frequent-Pattern-tree structure (FP-tree) efficiently mines association rules without generation of candidate item sets (Han & Fu, 1995).

But most databases in real world contains numerical, categorical and integer values. The binary algorithm cannot be applied directly. One way of mining quantitative rules is to treat them like categorical attributes and generate rules for all potential values. This would result in the explosion of number of rules generated and also specific numerical value will not appear frequently. So the domain of each quantitative attribute is divided into intervals and rules are generated from these intervals. This is called discretization (Srikant & Agrawal, 1996). Choosing intervals for numeric attributes is sensitive to support and confidence measures.

As it is possible that the data set may be skewed, intervals cannot be formed randomly. It was shown that if the range of the attribute is divided into equal intervals it leads to two problems of minsupport and minconfidence (Srikant & Agrawal, 1996). If large number of small intervals is chosen then the support of some of the interval becomes low. This is called “Minsupport” issue. Building larger intervals results in information loss and rules mined are different from that in original data. This is called “Minconfidence” issue. Another problem with discretization is sharp boundary problem. For example, consider the rule, if experience >=2 and bonus points >= 5000 then credit = approved. If a customer has a job for two years and bonus point of 4,990 then the application for credit approval may be rejected. Such a precise cut-off seems unfair. So, in order to avoid the sharp boundary problem the data is fuzzified.

In this example, the bonus point can be discretized into categories like low, medium, high and fuzzy logic can be applied to allow fuzzy threshold or boundaries to be defined for each category. Unlike the crisp sets where an element either belongs to a set S or its complement, in fuzzy set theory, each element can belong to more than one fuzzy set. In this example, the bonus point 4,990 belongs to both medium and high fuzzy sets, but to different degrees. In order to avoid sharp boundary problem numerous fuzzy mining approaches have been proposed to mine interesting association rules from quantitative values.

Complete Chapter List

Search this Book: