fbpx
Wikipedia

Rough set

In computer science, a rough set, first described by Polish computer scientist Zdzisław I. Pawlak, is a formal approximation of a crisp set (i.e., conventional set) in terms of a pair of sets which give the lower and the upper approximation of the original set. In the standard version of rough set theory (Pawlak 1991), the lower- and upper-approximation sets are crisp sets, but in other variations, the approximating sets may be fuzzy sets.

Definitions

The following section contains an overview of the basic framework of rough set theory, as originally proposed by Zdzisław I. Pawlak, along with some of the key definitions. More formal properties and boundaries of rough sets can be found in Pawlak (1991) and cited references. The initial and basic theory of rough sets is sometimes referred to as "Pawlak Rough Sets" or "classical rough sets", as a means to distinguish from more recent extensions and generalizations.

Information system framework

Let   be an information system (attribute–value system), where   is a non-empty, finite set of objects (the universe) and   is a non-empty, finite set of attributes such that   for every  .   is the set of values that attribute   may take. The information table assigns a value   from   to each attribute   and object   in the universe  .

With any   there is an associated equivalence relation  :

 

The relation   is called a  -indiscernibility relation. The partition of   is a family of all equivalence classes of   and is denoted by   (or  ).

If  , then   and   are indiscernible (or indistinguishable) by attributes from   .

The equivalence classes of the  -indiscernibility relation are denoted  .

Example: equivalence-class structure

For example, consider the following information table:

Sample Information System
Object          
  1 2 0 1 1
  1 2 0 1 1
  2 0 0 1 0
  0 0 1 2 1
  2 1 0 2 1
  0 0 1 2 2
  2 0 0 1 0
  0 1 2 2 1
  2 1 0 2 2
  2 0 0 1 0

When the full set of attributes   is considered, we see that we have the following seven equivalence classes:

 

Thus, the two objects within the first equivalence class,  , cannot be distinguished from each other based on the available attributes, and the three objects within the second equivalence class,  , cannot be distinguished from one another based on the available attributes. The remaining five objects are each discernible from all other objects.

It is apparent that different attribute subset selections will in general lead to different indiscernibility classes. For example, if attribute   alone is selected, we obtain the following, much coarser, equivalence-class structure:

 

Definition of a rough set

Let   be a target set that we wish to represent using attribute subset  ; that is, we are told that an arbitrary set of objects   comprises a single class, and we wish to express this class (i.e., this subset) using the equivalence classes induced by attribute subset  . In general,   cannot be expressed exactly, because the set may include and exclude objects which are indistinguishable on the basis of attributes  .

For example, consider the target set  , and let attribute subset  , the full available set of features. The set   cannot be expressed exactly, because in  , objects   are indiscernible. Thus, there is no way to represent any set   which includes   but excludes objects   and  .

However, the target set   can be approximated using only the information contained within   by constructing the  -lower and  -upper approximations of  :

 
 

Lower approximation and positive region

The  -lower approximation, or positive region, is the union of all equivalence classes in   which are contained by (i.e., are subsets of) the target set – in the example,  , the union of the two equivalence classes in   which are contained in the target set. The lower approximation is the complete set of objects in   that can be positively (i.e., unambiguously) classified as belonging to target set  .

Upper approximation and negative region

The  -upper approximation is the union of all equivalence classes in   which have non-empty intersection with the target set – in the example,  , the union of the three equivalence classes in   that have non-empty intersection with the target set. The upper approximation is the complete set of objects that in   that cannot be positively (i.e., unambiguously) classified as belonging to the complement ( ) of the target set  . In other words, the upper approximation is the complete set of objects that are possibly members of the target set  .

The set   therefore represents the negative region, containing the set of objects that can be definitely ruled out as members of the target set.

Boundary region

The boundary region, given by set difference  , consists of those objects that can neither be ruled in nor ruled out as members of the target set  .

In summary, the lower approximation of a target set is a conservative approximation consisting of only those objects which can positively be identified as members of the set. (These objects have no indiscernible "clones" which are excluded by the target set.) The upper approximation is a liberal approximation which includes all objects that might be members of target set. (Some objects in the upper approximation may not be members of the target set.) From the perspective of  , the lower approximation contains objects that are members of the target set with certainty (probability = 1), while the upper approximation contains objects that are members of the target set with non-zero probability (probability > 0).

The rough set

The tuple   composed of the lower and upper approximation is called a rough set; thus, a rough set is composed of two crisp sets, one representing a lower boundary of the target set  , and the other representing an upper boundary of the target set  .

The accuracy of the rough-set representation of the set   can be given (Pawlak 1991) by the following:

 

That is, the accuracy of the rough set representation of  ,  ,  , is the ratio of the number of objects which can positively be placed in   to the number of objects that can possibly be placed in   – this provides a measure of how closely the rough set is approximating the target set. Clearly, when the upper and lower approximations are equal (i.e., boundary region empty), then  , and the approximation is perfect; at the other extreme, whenever the lower approximation is empty, the accuracy is zero (regardless of the size of the upper approximation).

Objective analysis

Rough set theory is one of many methods that can be employed to analyse uncertain (including vague) systems, although less common than more traditional methods of probability, statistics, entropy and Dempster–Shafer theory. However a key difference, and a unique strength, of using classical rough set theory is that it provides an objective form of analysis (Pawlak et al. 1995). Unlike other methods, as those given above, classical rough set analysis requires no additional information, external parameters, models, functions, grades or subjective interpretations to determine set membership – instead it only uses the information presented within the given data (Düntsch and Gediga 1995). More recent adaptations of rough set theory, such as dominance-based, decision-theoretic and fuzzy rough sets, have introduced more subjectivity to the analysis.

Definability

In general, the upper and lower approximations are not equal; in such cases, we say that target set   is undefinable or roughly definable on attribute set  . When the upper and lower approximations are equal (i.e., the boundary is empty),  , then the target set   is definable on attribute set  . We can distinguish the following special cases of undefinability:

  • Set   is internally undefinable if   and  . This means that on attribute set  , there are no objects which we can be certain belong to target set  , but there are objects which we can definitively exclude from set  .
  • Set   is externally undefinable if   and  . This means that on attribute set  , there are objects which we can be certain belong to target set  , but there are no objects which we can definitively exclude from set  .
  • Set   is totally undefinable if   and  . This means that on attribute set  , there are no objects which we can be certain belong to target set  , and there are no objects which we can definitively exclude from set  . Thus, on attribute set  , we cannot decide whether any object is, or is not, a member of  .

Reduct and core

An interesting question is whether there are attributes in the information system (attribute–value table) which are more important to the knowledge represented in the equivalence class structure than other attributes. Often, we wonder whether there is a subset of attributes which can, by itself, fully characterize the knowledge in the database; such an attribute set is called a reduct.

Formally, a reduct is a subset of attributes   such that

  •   =  , that is, the equivalence classes induced by the reduced attribute set   are the same as the equivalence class structure induced by the full attribute set  .
  • the attribute set   is minimal, in the sense that   for any attribute  ; in other words, no attribute can be removed from set   without changing the equivalence classes  .

A reduct can be thought of as a sufficient set of features – sufficient, that is, to represent the category structure. In the example table above, attribute set   is a reduct – the information system projected on just these attributes possesses the same equivalence class structure as that expressed by the full attribute set:

 

Attribute set   is a reduct because eliminating any of these attributes causes a collapse of the equivalence-class structure, with the result that  .

The reduct of an information system is not unique: there may be many subsets of attributes which preserve the equivalence-class structure (i.e., the knowledge) expressed in the information system. In the example information system above, another reduct is  , producing the same equivalence-class structure as  .

The set of attributes which is common to all reducts is called the core: the core is the set of attributes which is possessed by every reduct, and therefore consists of attributes which cannot be removed from the information system without causing collapse of the equivalence-class structure. The core may be thought of as the set of necessary attributes – necessary, that is, for the category structure to be represented. In the example, the only such attribute is  ; any one of the other attributes can be removed singly without damaging the equivalence-class structure, and hence these are all dispensable. However, removing   by itself does change the equivalence-class structure, and thus   is the indispensable attribute of this information system, and hence the core.

It is possible for the core to be empty, which means that there is no indispensable attribute: any single attribute in such an information system can be deleted without altering the equivalence-class structure. In such cases, there is no essential or necessary attribute which is required for the class structure to be represented.

Attribute dependency

One of the most important aspects of database analysis or data acquisition is the discovery of attribute dependencies; that is, we wish to discover which variables are strongly related to which other variables. Generally, it is these strong relationships that will warrant further investigation, and that will ultimately be of use in predictive modeling.

In rough set theory, the notion of dependency is defined very simply. Let us take two (disjoint) sets of attributes, set   and set  , and inquire what degree of dependency obtains between them. Each attribute set induces an (indiscernibility) equivalence class structure, the equivalence classes induced by   given by  , and the equivalence classes induced by   given by  .

Let  , where   is a given equivalence class from the equivalence-class structure induced by attribute set  . Then, the dependency of attribute set   on attribute set  ,  , is given by

 

That is, for each equivalence class   in  , we add up the size of its lower approximation by the attributes in  , i.e.,  . This approximation (as above, for arbitrary set  ) is the number of objects which on attribute set   can be positively identified as belonging to target set  . Added across all equivalence classes in  , the numerator above represents the total number of objects which – based on attribute set   – can be positively categorized according to the classification induced by attributes  . The dependency ratio therefore expresses the proportion (within the entire universe) of such classifiable objects. The dependency   "can be interpreted as a proportion of such objects in the information system for which it suffices to know the values of attributes in   to determine the values of attributes in  ".

Another, intuitive, way to consider dependency is to take the partition induced by   as the target class  , and consider   as the attribute set we wish to use in order to "re-construct" the target class  . If   can completely reconstruct  , then   depends totally upon  ; if   results in a poor and perhaps a random reconstruction of  , then   does not depend upon   at all.

Thus, this measure of dependency expresses the degree of functional (i.e., deterministic) dependency of attribute set   on attribute set  ; it is not symmetric. The relationship of this notion of attribute dependency to more traditional information-theoretic (i.e., entropic) notions of attribute dependence has been discussed in a number of sources (e.g., Pawlak, Wong, & Ziarko 1988; Yao & Yao 2002; Wong, Ziarko, & Ye 1986, Quafafou & Boussouf 2000).

Rule extraction

The category representations discussed above are all extensional in nature; that is, a category or complex class is simply the sum of all its members. To represent a category is, then, just to be able to list or identify all the objects belonging to that category. However, extensional category representations have very limited practical use, because they provide no insight for deciding whether novel (never-before-seen) objects are members of the category.

What is generally desired is an intentional description of the category, a representation of the category based on a set of rules that describe the scope of the category. The choice of such rules is not unique, and therein lies the issue of inductive bias. See Version space and Model selection for more about this issue.

There are a few rule-extraction methods. We will start from a rule-extraction procedure based on Ziarko & Shan (1995).

Decision matrices

Let us say that we wish to find the minimal set of consistent rules (logical implications) that characterize our sample system. For a set of condition attributes   and a decision attribute  , these rules should have the form  , or, spelled out,

 

where   are legitimate values from the domains of their respective attributes. This is a form typical of association rules, and the number of items in   which match the condition/antecedent is called the support for the rule. The method for extracting such rules given in Ziarko & Shan (1995) is to form a decision matrix corresponding to each individual value   of decision attribute  . Informally, the decision matrix for value   of decision attribute   lists all attribute–value pairs that differ between objects having   and  .

This is best explained by example (which also avoids a lot of notation). Consider the table above, and let   be the decision variable (i.e., the variable on the right side of the implications) and let   be the condition variables (on the left side of the implication). We note that the decision variable   takes on two different values, namely  . We treat each case separately.

First, we look at the case  , and we divide up   into objects that have   and those that have  . (Note that objects with   in this case are simply the objects that have  , but in general,   would include all objects having any value for   other than  , and there may be several such classes of objects (for example, those having  ).) In this case, the objects having   are   while the objects which have   are  . The decision matrix for   lists all the differences between the objects having   and those having  ; that is, the decision matrix lists all the differences between   and  . We put the "positive" objects ( ) as the rows, and the "negative" objects   as the columns.

Decision matrix for  
Object          
           
           
           
           
           

To read this decision matrix, look, for example, at the intersection of row   and column  , showing   in the cell. This means that with regard to decision value  , object   differs from object   on attributes   and  , and the particular values on these attributes for the positive object   are   and  . This tells us that the correct classification of   as belonging to decision class   rests on attributes   and  ; although one or the other might be dispensable, we know that at least one of these attributes is indispensable.

Next, from each decision matrix we form a set of Boolean expressions, one expression for each row of the matrix. The items within each cell are aggregated disjunctively, and the individuals cells are then aggregated conjunctively. Thus, for the above table we have the following five Boolean expressions:

 

Each statement here is essentially a highly specific (probably too specific) rule governing the membership in class   of the corresponding object. For example, the last statement, corresponding to object  , states that all the following must be satisfied:

  1. Either   must have value 2, or   must have value 0, or both.
  2.   must have value 0.
  3. Either   must have value 2, or   must have value 0, or both.
  4. Either   must have value 2, or   must have value 0, or   must have value 0, or any combination thereof.
  5.   must have value 0.

It is clear that there is a large amount of redundancy here, and the next step is to simplify using traditional Boolean algebra. The statement   corresponding to objects   simplifies to  , which yields the implication

 

Likewise, the statement   corresponding to objects   simplifies to  . This gives us the implication

 

The above implications can also be written as the following rule set:

 

It can be noted that each of the first two rules has a support of 1 (i.e., the antecedent matches two objects), while each of the last two rules has a support of 2. To finish writing the rule set for this knowledge system, the same procedure as above (starting with writing a new decision matrix) should be followed for the case of  , thus yielding a new set of implications for that decision value (i.e., a set of implications with   as the consequent). In general, the procedure will be repeated for each possible value of the decision variable.

LERS rule induction system

The data system LERS (Learning from Examples based on Rough Sets) Grzymala-Busse (1997) may induce rules from inconsistent data, i.e., data with conflicting objects. Two objects are conflicting when they are characterized by the same values of all attributes, but they belong to different concepts (classes). LERS uses rough set theory to compute lower and upper approximations for concepts involved in conflicts with other concepts.

Rules induced from the lower approximation of the concept certainly describe the concept, hence such rules are called certain. On the other hand, rules induced from the upper approximation of the concept describe the concept possibly, so these rules are called possible. For rule induction LERS uses three algorithms: LEM1, LEM2, and IRIM.

The LEM2 algorithm of LERS is frequently used for rule induction and is used not only in LERS but also in other systems, e.g., in RSES (Bazan et al. (2004). LEM2 explores the search space of attribute–value pairs. Its input data set is a lower or upper approximation of a concept, so its input data set is always consistent. In general, LEM2 computes a local covering and then converts it into a rule set. We will quote a few definitions to describe the LEM2 algorithm.

The LEM2 algorithm is based on an idea of an attribute–value pair block. Let   be a nonempty lower or upper approximation of a concept represented by a decision-value pair  . Set   depends on a set   of attribute–value pairs   if and only if

 

Set   is a minimal complex of   if and only if   depends on   and no proper subset   of   exists such that   depends on  . Let   be a nonempty collection of nonempty sets of attribute–value pairs. Then   is a local covering of   if and only if the following three conditions are satisfied:

each member   of   is a minimal complex of  ,

 
  is minimal, i.e.,   has the smallest possible number of members.

For our sample information system, LEM2 will induce the following rules:

 

Other rule-learning methods can be found, e.g., in Pawlak (1991), Stefanowski (1998), Bazan et al. (2004), etc.

Incomplete data

Rough set theory is useful for rule induction from incomplete data sets. Using this approach we can distinguish between three types of missing attribute values: lost values (the values that were recorded but currently are unavailable), attribute-concept values (these missing attribute values may be replaced by any attribute value limited to the same concept), and "do not care" conditions (the original values were irrelevant). A concept (class) is a set of all objects classified (or diagnosed) the same way.

Two special data sets with missing attribute values were extensively studied: in the first case, all missing attribute values were lost (Stefanowski and Tsoukias, 2001), in the second case, all missing attribute values were "do not care" conditions (Kryszkiewicz, 1999).

In attribute-concept values interpretation of a missing attribute value, the missing attribute value may be replaced by any value of the attribute domain restricted to the concept to which the object with a missing attribute value belongs (Grzymala-Busse and Grzymala-Busse, 2007). For example, if for a patient the value of an attribute Temperature is missing, this patient is sick with flu, and all remaining patients sick with flu have values high or very-high for Temperature when using the interpretation of the missing attribute value as the attribute-concept value, we will replace the missing attribute value with high and very-high. Additionally, the characteristic relation, (see, e.g., Grzymala-Busse and Grzymala-Busse, 2007) enables to process data sets with all three kind of missing attribute values at the same time: lost, "do not care" conditions, and attribute-concept values.

Applications

Rough set methods can be applied as a component of hybrid solutions in machine learning and data mining. They have been found to be particularly useful for rule induction and feature selection (semantics-preserving dimensionality reduction). Rough set-based data analysis methods have been successfully applied in bioinformatics, economics and finance, medicine, multimedia, web and text mining, signal and image processing, software engineering, robotics, and engineering (e.g. power systems and control engineering). Recently the three regions of rough sets are interpreted as regions of acceptance, rejection and deferment. This leads to three-way decision making approach with the model which can potentially lead to interesting future applications.

History

The idea of rough set was proposed by Pawlak (1981) as a new mathematical tool to deal with vague concepts. Comer, Grzymala-Busse, Iwinski, Nieminen, Novotny, Pawlak, Obtulowicz, and Pomykala have studied algebraic properties of rough sets. Different algebraic semantics have been developed by P. Pagliani, I. Duntsch, M. K. Chakraborty, M. Banerjee and A. Mani; these have been extended to more generalized rough sets by D. Cattaneo and A. Mani, in particular. Rough sets can be used to represent ambiguity, vagueness and general uncertainty.

Extensions and generalizations

Since the development of rough sets, extensions and generalizations have continued to evolve. Initial developments focused on the relationship - both similarities and difference - with fuzzy sets. While some literature contends these concepts are different, other literature considers that rough sets are a generalization of fuzzy sets - as represented through either fuzzy rough sets or rough fuzzy sets. Pawlak (1995) considered that fuzzy and rough sets should be treated as being complementary to each other, addressing different aspects of uncertainty and vagueness.

Three notable extensions of classical rough sets are:

  • Dominance-based rough set approach (DRSA) is an extension of rough set theory for multi-criteria decision analysis (MCDA), introduced by Greco, Matarazzo and Słowiński (2001). The main change in this extension of classical rough sets is the substitution of the indiscernibility relation by a dominance relation, which permits the formalism to deal with inconsistencies typical in consideration of criteria and preference-ordered decision classes.
  • Decision-theoretic rough sets (DTRS) is a probabilistic extension of rough set theory introduced by Yao, Wong, and Lingras (1990). It utilizes a Bayesian decision procedure for minimum risk decision making. Elements are included into the lower and upper approximations based on whether their conditional probability is above thresholds   and  . These upper and lower thresholds determine region inclusion for elements. This model is unique and powerful since the thresholds themselves are calculated from a set of six loss functions representing classification risks.
  • Game-theoretic rough sets (GTRS) is a game theory-based extension of rough set that was introduced by Herbert and Yao (2011). It utilizes a game-theoretic environment to optimize certain criteria of rough sets based classification or decision making in order to obtain effective region sizes.

Rough membership

Rough sets can be also defined, as a generalisation, by employing a rough membership function instead of objective approximation. The rough membership function expresses a conditional probability that   belongs to   given  . This can be interpreted as a degree that   belongs to   in terms of information about   expressed by  .

Rough membership primarily differs from the fuzzy membership in that the membership of union and intersection of sets cannot, in general, be computed from their constituent membership as is the case of fuzzy sets. In this, rough membership is a generalization of fuzzy membership. Furthermore, the rough membership function is grounded more in probability than the conventionally held concepts of the fuzzy membership function.

Other generalizations

Several generalizations of rough sets have been introduced, studied and applied to solving problems. Here are some of these generalizations:

  • rough multisets (Grzymala-Busse, 1987)
  • fuzzy rough sets extend the rough set concept through the use of fuzzy equivalence classes(Nakamura, 1988)
  • Alpha rough set theory (α-RST) - a generalization of rough set theory that allows approximation using of fuzzy concepts (Quafafou, 2000)
  • intuitionistic fuzzy rough sets (Cornelis, De Cock and Kerre, 2003)
  • generalized rough fuzzy sets (Feng, 2010)
  • rough intuitionistic fuzzy sets (Thomas and Nair, 2011)
  • soft rough fuzzy sets and soft fuzzy rough sets (Meng, Zhang and Qin, 2011)
  • composite rough sets (Zhang, Li and Chen, 2014)

See also

References

  • Pawlak, Zdzisław (1982). "Rough sets". International Journal of Parallel Programming. 11 (5): 341–356. doi:10.1007/BF01001956. S2CID 9240608.
  • Bazan, Jan; Szczuka, Marcin; Wojna, Arkadiusz; Wojnarski, Marcin (2004). On the evolution of rough set exploration system. Proceedings of the RSCTC 2004. Lecture Notes in Computer Science. Vol. 3066. pp. 592–601. CiteSeerX 10.1.1.60.3957. doi:10.1007/978-3-540-25929-9_73. ISBN 978-3-540-22117-3.
  • Dubois, D.; Prade, H. (1990). "Rough fuzzy sets and fuzzy rough sets". International Journal of General Systems. 17 (2–3): 191–209. doi:10.1080/03081079008935107.
  • Herbert, J. P.; Yao, J. T. (2011). "Game-theoretic Rough Sets". Fundamenta Informaticae. 108 (3–4): 267–286. doi:10.3233/FI-2011-423.
  • Greco, Salvatore; Matarazzo, Benedetto; Słowiński, Roman (2001). "Rough sets theory for multicriteria decision analysis". European Journal of Operational Research. 129 (1): 1–47. doi:10.1016/S0377-2217(00)00167-3.
  • Grzymala-Busse, Jerzy (1997). "A new version of the rule induction system LERS". Fundamenta Informaticae. 31: 27–39. doi:10.3233/FI-1997-3113.
  • Grzymala-Busse, Jerzy; Grzymala-Busse, Witold (2007). An experimental comparison of three rough set approaches to missing attribute values. Transactions on Rough Sets. Lecture Notes in Computer Science. Vol. 6. pp. 31–50. doi:10.1007/978-3-540-71200-8_3. ISBN 978-3-540-71198-8.
  • Kryszkiewicz, Marzena (1999). "Rules in incomplete systems". Information Sciences. 113 (3–4): 271–292. doi:10.1016/S0020-0255(98)10065-8.
  • Pawlak, Zdzisław Rough Sets Research Report PAS 431, Institute of Computer Science, Polish Academy of Sciences (1981)
  • Pawlak, Zdzisław; Wong, S. K. M.; Ziarko, Wojciech (1988). "Rough sets: Probabilistic versus deterministic approach". International Journal of Man-Machine Studies. 29: 81–95. doi:10.1016/S0020-7373(88)80032-4.
  • Pawlak, Zdzisław (1991). Rough Sets: Theoretical Aspects of Reasoning About Data. Dordrecht: Kluwer Academic Publishing. ISBN 978-0-7923-1472-1.
  • Slezak, Dominik; Wroblewski, Jakub; Eastwood, Victoria; Synak, Piotr (2008). "Brighthouse: an analytic data warehouse for ad-hoc queries" (PDF). Proceedings of the VLDB Endowment. 1 (2): 1337–1345. doi:10.14778/1454159.1454174.
  • Stefanowski, Jerzy (1998). "On rough set based approaches to induction of decision rules". In Polkowski, Lech; Skowron, Andrzej (eds.). Rough Sets in Knowledge Discovery 1: Methodology and Applications. Heidelberg: Physica-Verlag. pp. 500–529.
  • Stefanowski, Jerzy; Tsoukias, Alexis (2001). "Incomplete information tables and rough classification". Computational Intelligence. 17 (3): 545–566. doi:10.1111/0824-7935.00162. S2CID 22795201.
  • Wong, S. K. M.; Ziarko, Wojciech; Ye, R. Li (1986). "Comparison of rough-set and statistical methods in inductive learning". International Journal of Man-Machine Studies. 24: 53–72. doi:10.1016/S0020-7373(86)80033-5.
  • Yao, J. T.; Yao, Y. Y. (2002). "Induction of classification rules by granular computing". Proceedings of the Third International Conference on Rough Sets and Current Trends in Computing (TSCTC'02). London, UK: Springer-Verlag. pp. 331–338.
  • Ziarko, Wojciech (1998). "Rough sets as a methodology for data mining". Rough Sets in Knowledge Discovery 1: Methodology and Applications. Heidelberg: Physica-Verlag. pp. 554–576.
  • Ziarko, Wojciech; Shan, Ning (1995). "Discovering attribute relationships, dependencies and rules by using rough sets". Proceedings of the 28th Annual Hawaii International Conference on System Sciences (HICSS'95). Hawaii. pp. 293–299.
  • Pawlak, Zdzisław (1999). "Decision rules, Bayes' rule and rough sets". New Direction in Rough Sets, Data Mining, and Granular-soft Computing: 1–9.
  • Pawlak, Zdzisław. "Rough relations, reports". 435. Institute of Computer Science. {{cite journal}}: Cite journal requires |journal= (help)
  • Orlowska, E. (1987). "Reasoning about vague concepts". Bulletin of the Polish Academy of Sciences. 35: 643–652.
  • Polkowski, L. (2002). "Rough sets: Mathematical foundations". Advances in Soft Computing.
  • Skowron, A. (1996). "Rough sets and vague concepts". Fundamenta Informaticae: 417–431.
  • Burgin M. (1990). Theory of Named Sets as a Foundational Basis for Mathematics, In Structures in mathematical theories: Reports of the San Sebastian international symposium, September 25–29, 1990 (http://www.blogg.org/blog-30140-date-2005-10-26.html)
  • Burgin, M. (2004). Unified Foundations of Mathematics, Preprint Mathematics LO/0403186, p39. (electronic edition: https://arxiv.org/ftp/math/papers/0403/0403186.pdf)
  • Burgin, M. (2011), Theory of Named Sets, Mathematics Research Developments, Nova Science Pub Inc, ISBN 978-1-61122-788-8
  • Cornelis, C., De Cock, M. and Kerre, E. (2003) Intuitionistic fuzzy rough sets: at the crossroads of imperfect knowledge, Expert Systems, 20:5, pp260–270
  • Düntsch, I. and Gediga, G. (1995) Rough Set Dependency Analysis in Evaluation Studies – An Application in the Study of Repeated Heart Attacks. University of Ulster, Informatics Research Reports No. 10
  • Feng F. (2010). Generalized Rough Fuzzy Sets Based on Soft Sets, Soft Computing, 14:9, pp 899–911
  • Grzymala-Busse, J. (1987). Learning from examples based on rough multisets, in Proceedings of the 2nd International Symposium on Methodologies for Intelligent Systems, pp. 325–332. Charlotte, NC, USA
  • Meng, D., Zhang, X. and Qin, K. (2011). Soft rough fuzzy sets and soft fuzzy rough sets, Computers & Mathematics with Applications, 62:12, pp4635–4645
  • Quafafou M. (2000). α-RST: a generalization of rough set theory, Information Sciences, 124:1–4, pp301–316.
  • Quafafou M. and Boussouf M. (2000). Generalized rough sets based feature selection. Journal Intelligent Data Analysis, 4:1 pp3 – 17
  • Nakamura, A. (1988) Fuzzy rough sets, ‘Notes on Multiple-valued Logic in Japan’, 9:1, pp1–8
  • Pawlak, Z., Grzymala-Busse, J., Slowinski, R. Ziarko, W. (1995). Rough Sets. Communications of the ACM, 38:11, pp88–95
  • Thomas, K. and Nair, L. (2011). Rough intuitionistic fuzzy sets in a lattice, International Mathematical Forum, 6:27, pp1327–1335
  • Zhang J., Li T., Chen H. (2014). Composite rough sets for dynamic data mining, Information Sciences, 257, pp81–100
  • Zhang J., Wong J-S, Pan Y, Li T. (2015). A parallel matrix-based method for computing approximations in incomplete information systems, IEEE Transactions on Knowledge and Data Engineering, 27(2): 326-339
  • Chen H., Li T., Luo C., Horng S-J., Wang G. (2015). A decision-theoretic rough set approach for dynamic data mining. IEEE Transactions on Fuzzy Systems, 23(6): 1958-1970
  • Chen H., Li T., Luo C., Horng S-J., Wang G. (2014). A rough set-based method for updating decision rules on attribute values' coarsening and refining, IEEE Transactions on Knowledge and Data Engineering, 26(12): 2886-2899
  • Chen H., Li T., Ruan D., Lin J., Hu C, (2013) A rough-set based incremental approach for updating approximations under dynamic maintenance environments. IEEE Transactions on Knowledge and Data Engineering, 25(2): 274-284

Further reading

  • Gianpiero Cattaneo and Davide Ciucci, "Heyting Wajsberg Algebras as an Abstract Environment Linking Fuzzy and Rough Sets" in J.J. Alpigini et al. (Eds.): RSCTC 2002, LNAI 2475, pp. 77–84, 2002. doi:10.1007/3-540-45813-1_10

External links

  • The International Rough Set Society
  • Rough set tutorial
  • Rough Set Exploration System
  • Rough Sets in Data Warehousing

rough, computer, science, rough, first, described, polish, computer, scientist, zdzisław, pawlak, formal, approximation, crisp, conventional, terms, pair, sets, which, give, lower, upper, approximation, original, standard, version, rough, theory, pawlak, 1991,. In computer science a rough set first described by Polish computer scientist Zdzislaw I Pawlak is a formal approximation of a crisp set i e conventional set in terms of a pair of sets which give the lower and the upper approximation of the original set In the standard version of rough set theory Pawlak 1991 the lower and upper approximation sets are crisp sets but in other variations the approximating sets may be fuzzy sets Contents 1 Definitions 1 1 Information system framework 1 2 Example equivalence class structure 1 3 Definition of a rough set 1 3 1 Lower approximation and positive region 1 3 2 Upper approximation and negative region 1 3 3 Boundary region 1 3 4 The rough set 1 3 5 Objective analysis 1 4 Definability 1 5 Reduct and core 1 6 Attribute dependency 2 Rule extraction 2 1 Decision matrices 2 2 LERS rule induction system 3 Incomplete data 4 Applications 5 History 6 Extensions and generalizations 6 1 Rough membership 6 2 Other generalizations 7 See also 8 References 9 Further reading 10 External linksDefinitions EditThe following section contains an overview of the basic framework of rough set theory as originally proposed by Zdzislaw I Pawlak along with some of the key definitions More formal properties and boundaries of rough sets can be found in Pawlak 1991 and cited references The initial and basic theory of rough sets is sometimes referred to as Pawlak Rough Sets or classical rough sets as a means to distinguish from more recent extensions and generalizations Information system framework Edit Let I U A displaystyle I mathbb U mathbb A be an information system attribute value system where U displaystyle mathbb U is a non empty finite set of objects the universe and A displaystyle mathbb A is a non empty finite set of attributes such that I U V a displaystyle I mathbb U rightarrow V a for every a A displaystyle a in mathbb A V a displaystyle V a is the set of values that attribute a displaystyle a may take The information table assigns a value a x displaystyle a x from V a displaystyle V a to each attribute a displaystyle a and object x displaystyle x in the universe U displaystyle mathbb U With any P A displaystyle P subseteq mathbb A there is an associated equivalence relation I N D P displaystyle mathrm IND P I N D P x y U 2 a P a x a y displaystyle mathrm IND P left x y in mathbb U 2 mid forall a in P a x a y right The relation I N D P displaystyle mathrm IND P is called a P displaystyle P indiscernibility relation The partition of U displaystyle mathbb U is a family of all equivalence classes of I N D P displaystyle mathrm IND P and is denoted by U I N D P displaystyle mathbb U mathrm IND P or U P displaystyle mathbb U P If x y I N D P displaystyle x y in mathrm IND P then x displaystyle x and y displaystyle y are indiscernible or indistinguishable by attributes from P displaystyle P The equivalence classes of the P displaystyle P indiscernibility relation are denoted x P displaystyle x P Example equivalence class structure Edit For example consider the following information table Sample Information System Object P 1 displaystyle P 1 P 2 displaystyle P 2 P 3 displaystyle P 3 P 4 displaystyle P 4 P 5 displaystyle P 5 O 1 displaystyle O 1 1 2 0 1 1O 2 displaystyle O 2 1 2 0 1 1O 3 displaystyle O 3 2 0 0 1 0O 4 displaystyle O 4 0 0 1 2 1O 5 displaystyle O 5 2 1 0 2 1O 6 displaystyle O 6 0 0 1 2 2O 7 displaystyle O 7 2 0 0 1 0O 8 displaystyle O 8 0 1 2 2 1O 9 displaystyle O 9 2 1 0 2 2O 10 displaystyle O 10 2 0 0 1 0When the full set of attributes P P 1 P 2 P 3 P 4 P 5 displaystyle P P 1 P 2 P 3 P 4 P 5 is considered we see that we have the following seven equivalence classes O 1 O 2 O 3 O 7 O 10 O 4 O 5 O 6 O 8 O 9 displaystyle begin cases O 1 O 2 O 3 O 7 O 10 O 4 O 5 O 6 O 8 O 9 end cases Thus the two objects within the first equivalence class O 1 O 2 displaystyle O 1 O 2 cannot be distinguished from each other based on the available attributes and the three objects within the second equivalence class O 3 O 7 O 10 displaystyle O 3 O 7 O 10 cannot be distinguished from one another based on the available attributes The remaining five objects are each discernible from all other objects It is apparent that different attribute subset selections will in general lead to different indiscernibility classes For example if attribute P P 1 displaystyle P P 1 alone is selected we obtain the following much coarser equivalence class structure O 1 O 2 O 3 O 5 O 7 O 9 O 10 O 4 O 6 O 8 displaystyle begin cases O 1 O 2 O 3 O 5 O 7 O 9 O 10 O 4 O 6 O 8 end cases Definition of a rough set Edit Let X U displaystyle X subseteq mathbb U be a target set that we wish to represent using attribute subset P displaystyle P that is we are told that an arbitrary set of objects X displaystyle X comprises a single class and we wish to express this class i e this subset using the equivalence classes induced by attribute subset P displaystyle P In general X displaystyle X cannot be expressed exactly because the set may include and exclude objects which are indistinguishable on the basis of attributes P displaystyle P For example consider the target set X O 1 O 2 O 3 O 4 displaystyle X O 1 O 2 O 3 O 4 and let attribute subset P P 1 P 2 P 3 P 4 P 5 displaystyle P P 1 P 2 P 3 P 4 P 5 the full available set of features The set X displaystyle X cannot be expressed exactly because in x P displaystyle x P objects O 3 O 7 O 10 displaystyle O 3 O 7 O 10 are indiscernible Thus there is no way to represent any set X displaystyle X which includes O 3 displaystyle O 3 but excludes objects O 7 displaystyle O 7 and O 10 displaystyle O 10 However the target set X displaystyle X can be approximated using only the information contained within P displaystyle P by constructing the P displaystyle P lower and P displaystyle P upper approximations of X displaystyle X P X x x P X displaystyle underline P X x mid x P subseteq X P X x x P X displaystyle overline P X x mid x P cap X neq emptyset Lower approximation and positive region Edit The P displaystyle P lower approximation or positive region is the union of all equivalence classes in x P displaystyle x P which are contained by i e are subsets of the target set in the example P X O 1 O 2 O 4 displaystyle underline P X O 1 O 2 cup O 4 the union of the two equivalence classes in x P displaystyle x P which are contained in the target set The lower approximation is the complete set of objects in U P displaystyle mathbb U P that can be positively i e unambiguously classified as belonging to target set X displaystyle X Upper approximation and negative region Edit The P displaystyle P upper approximation is the union of all equivalence classes in x P displaystyle x P which have non empty intersection with the target set in the example P X O 1 O 2 O 4 O 3 O 7 O 10 displaystyle overline P X O 1 O 2 cup O 4 cup O 3 O 7 O 10 the union of the three equivalence classes in x P displaystyle x P that have non empty intersection with the target set The upper approximation is the complete set of objects that in U P displaystyle mathbb U P that cannot be positively i e unambiguously classified as belonging to the complement X displaystyle overline X of the target set X displaystyle X In other words the upper approximation is the complete set of objects that are possibly members of the target set X displaystyle X The set U P X displaystyle mathbb U overline P X therefore represents the negative region containing the set of objects that can be definitely ruled out as members of the target set Boundary region Edit The boundary region given by set difference P X P X displaystyle overline P X underline P X consists of those objects that can neither be ruled in nor ruled out as members of the target set X displaystyle X In summary the lower approximation of a target set is a conservative approximation consisting of only those objects which can positively be identified as members of the set These objects have no indiscernible clones which are excluded by the target set The upper approximation is a liberal approximation which includes all objects that might be members of target set Some objects in the upper approximation may not be members of the target set From the perspective of U P displaystyle mathbb U P the lower approximation contains objects that are members of the target set with certainty probability 1 while the upper approximation contains objects that are members of the target set with non zero probability probability gt 0 The rough set Edit The tuple P X P X displaystyle langle underline P X overline P X rangle composed of the lower and upper approximation is called a rough set thus a rough set is composed of two crisp sets one representing a lower boundary of the target set X displaystyle X and the other representing an upper boundary of the target set X displaystyle X The accuracy of the rough set representation of the set X displaystyle X can be given Pawlak 1991 by the following a P X P X P X displaystyle alpha P X frac left underline P X right left overline P X right That is the accuracy of the rough set representation of X displaystyle X a P X displaystyle alpha P X 0 a P X 1 displaystyle 0 leq alpha P X leq 1 is the ratio of the number of objects which can positively be placed in X displaystyle X to the number of objects that can possibly be placed in X displaystyle X this provides a measure of how closely the rough set is approximating the target set Clearly when the upper and lower approximations are equal i e boundary region empty then a P X 1 displaystyle alpha P X 1 and the approximation is perfect at the other extreme whenever the lower approximation is empty the accuracy is zero regardless of the size of the upper approximation Objective analysis Edit Rough set theory is one of many methods that can be employed to analyse uncertain including vague systems although less common than more traditional methods of probability statistics entropy and Dempster Shafer theory However a key difference and a unique strength of using classical rough set theory is that it provides an objective form of analysis Pawlak et al 1995 Unlike other methods as those given above classical rough set analysis requires no additional information external parameters models functions grades or subjective interpretations to determine set membership instead it only uses the information presented within the given data Duntsch and Gediga 1995 More recent adaptations of rough set theory such as dominance based decision theoretic and fuzzy rough sets have introduced more subjectivity to the analysis Definability Edit In general the upper and lower approximations are not equal in such cases we say that target set X displaystyle X is undefinable or roughly definable on attribute set P displaystyle P When the upper and lower approximations are equal i e the boundary is empty P X P X displaystyle overline P X underline P X then the target set X displaystyle X is definable on attribute set P displaystyle P We can distinguish the following special cases of undefinability Set X displaystyle X is internally undefinable if P X displaystyle underline P X emptyset and P X U displaystyle overline P X neq mathbb U This means that on attribute set P displaystyle P there are no objects which we can be certain belong to target set X displaystyle X but there are objects which we can definitively exclude from set X displaystyle X Set X displaystyle X is externally undefinable if P X displaystyle underline P X neq emptyset and P X U displaystyle overline P X mathbb U This means that on attribute set P displaystyle P there are objects which we can be certain belong to target set X displaystyle X but there are no objects which we can definitively exclude from set X displaystyle X Set X displaystyle X is totally undefinable if P X displaystyle underline P X emptyset and P X U displaystyle overline P X mathbb U This means that on attribute set P displaystyle P there are no objects which we can be certain belong to target set X displaystyle X and there are no objects which we can definitively exclude from set X displaystyle X Thus on attribute set P displaystyle P we cannot decide whether any object is or is not a member of X displaystyle X Reduct and core Edit An interesting question is whether there are attributes in the information system attribute value table which are more important to the knowledge represented in the equivalence class structure than other attributes Often we wonder whether there is a subset of attributes which can by itself fully characterize the knowledge in the database such an attribute set is called a reduct Formally a reduct is a subset of attributes R E D P displaystyle mathrm RED subseteq P such that x R E D displaystyle x mathrm RED x P displaystyle x P that is the equivalence classes induced by the reduced attribute set R E D displaystyle mathrm RED are the same as the equivalence class structure induced by the full attribute set P displaystyle P the attribute set R E D displaystyle mathrm RED is minimal in the sense that x R E D a x P displaystyle x mathrm RED a neq x P for any attribute a R E D displaystyle a in mathrm RED in other words no attribute can be removed from set R E D displaystyle mathrm RED without changing the equivalence classes x P displaystyle x P A reduct can be thought of as a sufficient set of features sufficient that is to represent the category structure In the example table above attribute set P 3 P 4 P 5 displaystyle P 3 P 4 P 5 is a reduct the information system projected on just these attributes possesses the same equivalence class structure as that expressed by the full attribute set O 1 O 2 O 3 O 7 O 10 O 4 O 5 O 6 O 8 O 9 displaystyle begin cases O 1 O 2 O 3 O 7 O 10 O 4 O 5 O 6 O 8 O 9 end cases Attribute set P 3 P 4 P 5 displaystyle P 3 P 4 P 5 is a reduct because eliminating any of these attributes causes a collapse of the equivalence class structure with the result that x R E D x P displaystyle x mathrm RED neq x P The reduct of an information system is not unique there may be many subsets of attributes which preserve the equivalence class structure i e the knowledge expressed in the information system In the example information system above another reduct is P 1 P 2 P 5 displaystyle P 1 P 2 P 5 producing the same equivalence class structure as x P displaystyle x P The set of attributes which is common to all reducts is called the core the core is the set of attributes which is possessed by every reduct and therefore consists of attributes which cannot be removed from the information system without causing collapse of the equivalence class structure The core may be thought of as the set of necessary attributes necessary that is for the category structure to be represented In the example the only such attribute is P 5 displaystyle P 5 any one of the other attributes can be removed singly without damaging the equivalence class structure and hence these are all dispensable However removing P 5 displaystyle P 5 by itself does change the equivalence class structure and thus P 5 displaystyle P 5 is the indispensable attribute of this information system and hence the core It is possible for the core to be empty which means that there is no indispensable attribute any single attribute in such an information system can be deleted without altering the equivalence class structure In such cases there is no essential or necessary attribute which is required for the class structure to be represented Attribute dependency Edit One of the most important aspects of database analysis or data acquisition is the discovery of attribute dependencies that is we wish to discover which variables are strongly related to which other variables Generally it is these strong relationships that will warrant further investigation and that will ultimately be of use in predictive modeling In rough set theory the notion of dependency is defined very simply Let us take two disjoint sets of attributes set P displaystyle P and set Q displaystyle Q and inquire what degree of dependency obtains between them Each attribute set induces an indiscernibility equivalence class structure the equivalence classes induced by P displaystyle P given by x P displaystyle x P and the equivalence classes induced by Q displaystyle Q given by x Q displaystyle x Q Let x Q Q 1 Q 2 Q 3 Q N displaystyle x Q Q 1 Q 2 Q 3 dots Q N where Q i displaystyle Q i is a given equivalence class from the equivalence class structure induced by attribute set Q displaystyle Q Then the dependency of attribute set Q displaystyle Q on attribute set P displaystyle P g P Q displaystyle gamma P Q is given by g P Q i 1 N P Q i U 1 displaystyle gamma P Q frac sum i 1 N left underline P Q i right left mathbb U right leq 1 That is for each equivalence class Q i displaystyle Q i in x Q displaystyle x Q we add up the size of its lower approximation by the attributes in P displaystyle P i e P Q i displaystyle underline P Q i This approximation as above for arbitrary set X displaystyle X is the number of objects which on attribute set P displaystyle P can be positively identified as belonging to target set Q i displaystyle Q i Added across all equivalence classes in x Q displaystyle x Q the numerator above represents the total number of objects which based on attribute set P displaystyle P can be positively categorized according to the classification induced by attributes Q displaystyle Q The dependency ratio therefore expresses the proportion within the entire universe of such classifiable objects The dependency g P Q displaystyle gamma P Q can be interpreted as a proportion of such objects in the information system for which it suffices to know the values of attributes in P displaystyle P to determine the values of attributes in Q displaystyle Q Another intuitive way to consider dependency is to take the partition induced by Q displaystyle Q as the target class C displaystyle C and consider P displaystyle P as the attribute set we wish to use in order to re construct the target class C displaystyle C If P displaystyle P can completely reconstruct C displaystyle C then Q displaystyle Q depends totally upon P displaystyle P if P displaystyle P results in a poor and perhaps a random reconstruction of C displaystyle C then Q displaystyle Q does not depend upon P displaystyle P at all Thus this measure of dependency expresses the degree of functional i e deterministic dependency of attribute set Q displaystyle Q on attribute set P displaystyle P it is not symmetric The relationship of this notion of attribute dependency to more traditional information theoretic i e entropic notions of attribute dependence has been discussed in a number of sources e g Pawlak Wong amp Ziarko 1988 Yao amp Yao 2002 Wong Ziarko amp Ye 1986 Quafafou amp Boussouf 2000 Rule extraction EditThe category representations discussed above are all extensional in nature that is a category or complex class is simply the sum of all its members To represent a category is then just to be able to list or identify all the objects belonging to that category However extensional category representations have very limited practical use because they provide no insight for deciding whether novel never before seen objects are members of the category What is generally desired is an intentional description of the category a representation of the category based on a set of rules that describe the scope of the category The choice of such rules is not unique and therein lies the issue of inductive bias See Version space and Model selection for more about this issue There are a few rule extraction methods We will start from a rule extraction procedure based on Ziarko amp Shan 1995 Decision matrices Edit Let us say that we wish to find the minimal set of consistent rules logical implications that characterize our sample system For a set of condition attributes P P 1 P 2 P 3 P n displaystyle mathcal P P 1 P 2 P 3 dots P n and a decision attribute Q Q P displaystyle Q Q notin mathcal P these rules should have the form P i a P j b P k c Q d displaystyle P i a P j b dots P k c to Q d or spelled out P i a P j b P k c Q d displaystyle P i a land P j b land dots land P k c to Q d where a b c displaystyle a b c dots are legitimate values from the domains of their respective attributes This is a form typical of association rules and the number of items in U displaystyle mathbb U which match the condition antecedent is called the support for the rule The method for extracting such rules given in Ziarko amp Shan 1995 is to form a decision matrix corresponding to each individual value d displaystyle d of decision attribute Q displaystyle Q Informally the decision matrix for value d displaystyle d of decision attribute Q displaystyle Q lists all attribute value pairs that differ between objects having Q d displaystyle Q d and Q d displaystyle Q neq d This is best explained by example which also avoids a lot of notation Consider the table above and let P 4 displaystyle P 4 be the decision variable i e the variable on the right side of the implications and let P 1 P 2 P 3 displaystyle P 1 P 2 P 3 be the condition variables on the left side of the implication We note that the decision variable P 4 displaystyle P 4 takes on two different values namely 1 2 displaystyle 1 2 We treat each case separately First we look at the case P 4 1 displaystyle P 4 1 and we divide up U displaystyle mathbb U into objects that have P 4 1 displaystyle P 4 1 and those that have P 4 1 displaystyle P 4 neq 1 Note that objects with P 4 1 displaystyle P 4 neq 1 in this case are simply the objects that have P 4 2 displaystyle P 4 2 but in general P 4 1 displaystyle P 4 neq 1 would include all objects having any value for P 4 displaystyle P 4 other than P 4 1 displaystyle P 4 1 and there may be several such classes of objects for example those having P 4 2 3 4 e t c displaystyle P 4 2 3 4 etc In this case the objects having P 4 1 displaystyle P 4 1 are O 1 O 2 O 3 O 7 O 10 displaystyle O 1 O 2 O 3 O 7 O 10 while the objects which have P 4 1 displaystyle P 4 neq 1 are O 4 O 5 O 6 O 8 O 9 displaystyle O 4 O 5 O 6 O 8 O 9 The decision matrix for P 4 1 displaystyle P 4 1 lists all the differences between the objects having P 4 1 displaystyle P 4 1 and those having P 4 1 displaystyle P 4 neq 1 that is the decision matrix lists all the differences between O 1 O 2 O 3 O 7 O 10 displaystyle O 1 O 2 O 3 O 7 O 10 and O 4 O 5 O 6 O 8 O 9 displaystyle O 4 O 5 O 6 O 8 O 9 We put the positive objects P 4 1 displaystyle P 4 1 as the rows and the negative objects P 4 1 displaystyle P 4 neq 1 as the columns Decision matrix for P 4 1 displaystyle P 4 1 Object O 4 displaystyle O 4 O 5 displaystyle O 5 O 6 displaystyle O 6 O 8 displaystyle O 8 O 9 displaystyle O 9 O 1 displaystyle O 1 P 1 1 P 2 2 P 3 0 displaystyle P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 displaystyle P 1 1 P 2 2 P 1 1 P 2 2 P 3 0 displaystyle P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 P 3 0 displaystyle P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 displaystyle P 1 1 P 2 2 O 2 displaystyle O 2 P 1 1 P 2 2 P 3 0 displaystyle P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 displaystyle P 1 1 P 2 2 P 1 1 P 2 2 P 3 0 displaystyle P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 P 3 0 displaystyle P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 displaystyle P 1 1 P 2 2 O 3 displaystyle O 3 P 1 2 P 3 0 displaystyle P 1 2 P 3 0 P 2 0 displaystyle P 2 0 P 1 2 P 3 0 displaystyle P 1 2 P 3 0 P 1 2 P 2 0 P 3 0 displaystyle P 1 2 P 2 0 P 3 0 P 2 0 displaystyle P 2 0 O 7 displaystyle O 7 P 1 2 P 3 0 displaystyle P 1 2 P 3 0 P 2 0 displaystyle P 2 0 P 1 2 P 3 0 displaystyle P 1 2 P 3 0 P 1 2 P 2 0 P 3 0 displaystyle P 1 2 P 2 0 P 3 0 P 2 0 displaystyle P 2 0 O 10 displaystyle O 10 P 1 2 P 3 0 displaystyle P 1 2 P 3 0 P 2 0 displaystyle P 2 0 P 1 2 P 3 0 displaystyle P 1 2 P 3 0 P 1 2 P 2 0 P 3 0 displaystyle P 1 2 P 2 0 P 3 0 P 2 0 displaystyle P 2 0 To read this decision matrix look for example at the intersection of row O 3 displaystyle O 3 and column O 6 displaystyle O 6 showing P 1 2 P 3 0 displaystyle P 1 2 P 3 0 in the cell This means that with regard to decision value P 4 1 displaystyle P 4 1 object O 3 displaystyle O 3 differs from object O 6 displaystyle O 6 on attributes P 1 displaystyle P 1 and P 3 displaystyle P 3 and the particular values on these attributes for the positive object O 3 displaystyle O 3 are P 1 2 displaystyle P 1 2 and P 3 0 displaystyle P 3 0 This tells us that the correct classification of O 3 displaystyle O 3 as belonging to decision class P 4 1 displaystyle P 4 1 rests on attributes P 1 displaystyle P 1 and P 3 displaystyle P 3 although one or the other might be dispensable we know that at least one of these attributes is indispensable Next from each decision matrix we form a set of Boolean expressions one expression for each row of the matrix The items within each cell are aggregated disjunctively and the individuals cells are then aggregated conjunctively Thus for the above table we have the following five Boolean expressions P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 P 1 2 P 3 0 P 2 0 P 1 2 P 3 0 P 1 2 P 2 0 P 3 0 P 2 0 P 1 2 P 3 0 P 2 0 P 1 2 P 3 0 P 1 2 P 2 0 P 3 0 P 2 0 P 1 2 P 3 0 P 2 0 P 1 2 P 3 0 P 1 2 P 2 0 P 3 0 P 2 0 displaystyle begin cases P 1 1 lor P 2 2 lor P 3 0 land P 1 1 lor P 2 2 land P 1 1 lor P 2 2 lor P 3 0 land P 1 1 lor P 2 2 lor P 3 0 land P 1 1 lor P 2 2 P 1 1 lor P 2 2 lor P 3 0 land P 1 1 lor P 2 2 land P 1 1 lor P 2 2 lor P 3 0 land P 1 1 lor P 2 2 lor P 3 0 land P 1 1 lor P 2 2 P 1 2 lor P 3 0 land P 2 0 land P 1 2 lor P 3 0 land P 1 2 lor P 2 0 lor P 3 0 land P 2 0 P 1 2 lor P 3 0 land P 2 0 land P 1 2 lor P 3 0 land P 1 2 lor P 2 0 lor P 3 0 land P 2 0 P 1 2 lor P 3 0 land P 2 0 land P 1 2 lor P 3 0 land P 1 2 lor P 2 0 lor P 3 0 land P 2 0 end cases Each statement here is essentially a highly specific probably too specific rule governing the membership in class P 4 1 displaystyle P 4 1 of the corresponding object For example the last statement corresponding to object O 10 displaystyle O 10 states that all the following must be satisfied Either P 1 displaystyle P 1 must have value 2 or P 3 displaystyle P 3 must have value 0 or both P 2 displaystyle P 2 must have value 0 Either P 1 displaystyle P 1 must have value 2 or P 3 displaystyle P 3 must have value 0 or both Either P 1 displaystyle P 1 must have value 2 or P 2 displaystyle P 2 must have value 0 or P 3 displaystyle P 3 must have value 0 or any combination thereof P 2 displaystyle P 2 must have value 0 It is clear that there is a large amount of redundancy here and the next step is to simplify using traditional Boolean algebra The statement P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 P 3 0 P 1 1 P 2 2 displaystyle P 1 1 lor P 2 2 lor P 3 0 land P 1 1 lor P 2 2 land P 1 1 lor P 2 2 lor P 3 0 land P 1 1 lor P 2 2 lor P 3 0 land P 1 1 lor P 2 2 corresponding to objects O 1 O 2 displaystyle O 1 O 2 simplifies to P 1 1 P 2 2 displaystyle P 1 1 lor P 2 2 which yields the implication P 1 1 P 2 2 P 4 1 displaystyle P 1 1 lor P 2 2 to P 4 1 Likewise the statement P 1 2 P 3 0 P 2 0 P 1 2 P 3 0 P 1 2 P 2 0 P 3 0 P 2 0 displaystyle P 1 2 lor P 3 0 land P 2 0 land P 1 2 lor P 3 0 land P 1 2 lor P 2 0 lor P 3 0 land P 2 0 corresponding to objects O 3 O 7 O 10 displaystyle O 3 O 7 O 10 simplifies to P 1 2 P 2 0 P 3 0 P 2 0 displaystyle P 1 2 P 2 0 lor P 3 0 P 2 0 This gives us the implication P 1 2 P 2 0 P 3 0 P 2 0 P 4 1 displaystyle P 1 2 land P 2 0 lor P 3 0 land P 2 0 to P 4 1 The above implications can also be written as the following rule set P 1 1 P 4 1 P 2 2 P 4 1 P 1 2 P 2 0 P 4 1 P 3 0 P 2 0 P 4 1 displaystyle begin cases P 1 1 to P 4 1 P 2 2 to P 4 1 P 1 2 land P 2 0 to P 4 1 P 3 0 land P 2 0 to P 4 1 end cases It can be noted that each of the first two rules has a support of 1 i e the antecedent matches two objects while each of the last two rules has a support of 2 To finish writing the rule set for this knowledge system the same procedure as above starting with writing a new decision matrix should be followed for the case of P 4 2 displaystyle P 4 2 thus yielding a new set of implications for that decision value i e a set of implications with P 4 2 displaystyle P 4 2 as the consequent In general the procedure will be repeated for each possible value of the decision variable LERS rule induction system Edit The data system LERS Learning from Examples based on Rough Sets Grzymala Busse 1997 may induce rules from inconsistent data i e data with conflicting objects Two objects are conflicting when they are characterized by the same values of all attributes but they belong to different concepts classes LERS uses rough set theory to compute lower and upper approximations for concepts involved in conflicts with other concepts Rules induced from the lower approximation of the concept certainly describe the concept hence such rules are called certain On the other hand rules induced from the upper approximation of the concept describe the concept possibly so these rules are called possible For rule induction LERS uses three algorithms LEM1 LEM2 and IRIM The LEM2 algorithm of LERS is frequently used for rule induction and is used not only in LERS but also in other systems e g in RSES Bazan et al 2004 LEM2 explores the search space of attribute value pairs Its input data set is a lower or upper approximation of a concept so its input data set is always consistent In general LEM2 computes a local covering and then converts it into a rule set We will quote a few definitions to describe the LEM2 algorithm The LEM2 algorithm is based on an idea of an attribute value pair block Let X displaystyle X be a nonempty lower or upper approximation of a concept represented by a decision value pair d w displaystyle d w Set X displaystyle X depends on a set T displaystyle T of attribute value pairs t a v displaystyle t a v if and only if T t T t X displaystyle emptyset neq T bigcap t in T t subseteq X Set T displaystyle T is a minimal complex of X displaystyle X if and only if X displaystyle X depends on T displaystyle T and no proper subset S displaystyle S of T displaystyle T exists such that X displaystyle X depends on S displaystyle S Let T displaystyle mathbb T be a nonempty collection of nonempty sets of attribute value pairs Then T displaystyle mathbb T is a local covering of X displaystyle X if and only if the following three conditions are satisfied each member T displaystyle T of T displaystyle mathbb T is a minimal complex of X displaystyle X t T T X displaystyle bigcup t in mathbb T T X T displaystyle mathbb T is minimal i e T displaystyle mathbb T has the smallest possible number of members For our sample information system LEM2 will induce the following rules P 1 1 P 4 1 P 5 0 P 4 1 P 1 0 P 4 2 P 2 1 P 4 2 displaystyle begin cases P 1 1 to P 4 1 P 5 0 to P 4 1 P 1 0 to P 4 2 P 2 1 to P 4 2 end cases Other rule learning methods can be found e g in Pawlak 1991 Stefanowski 1998 Bazan et al 2004 etc Incomplete data EditRough set theory is useful for rule induction from incomplete data sets Using this approach we can distinguish between three types of missing attribute values lost values the values that were recorded but currently are unavailable attribute concept values these missing attribute values may be replaced by any attribute value limited to the same concept and do not care conditions the original values were irrelevant A concept class is a set of all objects classified or diagnosed the same way Two special data sets with missing attribute values were extensively studied in the first case all missing attribute values were lost Stefanowski and Tsoukias 2001 in the second case all missing attribute values were do not care conditions Kryszkiewicz 1999 In attribute concept values interpretation of a missing attribute value the missing attribute value may be replaced by any value of the attribute domain restricted to the concept to which the object with a missing attribute value belongs Grzymala Busse and Grzymala Busse 2007 For example if for a patient the value of an attribute Temperature is missing this patient is sick with flu and all remaining patients sick with flu have values high or very high for Temperature when using the interpretation of the missing attribute value as the attribute concept value we will replace the missing attribute value with high and very high Additionally the characteristic relation see e g Grzymala Busse and Grzymala Busse 2007 enables to process data sets with all three kind of missing attribute values at the same time lost do not care conditions and attribute concept values Applications EditThis section does not cite any sources Please help improve this section by adding citations to reliable sources Unsourced material may be challenged and removed July 2017 Learn how and when to remove this template message Rough set methods can be applied as a component of hybrid solutions in machine learning and data mining They have been found to be particularly useful for rule induction and feature selection semantics preserving dimensionality reduction Rough set based data analysis methods have been successfully applied in bioinformatics economics and finance medicine multimedia web and text mining signal and image processing software engineering robotics and engineering e g power systems and control engineering Recently the three regions of rough sets are interpreted as regions of acceptance rejection and deferment This leads to three way decision making approach with the model which can potentially lead to interesting future applications History EditThe idea of rough set was proposed by Pawlak 1981 as a new mathematical tool to deal with vague concepts Comer Grzymala Busse Iwinski Nieminen Novotny Pawlak Obtulowicz and Pomykala have studied algebraic properties of rough sets Different algebraic semantics have been developed by P Pagliani I Duntsch M K Chakraborty M Banerjee and A Mani these have been extended to more generalized rough sets by D Cattaneo and A Mani in particular Rough sets can be used to represent ambiguity vagueness and general uncertainty Extensions and generalizations EditSince the development of rough sets extensions and generalizations have continued to evolve Initial developments focused on the relationship both similarities and difference with fuzzy sets While some literature contends these concepts are different other literature considers that rough sets are a generalization of fuzzy sets as represented through either fuzzy rough sets or rough fuzzy sets Pawlak 1995 considered that fuzzy and rough sets should be treated as being complementary to each other addressing different aspects of uncertainty and vagueness Three notable extensions of classical rough sets are Dominance based rough set approach DRSA is an extension of rough set theory for multi criteria decision analysis MCDA introduced by Greco Matarazzo and Slowinski 2001 The main change in this extension of classical rough sets is the substitution of the indiscernibility relation by a dominance relation which permits the formalism to deal with inconsistencies typical in consideration of criteria and preference ordered decision classes Decision theoretic rough sets DTRS is a probabilistic extension of rough set theory introduced by Yao Wong and Lingras 1990 It utilizes a Bayesian decision procedure for minimum risk decision making Elements are included into the lower and upper approximations based on whether their conditional probability is above thresholds a displaystyle textstyle alpha and b displaystyle textstyle beta These upper and lower thresholds determine region inclusion for elements This model is unique and powerful since the thresholds themselves are calculated from a set of six loss functions representing classification risks Game theoretic rough sets GTRS is a game theory based extension of rough set that was introduced by Herbert and Yao 2011 It utilizes a game theoretic environment to optimize certain criteria of rough sets based classification or decision making in order to obtain effective region sizes Rough membership Edit Rough sets can be also defined as a generalisation by employing a rough membership function instead of objective approximation The rough membership function expresses a conditional probability that x displaystyle x belongs to X displaystyle X given R displaystyle textstyle mathbb R This can be interpreted as a degree that x displaystyle x belongs to X displaystyle X in terms of information about x displaystyle x expressed by R displaystyle textstyle mathbb R Rough membership primarily differs from the fuzzy membership in that the membership of union and intersection of sets cannot in general be computed from their constituent membership as is the case of fuzzy sets In this rough membership is a generalization of fuzzy membership Furthermore the rough membership function is grounded more in probability than the conventionally held concepts of the fuzzy membership function Other generalizations Edit Several generalizations of rough sets have been introduced studied and applied to solving problems Here are some of these generalizations rough multisets Grzymala Busse 1987 fuzzy rough sets extend the rough set concept through the use of fuzzy equivalence classes Nakamura 1988 Alpha rough set theory a RST a generalization of rough set theory that allows approximation using of fuzzy concepts Quafafou 2000 intuitionistic fuzzy rough sets Cornelis De Cock and Kerre 2003 generalized rough fuzzy sets Feng 2010 rough intuitionistic fuzzy sets Thomas and Nair 2011 soft rough fuzzy sets and soft fuzzy rough sets Meng Zhang and Qin 2011 composite rough sets Zhang Li and Chen 2014 See also EditAlgebraic semantics Alternative set theory Analog computer Description logic Fuzzy logic Fuzzy set theory Granular computing Near sets Rough fuzzy hybridization Type 2 fuzzy sets and systems Decision theoretic rough sets Version space Dominance based rough set approachReferences EditPawlak Zdzislaw 1982 Rough sets International Journal of Parallel Programming 11 5 341 356 doi 10 1007 BF01001956 S2CID 9240608 Bazan Jan Szczuka Marcin Wojna Arkadiusz Wojnarski Marcin 2004 On the evolution of rough set exploration system Proceedings of the RSCTC 2004 Lecture Notes in Computer Science Vol 3066 pp 592 601 CiteSeerX 10 1 1 60 3957 doi 10 1007 978 3 540 25929 9 73 ISBN 978 3 540 22117 3 Dubois D Prade H 1990 Rough fuzzy sets and fuzzy rough sets International Journal of General Systems 17 2 3 191 209 doi 10 1080 03081079008935107 Herbert J P Yao J T 2011 Game theoretic Rough Sets Fundamenta Informaticae 108 3 4 267 286 doi 10 3233 FI 2011 423 Greco Salvatore Matarazzo Benedetto Slowinski Roman 2001 Rough sets theory for multicriteria decision analysis European Journal of Operational Research 129 1 1 47 doi 10 1016 S0377 2217 00 00167 3 Grzymala Busse Jerzy 1997 A new version of the rule induction system LERS Fundamenta Informaticae 31 27 39 doi 10 3233 FI 1997 3113 Grzymala Busse Jerzy Grzymala Busse Witold 2007 An experimental comparison of three rough set approaches to missing attribute values Transactions on Rough Sets Lecture Notes in Computer Science Vol 6 pp 31 50 doi 10 1007 978 3 540 71200 8 3 ISBN 978 3 540 71198 8 Kryszkiewicz Marzena 1999 Rules in incomplete systems Information Sciences 113 3 4 271 292 doi 10 1016 S0020 0255 98 10065 8 Pawlak Zdzislaw Rough Sets Research Report PAS 431 Institute of Computer Science Polish Academy of Sciences 1981 Pawlak Zdzislaw Wong S K M Ziarko Wojciech 1988 Rough sets Probabilistic versus deterministic approach International Journal of Man Machine Studies 29 81 95 doi 10 1016 S0020 7373 88 80032 4 Pawlak Zdzislaw 1991 Rough Sets Theoretical Aspects of Reasoning About Data Dordrecht Kluwer Academic Publishing ISBN 978 0 7923 1472 1 Slezak Dominik Wroblewski Jakub Eastwood Victoria Synak Piotr 2008 Brighthouse an analytic data warehouse for ad hoc queries PDF Proceedings of the VLDB Endowment 1 2 1337 1345 doi 10 14778 1454159 1454174 Stefanowski Jerzy 1998 On rough set based approaches to induction of decision rules In Polkowski Lech Skowron Andrzej eds Rough Sets in Knowledge Discovery 1 Methodology and Applications Heidelberg Physica Verlag pp 500 529 Stefanowski Jerzy Tsoukias Alexis 2001 Incomplete information tables and rough classification Computational Intelligence 17 3 545 566 doi 10 1111 0824 7935 00162 S2CID 22795201 Wong S K M Ziarko Wojciech Ye R Li 1986 Comparison of rough set and statistical methods in inductive learning International Journal of Man Machine Studies 24 53 72 doi 10 1016 S0020 7373 86 80033 5 Yao J T Yao Y Y 2002 Induction of classification rules by granular computing Proceedings of the Third International Conference on Rough Sets and Current Trends in Computing TSCTC 02 London UK Springer Verlag pp 331 338 Ziarko Wojciech 1998 Rough sets as a methodology for data mining Rough Sets in Knowledge Discovery 1 Methodology and Applications Heidelberg Physica Verlag pp 554 576 Ziarko Wojciech Shan Ning 1995 Discovering attribute relationships dependencies and rules by using rough sets Proceedings of the 28th Annual Hawaii International Conference on System Sciences HICSS 95 Hawaii pp 293 299 Pawlak Zdzislaw 1999 Decision rules Bayes rule and rough sets New Direction in Rough Sets Data Mining and Granular soft Computing 1 9 Pawlak Zdzislaw Rough relations reports 435 Institute of Computer Science a href Template Cite journal html title Template Cite journal cite journal a Cite journal requires journal help Orlowska E 1987 Reasoning about vague concepts Bulletin of the Polish Academy of Sciences 35 643 652 Polkowski L 2002 Rough sets Mathematical foundations Advances in Soft Computing Skowron A 1996 Rough sets and vague concepts Fundamenta Informaticae 417 431 Burgin M 1990 Theory of Named Sets as a Foundational Basis for Mathematics In Structures in mathematical theories Reports of the San Sebastian international symposium September 25 29 1990 http www blogg org blog 30140 date 2005 10 26 html Burgin M 2004 Unified Foundations of Mathematics Preprint Mathematics LO 0403186 p39 electronic edition https arxiv org ftp math papers 0403 0403186 pdf Burgin M 2011 Theory of Named Sets Mathematics Research Developments Nova Science Pub Inc ISBN 978 1 61122 788 8 Cornelis C De Cock M and Kerre E 2003 Intuitionistic fuzzy rough sets at the crossroads of imperfect knowledge Expert Systems 20 5 pp260 270 Duntsch I and Gediga G 1995 Rough Set Dependency Analysis in Evaluation Studies An Application in the Study of Repeated Heart Attacks University of Ulster Informatics Research Reports No 10 Feng F 2010 Generalized Rough Fuzzy Sets Based on Soft Sets Soft Computing 14 9 pp 899 911 Grzymala Busse J 1987 Learning from examples based on rough multisets in Proceedings of the 2nd International Symposium on Methodologies for Intelligent Systems pp 325 332 Charlotte NC USA Meng D Zhang X and Qin K 2011 Soft rough fuzzy sets and soft fuzzy rough sets Computers amp Mathematics with Applications 62 12 pp4635 4645 Quafafou M 2000 a RST a generalization of rough set theory Information Sciences 124 1 4 pp301 316 Quafafou M and Boussouf M 2000 Generalized rough sets based feature selection Journal Intelligent Data Analysis 4 1 pp3 17 Nakamura A 1988 Fuzzy rough sets Notes on Multiple valued Logic in Japan 9 1 pp1 8 Pawlak Z Grzymala Busse J Slowinski R Ziarko W 1995 Rough Sets Communications of the ACM 38 11 pp88 95 Thomas K and Nair L 2011 Rough intuitionistic fuzzy sets in a lattice International Mathematical Forum 6 27 pp1327 1335 Zhang J Li T Chen H 2014 Composite rough sets for dynamic data mining Information Sciences 257 pp81 100 Zhang J Wong J S Pan Y Li T 2015 A parallel matrix based method for computing approximations in incomplete information systems IEEE Transactions on Knowledge and Data Engineering 27 2 326 339 Chen H Li T Luo C Horng S J Wang G 2015 A decision theoretic rough set approach for dynamic data mining IEEE Transactions on Fuzzy Systems 23 6 1958 1970 Chen H Li T Luo C Horng S J Wang G 2014 A rough set based method for updating decision rules on attribute values coarsening and refining IEEE Transactions on Knowledge and Data Engineering 26 12 2886 2899 Chen H Li T Ruan D Lin J Hu C 2013 A rough set based incremental approach for updating approximations under dynamic maintenance environments IEEE Transactions on Knowledge and Data Engineering 25 2 274 284Further reading EditGianpiero Cattaneo and Davide Ciucci Heyting Wajsberg Algebras as an Abstract Environment Linking Fuzzy and Rough Sets in J J Alpigini et al Eds RSCTC 2002 LNAI 2475 pp 77 84 2002 doi 10 1007 3 540 45813 1 10External links EditThe International Rough Set Society Rough set tutorial Rough Sets A Quick Tutorial Rough Set Exploration System Rough Sets in Data Warehousing Retrieved from https en wikipedia org w index php title Rough set amp oldid 1138326391, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.