Paper deep dive
A New Modeling to Feature Selection Based on the Fuzzy Rough Set Theory in Normal and Optimistic States on Hybrid Information Systems
Mohammad Hossein Safarpour, Seyed Majid Alavi, Mohammad Izadikhah, Hossein Dibachi
Intelligence
Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 94%
Last extracted: 3/13/2026, 12:56:07 AM
Summary
The paper introduces FSbuHD, a novel feature selection model for hybrid information systems that utilizes fuzzy rough set theory. By calculating a hybrid distance between objects to derive fuzzy equivalence relations, the model addresses the computational complexity and noise sensitivity of traditional intersection-based methods. It operates in normal and optimistic modes and is validated against UCI datasets, demonstrating superior efficiency and effectiveness.
Entities (4)
Relation Signals (2)
FSbuHD → performson → Hybrid Information System
confidence 95% · This research focuses on developing a feature selection approach utilizing fuzzy rough sets in hybrid information systems (HIS).
FSbuHD → utilizes → Fuzzy Rough Set Theory
confidence 95% · We proposed a new feature selection model that calculates the combined distance between objects and subsequently used this information to derive the fuzzy equivalence relation.
Cypher Suggestions (2)
Find all algorithms related to fuzzy rough set theory · confidence 90% · unvalidated
MATCH (a:Algorithm)-[:UTILIZES]->(f:Framework {name: 'Fuzzy Rough Set Theory'}) RETURN a.nameList datasets used for testing the model · confidence 85% · unvalidated
MATCH (a:Algorithm {name: 'FSbuHD'})-[:TESTED_ON]->(d:Dataset) RETURN d.nameAbstract
Abstract:Considering the high volume, wide variety, and rapid speed of data generation, investigating feature selection methods for big data presents various applications and advantages. By removing irrelevant and redundant features, feature selection reduces data dimensions, thereby facilitating optimal decision-making within decision systems. One of the key tools for feature selection in hybrid information systems is fuzzy rough set theory. However, this theory faces two significant challenges: First, obtaining fuzzy equivalence relations through intersection operations in high-dimensional spaces can be both time-consuming and memory-intensive. Additionally, this method may produce noisy data, complicating the feature selection process. The purpose and innovation of this paper are to address these issues. We proposed a new feature selection model that calculates the combined distance between objects and subsequently used this information to derive the fuzzy equivalence relation. Rather than directly solving the feature selection problem, this approach reformulates it into an optimization problem that can be tackled using appropriate meta-heuristic algorithms. We have named this new approach FSbuHD. The FSbuHD model operates in two modes - normal and optimistic - based on the selection of one of the two introduced fuzzy equivalence relations. The model is then tested on standard datasets from the UCI repository and compared with other algorithms. The results of this research demonstrate that FSbuHD is one of the most efficient and effective methods for feature selection when compared to previous methods and algorithms.
Tags
Links
- Source: https://arxiv.org/abs/2603.08900v2
- Canonical: https://arxiv.org/abs/2603.08900v2
PDF not stored locally. Use the link above to view on the source site.
Full Text
64,253 characters extracted from source content.
Expand or collapse full text
IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 Please cite this article as: Safarpour MH, Alavi SM, Izadikhah M, Dibachi H. A New Modeling to Feature Selection Based on the Fuzzy Rough Set Theory in Normal and Optimistic States on Hybrid Information Systems. International Journal of Engineering, Transactions B: Applications. 2025;38(11):2657-74. International Journal of Engineering J o u r n a l H o m e p a g e : w w w . i j e . i r A New Modeling to Feature Selection Based on the Fuzzy Rough Set Theory in Normal and Optimistic States on Hybrid Information Systems M. H. Safarpour, S. M. Alavi * , M. Izadikhah, H. Dibachi Department of Mathematics and Computer Science, Arak Branch, Islamic Azad University, Arak, Iran P A P E R I N F O Paper history: Received 08 August 2024 Received in revised form 24 October 2024 Accepted 10 November 2024 Keywords: Big Data Data Mining Feature Selection Fuzzy Rough Sets Hybrid Information System A B S T R A C T Considering the high volume, wide variety, and rapid speed of data generation, investigating feature selection methods for big data presents various applications and advantages. By removing irrelevant and redundant features, feature selection reduces data dimensions, thereby facilitating optimal decision- making within decision systems. One of the key tools for feature selection in hybrid information systems is fuzzy rough set theory. However, this theory faces two significant challenges: First, obtaining fuzzy equivalence relations through intersection operations in high-dimensional spaces can be both time- consuming and memory-intensive. Additionally, this method may produce noisy data, complicating the feature selection process.The purpose and innovation of this paper are to address these issues. We proposed a new feature selection model that calculates the combined distance between objects and subsequently used this information to derive the fuzzy equivalence relation. Rather than directly solving the feature selection problem, this approach reformulates it into an optimization problem that can be tackled using appropriate meta-heuristic algorithms. We have named this new approach FSbuHD. The FSbuHD model operates in two modes—normal and optimistic—based on the selection of one of the two introduced fuzzy equivalence relations. The model is then tested on standard datasets from the UCI repository and compared with other algorithms. The results of this research demonstrate that FSbuHD is one of the most efficient and effective methods for feature selection when compared to previous methods and algorithms. doi: 10.5829/ije.2025.38.11b.15 Graphical Abstract 1 *Corresponding Author Email: sm.alavi@iau.ac.ir (S. M. Alavi) 2658 M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 1. INTRODUCTION In today’s world, the exponential growth of data has led to the emergence of big data, characterized by four key attributes: volume, variety, velocity, and value (1-3). The immense volume of data generated encompasses diverse formats, including structured, unstructured, and semi- structured data from numerous sources. Furthermore, the high velocity at which data is produced presents significant challenges for efficient processing and analysis. Another challenge related to big data is ensuring its quality and reliability, as it is often obtained from external systems, which presents the risk of the data being inaccurate or outdated (4). Traditional production and operation modes from the industrial economy era have been fundamentally impacted by the rapid development of information technology and the Internet. The structure and operation of the network economy have evolved through the integration of data and information, involving a variety of entities, such as equipment, individuals, and enterprises. Big data technology and management tools enable enterprises to integrate information resources, transform data into knowledge, and accelerate knowledge creation (5). The potential value of big data lies in extracting valuable insights and information through accurate analysis and effective utilization (6). This potential has created a critical need for feature selection algorithms to enhance knowledge discovery across various domains. Feature selection, also known as variable selection or feature subset selection, is the process of selecting subsets of features used in model construction. Datasets often contain numerous redundant or correlated features that can be eliminated without significant information loss before applying feature selection techniques. Feature extraction, in contrast, generates new features from the original ones, producing a transformation of the original features. While the compression of new features can be more efficient, it may compromise the physical interpretability of the original features, which is a limitation (7). Feature selection is a fundamental technique in pattern recognition, machine learning, and data mining, focusing on identifying and selecting the most relevant and informative features within a dataset (8). Many feature selection methods exist, including filtering, wrapping, and embedding approaches, each with its own advantages, limitations, and applicability to various data types (9). Fuzzy rough set theory is a valuable tool for feature selection and dimensionality reduction in big data, and it has attracted significant interest. Considerable research has explored different methodologies and algorithms based on this theory. Rough set theory, specifically, provides a unique mathematical framework for identifying and selecting relevant features in datasets, accommodating both qualitative and quantitative attributes. Further examination of rough set theory reveals that many studies use it in conjunction with other computational techniques, such as neural networks or metaheuristic algorithms, thus enriching the research landscape. This interdisciplinary approach not only reflects the robustness of the theory but also highlights a trend toward integrating rough set principles with emerging methodologies in data science to enhance feature selection processes (10). Consequently, rough set theory (RST) has proven to be an effective tool for feature selection, with the capacity to handle multi- source data (11, 12). However, RST has two major drawbacks: using the similarity relation derived from rough set theory in high- dimensional spaces often leads to noise, requiring considerable time and cost. This paper aims to address these drawbacks by introducing a new fuzzy similarity relation derived from the hybrid distance between objects. This research focuses on developing a feature selection approach utilizing fuzzy rough sets in hybrid information systems (HIS). The objective is to improve the effectiveness of feature selection by incorporating the distance between elements in HIS, rather than relying solely on similarity relations. Data mining and machine learning (ML) have garnered academic attention worldwide as heuristic strategies yield promising results (13). By employing this approach, the most relevant and informative features within HIS can be identified and selected with greater accuracy and efficiency. This method has the potential to enhance the performance of various learning and recognition tasks by reducing the dataset’s dimensionality while preserving its essential characteristics. The structure of the paper is as follows: Section 2 describes related works. Section 3 introduces the fundamental concepts of fuzzy rough set theory and the computation of hybrid distance. Section 4 presents the feature selection method. Section 5 details a series of experiments performed on publicly available data, with results analyzed and compared. Finally, section 6 summarizes the findings. 2. RESEARCH BACKGROUND The classical rough set, introduced by Pawlak (14), is especially effective for processing symbolic or discrete data, allowing identification and selection of significant and relevant features for further analysis. However, in many practical applications, databases often become large and complex, resulting in a hybrid information system (HIS) that combines multiple attribute types (15, 16). In learning or recognition tasks, certain attributes M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 2659 may be irrelevant, and including these in learning algorithms can degrade performance by increasing dimensionality, leading to longer training and testing times and ultimately reducing system efficiency (17, 18). Consequently, identifying and eliminating irrelevant attributes from the HIS is crucial to improving the accuracy and efficiency of these tasks. In HIS, real-valued data usually require discretization before feature selection. However, this process risks losing the mutual information contained in the original real-valued features. To address this challenge, Dubois and Prade (19) introduced the classical fuzzy rough set (FRS), which can handle real-valued features directly. In FRS, crisp equivalence relations are replaced with membership degrees that indicate an object’s affiliation with decision classes. Over time, several extensions of FRS models have been developed (20, 21), introducing various operators to define fuzzy similarity relations. Examples include the T-similarity measure, the R- implication measure (22), and kernel-based similarity measures (23). The application of FRS models to feature selection has gained considerable attention, leading to numerous studies focusing on implementing FRS-based feature selection techniques (24, 25). Various attribute reduction algorithms have been developed to identify minimal attribute sets using discernibility matrices, such as the algorithm proposed by Chen et al. (26). In a similar approach, Chen and Yang [24] combined classical rough sets with FRS to define a discernibility relation for each symbolic and real-valued attribute, employing intersection operators to determine attribute relevance. Wang et al. (27) introduced the parameterized fuzzy relation, which enables the analysis of real-valued data by constructing a fuzzy neighborhood rough set model based on the relationship between fuzzy neighborhoods and fuzzy decisions. Fuzzy similarity relations are the core of fuzzy rough set models, with each model being defined by the specific type of fuzzy similarity relations it uses. Readers may refer to standard reference texts in discrete mathematics for further insights into various methods for establishing fuzzy similarity relations (28). The primary concept underlying these algorithms is the use of intersection operations to calculate fuzzy similarities between samples, which then serves to assess the dependence between attributes and decisions (29). For example, with a subset of attributes B and a fuzzy similarity relation 푅 퐵 induced by 퐵, the fuzzy similarity relation 푅 퐵 is determined by taking the intersection of multiple fuzzy similarity relations, specifically 푅 퐵 = ⋂ 푎∈퐵 푅 푎 , where 푅 푎 is a fuzzy similarity relation induced by attribute a. However, a potential issue with this approach of calculating the similarity between samples is that it may result in limited discrimination of membership degrees in a fuzzy relation, especially when noise is present in the dataset (30). The effect of noise can be amplified by repeated intersection operations, leading to a less accurate estimation of the fuzzy similarity relation. Revising feature reduction in fuzzy rough sets is thus motivated by the fact that the membership function of a fuzzy similarity relation, generated by intersection operations, may not accurately reflect the relationship between instances (31, 32). In traditional fuzzy similarity relations, intersection operations are used to compute the degree of similarity between samples. However, this approach has limitations in capturing the true relationship between samples accurately (33). To address this limitation, the use of distance measures is proposed. Distance measures are widely employed to characterize similarity between samples and do not depend on fuzzy sets to compute similarity relations. Therefore, fuzzy similarity relations defined through distance measures can overcome the shortcomings of those generated by intersection operations. A model for attribute reduction is constructed by introducing a distance measure into fuzzy rough sets to address the weaknesses arising from intersection operations (34). By utilizing distance measures, a more accurate and efficient approach to attribute reduction in fuzzy rough sets can be developed, improving performance in various learning and recognition tasks, such as pattern recognition, machine learning, and data mining. This approach, which incorporates hybrid distance and a new definition of similarity relationships, has led to the construction of a novel model, which is described in detail below. 3. PRELIMINARIES In this section, fundamental concepts related to fuzzy rough sets and the various attributes in hybrid information systems are reviewed. Additionally, calculating distances between these attributes is a critical step in data analysis and will be discussed in detail. 3. 1. Fuzzy Rough Set Definitin 1. An information system is defined as a ⟨ 푈,퐴∪퐷,푉 ⟩ , where 푈 is a non-empty finite set of objects, 퐴 is a non-empty finite set of attributes, and 퐷 denotes the set of decision attributes, 퐴∩퐷=∅. Each attribute 푎∈퐴 or 푑∈퐷 is associated with a set 푉 푎 or 푉 푑 of its value, called the domain of 푎 or 푑, respectively (34). The rough set theory describes a crisp subset of a universe by two definable subsets called lower and upper approximations. Definition 2. Let 푈 be a non-empty finite universe, then 푅⊆푈×푈 is called an equivalence relation on 푈, if for ∀푥,푦,푧∈푈, (35) 2660 M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 (1) 푅푒푓푙푒푥푖푣푖푡푦:푅(푥,푥)=1, (2) 푆푦푚푒푡푟푦:푅(푥,푦)=푅(푦,푥), (3) 푇푟푎푛푠푖푡푖푣푖푡푦:푅(푥,푦)=푅(푦,푧)⇒ 푅(푥,푦)= 푅(푥,푧) (1) The equivalence relation partitions the universe into a family of disjoint subsets called equivalence classes. [푥] 푅 denotes an equivalence class of 푅 containing an element 푥∈푈. Definition 3. Let 푈 be a non-empty finite universe, 푅⊆ 푈×푈 is an equivalence relation on 푈, 푈/푅 denotes the family of all equivalence classes 푅. For any 푋⊆푈, the lower and the upper approximations of 푋 are defined as follows, respectively (19): 푅푋=푥∈푈|[푥] 푅 ⊆푋 푅푋=푥∈푈|[푥] 푅 ∩푋≠∅. (2) Definition 4. Let 푈 be a non-empty finite universe, and 퐴(.):푈→[0,1] be a mapping. 퐴 is called a fuzzy set on 푈, for any 푥∈푈, and 퐴(푥) is called membership degree of 푥 to 퐴 (36). Definition 5. A fuzzy number 퐴 is a fuzzy set of the real line with a normal, convex, and continuous membership function of bounded support (37). Definition 6. A fuzzy set 퐴 is called trapezoidal fuzzy number with tolerance interval [푏,푐], left width (푏−푎) and right width (푑−푐) if its membership function has the following form (37, 38): 퐴(푡)= 1−(푏−푡)/푏−푎, 푖푓 푎≤푡≤푐, 1, 푖푓 푏≤푡≤푐, 1−(푡−푐)/푑−푐, 푖푓 푏≤푡≤푑, 0, 푖푓 표푡ℎ푒푟푤푖푠푒. (3) The notation A = (a, b, c, d) shows a trapezoidal fuzzy number, if b=c then the number fuzzy is called triangular fuzzy number. Definition 7. Let 푈 be a non-empty finite universe, and 푅∈퐹(푈×푈), 푅 is called a fuzzy equivalence relation on 푈, if for ∀푥,푦,푧∈푈, (35): (1) 푅푒푓푙푒푥푖푣푖푡푦:푅(푥,푥)=1, (2) 푆푦푚푒푡푟푦:푅(푥,푦)=푅(푦,푥), (3) 푀푖푛−푡푟푎푛푠푖푡푖푣푖푡푦:푚푖푛푅(푥,푦),푅(푦,푧)≤ 푅(푥,푧) (4) Definition 8. A binary operator 푇 on the unit interval [0,1]×[0,1] is said to be a triangular norm. If ∀푎,푏,푐∈ [0,1], the following conditions are satisfied (39): (1) 퐶표푚푢푡푎푡푖푣푖푡푦:푇(푎,푏)=푇(푏,푎), (2) 퐴푠표푐푖푎푡푖푣푖푡푦:푇(푇(푎,푏),푐)=푇(푎,푇(푏,푐)), (3) 푀표푛표푡표푛푖푐푖푡푦:푎≤푐,푏≤푑⇒푇(푎,푏)≤ 푇(푐,푑), (4) 퐵표푢푛푑푎푟푦 푐표푛푑푖푡푖표푛:푇(푎,1)=푎,푇(1,푎)=푎 (5) As an example, 푇 푝 (푎,푏)=푎푏 is called probability t- norm. A binary operator 푆 on the unit interval [0,1]×[0,1] is a triangular-conorm (shortly, t-conorm or s-norm) if it is increasing, associative, and commutative and satisfies the boundary condition of the form 푆(푎,0)=푎,∀푎∈[0,1]. Definition 9. Let 푈 be a non-empty finite universe, and 푅∈퐹(푈×푈), 푅 is called a T-fuzzy equivalence relation on 푈, if for ∀푥,푦,푧∈푈, (1) 푅푒푓푙푒푥푖푣푖푡푦:푅(푥,푥)=1, (2) 푆푦푚푒푡푟푦:푅(푥,푦)=푅(푦,푥), (3) 푇−푡푟푎푛푠푖푡푖푣푖푡푦:푇(푅(푥,푦),푅(푦,푧))≤ 푅(푥,푧). (6) Definition 10. Let 푅 be a fuzzy equivalence relation on 푈 and 푋 be a fuzzy subset on 푈. The fuzzy lower and upper approximations of 푋 is defined as follows, respectively (40): 푅푋 ( 푥 ) =푖푛푓 푦∈푈 푆 푝 1−푅 ( 푥,푦 ) ,푋 ( 푦 ) , 푅푋 ( 푥 ) =푠푢푝 푦∈푈 푇 푝 푅 ( 푥,푦 ) ,푋 ( 푦 ) . (7) Definition 11. Let ⟨ 푈,퐴∪퐷,푉 ⟩ be the information system, where 푈=푥 1 ,푥 2 ,...,푥 푛 is a finite set of objects and 퐴=푎 1 ,푎 2 ,...,푎 푚 is the condition attribute set. 푅 푎 is defined a fuzzy relation for each condition attributes 푎∈퐴, 퐷=푑 is the decision attributes, the sample set 푈 is divided into r crisp equivalence classes by 퐷, 푈/퐷= 퐷 1 ,퐷 2 ,...,퐷 푟 ( 1≤푗≤푟 ) (19). 퐵 is a subset of 퐴, denoted as 퐵⊆퐴, and 푅 푎 is the fuzzy similarity relation for each 푎∈퐵. Let 푅 퐵 = ⋂ 푎∈퐵 푅 푎 , (8) Then, 푅 퐵 is a fuzzy similarity relation on 푈. Furthermore, the membership function of the decision class 퐷 푗 is: 퐷 푗 (푥)= 1, 푥∈퐷 푗 , 0, 푂푡ℎ푒푟푤푖푠푒. (9) To handle the uncertainty of the decision feature and according to the Relation 7, the classical FRS introduces the definitions of the fuzzy rough approximations as follows: 푅 퐵 (퐷 푗 )(푥)=푖푛푓 푦∉퐷 푗 1−푅 퐵 (푥,푦),퐷 푗 ∈푈/퐷, 푅 퐵 (퐷 푗 )(푥)=푠푢푝 푦∈퐷 푗 푅 퐵 (푥,푦),퐷 푗 ∈푈/퐷 (10) where 푅 퐵 (퐷 푗 )(푥) and 푅 퐵 (퐷 푗 )(푥) are called the fuzzy lower and upper approximations of 퐷 푗 , respectively. 푅 퐵 (퐷 푗 )(푥) denotes that the membership of 푥 certainly belonging to 퐷 푗 equals to the minimum of the dissimilarities between 푥 and all objects from the sample domain 푈−퐷 푗 . 푅 퐵 (퐷 푗 )(푥) means that the membership M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 2661 of 푥 can belong to 퐷 푗 equals the maximum of the similarities between all objects of 퐷 푗 . Definition 12. The definition of the fuzzy positive region 푃푂푆 퐵 (퐷) of the condition attribute subset concerning the decision attribute 퐷 can be expressed as follows (41): 푃푂푆 퐵 (퐷)= ⋃ 푟 푗=1 푅 퐵 (퐷 푗 ) (11) Definition 13. The fuzzy dependency function can be calculated using the definition of the fuzzy positive region, as follows: The ratio of the positive region to all objects in the feature space is defined as 훾 퐵 (퐷), 훾 퐵 (퐷)= |푃푂푆 퐵 (퐷)| |푈| (12) where |.| denotes the cardinality of a set. It is important to note that 훾 퐵 (퐷) falls within the range of 0 to 1, and it is utilized to assess the importance of a subset of features. The fuzzy dependency, conversely, is a measure of the classification ability of a particular feature subset 퐵. Definition 14. A hybrid information system is defined as a 퐻퐼푆= ⟨ 푈,퐴∪퐷,푉 ⟩ , where 푈 is a non-empty finite set of objects, and 퐴 is a non-empty finite set and consists of different types of attributes, let 퐴=퐴 푟 ∪퐴 푠 ∪퐴 푐 ∪ 퐴 푏 ∪퐴 푙 , 퐴 푟 is the real-valued attribute set, 퐴 푠 is the set- valued attribute set, 퐴 푐 is the categorical attribute set, 퐴 푏 is the boolean attribute set and 퐴 푙 is the linguistic variable attribute set and, 퐴 푟 ∩퐴 푠 =∅, 퐴 푟 ∩퐴 푐 =∅, 퐴 푟 ∩퐴 푏 = ∅, 퐴 푟 ∩퐴 푙 =∅, 퐴 푠 ∩퐴 푐 =∅, 퐴 푠 ∩퐴 푏 =∅, 퐴 푠 ∩퐴 푙 = ∅, 퐴 푐 ∩퐴 푏 =∅, 퐴 푐 ∩퐴 푙 =∅, 퐴 푏 ∩퐴 푙 =∅, D denotes the set of decision attributes, 퐴∩퐷=∅, and 푉 shows the domain set of attributes (39). 3. 2. Hybrid Distance To construct an efficient distance function among objects in a Health Information System (HIS), it is essential to develop a hybrid distance function. This necessity arises because HIS may contain various types of attributes, including Boolean, real- valued, set-valued, and linguistic variable attributes. To address the value differences across these diverse attributes, the following definition for value difference is proposed. The concept of a linguistic variable provides a means to approximately characterize phenomena that are too complex or ill-defined to be adequately described in conventional quantitative terms. Specifically, treating the illness rate as a linguistic variable with values such as "minor," "moderate," and "severe" enables the application of fuzzy logic. In this research, the linguistic variable attribute distance (LD) is defined in two steps. In the first step, the conversion of the linguistic variable into a trapezoidal fuzzy number is carried out. For instance, the linguistic variable “T(illness rate) = Minor, Moderate, Severe” can be represented as a trapezoidal fuzzy number, as illustrated in Figure 1. Figure 1. Linguistic variable of the illness rate as the trapezoidal fuzzy number In the second step, defuzzification is performed, which involves converting a collection of fuzzy values or membership functions—known as a fuzzy set—into a single numerical or crisp value. While several methods are available for defuzzification, the centroid method is utilized in this study. This method calculates the crisp value as the center of gravity or the weighted average of the trapezoidal fuzzy number. The centroid method typically strikes a good balance between accuracy and simplicity, making it a widely used approach in various applications. Assuming that 휇 푁 ̃ be a fuzzy set then the formula for the centroid method is expressed as follows (42): 푐 휇 푁 ̃ = ∫ 푥휇 푁 ̃ (푥)푑푥 ∫ 휇 푁 ̃ (푥)푑푥 (13) In the special case, if A= (a, b, c, d) is a trapezoidal fuzzy number, then the centroid method can be written as follows: 푐 퐴 =(3푎+푏+3푐+2푑)/9 (14) Definition 15. Let 퐻퐼푆= ⟨ 푈,퐴∪퐷,푉 ⟩ , for each 푥,푦∈ 푈, then 푎(푥), 푎(푦) denote the values of 푥, 푦 under 푎, respectively. Relations 12-15 show the attributes distance in four different cases (43). 1) Let 푎 be a boolean attribute, for each 푎∈퐴. The boolean attribute distance (퐵퐷) is defined as follows: 퐵퐷(푎(푥),푎(푦))= 0, 푎(푥)=푎(푦), 1, 푎(푥)≠푎(푦). (15) 2) Let 푎 be a real-valued attribute, for each 푎∈퐴. The real-valued attribute distance (푅퐷) is defined as follows: 푅퐷(푎(푥),푎(푦))= |푎(푥)−푎(푦)| 4휎 푎 (16) where 휎 푎 is the standard deviation under the attribute 푎. 3) Let 푎 be a set-valued attribute, for each 푎∈퐴 , and 푉 푎 is the domain of 푎. The set-valued attribute distance (푆퐷) is defined as follows: 푆퐷(푎(푥),푎(푦))=1− 푎(푥)∩푎(푦) 푠 (17) 2662 M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 where 푠 is the maximum cardinal number (cardinality) in the subset of 푉 푎 . 4) Let 푎 be a linguistic variable attribute, for each 푎∈퐴. The linguistic variable attribute distance (퐿퐷) is defined as follows, there are different formulas to find the distance between two fuzzy numbers, and in this study, the following formula is used. 퐿퐷(푎(푥),푎(푦))= |푐 1 −푐 2 | 4휎 푎 (18) 푐 1 is a defuzzification of the linguistic variable of 푎(푥), and 푐 2 is a defuzzification of the linguistic variable of 푎(푦), respectively. 휎 푎 is the standard deviation under the attribute 푎 when the linguistic variables 푎(푥) and 푎(푦) are de-fuzzified. To handle hybrid attributes, a hybrid distance (퐻퐷) is defined based on the different attributes distance as mentioned earlier as follows. Definition 16. Given a 퐻퐼푆, for each 푥,푦∈푈, the hybrid distance (퐻퐷) is defined as (39): 퐻퐷(푥,푦)= √ ∑ 푚 푘=1 푑 2 (푎 푘 (푥),푎 푘 (푦)) (19) whereas the number of attributes is denoted by 푚, the above content can be summarized as follows (39, 43, 44): 푑(푎(푥),푎(푦))= 퐵퐷(푎(푥),푎(푦)),푎 푖푠 푎 푏표푙푒푎푛, 푆퐷(푎(푥),푎(푦)),푎 푖푠 푎 푠푒푡−푣푎푙푢푒푑, 푅퐷(푎(푥),푎(푦)),푎 푖푠 푎 푟푒푎푙−푣푎푙푢푒푑, 퐿퐷(푎(푥),푎(푦)),푎 푖푠 푎 푙푖푛푔푢푖푠푡푖푐 푣푎푟푖푎푏푙푒. (20) 4. PROPOSED METHOD In recent years, fuzzy rough set theory has found extensive applications across various fields due to its effectiveness in handling uncertainty and vagueness in data. However, a limitation of the current fuzzy dependency function is its inability to accurately classify subsets of attributes in high-dimensional data. This challenge arises from the intersection operation used to compute fuzzy similarity relations when determining rough approximations, which may not effectively discriminate the membership degrees of these relations in a high-dimensional sample space. As a result, the true relationship between samples may not be accurately represented by the membership function of a fuzzy similarity relation. 4. 1. Creating Fuzzy Information Granules Recently, Hu et al. (45) proposed a combination of the Gaussian kernel function and Euclidean distance to create fuzzy information granules. However, it was noted that the Euclidean distance was inadequate for directly handling hybrid attributes. To address this limitation, Zeng et al. (39) introduced a hybrid distance measure that integrates the Euclidean distance with HD (hybrid distance). This approach enabled the effective construction of fuzzy information granules capable of managing hybrid attributes. Let 푈 represent a non-empty finite universe, with samples expressed as m-dimensional vectors. ∀푥 푖 ,푥 푗 ∈푈, 푥 푖 = ⟨ 푥 푖1 ,푥 푖2 ,...,푥 푖푚 ⟩ , 푥 푗 = ⟨ 푥 푗1 ,푥 푗2 ,...,푥 푗푚 ⟩ . The Gaussian kernel function 푘(푥 푖 ,푥 푗 )=푒푥푝 (− ||푥 푖 −푥 푗 || 2 2휎 2 ) (21) can be used to compute the similarity between 푥 푖 and 푥 푗 in wich ||푥 푖 −푥 푗 || is the 퐻퐷 distance between 푥 푖 and 푥 푗 . The following relations hold: (1) 푘(푥 푖 ,푥 푗 )∈[0,1], (2) 푘(푥 푖 ,푥 푗 )=푘(푥 푗 ,푥 푖 ), (3) 푘(푥 푖 ,푥 푖 )=1. and the fuzzy relation 푅 퐺 computed with Gaussian kernel is given as follows: 푅 퐺 (푥 푖 ,푥 푗 )=푒푥푝 (− ∑ 푚 푘=1 푑 2 (푎 푘 (푥 푖 ),푎 푘 (푥 푗 )) 2휎 2 ) (22) Theorem 1. According to Sun et al. (21) the Gaussian kernel 푘:푈×푈→[0,1] satisfies 푇 푝 −푡푟푎푛푠푖푡푖푣푒, and the fuzzy relation 푅 퐺 is a 푇 푝 −푒푞푢푖푣푎푙푒푛푐푒 relation. Proof. 1) 푅 퐺 (푥 푖 ,푥 푖 )=1 then 푅푒푓푙푒푥푖푣푖푡푦 2) 푅 퐺 (푥 푖 ,푥 푗 )=푅 퐺 (푥 푗 ,푥 푖 ) then 푆푦푚푒푡푟푖푐 3) ||푥 푖 −푥 푗 +푥 푗 −푥 푟 || 2 <||푥 푖 −푥 푗 || 2 +||푥 푗 −푥 푟 || 2 ⇒ −||푥 푖 −푥 푗 || 2 −||푥 푗 −푥 푟 || 2 <−||푥 푖 −푥 푟 || 2 ⇒ 푒푥푝( −||푥 푖 −푥 푗 || 2 2휎 2 ).푒푥푝( −||푥 푗 −푥 푟 || 2 2휎 2 )<푒푥푝( −||푥 푖 −푥 푟 || 2 2휎 2 ) ⇒ 푅 퐺 (푥 푖 ,푥 푗 ).푅 퐺 (푥 푗 ,푥 푟 )<푅 퐺 (푥 푖 ,푥 푟 ) ⇒ 푇 푝 (푅 퐺 (푥 푖 ,푥 푗 ),푅 퐺 (푥 푗 ,푥 푟 ))<푇 푝 (푅 퐺 (푥 푖 ,푥 푟 )) then 푇− 푡푟푎푛푠푖푡푖푣푖푡푦 4. 2. Creating Similarity Relations As previously mentioned, determining similarity between samples is a significant challenge in fuzzy rough set theory, and addressing this challenge requires the use of distance measures between samples. The decision- making information system employed in this study utilizes a hybrid system that incorporates linguistic variables along with multiple features. To fully understand the proposed model and accurately compute the similarity distance between samples, let us examine a detailed example that utilizes the formulas discussed earlier. Based on Table 1, the HD matrix can be calculated. According to the introduced formula, the results hold: 1) 푎 1 is the linguistic variable attribute. Convert linguistic variable into a trapezoidal fuzzy number: M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 2663 TABLE 1. A hybrid information system U Illness rate (풂 ퟏ ) Pain (풂 ퟐ ) Fever temperature (풂 ퟑ ) Syndrome (풂 ퟒ ) D 푥 1 Minor Yes 39.2 퐶,푅,퐴 Flu 푥 2 Minor Yes 39.5 퐶,푅,퐴 Flu 푥 3 Moderate No 39.0 퐶 Flu 푥 4 Minor Yes 38.8 퐶,푅,퐴 Rhinitis 푥 5 Severe No 37.3 푅 Rhinitis 푥 6 Severe No 36.8 푅,퐴 Health 푥 7 Moderate Yes 39.0 퐶,푅,퐴 Health Minor= (0, 1, 1, 3), Moderate= (2, 4, 6, 6), Severe= (4, 7, 10, 10), 퐶 1 =(3푎+푏+3푐+2푑)/9=1.11, 퐶 4 =(3푎+푏+3푐+ 2푑)/9=1.11, 퐿퐷(푎 1 (푥 1 ),푎 1 (푥 4 ))= |퐶 1 −퐶 4 | 4휎 푎 = |1.11−1.11| 4×2.94 =0, 2) 푎 2 is a boolean attribute, 퐵퐷(푎 2 (푥 1 ),푎 2 (푥 4 ))=0, 3) 푎 3 is a real-valued attribute, 푅퐷(푎 3 (푥 1 ),푎 3 (푥 4 ))= |푎 3 (푥 1 )−푎 3 (푥 4 )| 4휎 푎 = |39.2−38.8| 4×1.03 =0.09, 4) 푎 4 is a set-valued attribute, 푆퐷(푎 4 (푥 1 ),푎 4 (푥 4 ))= 1− 푎 4 (푥 1 )∩푎 4 (푥 4 ) 푠 =1− 3 3 =0, And finally, 퐻퐷(푥 1 ,푥 4 )= √ ∑ 4 푘=1 푑 2 (푎 푘 (푥 1 ),푎 푘 (푥 4 ))=0.09. At the end of this step,an HD matrix is obtained that displays the hybrid distance between the objects in the U set. This matrix is a valuable tool for determining the similarity between the objects in the 푈 set. 퐻퐷= [ 00.071.230.091.391.320.28 01.230.681.421.350.72 01.441.111.161.41 01.191.331.04 00.511.49 01.26 0 ] In the next step, the 푇 푃 −푒푞푢푖푣푎푙푒푛푐푒 relation 푅 퐺 , is calculated using the Gaussian kernel function. The fuzzy relation between each two samples can be computed by Equation 22. Let 휎 2 =0.4. For example: 푅 퐺 (푥 1 ,푥 4 )=푘(푥 1 ,푥 4 )=푒푥푝(− 퐻퐷 2 (푥 1 ,푥 4 ) 2×휎 2 )= 푒푥푝(− 0.09 2 2×0.4 )=0.98 Therefore, 푅 퐺 = [ 10.990.150.980.080.110.90 10.140.550.070.090.51 10.070.210.180.08 10.160.100.25 10.710.06 10.13 1 ] Then, the equivalence classes are: [푥 1 ] 푅 퐺 = 1 푥 1 + 0.99 푥 2 + 0.15 푥 3 + 0.98 푥 4 + 0.08 푥 5 + 0.11 푥 6 + 0.90 푥 7 [푥 2 ] 푅 퐺 = 0.99 푥 1 + 1 푥 2 + 0.14 푥 3 + 0.55 푥 4 + 0.07 푥 5 + 0.09 푥 6 + 0.51 푥 7 . . . [푥 7 ] 푅 퐺 = 0.90 푥 1 + 0.51 푥 2 + 0.08 푥 3 + 0.25 푥 4 + 0.06 푥 5 + 0.13 푥 6 + 1 푥 7 According to Relation 7, the fuzzy upper approximation of the relation of 푅 퐺 is defined as follows: 푅 퐺 = [ 10.990.150.980.160.120.90 10.150.970.90.110.89 10.150.210.180.14 10.160.110.88 10.710.09 10.13 1 ] The fuzzy lower approximation of the relation of 푅 퐺 (푅 퐺 ) can be obtained in a similar way, elements of 푅 퐺 indicate certaintly version of [푥] under fuzzy relation 푅 and elements of 푅 퐺 indicate possiblity version of [푥] under fuzzy relation 푅. The important point is that 푅 퐺 is not a fuzzy equivalence relation. Any of the above relations (푅 퐺 ,푅 퐺 ) can be used to create a new model based on the opinion of the decision- maker. The state of model is called normal when 푅 퐺 is used for modeling and state is called optimistic when 푅 퐺 is used for modeling. 2664 M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 4. 3. Creating a Set of Acceptable Indexes of the Problem Since 푇 푝 is a special 푇−푛표푟푚, according to Theorem 1, it can be used 푅 퐺 as a similarity relation. According to the obtained similarity equivalence relation and considering the classification of objects, Table 2 can be drawn as follows: In Table 2, the similarity of objects can be seen by considering their classification. The final goal of this research is to select the features, and it is intended to select the features in such a way that the classification of the objects does not deteriorate, and calculations are applied in such a way that fewer calculations are done in the final model presented. Therefore, for making a model, the objects that are in the same class will not be considered and a constant number is considered as the degree of similarity criterion, which is displayed with 훿 (0≤훿≤1). The choice of delta size depends on the opinion of the decision-maker, according to the 푅 퐺 relation and delta size, two sets 퐺 1 and 퐺 2 which are given below are made: Let 푈 be a non-empty finite universe and divided into r crisp equivalence classes by 퐷, 퐷=푑 is the decision attributes, 푈/퐷=퐷 1 ,퐷 2 ,...,퐷 푟 . If (푥 푖 ,푥 푗 )∈퐷 푙 ×퐷 푚 (푙≠푚) then, 퐺 1 =(푥 푖 ,푥 푗 )|푅(푥 푖 ,푥 푗 )≤훿, 퐺 2 =(푥 푖 ,푥 푗 )|푅(푥 푖 ,푥 푗 )>훿 (23) For example, according to Table 2, 퐷 1 =푥 1 ,푥 2 ,푥 3 ,퐷 2 =푥 4 ,푥 5 ,퐷 3 =푥 6 ,푥 7 . if 훿=0.85, then 퐺 1 =(푥 1 ,푥 5 ),(푥 1 ,푥 6 ),(푥 2 ,푥 4 ),(푥 2 ,푥 5 ),(푥 2 ,푥 6 ),(푥 2 ,푥 7 ),(푥 3 ,푥 4 ),(푥 3 ,푥 5 ),(푥 3 ,푥 6 ), (푥 3 ,푥 7 ),(푥 4 ,푥 6 ),(푥 4 ,푥 7 ),(푥 5 ,푥 6 ),(푥 5 ,푥 7 ). 퐺 2 =(푥 1 ,푥 4 ),(푥 1 ,푥 7 ). Considering that the set of 퐺 2 represents the objects that belong to the fuzzy rough set and the similarity of the features becomes fixed or more by feature selection, 퐺 2 is to be ignored. To describe the proposed new method, the set 푆 훿 is defined: 푆 훿 =(푖,푗)|(푥 푖 ,푥 푗 )∈퐺 1 (24) 4. 4. Creating Constraint and Objective Function of the Problem By reducing each feature, the distance between objects can be decreased or maintained at the same level, and this can be utilized as an effective method for measuring the similarity between objects. The greater the similarity between the objects, the smaller the distance value. Therefore, the similarity relation between objects can be increased or maintained at the same level by reducing each feature. This is because each feature represents a certain aspect or characteristic of the objects being analyzed, and features that do not contribute significantly to the similarity between objects can be removed. By focusing on the most informative and relevant features, we can better understand the underlying patterns and relations in the data, and more accurate and efficient models can be developed. By assuming a feature reduction and removing one or several features, the objects are in the same classes, and the relations whose value is greater than the measure of the degree of similarity 훿 are ignored. But other relations would still have to be inferior to the criterion of the degree of similarity. This article is presented as a constraint in problem modeling as below: 푒 (− 휒 푘 퐻퐷 2 (푥 푖 ,푥 푗 ) 2휎 2 ) ≤훿→− 휒 푘 퐻퐷 2 (푥 푖 ,푥 푗 ) 2휎 2 ≤ln훿,(푖,푗)∈푆 훿 (25) TABLE 2. The similarity between objects with considering their classes 1 2 3 Class Object 풙 ퟏ 풙 ퟐ 풙 ퟑ 풙 ퟒ 풙 ퟓ 풙 ퟔ 풙 ퟕ 1 푥 1 1 0.99 0.15 0.98 0.08 0.11 0.90 푥 2 1 0.14 0.55 0.07 0.09 0.51 푥 3 1 0.07 0.21 0.18 0.08 2 푥 4 1 0.16 0.10 0.25 푥 5 1 0.71 0.06 3 푥 6 1 0.13 푥 7 1 M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 2665 Let 휒 푘 ∈0,1,휒 푘 =1 means that the feature is selected and 휒 푘 =0 means that the feature is not selected. 푘= 1,2,...,푝, and 푝 represents the number of attributes. The goal of this problem is to reduce features, so the objective function will be formulate as follows: 푀푖푛 ∑ 푝 푘=1 휒 푘 (26) 4. 5. Obtaining Optimization Problem Based on the information provided above, the problem will be modeled as follows: The set of objects and their corresponding features will be defined, and a hybrid distance measure will be employed to determine the relationships between these objects. Additionally, excluding objects that belong to the same classes, as well as relationships with values greater than the degree of similarity, can simplify the model and help prevent overfitting. Subsequently, the constraints and objective function of the problem will be generated, and a novel model will be defined as follows: 푀푖푛 ∑ 푝 푘=1 휒 푘 푠.푡.− ∑ 푃 푘=1 휒 푘 푑 2 (푎 푘 (푥 푖 ),푎 푘 (푥 푗 )) 2휎 2 ≤ln훿 (푖,푗)∈푆 훿 ,휒 푘 ∈0,1 (27) The problem can now be addressed by selecting an appropriate meta-heuristic algorithm to find an approximate solution to the proposed model. In the following sections, we will apply the black hole (BH) algorithm, which has been assessed as suitable for such problems. 5. EXPERIMENTAL EVALUATION In this research, we will validate the effectiveness of the proposed model using public datasets. To achieve this, datasets will be downloaded from the UC-Irvine Machine Learning Repository, and classification algorithms will be applied to these datasets after employing various feature selection algorithms to evaluate the model's effectiveness. Three types of experimental results will be presented: 1. The feature selection results of FSbuHD in two states: normal and optimistic. 2. A comparison of the previously tested algorithms with FSbuHD. 3. A comparison of accuracy, precision, recall, and the Matthews correlation coefficient. Detailed explanations of these results are provided below. 5. 1. Metaheuristic Algorithms The goal of feature selection is to identify the smallest possible number of attributes that can achieve the same or similar classification accuracy as the entire set of attributes. While there may be several possible subsets of attributes that can accomplish this, it is sufficient to find just one of them. The black hole optimization algorithm is a type of meta-heuristic algorithm inspired by nature. It simulates the phenomenon of black holes, which are formed from massive stars with strong gravitational forces. The algorithm begins with a specific population size of potential solutions, which are then evaluated. The best solution is selected as the "black hole" (46, 47). A total of eight datasets were utilized from the UCI machine learning repository to assess the effectiveness of the model. These datasets comprise a variety of features including single and mixed types, and also consist of both binary and multiple classes. A summary of fundamental descriptive characteristics for each of the datasets used is provided in Table 3. Evaluation of feature selection using different algorithms (FARNem, WARA, CfsSubsetEval, and RSFSAID) is shown in Table 4. FARNeM is used, 훿, to determine the neighborhood set of objects, WARA is used a parameter, 휖, which serves as a threshold for feature selection, CfsSubsetEval, can be implemented without any parameters, and RSFSAID, is used a parameter, 훿, which is determined in the same way as in the FARNeM. For the experiments, the values of 훿 and 휖 were fixed to increase from 0 to 0.99 in increments of 0.01. The best results were registered (48). The results of the proposed model using 훿 and 휎= 0.2 in different data sets in two normal and optimistic states are shown in the last two columns of Table 4 for comparison with other algorithms. The value of 훿 is considered from 0 to 0.9 incrementally, and with step 0.1, the best results were registered. In Table 5, the set of selected features in different datasets that were chosen by the FSbuHD method can be seen. 5. 2. Result of Feature Selection Tables 4 and 5 present the results of FSbuHD in both normal and optimistic states. The evaluation of feature selection using FARNeM, WARA, CfsSubsetEval, RSFSAID, and FSbuHD for different datasets is detailed in Table 4. The results displayed in Table 4 indicate that FSbuHD is a highly effective method for selecting a subset of features. It is important to note that the features selected by different algorithms may vary, as the feature evaluation procedures and selection algorithms employed can differ. However, the efficacy of any feature selection algorithm can only be verified through the use of classification algorithms. By evaluating the performance of the selected features with classification algorithms, we can assess the effectiveness of the feature selection algorithm in question. 2666 M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 Figure 2. Framework of feature selection TABLE 3. Summary of the datasets NO. Name # I # F # C Data Type 1 crx 690 15 2 Hybrid 2 australian 690 14 2 Hybrid 3 heart 270 13 2 Hybrid 4 ionosphere 351 34 2 Hybrid 5 segment 2308 19 7 Numerical 6 wpbc 198 33 2 Numerical 7 zoo-3 101 16 7 Numerical 8 wdbc 569 30 2 Numerical #I, denotes the number of instances. #F, denotes the number of features. #C, denotes the number of classes. TABLE 4. Feature selection using different algorithms NO. Dataset FARNem-F (#) WARA-F (#) CfsSubset Eval-F (#) RSFSAID -F (#) FSbuHD-F(#) Normal state Optimistic state 1 crx 14 13 7 9 6 6 2 australian 14 12 7 6 6 4 3 heart 11 11 8 5 7 6 4 ionosphere 16 8 17 8 15 14 5 segment 10 11 3 3 10 9 6 wpbc 12 1 13 12 8 5 7 zoo-3 5 5 7 6 5 4 8 wdbc 8 5 11 8 13 13 Define problem and black hole parameters Create initial population For every star or candidate, calculate the objective function Select the best candidates as black hole Location of each star is updated Replacement of new star with old star Stopping criteria? Report solution M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 2667 TABLE 5. Set of feature selection using FSbuHD NO. Dataset State FSbuHD-F(#) FSbuHD-F(S) 1 crx Normal 6 [4,9,10,11,12,15] Optimistic 6 [2,5,9,10,11,14] 2 australian Normal 6 [4,5,8,10,12,14] Optimistic 4 [5,8,12,14] 3 heart Normal 7 [1,2,7,8,9,12,13] Optimistic 6 [2,3,7,9,12,13] 4 ionosphere Normal 15 [2,3,5,6,7,8,11,18,20,21,22,23,24,29,34] Optimistic 14 [1,2,3,5,7,11,13,16,18,19,21,22,33,34] 5 segment Normal 10 [1,2,3,5,8,11,14,16,18,19] Optimistic 9 [2,3,13,14,15,16,17,18,19] 6 wpbc Normal 8 [1,2,8,21,23,24,26,31] Optimistic 5 [2,7,13,17,20] 7 zoo-3 Normal 5 [2,3,4,6,9] Optimistic 4 [4,6,9,10] 8 wdbc Normal 13 [1,2,3,7,8,10,14,15,18,23,24,27,28] Optimistic 13 [1,2,3,7,8,13,14,15,18,23,25,26,28] F(S), denotes the set of features that are selected. Therefore, it is imperative to conduct a thorough analysis of the selected features using various classification techniques to ensure that the proposed feature selection approach is both accurate and reliable. Overall, the results presented in Table 4 suggest that FSbuHD is a promising method for feature selection and merits further investigation. 5. 3. Evaluation by Using the Classification Performance In this research, three different algorithms—namely, Linear SVM, KNN, and Complex Tree—have been employed to evaluate the effectiveness of the proposed model for feature selection. The results of feature selection across different algorithms and various datasets have been analyzed and compared. Linear SVM is utilized for linearly separable data, while K-nearest neighbors (KNN) is a non-parametric, supervised learning classifier that makes classifications or predictions based on the proximity to other data points (with k=3 considered in this experiment). The Complex Tree is a hierarchical decision support model that employs decision trees to represent an algorithm containing conditional control statements, chance event outcomes, resource costs, and utility. Experiments were conducted using a five-fold cross- validation framework. Each dataset was divided into five sections, with one section designated as the testing set and the other four used for training. Each section is rotated as the testing set, resulting in five experiments for each dataset. For each experiment, the training set undergoes preprocessing using feature selection algorithms. A classifier is then constructed based on the training set after feature selection. Finally, the performance of the classifier is evaluated by applying it to the testing set. In a binary classification problem, the classification results are labeled as positive (P) and negative (N), leading to the creation of a confusion matrix based on these labels. The confusion matrix is a standardized table that compares the number of instances in different groups according to binary classification. Confusion Matrix Prediction Actual Positive TP FN Class Negative FP TN The accuracy, precision, recall, and Matthews correlation coefficient (MCC) are used to evaluate different algorithms and determine which one is most effective in comparison to others. All the metrics mentioned above are derived from the confusion matrix. This section analyzes the accuracy, precision, recall, and MCC for FARNeM, WARA, CfsSubsetEval, RSFSAID, and FSbuHD (normal state). From the eight datasets presented in Table 3, three datasets are selected at random for evaluation, and the results are obtained for these subsets. It is possible to calculate the performance metrics for all the datasets 2668 M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 listed in the table, although some calculations may have been omitted for brevity. The Classification Learner app in MATLAB R2017a is utilized to prepare Tables 6, 7, 8, and 9. 5. 4. Performance Measurement 1) Accuracy: The percentage of events that were successfully predicted is compared with all the predictions. 퐴퐶= 푇푃+푇푁 푇푃+푇푁+퐹푃+퐹푁 (28) 2) Precision: The ratio of all true positives to all positive predictions is presented as follows, 푃푟푒푐푖푠푖표푛= 푇푃 푇푃+퐹푃 (29) 3) Recall: The division of true positives by positive results indicates how many potential positives were found by the model (49). 푅푒푐푎푙= 푇푃 푇푃+퐹푁 (30) 4) Matthew correlation coefficient: Various indicators have been introduced to evaluate the classification, one of them is Matthew correlation coefficient (MMC) index. MMC produces a more informative and truthful score in evaluating binary classification than accuracy and F1 score (50, 51), which is calculated as follows; 푀퐶= 푇푃×푇푁−퐹푃×퐹푁 √ (푇푃+퐹푃)(푇푃+퐹푁)(푇푁+퐹푃)(푇푁+퐹푁) (31) Four tables and figures will be presented in the following, drawn based on each of them (Tables 6 to 9 and Figures 3 to 14). The comparison of accuracy (Table 6, Figures 3, 4 and 5), precision (Tables 7, Figures 6, 7 and 8), recall (Table 8, Figures 9, 10 and11), MMC (Table 9, Figures 12, 13 and 14) according to three different data sets and the UCI repository (52) based on three classification algorithms will be shown. TABLE 6. Compare accuracy Dataset Result of train Original FARNem-F(#) WARA-F(#) CfsSubset Eval-F(#) RSFSAID-F(#) FSbuHD-F(#) wpbc Linear SVM 0.7626 0.7828 0.7626 0.7677 0.7626 0.7778 KNN 0.7121 0.7374 0.6717 0.7071 0.6919 0.7223 Complex Tree 0.6869 0.7020 0.6313 0.7222 0.6869 0.7222 heart Linear SVM 0.8481 0.8370 0.8407 0.8556 0.8370 0.8370 KNN 0.7852 0.7667 0.8074 0.8111 0.8037 0.8111 Complex Tree 0.7556 0.7556 0.7741 0.7963 0.7815 0.7963 australian Linear SVM 0.8536 0.8536 0.8522 0.8551 0.8551 0.8522 KNN 0.8290 0.8290 0.8319 0.8609 0.8304 0.8435 Complex Tree 0.8304 0.8304 0.8348 0.8275 0.8319 0.8435 The values for different algorithms are underlined if greater than the original, and the greatest value is bold. Figure 3. The values of ACC when using SVM classifier Figure 4. The values of ACC when using KNN classifier M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 2669 Figure 5. The values of ACC when using Complex Tree classifier TABLE 7. Compare precision Dataset Result of train Original FARNem-F(#) WARA-F(#) CfsSubset Eval-F(#) RSFSAID-F(#) FSbuHD-F(#) wpbc Linear SVM 0.5000 0.6111 _ 0.5185 _ 0.5882 KNN 0.3750 0.4242 0.2188 0.3429 0.3056 0.3939 Complex Tree 0.3404 0.3636 0.1750 0.4000 0.3256 0.4000 heart Linear SVM 0.8496 0.8393 0.8348 0.8649 0.8167 0.8455 KNN 0.7500 0.7395 0.7881 0.7899 0.7913 0.7899 Complex Tree 0.7213 0.7213 0.7438 0.7982 0.7607 0.7600 australian Linear SVM 0.7845 0.7845 0.7839 0.7867 0.7867 0.7839 KNN 0.8161 0.8161 0.8013 0.8328 0.8105 0.8241 Complex Tree 0.8167 0.8167 0.8185 0.8176 0.8032 0.7867 The values for different algorithms are underlined if greater than the original, and the greatest value is bold. Figure 6. The values of precision when using SVM classifier Figure 7. The values of precision when using KNN classifier 2670 M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 Figure 8. The values of precision when using Complex Treeclassifier TABLE 8. Compare recall Dataset Result of train Original FARNem-F(#) WARA-F(#) CfsSubsetEval-F(#) RSFSAID-F(#) FSbuHD-F(#) wpbc Linear SVM 0.1702 0.2340 0 0.2979 0 0.2128 KNN 0.3191 0.2979 0.1489 0.2553 0.2340 0.2766 Complex Tree 0.3404 0.3404 0.1489 0.3404 0.2979 0.3404 heart Linear SVM 0.8000 0.7833 0.8000 0.8000 0.8167 0.7750 KNN 0.7750 0.7333 0.7750 0.7833 0.7583 0.7833 Complex Tree 0.7333 0.7333 0.7500 0.7250 0.7417 0.7917 australian Linear SVM 0.9251 0.9251 0.9218 0.9251 0.9251 0.9218 KNN 0.7948 0.7948 0.8274 0.8599 0.8078 0.8241 Complex Tree 0.7980 0.7980 0.8078 0.7883 0.8241 0.8893 The values for different algorithms are underlined if greater than the original, and the greatest value is bold. Figure 9. The values of recall when using SVM classifier Figure 10. The values of recall when using KNN classifier M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 2671 Figure 11. The values of recall when using Complex Tree classifier TABLE 9. Compare MMC Dataset Result of train Original FARNem-F(#) WARA-F(#) CfsSubsetEval-F(#) RSFSAID-F(#) FSbuHD-F(#) wpbc Linear SVM 0.1830 0.2778 - 0.2626 - 0.2527 KNN 0.1628 0.1964 −0.0192 0.1149 0.0755 0.1646 Complex Tree 0.1351 0.1586 −0.0738 0.1923 0.1092 0.1923 heart Linear SVM 0.6917 0.6690 0.6766 0.7069 0.6700 0.6691 KNN 0.5667 0.5271 0.6094 0.6172 0.6013 0.6172 Complex Tree 0.5059 0.5059 0.5429 0.5857 0.5565 0.5896 australian Linear SVM 0.7179 0.7179 0.7146 0.7204 0.7204 0.7146 KNN 0.6531 0.6531 0.6610 0.7195 0.6566 0.6831 Complex Tree 0.6561 0.6561 0.6651 0.6499 0.6607 0.6918 The values for different algorithms are underlined if greater than the original, and the greatest value is bold. Figure 12. The values of MMC when using SVM classifier Figure 13. The values of MMC when using KNN classifier 2672 M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 Figure 14. The values of MMC when using Complex Tree classifier Overall, it can be observed that the performance and efficiency of the presented model are superior to those of previous models. This conclusion is based on the results obtained from the tables above and the analysis of the relevant results (underlined numbers) across different datasets. 6. CONCLUSIONS The present study introduces a novel model for feature selection based on fuzzy rough set theory, focusing on its application within a hybrid information system. This model addresses the limitations of traditional feature selection methods, particularly in high-dimensional spaces and hybrid information systems, where conventional techniques may fall short. The newly proposed model is named FSbuHD. In this model, a hybrid distance measure is employed to establish a similarity relation between objects. This similarity relation is utilized in both normal and optimistic modes, enhancing the effectiveness of feature selection. To solve the FSbuHD model, we apply a meta- heuristic algorithm known as the black hole algorithm, which is inspired by natural phenomena. After implementing the algorithm, we present the feature selection results obtained from eight different datasets sourced from the UCI Machine Learning Repository. The acceptability of the FSbuHD model is demonstrated through a comparison of its selected features with those derived from previously established algorithms. To evaluate the performance of the proposed model against other algorithms, three datasets were randomly selected. The performance metrics—accuracy, precision, recall, and Matthews correlation coefficient— were computed using three classification algorithms: SVM, KNN, and Complex Tree. The results indicate that the FSbuHD model outperforms previous algorithms, highlighting its effectiveness in feature selection. For future research, several avenues can be explored to further enhance the FSbuHD model. One potential direction is to solve the model using various other meta- heuristic algorithms, allowing for a comprehensive comparison of their performance in terms of both the quantity of selected features and the quality of the results obtained. Additionally, investigating alternative equivalence relations could represent a significant advancement in the field of feature selection and warrant further study. In conclusion, this research contributes to the growing body of knowledge in feature selection methods by introducing a robust model that effectively leverages fuzzy rough set theory within hybrid information systems. The promising results obtained from the FSbuHD model suggest its potential for wider application and further development in future studies. 7. ACKNOWLEDGEMENT Data for this study come from the UCI repository. 8. REFERENCES 1. Camacho J. Visualizing big data with compressed score plots: approach and research challenges. Chemometrics and Intelligent Laboratory Systems. 2014;135:110-25. https://doi.org/10.1016/j.chemolab.2014.04.011 2. Dobre C, Xhafa F. Intelligent services for big data science. Future generation computer systems. 2014;37:267-81. https://doi.org/10.1016/j.future.2013.07.014 3. Gandomi A, Haider M. Beyond the hype: Big data concepts, methods, and analytics. International journal of information management. 2015;35(2):137-44. https://doi.org/10.1016/j.ijinfomgt.2014.10.007 4. Theodorakopoulos L, Thanasas G, Halkiopoulos C. Implications of big data in accounting: challenges and opportunities. Emerging Science Journal. 2024;8(3):1201-14. http://dx.doi.org/10.28991/ESJ-2024-08-03-024 5. Botao L, Kongruang C. Revealing the Effect of Big Data Capabilities on Efficiency-Based Business Model Innovation. Emerging Science Journal. 2024;8(4):1506-19. http://dx.doi.org/10.28991/ESJ-2024-08-04-016 6. Arena F, Pau G. An overview of big data analysis. Bulletin of Electrical Engineering and Informatics. 2020;9(4):1646-53. https://doi.org/10.11591/eei.v9i4.2359 7. Yao E, Li D, Zhai Y, Zhang C. Multilabel feature selection based on relative discernibility pair matrix. IEEE Transactions on Fuzzy Systems. 2021;30(7):2388-401. 10.1109/TFUZZ.2021.3082171 8. Parmezan ARS, Lee HD, Spolaôr N, Wu FC. Automatic recommendation of feature selection algorithms based on dataset characteristics. Expert Systems with Applications. 2021;185:115589. https://doi.org/10.1016/j.eswa.2021.115589 9. Chandrashekar G, Sahin F. A survey on feature selection methods. Computers & electrical engineering. 2014;40(1):16-28. https://doi.org/10.1016/j.compeleceng.2013.11.024 M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 2673 10. Pawlak Z. Rough sets. International journal of computer & information sciences. 1982;11:341-56. https://doi.org/10.1007/BF01001956 11. Yang F, Zhang P. MSIF: Multi-source information fusion based on information sets. J Intell Fuzzy Syst. 2023;44(3):4103-12. https://doi.org/10.3233/JIFS-222210 12. Zhang X, Mei C, Li J, Yang Y, Qian T. Instance and feature selection using fuzzy rough sets: A bi-selection approach for data reduction. IEEE Transactions on Fuzzy Systems. 2022;31(6):1981-94. https://doi.org/10.1109/TFUZZ.2022.3216990 13. Abbassi H, El Mendili S, Gahi Y. Real-Time Online Banking Fraud Detection Model by Unsupervised Learning Fusion. HighTech and Innovation Journal. 2024;5(1):185-99. http://dx.doi.org/10.28991/HIJ-2024-05-01-014 14. Raman MG, Somu N, Kirthivasan K, Liscano R, Sriram VS. An efficient intrusion detection system based on hypergraph-Genetic algorithm for parameter optimization and feature selection in support vector machine. Knowledge-Based Systems. 2017;134:1- 12. https://doi.org/10.1016/j.knosys.2017.07.005 15. Wang S, Wang Y. Uncertainty measurement for a hybrid information system with images: an application in attribute reduction. IEEE Access. 2020;8:180491-509. https://doi.org/10.1109/ACCESS.2020.3027295 16. Zeng J, Li Z, Zhang P, Wang P. Information structures and uncertainty measures in a hybrid information system: Gaussian kernel method. International Journal of Fuzzy Systems. 2020;22(1):212-31. https://doi.org/10.1007/s40815-019-00779-8 17. Kwak N, Choi C-H. Input feature selection for classification problems. IEEE transactions on neural networks. 2002;13(1):143- 59. https://doi.org/10.1109/72.977291 18. Torkamana A, Badieb K, Salajegheh A, Bokaeic M, Ardestanid SF. A Hybrid Approach to Detect Researchers’ Communities Based on Deep Learning and Game Theory. International Journal of Engineering, Transactions B: Applications. 2023;36(11):2052- 62. https://doi.org/10.5829/ije.2023.36.11b.10 19. Dubois D, Prade H. Rough fuzzy sets and fuzzy rough sets. International Journal of General System. 1990;17(2-3):191-209. https://doi.org/10.1080/03081079008935107 20. Moser B. On the T-transitivity of kernels. Fuzzy Sets and Systems. 2006;157(13):1787-96. https://doi.org/10.1016/j.fss.2006.01.007 21. Sun B, Ma W, Qian Y. Multigranulation fuzzy rough set over two universes and its application to decision making. Knowledge- Based Systems. 2017;123:61-74. https://doi.org/10.1016/j.knosys.2017.01.036 22. Morsi N, Yakout M. Axiomatics for fuzzy rough sets. Fuzzy sets and Systems. 1998;100(1-3):327-42. https://doi.org/10.1016/S0165-0114(97)00104-8 23. Tamaddondar M, Noori N. Hybrid massive mimo channel model based on edge detection of interacting objects and cluster concept. International Journal of Engineering, Transactions B: Applications. 2022;35(2):471-80. https://doi.org/10.5829/ije.2022.35.02b.23 24. Chen D, Yang Y. Attribute reduction for heterogeneous data based on the combination of classical and fuzzy rough set models. IEEE Transactions on Fuzzy Systems. 2013;22(5):1325-34. https://doi.org/10.1109/TFUZZ.2013.2291570 25. Qian Y, Liang J, Wei-zhi ZW, Dang C. Information granularity in fuzzy binary GrC model. IEEE Transactions on Fuzzy Systems. 2010;19(2):253-64. https://doi.org/10.1109/TFUZZ.2010.2095461 26. Chen D, Zhang L, Zhao S, Hu Q, Zhu P. A novel algorithm for finding reducts with fuzzy rough sets. IEEE Transactions on Fuzzy Systems. 2011;20(2):385-9. https://doi.org/10.1109/TFUZZ.2011.2173695 27. Wang C, Shao M, He Q, Qian Y, Qi Y. Feature subset selection based on fuzzy neighborhood rough sets. Knowledge-Based Systems. 2016;111:173-9. https://doi.org/10.1016/j.knosys.2016.08.009 28. Zhang K, Zhan J, Wu W-Z. Novel fuzzy rough set models and corresponding applications to multi-criteria decision-making. Fuzzy Sets and Systems. 2020;383:92-126. https://doi.org/10.1016/j.fss.2019.06.019 29. Luo M, Zhang Y, Fu L. A new similarity measure for picture fuzzy sets and its application to multi-attribute decision making. Informatica. 2021;32(3):543-64. https://doi.org/10.15388/21- INFOR452 30. Kumar A, Prasad PS. Scalable fuzzy rough set reduct computation using fuzzy min–max neural network preprocessing. IEEE Transactions on Fuzzy Systems. 2020;28(5):953-64. https://doi.org/10.1109/TFUZZ.2020.2965899 31. Greco S, Matarazzo B, Slowinski R, editors. Fuzzy similarity relation as a basis for rough approximations. Rough Sets and Current Trends in Computing: First International Conference, RSCTC’98 Warsaw, Poland, June 22–26, 1998 Proceedings 1; 1998: Springer. 32. Wang C, Wang Y, Shao M, Qian Y, Chen D. Fuzzy rough attribute reduction for categorical data. IEEE Transactions on Fuzzy Systems. 2019;28(5):818-30. https://doi.org/10.1109/TFUZZ.2019.2949765 33. El Alaoui M, El Yassini K. Fuzzy similarity relations in decision making. Handbook of research on emerging applications of fuzzy algebraic structures: IGI Global; 2020. p. 369-85. 34. Wang C, Huang Y, Shao M, Fan X. Fuzzy rough set-based attribute reduction using distance measures. Knowledge-Based Systems. 2019;164:205-12. https://doi.org/10.1016/j.knosys.2018.10.038 35. Jayaram B, Mesiar R. I-fuzzy equivalence relations and I-fuzzy partitions. Information Sciences. 2009;179(9):1278-97. https://doi.org/10.1016/j.ins.2008.12.027 36. Bourjandi M, Yadollahzadeh Tabari M, Golsorkhtabaramiri M. Fuzzy centralized coordinate learning and hybrid loss for human activity recognition. International Journal of Engineering, Transactions A: Basics. 2022;35(1):130-41. https://doi.org/10.5829/ije.2022.35.01A.12 37. de Hierro AFRL, Tíscar MÁ, Roldán C, Bustince H. A fuzzy methodology for approaching fuzzy sets of the real line by fuzzy numbers. Fuzzy Sets and Systems. 2022;435:55-77. https://doi.org/10.1016/j.fss.2021.08.024 38. Mashadi, Safitri Y, Sukono, Prihanto IG, Johansyah MD, Saputra MPA. The Inverse and General Inverse of Trapezoidal Fuzzy Numbers with Modified Elementary Row Operations. Mathematics. 2024;12(7):946. https://doi.org/10.3390/math12070946 39. Zeng A, Li T, Liu D, Zhang J, Chen H. A fuzzy rough set approach for incremental feature selection on hybrid information systems. Fuzzy Sets and Systems. 2015;258:39-60. https://doi.org/10.1016/j.fss.2014.08.014 40. Yeung DS, Chen D, Tsang EC, Lee JW, Xizhao W. On the generalization of fuzzy rough sets. IEEE Transactions on fuzzy systems. 2005;13(3):343-61. https://doi.org/10.1109/TFUZZ.2004.841734 41. Liang P, Lei D, Chin K, Hu J. Feature selection based on robust fuzzy rough sets using kernel-based similarity and relative classification uncertainty measures. Knowledge-Based Systems. 2022;255:109795. https://doi.org/10.1016/j.knosys.2022.109795 42. Kodukulla SM, Sireesha V. A Novel Ranking Approach to Rank Interval-Valued Trapezoidal Intuitionistic Fuzzy Sets by Using 2674 M. H. Safarpour et al. / IJE TRANSACTIONS B: Applications Vol. 38, No. 11, (November 2025) 2657-2674 the Concept of Centroids. SN Computer Science. 2024;5(5):493. https://doi.org/10.1007/s42979-024-02861-x 43. Wilson DR, Martinez TR. Improved heterogeneous distance functions. Journal of artificial intelligence research. 1997;6:1-34. https://doi.org/10.1613/jair.346 44. Ganesh AH, Suresh M, Sivakumar G. On solving fuzzy transportation problem based on distance based defuzzification method of various fuzzy quantities using centroid. Malaya J Mat MJM. 2020;1(1):410-26. https://doi.org/10.26637/MJM0S20/0078 45. Hu Q, Zhang L, Chen D, Pedrycz W, Yu D. Gaussian kernel based fuzzy rough sets: model, uncertainty measures and applications. International Journal of Approximate Reasoning. 2010;51(4):453-71. https://doi.org/10.1016/j.ijar.2010.01.004 46. Akbarzadeh Z, Ghadikolaei AS, Madhoushi M, Aghajani H. A hybrid fuzzy multicriteria decision making model based on fuzzy DEMATEL with fuzzy analytical network process and interpretative structural model for prioritizing LARG supply chain practices. International Journal of Engineering, Transactions C: Aspects. 2019;32(3):413-23. https://doi:10.5829/ije.2019.32.03c.09 47. Demiral MF. Perspective Chapter: Experimental Analysis of Black Hole Algorithm with Heuristic Algorithms in Traveling Salesman Problem. Response Surface Methods-Theory, Applications and Optimization Techniques: IntechOpen; 2024. 48. Chen H, Li T, Fan X, Luo C. Feature selection for imbalanced data based on neighborhood rough sets. Information sciences. 2019;483:1-20. https://doi.org/10.1016/j.ins.2019.01.041 49. Tuleuna DT, Ranabhat NAS. Anomalies Detection in Social Media News Using Machine Learning Approach. 2023. https://doi.org/10.48084/etasr.7384 50. Chicco D, Jurman G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC genomics. 2020;21:1-13. https://doi.org/10.1186/s12864-019-6413-7 51. Selvana J, Manavalla S. An Innovative and Reliable Hybrid Cooling Method for Electric Vehicle Motors. International Journal of Engineering, Transactions A: Basics. 2024;37(07):1263. https://doi.org/10.5829/ije.2024.37.07a.06 52. Kelly M, Longjohn R, Nottingham K. The UCI machine learning repository. 2023. COPYRIGHTS ©2025 The author(s). This is an open access article distributed under the terms of the Creative Commons Attribution (C BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, as long as the original authors and source are cited. No permission is required from the authors or the publishers . Persian Abstract چکیده با توجه به حجم بالا، تنوع زیاد و سرعت بالا ی تولید داده ها، برسی روش های انتخاب ویژگی از کلان داده کاربردها و مزایای مختلفی دارد. انتخاب ویژگی با حذف ویژگی های نامرتبط و تکراری منجر به کاهش ابعاد داده می شود که به تصمیم گیری بهینه در ی ک سیستم تصمیم گی ری کمک می کند. یک ی از ابزارهای مهم برای انتخاب ویژگی در سیستم های اطلاعات ترکی بی، نظریه مجموعه های راف فازی است. با این حال، از آنجا که این نظریه از روابط هم ارزی فازی برای انتخاب ویژگی استفاده می کند، دو مشکل مهم دارد. نخست اینکه، به دست آوردن روابط هم ارزی فازی از طریق عملیات های متد اشتراک در ابعاد بالا نیازمند صرف زمان و حافظه زیاد می باشد و ضعف دیگرای ن روش نیز تولید بسیار زیاد داده های نویزی در ابعاد بالاست. هدف و نوآوری این مقاله غلبه بر این مشکلات است. ای ن مقاله یک مدل جدید برای انتخاب ویژگی پیشنهاد می کند. در این مدل فاصله هیبریدی بین اشیا محاسبه و سپس از آن برای به دست آوردن رابطه هم ارزی فازی استفاده می شود. در واقع، به جای حل مسئله انتخاب ویژگی بصورت مستقیم، این روش آن را به یک مسئله بهینه سازی تبدیل می کند که می توان آن را با استفاده از یکی از الگوریتمهای فرا ابتکاری مناسب حل کرد. نویسندگان، این رویکرد جدید را FSbuHD نامگذاری کرده اند. مدل جدید، بر اساس انتخاب یکی از دو رابطه هم ارزی فازی معرفی شده، دارای دو حالت عادی و خوشبینانه خواهد بود . مدل FSbuHD سپس بر روی داده های استاندارد UCI آزمایش و با سای ر الگوری تمها مقایسه شده است. نتایج ای ن تحق یق نشان می دهد که این روش یکی از کارآمدترین و مؤثرترین روشها برای انتخاب ویژگی در مقای سه با روشها و الگوریتمهای قبلی است.