Results 1 to 10 of about 703,161 (279)
Threshold Adaptation for Improved Wrapper-Based Evolutionary Feature Selection [PDF]
Feature selection is essential for enhancing classification accuracy, reducing overfitting, and improving interpretability in high-dimensional datasets.
Uroš Mlakar, Iztok Fister, Iztok Fister
doaj +2 more sources
Agnostic Feature Selection [PDF]
Unsupervised feature selection is mostly assessed along a supervised learning setting, depending on whether the selected features efficiently permit to predict the (unknown) target variable. Another setting is proposed in this paper: the selected features aim to efficiently recover the whole dataset.
Doquet, Guillaume Florent +1 more
openaire +2 more sources
Online Feature Selection with Streaming Features [PDF]
We propose a new online feature selection framework for applications with streaming features where the knowledge of the full feature space is unknown in advance. We define streaming features as features that flow in one by one over time whereas the number of training examples remains fixed.
Xindong, Wu +4 more
openaire +3 more sources
Hybrid-Recursive Feature Elimination for Efficient Feature Selection
As datasets continue to increase in size, it is important to select the optimal feature subset from the original dataset to obtain the best performance in machine learning tasks.
Hyelynn Jeon, Sejong Oh
doaj +1 more source
In this paper, we introduce a novel unsupervised, graph-based filter feature selection technique which exploits the power of topologically constrained network representations. We model dependency structures among features using a family of chordal graphs (the Triangulated Maximally Filtered Graph), and we maximise the likelihood of features' relevance ...
Briola, Antonio, Aste, Tomaso
openaire +3 more sources
Feature selection, as a data preprocessing strategy, has been proven to be effective and efficient in preparing data (especially high-dimensional data) for various data-mining and machine-learning problems. The objectives of feature selection include building simpler and more comprehensible models, improving data-mining performance, and preparing clean,
Jundong Li +6 more
openaire +2 more sources
Biogeography-based optimization for feature selection
Data clustering has many applications in medical sciences, banking, and data mining. K-means is the most popular data clustering algorithm due to its efficiency and simplicity of implementation. However, K-means has some limitations, which may affect its
Mandana Gholami +2 more
doaj +1 more source
Ontology-Based Feature Selection: A Survey
The Semantic Web emerged as an extension to the traditional Web, adding meaning (semantics) to a distributed Web of structured and linked information.
Konstantinos Sikelis +2 more
doaj +1 more source
Feature Selection Embedded Robust K-Means
Clustering is one of the most important unsupervised learning problems in machine learning. As one of the most widely used clustering algorithms, K-means has been studied extensively.
Qian Zhang, Chong Peng
doaj +1 more source
Redundancy Is Not Necessarily Detrimental in Classification Problems
In feature selection, redundancy is one of the major concerns since the removal of redundancy in data is connected with dimensionality reduction. Despite the evidence of such a connection, few works present theoretical studies regarding redundancy.
Sebastián Alberto Grillo +9 more
doaj +1 more source

