Boosting initial population in multiobjective feature selection with knowledge-based partitioning

DENIZ, Ayca and KIZILOZ, Hakan Ezgi (2022). Boosting initial population in multiobjective feature selection with knowledge-based partitioning. In: 2022 International Joint Conference on Neural Networks (IJCNN). IEEE.

[img]
Preview
PDF
Boosting_Initial_Population_in_Multiobjective_Feature_Selection_with_Knowledge_Based_Partitioning.pdf - Accepted Version
All rights reserved.

Download (265kB) | Preview
Official URL: https://ieeexplore.ieee.org/document/9892123
Link to published version:: https://doi.org/10.1109/ijcnn55064.2022.9892123
Related URLs:

    Abstract

    The quality of features is one of the main factors that affect classification performance. Feature selection aims to remove irrelevant and redundant features from data in order to increase classification accuracy. However, identifying these features is not a trivial task due to a large search space. Evolutionary algorithms have been proven to be effective in many optimization problems, including feature selection. These algorithms require an initial population to start their search mechanism, and a poor initial population may cause getting stuck in local optima. Diversifying the initial population is known as an effective approach to overcome this issue; yet, it may not suffice as the search space grows exponentially with increasing feature sizes. In this study, we propose an enhanced initial population strategy to boost the performance of the feature selection task. In our proposed method, we ensure the diversity of the initial population by partitioning the candidate solutions according to their selected number of features. In addition, we adjust the chances of features being selected into a candidate solution regarding their information gain values, which enables wise selection of features among a vast search space. We conduct extensive experiments on many benchmark datasets retrieved from UCI Machine Learning Repository. Moreover, we apply our algorithm on a real-world, large-scale dataset, i.e., Stanford Sentiment Treebank. We observe significant improvements after the comparisons with three off-the-shelf initialization strategies.

    Item Type: Book Section
    Additional Information: Electronic ISSN: 2161-4407
    Identification Number: https://doi.org/10.1109/ijcnn55064.2022.9892123
    SWORD Depositor: Symplectic Elements
    Depositing User: Symplectic Elements
    Date Deposited: 18 Nov 2022 13:58
    Last Modified: 18 Nov 2022 17:32
    URI: https://shura.shu.ac.uk/id/eprint/31056

    Actions (login required)

    View Item View Item

    Downloads

    Downloads per month over past year

    View more statistics