An adaptive multi-level sequential floating feature selection
Files
Issued Date
2020
Available Date
Copyright Date
Resource Type
Series
Edition
Language
eng
File Type
application/pdf
No. of Pages/File Size
87 leaves
ISBN
ISSN
eISSN
Other identifier(s)
b212172
Identifier(s)
Access Rights
Access Status
Rights
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Rights Holder(s)
Physical Location
National Institute of Development Administration. Library and Information Center
Bibliographic Citation
Citation
Knitchepon Chotchantarakun (2020). An adaptive multi-level sequential floating feature selection. Retrieved from: https://repository.nida.ac.th/handle/662723737/5536.
Title
An adaptive multi-level sequential floating feature selection
Alternative Title(s)
Author(s)
Editor(s)
Advisor(s)
Advisor's email
Contributor(s)
Contributor(s)
Abstract
Dealing with a large amount of available data becomes a major challenge in data mining and machine learning. Feature selection is a significant preprocessing step for selecting the most informative features by removing irrelevant and redundant features, especially for large datasets. These selected features play an important role in information searching and enhancing the performance of machine learning models such as classification and prediction. There have been several strategies proposed in the past few decades. In this dissertation, we propose a new technique called An Adaptive Multi-level Sequential Floating Feature Selection (AMFFS). AMFFS consists of three proposed algorithms, which are One Level Forward Inclusion (OLFI), One-level Forward Multi-level Backward Selection (OFMB) and Multi-level Forward Inclusion (MLFI). Our proposed methods are considered to be deterministic algorithms related to sequential feature selection under the supervised learning model. The OFMB algorithm consists of two parts. The first part aims to create preliminarily selected subsets. These subsets have similar performance to the Improved Forward Floating Selection (IFFS). This part contains the same procedure as the OLFI algorithm. The second part provides an improvement on the previous result using the multi-level backward searching technique. The idea is to apply an improved step during the feature addition and the adaptive search method on the backtracking step. However, we need to limit the level of backwards-searching to maintain lower execution time by introducing an adaptive variable called the generalization limit. The MLFI algorithm also consists of two parts. The first part aims to search for the maximum classification accuracy by applying the multi-level forward-searching technique. The second part provides an improvement on the previous result by replacing the week feature technique. The idea is to apply an adaptive multi-level forward search method with the replacement step during the feature addition without any backtracking search. Similar to OFMB, we also need to limit the level of forward-searching by the generalization limit. In the experiments, we applied KNN, Naive Bayes, and Decision Tree for our criterion functions. We tested our algorithms on fourteen standard UCI datasets and compared their classification accuracy with other popular methods. Our proposed algorithms showed better results than the other sequential feature selection techniques for the majority of the tested datasets. The OFMB and MLFI algorithms spend more computational time than the other methods due to the complexity of the program.
Table of contents
Description
Thesis (Ph.D. (Computer Science and Information Systems))--National Institute of Development Administration, 2020