An adaptive multi-level sequential floating feature selection

dc.contributor.advisorOhm Sornilth
dc.contributor.authorKnitchepon Chotchantarakunth
dc.descriptionThesis (Ph.D. (Computer Science and Information Systems))--National Institute of Development Administration, 2020th
dc.description.abstractDealing with a large amount of available data becomes a major challenge in data mining and machine learning. Feature selection is a significant preprocessing step for selecting the most informative features by removing irrelevant and redundant features, especially for large datasets. These selected features play an important role in information searching and enhancing the performance of machine learning models such as classification and prediction. There have been several strategies proposed in the past few decades. In this dissertation, we propose a new technique called An Adaptive Multi-level Sequential Floating Feature Selection (AMFFS). AMFFS consists of three proposed algorithms, which are One Level Forward Inclusion (OLFI), One-level Forward Multi-level Backward Selection (OFMB) and Multi-level Forward Inclusion (MLFI). Our proposed methods are considered to be deterministic algorithms related to sequential feature selection under the supervised learning model. The OFMB algorithm consists of two parts. The first part aims to create preliminarily selected subsets. These subsets have similar performance to the Improved Forward Floating Selection (IFFS). This part contains the same procedure as the OLFI algorithm. The second part provides an improvement on the previous result using the multi-level backward searching technique. The idea is to apply an improved step during the feature addition and the adaptive search method on the backtracking step. However, we need to limit the level of backwards-searching to maintain lower execution time by introducing an adaptive variable called the generalization limit. The MLFI algorithm also consists of two parts. The first part aims to search for the maximum classification accuracy by applying the multi-level forward-searching technique. The second part provides an improvement on the previous result by replacing the week feature technique. The idea is to apply an adaptive multi-level forward search method with the replacement step during the feature addition without any backtracking search. Similar to OFMB, we also need to limit the level of forward-searching by the generalization limit. In the experiments, we applied KNN, Naive Bayes, and Decision Tree for our criterion functions. We tested our algorithms on fourteen standard UCI datasets and compared their classification accuracy with other popular methods. Our proposed algorithms showed better results than the other sequential feature selection techniques for the majority of the tested datasets. The OFMB and MLFI algorithms spend more computational time than the other methods due to the complexity of the
dc.format.extent87 leavesth
dc.publisherNational Institute of Development Administrationth
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International
dc.subjectClassification accuracyth
dc.subjectFeature selectionth
dc.subjectSequential searchth
dc.subject.otherDimension reduction (Statistics)th
dc.subject.otherDimensional analysisth
dc.subject.otherSequential analysisth
dc.subject.otherData miningth
dc.subject.otherSupervised learning (Machine learning)th
dc.titleAn adaptive multi-level sequential floating feature selectionth
dc.typetext--thesis--doctoral thesisth
mods.physicalLocationNational Institute of Development Administration. Library and Information Centerth of Applied Statisticsth Science and Information Systemsth Institute of Development Administrationth of Philosophyth
Original bundle
Now showing 1 - 1 of 1
Thumbnail Image
2.25 MB
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
Thumbnail Image
202 B
Plain Text