Author: Perera, Kushani; Chan, Jeffrey; Karunasekera, Shanika
Title: Group Based Unsupervised Feature Selection Cord-id: gycz4tyd Document date: 2020_4_17
ID: gycz4tyd
Snippet: Unsupervised feature selection is an important task in machine learning applications, yet challenging due to the unavailability of class labels. Although a few unsupervised methods take advantage of external sources of correlations within feature groups in feature selection, they are limited to genomic data, and suffer poor accuracy because they ignore input data or encourage features from the same group. We propose a framework which facilitates unsupervised filter feature selection methods to e
Document: Unsupervised feature selection is an important task in machine learning applications, yet challenging due to the unavailability of class labels. Although a few unsupervised methods take advantage of external sources of correlations within feature groups in feature selection, they are limited to genomic data, and suffer poor accuracy because they ignore input data or encourage features from the same group. We propose a framework which facilitates unsupervised filter feature selection methods to exploit input data and feature group information simultaneously, encouraging features from different groups. We use this framework to incorporate feature group information into Laplace Score algorithm. Our method achieves high accuracy compared to other popular unsupervised feature selection methods ([Formula: see text]30% maximum improvement of Normalized Mutual Information (NMI)) with low computational costs ([Formula: see text]50 times lower than embedded methods on average). It has many real world applications, particularly the ones that use image, text and genomic data, whose features demonstrate strong group structures.
Search related documents:
Co phrase search for related documents- accuracy gain and machine learning task: 1
- adjacency matrix and machine learning: 1
- low accuracy and machine learning: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11
- low computational cost and machine learning: 1, 2, 3, 4, 5
- low distance and machine learning: 1
- low similarity and machine learning: 1, 2
Co phrase search for related documents, hyperlinks ordered by date