Resumen
Extracting features from high-dimensional data is a critically important task for pattern recognition and machine learning applications. High-dimensional data typically have much more variables than observations, and contain significant noise, missing components, or outliers. Features extracted from high-dimensional data need to be discriminative, sparse, and can capture essential characteristics of the data. In this paper, we present a way to constructing multivariate features and then classify the data into proper classes. The resulting small subset of features is nearly the best in the sense of Greenshtein's persistence; however, the estimated feature weights may be biased. We take a systematic approach for correcting the biases. We use conjugate gradient-based primal-dual interior-point techniques for large-scale problems. We apply our procedure to microarray gene analysis. The effectiveness of our method is confirmed by experimental results.
| Idioma original | English |
|---|---|
| Número de artículo | 4770093 |
| Páginas (desde-hasta) | 636-646 |
| Número de páginas | 11 |
| Publicación | IEEE/ACM Transactions on Computational Biology and Bioinformatics |
| Volumen | 7 |
| N.º | 4 |
| DOI | |
| Estado | Published - 2010 |
Financiación
| Financiadores | Número del financiador |
|---|---|
| National Science Foundation (NSF) | 0845951 |
ODS de las Naciones Unidas
Este resultado contribuye a los siguientes Objetivos de Desarrollo Sostenible
-
Good health and well being
ASJC Scopus subject areas
- Biotechnology
- Genetics
- Applied Mathematics
Huella
Profundice en los temas de investigación de 'A Sparse learning machine for high-dimensional data with application to microarray gene analysis'. En conjunto forman una huella única.Citar esto
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver