• English
  • Deutsch
  • Log In
    or
  • Research Outputs
  • Projects
  • Researchers
  • Institutes
  • Statistics
Repository logo
Fraunhofer-Gesellschaft
  1. Home
  2. Fraunhofer-Gesellschaft
  3. Artikel
  4. A sparse grid based method for generative dimensionality reduction of high-dimensional data
 
  • Details
  • Full
Options
2016
  • Zeitschriftenaufsatz

Titel

A sparse grid based method for generative dimensionality reduction of high-dimensional data

Abstract
Generative dimensionality reduction methods play an important role in machine learning applications because they construct an explicit mapping from a low-dimensional space to the high-dimensional data space. We discuss a general framework to describe generative dimensionality reduction methods, where the main focus lies on a regularized principal manifold learning variant. Since most generative dimensionality reduction algorithms exploit the representer theorem for reproducing kernel Hilbert spaces, their computational costs grow at least quadratically in the number n of data. Instead, we introduce a grid-based discretization approach which automatically scales just linearly in n. To circumvent the curse of dimensionality of full tensor product grids, we use the concept of sparse grids. Furthermore, in real-world applications, some embedding directions are usually more important than others and it is reasonable to refine the underlying discretization space only in these directions. To this end, we employ a dimension-adaptive algorithm which is based on the ANOVA (analysis of variance) decomposition of a function. In particular, the reconstruction error is used to measure the quality of an embedding. As an application, the study of large simulation data from an engineering application in the automotive industry (car crash simulation) is performed.
Author(s)
Bohn, Bastian
Garcke, Jochen
Griebel, Michael
Zeitschrift
Journal of computational physics
Thumbnail Image
DOI
10.1016/j.jcp.2015.12.033
Language
Englisch
google-scholar
SCAI
  • Cookie settings
  • Imprint
  • Privacy policy
  • Api
  • Send Feedback
© 2022