Multi-scale similarities in stochastic neighbour embedding: Reducing dimensionality while preserving both local and global structure. J.A., L., D.H., P., & Verleysen M. AO - Lee, J., A., O., h. 2015.
Multi-scale similarities in stochastic neighbour embedding: Reducing dimensionality while preserving both local and global structure [link]Website  abstract   bibtex   
Stochastic neighbour embedding (SNE) and its variants are methods of nonlinear dimensionality reduction that involve soft Gaussian neighbourhoods to measure similarities for all pairs of data. In order to build a suitable embedding, these methods try to reproduce in a low-dimensional space the neighbourhoods that are observed in the high-dimensional data space. Previous works have investigated the immunity of such similarities to norm concentration, as well as enhanced cost functions, like sums of Jensen-Shannon divergences. This paper proposes an additional refinement, namely multi-scale similarities, which are averages of soft Gaussian neighbourhoods with exponentially growing bandwidths. Such multi-scale similarities can replace the regular, single-scale neighbourhoods in SNE-like methods. Their objective is then to maximise the embedding quality on all scales, with the best preservation of both local and global neighbourhoods, and also to exempt the user from having to fix a scale arbitrarily. Experiments with several data sets show that the proposed multi-scale approach captures better the structure of data and improves significantly the quality of dimensionality reduction.
@misc{
 title = {Multi-scale similarities in stochastic neighbour embedding: Reducing dimensionality while preserving both local and global structure},
 type = {misc},
 year = {2015},
 source = {Neurocomputing},
 identifiers = {[object Object]},
 keywords = {*dimensionality reduction,*machine learning,*mathematical parameters,*stochastic neighbour embedding,accuracy,article,data analysis,intermethod comparison,kernel method,mathematical computing,mathematical model,nonlinear system,priority journal,process development,process optimization,quality control},
 websites = {https://www.sciencedirect.com/science/article/abs/pii/S0925231215003641},
 id = {517b6ffc-8514-3ae2-be3c-e634f3fa4026},
 created = {2020-12-29T22:52:04.531Z},
 file_attached = {false},
 profile_id = {aba9653c-d139-3f95-aad8-969c487ed2f3},
 last_modified = {2020-12-30T00:45:51.974Z},
 read = {false},
 starred = {false},
 authored = {true},
 confirmed = {true},
 hidden = {false},
 private_publication = {false},
 abstract = {Stochastic neighbour embedding (SNE) and its variants are methods of nonlinear dimensionality reduction that involve soft Gaussian neighbourhoods to measure similarities for all pairs of data. In order to build a suitable embedding, these methods try to reproduce in a low-dimensional space the neighbourhoods that are observed in the high-dimensional data space. Previous works have investigated the immunity of such similarities to norm concentration, as well as enhanced cost functions, like sums of Jensen-Shannon divergences. This paper proposes an additional refinement, namely multi-scale similarities, which are averages of soft Gaussian neighbourhoods with exponentially growing bandwidths. Such multi-scale similarities can replace the regular, single-scale neighbourhoods in SNE-like methods. Their objective is then to maximise the embedding quality on all scales, with the best preservation of both local and global neighbourhoods, and also to exempt the user from having to fix a scale arbitrarily. Experiments with several data sets show that the proposed multi-scale approach captures better the structure of data and improves significantly the quality of dimensionality reduction.},
 bibtype = {misc},
 author = {J.A., Lee and D.H., Peluffo-Ordonez and Verleysen M. AO - Lee, John A.; ORCID: http://orcid.org/0000-0001-5218-759X}
}

Downloads: 0