arXiv Analytics

Sign in

arXiv:1912.01098 [cs.LG]AbstractReferencesReviewsResources

Using Dimensionality Reduction to Optimize t-SNE

Rikhav Shah, Sandeep Silwal

Published 2019-12-02Version 1

t-SNE is a popular tool for embedding multi-dimensional datasets into two or three dimensions. However, it has a large computational cost, especially when the input data has many dimensions. Many use t-SNE to embed the output of a neural network, which is generally of much lower dimension than the original data. This limits the use of t-SNE in unsupervised scenarios. We propose using \textit{random} projections to embed high dimensional datasets into relatively few dimensions, and then using t-SNE to obtain a two dimensional embedding. We show that random projections preserve the desirable clustering achieved by t-SNE, while dramatically reducing the runtime of finding the embedding.

Comments: 11th Annual Workshop on Optimization for Machine Learning (OPT2019 )
Categories: cs.LG, stat.ML
Related articles: Most relevant | Search more
arXiv:2310.03398 [cs.LG] (Published 2023-10-05)
Interpolating between Clustering and Dimensionality Reduction with Gromov-Wasserstein
arXiv:2206.13891 [cs.LG] (Published 2022-06-28)
Feature Learning for Dimensionality Reduction toward Maximal Extraction of Hidden Patterns
arXiv:2007.13185 [cs.LG] (Published 2020-07-26)
Dimensionality Reduction for $k$-means Clustering