• Stars
    star
    106
  • Rank 325,871 (Top 7 %)
  • Language
    Jupyter Notebook
  • License
    MIT License
  • Created over 4 years ago
  • Updated 4 months ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Uniform Manifold Approximation with Two-phase Optimization (IEEE VIS 2022 short)

UMATO

Uniform Manifold Approximation with Two-phase Optimization


Uniform Manifold Approximation with Two-phase Optimization (UMATO) is a dimensionality reduction technique, which can preserve the global as well as the local structure of high-dimensional data. Most existing dimensionality reduction algorithms focus on either of the two aspects, however, such insufficiency can lead to overlooking or misinterpreting important global patterns in the data. Moreover, the existing algorithms suffer from instability. To address these issues, UMATO proposes a two-phase optimization: global optimization and local optimization. First, we obtain the global structure by selecting and optimizing the hub points. Next, we initialize and optimize other points using the nearest neighbor graph. Our experiments with one synthetic and three real world datasets show that UMATO can outperform the baseline algorithms, such as PCA, t-SNE, Isomap, UMAP, LAMP and PacMAP, in terms of accuracy, stability, and scalability.

System Requirements

  • Python 3.9 or greater
  • scikit-learn
  • numpy
  • scipy
  • numba
  • pandas (to read csv files)

Installation

UMATO is available via pip.

pip install umato
import umato
from sklearn.datasets import load_iris

X, y = load_iris(return_X_y=True)
emb = umato.UMATO(hub_num=50).fit_transform(X)

API

Default Class: UMATO

Umato has only one class UMATO. Note that UMATO shares a bunch of parameters with UMAP. For more details, please refer to UMAP API.

class UMATO(BaseEstimator):
    def __init__(
        self,
        n_neighbors=50,
	min_dist=0.1,
        n_components=2,
        hub_num=300,
        metric="euclidean",
        global_n_epochs=None,
        local_n_epochs=None,
        global_learning_rate=0.0065,
        local_learning_rate=0.01,
        spread=1.0,
        low_memory=False,
        set_op_mix_ratio=1.0,
        local_connectivity=1.0,
        gamma=0.1,
        negative_sample_rate=5,
        random_state=None,
        angular_rp_forest=False,
        init="pca",
	verbose=False
    ):

Parameters

n_neighbors = 50

The size of the local neighborhood (defined by the number of nearby sample points) used for manifold approximation. Bigger values lead to a more comprehensive view of the manifold, whereas smaller values retain more local information. Generally, values should fall within the range of 2 to 100. It must be an integer greater than 1. Same effect as it does in UMAP.

min_dist = 0.1

The minimum distance between embedded points. Same effect as it does in UMAP.

n_components = 2

The dimensionality of the output embedding space. It must be a positive integer. This defaults to 2, but can reasonably be set to any integer value below the number of the original dataset dimensions.

hub_num = 300

Number of hub points to use for the embedding. It must be a positive integer or -1 (None).

metric = "euclidean"

The metric to use to compute distances in high dimensional space. If a string is passed it must match a valid predefined metric. If a general metric is required a function that takes two 1d arrays and returns a float can be provided. For performance purposes it is required that this be a numba jitโ€™d function.

The default distance function is a Euclidean. The list of available options can be found in the source code.

global_n_epochs = None

The number of epochs for the global optimization phase. It must be a positive integer of at least 10. If not defiend, it will be set to 100.

local_n_epochs = None 

The number of epochs for the local optimization phase. It must be a positive integer of at least 10. If not defined, it will be set to 50.

global_learning_rate = 0.0065,

The learning rate for the global optimization phase. It must be positive.

local_learning_rate = 0.01

The learning rate for the local optimization phase. It must be positive.

spread = 1.0

Determines the scale at which embedded points will be spread out. Higher values will lead to more separation between points. min_dist must be less than or equal to spread.

low_memory = False

Whether to use a lower memory, but more computationally expensive approach to construct k-nearest neighbor graph

set_op_mix_ratio

Interpolation parameter for combining the global and local structures in the fuzzy simplicial set. It must be between 0.0 and 1.0. A value of 1.0 will use a pure fuzzy union, while 0.0 will use a pure fuzzy intersection.

local_connectivity = 1.0

The local connectivity required โ€“ i.e. the number of nearest neighbors that should be assumed to be connected at a local level. The higher this value the more connected the manifold becomes locally.

gamma = 0.1

The gamma parameter used in local optimization for adjusting the balance between attractive and repulsive forces. It must be non-negative.

negative_sample_rate = 5

The number of negative samples to select per positive sample in the optimization process. Increasing this value will result in greater repulsive force being applied, greater optimization cost, but slightly more accuracy.

random_state = None

The seed of the pseudo random number generator to use when shuffling the data.

angular_rp_forest = False

A boolean flag that indicates whether to use angular random projection forest for approximate nearest neighbor search. It is set to True if the self.metric is in the set {"cosine", "correlation", "dice", "jaccard", "ll_dirichlet", "hellinger"}.

These parameters, along with their conditions and constraints, control various aspects of the embedding process, including the distance metric, optimization settings, and the structure of the resulting low-dimensional space.

Whether to utilize an angular random projection forest for initializing the approximate nearest neighbor search. This approach can offer improved speed, but it is primarily beneficial for metrics employing an angular-style distance, such as cosine, correlation, and others. For these metrics, angular forests will be automatically selected.

init = "pca"

The initialization method to use for the embedding. It must be a string or a numpy array. If a string is passed it must match one of the following: init, random, spectral.

verbose = False

Whether to print information about the optimization process.

Function fit

def fit(self, X):

This fit function embeds the input data X into a lower-dimensional space. It handles optional arguments, validates parameters, checks the data sparsity, and builds the nearest neighbor graph structure. It also computes global and local optimization, initializes the embedding using the original hub information, and embeds outliers.

After validating the input data and setting default values for optional arguments, the function checks if the metric is supported by PyNNDescent and computes the nearest neighbors accordingly. It then builds the k-nearest neighbor graph structure, runs global and local optimization, and embeds the remaining data points. Finally, it embeds outliers and returns the fitted model.

Parameters

X : array, shape (n_samples, n_features) or (n_samples, n_samples)

If the metric is 'precomputed' X must be a square distance matrix. Otherwise it contains a sample per row. If the method is 'exact', X may be a sparse matrix of type 'csr', 'csc' or 'coo'.

Function fit_transform

def fit_transform(self, X):

Fit X into an embedded space and return that transformed output.

Parameters

X : array, shape (n_samples, n_features) or (n_samples, n_samples)

If the metric is โ€˜precomputedโ€™ X must be a square distance matrix. Otherwise it contains a sample per row.

Returns

X_new : array, shape (n_samples, n_components) Embedding of the training data in low-dimensional space.

Citation

UMATO can be cited as follows:

@inproceedings{jeon2022vis,
  title={Uniform Manifold Approximation with Two-phase Optimization},
  author={Jeon, Hyeon and Ko, Hyung-Kwon and Lee, Soohyun and Jo, Jaemin and Seo, Jinwook},
  booktitle={2022 IEEE Visualization and Visual Analytics (VIS)},
  pages={80--84},
  year={2022},
  organization={IEEE}
}

Jeon, H., Ko, H. K., Lee, S., Jo, J., & Seo, J. (2022, October). Uniform Manifold Approximation with Two-phase Optimization. In 2022 IEEE Visualization and Visual Analytics (VIS) (pp. 80-84). IEEE.