Great news, we have a new version of the algorithm 🥳🥳🥳
🎉 New in h-NNE v2: a fast, geometry-aware packing layer for the top (coarse) levels that allocates space proportional to cluster size and preserves local anchor relations. This fixes the “dotifying” look (tiny, collapsed clusters) and the “squeezed big cluster” issue that could appear in v1 when many anchors sat too close.
- Space by cluster mass. Coarse-level anchors are laid out via a lightweight circle-packing step guided by PCA anchors and 1-NN relations to produce acompact, overlap-free arrangement. Larger clusters get more area; small ones remain visible.
- Better global framing. A vectorized overlap resolver plus relaxed/capped k-NN edge targets yields tight layouts without blow-ups, reducing unused whitespace and artifacts.
- Drop-in for h-NNE. v2 runs on the top few FINCH levels, then hands the result to the standard h-NNE refinement. Deeper (fine) levels still use the original fast point-to-anchor updates.
- Note: When running v2 on large datasets (>= 1M points), it starts the tree layout by default after some level of FINCH with a minimum number of clusters (10-10,000) to enhance point spread.
v2 is on by default. To use the legacy pipeline:
HNNE(..., hnne_version="v1")
Defaults generally work well. Optional advanced knobs (e.g.,
start_cluster_view) let you steer the look: choose a deeper level for a more uniform, zoomed-in spread, or a coarser (top) level for a global, zoomed-out view before refinement.
More technical details (algorithms, complexity, limitations) are in our arXiv paper (linked soon).
h-NNE is a hierarchical dimensionality reduction method—akin in spirit to t-SNE/UMAP—but designed for speed, simplicity, and structured views. It first builds a clustering hierarchy with FINCH, then embeds cluster anchors level-by-level and maps points to their anchors with simple, vectorized updates. This preserves local neighborhoods while providing a natural coarse→fine “zoom” over the data.
- Fast & scalable. No global nonconvex optimization; point updates are vectorized and memory-friendly.
- Structure-aware zoom. Choose a parent level for global context, then expand children for detail.
- Labels included. FINCH provides cluster labels out of the box—useful for unlabeled data and structured visualization.
- Parameter-light. FINCH is parameter-free; h-NNE uses robust defaults.
See our corresponding CVPR 2022 paper for the original algorithm and the arXiv addendum for the v2 packing layer.
M. Saquib Sarfraz*, Marios Koulakis*, Constantin Seibold, Rainer Stiefelhagen. Hierarchical Nearest Neighbor Graph Embedding for Efficient Dimensionality Reduction. CVPR 2022.
More details are available in the project documentation.
The project is available in PyPI. To install run:
pip install hnne
The HNNE class implements the common methods of the sklearn interface.
Below a dataset of dimensionality 256 is projected to 2 dimensions.
import numpy as np
from hnne import HNNE
data = np.random.random(size=(1000, 256))
hnne = HNNE(n_components=2)
projection = hnne.fit_transform(data)Once a dataset has been projected, one can apply the transform and project new points to the same dimension.
hnne = HNNE()
projection = hnne.fit_transform(data)
new_data_projection = hnne.transform(new_data)The following demo notebooks are available:
If you make use of this project in your work, it would be appreciated if you cite the hnne paper:
@article{hnne,
title={Hierarchical Nearest Neighbor Graph Embedding for Efficient Dimensionality Reduction},
author={M. Saquib Sarfraz, Marios Koulakis, Constantin Seibold, Rainer Stiefelhagen},
booktitle = {Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
year = {2022}
}If you make use of the clustering properties of the algorithm please also cite:
@inproceedings{finch,
author = {M. Saquib Sarfraz and Vivek Sharma and Rainer Stiefelhagen},
title = {Efficient Parameter-free Clustering Using First Neighbor Relations},
booktitle = {Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
pages = {8934--8943},
year = {2019}
}Contributions are very welcome :-) Please check the contributions guide for more details.