Visible to the public K-nearest Neighbor Search by Random Projection Forests

TitleK-nearest Neighbor Search by Random Projection Forests
Publication TypeConference Paper
Year of Publication2018
AuthorsYan, Donghui, Wang, Yingjie, Wang, Jin, Wang, Honggang, Li, Zhenpeng
Conference Name2018 IEEE International Conference on Big Data (Big Data)
Date Publisheddec
KeywordsBig Data, clustered computers, Clustering algorithms, computational complexity, data mining, Ensemble, ensemble methods, ensemble random projection trees, ensemble size, exponential decay, Forestry, k-nearest neighbor search, k-nearest neighbors, kNN distances, kNN search, learning (artificial intelligence), machine learning, machine learning algorithms, Measurement, Metrics, multicore computers, nearest neighbor search, nearest neighbour methods, pubcrawl, random forests, random projection forests, rpForests, search problems, tree-based methodology, unsupervised learning, Vegetation
AbstractK-nearest neighbor (kNN) search has wide applications in many areas, including data mining, machine learning, statistics and many applied domains. Inspired by the success of ensemble methods and the flexibility of tree-based methodology, we propose random projection forests, rpForests, for kNN search. rpForests finds kNNs by aggregating results from an ensemble of random projection trees with each constructed recursively through a series of carefully chosen random projections. rpForests achieves a remarkable accuracy in terms of fast decay in the missing rate of kNNs and that of discrepancy in the kNN distances. rpForests has a very low computational complexity. The ensemble nature of rpForests makes it easily run in parallel on multicore or clustered computers; the running time is expected to be nearly inversely proportional to the number of cores or machines. We give theoretical insights by showing the exponential decay of the probability that neighboring points would be separated by ensemble random projection trees when the ensemble size increases. Our theory can be used to refine the choice of random projections in the growth of trees, and experiments show that the effect is remarkable.
DOI10.1109/BigData.2018.8622307
Citation Keyyan_k-nearest_2018