nearest-neighbor

How to understand Locality Sensitive Hashing?

半世苍凉 提交于 2019-11-26 14:50:24
I noticed that LSH seems a good way to find similar items with high-dimension properties. After reading the paper http://www.slaney.org/malcolm/yahoo/Slaney2008-LSHTutorial.pdf , I'm still confused with those formulas. Does anyone know a blog or article that explains that the easy way? The best tutorial I have seen for LSH is in the book: Mining of Massive Datasets. Check Chapter 3 - Finding Similar Items http://infolab.stanford.edu/~ullman/mmds/ch3a.pdf Also I recommend the below slide: http://www.cs.jhu.edu/%7Evandurme/papers/VanDurmeLallACL10-slides.pdf . The example in the slide helps me a

Nearest neighbors in high-dimensional data?

心已入冬 提交于 2019-11-26 12:34:43
I have asked a question a few days back on how to find the nearest neighbors for a given vector. My vector is now 21 dimensions and before I proceed further, because I am not from the domain of Machine Learning nor Math, I am beginning to ask myself some fundamental questions: Is Euclidean distance a good metric for finding the nearest neighbors in the first place? If not, what are my options? In addition, how does one go about deciding the right threshold for determining the k-neighbors? Is there some analysis that can be done to figure this value out? Previously, I was suggested to use kd

Nearest-neighbor interpolation algorithm in MATLAB

北城以北 提交于 2019-11-26 06:07:30
问题 I am trying to write my own function for scaling up an input image by using the Nearest-neighbor interpolation algorithm. The bad part is I am able to see how it works but cannot find the algorithm itself. I will be grateful for any help. Here\'s what I tried for scaling up the input image by a factor of 2: function output = nearest(input) [x,y]=size(input); output = repmat(uint8(0),x*2,y*2); [newwidth,newheight]=size(output); for i=1:y for j=1:x xloc = round ((j * (newwidth+1)) / (x+1));

How to understand Locality Sensitive Hashing?

你离开我真会死。 提交于 2019-11-26 04:02:05
问题 I noticed that LSH seems a good way to find similar items with high-dimension properties. After reading the paper http://www.slaney.org/malcolm/yahoo/Slaney2008-LSHTutorial.pdf, I\'m still confused with those formulas. Does anyone know a blog or article that explains that the easy way? 回答1: The best tutorial I have seen for LSH is in the book: Mining of Massive Datasets. Check Chapter 3 - Finding Similar Items http://infolab.stanford.edu/~ullman/mmds/ch3a.pdf Also I recommend the below slide:

Nearest neighbors in high-dimensional data?

天涯浪子 提交于 2019-11-26 02:26:53
问题 I have asked a question a few days back on how to find the nearest neighbors for a given vector. My vector is now 21 dimensions and before I proceed further, because I am not from the domain of Machine Learning nor Math, I am beginning to ask myself some fundamental questions: Is Euclidean distance a good metric for finding the nearest neighbors in the first place? If not, what are my options? In addition, how does one go about deciding the right threshold for determining the k-neighbors? Is