I have been reading this paper for a few days. There is one section (Section 3.3) that confuses me.
We start by gathering local features from training images of a particular class into a single set. Then for every local feature, a Gaussian kernel is placed in the feature space with its mean at the feature. The probability density function (PDF) of the class is then defined as the normalized sum of all the kernels.
To simplify the discussion, we can assume we have $N$ features with each feature has dimensionality 128. Now, I am lost when the author said
a Gaussian kernel is placed in the feature space with its mean at the feature
It seems that this means using kernel method but it also seems the author wants to use kernel density estimator, and what exactly does with its mean at the feature means?
Any suggestions to clarify these confusions are welcome!
with its mean at the feature. As far as I know, the only parameter for KDE is the bandwidth, please see here Section 2.2. I am not sure what role ismeanplaying here. – yiping Sep 17 '19 at 16:39We set the covariance of the kernels using Parzen Windows(In the next paragraph of the paper). I believe Parazen Windows is KDE (from Wikipedia), so I am not sure what is he referring to. – yiping Sep 17 '19 at 17:05