Talk:Variable kernel density estimation

Latest comment: 3 years ago by 2001:16B8:A07F:C600:D80B:18C9:FA99:B492 in topic Kernel shape variation

There seems to be a factor 1/h missing in the first section called Rationale. Compare here: http://en.wikipedia.org/wiki/Kernel_density_estimation#Definition

Good call. In my own paper, the h is a sigma and it is absorbed into K while the K is not normalized and there is a separate normalization coefficient. However, I decided in this article to pull out the bandwidth, change the symbol and normalize K to bring my notation more in line with others, not realizing that you also have to pull h out of the normalization coefficient. Thanks. Peteymills (talk) 23:16, 13 May 2013 (UTC)Reply

nested kernel estimators : edit

nested kernel estimators = multilayer perceptron (or feedforward neural networks)

the "multi-variate kernel density estimation" is the special case where the kernel's parameter depends on space directions, to be general in fact the kernel could depend on the point of space, where it's parameters are given by another "kernel density estimator",

this way each layer (each kernel density estimator) is the layer of a (feed-forward) neural network,

and we also include hidden markov models (for discrete random variables)

78.227.78.135 (talk) 01:10, 30 October 2015 (UTC)Reply

Kernel shape variation edit

"For multivariate estimators, the parameter, h, can be generalized to vary not just the size, but also the shape of the kernel. This more complicated approach will not be covered here."

It would be helpful if the article either covered the part about shape variation, or provided a reference on where to find more information about that.

2001:16B8:A07F:C600:D80B:18C9:FA99:B492 (talk) 07:39, 13 May 2020 (UTC)Reply