Abstract: We study the problem of learning Gaussian Mixture Model (GMM) in one dimension. Given samples access to a mixture f of k Gaussians and an accurate parameter $$\epsilon >0$$ , our algorithm takes $$\tilde{O}{(\frac{k}{\epsilon ^5})}$$ samples, runs in polynomial time and outputs a mixture g of at most $$\tilde{O}(\min \{\frac{k^2}{\epsilon ^2},\frac{k}{\epsilon ^3}\})$$ Gaussians such that the total variation distance between f and g is at most $$\epsilon $$ . This improves the previous result by [4], which uses $$O(\frac{k^2}{\epsilon ^6})$$ samples and outputs a mixture of $$O(\frac{k}{\epsilon ^3})$$ Gaussians. Our algorithm uses LP rounding technique to find the sparse solution of a linear programming. The main technical contribution of us is a non-trivial inequality for Gaussians, which may be interesting in its own right. We also consider the problem of properly learning mixture of two Gaussians. We show how to reduce the learning task to the closest pair problem in $$L_{\infty }$$ -norm. Our algorithm takes $$\tilde{O}(\frac{1}{\epsilon ^2})$$ samples and runs in time $$\tilde{O}(\frac{1}{\epsilon ^{4.001}})$$ . Our result improves the previous result by [7], which uses $$\tilde{O}(\frac{1}{\epsilon ^2})$$ samples and runs in time $$\tilde{O}(\frac{1}{\epsilon ^5})$$ .
0 Replies
Loading