Cited by
- BibTex
- RIS
- TXT
A fundamental task in phase retrieval is to recover an unknown signal $x\in\mathbb{R}^n$ from a set of magnitude-only measurements $y_i=|\langle a_i,x\rangle|,$ $ i=1,\cdots,m$. In this paper, we propose two novel perturbed amplitude models (PAMs) which have a non-convex and quadratic-type loss function. When the measurements $ a_i \in \mathbb{R}^n$ are Gaussian random vectors and the number of measurements $m\ge Cn$, we rigorously prove that the PAMs admit no spurious local minimizers with high probability, i.e., the target solution $ x$ is the unique local minimizer (up to a global phase) and the loss function has a negative directional curvature around each saddle point. Thanks to the well-tamed benign geometric landscape, one can employ the vanilla gradient descent method to locate the global minimizer $x$ (up to a global phase) without spectral initialization. We carry out extensive numerical experiments to show that the gradient descent algorithm with random initialization outperforms state-of-the-art algorithms with spectral initialization in empirical success rate and convergence speed.
}, issn = {}, doi = {https://doi.org/10.4208/aam.OA-2021-0009}, url = {http://global-sci.org/intro/article_detail/aam/20092.html} }A fundamental task in phase retrieval is to recover an unknown signal $x\in\mathbb{R}^n$ from a set of magnitude-only measurements $y_i=|\langle a_i,x\rangle|,$ $ i=1,\cdots,m$. In this paper, we propose two novel perturbed amplitude models (PAMs) which have a non-convex and quadratic-type loss function. When the measurements $ a_i \in \mathbb{R}^n$ are Gaussian random vectors and the number of measurements $m\ge Cn$, we rigorously prove that the PAMs admit no spurious local minimizers with high probability, i.e., the target solution $ x$ is the unique local minimizer (up to a global phase) and the loss function has a negative directional curvature around each saddle point. Thanks to the well-tamed benign geometric landscape, one can employ the vanilla gradient descent method to locate the global minimizer $x$ (up to a global phase) without spectral initialization. We carry out extensive numerical experiments to show that the gradient descent algorithm with random initialization outperforms state-of-the-art algorithms with spectral initialization in empirical success rate and convergence speed.