An asymptotically optimal, machine learning based, adaptive single-photon phase estimation protocol that allows us to reach the standard quantum limit when a very limited number of photons is employed.
Phase estimation represents a significant example to test the application of quantum theory for enhanced measurements of unknown physical parameters. Several recipes have been developed, allowing to define strategies to reach the ultimate bounds in the asymptotic limit of a large number of trials. However, in certain applications it is crucial to reach such bound when only a small number of probes is employed. Here, we discuss an asymptotically optimal, machine learning based, adaptive single-photon phase estimation protocol that allows us to reach the standard quantum limit when a very limited number of photons is employed.