Chinese Physics Letters, 2020, Vol. 37, No. 10, Article code 103201 Active Learning Approach to Optimization of Experimental Control Yadong Wu (吴亚东)1, Zengming Meng (孟增明)2, Kai Wen (文凯)2, Chengdong Mi (米成栋)2, Jing Zhang (张靖)2*, and Hui Zhai (翟荟)1* Affiliations 1Institute for Advanced Study, Tsinghua University, Beijing 100084, China 2State Key Laboratory of Quantum Optics and Quantum Optics Devices, and Institute of Opto-Electronics, Collaborative Innovation Center of Extreme Optics, Shanxi University, Taiyuan 030006, China Received 28 July 2020; accepted 24 August 2020; published online 29 September 2020 Supported by the Beijing Outstanding Young Scientist Program (HZ), the National Key R&D Program of China (Grant Nos. 2016YFA0301600, 2016YFA0301602, and 2018YFA0307600), and the National Natural Science Foundation of China (Grant Nos. 11734010 and 11804203).
*Corresponding authors. Email: jzhang74@sxu.edu.cn; hzhai@tsinghua.edu.cn
Citation Text: Wu Y D, Meng Z M, Wen K, Mi C D and Zhang J et al. 2020 Chin. Phys. Lett. 37 103201    Abstract We present a general machine learning based scheme to optimize experimental control. The method utilizes the neural network to learn the relation between the control parameters and the control goal, with which the optimal control parameters can be obtained. The main challenge of this approach is that the labeled data obtained from experiments are not abundant. The central idea of our scheme is to use the active learning to overcome this difficulty. As a demonstration example, we apply our method to control evaporative cooling experiments in cold atoms. We have first tested our method with simulated data and then applied our method to real experiments. It is demonstrated that our method can successfully reach the best performance within hundreds of experimental runs. Our method does not require knowledge of the experimental system as a prior and is universal for experimental control in different systems. DOI:10.1088/0256-307X/37/10/103201 PACS:32.80.Hd, 37.10.De, 07.05.Fb, 07.05.-t © 2020 Chinese Physics Society Article Text Machine learning can find broad applications in physics research. One of the major advantages of machine learning algorithms is their excellent performance for optimization. On the other hand, optimization is one of the most common tasks in physical experiments, as experimentalists always need to adjust control parameters to reach the best performance. Therefore, the advantage of the machine learning algorithms fits perfectly the demand of experiments in physical sciences. In recent years, there are growing interests in applying various kinds of machine learning based algorithms for optimizing experimental controls.[1–25] Here we focus on a class of experimental control problems as described in the following. First of all, the entire control process can be quantified. All the control parameters, say, denoted by ${\alpha}=\{\alpha_1, \dots, \alpha_n\}$, can be quantitatively determined. Once these control parameters are fixed, the performance $\mathcal{F}$ is also quantitatively determined and measuring $\mathcal{F}$ is repeatable. Secondly, the goal of the experimental control is unique, which is to optimize the performance $\mathcal{F}$. In other words, one needs to reach either the maximum or the minimum of $\mathcal{F}$. Thirdly, if the experimental process is not complicated and is well understood theoretically, the function relation between the control parameters ${\alpha}$ and the control output $\mathcal{F}$ can be calculated theoretically, and the theoretical simulation can be used to guide the experimental optimization. However, for many cases as we consider here, because of various complicated facts in reality, it is difficult to reliably determine the function theoretically, and in some cases, it is hard even to determine its general functional form. Therefore, for each given ${\alpha}$, we can only determine the performance by performing experimental measurements. Finally, the experimental measurements require resources. In many cases, it is possible to carry out hundreds of experimental runs but it is impossible to obtain large datasets including hundreds of thousands of data. Since there exists a deterministic function relation between ${\alpha}$ and $\mathcal{F}$, it is conceivable that we can empirically determine this function by fitting the data. Hence, the neural network (NN) shows its advantage here. Because of the great expressibility of the NN, a sufficiently deep NN can express a general function without assuming its explicit form as a prior. Therefore, a natural idea is to fit the functional mapping between ${\alpha}$ and $\mathcal{F}$ with an NN. Then, with this NN, we can find out the extreme of this function that gives rise to the optimal control parameter. However, there is a main obstacle to this approach. In order to learn a multi-dimensional function, usually one needs a lot of data. The number of data that can be obtained from experiments limits the use of the NN based approach. In other words, it is a major challenge of applying NN based algorithms to optimize experimental control that the total numbers of labeled data are not abundant. However, on the second thought, one realizes that we actually do not need to fit the entire function in order to determine its extreme. In fact, we only need to know the function in the neighborhood of its extreme. Therefore, we only need to measure data in this neighborhood regime and the NN also only needs to fit the data accurately in this regime, with which we are able to determine the optimal parameters. This can certainly significantly reduce the demand of data. However, it comes to a “paradox”. On the one hand, to reach our goal with the minimal number of data, we should try to sample the data points nearby the extreme. On the other hand, since our goal is to find out the extreme, we do not know where the extreme regime locates before we reach our goal. In this work, we develop an active learning approach to solve this paradox. Active learning is a stratagem that guides sampling data iteratively.[26,27] In recent years, active learning has found more and more applications in physical science,[2–4,7,8,14,20–25,28–43] especially in computation problem for material science and chemistry.[35–43] In the following, we will first describe our general active learning scheme for controlling experiments, and then as an example, we apply our method to optimize evaporative cooling experiments in cold atomic gases.
cpl-37-10-103201-fig1.png
Fig. 1. Active learning protocol for optimizing experimental control. Here NN represents a “neural network”. $M$ is a large number such as a few hundred thousands or a few millions, and both $m_0$ and $m_1$ are small numbers such as a few tens. The circle with arrow indicates iteration of steps 1–4 until it converges.
General Scheme. We present our general scheme in Fig. 1. The basic idea is to first train an NN with a small number of randomly sampled labelled data and then use this NN to guess where the possible extreme regime locates. This estimation guides further query of actual data, with which one refines the fitting nearby the extreme. This process continues iteratively until the prediction of the NN and the actual measurements converge in the neighborhood of the extreme. More concretely, the entire process contains the following five steps: (0) Initially, we randomly generate a $m_0$ number of different control parameters ${\alpha}$, and experimentally measure the performance $\mathcal{F}$ for each control. Here $m_0$ is a small number usually taken as a few tens. This generates an initial dataset $\mathcal{L}^0=\{({\alpha}^j, \mathcal{F}^j),j=1,\dots,m_0\}$. Especially, we shall pay attention to a sub-set $\bar{\mathcal{L}}^0\subset \mathcal{L}^0$, which contains the $\bar{m}_0$ ($\bar{m}_0 < m_0$) number of cases with the best performance. (1) We design an NN and train it with the initial dataset $\mathcal{L}^0$. Since the number of data is small, the training results can depend on the initialization. Therefore, we run different initializations and choose one with the smallest loss. Especially, we should pay special attention to the loss on the sub-set $\bar{\mathcal{L}}^0$ when we choose an NN structure and an initialization. (2) We use the trained NN to make predictions on a very large dataset with $M$ different control parameters ${\alpha}$. Here $M$ is a large number, say, a few hundred thousands or a few millions. When sampling these parameters, we simultaneously take two different kinds of probability distributions. Part of the parameters are sampled with a probability obeying the Gaussian distribution around the parameters in $\bar{\mathcal{L}}^0$, and part of the parameters are sampled with a probability obeying the uniform random distribution in the entire parameter space. The latter is important to ensure that we will not miss other possible extremes that are not covered in the initial dataset. The ratio between two different ways of sampling can be adjusted as the process continues. The weight on the former can increase in the later stage when the results gradually converge. (3) We select out the $m_1$ number of control parameters among all $M$ control parameters explored by the NN in step 2, and these $m_1$ numbers of control parameters have the best performance $\mathcal{F}$ predicted by the NN. Here, again $m_1$ is a small number of a few tens. However, we should keep in mind that since the NN is trained with a small dataset $\mathcal{L}^0$, the predication of this NN may not be accurate. (4) We experimentally measure the actual performance of these $m_1$ numbers of control parameters proposed by the NN in step 3, and the measurements return the actual performance $\mathcal{F}$. To quantify whether the NN converges in the neighborhood of the extreme, we compare the actual value with the prediction of NN for a few best ones. If not converged, we will add this $m_1$ number of dataset $\mathcal{L}^1=\{({\alpha}^j, \mathcal{F}^j),j=1,\dots,m_1\}$ into the training set $\mathcal{L}^0$ and train the NN again. It forms a circle from step 1 to step 4, and each of such circles is known as an Epoch. We continue the circle until the prediction of NN and the actual measurements agree with each other. Note that here we do not require the agreement in the entire parameter regime, and only require the agreement in the regime around the extreme. Suppose we can reach convergence in a few tens of epochs and the number of epochs is denoted by $\mathcal{S}$. The total number of data required is $m_0+m_1\times \mathcal{S}$, which is of the order of a few hundreds. In this way, we can reach the optimal experimental control with a few hundreds of runs. Here we remark that in this algorithm, the NN needs to sample a huge number of possible control parameters at step 3 of each epoch, which causes computational resources of the NN. That is to say, this algorithm makes a trade-off between experimental resources and computational resources. Here we should also remark that this algorithm is reminiscent of how human reaches an optimal control. Suppose a new student Bob has no prior knowledge of the experiment that he will work on and unfortunately he also has no help from others. What Bob can do is to first randomly try some controls. With this random tries, he builds up some intuitions in his mind. This intuition is an analogy of the first NN trained by the initial dataset. This intuition tells him which regime is probably more favorable for reaching optimal control and which regime may be unfavorable. Next, he will try more experiments in the favorable regime and fewer experiments in the unfavorable regime. These further tries correct his intuition and make it more and more accurate. When Bob has a good intuition about his experiment, it means that the prediction of his brain agrees very well with the actual measurement, at least in the regime with good performance. Example: Evaporative Cooling. As a demonstration example, we apply our method to optimize evaporative cooling for cold atoms. Evaporative cooling is a universal scheme to reach quantum degeneracy for cold Bose and Fermi atomic gases.[44–55] Cold atoms are confined in a trap created by either magnetic fields or laser fields. The basic idea is to reduce the trap depth as a function of time, such that atoms with the kinetic energy higher than the trap depth will escape from the trap. Therefore, the averaged kinetic energy of the remaining atoms will be reduced such that the temperature will be lowered after the system rethermalizes. Nevertheless, we should notice that the evaporative cooling pays the price of loss atoms in order for decreasing temperature. When the number of atoms is reduced, the characteristic temperature scale for quantum degeneracy also decreases. This characteristic temperature scale is the Bose–Einstein condensation temperature $T_{\rm c}$ for bosons and the Fermi temperature $T_{\rm F}$ for fermions. To characterize quantum degeneracy, what really matters is how small the ratio $T/T_{\rm c}$ or $T/T_{\rm F}$ at which one can achieve with the evaporative cooling. Suppose we start with a fixed initial trap depth $K_0$ and an initial atoms number, what one can control during the evaporative cooling process is how the trap depth $K$ changes as a function of time $t$. We parameterize the $K(t)$ trajectory as follows: $$ K(t)=K_0\Big\{1+\sum\limits_{i=1}^{l}\Big[a_i\Big(\frac{t}{t_{\rm f}}\Big)^i+b_i\Big(\frac{t}{t_{\rm f}}\Big)^{1/(i+1)}\Big]\Big\},~~ \tag {1} $$ with $t\subset[0,t_{\rm f}]$. Here the control parameter ${\alpha}=\{a_{i}, b_{i}, (i=1,\dots,l),t_{\rm f}\}$. The evaporative cooling is therefore a typical optimization problem as described above. We need to find the best control parameter ${\alpha}$, and the control goal is to reach the lowest $T/T_{\rm c}$ or $T/T_{\rm F}$. In the following, we will apply the active learning scheme described above to optimize the evaporative cooling process. We note that there have already been several works using machine learning based methods to optimize evaporative cooling,[22–25] but their methods are different from ours. Some of these works used Bayesian optimization, in which learning and querying were also carried out iteratively. However, in the Bayesian optimization, the sampling is guided by the joint Gaussian probability of the control parameters and control goal after training, while in our method, the sampling is guided by the possible extreme predicted by the neural network. Simulated Data. Since the evaporative cooling process is a complicated non-equilibrium dynamics of an interacting many-body system, it is difficult to accurately simulate this process theoretically. However, there have been efforts taken to understanding the evaporative cooling process, for instance, based on the quantum kinetic equations.[55–57] Here, to test our method, we first use the theoretical simulation of the evaporative cooling of two-component fermions introduced in Ref. [56] and compute the final $T/T_{\rm F}$ for each given evaporation curve of Eq. (1). Hence, we use $T/T_{\rm F}$ as $\mathcal{F}$, and the goal is to find the smallest $T/T_{\rm F}$. We follow the general scheme discussed above, except that the experimental measurements are replaced by numerical simulation. Here we choose $m_0=m_1=10$.
cpl-37-10-103201-fig2.png
Fig. 2. The lowest $T/T_{\rm F}$ as a function of training epoch $\mathcal{S}$. The horizontal axis is the number of epochs. The red point is the lowest $T/T_{\rm F}$ in all $m_1$ control trajectories proposed at each epoch, and the blue point is the average of the lowest five $T/T_{\rm F}$. The error bar indicates the variance of this average. Here we use the simulated results with theory provided by Ref. [56] to determine temperature.
As we show in Fig. 2, we plot $T/T_{\rm F}$ as the lowest temperature, or averaged over the lowest five temperatures, among the actual values of all $m_1$ control trajectories proposed by the NN at each epoch. One can see that the temperature $T/T_{\rm F}$ decreases and the variance among the five lowest temperatures is also suppressed as the training epoch increases. Eventually it achieves the lowest temperature $T/T_{\rm F}\sim 0.03$. In Fig. 3, we compare our best trajectory with the one found in Ref. [56]. Holland et al.[56] proposed a classical method to optimize the trajectory that maximizes the kinetic energy removed per atom from the system at each small time interval. Figure 3 shows that although at the end of the evaporation, both two methods yield similar $T/T_{\rm F}$ and atom number, these two trajectories are quite different, and the evaporation time for the trajectory obtained by our method is only about half of the trajectory obtained in Ref. [56].
cpl-37-10-103201-fig3.png
Fig. 3. The trap depth $K/E_{\rm F}$ (a), the temperature $T/T_{\rm F}$ (b), the total atom number $N$ (c), and the chemical potential $\mu/E_{\rm F}$ (d) as a function of evaporative cooling time. The blue lines are the optimal control found in Ref. [56] and the red lines are the optimal control found by the active learning scheme.
Experimental Data. With the success in the test with simulated data, we now apply our method to real experiment. In this case, we consider evaporative cooling of ${}^{87}$Rb Bose gas, and the experimental apparatus was described in Refs. [58–60]. Though one can measure temperature in cold atomic gases, it is more direct to measure the phase space density. More precisely, since the density of cold atomic gases is measured by absorption imaging of laser after the time of flight, a dimensionless quantity called the optical density (OD) is used for quantifying the phase space density.[61,62] Larger OD means higher phase space density, and the higher the OD, the lower the $T/T_{\rm c}$. Hence, we take the OD as $\mathcal{F}$ and the control goal is to maximize the OD after a sufficiently long time of flight. Here we take $m_0=m_1=20$. The results are shown in Fig. 4. Figure 4(a) shows the highest OD, or the average over the highest five OD, among all $20$ trajectories proposed by the NN at each epoch. Similar to Fig. 2, one can see that the OD increases and the variance is suppressed as the training epoch increases. In Fig. 4(b) we compare the prediction by the NN ${\rm OD}^{\rm NN}$ with the actual measured results ${\rm OD}^{\rm expt}$ for the best five trajectories in each epoch. Here $\delta {\rm OD}$ is defined as $\sum_{i=1}^{5}|{\rm OD}_i^{\rm NN}-{\rm OD}_i^{\rm expt}|/{\rm OD}_i^{\rm expt}$. Figure 4(b) shows $\delta {\rm OD}$ approaches zero as training epoch continues. In Fig. 5, we show the three trajectories with the highest measured OD at different epochs. One can see that initially these trajectories are quite different, and they gradually converge in the later stage. These two figures show that the NN successful finds the extreme around $\sim $15 epochs and the NN converges in the neighborhood of the extreme. The total cost of experimental measures is about $360$.
cpl-37-10-103201-fig4.png
Fig. 4. (a) The dimensionless optical density (OD) of a Bose condensate. (b) The relative difference $\delta {\rm OD}$ between the prediction of the NN and the actual experimental measurements. Both are plotted as a function of the training epoch. The red points in (a) are the best OD for all $m_1$ control trajectories proposed by the NN at each epoch, and the blue points are averaged results of the five highest ones. The points in (b) are also averaged over the five trajectories with the best performance. Error bars indicate the variance for taking the average. Here the actual results are obtained by experimental measurements. The OD is measured after time of flight for $53$ ms.
cpl-37-10-103201-fig5.png
Fig. 5. The three trajectories with the highest optical density at different epochs, as marked by arrows in Fig. 4(b).
In summary, we have presented a general NN based scheme to optimize the experimental control, and the emphasis of this scheme is that we use the active learning to reduce the query of experimental measurements to the minimum. Our scheme is quite universal that can be applied to experiments in different areas, for instance, growing samples in material science and preparing quantum states or quantum gates in quantum science. More importantly, our method does not require any knowledge of the physical system as a prior, and therefore, it can be used for controlling new experiments in the future.
References Quantum adiabatic algorithm design using reinforcement learningMultiparameter optimisation of a magneto-optical trap using deep learningApproaching the adiabatic timescale with machine learningA machine learning approach for online automated optimization of super-resolution optical microscopyReinforcement learning for autonomous preparation of Floquet-engineered states: Inverting the quantum Kapitza oscillatorDeep learning for accelerated all-dielectric metasurface designA Novel Modeling Approach for All-Dielectric Metasurfaces Using Deep Neural NetworksDispersion characterization and pulse prediction with machine learningDeep reinforcement learning for quantum gate controlReinforcement Learning in Different Phases of Quantum ControlMachine learning-based pulse characterization in figure-eight mode-locked lasersExtreme spin squeezing from deep reinforcement learningIntegrating Neural Networks with a Quantum Simulator for State ReconstructionMachine Learning Methods for Control of Fibre Lasers with Double Gain Nonlinear Loop MirrorWhen does reinforcement learning stand out in quantum control? A comparative study on state preparationQuantum Lyapunov control with machine learningPolicy Gradient based Quantum Approximate Optimization AlgorithmA differentiable programming method for quantum controlExperimental neural network enhanced quantum tomographyA Deep Ultraviolet Mode-locked Laser Based on a Neural NetworkToward a self-driving ultrafast fiber laserFast machine-learning online optimization of ultra-cold-atom experimentsApplying machine learning optimization methods to the production of a quantum gasNon-standard trajectories found by machine learning for evaporative cooling of 87 Rb atomsMachine learning for achieving Bose-Einstein condensation of thulium atomsActive LearningRecommender Systems HandbookEfficient Phase Diagram Sampling by Active LearningMachine Learning for Molecular Dynamics on Long TimescalesActive Learning of Spin Network ModelsA quantum active learning algorithm for sampling against adversarial attacksActive emulation of computer codes with Gaussian processes – Application to remote sensingRetrieving Quantum Information with Active LearningActive learning algorithm for computational physicsLess is more: Sampling chemical space with active learningAccelerating high-throughput searches for new alloys with active learning of interatomic potentialsFast and Accurate Uncertainty Estimation in Chemical Machine LearningActive learning of uniformly accurate interatomic potentials for materials simulationMolecular Dynamics with Neural-Network PotentialsMachine Learning Inter-Atomic Potentials Generation Driven by Active Learning: A Case Study for Amorphous and Liquid Hafnium dioxideScale bridging materials physics: Active learning workflows and integrable deep neural networks for free energy function representations in alloysAutomatically growing global reactive neural network potential energy surfaces: A trajectory-free active learning strategyPlasmonic nanostructure design and characterization via Deep LearningEvaporative cooling of magnetically trapped and compressed spin-polarized hydrogenBose-Einstein Condensation of Metastable HeliumDynamics and spatial order of cold cesium atoms in a periodic optical potentialEvidence of Bose-Einstein Condensation in an Atomic Gas with Attractive InteractionsBose-Einstein Condensation of Lithium: Observation of Limited Condensate NumberNobel Lecture: Laser cooling and trapping of neutral atomsAn analytical model for evaporative cooling of atomsBose-Einstein Condensation of Atomic HydrogenNobel Lecture: The manipulation of neutral particlesKinetic theory of the evaporative cooling of a trapped gasEvaporative cooling of a two-component degenerate Fermi gasEvaporative cooling of unitary Fermi gas mixtures in optical trapsBose-Einstein condensate in a light-induced vector gauge potential using 1064-nm optical-dipole-trap lasersEvaporative cooling of 87 Rb atoms into Bose-Einstein condensate in an optical dipole trapGlossary of terms used in photochemistry (IUPAC Recommendations 1996)
[1] Lin J, Lai Z Y and Li X 2020 Phys. Rev. A 101 052327
[2] Tranter A D, Slatyer H J, Hush M R, Leung A C, Everett J L, Paul K V, Vernaz-Gris P, Lam P K, Buchler B C and Campbell G T 2018 Nat. Commun. 9 4360
[3] Henson B M, Shin D K, Thomas K F, Ross J A, Hush M R, Hodgman S S and Truscott A G 2018 Proc. Natl. Acad. Sci. USA 115 13216
[4] Durand A, Wiesner T, Gardner M A, Robitaille E, Bilodeau A, Gagné C, De Koninck P and Lavoie-Cardinal F 2018 Nat. Commun. 9 5247
[5] Bukov M 2018 Phys. Rev. B 98 224305
[6] Nadell C C, Huang B, Malof J M and Padilla W J 2019 Opt. Express 27 27523
[7] An S, Fowler C, Zhang H et al. 2019 arXiv:1906.03387 [physics.optics]
[8] Lohani S, Knutson E M, Zhang W and Glasser R T 2019 OSA Continuum 2 3438
[9] An Z and Zhou D L 2019 Europhys. Lett. 126 60002
[10] Bukov M, Day A G R, Sels D, Weinberg P, Polkovnikov A and Mehta P 2018 Phys. Rev. X 8 031086
[11] Kokhanovskiy A, Bednyakova A, Kuprikov E, Ivanenko A, Dyatlov M, Lotkov D, Kobtsev S and Turitsyn S 2019 Opt. Lett. 44 003410
[12] Chen F, Chen J J, Wu L N, Liu Y C and You L 2019 Phys. Rev. A 100 041801
[13] Torlai G, Timar B, Van Nieuwenburg E P L, Levine H, Omran A, Keesling A, Bernien H, Greiner M, Vuletić V, Lukin M D, Melko R G and Endres M 2019 Phys. Rev. Lett. 123 230504
[14] Kokhanovskiy A, Ivanenko A, Kobtsev S, Smirnov S and Turitsyn S 2019 Sci. Rep. 9 2916
[15] Zhang X M, Wei Z, Asad R, Yang X C and Wang X 2019 npj Quantum Inf. 5 85
[16] Hou S C and Yi X X 2020 Quantum Inf. Process. 19 8
[17] Yao J, Bukov M and Lin L 2020 arXiv:2002.01068 [quant-ph]
[18] Schäfer F, Kloc M, Bruder C and Lörch N 2020 arXiv:2002.08376 [quant-ph]
[19] Palmieri A M, Kovlakov E, Bianchi F, Yudin D, Straupe S, Biamonte J D and Kulik S 2020 npj Quantum Inf. 6 20
[20] Lu H, Xu H, Zhao J and Hou D 2020 Sci. Rep. 10 116
[21] Meng F and Dudley J M 2020 Light: Sci. & Appl. 9 26
[22] Wigley P B, Everitt P J, van den Hengel A, Bastian J W, Sooriyabandara M A, McDonald G D, Hardman K S, Quinlivan C D, Manju P, Kuhn C C N, Petersen I R, Luiten A N, Hope J J, Robins N P and Hush M R 2016 Sci. Rep. 6 25890
[23] Barker A J, Style H, Luksch K, Sunami S, Garrick D, Hill F, Foot C J and Bentine E 2020 Mach. Learn.: Sci. Technol. 1 015007
[24] Nakamura I, Kanemura A, Nakaso T, Yamamoto R and Fukuhara T 2019 Opt. Express 27 20435
[25] Davletov E T, Tsyganok V V, Khlebnikov V A, Pershin D A, Shaykin D V and Akimov A V 2020 Phys. Rev. A 102 011302
[26] Settles B 2012 Synthesis Lectures on Artificial Intelligence and Machine Learning 6 1
[27] Rubens N, Elahi M, Sugiyama M and Kaplan D 2015 Recommender Systems Handbook (Boston: Springer) pp 809–846
[28] Dai C and Glotzer S C 2020 J. Phys. Chem. B 124 1275
[29] Noé F 2018 arXiv:1812.07669 [physics.chem-ph]
[30] Jiang J, Sivak D A and Thomson M 2019 arXiv:1903.10474 [cond-mat.dis-nn]
[31] Casares P A M and Martin-Delgado M A 2020 New J. Phys. 22 073026
[32] Svendsen D H, Martino L and Camps-Valls G 2020 Pattern Recognit. 100 107103
[33] Ding Y, Martín-Guerrero J D, Sanz M, Magdalena-Benedicto R, Chen X and Solano E 2020 Phys. Rev. Lett. 124 140504
[34] Yao J, Wu Y, Koo J, Yan B and Zhai H 2020 Phys. Rev. Res. 2 013287
[35] Smith J S, Nebgen B, Lubbers N, Isayev O and Roitberg A E 2018 J. Chem. Phys. 148 241733
[36] Gubaev K, Podryabinkin E V, Hart G L W and Shapeev A V 2019 Comput. Mater. Sci. 156 148
[37] Musil F, Willatt M J, Langovoy M A and Ceriotti M 2019 J. Chem. Theory Comput. 15 906
[38] Zhang L, Lin D Y, Wang H, Car R and W E 2019 Phys. Rev. Mater. 3 023804
[39] Gastegger M and Marquetand P 2018 arXiv:1812.07676 [physics.chem-ph]
[40] Sivaraman G, Krishnamoorthy A N, Baur M, Holm C, Stan M, Csányi G, Benmore C and Mayagoitia V 2019 arXiv:1910.10254 [cond-mat.mtrl-sci]
[41] Teichert G H, Natarajan A R, Van der Ven A and Garikipati K 2020 arXiv:2002.02305 [cs.LG]
[42] Lin Q, Zhang Y, Zhao B and Jiang B 2020 J. Chem. Phys. 152 154104
[43] Malkiel I, Mrejen M, Nagler A, Arieli U, Wolf L and Suchowski H 2018 Light: Sci. & Appl. 7 60
[44]Ketterle W and Druten N J V 1996 Adv. At. Mol. Opt. Phys. 37 181
[45] Hess H F 1986 Phys. Rev. B 34 3476
[46] Pereira Dos Santos F, Léonard J, Wang J, Barrelet C J, Perales F, Rasel E, Unnikrishnan C S, Leduc M and Cohen-Tannoudji C 2001 Phys. Rev. Lett. 86 3459
[47] Verkerk P, Lounis B, Salomon C, Cohen-Tannoudji C, Courtois J Y and Grynberg G 1992 Phys. Rev. Lett. 68 3861
[48] Bradley C C, Sackett C A, Tollett J J and Hulet R G 1995 Phys. Rev. Lett. 75 1687
[49] Bradley C C, Sackett C A and Hulet R G 1997 Phys. Rev. Lett. 78 985
[50] Phillips W D 1998 Rev. Mod. Phys. 70 721
[51] Davis K B, Mewes M O and Ketterle W 1995 Appl. Phys. B 60 155
[52] Fried D G, Killian T C, Willmann L, Landhuis D, Moss S C, Kleppner D and Greytak T J 1998 Phys. Rev. Lett. 81 3811
[53]Greytak T J 1995 Bose–Einstein Condensation ed Griffin A, Snoke D W and Stringari S (Cambridge: Cambridge University Press) p 131
[54] Chu S, Cohen-Tannoudji C and Phillips W D 1998 Rev. Mod. Phys. 70 685
[55] Luiten O J, Reynolds M W and Walraven J T M 1996 Phys. Rev. A 53 381
[56] Holland M J, DeMarco B and Jin D S 2000 Phys. Rev. A 61 053610
[57] Luo L, Clancy B, Joseph J, Kinast J, Turlapov A and Thomas J E 2006 New J. Phys. 8 213
[58] Fu Z, Wang P, Chai S, Huang L and Zhang J 2011 Phys. Rev. A 84 043609
[59] Xiong D, Wang P, Fu Z, Chai S and Zhang J 2010 Chin. Opt. Lett. 8 627
[60]Chai S, Wang P, Fu Z, Huang L and Zhang J 2012 Acta Sin. Quantum Opt. 18 171
[61]McNaught A D and Wilkinson A 1997 IUPAC Compendium of Chemical Terminology 2nd edn (Oxford: Blackwell Science)
[62] Verhoeven J W 1996 Pure Appl. Chem. 68 2223