Principal Component Analysis of Diffuse Magnetic Scattering: a Theoretical Study

We present a theoretical study of the potential of Principal Component Analysis to analyse magnetic diffuse neutron scattering data on quantum materials. To address this question, we simulate the scattering function $S\left(\mathbf{q}\right)$ for a model describing a cluster magnet with anisotropic spin-spin interactions under different conditions of applied field and temperature. We find high dimensionality reduction and that the algorithm can be trained with surprisingly small numbers of simulated observations. Subsequently, observations can be projected onto the reduced-dimensionality space defined by the learnt principal components. Constant-field temperature scans corresponds to trajectories in this space which show characteristic bifurcations at the critical fields corresponding to ground-state phase boundaries. Such plots allow the ground-state phase diagram to be accurately determined from finite-temperature measurements.


I. INTRODUCTION
The study of quantum matter has emerged in recent decades as a major field of scientific endeavour.The behaviour of many-body systems is quite well understood at relatively high temperatures where it is dominated by classical forces and entropy and where it can be simulated efficiently using classical computers.However quantum effects including entanglement and particle indistinguishability make the equivalent, low-temperature problem much harder in principle.Even so, a good understanding has emerged for various ordered ground states, including the Landau Fermi liquid and states showing magnetic, superconducting or topological order [1,2].Strongly-correlated quantum matter [3], on the other hand, shows quantum correlations persisting at intermediate energy scales and is less well understood with many outstanding questions.For example, the precise relationship between the intermediate-temperature, "liquid" states and the various ground states in its phase diagrams remains unknown [4,5].
In recent years the arsenal available to tackle such challenging problems has been enlarged by by the application of Machine Learning (ML).For instance, artificial neural networks have been used to efficiently encode the wave function of a many-body Hamiltonian, searching for the ground state by reinforcement learning [6]; to predict the properties of one material from those of other substances, without involving a model Hamiltonian [7]; and to detect phase transitions from piezoelectric relaxation measurements [8] and spectroscopic imaging scanning tunnelling microscopy [9].
Here we propose an application of ML to magnetic neutron scattering (NS).The neutron's intrinsic magnetic moment and availability of neutron beams with wavelengths of the order of an angstrom make NS one of the most powerful probes of magnetism in materials [10].NS has provided, for instance: a thorough characterisation of the magnetic excitation spectrum of the cuprates [11][12][13][14][15]; strong evidence of magnetic monopole excitations in 'spin ice' frustrated magnets [16,17]; and a quantitative understanding of quantum phase transitions in magnetic insulators [18][19][20][21].
Our approach to magnetic NS is based on Principal Component Analysis (PCA), a well-established technique for dimensionality reduction [22,23] that can be regarded as a form of unsupervised machine learning.Formally, PCA is equivalent to a linear autoencoder [24] and is a suitable initial step for a wide range of classification problems.In recent years PCA and auto-encoders have been applied to data obtained through numerical simulation of many-body systems [25,26].It has been shown that, when provided with detailed information on large representative samples of microstates of such systems, such algorithms are capable of "discovering" important features in their phase diagrams, including order parameters and phase transitions [25].On the other hand experimentallyaccessible information is normally limited and does not provide access to individual microstates, consisting instead of thermal averages.The question emerges: can PCA still identify important features from such averages?
Recently an autoencoder-based approach to magnetic diffuse neutron scattering data on the "spin-ice" material Dy 2 Ti 2 O 7 has been demonstrated [26].The autoencoder is trained on a set of simulated neutron-scattering images.The simulations correspond to a class of candidate model Hamiltonians.The trained autoencoder is then used to describe real experimental data.It was found that the autoencoder provides a compact description of the experimental data, facilitating the identification of optimal model Hamiltonians, and that it can also recognise distinct physical regimes in the simulations.The latter suggests the question highlighted above can be answered in the affirmative.
Here we ask whether PCA can be used to infer relevant features from the data even in the absence of prior knowledge of a class of applicable Hamiltonians (or in cases where the Hamiltonians might not be tractable).This would require training the algorithm directly on the experimental data.Given the scarcity of neutron flux, the training set would need to consist of a limited number of scattering images, each with limited resolution.The trained algorithm would then be used to filter additional (but similarly limited) data sets and it would have to produce qualitative signatures of any relevant features.Of particular interest is the ability to infer ground-state phase boundaries from finite-temperature data.
We perform a theoretical study to address the above questions, focusing on a class of models describing spin-1/2, anti-ferromagnetic, ring-shaped molecular magnets.The field-dependent phase diagrams of all instances of our model feature one or more level crossings (LC) where the nature of the ground state changes discontinuously.One of these LCs is a so-called "entanglement transition" (ET).At the ET the ground state factorises exactly.These changes in the system's ground state have clear signatures in simulated low-temperature, high-resolution diffuse magnetic neutron scattering cross-sections [27].The main question we tackle here is whether a PCA can detect these features using a more limited number of lower-resolution images.We will see that this is indeed possible.Moreover the ground-state phase boundaries can be accurately determined from finite-temperature data.In order to achieve all this we introduce the notion of a "score bifurcation plot" for principal components (PCs).This tool is only a small variation of the usual "score plots" in general use with PCA algorithms but it shows in a particularly transparent way how the different quantum ground states in the model emerge from the high-temperature phase as the temperature is lowered.We argue that this can be a useful tool for the identification of quantum ground states from experimental data.

II. MODEL
For the purpose of our study we consider a spin-1/2, anisotropic Heisenberg ring in an applied magnetic field perpendicular to the plane of the ring (for an illustration, see Fig. 1).Assuming nearest-neighbour interactions only, the system has the Hamiltonian Here N is the number of magnetic ions in the ring, which we assume to be even, J and h are, respectively, the exchange and field energies, and γ and ∆ are two dimensionless parameters describing the anisotropy of the spinspin interaction.The operator Ŝα j represents the α th component of the spin at the j th magnetic site and the labels x, y, z refer to the local mangetic axes at that site.The x and y axes rotate from site to site so as to preserve the C N rotational symmetry of the molecule around the z axis, which is fixed.Note that we have assumed that the interaction is diagonal in this basis.An illustration of the geometry of the model can be found in [Ref.27,Fig. 1].We take all three components of the interaction to be anti-ferromagnetic, corresponding to J < 0, and assume without loss of generality 0 ≤ γ ≤ 1 and ∆ > 0 1 .The boundary conditions are enforced by setting N + 1 ≡ 1.
For fixed J, γ, ∆ it has N/2 ground state degeneracies at h = h 1 , h 2 , . . ., h N/2 = h f , where the last degeneracy takes place at the N −independent factorisation field h f = J (1 + ∆) 2 − γ 2 .The simulated diffuse magnetic neutron scattering function S (q) for the geometry under consideration and with the scattering vector q within the plane of the molecule [27] shows qualitative changes from anti-ferromagnetic correlations for h < h f to ferromagnetic ones for h > h f .This is consistent with an entan- 1 Ignoring the spatial arrangement of the atoms, the Hamiltonian in Eq. ( 1) can correspond to a number of distinct universality classes: Heisenberg for ∆ = γ = 0; XY for γ = 0 = ∆; and Ising for ∆ = 0 = γ.
glement transition from anti-parallel Bell states to parallel Bell states, respectively, known to take place at h f Amico et al. [33].Less striking, but well-defined changes also occur at the other level crossings.Specifically, numerical evidence for a jump of S (0) in the ground state taking place at all N/2 level crossings has been obtained for N = 4, 6 [27], 8 and 10 [34].At finite temperatures the jumps become crossovers which get smoother as the temperature is raised further.
The codes we used for this study are freely available as open source from Refs.[35] (neutron scattering simulations) and [36] (PCA).They require only the Octave computer language [37].Further details are given in the appendix.

III. DIMENSIONALITY REDUCTION
For any given set of values of the parameters of our model, the scattering function S (q) mentioned above can be interpreted as an image2 .Giving the parameters different values allows us to generate different images which can be subject to PCA.Quite generally, the result of a PCA of any set of images is a complete, orthogonal basis set that can be used to reconstruct exactly, through linear superposition, any of the images in the original training set.The advantage of this new basis is that the PCs are ordered, with the first basis element capturing the largest amount of variance within the original data set, the second capturing the second largest amount, and so forth.For images comprising solely random pixel values this would offer no advantage but if the images are strongly-correlated then a very good approximation to all the images in the training set can be obtained using just the first few PCs.PCA can thus be regarded as a technique for dimensionality-reduction and this forms the basis of its application to problems such as face recognition [23].Its effectiveness relies on correlations within the training set: if correlation is high (e.g.all images represent human faces) then a small number M of PCs can capture most of the variance in the data set.The details of our PCA procedure are given in the appendix.
In our problem we expect to achieve significant reduction because all images have been derived from instances of the same class of Hamiltonians.Let us fix the number of magnetic moments in the molecule N and vary the parameters γ, ∆, h and temperature T (all four energies Figure 2: Fraction of the variance in a set of simulated diffuse magnetic neutron scattering images that is captured by the first 10 PCs.Each curve was obtained by PCA of 500 random observations.Each observation is a 24 × 24 pixel image obtained by computing the scattering function S (q) of our cluster-magnet model.A uniform mesh of q-vectors with components q x , q y ranging from −9π/a to 9π/a, where a is the distance between nearest neighbours within the cluster, was used.Each curve corresponds to a particular number N of magnetic ions in the cluster, as indicated.The insets show the same data on a logarithmic scale.(a) System parameters γ, ∆, h, T varied randomly within the ranges 0 < γ < 1, 0 < ∆ < 1, 0 < h < 2J, and 0.01 < T < 2J (variation with respect to the exchange energy scale J is not necessary as it merely sets the overall energy scale).
(b) h, T varied randomly within the same ranges as before with fixed γ = 0.6 and ∆ = 0.
are measured in units of J).For each set of values we can use the method in Refs.[27,34] to compute S (q) for a fixed set of wave vectors q.This results in a set of images which can then be classified by a standard PCA algorithm.Our expectation is borne out by the scree plots in Fig. 2 (a).Specifically, we find that M = 4 PCs suffice to capture 99% of the variance in the data set for the range of values of N shown in the graph.
In an experimental situation, we expect the parameters defining the strength and anisotropy of magnetic interactions to be fixed for a given material, while the strength of the externally-applied magnetic field and temperature can vary.Scree plots for a representative case (γ = 0.6, ∆ = 0) are shown in Fig. 2 (b). 3 We find that now M = 2 captures 99% of the variance in the training set.
Our results indicate that the number M of PCs necessary to reproduce to very high accuracy all the images in the training set is bound by the number M p of free parameters used to generate the training images.This might suggest an unbiased (model-independent) way to constrain experimentally the number of independent parameters describing a class of related materials -an important step in the derivation of a model Hamiltonian.We note, however, that M is more likely related to the number of distinct states in the phase diagram which, although related to M p , can potentially be larger than it.Elucidating this will require the analysis of a broader range of models and is beyond the scope of the present work.
We highlight that unlike an autoencoder, where the number M n of neurons in the hidden layer has to be fixed a priori, our approach enables us to find out a posteriori the number M of PCs needed to describe the data accurately.One would in principle expect M n ∼ M and thus for our model M n ∼ M p .In contrast, for the model of Ref. Samarakoon et al. [26] with M p = 4 dimensionless parameters M n = 30 was found to strike a good balance between overfitting and underfitting.This would suggest that in our case we achieve greater dimensionality reduction.
Although PCA is equivalent to an autoencoder in the linear limit, we must note that the two studies are rather different.In Samarakoon et al. [26] the aim is to automatically interpret structure factor data, from noisy measurements and to disambiguate among many possible solutions of the inverse scattering problem, whereas our approach is to achieve an embedding of the data that is amenable to human interpretation.Our work does not explicitly address experimental noise.However, we note that both autoencoder and PCA are known to have noise suppressing properties and therefore we anticipate that our method would also be applicable to experimental data.

IV. SIZE OF TRAINING SET
Once the value of M has been set, other images not included in the original training set can be accurately reconstructed using the first M PCs.For this to be possible, the following two conditions must be met: i) the new images must be of the same type as those in the training set; ii) the training set must be sufficiently representative.The latter criterion will only be met if the number of images in the training set M t is sufficiently large.This limits the feasibility of obtaining training sets experimentally.Fig. 3 shows three versions of the first 2 PCs for a particular instance of our model (N = 4,γ = 0.6, ∆ = 0).Each version has been obtained using a different training set, shown graphically in the first row of panels: a large training set obtained from 500 randomly-chosen values of (h, T ); a much smaller training set generated from 9 randomly-chosen values of (h, T ); and a minimal set formed by 3 values of (h, T ), chosen strategically (one with high T one with low T and low h, and one with low T and high h) 4 4 .Appendix B gives further details about the smaller of the two random sets, including a display of the individual simulated neutron scattering images that compose it.Our results indicate that M t can indeed be very small.This is consistent with the small number of principal components necessary to describe the data accurately (see previous section) and is confirmed by Fig. 4 which shows the reconstruction of a particular instance of S (q) using the three sets of PCs.We note that the reconstructed image is not present in any of the training sets.

V. SCORE BIFURCATION PLOTS
Figs. 2, 3 and 4, taken together, imply that a relatively small number of initial observations can be used to determine a few PCs in terms of which data obtained subsequently can be accurately described -in other words, by projecting new observations onto the low-dimensional space spanned by the PCs, we obtain a low-dimensional representation of the data in terms of the PC scores.The dependencies of such scores on system parameters such as magnetic field h or temperature T can then be used to identify features in the phase diagram.Such approach, when applied to microscopic data on classical states, has been shown capable of detecting, for example, symmetrybreaking phase transitions [38].Here we ask whether the same benefit can be obtained when working with observable averages such as S (q) for our quantum magnet model.We will answer in the affirmative and moreover present a useful analytical tool based on this idea, which we call a "score bifurcation plot".Our simulations indicate that this technique may facilitate the detection of qualitative changes in the ground state of a quantum system, even from data taken at relatively high temperatures.
To illustrate our method we consider first the simplest case of our model where N = 2.Such S = 1/2 dimer has two possible ground states: a low-field anti-ferromagnetic state | ↑↓ − | ↓↑ with anti-parallel entanglement and a high-field ferromagnetic state | ↑↑ + δ| ↓↓ with parallel entanglement (as h → ∞ the amplitude δ → 0 resulting in the classical state | ↑↑ ).The two states are degenerate at the factorisation field h f .Panels (a) and (b) of Fig. 5 show the weights, or scores, of the first two principal components, wPC1,wPC2, obtained by projecting the PCs onto the simulated neutron scattering function S (q), for two different training sets.In panel (a) the weights have been obtained for the scattering functions in the original training set.We can appreciate a marked difference between the high-temperature data, concentrated in a small region of wPC1-wPC2 space, and the low-temperature data which cover a much wider area.This is reminiscent of results obtained for microstates of Ising-type models -c.f.Fig. 3 (b) in Ref. [38], where we also see that, at high temperatures, the data are clustered in a small, high-entropy region of PC space while, at lower temperatures, the data fan out as distinct ground states are selected by their energies.In our case, however, we are examining the statistical average S (q) which is a function of parameters h, T that can, in an experimental situation, be controlled externally.It is therefore possible to explore the space of scattering functions systematically by varying h and T and projecting the new measurements onto the PCs deduced from the training set.A simula-   ,c) correspond to additional values of (h, T ) not present in the training set.These have been obtained by varying h from 0 to 2J in steps of ∆h = 0.05J and T from 2J to 0.01J in steps of ∆T = 0.01J.Each line corresponds to a fixed value of the field h.The isolines corresponding to the lowest and highest field values are the the ones reaching furthest to the right and furthest to the top of the graph, respectively.The curves highlighted in cyan are those for which the field is within ∆h/2 of the factorisation field h f = 4J/5 (b,c) or the additional ground-state level-crossing field h 1 = 0.35J (c).The inset to panel (c) shows in detail the low-temperature behaviour near h ≈ h 1 .The colour in all panels encodes temperature, as indicated.Note that the axes limits differ in each case since (a) and (b) are plots obtained using principal components for the same physical system but generated using 9 and 500 images respectively and (c) represents a different system altogether.
tion of that approach is presented in panel (b), which shows the evolution of wPC1 and wPC2 with temperature for different fixed values of the magnetic field (fixedfield temperature scans).We observe a marked difference between the curves corresponding to fields lower than the factorisation field h f and those greater than that field.Each constant-field temperature scan is represented by a single curve in the space defined by the two principal components.As the temperature is lowered, the curve starts to bend in a direction that indicates the nature of the ground state (ferromagnetic if h > h f and antiferromagnetic if h < h f ).This manifests as a marked bifurcation, with the curve corresponding to h = h f (high-lighted in cyan) marking the bifurcation where the directions of this bending changes sign.At this field, below some finite temperature T * the system gets "stuck" at a particular point (wPC1 * ,wPC2 * ) and does not evolve further.This suggests that such "score bifurcation plots" can be used to elucidate systematically the ground-state phase diagram from finite-temperature data, even in systems such as the one we model where there are no finitetemperature phase transitions.Appendix D describes a geometric construction that can be used to determine h f accurately from limited finite-temperature data.
Further evidence of the above hypothesis is provided in Fig. 5 (c) where similar data are presented for N = 4.
As we reviewed in Sec.II, two special values of the field h 1 , h f are expected to emerge at low temperatures in this case.Indeed we find two bifurcations occurring at those fields, within the resolution given by our field increment ∆h = 0.05J.We note however that the bifurcation at h f , where the ground state factorises leading to the vanishing of entanglement measures, is detectable at a higher temperature than that at the level crossing field, where entanglement is suppressed but does not vanish [27].We have verified that the factorisation field is also seen in similar score bifurcation plots obtained for N = 6.

VI. EXPERIMENTAL RESOLUTION
In Sec.III we showed that PCA could achieve good dimensionality reduction for neutron scattering data simulated using our model.In Sec.IV we further showed that this could be achieved using surprisingly small training sets, in the sense of containing very few observations.We will now address the question of experimental resolutionspecifically, how many pixels each individual observation needs to have for the score bifurcation plots introduced in the last section to yield accurate values of the critical fields.We note that modern neutron scattering instruments allow a trade off between neutron flux and resolution at the time when the measurement is being made [39].In addition, one can always group pixels together, post-measurement, to form a more coarse-grained, but less noisy image.We note that the authors of Ref. 26 have addressed the question of noise in a different way, namely by applying their methodology directly to noisy images, and their conclusions are similar to ours.
To address our question we have repeated some of our previous calculations using lower-resolution images both at the training and analysis stages.Specifically, we replace our previous 24 × 24 pixel matrices with 8 × 8 matrices which corresponds to nearly an order of magnitude reduction in the amount of data in each observation.The results, for the N = 4 case, are shown in Fig. 6.Clearly, the score bifurcation plot obtained from these lower-resolution images is as useful as that obtained before, and in particular it allows us to pinpoint the critical fields h 1 and h f to the same values, within our field-scan accuracy ∆h = ±0.025J.

VII. DISCUSSION AND CONCLUSIONS
In this work we have presented a method to obtain quantitative information about the phase diagram of a quantum system from experimentally observable data, namely the diffuse magnetic neutron scattering function S (q).Our method is based on a simple form of unsupervised machine learning, PCA, and provides a visual representation of the data that facilitates a greater under- standing of the underlying physics.We addressed our research question theoretically by analysing simulated scattering functions for a simple model of a cluster quantum magnet.
Our method is based on using a small training set of observations to determine PCs describing the data.We then analyse subsequent observations by projection onto those PCs.We found that this procedure can achieve a large degree of dimensionality reduction i.e. very few PCs suffice for an accurate description of subsequent observations.Consequently, we found that effective training requires only a small number of observations.Moreover, each observation may be a very low-resolution image, which should facilitate greatly the implementation of our methodology in a real experimental setting.Finally, we devised a way to use the trained PCA algorithm to characterise the system's phase diagram.
Our method to investigate phase diagrams relies on temperature scans at fixed values of another control parameter (in our case, magnetic field h -however the method can be trivially generalised to other control parameters such as pressure).By plotting the evolution of the system in PC space we find paths with bifurcations occurring at the values of the field that are known to correspond to changes in the system's ground state.
It is interesting to speculate why a linear technique, PCA, can provide such a compact and enlightening description of a system dominated by strong correlations.In this respect we note that our calculation of the scattering function S(q) was carried out in the linear-response regime [27,35].This is standard in the theory of magnetic neutron scattering and is justified by the fact that the neutron acts as a weak perturbation [10].Whether this can be used as the starting point for a justification of the applicability of PCA is an interesting question but lies outside the scope of the present work.
We hasten to add that the linear dependence of S(q) on the principal components does not imply that S(q) is a linear function of the model parameters γ, ∆, T, h.The scattering function is given by where the score w n (γ, ∆, h, T ) of the n th PC is in general a highly non-linear function of γ, ∆, h and T .Thus in our dimensionally-compressed representation of the data S(q) is a linear function of just a few S n (q)'s but the nonlinearity of the model is still reflected in the dependence on system parameters of the expansion coefficients.It is nevertheless remarkable that the non-linear features of the model can be accurately captured in this way.An explicit example of such non-linear behaviour of the PC scores is provided in Appendix C.
Our method provides an efficient way to extrapolate the ground-state phase diagram from finite-temperature data, even in a system that is effectively of finite size and therefore lacks well-defined, finite-temperature phase boundaries.One may speculate that applying such methodology to poorly-understood real systems such as copper-based high-temperature superconductors might offer a fresh perspective on an old conundrum in the theory of strongly-correlated electron systems: are the correlated quantum "liquid" phases found at finite temperature best thought of as manifestations of the quantum critical points separating the different quantum-ordered phases (as proposed by Laughlin and co-workers [4])?Or are they best regarded, instead, as condensations of the "gaseous" phase existing at higher temperatures, which become susceptible to different quantum ordering transitions as the temperature is lowered further (as put forward by Anderson [5])?Addressing this question will require applying our method to experimental data on real systems.For instance, a score bifurcation plot of measurements taken in the "strange metal" phase of the cuprate phase diagram may provide evidence for proposals that puzzling crossovers observed in that region are due to a Widom line associated with a hidden, lowtemperature critical endpoint [40].Here we discuss briefly the field-and temperaturedependences of the individual principal component scores wPC1, wPC2.These are shown in Fig. 8 for the same calculation used to obtain Fig. 5 (b).As we can see the temperature-dependence of the PC scores changes qualitatively at the factorisation field -for instance at low temperatures the derivative of wPC1 with respect to T is positive for h > h f and negative for h < h f .At exactly h = h f (highlighted in cyan) this derivative vanishes in the limit T → 0. We note, however, that in a general situation we will not have a constant-field scan taken at exactly h = h f (see footnote 3 on page 4).Because of this, in general locating the value of the field where the PC scores become temperature-independent involves an extrapolation.This must be contrasted with the score bifurcation plot where wPC1 and wPC2 are shown in combination and the factorisation field can be determined precisely from finite-temperature data.Here we describe a simple geometric construction for determining precisely the location of the factorisation field from finite-temperature data.We illustrate this by focusing on the N = 2 data in Fig. 5.The same data is reproduced in Fig. 9, but with additional symbols showing the points along the constant-field curves where the temperature reaches two particular values, namely T = 0.10 and T = 0.15.By inspection of that plot it is quite clear that the two curves will cross at the point where the bifurcation takes place, whose precise coincidence with the factorisation field we have determined numerically.It is evident that this construction will work, and give the same value, irrespective of which two temperatures we choose, as long as they are low enough to show the inflection of the curve to the right of the bifurcation.A similar construction can be made for the N = 4 plots in Figs. 9 (c) and 6.I.The panel labels (a-i) correspond to the index on the first column of Table I.

Figure 1 :
Figure 1: Our simple model of a planar, ring-shaped cluster magnet for the cases with N = 4 (left) and N = 6 (right) magnetic ions in the cluster.Reproduced from Ref. [27].

Figure 3 :
Figure 3: Dependence of the PCs on the choice of training set for fixed N = 4, γ = 0.6, ∆ = 0. (a-c) Values of the magnetic field h and temperature T used to generate each training set.(d-f) First PC for each of the respective training sets.(g-i) Second PCs.

Figure 4 :Figure 5 :
Figure 4: Neutron scattering cross-section S (q) predicted by our model for γ = 0.6, ∆ = 0, h = 0.72J, and T = 0.36J (a) and its reconstruction using only two PCs (b-d).The training sets and their corresponding PCs are the ones shown in Fig. 3 consisting of 3 training images (b), 9 training images (c) and 500 training images (d).

Figure 6 :
Figure 6: (a) Score bifurcation plot for the same parameters as in Fig. 5 (c) except that here all observations (both those used in training as well as all subsequnet observations) consist of much lower-resolution, namely 8 × 8 pixel images.(b) An example of a scattering function obtained by exact diagonalisation at that resolution (the other parameters are as in Fig. 4).
tron scattering functions.Each image constitutes one element of the training set (an "observation").This 9image training set was used to obtain the principal components shown in Figs. 3 (e) and (h).Projecting the 9 images shown in Fig. 7 onto those principal components we obtain the coordinates of the 9 filled circles shown in Fig. 5 (b).The lines shown in the latter figure are obtained by projecting other images (not forming part of the training set) onto the same principal components.A similar procedure was followed when using the 3-image training set corresponding to the (h, T ) values in Fig. 3 (a) and the 500-image training set corresponding to Fig. 3 (c).Appendix C: Temperature-and field-dependence of individual principal component scores Appendix D: Geometric construction for determining the factorisation field from finite-temperature data

Figure 7 :
Figure 7: Training set generated using the values of magnetic field h and temperature T shown in Fig. 3 (b) and TableI.The panel labels (a-i) correspond to the index on the first column of TableI.

Figure 8 :
Figure 8: Temperature-and field-dependence of the individual PC scores wPC1 (a) and wPC2 (b).The data shown are from the same calculation as Fig. 5 (b).

Figure 9 :
Figure 9: Here we reproduce the plot in Fig. 5 (b) but with the addition of green and blue circles marking the points along the constant-field lines corresponding to temperatures T = 0.10 and 0.15, respectively.

Table I :
Numerical values of the combinations of magnetic field h and temperature T represented graphically in Fig.3 (b).The values are given in units of the exchange constant J, as indicated in the table headings.