Figures
Abstract
How to distribute a set of points uniformly on a spherical surface is a longstanding problem that still lacks a definite answer. In this work, we introduce a physical measure of uniformity based on the distribution of distances between points, as an alternative to commonly adopted measures based on interaction potentials. We then use this new measure of uniformity to characterize several algorithms available in the literature. We also study the effect of optimizing the position of the points through the minimization of different interaction potentials via a gradient descent procedure. In this way, we can classify different algorithms and interaction potentials to find the one that generates the most uniform distribution of points on the sphere.
Citation: Del Bono LM, Nicoletti F, Ricci-Tersenghi F (2024) The most uniform distribution of points on the sphere. PLoS ONE 19(12): e0313863. https://doi.org/10.1371/journal.pone.0313863
Editor: Mattia Miotto, Italian Institute of Technology, ITALY
Received: July 16, 2024; Accepted: October 31, 2024; Published: December 27, 2024
Copyright: © 2024 Del Bono et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data are available on Zenodo, with DOI: 10.5281/zenodo.14198415.
Funding: The research has received financial support from ICSC - Italian Research Center on High-Performance Computing, Big Data, and Quantum Computing, funded by the European Union - NextGenerationEU. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
1 Introduction
Distributing a finite set of points uniformly on a spherical surface is a fundamental problem, dating back to at least the beginning of the XX century [1, 2]. It has applications, among others, in biology [3, 4], engineering [5], theoretical physics [6–8] and astronomy [9] and is, in general, a relevant question every time a discretization of the sphere is required. However, the first difficulty one encounters when dealing with this kind of problem is how to properly define uniformity.
A first definition of uniform distribution of a set of points on the sphere focuses on the approximation of integrals on the sphere [10]. Indeed, one can search for a distribution of points such that the difference between the numerical integration carried on using these points and the analytic result is small. Expanding on this idea, one can define t−spherical designs as sets of points that allow for perfect integration (i.e. the numerical and analytical results coincide) for all polynomials up to degree t [11].
Another very studied and rather classical definition of uniformity for a set of Np points is based on the minimization of the interaction potential among these points. The first problem of this kind was proposed by Thomson who considered Np electrons on the sphere interacting through a Coulomb electric potential [1]. The Thomson problem has been recently generalized by taking into account the possibility of electrons being inside the sphere [12] and interacting through different potentials decaying as r−α at large distance [13]. A more geometrical definition of uniformity focuses on packing a given number of circles on the sphere such that the minimum radius is maximized: this is the so-called Tammes problem [2], and can be seen as the α → ∞ limit of the Thomson problem.
On the one hand, the use in the Thomson problem of smooth and soft potentials (e.g. the Coulomb one or anyone with α not too large) allows for a direct application of standard optimization techniques, such as the Gradient Descent (GD) and the quasi-Newton method. On the other hand, the sphere packing approach in the Tammes problem would be better to generate points uniformly on the sphere, as it does not require the introduction of an arbitrary potential, but its solution has a much higher computational cost. Our analysis suggests a way to bridge the gap between the theoretical optimum and what is computationally achievable.
Given the arbitrariness in the definition of uniformity, over the years many sphere discretization algorithms have been proposed, all developed with different goals in mind. Algorithms for finding the ground state configurations of a given potential range from more geometrical approaches [14] to a brute-force optimization of many randomly generated configurations [15], passing through more sophisticated procedures such as Simulated Annealing [16]. Algorithms developed for practical applications may have additional constraints due to the specific use they are intended for. This is the case of sphere discretizations used in the analysis of astronomical data [17, 18]. In the present work, we are going to analyze only a small number of such algorithms, but our criteria can be applied to any other procedure.
Our definition of uniformity is based on quantifying how crystalline-like the distribution of points is. In the Euclidean space, a regular lattice can be straightforwardly defined through discretized transformations, e.g. translation by lattice vectors. However, on a sphere such regular lattices do not exist, but for few values of Np. Indeed, the only truly regular lattices are the sets of vertices of the five platonic solids. So, in general, the most uniform distribution of points on a sphere is not a perfectly regular lattice and contains defects. This is the essential reason why the definition of the most uniform distribution of points on a sphere is not easy to provide.
Our criteria for uniformity will consider first the minimization of defects and then the distance between points, especially nearest neighbors. Indeed we aim primarily at controlling the discretization of the sphere locally. This is a crucial requirement when one is interested in approximating a function defined on the sphere with function values computed on a discrete set of points. An illuminating application requiring such a uniform local discretization is the approximation of marginals for Heisenberg models that we have solved recently [19].
As expected, if one is interested in optimizing the nearest neighbors’ distances, short-ranged potentials (i.e. large α values) are better and provide locally improved crystalline order. Under this new uniformity criterion, long-ranged potentials (e.g. the Coulomb one) are very inefficient, because any given point interacts too much with far away points and this is not useful at all to optimize the discretization locally. Thanks to our analysis, we can provide a recipe to both generate distributions at the state-of-the-art level and to benchmark future methods, thus allowing us to study the most uniform distribution of points on the sphere.
It is important to stress that spatially uniform distributions are different from the so-called hyperuniform distributions. Indeed, the concept of hyperuniformity on the sphere, which will be described in detail in Sec. 2.3, is a long-distance property of the system, which does not take much into account local properties of the systems, as proven by the huge variety of hyperuniform points distributions found in previous works [20, 21]. Our results further confirm this claim.
This paper is organized as follows. In Sec. 2 we summarise known results and give some definitions that will be used throughout the rest of the manuscript. In particular, we define a defect in a distribution of points on the sphere (Sec. 2.1), we define the different measures of uniformity one can introduce (Sec. 2.2) as well as the concept of hyperuniformity (Sec. 2.3). In Sec. 3 we introduce the methods we use in this work. Specifically, we briefly describe the algorithms analyzed in this paper (Sec. 3.1) and some possible choices of the potentials for applying a gradient descent optimization procedure (Sec. 3.2). In Sec. 4 we report the original results we have obtained and point out what is the best method to distribute points uniformly on the sphere. Specifically, we compare the different algorithms to a random distribution (Sec. 4.1), we compute the distribution of distances (Sec. 4.2), we check whether the methods satisfy hyperuniformity criteria (Sec. 4.3), we study thoroughly the first peak of the distribution of distances to find the most uniform grid (Sec. 4.4), we perform extrapolation to infinitely sharp potentials and to an infinite number of points (Sec. 4.5) and we check stability against perturbations (Sec. 4.6). Finally, in Sec. 5 we draw our conclusion and summarize the main results of the paper.
2 Theoretical background
2.1 Defects
It is well-known that in the plane the highest density packing of circles is the hexagonal one [22]. A perfectly hexagonal lattice cannot, however, be embedded on the sphere surface. Indeed [11], from Euler’s formula
(1)
(
,
and
being the number of faces, vertices and edges, respectively) and assuming that faces are triangles whose vertices only have 4, 5, 6 or 7 neighbors, it follows that
(2)
where
is the number of vertices with i neighbours. Here, the neighborhood can be defined in terms of the Voronoi diagrams. Given a set of points
on the sphere, the Voronoi decomposition is obtained associating to each point the region
of all the points on the sphere which are closer to
than to any other
. As a consequence, it is impossible not to have any defects (points with more or less than six neighbors, also called disclinations [23]). Moreover, from Eq (2) follows that, if
then
. The most symmetric property that we can require from these twelve minimal defects is that they are placed at the vertices of a regular icosahedron. Configurations that have this property are called icosadeltahedral. Algorithms such as the Lattice Points (LP) method (see Sec. 3.1) generate configuration with this minimal number of defects, while the Polar Coordinates Subdivision method does not.
It has been shown that icosadeltahedral configuration, and in particular configurations generated through LP, are not generally the lowest energy configuration, e.g. for the Coulomb potential [24, 25], despite the low-energy configurations still displaying symmetries when defects are taken into account [26]. Indeed, optimal configurations start to develop more defects the more points are placed on the sphere, usually near each other in the form of dislocations [27], and for large values of Np only a very small fraction of configurations is defect-free [28].
Since grids that minimize Coulomb energy usually have many defects, and since we expect that a uniform mesh should have a structure that is locally as close as possible to that of a hexagonal lattice, it follows that minimizing the Coulomb energy might not be the best descriptor of uniformity for practical purposes. One possible reason could be that the Coulomb interaction is long-ranged and, thus, allows points that are very far away to interact with each other. On the other hand, in a hexagonal lattice points should mainly interact with neighbors. We shall later show evidence in favor of this intuition by comparing different potentials.
2.2 Measures of uniformity
As already mentioned, there is no unique way to define how uniform a distribution of points on the sphere is. One common approach is to define a potential and then try to find the distribution that minimizes (or, in some special cases, maximizes) it. Common choices for the energy [29] are the aforementioned Coulomb potential
(3)
where the sum runs over all couples of different points and the distance is the Euclidean distance; its generalization (Riesz s−energy)
(4)
and the minimum spacing (related to the Tammes problem)
(5)
Other frequently used potentials are the logarithmic and harmonic energies:
(6)
(7)
The problem then becomes that of finding the set of points that minimizes (in the minimum spacing and harmonic cases, maximizes) the energy. One of the advantages of these procedures is that there are results (such as the Poppy-seed Bagel theorem) that guarantee that minimal energy configurations of suitably chosen potentials are well-behaved in the asymptotic limit.
In this work, we have instead chosen to focus mainly on the distribution g(r) of distances r between pairs of different points,
(8)
From a practical point of view, g(r) is more commonly studied as a histogram, so that the δ centered at close-by distances are coarse-grained together, resulting in a smoothed-out behavior of the function. The use of g(r) is connected to the theory of crystallography, in which the structure factor is connected to the charge distribution of ions in the crystal [30]. Indeed, in a perfect crystal, despite the coarse-graining, g(r) is still the sum of many δ functions centered at the distances between first neighbors, second neighbors, third neighbors, and so on. On the sphere, on the other hand, the peaks are widened due to the curvature and the presence of defects. Despite this, we expect that the more the distribution of points is regular, i.e. the more it locally resembles a crystal, the narrower the peaks will be. Since a crystalline-like configuration would be the most uniform, we expect that the better-behaving distributions are those for which g(r) is closer to that of a crystal.
To quantify this approach, we have studied the first peak of the g(r), i.e. the one connected to distances between first neighbors. Subsequent peaks can be studied, but the analysis becomes noisier due to the increasing number of neighbors (e.g. the second peak includes both points at distance 2 and at distance , in units of first neighbors distance) and for large r it becomes altogether impossible to separate different peaks.
2.3 Hyperuniformity
The concept of hyperuniformity was first introduced approximately twenty years ago in the Euclidean space setup. It has recently been applied to distributions of points on the sphere.
In Euclidean space, hyperuniformity is identified by a vanishing structure factor S at low wave vectors , i.e. by
(9)
Intuitively, this property corresponds to the requirement that the systems does not have infinite-wavelength fluctuations, that is, that the system is uniform on the large scale. On the sphere, the structure factor of a distribution of Np points is [20]:
(10) γij being the spherical distance between points i and j. Pℓ are the Legendre polynomials. In this case, the wave vector
is substituted by the wave number ℓ and therefore the condition in Eq (9) becomes the requirement that the structure factor vanishes at low ℓ. It is important to notice that, while
is a continuous variable, ℓ is discrete-valued and therefore the limit ℓ → 0 is not well-defined, so the uniformity criterion used in practice is that the structure factor in Eq (10) becomes zero for sufficiently small ℓ.
An additional definition of hyperuniformity on the sphere can be given in terms of the spherical cap variance, . For a given angle θ, the spherical cap variance is defined as
(11)
where
is the number of points in the spherical cap of opening angle θ and centered in the point
and the averages are taken over all possible centers on the
sphere.
It can be shown [20] that the spherical cap variance is connected to the spherical structure factor by
(12)
Eq (12) implies a vanishing spherical cap variance for θ = π/2 for distributions of points symmetric under parity (such as the distributions generated using the LP method).
Moreover, it was suggested by Božič and Čopar that the spherical cap variance behaves as
(13)
where
and
are numerical coefficients that identify the presence of hyperuniformity (namely, hyperuniformity is reached for
) and ξ(θ) is an additional oscillatory contribution, stronger the more crystalline-like the distribution is.
Hyperuniformity can be defined in terms of the limit for increasing θ, where s(θ) is the surface of a spherical cap with opening angle θ [21]. Of course, such a limit cannot be truly achieved since we are working on a finite-size sphere and, consequently, the requirement becomes that the ratio decreases sufficiently rapidly for θ → π/2.
It is to be noted that both the definitions of hyperuniformity given above rely on long-range, large-scale properties of the system. Hence, we expect that hyperuniformity will not give much information on the local structure of the set of points. Indeed, as shown in [20, 21], a huge variety of different distributions satisfy the criteria of hyperuniformity. This result is in agreement with the results we obtained in this work, which are presented in Sec. 4.3.
3 Methods
3.1 Algorithms
The procedures we have considered to generate points uniformly on the unitary sphere are listed below.
- Lattice Points (LP) method [14, 25] consisting in the following steps:
- first, an icosahedron inscribed in the sphere is considered;
- then, points are added on a face of the icosahedron by mapping the face to an equilateral triangle in the complex plane
with vertices at 0, m + nζ and mζ + nζ2, where
and m, n non-negative integers such that m ≥ n;
- points on the complex plane of coordinates k + lζ, with k and l integers, which lay inside or on the boundary of the triangle are added to the face;
- points on the other faces of the icosahedron are obtained through a similar mapping and appropriate rotations of the points to match them on the edges;
- the final distribution of points is obtained by projecting the points from the faces of the icosahedron onto the sphere.
The number of points Np generated for a given pair m, n is Np = 10(m2 + n2 + mn) + 2; - Mathematica’s SpherePoints (MSP) function [31], which takes as input an integer Np and gives as output an approximately uniform distribution of Np points on the sphere;
- Polar Coordinates Subdivision (PCS) method [25]. Denoting by φ, θ the polar coordinates on the sphere, the angles
(n is a parameter that fixes the number of points generated) are considered. For each φj,
equally spaced points are placed at angles
. One can add the two poles (we did not). Finally, a shift on alternate latitudes is introduced to symmetrize the distribution. This method can be slightly modified to increase the number of Np available, but we did not consider this modified version in this work;
- the HEALPix package (https://healpix.sourceforge.io/), a structure to obtain iso-latitude, equal-area pixels on the sphere organized in a hierarchical way, which is widely used in the analysis of astronomical data [17]. HEALPix creates pixels by subsequent subdivision of twelve original pixels. At each step, each pixel is divided into four sub-pixels. At the end of the procedure, a discretization of the sphere can be obtained by taking each pixel’s center;
- Icosahedron-based Equal Area Pixelization (IEAP) method [32], which consists of the following steps:
- first, the faces of an icosahedron with a regular triangular grid are pixeled. Each pixel is identified by a point which represents its center;
- the points obtained in the previous step are projected onto the unit sphere;
- the points are shifted around slightly so that each pixel has an equal area.
All the above methods take a time linear in the number of points. A Gradient Descent (GD) procedure can then be applied to the sets of points obtained by the above algorithms to further improve the final result. We will discuss some possible choices of potentials to perform the GD optimization and their effects. In general, when considering potentials without any cutoff like the ones described in Sec. 2.2, the computational cost of a single gradient descent step scales quadratically in the number of points, since each point feels a force caused by all the other points. However, if the potential decays to zero fast enough (for example, in the case of a Riesz with s large), one can introduce a cutoff rc, setting to zero the interaction between points that are more than rc apart. This allows us to reduce the computational cost to a linear cost in the number of points.
More sophisticated forms of optimization can be used, such as quasi-Newton methods [14] or more complicated procedures [27], at a higher computational cost. However, given that the number of GD steps to optimize the potentials is very small, we do not see good reasons for using more complicated optimization algorithms. Thus we limit ourselves to the study of GD and its simplicity allows us to perform a thorough study of many different potentials in the optimization procedure.
Examples of point distribution obtained using the five algorithms described above are shown in Figs 1 and 2.
Examples of point distributions generated using different algorithms (Lattice Points and Mathematica’s SpherePoints), together with the corresponding Voronoi tessellation. Blue points have six neighbors, while red points are defects with less or more than six neighbors.
More examples of point distributions generated using different algorithms (HEALPix, Icosahedron-based Equal Area Pixelization and Polar Coordinates Subdivision), together with the corresponding Voronoi tessellation. As in Fig 1, blue points have six neighbors, while red points are defects with fewer or more than six neighbors.
3.2 Potentials for the GD
To study the effects that different potentials have on the GD procedure, we consider 4 kinds of potentials:
- Power-law potentials:
(14)
- Power-law potentials with an exponential cutoff:
(15)
- Scaled power-law potentials:
(16)
- Scaled power-law potentials with an exponential cutoff:
(17)
The characteristic length scale r0 has been chosen as the average distance between two nearest neighbors, estimated assuming a hexagonal lattice local structure for the distribution:
(18) η being the packing density of the 2D hexagonal lattice,
. The correctness of expression in Eq (18) has been verified by looking at both non-optimized and optimized configurations (see Fig 8 in the Results section).
Since we are interested in distributions that are uniform at a local level, we have not considered potentials like the logarithmic one. Indeed, it has been shown that the logarithmic potential allows for the screening of the defects [33, 34] (that is, an arrangement of the points surrounding a defect such that the energy cost of such a defect is vanishingly small in the thermodynamic limit), thus allowing defect-abundant configuration to exist at relatively low energies.
4 Results
Now we present the results obtained using different algorithms and optimizing potentials. Distances are always Euclidean, given that the difference with the geodesic distances is very small when considering nearby points and can be safely neglected.
4.1 Comparison with the random case
First, we compare the probability distribution of distances found using the algorithms LP, MSP, PCS and IEAP with the theoretical prediction for points distributed at random according to a uniform distribution. For a unit hypersphere in dimension m, the probability distribution of distances, in the random case, is given by
(19)
where
(20)
which simply reduces to Prand(r) = r/2 in our m = 3 case. The simple derivation of the previous formulas is shown in S1 Appendix. The comparison is shown in Fig 3. For all algorithms, fluctuations from the linear behavior of the random case are present, signaling the presence of structure. In particular, these fluctuations appear to be bigger for the LP algorithm at smaller r (the region we are most interested in when looking for crystalline-like order) and for the IEAP algorithm at larger r. Similarly, in Fig 3 we also considered the scalar products and the angles between points. For random points on the sphere we have
(21)
for scalar products and
(22)
for angles, which reduce to Pu(u) = 1/2 and Pθ(θ) = sin(θ)/2 in the m = 3 case we are considering.
Comparison between the probability distributions of distances, scalar products and angles between points obtained via LP, MSP, PCS and IEAP methods and the theoretical prediction in the case of points placed uniformly at random on the sphere.
4.2 Distributions of distances
In Fig 4, the histograms of the average number of points at a distance r from a given point, obtained through algorithms LP, MSP, PCS and IEAP are shown. In the LP and IEAP cases, the first peak (corresponding to nearest neighbors) can be easily identified, whereas in the MSP and PCS the first peak mixes with the second one. The HEALPix distribution has a behavior qualitatively similar to that of PCS and MSP, with a first peak not well separated from the second one. Moreover, it is a rather sparse algorithm in this range of Np. That is, there is only one configuration with a number of points between 200 and 3000. This makes HEALPix unsuited for analyzing the effects of discretization, e.g. to study finite size effects. For these two reasons, it was not considered further.
LP, MSP, PCS and IEAP methods are considered. Red and blue bars represent the distribution before and after Coulomb GD optimization, respectively. The first peak is well separated for LP and IEAP, but not for MPS and PCS. Notably, GD makes the first peak sharper.
Fig 5 shows the histograms obtained using the different methods (followed by a Coulomb GD) while the type of points (normal points with six neighbors or defects) is taken into account. PCS and MSP produce a huge number of defects.
Bars are placed on top of each other, so the total distribution of points is given by the highest contour. Different colors identify different kinds of points interacting. Bottom, teal: normal-normal (NN). Middle, red: normal-defect (ND). Top, yellow: defect-defect (DD). In the LP and IEAP methods, the number of defects is so small compared to the number of normal points that defect-defect interactions do not appear visually in the figure.
It is clear that the first peak of the distribution is the one that carries the most information about the local structure of the distribution, so we will analyze it more carefully. Before doing so, however, we present the results obtained testing hyperuniformity criteria on the different distributions.
4.3 Hyperuniformity results
We calculated both the structure factor and the spherical cap variance for three distributions of points obtained through the LP (Np = 482), MSP (Np = 500) and PCS (Np = 538) methods. The spherical cap variance and the spherical cap variance normalized using the surface of the cap are shown in Fig 6. It is clear that, in all three cases, the ratio goes to zero for increasing θ, thus suggesting that all the distributions of points satisfy hyperuniformity. This is further confirmed by the behavior of the structure factor, shown in Fig 7. Indeed, a vanishing S(ℓ) is observed at small ℓ, together with a well-defined first peak.
(a) spherical cap variance and (b) spherical cap variance normalized by spherical cap area, as functions of the angle, for different point distribution: LP (Np = 482, with no optimization), MSP (Np = 500) and PCS (Np = 538). ρ is the density of points, i.e. Np/4π. The spherical cap variance was estimated taking the average over 20000 random centers of the cap for every angle. Notice the vanishing for θ = π/2 for the LP method. This is a consequence of the symmetry of the distribution under space inversion, which implies a zero structure factor for odd ℓ, which in turn means that only odd Legendre polynomials enter in the sum of Eq (12).
Main plot: structure factor as a function of the wave number ℓ for different point distribution: LP (Np = 482, with no optimization), MSP (Np = 500) and PCS (Np = 538). Inset: zoom in on the ℓ ≤ 100 region. For the LP method, only even ℓ are considered, since the structure factor vanishes at odd ℓ for symmetry reasons.
We notice that in Fig 6 the behaviors of the three distributions appear to follow Eq (13). The LP configuration, however, shows bigger fluctuations and actually seems to possess two natural frequencies, compared to only one as in the MSP and PCS cases. Since the oscillatory term ξ(θ) in Eq (13) is connected to crystalline-like ordering (it was observed that random distributions of points do not fluctuate at all), this might be additional proof that LP method indeed produces more ordered structures than the other two methods.
As previously stated, hyperuniformity does not seem to be a strong enough criterion to identify spatially uniform distributions. It is, however, interesting to notice that, while PCS and MSP give distributions that reach a unitary structure factor relatively quickly, it takes LP much longer to reach the asymptotic regime. The structure factor must go to 1 for ℓ → ∞. Indeed, S = 1 for a completely random distribution of points. Since large ℓ corresponds to large wave vectors in the Euclidean case, it seems reasonable to consider this slower decay as an indication of the presence of local structure.
4.4 Study of the first peak
In order to study the uniformity of the distribution, we studied the first peak of the g(r), which, apart from a normalization factor, coincides with the histograms previously obtained.
First of all, we can introduce a measure of how well-separated the first peak is, in order to support the visual evidence presented in Figs 4 and 5. We notice that the rightmost point in the first peak should, in the icosadeltahedral case, correspond to the n* = 6(Np − 2)-th point of the distribution. Moreover, it should be well separated from the points on the right and very close to the points on the left. So we can consider the ratio
(23)
where ri is the i-th smallest distance between 2 points and ν is a parameter chosen to smooth the behaviour. As an alternative, we can also consider the ratio (rn*+ν − rn*)/(rn* − rn*−ν). In the well-separated case, R should be close to 0, because the terms in the numerator sum are small, since the corresponding points belong to the first peak (rj ≃ rn*, j = n* − ν, …, n* − 1), while the terms in the denominator sum are relatively large, since the corresponding points belong to the second peak (rj > rn*, j = n* + 1, …, n* + ν). On the other hand, if the first peak is not well-separated from the second one, then there is no particular reason why rj should be much larger than rn* for j > n*, so the sums are approximately equal and R should be close to 1 (or at least of the same order of magnitude). In Tables 1 and 2 the values of R for different choices of parameters in the LP and MSP cases are presented. As expected, in the case of a well-defined peak R ≈ 0, while in the case of a crossover behaviour R ≈ 1.
From the discussion made in Sec. 2.1 and Sec. 2.2, it is clear that, of all the algorithms considered, only those that produce icosadeltahedral configurations, or that at least keep the first peaks well-separated, have a good distribution of points. Of all the methods considered in the previous section, only LP and IEAP satisfy this criterion. The latter produces configurations for fewer values of the number of points. Indeed, LP generates configurations with Np = 10(m2 + n2 + mn) + 2, for , as previously mentioned, while IEAP produces configurations with Np = 40n(n − 1) + 12,
. and has a slightly larger value, after the GD, of the STD/mean observable (as described later). We therefore focus on the LP method. Since we found a good option for the starting algorithm, the next logical step is to consider different potentials for the GD procedure and see how things change.
Now we study the effect of optimizing the points positions by minimizing different interaction potentials. We analyzed the first peak of the distribution obtained after a GD optimization, considering its mean, the difference between the mean and the minimum, the standard deviation STD and the ratio STD/mean. Results are shown in Figs 8 to 11.
Mean distance inside the first peak as a function of the number Np of points placed on the sphere, both for the original configuration (left) and for the configuration optimized by GD (right). In the latter case, the V3 and V4 potentials are considered. The dotted line corresponds to the characteristic length scale r0 (that is, the estimated average distance between two neighbours) and is given by Eq (18).
Difference between the mean of the distances inside the first peak and the minimum distance as a function of the number Np of points placed on the sphere, both for the original configuration (left) and for the configuration optimized by GD (right). Insets show zoom-ins of the region at larger values of Np. Top: V1 and V2. Bottom: V3.
Standard deviation of distances inside the first peak as a function of the number Np of points placed on the sphere, both for the original configuration (left) and for the configuration optimized by GD (right). Insets show zoom-ins of the region at larger values of Np. Top: V1 and V2. Bottom: V3 and V4.
Ratio between the standard deviation and the mean of distances inside the first peak as a function of the number Np of points placed on the sphere. and V4 potentials are considered. Both graphs share the same y axis.
The mean appears to be largely unaffected by the GD (Fig 8), as can be expected since the total surface of the sphere is fixed. Moreover, it follows the expected curve given by Eq (18).
The difference between the mean and the minimal value in the first peak is generally a non-monotonic function of α, especially in the V3 case (Fig 9). It is however important to remember that this observable, taking into account the minimum distance between two points, measures one extreme of the distribution, thus it is strongly affected by outliers. It is nonetheless interesting to notice that the GD procedure greatly reduces the difference between the minimum and the average, that is, it makes outliers much closer to the typical, average value, thus acting as a sort of regularizer.
A more stable measure is the standard deviation (Fig 10). Forcing the potential to be more and more short-ranged (i.e. increasing α) makes the peak sharper and sharper, decreasing the STD. A comparison of the values obtained shows that V3 and V4 perform better. This is reasonable since these potentials cut interactions between far-away points, which we expect should matter little in the formation of a crystalline structure.
A clearer understanding is obtained considering the ratio STD/mean (Fig 11). Of all the potentials taken into account, V3 with α = 7 is the one performing better. In general, for equal α, V4 performs better than V3. However, due to the computational complications of dealing with exponentials, the former requires more care to be implemented at higher values of α. All in all, it would seem that the V3 is particularly well suited for the task at hand and it is, therefore, the one we studied more in-depth.
The metrics presented in this section can be better suited for discriminating locally uniform distribution of points than simply evaluating the energy of a configuration. As an example, the distributions of points in Fig 1 generated using the LP method have intensive energy () equal to 0.4757 for Np = 492 and equal to 0.4838 for Np = 1082. Instead, those generated using SpherePoints have lower energies: 0.4753 for Np = 492 and 0.4833 for Np = 1082, despite having a huge number of defects. This makes clear that lower energy does not imply a smaller number of defects, and so using the energy as a proxy of uniformity is wrong. On the other hand, the LP-generated distributions have values R = 5.94 ⋅ 10−15 (ν = 100) and STD/mean = 0.074 for Np = 492 and R = 8.15 ⋅ 10−15 and STD/mean = 0.073 for Np = 1082, while the SpherePoints-generated distributions have values R = 3.37 and STD/mean = 0.136 for Np = 492 and R = 1.11 and STD/mean = 0.120 for Np = 1082. Thus the use of R and STD/mean clearly discriminates the most uniform distribution of points.
4.5 The α → ∞, Np → ∞ extrapolation
Since, in principle, we are interested in the limit of large Np and infinite α (sharp potential connected to Tammes problem), we want to quantify the limitations that arise from considering only finite values of Np and α, and what is the difference from the asymptotic case. First of all, we plot the ratio STD/mean as a function of α for different Np, corresponding to different m and n (Fig 12). We considered only cases with Np large enough for the behavior to be monotonically decreasing and the cases with m = n or at least similar since it has been conjectured that configurations with m much greater than n have high (Coulomb) energy and are far from optimal [14]. We then performed a sigmoidal fit using a function f(x;A, B, C) = C + B[1 − S(x − A)], where S = 1/(1 + e−x). Extrapolation to large values of α gives variations of the order of 0.1% with respect to the α = 7 case for all Np.
Ratio STD/mean as a function of α for different Np, together with sigmoidal fits and extrapolation at Np → ∞.
For each α we then performed a fit using the function:
(24)
The results are presented in Fig 13. We then considered the behaviour of
and c(α) as functions of α (Fig 14) and performed sigmoidal (as before) and exponential fits, respectively.
Ratio STD/mean as a function of for different α, together with fits and extrapolations at Np → ∞.
STD/mean(∞;α) and c(α) coefficients as a function of α, together with sigmoidal and exponential fits (dotted lines).
Finally, in Fig 12 is plotted together with the values obtained at finite Np.
In the end, the effect of finite Np is the one that seems to play the greatest role, with effects of the order of 10%, while the effect of finite α is small. Thus, using values of α ≥ 7 should already yield good results, close enough to the infinitely sharp case.
4.6 Stability
Another question that arises naturally when analyzing these distributions is whether they are stable under the application of a small perturbation. We studied the stability of the distributions obtained using the LP procedure and different values of α. In order to do so, we used the following procedure:
- we generated Np points using the LP algorithm;
- to all points we applied a random Gaussian perturbation with zero mean and standard deviation equal to the 5% of the mean distance between first neighbours, as obtained through Eq (18). Then we projected the points back on the sphere;
- we performed GD using V3 with different α and saw if the distribution of the standard deviation and of the ratio
was the same as the one obtained without the addition of the perturbation.
We did these steps for α = 1, 7, repeating the process 5 times for each value of α. Results are presented in Fig 15. The choice α = 7 cancels differences much faster (all cases basically collapse on each other after GD) and recovers quickly the values obtained in absence of perturbation. We also performed the same procedure with a preliminary GD optimization between the first and the second steps. Results present no significant difference with respect to the shown case.
STD (a) and STD/mean (b) as a function of Np for distribution of points with added noise, before and after GD. Points (circles and squares) are perturbed configurations. Crosses are distributions of points after LP algorithm (left) and distributions after GD (right).
All things considered, the α = 7 case not only demonstrates a superior capability in reducing the ratio STD/mean, but it also appears to be considerably more stable.
5 Conclusions
In conclusion, in this work, we have introduced a framework to analyze the uniformity of distributions of point on the sphere. We introduce new measures alternative to those commonly found in literature. We believe the observables described in this paper are better descriptors of the proprieties of uniformity usually required in practical applications.
We have used these new tools to study a variety of algorithms that are aimed at producing uniform points on the sphere. Further we have optimized these point distributions through the minimization of several interaction potentials via the gradient descent algorithm. Finally, we have characterized finite size effects in an extensive way.
In the end, we found that the Lattice Points algorithm followed by a GD optimization with power-law potential with a cutoff performs particularly well, and already the potential with α = 7 seems to yield very good results. This is a very good compromise between the optimal potentials (α ≫ 1) and potentials smooth enough to be optimized quickly.
The methods used in this paper can be further generalized to study different algorithms and optimization methods, therefore they allow us to find the most uniform distribution of points on the sphere.
Supporting information
S1 Appendix. Computation of g(r) in the random case.
https://doi.org/10.1371/journal.pone.0313863.s001
(PDF)
References
- 1. Thomson JJ. XXIV. On the structure of the atom: an investigation of the stability and periods of oscillation of a number of corpuscles arranged at equal intervals around the circumference of a circle; with application of the results to the theory of atomic structure. The London, Edinburgh, and Dublin Philosophical Magazine and Journal of Science. 1904;7(39):237–265.
- 2. Tammes P. On the origin of number and arrangement of the places of exit on the surface of pollen-grains; 1930.
- 3. Barreira R, Elliott CM, Madzvamuse A. The surface finite element method for pattern formation on evolving biological surfaces. Journal of mathematical biology. 2011;63:1095–1119. pmid:21274536
- 4. Lorman VL, Rochal SB. Density-Wave Theory of the Capsid Structure of Small Icosahedral Viruses. Phys Rev Lett. 2007;98:185502. pmid:17501583
- 5.
Verhaevert J, Van Lil E, Van de Capelle A. Uniform spherical distributions for adaptive array applications. In: IEEE VTS 53rd Vehicular Technology Conference, Spring 2001. Proceedings (Cat. No.01CH37202). vol. 1; 2001. p. 98–102 vol.1.
- 6. Alcaraz F, Herrmann H. Numerical difficulties in obtaining 3D critical exponents from Platonic solids. Journal of Physics A: Mathematical and General. 1987;20(16):5735.
- 7. Weigel M, Janke W. Universal amplitude-exponent relation for the Ising model on sphere-like lattices. Europhysics Letters. 2000;51(5):578.
- 8. Zhu W, Han C, Huffman E, Hofmann JS, He YC. Uncovering conformal symmetry in the 3D Ising transition: state-operator correspondence from a quantum fuzzy sphere regularization. Physical Review X. 2023;13(2):021009.
- 9. Bauer R. Distribution of points on a sphere with application to star catalogs. Journal of Guidance, Control, and Dynamics. 2000;23(1):130–137.
- 10.
Beentjes CH. Quadrature on a spherical surface. Working note available on the website http://people.maths.ox.ac.uk/beentjes/Essays. 2015;.
- 11. Saff EB, Kuijlaars AB. Distributing many points on a sphere. The mathematical intelligencer. 1997;19:5–11.
- 12. Levin Y, Arenzon JJ. Why charges go to the surface: A generalized Thomson problem. Europhysics Letters. 2003;63(3):415.
- 13. Bowick M, Cacciuto A, Nelson DR, Travesset A. Crystalline Order on a Sphere and the Generalized Thomson Problem. Phys Rev Lett. 2002;89:185502. pmid:12398614
- 14. Altschuler EL, Williams TJ, Ratner ER, Tipton R, Stong R, Dowla F, et al. Possible global minimum lattice configurations for Thomson’s problem of charges on a sphere. Physical Review Letters. 1997;78(14):2681.
- 15. Erber T, Hockney G. Complex Systems: Equilibrium Configurations of N Equal Charges on a Sphere (2 < = N > = 112). Advances in chemical physics. 1997;98:495–594.
- 16. Xiang Y, Gubian S, Suomela B, Hoeng J. Generalized Simulated Annealing for Global Optimization: The GenSA Package. The R Journal Volume 5(1):13-29, June 2013. 2013;5.
- 17. Gorski KM, Hivon E, Banday AJ, Wandelt BD, Hansen FK, Reinecke M, et al. HEALPix: A Framework for High-Resolution Discretization and Fast Analysis of Data Distributed on the Sphere. The Astrophysical Journal. 2005;622(2):759–771.
- 18. Malkin Z. A new equal-area isolatitudinal grid on a spherical surface. The Astronomical Journal. 2019;158(4):158.
- 19.
Del Bono LM, Nicoletti F, Ricci-Tersenghi F. How to compute efficiently the analytical solution to Heisenberg spin glass models on sparse random graphs and their de Almeida-Thouless line. arXiv preprint arXiv:240616836. 2024;.
- 20. Božič AL, Čopar S. Spherical structure factor and classification of hyperuniformity on the sphere. Physical Review E. 2019;99(3):032601.
- 21. Meyra AG, Zarragoicoechea GJ, Maltz AL, Lomba E, Torquato S. Hyperuniformity on spherical surfaces. Physical Review E. 2019;100(2):022107. pmid:31574707
- 22.
Chang HC, Wang LC. A simple proof of Thue’s theorem on circle packing. arXiv preprint arXiv:10094322. 2010;.
- 23. Wales DJ, Ulker S. Structure and dynamics of spherical crystals characterized for the Thomson problem. Physical Review B. 2006;74(21):212101.
- 24. Pérez-Garrido A, Dodgson M, Moore M. Influence of dislocations in thomson’s problem. Physical Review B. 1997;56(7):3640.
- 25. Katanforoush A, Shahshahani M. Distributing points on the sphere, I. Experimental Mathematics. 2003;12(2):199–209.
- 26. Pérez-Garrido A, Moore M. Symmetric patterns of dislocations in Thomson’s problem. Physical Review B. 1999;60(23):15628.
- 27. Wales DJ, McKay H, Altschuler EL. Defect motifs for spherical topologies. Physical Review B. 2009;79(22):224115.
- 28. Altschuler EL, Pérez-Garrido A. Defect-free global minima in Thomson’s problem of charges on a sphere. Physical Review E. 2006;73(3):036108. pmid:16605599
- 29. Alishahi K, Zamani M. The spherical ensemble and uniform distribution of points on the sphere. 2015;.
- 30.
Ashcroft NW, Mermin ND. Solid State Physics. Holt-Saunders; 1976.
- 31.
Research W. SpherePoints; 2022. https://reference.wolfram.com/language/ref/SpherePoints.html.
- 32. Tegmark M. An Icosahedron-based Method for Pixelizing the Celestial Sphere. The Astrophysical Journal. 1996;470(2):L81.
- 33. Moore M, Pérez-Garrido A. Absence of a finite-temperature melting transition in the classical two-dimensional one-component plasma. Physical review letters. 1999;82(20):4078.
- 34. Dodgson M, Moore M. Vortices in a thin-film superconductor with a spherical geometry. Physical Review B. 1997;55(6):3816.