Preprint
Article

This version is not peer-reviewed.

Fisher Information in Helmholtz-Boltzmann Thermodynamics of Mechanical Systems

A peer-reviewed article of this preprint also exists.

Submitted:

12 May 2025

Posted:

14 May 2025

You are already at the latest version

Abstract
In this paper we review Helmholtz-Boltzmann thermodynamics for mechanical systems depending on parameters and we compute the Fisher information matrix for the associated probability density. The divergence of Fisher information has been used as a signal for the existence of phase transitions in finite systems even in the absence of a thermodynamic limit. We investigate through examples if qualitative changes in the dynamic of mechanical systems described by Helmholtz-Boltzmann thermodynamic formalism can be detected using Fisher information.
Keywords: 
;  ;  ;  

1. Introduction

The first part of the paper is devoted to an introduction of Helmholtz-Boltzmann (HB) thermodynamics for mechanical systems. In [1,2] Helmholtz proved that: 1) for a one-dimensional conservative mechanical system with potential energy u ( x , λ ) – where λ is a parameter– a probability density p = f ( e u ( x , λ ) ) can be defined on the region of configuration space corresponding to motions with energy e; moreover: 2) mechanical analogs of temperature T, pressure P and entropy S can be introduced such that the first principle of thermodynamics in the form T d S = d e + P d λ holds for these systems. In the same years, Boltzmann [3] derived the analog of the probability density p for a mechanical system of n dimensions, and investigated the ergodic hypothesis for these systems. Interestingly, this picture of Statistical Mechanics based on a probability density defined in the configuration space preceded the Gibbs formulation of statistical mechanics [4] whose ensembles are defined on phase space. In recent times HB thermodynamics has been investigated in a series of papers [5,6,7,8].
In the second part of this paper we compute the Fisher Information matrix for mechanical systems described by HB thermodynamics. Fisher information (FI) matrix [9,10], a notion originally developed in statistical estimation theory, is nowadays at the crossroad between information theory [11,12], differential geometry [13], and Statistical Mechanics [14]. Recently, FI has been used to detect phase transitions in machine learning [15], neural networks [16], and in quantum systems even in the absence of a thermodynamic limit [17]. For finite systems, the existence of a phase transition is related to the divergence of one or more of the FI matrix elements, or more simply, it is located in correspondence to their maxima. Within this approach, which merges statistical mechanics and information theory, divergence of FI matrix are used to single out order parameters and reveal phase transitions, see [18].
However, for systems described by HB thermodynamics, all the information on the dynamics is encoded in the form of the potential energy u ( x , λ ) where x R d and λ R k so clearly we are dealing with a finite-dimensional system which is not composed of N interacting elementary units as in Statistical Mechanics. Moreover, there is no clear indication for the identification of one or more parameters λ as system’ volume, therefore, we cannot carry over for these systems the thermodynamic limit. We thus necessarily need to understand a possible notion of "phase transition" in a broader sense, as a qualitative change in some of the system’ features or patterns when one or more tunable parameters cross a threshold, as in the bifurcation theory for dynamical systems. Essentially, with HB thermodynamics, which is a statistical theory based on a probability density p = f ( e u ( x , λ ) ) , we are between the statistical mechanics and bifurcation theory approaches, and this drives the interest for these systems and shapes the tools that we are going to use.
In Section 3, we state the conditions under which the probability density p = f ( e u ( x , λ ) ) can define a statistical model and compute the related FI matrix. We examine the relationship between the elements of the FI matrix and the second-order derivatives of the free energy for the density p = f ( e u ( x , λ ) ) which is not an exponential family and make a comparison with the corresponding notions for the canonical density (an exponential family) used in Statistical Mechanics. In Section 5 we show with paradigmatic examples that the divergence in the FI elements can detect qualitative changes in the form of probability density that describe the system (which can be interpreted as a signature of phase transition in finite systems) or it can locate the transition from negative to positive specific heat. We are aware that these results are not enough to shape a general theory, but we hope that this first step may help to renew interest for Helmholtz-Boltzmann approach to thermodynamics of mechanical systems.

2. Helmholtz-Boltzmann Thermodynamics of Mechanical Systems

In this Section we give a short exposition of Helmholtz-Boltzmann thermodynamics. Let us consider the unidimensional motion of a point of mass m under the action of a positional force f ( x ) = u ( x ) where u is the potential energy. This mechanical system is conservative and the energy integral
m 2 v 2 + u ( x ) = e
defines a 1-dimensional orbit in phase plane ( x , v ) which can be locally described as v ( x , e ) = 2 m φ where φ = e u is the kinetic energy. If the potential energy is convex, the motion corresponding to a fixed value e of the energy is confined in the convex interval
H ( e ) = { x R : φ = e u > 0 } = ( x 1 ( e ) , x 2 ( e ) )
and it is periodic. Denoting with d t = d x / v ( x , e ) the time needed to travel a space interval d x , the semi-period of the motion is expressed by the generalized integral
T ( e ) = H ( e ) d x v ( x , e ) = m 2 H ( e ) φ 1 2 d x = m 2 Z ( e ) .
We have that v = 0 on the boundary of H ( e ) but the generalized integral T ( e ) is convergent if the orbit corresponding to the value of e it is not a separatrix. We can thus define a probability density on H ( e ) as (the value of m is not relevant)
p ( x , e ) = φ c ( x , e ) Z ( e ) = φ c ( x , e ) H ( e ) φ c ( x , e ) d x where c = 1 2 .
As an example, if u ( x ) = x 2 is the elastic potential energy and e = a 2 , a > 0 , we have φ = a 2 x 2 and p is the arc-sine probability density defined in ( a , a )
p ( x , a ) = 1 Z ( a 2 x 2 ) 1 2 = 1 π a 2 x 2 .
We denote with f the average of a function defined on H ( e ) with respect to the probability p in (3).
The aim of Helmholtz and Boltzmann (see [5] for an historical recognition of the theory) was to provide a mechanical analogy of the first principle of thermodynamic in Gibbs form
T d S = d e + P d V
where S is the entropy, T is the temperature, P is the pressure and V is the volume.
This can be obtained by supposing that the potential energy depends on a parameter λ as u = u ( x , λ ) and by defining the temperature and the pressure as the averages of φ (which is the kinetic energy) and of φ λ = φ / λ with respect to the probability density p in (3) –which now depends on λ
T ( e , λ ) = 2 φ , P ( e , λ ) = φ λ .
The entropy S is defined as the logarithm of the area in phase space enclosed by the orbit of energy e which gives
S ( e , λ ) = ln ( 2 H v ( x , e , λ ) d x = ln Z ( e , λ ) φ + c o n s t .
One can check that the above defined thermodynamic analogs of T , P , S for a mechanical system do satisfy the following relation, which is the mechanical equivalent of the first principle of thermodynamics
d S = S e d e + S λ d λ = 1 2 φ [ d e + φ λ d λ ] = 1 T [ d e + P d λ ]
The proof is a straightforward application of the formula of derivation of a integral depending on a parameter, see [5,7].

2.1. The Multidimensional Probability Density

In [7] an extension of the above result to the case of x R d , d 1 and λ U R k is provided (see also [8] for a different approach). For the sake of completeness and to provide a basis for further development we give an account of it here. The first step is to generalize the entropy formula (5) and the probability density (3) to the d-dimensional case. To this aim we consider a system of n point masses of mass m moving in R 3 referred to coordinates x R d , d = 3 n and subject to potential energy u ( x , λ ) . The total energy is m 2 v 2 + u ( x , λ ) = e and, analogously to what we have done before, we define
Q ( e , λ ) = { ( x , v ) R 2 d : m 2 v 2 + u ( x , λ ) e }
to be the portion of phase space where the energy is not greater than e. We define as in the microcanonical approach of Statistical Mechanics (see [5,8] for a justification of this definition) the entropy of the system as the logarithm of the measure of Q in phase space R 2 d
S ( e , λ ) = ln Σ ( e , λ ) = ln Q ( e , λ ) d x d v .
If we set as before
H ( e , λ ) = { x R d : φ = e u ( x , λ ) > 0 }
then
Q ( e , λ ) = x H ( e , λ ) { ( x , v ) : v 2 2 m φ } = x H ( e , λ ) { ( x , v ) : v B d ( 2 m φ ) }
where B d ( r ) is the ball of radius r in R d . Let R ( x , e , λ ) = 2 m φ . Using Fubini theorem we can factorize the above entropy integral as
Σ = e S = Q ( e , λ ) d x d v = H ( e , λ ) B d ( R ) d v d x
and the inner integral can be computed by "integrating on spheres" as
B d ( R ) d v = α d 0 R r d 1 d r = α d d R d = ω d R d = ω d ( 2 m φ ) d 2
where ω d is the measure of the unit ball in R d . Therefore, up to unessential constant term the entropy is
S ( e , λ ) = ln H ( e , λ ) φ d 2 d x .
Setting
H φ d 2 d x = H Z φ φ d 2 1 Z d x
if we define the probability density on the subset H ( e , λ ) R d of configuration space as –see (3)–
p ( x , e , λ ) = φ c Z ( e , λ ) = φ c H φ c d x , where c = d 2 1
then the entropy integral (9) can be expressed in the same form of (5)
S = ln ( Z φ ) .
Remark. We will use the above introduced probability density (10) for a generic system whose configuration space has dimension d without making reference to a point particle system. Note that for d 2 we have c = d 2 1 > 0 hence a definite integral in (10) and φ c defines a normalizable probability density.
For later use, it will be useful to express the probability density (10) in exponential form by setting
h = ln φ ( x , e , λ ) = ln ( e u ( x , λ ) ) , Ψ = ln Z ( e , λ )
therefore (10) can be given the form
p ( x , e , λ ) = e c h H e c h d x = e c h Z ( e , λ ) = e c h ( x , e , λ ) Ψ ( e , λ )
and the entropy in (11) becomes
S = ln ( Z φ ) = Ψ + ln φ .

2.2. Multidimensional HB Thermodynamics

Our aim is to prove that for a mechanical system described by a potential energy u ( x , λ ) with x R d , λ U R k a mechanical analog of the first principle of thermodynamics holds. In this section, we recall a number of results that will be used in the proof of the multidimensional HB thermodynamics result.
We start with recalling the Leibnitz integral rule, also called the Reynold transport theorem in continuum mechanics, which generalizes to R d the formula of derivation for the integral depending on parameters.
Proposition 1.
Let f = f ( x , α ) be a real function depending on one or more parameters α to be integrated on a domain D ( α ) R d . Then
α D ( α ) f d x = D ( α ) f α d x + D f n · v d σ
where D is the boundary of D, n is its outer normal unit vector and v is the speed of the boundary. If D = { x R d : φ ( x , α ) 0 } then D = { x R d : φ ( x , α ) = 0 } and it holds that
n = φ | φ | , v = φ α | φ | n .
Using the above Proposition 1, we want to compute the partial derivatives of the free entropy Ψ in (12) with respect to α where α denotes the parameters e or λ . If we assume that c > 0 so that e c h = φ c = 0 on H , the boundary term in (15) is vanishing and we have (here and in the following we set f , α = f / α )
Ψ , α = 1 Z α H e c h d x = 1 Z H e c h ( c h , α ) d x = c h , α .
If we introduce the log-likelihood function associated to the probability density in (13)
l = ln p = c h Ψ
we get the result
l , α = c h , α Ψ , α = c ( h , α h , α ) = p , α p
and we can prove the following useful formula:
Proposition 2.
Let f = f ( x , α ) be a real function such that f p = 0 on H . Then we have that
f , α = α H f p d x = H ( f p ) , α d x = H ( f , α p + f p , α ) d x = f , α + c H f p ( h , α h , α ) d x = f , α + c ( f h , α f h , α ) = f , α + c c o v ( f , h , α ) = f , α + c c o v ( f , φ , α φ )
As a straightforward application of Proposition 2 for f = φ we can compute (here h , α = φ , α φ )
φ , α = φ , α + c φ h , α φ h , α = ( 1 + c ) φ , α c φ h , α .
Now all the elements to prove the multidimensional version of HB thermodynamics have been introduced .
Proposition 3.
Let the entropy function for a mechanical system with potential energy u ( x , λ ) be
S ( e , λ ) = ln H ( e , λ ) φ d 2 d x = ln Z + ln φ = Ψ + ln φ
and define the temperature T ( e , λ ) and the i type pressure P i ( e , λ ) as
T = φ 1 + c = 2 d φ , P i = φ , λ i .
Then it holds that
d S = 1 T [ d e + i P i d λ i ]
Proof. 
Let α = e or α = λ i . We have from (16) and (19) that
S , α = Ψ , α + ( ln φ ) , α = c h , α + 1 φ [ ( 1 + c ) φ , α c φ h , α ] = 1 + c φ φ , α
and, since φ , e = 1 we derive that for S ( e , λ )
d S = S e d e + i = 1 k S λ i d λ i = 1 + c φ [ d e + i = 1 k φ , λ i d λ i ]
= 1 T [ d e + i P i d λ i ]
From the above result, temperature and pressure can be computed using the relations
T 1 = S e , P i T = S λ i .

2.3. Relation with Microcanonical Entropy

In HB thermodynamics, the entropy is the volume entropy defined as S = ln Σ as in (6). If we choose to adopt the definition of the entropy of the microcanonical ensemble we introduce the so-called density of states g ( e , λ ) and the microcanonical entropy S μ
g ( e , λ ) = Σ ( e , λ ) e , S μ ( e , λ ) = ln g ( e , λ ) .
Since from (9)
Σ ( e , λ ) = ω d H φ c + 1 d x
we have
g ( e , λ ) = Σ ( e , λ ) e = ( c + 1 ) ω d H φ c d x = ( c + 1 ) ω d Z ( e , λ )
therefore, up to constant terms, the microcanonical entropy S μ is
S μ ( e , λ ) = ln g ( e , λ ) = ln Z ( e , λ ) = Ψ ( e , λ )
and we have from (14) the relation between HB and microcanonical entropy
S = S μ + ln φ .
For most thermodynamic systems the two definitions of entropy are equivalent in the limit of n , where n is the number of particles of the system, see e.g. [19]. If we define the microcanonical temperature as T μ 1 = S μ / e we have, from (22) and (16)
1 T μ = S μ e = Ψ , e = c 1 φ
which is different from the definition of temperature in HB thermodynamics, T = 2 φ / d , see (20). So the two ensembles, for finite number of degree of freedom systems, are non-equivalent.
We now compute the specific heat in the HB and microcanonical cases. Note that since the temperature in HB thermodynamics is defined as the average value of the kinetic energy, c v is a dimensionless quantity. We define
1 c v = d T d e = 2 d φ , e
and using again Proposition 2 we get
1 c v = 2 d [ φ , e + c c o v ( φ , h , e ) ] = 2 d [ 1 + c c φ 1 φ ] = 1 T T μ
so the specific heat c v diverges when T = T μ . Using the microcanonical temperature (23) we have
d d e ( T μ 1 ) = 1 T μ 2 d T μ d e = 1 T μ 2 1 c v μ = c 1 φ , e = c 1 φ 2 + c 2 v a r ( 1 φ )
therefore
1 c v μ = T μ 2 c 1 φ 2 c 2 v a r ( 1 φ )
Note that, unlike what happens with the canonical ensemble, where the specific heat is necessarily positive (see (36) below), in the HB and microcanonical case the specific heat can be negative.

3. Statistical Models and Fisher Matrix

The notion of Fisher information matrix [9,10] is fundamental to describes the geometry of statistical models. We start by recalling the definition of a regular statistical model; see [12,13]. Let p θ ( x ) be a probability density defined on state space X depending on finitely many parameters θ Z R d . We introduce the set
S = { p θ = p ( x , θ ) : θ Z } L 1 ( X )
Definition 1.
S is a regular statistical model if the following conditions 1) and 2) hold:
1)
(injectivity) the map f : Z S , θ f ( θ ) = p θ is one to one and
2)
(regularity) the d functions defined on X
p i ( x , θ ) = p θ i ( x , θ ) , i = 1 , , d
are linearly independent as functions on X for every θ Z .
The inverse φ : S Z , φ ( p θ ) = θ of the map f, which exists by 1), defines a global coordinate system for S . It is convenient to introduce the log-likelihood l = ln p and the score basis l i = l / θ i . Note that since l i = ( 1 / p ) p i the function p i and l i are proportional, therefore, the regularity condition 2 ) above holds if and only if the elements of the score basis are independent functions over X. The element of the Fisher matrix are defined as follows
g i j ( θ ) = l i l j = X l θ i l θ j p ( x , θ ) d x .
The Fisher matrix is symmetric and positive definite; therefore, it defines a Riemannian metric on Z . (see [13], p.24). In fact we have (sum over repeated indices is understood)
g i j v i v j = l i l j v i v j = ( l i v i ) 2 = 0 l i v i = 0 v i = 0 i
since the score vectors l i are linearly independent over X. Note also that g is invariant with respect to change of coordinates in the state space X and covariant (as an order 2 tensor) with respect to change of coordinates in the parameter space [13].

3.1. Statistical Models in HB Thermodynamics

In this section we consider the HB-type probability density introduced in (13) and we compute the related Fisher matrix. We define the parameters θ Z R 1 + k where
θ 0 = e , θ i = λ i , i = 1 , k .
and the probability density –see (13)–
p ( x , θ ) = e c h ( θ ) Ψ ( θ ) = φ c ( x , θ ) H ( θ ) φ c ( x , θ ) d x , c = d 2 1
defined on the set
H ( θ ) = { x R d : φ = θ 0 u ( x , θ ) > 0 } .
Unlike the theory exposed above in Section 3, the probability densities p θ , θ Z in (29) are not defined in the same sample space X. To embed this case in the previously exposed theory we need to embed every probability density in the same space R d by setting
p ˜ ( x , θ ) = p ( x , θ ) if x H ( θ ) 0 if x H ( θ )
In the following we will spare the tilde symbol in p ˜ ( x , θ ) . We suppose that conditions 1) and 2) do hold for p ( x , θ ) . If we compute the score basis we have from (17) and (18)
l i = l θ i = c h , i Ψ , i = c ( h , i h , i ) = c φ , i φ φ , i φ , i = 0 , k
and the elements of the Fisher matrix (26) are
g i j ( θ ) = l i l j = c 2 ( h , i h , i ) ( h , j h , j ) = c 2 c o v ( h , i , h , j ) .
Another useful expression for the elements of the Fisher matrix can be deduced from Proposition 2: we have from (16) that Ψ , i = c h , i hence
Ψ , i j = c h , i , j = c [ h , i j + c c o v ( h , i , h , j ) ] = c h , i j + g i j
therefore
g i j = Ψ , i j c h , i j .
Note that unlike what happens with exponential families, the Fisher matrix does not coincide with second-order derivatives of the free entropy function Ψ . However, the above formula coincides with the one obtained for Fisher matrix for generalized exponential families, [20,21]. Below we give detailed formulae for the elements of the Fisher matrix associated to HB thermodynamics. From (28) and (12) we have
h , 0 = φ , 0 φ = 1 φ , h , i = φ , i φ = u , i φ
so that
g 00 ( θ ) = c 2 c o v ( h , 0 , h , 0 ) = c 2 1 φ 2 1 φ 2 = c 2 v a r ( 1 φ )
and
g 0 i ( θ ) = c 2 c o v ( h , 0 , h , i ) = c 2 u , i φ 2 1 φ u , i φ
and
g i j ( θ ) = c 2 c o v ( h , i , h , j ) = c 2 u , i u , j φ 2 u , i φ u , j φ
More in detail, the explicit expression of the Fisher matrix elements involves the computation of integrals of the type
1 φ 2 = H ( θ ) p ( x , θ ) 1 φ 2 d x = 1 Z ( θ ) H ( θ ) φ c 2 d x
where f = φ c 2 satisfies the condition f = 0 on H if c 2 = ( d 6 ) / 2 0 .

4. Fisher Matrix for Canonical Ensemble

In this section we recall a well known relation between Fisher matrix and second order derivatives of the canonical free entropy function. In statistical mechanics a system of N microscopic units can be described by the Boltzmann-Gibbs (BG) or canonical probability density
p ( x ) = e β H ( x ) Ψ ( β ) , Ψ ( β ) = ln Z ( β ) = e β H ( x ) d x
where H ( x ) is the Hamiltonian and the integral (or sum) is extended over the set of system states x. The inverse temperature β = 1 / k T can be related to the system energy e = H by inverting the equation e ( β ) = Ψ , β ( β ) so that β = β ( e ) . As a consequence
2 Ψ β 2 = v a r ( H ) = d e d β = ( d β d e ) 1 = ( 1 k T 2 d T d e ) 1 = k T 2 c v > 0 .
Second order phase transitions are characterized by discontinuities in the function Ψ , β β . Ψ is an analytic function for finite systems but in the thermodynamic limit f = lim N 1 N Ψ can develop discontinuities. If now we consider the log-likelihood of the canonical probability density l = ln p = β H Ψ in (35) we have
l β = l β = H Ψ , β = H + H
and the Fisher information of the canonical density (35) is
g β β = l β 2 = v a r ( H ) = 2 Ψ β 2 .
so from (36) and (37) we get the important result that for an exponential family density the Fisher matrix coincides with the second order derivatives of the free entropy and that the diagonal elements of the Fisher matrix diverge at the phase transition when c v tends to infinity.
If the system Hamiltonian is in the form θ · X = i = 1 k θ i X i the above relations are generalized as
X i = Ψ θ i , g i j = X i θ j = 2 Ψ θ i θ j = c o v ( X i , X j ) .
In [18] the above argument is reversed and the claim is that the divergence of the diagonal elements of the Fisher matrix is a signature of a phase transition even if the underlying probability density is not the canonical one. The order parameter is defined by the relation ( 38 ) 1 and ( 38 ) 2 defines the generalized susceptibility. The same argument is developed in [15] in relation to the study of phase transitions using machine learning methods. To conclude, we recall that phase transitions have been defined for finite systems or for systems with long-range interactions using the non convexity of the entropy function or the non equivalence of canonical and microcanonical ensembles (see [22,23]).

5. Fisher Matrix in HB Thermodynamics. Examples

The HB thermodynamic formalism allows us to define a probability density see (10) which is related to the microcanonical one but it is "projected" on the configuration space. Instead of prescribing the interaction between the elementary units of the system, all the information about the system is encoded in its potential energy u ( x , θ ) depending on multiple parameters and the energy θ 0 . So we are using a formalism, which loosely speaking is in between statistical mechanics and catastrophe theory, and a probability density, which is not an exponential family. Nevertheless we still have, see (16) and (32) that for multidimensional HB thermodynamics –compare with (38)–
Ψ , α = c h , α , g i j ( θ ) = c 2 c o v ( h , i , h , j ) = 2 Ψ θ i θ j c h , i j .
We want to investigate whether the divergence of elements of the Fisher matrix for particular values of parameters θ can detect qualitative changes in the dynamics in mechanical systems described by potential energy u ( x , θ ) .
In the following, we consider some examples where the potential energy u ( x , θ ) , x R d is described by a radial function such that u ( x , λ ) = w ( | x | , λ ) . This allows us to investigate our claim while keeping the computations at a tractable level. Note that, apart from the elastic chain, all the examples below define a statistical model, as it is easy to check using Definition 1.

5.1. Harmonic Potential Energy

Let us consider the simplest potential energy with no external parameters w ( | x | ) = x · x = x 2 with x R d . This energy represents d independent one-dimensional harmonic oscillators. Then φ = e x 2 and H ( e ) = B d ( e ) , the ball of radius e in R d . Let us compute
Z ( e ) = H ( e ) φ c ( x , e ) d x = α d 0 e ( e r 2 ) c r d 1 d r , Σ ( e ) = H ( e ) φ c + 1 ( x , e ) d x = α d 0 e ( e r 2 ) c + 1 r d 1 d r
so that
Z ( e ) = π Γ ( d 2 ) 2 d Γ ( d + 1 2 ) e d 1 , Σ ( e ) = π Γ ( d 2 ) 2 d + 1 Γ ( d + 1 2 ) e d
where Γ is the Gamma function and the entropy function are
S ( e ) = ln Σ = d ln e + c o s t ( d ) , S μ = ln Z = ( d 1 ) ln e + c o s t ( d ) .
By direct computation we get the temperature
T = ( S e ) 1 = e d = 2 d φ , T μ = ( S μ e ) 1 = e d 1 = 1 c 1 φ
and the specific heat can be computed from c v 1 = d T / d e
c v = d , c v μ = d 1
a result which holds also in Boltzmann Gibbs statistical mechanics (see [19]). The one dimensional Fisher matrix, also called Fisher information is, see (33)
g 00 = c 2 v a r ( 1 φ ) = c 2 1 φ 2 1 φ 2 = c 2 1 φ 2 1 T μ 2
where the probability density (29) is p ( x , e ) = φ c / Z ( e ) and
c 2 1 φ 2 = c 2 Z ( e ) H ( e ) φ c 2 d x = 4 ( d 1 ) ( d 4 ) e 2
thereby obtaining (note that the average value exists for d > 4 )
g 00 ( e ) = d ( d 1 ) ( d 4 ) 1 e 2 = 1 ( d 4 ) 1 T T μ
We see that g 00 ( e ) diverges for e 0 + , that is, when the system energy e tends to the minimum of the potential energy u.
Remark. This quadratic potential energy system is the prototype of the description of a mechanical system in the vicinity of a minimum of the potential energy. By a translation of the reference frame we can suppose that the minimum is in x = 0 and that u ( 0 ) = 0 so that using a Taylor expansion we have
u ( x ) = u ( 0 ) + u ( 0 ) · x + 1 2 u ( 0 ) x · x + O ( 3 )
and
m i n ( λ ) x 2 u ( 0 ) x · x m a x ( λ ) x 2
where m i n ( λ ) and m a x ( λ ) are the smallest and greatest (real) eigenvalues of the symmetric matrix u ( 0 ) . So we can conclude that the Fisher information diverges when the energy e approaches the minimum of the potential energy from above. This can be interpreted by recalling that the Fisher information measures the change in the "shape" of the probability distribution p ( x , e ) with respect to a change in the parameter e; when the energy becomes equal to the minimum of the potential energy, the probability density becomes a Dirac delta concentrated at the minimum x = 0 . We can say that the divergence of the Fisher information g 00 at e = 0 corresponds to the qualitative "phase transition" between a point orbit and a set of d closed orbits for the d oscillators when e > 0 .

5.2. The Elastic Chain

The following example shows that for a n-dimensional mechanical systems with convex potential energy, HB thermodynamics gives a sound description of the system’s behavior (see also [7]).
Let us consider a chain of n + 1 point particles of mass m linked by linearly elastic springs and constrained to move on the real axis. Suppose that particle 0 is fixed in the origin and that the coordinate of the n-th particle is λ n so that the length λ n of the chain is a controlled parameter λ corresponding to the equilibrium length of the n springs. Let the elongation of the spring i = 1 , n be written as u i = λ + x i . The length constraint is
i = 1 n u i = i = 1 n ( λ + x i ) = n λ , t h a t i s i = 1 n x i = 0 .
Due to the length constraint the chain potential energy of the resulting n 1 dimensional system can be written as
u ( x ) = u ( x 1 , , x n ) = h 2 i = 1 n ( x i + λ ) 2 = h 2 ( x 2 + n λ 2 ) , i = 1 n x i = 0
where x 2 = i = 1 n x i 2 . Therefore the region H ( e , λ ) is
H ( e , λ ) = { x R n φ = e u ( x ) = e h 2 ( x 2 + n λ 2 ) > 0 , i = 1 n x i = 0 }
i.e.
H ( e , λ ) = { x R n : x 2 < 2 e h n λ 2 , i = 1 n x i = 0 } = B n ( a ) L = B n 1 ( a )
where B n ( a ) is the ball in R n of radius a,
a 2 ( e , λ ) = 2 e h n λ 2
and L is the n 1 dimensional hyperplane i = 1 n x i = 0 . We compute the entropy S = ln Σ by setting
φ = e u = h 2 ( a 2 x 2 )
and (here c = d 2 1 = n 3 2 )
S = ln Σ = ln H ( e , λ ) φ c + 1 d x = ln B n 1 ( a ) h 2 ( a 2 x 2 ) c + 1 d x .
By standard computation on radial function we get
S ( e , λ ) = ( n 1 ) ln a 2 ( e , λ ) + c o s t ( n ) = ( n 1 ) ln ( e n h 2 λ 2 ) + c o s t ( n )
Let us compute the temperature of this elongated chain using
T = ( S e ) 1 = a 2 n 1 = 1 n 1 ( e n h 2 λ 2 )
which coincides with the kinetic energy e u ( λ ) per degree of freedom. Let us compute the pressure P corresponding to the controlled parameter λ which is the elongation on the elastic chain using ( 20 ) 2 . We have
P = φ λ = n h λ = n h λ
which corresponds to our physical intuition of the reaction force exerted by the chain on its controlled end. The specific heat c v = d e / d T is
1 c v = d T d e = 1 n 1 = 1 d .
Remark. Note that the elastic chain model fails to define a statistical model because conditions 1) and 2) in Definition 1 are not met. Indeed concerning condition 2) we have
l e = 1 φ , l λ = φ , λ φ = h λ φ = h λ l e
so the two score vectors l e and l λ are not independent functions on H ( e , λ ) . Therefore we can not compute the Fisher matrix for this mechanical system. If we consider only the energy parameter e, the system is similar to the previous example 1.

5.3. Two-Body System

In this example we consider an isolated system of two points of equal mass m subject to gravitational forces. If x = ( x 1 , x 2 ) , x i R 3 , i = 1 , 2 denote the positions of the two points, the potential energy is u ( x , λ ) = λ | x 2 x 1 | , where λ = G m 2 can be considered as a parameter. A system of points interacting via gravity requires a specific description in Statistical Mechanics due to the long-range nature of the force which prevents these systems to display the extensive character of the total energy. Also, it is known that gravitating systems exhibit the phenomenon of negative specific heat [24], a possibility which is excluded in the canonical description, see (36). Therefore, the correct statistical ensemble to adopt is the microcanonical one, see [25]. In [25] the above two body system is modified to construct a toy model which exhibits all the features of a many-body self-gravitating system. This is obtained by imposing a short range cutoff on the particles inter-distance | x 2 x 1 | > a , which behaves as hard spheres of radius a / 2 , and a long range cutoff | x 2 x 1 | < R . This modified system display a phase transition between phases of positive and negative specific heat when the system energy is varied. We want to discuss this mechanical system using HB thermodynamics. See also [5] for a different analysis of a two body system with HB thermodynamics. Therefore we compute the probability density (10) and the Fisher information for this system and discuss its ability to describe phase transitions. Using (12) we have for e < 0 ,
H ( e , λ ) = { x R 6 : φ ( x , e , λ ) = e u ( x , λ ) = e + λ | x 2 x 1 | > 0 }
and from (10) with d = 3 n = 6 , c = d 2 1 = 2 ,
Z ( e , λ ) = H ( e , λ ) φ c d x = H ( e , λ ) e + λ | x 2 x 1 | 2 d x .
To compute the last integral it is useful to perform the change of variables χ ( x 1 , x 2 ) = ( x 1 , s + x 1 ) where s = x 2 x 1 is the inter-particle vector. We have det d χ = 1 and H ( e , λ ) = χ ( Λ ) where
Λ = { ( x 1 , s ) R 6 : | s | < λ e } = R x 1 3 × B 3 ( λ e )
therefore, by using the change of variable and Fubini theorem, we can write
Z ( e , λ ) = Λ e + λ | s | 2 det d χ d x 1 d s = R x 1 3 d x B 3 ( λ e ) e + λ | s | 2 d s .
The integral in d x is infinite unless we assume that the particle x is confined in a bounded region of volume V. Therefore we can write Z = V Z s and, since we deal with a function radial in s we have that Z s is finite and
Z s ( e , λ ) = 0 λ e e + λ s 2 s 2 d s = V α 2 λ 3 3 e , e < 0 .
The probability density (10) is thus factorized into the product of two densities
p ( e , λ , x , s ) d x d s = φ c Z ( e , λ ) d x d s = d x V · φ 2 ( e , λ , s ) Z s d s
showing that the positions of the two particles are independent random variables. Therefore we can forget about the x 1 particle and perform all the computations in the s variable.
If we let the energy tend to zero, so that the inter-particle distance s can be arbitrarily large, we make Z in (40), hence the microcanonical entropy of the system S μ = ln Z , infinite. The HB or volume entropy is
S = S μ + ln φ
and it is easy to see that φ is infinite if we let the inter-particle distance s go to zero, so the entropy S is infinite when stend to zero or to infinity, while S μ is infinite only in the latter case.
Therefore, it is necessary to assume, as in [25], that there are two cutoffs a < | s | < R , usually with R a , so that the two body system is confined in a bounded region of space. Moreover the system displays two energy scales
e m i n = u ( a ) = λ a < 0 , e R = u ( r ) = λ R < 0
Also, it is stipulated in [25] that if the energy is in the range e m i n e e R , the integration in the s variable is performed between extrema a and λ e , while for e e R the integration is between fixed extrema a and R. Due to the presence of the cutoff a, the system displays a negative specific heat phase (see again [25]).
We can now compute the temperature T, the microcanonical temperature T μ and the pressure P using (21) for our system in the two energy ranges and then juxtapose the plots that have continuous junction at e = e R . See Appendix 1 for computations. We also compute the Fisher information g 00 ( e , λ ) from (33). See Figure 1 for the plot of T, T μ and g 00 .
The temperature T = T ( e , λ ) (blue curve of Figure 1) presents a gentle maximum for negative energy (point B) and a sharp minimum (point C) for a positive value of the energy. We see that for e = e m i n the temperature is zero as in the previous example 1. The temperature curve of this simple mechanical system displays the same features of more complex and realistic models of gravitating many particle systems: a phase (A to B) of positive specific heat followed by a phase (B to C) of negative specific heat and again a phase of positive specific heat (after C). The microcanonical temperature T μ (orange curve) has a similar behavior of T. The points where the two temperature curves cross each other are the points where the specific heat c v diverges; see (24). See [25] for a physical interpretation of these phases.
We see that the Fisher information g 00 = g 00 ( e , λ ) (red curve of Figure 1) diverges for e = e m i n and has a peak for e = e R . The divergence in e = e m i n is located at the minimum of the potential energy (due to the cutoff at s = a ) and it is similar to the one found in the previous example 1. Here we are concerned with the assessment of the ability of Fisher information to locate the phase transitions at B and C. It seems that the phase transition in B is not detected, and the one in C is not exactly located. This is due to the presence of the cutoffs a < s < R (which are essential for the existence of the negative specific heat phase). As far as the phase transition in B is concerned, the shape of the curve of temperature T does not change with a and in the limit a 0 entropy S becomes infinite. On the other hand, the microcanonical entropy S μ = ln Z is defined for a = 0 and the curve of the microcanonical temperature computed for a = 0 (dashed line of Figure 2) does not show the phase A to B of positive specific heat. Therefore, for a = 0 , the phase transition in B is removed in the microcanonical description of the system, while the remaining one is detected by the divergence of Fisher information.
We will show that if the bounds are removed, letting a 0 and R + the minimum C and the peak in Fisher information g 00 coincide at e = 0 . In fact, the energy corresponding to point C can be determined using the condition T ( e , λ ) / e = 0 . By computing T ( e , λ ) / e = 0 at the leading order in 1 a / R we deduce that the minimum in C in Figure 1 is located at
e 2 * = λ R 6 ln R 3 .
The peak of Fisher information is located at e = e R = λ / R and it height is
g 00 ( λ R , λ ) = 3 R 4 λ 2 ( a R ) 2 3 ( R λ ) 2
In the limit R + both e R and e 2 * tends to 0. So we can say that the divergence of Fisher information g 00 correctly detects the phase transition from negative to positive specific heat located at e = 0 .

6. Conclusions

The formulation of HB thermodynamics defines the analogs of temperature, pressure, and entropy for a mechanical system whose potential energy depends on a parameter λ . This scheme has been generalized to n degree-of-freedom systems with multiple parameters, but generalization to an infinite number of degrees of freedom (and hence the operation of thermodynamic limit) seems to be beyond the possibilities of the theory. It is thus interesting (and it is the aim of this work) to ask if the thermodynamic description of these mechanical systems is capable of displaying different patterns of organization (driven by qualitative changes in the underlying dynamics ) and if these patterns can be signaled by an order parameter. Given the probabilistic nature of the theory, it has been natural to investigate if the HB thermodynamic scheme can be read as a statistical model and to compute the associated Fisher matrix, which has previously been used in the literature to describe systems which are of more general nature than those considered in Statistical Mechanics. We have shown with some paradigmatic examples that there are cases in which the Fisher information computed from HB thermodynamics is capable of locating phase transitions in the generalized sense exposed above.

Funding

This research received no external funding.

Appendix

Concerning Section 5.3, all the computations are performed in the two separated energy ranges:
A) e m i n < e < e R , and B) e e R , where e m i n = λ / a , e R = λ / R , and R a .
In the A) case the integration in the s variable is in the interval ( a , λ / e ) while in the B) case the interval is ( a , R ) . Note that the entropy is, see (9)
S ( e , λ ) = ln Σ = ln H ( e , λ ) φ c + 1 d x + c o s t
and it can be computed in the two energy ranges as:
A ) S ( e , λ ) = ln a λ e ( e + λ s ) 3 s 2 d s = ln [ 1 6 ( a e + λ ) ( 2 a 2 λ 2 + 7 a e λ + 11 λ 2 ) + λ 3 ln ( λ a e ) ]
B ) S ( e , λ ) = ln a R ( e + λ s ) 3 s 2 d s = ln [ 3 e λ 2 ( R a ) + 3 2 e 2 λ ( R 2 a 2 ) + 1 3 ( R 3 a 3 ) + λ 3 ln ( R a ) ]
Note that the entropy S is not defined for a = 0 . Temperature T and pressure P can be computed from (21). Microcanonical entropy is
S μ ( e , λ ) = ln Z = ln H ( e , λ ) φ c d x + c o s t
and it can be computed in the two energy ranges as:
A ) S μ ( e , λ ) = ln Z = ln a λ e ( e + λ s ) 2 s 2 d s = ln [ ( a e + λ ) 3 3 e ]
B ) S μ ( e , λ ) = ln a R ( e + λ s ) 2 s 2 d s = ln [ 1 3 ( R a ) ( a 2 e 2 + 3 λ 2 + 3 e λ R + e 2 R 2 + a e ( 3 λ + e R ) ) ]
Note that the entropy S μ is defined also for a = 0 . See plot of the entropies S and S μ in Figure 3 below. Temperature T μ and pressure P μ can be computed by derivation using (21). See plot of T and T μ in Figure 1 in the main text and plot of P and P μ in Figure 4 below.
For the Fisher information (see (33)) we have
g 00 ( e , λ ) = c 2 1 φ 2 1 φ 2 = c 2 1 φ 2 1 T μ 2 = c 2 Z H ( e , λ ) φ c 2 d x 1 T μ 2
therefore:
A ) g 00 ( e , λ ) = 3 λ 2 e 2 ( a e λ )
B ) g 00 ( e , λ ) = 3 λ 2 ( R a ) 2 ( a 2 e 2 + 3 λ 2 + 3 e λ R + e 2 R 2 + a e ( 3 λ + e R ) ) 2
Note that g 00 is defined for a = 0 . See plot of g 00 in Figure 1 in main text.
Figure 3. Plot of entropy S (blue curve), microcanonical entropy S μ (orange curve) and microcanonical entropy S μ in the a = 0 case (dashed curve) as a function of the energy e. The non convexity of the entropy function in the a > 0 case (orange and blue solid curves) is a signature of a phase transition. See corresponding plot of temperatures in Figure 1, and in Figure 2 for the case a = 0 .
Figure 3. Plot of entropy S (blue curve), microcanonical entropy S μ (orange curve) and microcanonical entropy S μ in the a = 0 case (dashed curve) as a function of the energy e. The non convexity of the entropy function in the a > 0 case (orange and blue solid curves) is a signature of a phase transition. See corresponding plot of temperatures in Figure 1, and in Figure 2 for the case a = 0 .
Preprints 159189 g003
Figure 4. Plot of pressure P (blue curve) and microcanonical pressure P μ (orange curve) as a function of the energy e. Inlet: detail of the pressure plot for positive values of energy
Figure 4. Plot of pressure P (blue curve) and microcanonical pressure P μ (orange curve) as a function of the energy e. Inlet: detail of the pressure plot for positive values of energy
Preprints 159189 g004

Acknowledgments

Conflicts of Interest

The author declares no conflicts of interest

References

  1. H. Helmholtz (1884) Principien der Statik monocyklischer Systeme, Crelle’s Journal 97 S. 111–140, reprinted in “Wissenschaftliche Abhandlungen", vol. III, p.142-162 and p.179-202, Leipzig, 1895.
  2. H. Helmholtz (1884) Studien zur Statik monocyklischer Systeme, Akademie der Wissenschaften zu Berlin, S. 159–177 , reprinted in “Wissenschaftliche Abhandlungen", vol. III, p.163-172 and p. 173-178, Leipzig, 1895.
  3. L. Boltzmann (1884) Über die Eigenschaften monozyklischer und anderer damit verwandter Systeme, Crelle’s Journal 98, S.68–94, reprinted in F. Hasenöhrl Wissenschaftliche Abhandlungen von Ludwig Boltzmann Band III, p.122 (1909), reprinted by Chelsea Publ. Company New York, N.Y. (1968).
  4. Gibbs, JW. Elementary principles in statistical mechanics: developed with especial reference to the rational foundations of thermodynamics. C. Scribner’s sons; 1902.
  5. G. Gallavotti (1999) Statistical Mechanics. A short treatise Texts and Monographs in Physics. Springer-Verlag, Berlin.
  6. Porporato, A., Rondoni, L. "Deterministic engines extending Helmholtz thermodynamics." Physica A: Statistical Mechanics and its Applications 640 (2024): 129700.
  7. Cardin F., Favretti M. On the Helmholtz-Boltzmann thermodynamics of mechanical systems. Continuum mechanics and thermodynamics. 2004 Feb;16(1):15-29. [CrossRef]
  8. Campisi, M. (2005). On the mechanical foundations of thermodynamics: The generalized Helmholtz theorem. Studies in History and Philosophy of Science Part B: Studies in History and Philosophy of Modern Physics, 36(2), 275-290.
  9. Fisher, R. A. On the mathematical foundations of theoretical statistics. Philosophical Transactions of the Royal Society of London. Series A, Containing Papers of a Mathematical or Physical Character 1922.
  10. Rao, C. R. Information and the accuracy attainable in the estimation of statistical parameters. In Breakthroughs in statistics Springer, New York, NY, 1992; pp. 235-247.
  11. Amari, S. Nagaoka, H. Methods of Information Geometry AMS and Oxford University Press, 2000.
  12. Amari, S. Information geometry and its applications Vol. 194. Springer, 2016.
  13. Calin, O., Udriste, C. Geometric modeling in probability and statistics Springer, Berlin, 2014.
  14. Crooks, G. E. Fisher information and statistical mechanics. Tech. Rep. 2011.
  15. Arnold J, Lörch N, Holtorf F, Schäfer F. Machine learning phase transitions: Connections to the Fisher information. arXiv preprint arXiv:2311.10710. 2023 Nov 17.
  16. Brunel, N., Nadal,JP. "Mutual information, Fisher information, and population coding." Neural computation 10, no. 7 (1998): 1731-1757.
  17. Petz, D. Quantum information theory and quantum statistics, Springer, Berlin, Heidelberg, 2008.
  18. Prokopenko, M., et al. Relating Fisher information to order parameters. Physical Review E 2011 84.4 041116. [CrossRef]
  19. Huang, K. (2008). Statistical mechanics. John Wiley & Sons.
  20. Favretti, M. Geometry and control of thermodynamic systems described by generalized exponential families. Journal of Geometry and Physics 176 (2022): 104497. [CrossRef]
  21. Favretti, M. (2022). Exponential Families with External Parameters. Entropy, 24(5), 698. [CrossRef]
  22. Dauxois, T., Ruffo,S., Arimondo,E. and Wilkens,M. Dynamics and thermodynamics of systems with long-range interactions: An introduction. Springer Berlin Heidelberg, 2002.
  23. Campa, A., Dauxois, T. and Ruffo, S., 2009. Statistical mechanics and dynamics of solvable models with long-range interactions. Physics Reports, 480(3-6), pp.57-159. [CrossRef]
  24. Thirring, W. (1970) Z. Phys. 235 p. 339.
  25. Padmanabhan, T., 1990. Statistical mechanics of gravitating systems. Physics Reports, 188(5), pp.285-362. [CrossRef]
Figure 1. Plot of Fisher information g 00 (red curve), Temperature T (blue curve) and microcanonical temperature T μ (orange curve) as a function of the energy e.
Figure 1. Plot of Fisher information g 00 (red curve), Temperature T (blue curve) and microcanonical temperature T μ (orange curve) as a function of the energy e.
Preprints 159189 g001
Figure 2. Plot of Fisher information g 00 (red curve), microcanonical temperature T μ (orange curve) and microcanonical temperature for a = 0 (dashed curve) as a function of the energy e. In the a = 0 case the phase of negative specific heat is not present and the remaining phase transition is located at e = 0 . Note that for greater clarity the plot is not in the whole energy range e > λ / a .
Figure 2. Plot of Fisher information g 00 (red curve), microcanonical temperature T μ (orange curve) and microcanonical temperature for a = 0 (dashed curve) as a function of the energy e. In the a = 0 case the phase of negative specific heat is not present and the remaining phase transition is located at e = 0 . Note that for greater clarity the plot is not in the whole energy range e > λ / a .
Preprints 159189 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated