### 7 Entropy as a measure of disorder ?

How then to determine the unknown function [and ] ? According to the last section, all the required information on may be obtained from a knowledge of the term in the differential equation (26) for . We shall try to solve this problem my means of the following two-step strategy: (i) Find an additional physical condition for the fundamental probability density , (ii) determine the shape of [as well as that of and ] from this condition.

At this point it may be useful to recall the way probability densities are determined in classical statistical physics. After all, the present class of theories is certainly not of a deterministic nature and belongs fundamentally to the same class of statistical (i.e. incomplete with regard to the description of single events) theories as classical statistical physics; no matter how important the remaining differences may be.

The physical condition for which determines the behavior of ensembles in classical statistical physics is the principle of maximal (Boltzmann) entropy. It agrees essentially with the information-theoretic measure of disorder introduced by Shannon . Using this principle both the micro-canonical and the canonical distribution of statistical thermodynamics may be derived under appropriate constraints. Let us discuss this classical extremal principle in some detail in order to see if it can be applied, after appropriate modifications, to the present problem. This question also entails a comparison of different types of statistical theories.

The Boltzmann-Shannon entropy is defined as a functional of an arbitrary probability density . The statistical properties characterizing disorder, which may be used to define this functional, are discussed in many publications . Only one of these conditions will, for later use, be written down here, namely the so-called “composition law”: Let us assume that may be written in the form where depends only on points in a subspace of our dimensional sample space and let us further assume that is the direct product of and . Thus, this system consists of two independent subsystems. Then, the composition law is given by (41)

where operates only on .

For a countable sample space with events labeled by indices from an index set and probabilities , the entropy is given by (42)

where is a constant. To obtain meaningful results the extrema of (42) under appropriate constraints, or subsidiary conditions, must be found. The simplest constraint is the normalization condition . In this case the extrema of the function (43)

with respect to the variables must be calculated. One obtains the reasonable result that the minimal value of is (one of the equal to , all other equal to ) and the maximal value is (all equal, ).

For most problems of physical interest the sample space is non-denumerable. A straightforward generalization of Eq. (42) is given by (44)

where the symbol denotes now a point in the appropriate (generally dimensional) sample space. There are some problems inherent in the this straightforward transition to a continuous set of events which will be mentioned briefly in the next section. Let us put aside this problems for the moment and ask if (44) makes sense from a physical point of view. For non-denumerable problems the principle of maximal disorder leads to a variational problem and the method of Lagrange multipliers may still be used to combine the requirement of maximal entropy with other defining properties (constraints). An important constraint is the property of constant temperature which leads to the condition that the expectation value of the possible energy values is given by a fixed number , (45)

If, in addition, normalizability is implemented as a defining property, then the true distribution should be an extremum of the functional (46)

It is easy to see that the well-known canonical distribution of statistical physics is indeed an extremum of . Can we use a properly adapted version of this powerful principle of maximal disorder (entropy) to solve our present problem ?

Let us compare the class of theories derived in section 5 with classical theories like (46). This may be of interest also in view of a possible identification of ’typical quantum mechanical properties’ of statistical theories. We introduce for clarity some notation, based on properties of the sample space. Classical statistical physics theories like (46) will be referred to as ”phase space theories”. The class of statistical theories, derived in section 5, will be referred to as ”configuration space theories”.

The most fundamental difference between phase space theories and configuration space theories concerns the physical meaning of the coordinates. The coordinates of phase space theories are (generally time-dependent) labels for particle properties. In contrast, configuration space theories are field theories; individual particles do not exist and the (in our case one-dimensional) coordinates are points in space.

A second fundamental difference concerns the dimension of the sample space. Elementary events in phase space theories are points in phase space (of dimension for a particle system) including configuration-space and momentum-space (particle) coordinates while the elementary events of configuration space theories are (space) points in configuration space (which would be of dimension for a particle system in three spatial dimensions). This fundamental difference is a consequence of a (generally nonlocal) dependence between momentum coordinates and space-time points contained in the postulates of the present theory, in particular in the postulated form of the probability current [see (7)]. This assumption, a probability current, which takes the form of a gradient of a function (multiplied by ) is a key feature distinguishing configuration space theories, as potential quantum-like theories, from the familiar (many body) phase space theories. The existence of this dependence per se is not an exclusive feature of quantum mechanics, it is a property of all theories belonging to the configuration class, including the theory characterized by , which will be referred to as ”classical limit theory”. What distinguishes the classical limit theory from quantum mechanics is the particular form of this dependence; for the former it is given by a conventional functional relationship (as discussed in section 4) for the latter it is given by a nonlocal relationship whose form is still to be determined.

This dependence is responsible for the fact that no ”global” condition [like (45) for the canonical distribution] must be introduced for the present theory in order to guarantee conservation of energy in the mean - this conservation law can be guaranteed ”locally” for arbitrary theories of the configuration class by adjusting the relation between (the form of the dynamic equation) and (the definition of expectation values). In phase space theories the form of the dynamical equations is fixed (given by the deterministic equations of classical mechanics). Under constraints like (45) the above principle of maximal disorder creates - basically by selecting appropriate initial conditions - those systems which belong to a particular energy; for non-stationary conditions the deterministic differential equations of classical mechanics guarantee then that energy conservation holds for all times. In contrast, in configuration space theories there are no initial conditions (for particles). The conditions which are at our disposal are the mathematical form of the expectation values (the function ) and/or the mathematical form of the differential equation (the function ). Thus, if something like the principle of maximal disorder can be used in the present theory it will determine the form of the differential equation for rather than the explicit form of .

These considerations raise some doubt as to the usefulness of an measure of disorder like the entropy (44) - which depends essentially on instead of and does not contain derivatives of - for the present problem. We may still look for an information theoretic extremal principle of the general form (47)

Here, the functional attains its maximal value for the function which describes - under given constraints - the maximal disorder. But will differ from the entropy functional and appropriate constraints , reflecting the local character of the present problem, have still to be found. Both terms in (47) are at our disposal and will be defined in the next sections.