Essential Graduate Physics by Konstantin K. Likharev - HTML preview

PLEASE NOTE: This is an HTML preview only and some elements such as links or page numbers may be incorrect.
Download the book in PDF, ePub, Kindle for a complete version.

© K. Likharev

Image 2418

Image 2419

Image 2420

Image 2421

Image 2422

Image 2423

Image 2424

Image 2425

Image 2426

Image 2427

Image 2428

Image 2429

Image 2430

Image 2431

Image 2432

Image 2433

Image 2434

Image 2435

Image 2436

Image 2437

Essential Graduate Physics

SM: Statistical Mechanics

Now let us have a fast look at the typical macroscopic variables the statistical physics and thermodynamics should operate with. Since I have already mentioned pressure P and volume V, let me start with this famous pair of variables. First of all, note that volume is an extensive variable, i.e. a variable whose value for a system consisting of several non-interacting parts is the sum of those of its parts. On the other hand, pressure is an example of an intensive variable whose value is the same for different parts of a system – if they are in equilibrium. To understand why P and V form a natural pair of variables, let us consider the classical playground of thermodynamics, a portion of a gas contained in a cylinder, closed with a movable piston of area A (Fig. 1).

F

A

P, V

Fig. 1.1. Compressing gas.

x

Neglecting the friction between the walls and the piston, and assuming that it is being moved so slowly that the pressure P is virtually the same for all parts of the volume at any instant, the elementary work of the external force F = PA, compressing the gas, at a small piston displacement dx = –dV/ A, is

F 

d

Work

W  F dx  

 Adx   PdV .

(1.1)

A

on a gas

Of course, the last expression is more general than the model shown in Fig. 1, and does not depend on the particular shape of the system’s surface.5 (Note that in the notation of Eq. (1), which will be used through the course, the elementary work done by the gas on the external system equals – d W.) From the point of analytical mechanics,6 V and (– P) is just one of many possible canonical pairs of generalized coordinates qj and generalized forces F j, whose products d W j = F jdqj give independent contributions to the total work of the environment on the system under analysis. For example, the reader familiar with the basics of electrostatics knows that if the spatial distribution E(r) of an external electric field does not depend on the electric polarization P(r) of a dielectric medium placed into the field, its elementary work on the medium is

3

d W  E r d P rd 3 r

3

E r P r

.

(1.2a)

j   d

j  



d r

j1

The most important cases when this condition is fulfilled (and hence Eq. (2a) is valid) are, first, long cylindrical samples in a parallel external field (see, e.g., EM Fig. 3.13) and, second, the polarization of a sample (of any shape) due to that of discrete electric dipoles p k, whose electric interaction is negligible.

In the latter case, Eq. (2a) may be also rewritten as the sum over the single dipoles, located at points r k: 7

5 In order to prove that, it is sufficient to integrate the scalar product d W = d F  dr, with d F = –n Pd 2 r, where dr is the surface displacement vector (see, e.g., CM Sec. 7.1), and n is the outer normal, over the surface.

6 See, e.g., CM Chapters 2 and 10.

7 Some of my students needed an effort to reconcile the positive signs in Eqs. (2) with the negative sign in the well-known relation dUk = –E(r k) d p k for the potential energy of a dipole in an external electric field – see, e.g., Chapter 1

Page 2 of 24

Essential Graduate Physics

SM: Statistical Mechanics

.

d W   d W ,

d

with W  E r d .

(1.2b)

k

k

k k p

k

Very similarly, and at the similar conditions on the external magnetic field H(r), its elementary work on a magnetic medium may be also represented in either of two forms:8

3

d W   H r

M r

H r M r

,

(1.3a)

0

  d   d 3 r

3

0

j   d

j  



d r

j1

d W   d W ,

d

with W  

.

(1.3b)

0H r

d

k

k

k  m k

k

where M and m are the vectors of, respectively, the medium’s magnetization and the magnetic moment of a single dipole. Formulas (2) and (3) show that the roles of generalized coordinates may be played by Cartesian components of the vectors P (or p) and M (or m), with the components of the electric and magnetic fields playing the roles of the corresponding generalized forces. This list may be extended to other interactions (such as gravitation, surface tension in fluids, etc.). Following tradition, I will use the

{– P, V } pair in almost all the formulas below, but the reader should remember that they all are valid for any other pair {F j, qj}.9

Again, the specific relations between the variables of each pair listed above may depend on the statistical properties of the system under analysis, but their definitions are not based on statistics. The situation is very different for a very specific pair of variables, temperature T and entropy S, although these “sister variables” participate in many formulas of thermodynamics exactly as if they were just one more canonical pair {F j, qj}. However, the very existence of these two notions is due to statistics.

Namely, temperature T is an intensive variable that characterizes the degree of thermal “agitation” of the system’s components. On the contrary, the entropy S is an extensive variable that in most cases evades immediate perception by human senses; it is a qualitative measure of the disorder of the system, i.e. the degree of our ignorance about its exact microscopic state.10

The reason for the appearance of the { T, S} pair of variables in formulas of thermodynamics and statistical mechanics is that the statistical approach to large systems of particles brings some qualitatively new results, most notably the notion of the irreversible time evolution of collective ( macroscopic) variables describing the system. On one hand, the irreversibility looks absolutely natural in such phenomena as the diffusion of an ink drop in a glass of water. In the beginning, the ink molecules are located in a certain small part of the system’s volume, i.e. to some extent ordered, while at the late stages of diffusion, the position of each molecule in the glass is essentially random. However, as a second thought, the irreversibility is rather surprising, taking into account that the laws governing the EM Eqs. (3.15). The resolution of this paradox is simple: each term of Eq. (2b) describes the work d W k of the electric field on the internal degrees of freedom of the k th dipole, changing its internal energy Ek: dEk = d W k. This energy change may be viewed as coming from the dipole’s potential energy in the field: dEk = – dUk.

8 Here, as in all my series, I am using the SI units; for their translation to the Gaussian units, I have to refer the reader to the EM part of the series.

9 Note that in systems of discrete particles, most generalized forces, including the fields E and H, differ from the pressure P in the sense that their work may be explicitly partitioned into single-particle components – see Eqs.

(2b) and (3b). This fact gives some discretion for the calculations based on thermodynamic potentials – see Sec.4.

10 The notion of entropy was introduced into thermodynamics in 1865 by Rudolf Julius Emanuel Clausius on a purely phenomenological basis. In the absence of a clue about the entropy’s microscopic origin (which had to wait for the works by L. Boltzmann and J. Maxwell), this was an amazing intellectual achievement.

Chapter 1

Page 3 of 24

Essential Graduate Physics

SM: Statistical Mechanics

motion of the system’s components are time-reversible – such as the Newton laws or the basic laws of quantum mechanics.11 Indeed, if at a late stage of the diffusion process, we reversed the velocities of all molecules exactly and simultaneously, the ink molecules would again gather (for a moment) into the original spot.12 The problem is that getting the information necessary for the exact velocity reversal is not practicable. This example shows a deep connection between statistical mechanics and information theory.

A qualitative discussion of the reversibility-irreversibility dilemma requires a strict definition of the basic notion of statistical mechanics (and indeed of the probability theory), the statistical ensemble, and I would like to postpone it until the beginning of Chapter 2. In particular, in that chapter, we will see that the basic law of irreversible behavior is an increase of the entropy S in any closed system. Thus, the statistical mechanics, without defying the “microscopic” laws governing the evolution of system’s components, introduces on top of them some new “macroscopic” laws, intrinsically related to the evolution of information, i.e. the degree of our knowledge of the microscopic state of the system.

To conclude this brief discussion of variables, let me mention that as in all fields of physics, a very special role in statistical mechanics is played by the energy E. To emphasize the commitment to disregard the motion of the system as a whole in this subfield of physics, the E considered in thermodynamics it is frequently called the internal energy, though just for brevity, I will skip this adjective in most cases. The simplest example of such E is the sum of kinetic energies of molecules in a dilute gas at their thermal motion, but in general, the internal energy also includes not only the individual energies of the system’s components but also their interactions with each other. Besides a few

“pathological” cases of very-long-range interactions, these interactions may be treated as local; in this case the internal energy is proportional to N, i.e. is an extensive variable. As will be shown below, other extensive variables with the dimension of energy are often very useful as well, including the (Helmholtz) free energy F, the Gibbs energy G, the enthalpy H, and the grand potential . (The collective name for such variables is thermodynamic potentials.)

Now, we are ready for a brief discussion of the relationship between statistical physics and thermodynamics. While the task of statistical physics is to calculate the macroscopic variables discussed above13 for various microscopic models of the system, the main role of thermodynamics is to derive some general relations between the average values of the macroscopic variables (also called thermodynamic variables) that do not depend on specific models. Surprisingly, it is possible to accomplish such a feat using just a few either evident or very plausible general assumptions (sometimes called the laws of thermodynamics), which find their proof in statistical physics.14 Such general relations allow for a substantial reduction of the number of calculations we have to do in statistical physics: in most cases, it is sufficient to calculate from the statistics just one or two variables, and then use general 11 Because of that, the possibility of the irreversible macroscopic behavior of microscopically reversible systems was questioned by some serious scientists as recently as in the late 19th century – notably by J. Loschmidt in 1876.

12 While quantum-mechanical effects, with their intrinsic uncertainty, may be quantitatively important in this example, our qualitative discussion does not depend on them. Another classical example is the chaotic motion of a ball on a 2D Sinai billiard – see CM Chapter 9 and in particular Fig. 9.8 and its discussion.

13 Several other quantities, for example the heat capacity C, may be calculated as partial derivatives of the basic variables discussed above. Also, at certain conditions, the number of particles N in a certain system may be not fixed and also considered as an (extensive) variable – see Sec. 5 below.

14 Admittedly, some of these proofs are based on other plausible but deeper postulates, for example the central statistical hypothesis (Sec. 2.2), whose best proof, to my knowledge, is just the whole body of experimental data.

Chapter 1

Page 4 of 24

Image 2438

Image 2439

Image 2440

Image 2441

Image 2442

Image 2443

Image 2444

Image 2445

Image 2446

Image 2447

Image 2448

Image 2449

Essential Graduate Physics

SM: Statistical Mechanics

thermodynamic relations to get all other properties of interest. Thus the thermodynamics, sometimes snubbed as a phenomenology, deserves every respect not only as a useful theoretical tool but also as a discipline more general than any particular statistical model. This is why the balance of this chapter is devoted to a brief review of thermodynamics.

1.2. The 2nd law of thermodynamics, entropy, and temperature

Thermodynamics accepts a phenomenological approach to the entropy S, postulating that there is such a unique extensive measure of the aggregate disorder, and that in a closed system (defined as a system completely isolated from its environment, i.e. the system with its internal energy fixed) it may only grow in time, reaching its constant (maximum) value at equilibrium:15

2nd law of

thermo-

dS  0 .

(1.4)

dynamics

This postulate is called the 2nd law of thermodynamics – arguably its only substantial new law.16

Rather surprisingly, this law, together with the additivity of S in composite systems of non-interacting parts (as an extensive variable), is sufficient for a formal definition of temperature, and a derivation of its basic properties that comply with our everyday notion of this key variable. Indeed, let us consider a closed system consisting of two fixed-volume subsystems (Fig. 2) whose internal relaxation is very fast in comparison with the rate of the thermal flow (i.e. the energy and entropy exchange) between the parts. In this case, on the latter time scale, each part is always in some quasi-equilibrium state, which may be described by a unique relation E( S) between its energy and entropy.17

E , S

1

1

dE, dS

E , S

2

2

Fig. 1.2. A. composite thermodynamic system.

Neglecting the energy of interaction between the parts (which is always possible at N >> 1, and in the absence of long-range interactions), we may use the extensive character of the variables E and S

to write

E E S E S

S S S

(1.5)

1  1 

2  2 ,

,

1

2

for the full energy and entropy of the system. Now let us use them to calculate the following derivative: 15 Implicitly, this statement also postulates the existence, in a closed system, of thermodynamic equilibrium, an asymptotically reached state in which all macroscopic variables, including entropy, remain constant. Sometimes this postulate is called the 0th law of thermodynamics.

16 Two initial formulations of this law, later proved equivalent, were put forward independently by Lord Kelvin (born William Thomson) in 1851 and by Rudolf Clausius in 1854.

17 Here we strongly depend on a very important (and possibly the least intuitive) aspect of the 2nd law, namely that the entropy is a unique measure of disorder.

Chapter 1

Page 5 of 24

Essential Graduate Physics

SM: Statistical Mechanics

dS

dS

dS

dS

dS dE

dS

dS d( E E )

1

2

1

2

2

1

2

1

.

(1.6)

dE

dE

dE

dE

dE dE

dE

dE

dE

1

1

1

1

2

1

1

2

1

Since the total energy E of the closed system is fixed and hence independent of its re-distribution between the subsystems, we have to take dE/ dE 1 =0, and Eq. (6) yields

dS

dS

dS

1

2

.

(1.7)

dE

dE

dE

1

1

2

According to the 2nd law of thermodynamics, when the two parts have reached the thermodynamic equilibrium, the total entropy S reaches its maximum, so that dS/ dE 1 = 0, and Eq. (7) yields dS

dS

1

2

.

(1.8)

dE

dE

1

2

This equality shows that if a thermodynamic system may be partitioned into weakly interacting macroscopic parts, their derivatives dS/ dE should be equal in the equilibrium. The reciprocal of this derivative is called temperature. Taking into account that our analysis pertains to the situation (Fig. 2) when both volumes V 1,2 are fixed, we may write this definition as

E

 

  T ,

(1.9)

Definition of

S

 

temperature

V

the subscript V meaning that volume is kept constant at the differentiation. (Such notation is common and very useful in thermodynamics, with its broad range of variables.)

Note that according to Eq. (9), if the temperature is measured in energy units18 (as I will do in this course for the brevity of notation), then S is dimensionless. The transfer to the SI or Gaussian units, i.e. to the temperature T K measured in kelvins (not “Kelvins”, and not “degrees Kelvin”, please!), is given by the relation T = k B T K, where the Boltzmann constant k B  1.38×10-23 J/K = 1.38×10-16 erg/K.19

In those units, the entropy becomes dimensional: S K = k B S.

The definition of temperature, given by Eq. (9), is of course in sharp contrast with the popular notion of T as a measure of the average energy of one particle. However, as we will repeatedly see below, in many cases these two notions may be reconciled, with Eq. (9) being more general. In particular, the so-defined T is in semi-quantitative agreement with our everyday notion of temperature:20

(i) according to Eq. (9), the temperature is an intensive variable (since both E and S are extensive), i.e., in a system of similar particles, it is independent of the particle number N; 18 Here I have to mention a traditional unit of thermal energy, the calorie, still being used in some applied fields.

In the most common modern definition (as the so-called thermochemical calorie) it equals exactly 4.148 J.

19 For the more exact values of this and other constants, see appendix CA: Selected Physical Constants. Note that both T and T K define the natural absolute (also called “thermodynamic”) scale of temperature, vanishing at the same point – in contrast to such artificial scales as the degrees Celsius (“centigrades”), defined as T C  T K +

273.15, or the degrees Fahrenheit: T F  (9/5) T C + 32.

20 Historically, such notion was initially qualitative – just as something distinguishing “hot” from “cold”. After the invention of thermometers (the first one by Galileo Galilei in 1592), mostly based on thermal expansion of fluids, this notion had become quantitative but not very deep: being understood as something “what the thermometer measures” – until its physical sense as a measure of thermal motion’s intensity, was revealed in the 19th century.

Chapter 1

Page 6 of 24

Essential Graduate Physics

SM: Statistical Mechanics

(ii) temperatures of all parts of a system are equal at equilibrium – see Eq. (8);

(iii) in a closed system whose parts are not in equilibrium, thermal energy ( heat) always flows from a warmer part (with higher T) to the colder part.

In order to prove the last property, let us revisit the closed, composite system shown in Fig. 2, and consider another derivative:

dS

dS

dS

dS dE

dS dE

1

2

1

1

2

2

.

(1.10)

dt

dt

dt

dE dt

dE

dt

1

2

If the internal state of each part is very close to equilibrium (as was assumed from the very beginning) at each moment of time, we can use Eq. (9) to replace the derivatives dS 1,2/ dE 1,2 with 1/ T 1,2, getting dS

1 dE

1 dE

1

2

.

(1.11)

dt

T dt

T dt

1

2

Since in a closed system E = E 1 + E 2 = const, these time derivatives are related as dE 2/ dt = – dE 1/ dt, and Eq. (11) yields

dS

 1

1  dE

1

.

(1.12)

dt

 T T  dt

 1

2 

But according to the 2nd law of thermodynamics, this derivative cannot be negative: dS/ dt ≥ 0. Hence,

 1

1  dE

1

 .

0





(1.13)

T

T

dt

1

2 

For example, if T 1 > T 2, then dE 1/ dt  0, i.e. the warmer part gives energy to its colder counterpart.

Note also that at such a heat exchange, at fixed volumes V 1,2, and T 1  T 2, increases the total system’s entropy, without performing any “useful” mechanical work – see Eq. (1).

1.3. The 1st and 3rd laws of thermodynamics, and heat capacity

Now let us consider a thermally insulated system whose volume V may be changed by force –

see, for example, Fig. 1. Such a system is different from the fully closed one, because its energy E may be changed by the external force’s work – see Eq. (1):

dE d W   PdV .

(1.14)

Let the volume change be so slow ( dV/ dt → 0) that the system is virtually at equilibrium at any instant.

Such a slow process is called reversible, and in the particular case of a thermally insulated system, it is also called adiabatic. If the pressure P (or any generalized external force F j) is deterministic, i.e. is a predetermined function of time, independent of the state of the system under analysis, it may be considered as coming from a fully ordered system, i.e. the one having zero entropy, with the total system (the system under our analysis plus the source of the force) completely closed. Since the entropy of the total closed system should stay constant (see the second of Eqs. (5) above), S of the system under analysis should stay constant on its own. Thus we arrive at a very important conclusion: at an adiabatic process, the entropy of a system cannot change. (Sometimes such a process is called isentropic.) This means that we may use Eq. (14) to write

Chapter 1

Page 7 of 24

Image 2450

Image 2451

Image 2452

Image 2453

Image 2454

Image 2455

Essential Graduate Physics

SM: Statistical Mechanics

E

 

P  

 .

(1.15)

V

  S

Now let us consider a more general thermodynamic system that may also exchange thermal energy (“heat”) with its environment (Fig. 3).

dQ

d W

Fig. 1.3. An example of the thermodynamic

E( S, V)

process involving both the mechanical work by

the environment, and the heat exchange with it.

For such a system, our previous conclusion about the entropy’s constancy is not valid, so that S, in equilibrium, may be a function of not only the system’s energy E, but also of its volume: S = S( E, V).

Let us consider this relation resolved for energy: E = E( S, V), and write the general mathematical expression for the full differential of E as a function of these two independent arguments:

E

 

E

 

dE  

dS  

dV.

(1.16)

S

 

V

V

S

This formula, based on the stationary relation E = E( S, V), is evidently valid not only in equilibrium but also for all very slow, reversible21 processes. Now, using Eqs. (9) and (15), we may rewrite Eq. (16) as dE TdS PdV .

(1.17) Energy:

differential

According to Eq. (1), the second term on the right-hand side of this equation is just the work of the external force, so that due to the conservation of energy,22 the first term has to be equal to the heat dQ

transferred from the environment to the system (see Fig. 3):

dE dQ  W

d

,

(1.18) 1st law of

thermo-

dQ TdS .

(1.19) dynamics

The last relation, divided by T and then integrated along an arbitrary (but reversible!) process, dQ

S

 const,

(1.20)

T

is sometimes used as an alternative definition of entropy S – provided that temperature is defined not by Eq. (9), but in some independent way. It is useful to recognize that entropy (like energy) may be defined 21 Let me emphasize again that any adiabatic process is reversible, but not vice versa.

22 Such conservation, expressed by Eqs. (18)-(19), is commonly called the 1st law of thermodynamics. While it (in contrast with the 2nd law) does not present any new law of nature, and in particular was already used de-facto to write the first of Eqs. (5) and also Eq. (14), such a grand name was absolutely justified in the 19th century when the mechanical nature of the internal energy (the thermal motion) was not at all clear. In this context, the names of three scientists, Benjamin Thompson (who gave, in 1799, convincing arguments that heat cannot be anything but a form of particle motion), Julius Robert von Mayer (who conjectured the conservation of the sum of the thermal and macroscopic mechanical energies in 1841), and James Prescott Joule (who proved this conservation experimentally two years later), have to be reverently mentioned.

Chapter 1

Page 8 of 24

Image 2456

Image 2457

Image 2458

Image 2459

Image 2460

Image 2461

Image 2462

Image 2463

Image 2464

Essential Graduate Physics

SM: Statistical Mechanics

to an arbitrary constant, which does not affect any other thermodynamic observables. The common convention is to take

S  ,

0

at T  0 .

(1.21)

This condition is sometimes called the “3rd law of thermodynamics”, but it is important to realize that this is just a convention rather than a real law.23 Indeed, the convention corresponds well to the notion of the full order at T = 0 in some systems (e.g., separate atoms or perfect crystals), but creates ambiguity for other systems, e.g., amorphous solids (like the usual glasses) that may remain highly disordered for

“astronomic” times, even at T  0.

Now let us discuss the notion of heat capacity that, by definition, is the ratio dQ/ dT, where dQ is the amount of heat that should be given to a system to raise its temperature by a small amount dT. 24

(This notion is important because the heat capacity may be most readily measured experimentally.) The heat capacity depends, naturally, on whether the heat dQ goes only into an increase of the internal energy dE of the system (as it does if its volume V is constant), or also into the mechanical work (– d W)

performed by the system at its expansion – as it happens, for example, if the pressure P, rather than the volume V, is fixed (the so-called isobaric process – see Fig. 4).

A

g

M

dQ

Mg

P

const

Fig. 1.4. The simplest example of

A

the isobaric process.

Hence we should discuss at least two different quantities, 25 the heat capacity at fixed volume,

Q

 

C  

(1.22)

V

T

  V

Heat

capacity: and the heat capacity at fixed pressure

definitions

Q

 

C  

 ,

(1.23)

P

T

  P

23Actually, the 3rd law (also called the Nernst theorem) as postulated by Walter Hermann Nernst in 1912 was different – and really meaningful: “It is impossible for any procedure to lead to the isotherm T = 0 in a finite number of steps.” I will discuss this theorem at the end of Sec. 6.

24 By this definition, the full heat capacity of a system is an extensive variable, but it may be used to form such intensive variables as the heat capacity per particle, called the specific heat capacity, or just the specific heat.

(Please note that the last terms are rather ambiguous: they are used for the heat capacity per unit mass, per unit volume, and sometimes even for the heat capacity of the system as the whole, so that some caution is in order.) 25 Dividing both sides of Eq. (19) by dT, we get the general relation dQ/ dT = TdS/ dT, which may be used to rewrite the definitions (22) and (23) in the following forms:

S

 

S

 

C T

 ,

C T

 ,

V

P

T

 

T

V

P

more convenient for some applications.

Chapter 1

Page 9 of 24

Essential Graduate Physics

SM: Statistical Mechanics

and expect that for all “normal” (mechanically stable) systems, CPCV. The difference between CP and CV is rather minor for most liquids and solids, but may be very substantial for gases – see Sec. 4.

1.4. Thermodynamic potentials

Since for a fixed volume, d W = – PdV = 0, and Eq. (18) yields dQ = dE, we may rewrite Eq.

(22) in another convenient form

E

 

C  

 .

(1.24)

V

T

  V

so that to calculate CV from a certain statistical-physics model, we only need to calculate E as a function of temperature and volume. If we want to obtain a similarly convenient expression for CP, the best way is to introduce a new notion of so-called thermodynamic potentials – whose introduction and effective use is perhaps one of the most impressive techniques of thermodynamics. For that, let us combine Eqs.

(1) and (18) to write the 1st law of thermodynamics in its most common form

dQ dE PdV.

(1.25)

At an isobaric process (Fig. 4), i.e. at P = const, this expression is reduced to

dQ  dE d( PV )  d( E PV ) .

(1.26)

P

P

P

P

Thus, if we introduce a new function with the dimensionality of energy:26

H E PV ,

(1.27) Enthalpy:

definition

called enthalpy (or, sometimes, the “heat function” or the “heat contents”),27 we may rewrite Eq. (23) as

H

 

C  

 .

(1.28)

P

T

  P

Comparing Eqs. (28) and (24) we see that for the heat capacity, the enthalpy H plays the same role at fixed pressure as the internal energy E plays at fixed volume.

Now let us explore properties of the enthalpy at an arbitrary reversible process, i.e. lifting the restriction P = const, but keeping the definition (27). Differentiating this equality, we get dH dE PdV VdP .

(1.29)

Plugging into this relation Eq. (17) for dE, we see that the terms  PdV cancel, yielding a very simple expression

dH TdS VdP ,

(1.30) Enthalpy:

differential

whose right-hand side differs from Eq. (17) only by the swap of P and V in the second term, with the simultaneous change of its sign. Formula (30) shows that if H has been found (say, experimentally 26 From the point of view of mathematics, Eq. (27) is a particular case of the so-called Legendre transformations.

27 This function (as well as the Gibbs free energy G, see below), had been introduced in 1875 by J. Gibbs, though the term “enthalpy” was coined (much later) by H. Onnes.

Chapter 1

Page 10 of 24

Image 2465

Image 2466

Image 2467

Image 2468

Image 2469

Image 2470

Image 2471

Image 2472

Image 2473

Image 2474

Image 2475

Image 2476

Image 2477

Image 2478

Image 2479

Image 2480

Image 2481

Image 2482

Image 2483

Image 2484

Image 2485

Essential Graduate Physics

SM: Statistical Mechanics

measured or calculated for a certain microscopic model) as a function of the entropy S and the pressure P of a system, we can calculate its temperature T and volume V by simple partial differentiation:

H

 

H

 

T  

 ,

V  

 .

(1.31)

S

 

P

P

S

The comparison of the first of these relations with Eq. (9) shows that not only for the heat capacity but for temperature as well, enthalpy plays the same role at fixed pressure, as played by internal energy at fixed volume.

This success immediately raises the question of whether we could develop this idea further on, by defining other useful thermodynamic potentials the variables with the dimensionality of energy that would have similar properties – first of all, a potential that would enable a similar swap of T and S in its full differential, in comparison with Eq. (30). We already know that an adiabatic process is the reversible process with fixed entropy, inviting analysis of a reversible process with fixed temperature.

Such an isothermal process may be implemented, for example, by placing the system under consideration into thermal contact with a much larger system (called either the heat bath, or “heat reservoir”, or “thermostat”) that remains in thermodynamic equilibrium at all times – see Fig. 5.

heat bath

T

Fig. 1.5. The simplest

dQ

example of the isothermal

process.

Due to its very large size, the heat bath temperature T does not depend on what is being done with our system, and if the change is being done sufficiently slowly (i.e. reversibly), that this temperature is also the temperature of our system – see Eq. (8) and its discussion. Let us calculate the elementary mechanical work d W (1) at such a reversible isothermal process. According to the general Eq. (18), d W = dEdQ. Plugging dQ from Eq. (19) into this equality, for T = const we get

d W   dE TdS d( E TS)  dF, (1.32)

T

where the following combination,

Free energy:

F E TS ,

(1.33)

definition

is called the free energy (or the “Helmholtz free energy”, or just the “Helmholtz energy”28). Just as we have done for the enthalpy, let us establish properties of this new thermodynamic potential for an arbitrarily small, reversible (now not necessarily isothermal!) variation of variables, while keeping the definition (33). Differentiating this relation and then using Eq. (17), we get

Free energy:

differential

dF   SdT PdV.

(1.34)

28 It was named after Hermann von Helmholtz (1821-1894). The last of the listed terms for F was recommended by the most recent (1988) IUPAC’s decision, but I will use the first term, which prevails is physics literature. The origin of the adjective “free” stems from Eq. (32): F is may be interpreted as the internal energy’s part that is

“free” to be transferred to the mechanical work, at the (most common) reversible, isothermal process.

Chapter 1

Page 11 of 24

Essential Graduate Physics

SM: Statistical Mechanics

Thus, if we know the function F( T, V), we can calculate S and P by simple differentiation:

F

 

F

 

S  

 , P  

 .

(1.35)

T

 

V

V

T

Now we may notice that the system of all partial derivatives may be made full and symmetric if we introduce one more thermodynamic potential. Indeed, we have already seen that each of the three already introduced thermodynamic potentials ( E, H, and F) has an especially simple full differential if it is considered as a function of its two canonical arguments: one of the “thermal variables” (either S or T) and one of the “mechanical variables” (either P or V):29

E E( S, V );

H H ( S, P);

F F( T , V ).

(1.36)

In this list of pairs of four arguments, only one pair is missing: { T, P}. The thermodynamic function of this pair, which gives the two remaining variables ( S and V) by simple differentiation, is called the Gibbs energy (or sometimes the “Gibbs free energy”): G = G( T, P). The way to define it in a symmetric way is evident from the so-called circular diagram shown in Fig. 6.

(a)

(b)

S

H

P

S

H

P

+PV -TS

Fig. 1.6. (a) The circular diagram and

E

G

E

G

(b) an example of its use for variable

calculation. The thermodynamic

-TS +PV

potentials are typeset in red, each

flanked with its two canonical

V

F

T

V

F

T

arguments.

In this diagram, each thermodynamic potential is placed between its two canonical arguments –

see Eq. (36). The left two arrows in Fig. 6a show the way the potentials H and F have been obtained from energy E – see Eqs. (27) and (33). This diagram hints that G has to be defined as shown by either of the two right arrows on that panel, i.e. as

G F PV H TS E TS PV.

(1.37) Gibbs energy:

definition

In order to verify this idea, let us calculate the full differential of this new thermodynamic potential, using, e.g., the first form of Eq. (37) together with Eq. (34):

dG dF d( PV )  ( SdT PdV )  ( PdV VdP)   SdT VdP, (1.38) Gibbs energy:

differential

so that if we know the function G( T, P), we can indeed readily calculate both entropy and volume:

G

 

G

 

S  

 , V  

 .

(1.39)

T

 

P

P

T

29 Note the similarity of this situation with that is analytical mechanics (see, e.g., CM Chapters 2 and 10): the Lagrangian function may be used to derive the equations of motion if it is expressed as a function of generalized coordinates and their velocities, while to use the Hamiltonian function in a similar way, it has to be expressed as a function of the generalized coordinates and the corresponding momenta.

Chapter 1

Page 12 of 24

Essential Graduate Physics

SM: Statistical Mechanics

The circular diagram completed in this way is a good mnemonic tool for describing Eqs. (9), (15), (31), (35), and (39), which express thermodynamic variables as partial derivatives of the thermodynamic potentials. Indeed, the variable in any corner of the diagram may be found as a partial derivative of any of two potentials that are not its immediate neighbors, over the variable in the opposite corner. For example, the green line in Fig. 6b corresponds to the second of Eqs. (39), while the blue line, to the second of Eqs. (31). At this procedure, all the derivatives giving the variables of the upper row ( S

and P) have to be taken with negative signs, while those giving the variables of the bottom row ( V and T), with positive signs.30

Now I have to justify the collective name “thermodynamic potentials” used for E, H, F, and G.

For that, let us consider an irreversible process, for example, a direct thermal contact of two bodies with different initial temperatures. As was discussed in Sec. 2, at such a process, the entropy may grow even without the external heat flow: dS  0 at dQ = 0 – see Eq. (12). This means that at a more general process with dQ  0, the entropy may grow faster than predicted by Eq. (19), which has been derived for a reversible process, so that

dQ

dS

,

(1.40)

T

with the equality approached in the reversible limit. Plugging Eq. (40) into Eq. (18) (which, being just the energy conservation law, remains valid for irreversible processes as well), we get

dE TdS PdV.

(1.41)

We can use this relation to have a look at the behavior of other thermodynamic potentials in irreversible situations, still keeping their definitions given by Eqs. (27), (33), and (37). Let us start from the (very common) case when both the temperature T and the volume V of a system are kept constant. If the process is reversible, then according to Eq. (34), the full time derivative of the free energy F would equal zero. Eq. (41) says that at an irreversible process, this is not necessarily so: if dT = dV =0, then dF

d

dE

dS

( E TS) 

T

 .

0

(1.42)

dt

dt

T

dt

dt

Hence, in the general (irreversible) situation, F can only decrease, but not increase in time. This means that F eventually approaches its minimum value F( T, S), given by the equations of reversible thermodynamics. To re-phrase this important conclusion, in the case T = const, V = const, the free energy F, i.e. the difference E – TS, plays the role of the potential energy in the classical mechanics of dissipative processes: its minimum corresponds to the (in the case of F, thermodynamic) equilibrium of the system. This is one of the key results of thermodynamics, and I invite the reader to give it some thought. One of its possible handwaving interpretations of this fact is that the heat bath with fixed T > 0, i.e. with a substantial thermal agitation of its components, “wants” to impose thermal disorder in the system immersed into it, by “rewarding” it with lower F for any increase of disorder.

30 There is also a wealth of other relations between thermodynamic variables that may be represented as second derivatives of the thermodynamic potentials, including four Maxwell relations such as ( S/ V) T = ( P/ T) V, etc.

(They may be readily recovered from the well-known property of a function of two independent arguments, say, f( x, y): ( f/ x)/ y = ( f/ y)/ x.) In this chapter, I will list only the thermodynamic relations that will be used later in the course; a more complete list may be found, e.g., in Sec. 16 of the book by L. Landau and E. Lifshitz, Statistical Physics, Part 1, 3rd ed., Pergamon, 1980 (and its later re-printings).

Chapter 1

Page 13 of 24

Essential Graduate Physics

SM: Statistical Mechanics

Repeating the calculation for a different case, T = const, P = const, it is easy to see that in this case the same role is played by the Gibbs energy:

dG

d

dE

dS

dV

dS

dV

dS

dV

( E TS PV ) 

T

P

 ( T

P

)  T

P

 ,

0

(1.43)

dt

dt

dt

dt

dt

dt

dt

dt

dt

so that the thermal equilibrium now corresponds to the minimum of G rather than F.

For the two remaining thermodynamic potentials, E and H, the calculations similar to Eqs. (42) and (43) make less sense because that would require keeping S = const (with V = const for E, and P =

const for H) for an irreversible process, but it is usually hard to prevent the entropy from growing if initially it had been lower than its equilibrium value, at least on the long-term basis.31 Thus the circular diagram is not so symmetric after all: G and F are somewhat more useful for most practical calculations than E and H.

Note that the difference GF = PV between the two “more useful” potentials has very little to do with thermodynamics at all because this difference exists (although is not much advertised) in classical mechanics as well.32 Indeed, the difference may be generalized as G – F = – F q, where q is a generalized coordinate, and F is the corresponding generalized force. The minimum of F corresponds to the equilibrium of an autonomous system (with F = 0), while the equilibrium position of the same system under the action of external force F is given by the minimum of G. Thus the external force

“wants” the system to subdue to its effect, “rewarding” it with lower G.

Moreover, the difference between F and G becomes a bit ambiguous (approach-dependent) when the product F q may be partitioned into single-particle components – just as it is done in Eqs. (2b) and (3b) for the electric and magnetic fields. Here the applied field may be taken into account on the microscopic level, including its effect directly into the energy  k of each particle. In this case, the field contributes to the total internal energy E directly, and hence the thermodynamic equilibrium (at T =

const) is described as the minimum of F. (We may say that in this case F = G, unless a difference between these thermodynamic potentials is created by the actual mechanical pressure P.) However, in some cases, typically for condensed systems, with their strong interparticle interactions, the easier (and sometimes the only one practicable33) way to account for the field is on the macroscopic level, taking G

= F F q. In this case, the same equilibrium state is described as the minimum of G. (Several examples of this dichotomy will be given later in this course.) Whatever the choice, one should mind not take the same field effect into account twice.

31 There are a few practicable systems, notably including the so-called adiabatic magnetic refrigerators (to be discussed in Chapter 2), where the unintentional growth of S is so slow that the condition S = const may be closely approached.

32 It is convenient to describe it as the difference between the “usual” (internal) potential energy U of the system to its “Gibbs potential energy” U G – see CM Sec. 1.4. For the readers who skipped that discussion: my pet example is the usual elastic spring with U = kx 2/2, under the effect of an external force F, whose equilibrium position ( x 0 = F/ k) evidently corresponds to the minimum of U G = U – F x, rather than just U.

33 An example of such an extreme situation is the case when an external magnetic field H is applied to a superconductor in its so-called intermediate state, in which the sample partitions into domains of the “normal”

phase with B = 0H, and the superconducting phase with B = 0. In this case, the field is effectively applied to the interfaces between the domains, very similarly to the mechanical pressure applied to a gas portion via a piston

– see Fig. 1 again.

Chapter 1

Page 14 of 24

Essential Graduate Physics

SM: Statistical Mechanics

One more important conceptual question I would like to discuss here is why usually statistical physics pursues the calculation of thermodynamic potentials, rather than just of a relation between P, V, and T. (Such relation is called the equation of state of the system.) Let us explore this issue on the particular but important example of an ideal classical gas in thermodynamic equilibrium, for which the equation of state should be well known to the reader from undergraduate physics:34

Ideal gas:

equation

PV NT ,

(1.44)

of state

where N is the number of particles in volume V. (In Chapter 3, we will derive Eq. (44) from statistics.) Let us try to use it for the calculation of all thermodynamic potentials, and all other thermodynamic variables discussed above. We may start, for example, from the calculation of the free energy F. Indeed, integrating the second of Eqs. (35) with the pressure calculated from Eq. (44), P = NT/ V, we get dV

d( V / N )

V

F   PdV

  NT

  NT

  NT ln

Nf ( T ),

(1.45)

T const

V

( V / N )

N

where V has been divided by N in both instances just to represent F as a manifestly extensive variable, in this uniform system proportional to N. The integration “constant” f( T) is some function of temperature, which cannot be recovered from the equation of state. This function affects all other thermodynamic potentials, and the entropy as well. Indeed, using the first of Eqs. (35) together with Eq. (45), we get

  F

V

df T 

S  

  N ln 

 ,

(1.46)

  T V

N

dT

and now may combine Eqs. (33) with (46) to calculate the (internal) energy of the gas,35

V

V

df T

df T

E F TS   NT ln

Nf T

 

T N ln

N

N f T

 

T

,

(1.47)

N

N

dT

dT

then use Eqs. (27), (44) and (47) to calculate its enthalpy,

df T

H E PV E NT N f T

 

T

T ,

(1.48)

dT

and, finally, plug Eqs. (44) and (45) into Eq. (37) to calculate the Gibbs energy

V

G F PV N T ln

f T

T  .

(1.49)

N

34 The long history of the gradual discovery of this relation includes the very early (circa 1662) work by R. Boyle and R. Townely, followed by contributions from H. Power, E. Mariotte, J. Charles, J. Dalton, and J. Gay-Lussac.

It was fully formulated by Benoît Paul Émile Clapeyron in 1834, in the form PV = nRT K, where n is the number of moles in the gas sample, and R  8.31 J/moleK is the so-called gas constant. This form is equivalent to Eq. (44), taking into account that Rk B N A, where N A = 6.022 140 761023 mole-1 is the Avogadro number, i.e. the number of molecules per mole. (By the mole’s definition, N A is just the reciprocal mass, in grams, of the 1/12th part of the 12C atom, which is close to the mass of one proton or neutron – see Appendix CA: Selected Physical Constants.) Historically, this equation of state was the main argument for the introduction of the absolute temperature T, because only with it, the equation acquires the spectacularly simple form (44).

35 Note that Eq. (47), in particular, describes a very important property of the ideal classical gas: its energy depends only on temperature (and the number of particles), but not on volume or pressure.

Chapter 1

Page 15 of 24

Image 2486

Image 2487

Image 2488

Image 2489

Image 2490

Image 2491

Image 2492

Image 2493

Image 2494

Image 2495

Image 2496

Image 2497

Image 2498

Image 2499

Image 2500

Image 2501

Image 2502

Image 2503

Image 2504

Image 2505

Essential Graduate Physics

SM: Statistical Mechanics

One might ask whether the function f( T) is physically significant, or it is something like the inconsequential, arbitrary constant – like the one that may be always added to the potential energy in non-relativistic mechanics. In order to address this issue, let us calculate, from Eqs. (24) and (28), both heat capacities, which are evidently measurable quantities:

2

E

 

d f

C  

   NT

,

(1.50)

V

2

T

 

dT

V

2

H

 

d f

C  

  N T

1  C N.

(1.51)

P



2

T

 

dT



V

P

We see that the function f( T), or at least its second derivative, is measurable.36 (In Chapter 3, we will calculate this function for two simple “microscopic” models of the ideal classical gas.) The meaning of this function is evident from the physical picture of the ideal gas: the pressure P exerted on the walls of the containing volume is produced only by the translational motion of the gas molecules, while their internal energy E (and hence other thermodynamic potentials) may be also contributed by the internal dynamics of the molecules – their rotations, vibrations, etc. Thus, the equation of state does not give us the full thermodynamic description of a system, while the thermodynamic potentials do.

1.5. Systems with a variable number of particles

Now we have to consider one more important case: when the number N of particles in a system is not rigidly fixed, but may change as a result of a thermodynamic process. A typical example of such a system is a gas sample separated from the environment by a penetrable partition – see Fig. 7.37

environment

system

dN

Fig. 1.7. An example of a system

with a variable number of particles.

Let us analyze this situation for the simplest case when all the particles are similar. (In Sec. 4.1, this analysis will be extended to systems with particles of several sorts). In this case, we may consider N

as an independent thermodynamic variable whose variation may change the energy E of the system, so that (for a slow, reversible process) Eq. (17) should be now generalized as

Chemical

dE TdS PdV dN

 ,

(1.52) potential:

definition

36 Note, however, that the difference CPCV = N is independent of f( T). (If the temperature is measured in kelvins, this relation takes a more familiar form CP – CV = nR. ) It is straightforward (and hence left for the reader’s exercise) to show that the difference CP CV of any system is fully determined by its equation of state.

37 Another important example is a gas in a contact with the open-surface liquid of similar molecules.

Chapter 1

Page 16 of 24

Essential Graduate Physics

SM: Statistical Mechanics

where  is a new function of state, called the chemical potential.38 Keeping the definitions of other thermodynamic potentials, given by Eqs. (27), (33), and (37), intact, we see that the expressions for their differentials should be generalized as

dH TdS VdP dN

 ,

(1.53a)

dF   SdT PdV dN

 ,

(1.53b)

dG   SdT VdP dN

 ,

(1.53c)

so that the chemical potential may be calculated as either of the following partial derivatives:39

E

 

H

 

F

 

G

 

  

 

 

 

 .

(1.54)

N

 

N

 

N

 

N

S , V

S , P

T , V

T, P

Despite the formal similarity of all Eqs. (54), one of them is more consequential than the others.

Indeed, the Gibbs energy G is the only thermodynamic potential that is a function of two intensive parameters, T and P. However, as all thermodynamic potentials, G has to be extensive, so that in a system of similar particles it has to be proportional to N:

G Ng ,

(1.55)

where g is some function of T and P. Plugging this expression into the last of Eqs. (54), we see that 

equals exactly this function, so that

G

 as Gibbs

 

,

(1.56)

energy

N

i.e. the chemical potential is just the Gibbs energy per particle.

In order to demonstrate how vital the notion of chemical potential may be, let us consider the situation (parallel to that shown in Fig. 2) when a system consists of two parts, with equal pressure and temperature, that can exchange particles at a relatively slow rate (much slower than the speed of the internal relaxation of each part). Then we can write two equations similar to Eqs. (5):

N N N ,

G G G ,

(1.57)

1

2

1

2

where N = const, and Eq. (56) may be used to describe each component of G: G   N   N .

(1.58)

1

1

2

2

Plugging the N 2 expressed from the first of Eqs. (57), N 2 = NN 1, into Eq. (58), we see that dG     ,

(1.59)

1

2

dN 1

so that the minimum of G is achieved at 1 = 2. Hence, in the conditions of fixed temperature and pressure, i.e. when G is the appropriate thermodynamic potential, the chemical potentials of the system parts should be equal – the so-called chemical equilibrium.

38 This name, of a historic origin, is misleading: as evident from Eq. (52),  has a clear physical sense of the average energy cost of adding one more particle to the system of N >> 1 particles.

39 Note that strictly speaking, Eqs. (9), (15), (31), (35). and (39) should be now generalized by adding another lower index, N, to the corresponding derivatives; I will just imply this.

Chapter 1

Page 17 of 24

Image 2506

Image 2507

Image 2508

Image 2509

Image 2510

Image 2511

Image 2512

Image 2513

Image 2514

Image 2515

Image 2516

Image 2517

Essential Graduate Physics

SM: Statistical Mechanics

Finally, later in the course, we will also run into several cases when the volume V of a system, its temperature T, and the chemical potential  are all fixed. (The last condition may be readily implemented by allowing the system of our interest to exchange particles with an environment so large that its  stays constant.) The thermodynamic potential appropriate for this case may be obtained by subtraction of the product  N from the free energy F, resulting in the so-called grand thermodynamic (or “Landau”) potential:

G

Grand

Ω  F N

  F

N F G   PV .

(1.60) potential:

N

definition

Indeed, for a reversible process, the full differential of this potential is

Grand

d  dF d( N)  ( SdT PdV   dN)  ( dN

Nd )   SdT PdV

Nd , (1.61) potential:

differential

so that if  has been calculated as a function of T, V, and , other thermodynamic variables may be found as

 

 

 

 

 

 

S  

 , P  

 , N  

.

(1.62)

T

  

V





 

V ,

T ,

T, V

Now acting exactly as we have done for other potentials, it is straightforward to prove that an irreversible process with fixed T, V, and , provides d/ dt  0, so that system’s equilibrium indeed corresponds to the minimum of the grand potential . We will repeatedly use this fact in this course.

1.6. Thermal machines

In order to complete this brief review of thermodynamics, I cannot completely pass the topic of thermal machines – not because it will be used much in this course, but mostly because of its practical and historic significance.40 Figure 8a shows the generic scheme of a thermal machine that may perform mechanical work on its environment (in our notation, equal to –W) during each cycle of the expansion/compression of some “working gas”, by transferring different amounts of heat from a high-temperature heat bath ( Q H) and to the low-temperature bath ( Q L).

(a)

(b)

T

H

P

Q

H

W

“working gas”

 W

Q L

T L

0

V

Fig. 1.8. (a) The simplest implementation of a thermal machine, and (b) the graphic representation of the mechanical work it performs. On panel (b), the solid arrow indicates the heat engine cycle direction, while the dashed arrow, the refrigerator cycle direction.

40 The whole field of thermodynamics was spurred by the famous 1824 work by Nicolas Léonard Sadi Carnot, in which he, in particular, gave an alternative, indirect form of the 2nd law of thermodynamics – see below.

Chapter 1

Page 18 of 24

Essential Graduate Physics

SM: Statistical Mechanics

One relation between the three amounts Q H, Q L, and W is immediately given by the energy conservation (i.e. by the 1st law of thermodynamics):

Q Q

.

(1.63)

H

L

W

From Eq. (1), the mechanical work during the cycle may be calculated as

W   PdV ,

(1.64)

and hence represented by the area circumvented by the state-representing point on the [ P, V] plane – see Fig. 8b. Note that the sign of this circular integral depends on the direction of the point’s rotation; in particular, the work (–W) done by the working gas is positive at its clockwise rotation (pertinent to heat engines) and negative in the opposite case (implemented in refrigerators and heat pumps – see below).

Evidently, the work depends on the exact form of the cycle, which in turn may depend not only on T H

and T L, but also on the working gas’ properties.

An exception from this rule is the famous Carnot cycle, consisting of two isothermal and two adiabatic processes (all reversible!). In its heat engine’s form, the cycle may start, for example, from an isothermic expansion of the working gas in contact with the hot bath (i.e. at T = T H). It is followed by its additional adiabatic expansion (with the gas being disconnected from both heat baths) until its temperature drops to T L. Then an isothermal compression of the gas is performed in its contact with the cold bath (at T = T L), followed by its additional adiabatic compression to raise T to T H again, after which the cycle is repeated again and again. Note that during this cycle the working gas is never in contact with both heat baths simultaneously, thus avoiding the irreversible heat transfer between them. The cycle’s shape on the [ V, P] plane (Fig. 9a) depends on the exact properties of the working gas and may be rather complicated. However, since the system’s entropy is constant at any adiabatic process, the Carnot cycle’s shape on the [ S, T] plane is always rectangular – see Fig. 9b.

(a)

T

(b)

P

T T

H

T H

Fig. 1.9. Representation of the

S S

1

Carnot cycle: (a) on the [ V, P]

S S 2

plane (schematically), and (b) on

T L

the [ S, T] plane. The meaning of

T T L

the arrows is the same as in Fig. 8.

0

0

V

S

S

1

2

S

Since during each isotherm, the working gas is brought into thermal contact only with the corresponding heat bath, i.e. its temperature is constant, the relation (19), dQ = TdS, may be immediately integrated to yield

Q T ( S S ),

Q T ( S S ).

(1.65)

H

H

2

1

L

L

2

1

Hence the ratio of these two heat flows is completely determined by their temperature ratio: Q

T

H

H

,

(1.66)

Q

T

L

L

Chapter 1

Page 19 of 24

Essential Graduate Physics

SM: Statistical Mechanics

regardless of the working gas properties. Formulas (63) and (66) are sufficient to find the ratio of the work (–W) to any of Q H and Q L. For example, the main figure-of-merit of a thermal machine used as a heat engine ( Q H > 0, Q L > 0, –W =  W  > 0), is its efficiency W

Q Q

Q

Heat

H

L

 

 1 L  1.

(1.67) engine’s

Q

Q

Q

efficiency:

H

H

H

definition

For the Carnot cycle, this definition, together with Eq. (66), immediately yield the famous relation41

T

Carnot

L

 1

,

(1.68) cycle’s

Carnot

T

efficiency

H

which shows that at a given T L (that is typically the ambient temperature ~300 K), the efficiency may be increased, ultimately to 1, by raising the temperature T H of the heat source.42

The unique nature of the Carnot cycle (see Fig. 9b again) makes its efficiency (68) the upper limit for any heat engine.43 Indeed, in this cycle, the transfer of heat between any heat bath and the working gas is performed reversibly, when their temperatures are equal. (If this is not so, some heat might flow from a hotter to colder bath without performing any work.) In particular, it shows that max =

0 at T H = T L, i.e., no heat engine can perform mechanical work in the absence of temperature gradients.44

On the other hand, if the cycle is reversed (see the dashed arrows in Figs. 8 and 9), the same thermal machine may serve as a refrigerator, providing heat removal from the low-temperature bath ( Q L

< 0) at the cost of consuming external mechanical work: W > 0. This reversal does not affect the basic relation (63), which now may be used to calculate the relevant figure-of-merit, called the cooling coefficient of performance (COPcooling):

Q

Q

L

L

COP

.

(1.69)

cooling

W

Q Q

H

L

Notice that this coefficient may be above unity; in particular, for the Carnot cycle we may use Eq. (66) (which is also unaffected by the cycle reversal) to get

T

L

(COP

)

,

(1.70)

cooling Carnot

T T

H

L

41 Curiously, S. Carnot derived his key result still believing that heat is some specific fluid (“caloric”), whose flow is driven by the temperature difference, rather than just a form of particle motion.

42 Semi-quantitatively, such trend is valid also for other, less efficient but more practicable heat engine cycles –

see Problems 13-16. This trend is the leading reason why internal combustion engines, with T H of the order of 1,500 K, are more efficient than steam engines, with the difference T H – T L of at most a few hundred K.

43 In some alternative axiomatic systems of thermodynamics, this fact is postulated and serves the role of the 2nd law. This is why it is under persisting (dominantly, theoretical) attacks by suggestions of more efficient heat engines – recently, mostly of quantum systems using sophisticated protocols such as the so-called shortcut-to-adiabaticity – see, e.g., the recent paper by O. Abah and E. Lutz, Europhysics Lett. 118, 40005 (2017), and references therein. To the best of my knowledge, reliable analyses of all the suggestions put forward so far have confirmed that the Carnot efficiency (68) is the highest possible even in quantum systems.

44 Such a hypothetical heat engine, which would violate the 2nd law of thermodynamics, is called the “perpetual motion machine of the 2nd kind” – in contrast to any (also hypothetical) “perpetual motion machine of the 1st kind” that would violate the 1st law, i.e., the energy conservation.

Chapter 1

Page 20 of 24

Essential Graduate Physics

SM: Statistical Mechanics

so that this value is larger than 1 at T H < 2 T L, and even may be much larger than that when the temperature difference ( T H – T L) sustained by the refrigerator, tends to zero. For example, in a typical air-conditioning system, this difference is of the order of 10 K, while T L ~ 300 K, so that ( T H – T L) ~

T L/30, i.e. the Carnot value of COPcooling is as high as ~30. (In the state-of-the-art commercial HVAC

systems it is within the range of 3 to 4.) This is why the term “cooling efficiency”, used in some textbooks instead of (COP)cooling, may be misleading.

Since in the reversed cycle Q H = –W + Q L < 0, i.e. the system provides heat flow into the high-temperature heat bath, it may be used as a heat pump for heating purposes. The figure-of-merit appropriate for this application is different from Eq. (69):

Q H

Q

COP

H

,

(1.71)

heating

W

Q Q

H

L

so that for the Carnot cycle, using Eq. (66) again, we get

T

H

(COP

)

.

(1.72)

heating Carnot

T T

H

L

Note that this COP is always larger than 1, meaning that the Carnot heat pump is always more efficient than the direct conversion of work into heat (when Q H = –W, so that COPheating = 1), though practical electricity-driven heat pumps are substantially more complex, and hence more expensive than simple electric heaters. Such heat pumps, with the typical COPheating values around 4 in summer and 2 in winter, are frequently used for heating large buildings.

Finally, note that according to Eq. (70), the COPcooling of the Carnot cycle tends to zero at T L 

0, making it impossible to reach the absolute zero of temperature, and hence illustrating the meaningful (Nernst’s) formulation of the 3rd law of thermodynamics, cited in Sec. 3. Indeed, let us prescribe a finite but very large heat capacity C( T) to the low-temperature bath, and use the definition of this variable to write the following expression for the relatively small change of its temperature as a result of dn similar refrigeration cycles:

C T

( ) dT Q dn .

(1.73)

L

L

L

Together with Eq. (66), this relation yields

C T

( ) dT

Q

L

L

H

 

dn .

(1.74)

T

T

L

H

If T L  0, so that T H >> T L and  Q H   –W = const, the right-hand side of this equation does not depend on T L, so that if we integrate it over many ( n >> 1) cycles, getting the following simple relation between the initial and final values of T L:

T fin C T

( ) dT

Q

H

 

n

.

(1.75)

T

T

T

H

ini

For example, if C( T) is a constant, Eq. (75) yields an exponential law,

Q

T T exp

,

(1.76)

fin

ini



H n

CT H 

Chapter 1

Page 21 of 24

Essential Graduate Physics

SM: Statistical Mechanics

with the absolute zero of temperature not reached as any finite n. Even for an arbitrary function C( T) that does not vanish at T  0, Eq. (74) proves the Nernst theorem, because dn diverges at T L  0. 45

1.7. Exercise problems

1.1. Two bodies, with temperature-independent heat capacities C 1 and C 2, and different initial temperatures T 1 and T 2, are placed into a weak thermal contact. Calculate the change of the total entropy of the system before it reaches the thermal equilibrium.

1.2. A gas portion has the following properties:

(i) its heat capacity CV = aT b, and

(ii) the work W T needed for its isothermal compression from V 2 to V 1 equals cT ln( V 2/ V 1), where a, b, and c are some constants. Find the equation of state of the gas, and calculate the temperature dependence of its entropy S and thermodynamic potentials E, H, F, G, and .

1.3. A closed volume with an ideal classical gas of similar molecules is separated with a partition in such a way that the number N of molecules in each part is the same, but their volumes are different.

The gas is initially in thermal equilibrium, and its pressure in one part is P 1, and in the other part, P 2.

Calculate the change of entropy resulting from a fast removal of the partition, and analyze the result.

1.4. An ideal classical gas of N particles is initially confined to volume V, and is in thermal equilibrium with a heat bath of temperature T. Then the gas is allowed to expand to volume V’ > V in one of the following ways:

(i) The expansion is slow, so that due to the sustained thermal contact with the heat bath, the gas temperature remains equal to T.

(ii) The partition separating the volumes V and ( V’V) is removed very fast, allowing the gas to expand rapidly.

For each process, calculate the eventual changes of pressure, temperature, energy, and entropy of the gas at its expansion.

1.5. For an ideal classical gas with temperature-independent specific heat, derive the relation between P and V at an adiabatic expansion/compression.

1.6. Calculate the speed and the wave impedance of acoustic waves propagating in an ideal classical gas with temperature-independent specific heat , in the limits when the propagation may be treated as:

(i) an isothermal process, and

(ii) an adiabatic process.

45 Note that for such metastable systems as glasses the situation may be more complicated. (For a detailed discussion of this issue see, e.g., J. Wilks, The Third Law of Thermodynamics, Oxford U. Press, 1961.) Fortunately, this issue does not affect other aspects of statistical physics – at least those to be discussed in this course.

Chapter 1

Page 22 of 24

Essential Graduate Physics

SM: Statistical Mechanics

Which of these limits is achieved at higher wave frequencies?

1.7. As will be discussed in Sec. 3.5, the so-called “hardball” models of classical particle interaction yield the following equation of state of a gas of such particles:

P T n,

where n = N/ V is the particle density, and the function ( n) is generally different from that (ideal( n) = n) of the ideal gas, but still independent of temperature. For such a gas, with temperature-independent cV, calculate:

(i) the energy of the gas, and

(ii) its pressure as a function of n at the adiabatic compression.

1.8. For an arbitrary thermodynamic system with a fixed number of particles, prove the following four Maxwell relations (already mentioned in Sec. 4):

  S

 

P

 

V

 

T

 

i

: 

  

 ,

ii :     ,

V

 

T

 

S

 

P

T

V

P

S

S

 

V

 

P

 

T

 

iii

: 

  

 , iv

: 

  

 ,

P

 

T

 

S

 

V

T

P

V

S

and also the following relation:

E

 

P

 

  T

  .

P

V

 

T

T

V

1.9. Express the heat capacity difference, CPCV, via the equation of state P = P( V, T) of the system.

1.10. Prove that the isothermal compressibility 46

1  V

 

   

T

V P

  T, N

in a single-phase system may be expressed in two different ways:

V 2

2

  P

V N

 

 

.

T

N 2 

2 

  

N 2 



  

T

T V

,

P

1.11. A reversible process, performed with a fixed portion of an ideal 3 P

classical gas, may be represented on the [ V, P] plane with the straight line 0

shown in the figure on the right. Find the point at which the heat flow into/out P

of the gas changes its direction.

0

0

V / 2

V

V

1.12. Two bodies have equal, temperature-independent heat capacities

0

0

46 Note that the compressibility is just the reciprocal bulk modulus,  = 1/ K – see, e.g., CM Sec. 7.3.

Chapter 1

Page 23 of 24

Essential Graduate Physics

SM: Statistical Mechanics

C, but different temperatures, T 1 and T 2. Calculate the maximum mechanical work obtainable from this system, using a heat engine.

P

1.13. Express the efficiency  of a heat engine that uses the so- P

called Joule cycle, consisting of two adiabatic and two isobaric

max

S  const

processes (see the figure on the right), via the minimum and maximum

S  const

values of pressure, and compare the result with Carnot. Assume an ideal P min

classical working gas with temperature-independent CP and CV.

0

V

1.14. Calculate the efficiency of a heat engine using the Otto P

S  const

cycle,47 which consists of two adiabatic and two isochoric (constant-

volume) reversible processes – see the figure on the right. Explore how

the efficiency depends on the ratio r V max/ V min, and compare it with the Carnot cycle’s efficiency. Assume an ideal classical working gas with

S

temperature-independent heat capacity.

const

0

V

rV

0

0 V

1.15. A heat engine’s cycle consists of two isothermal ( T =

T

const) and two isochoric ( V = const) reversible processes – see the figure T H

on the right.48

(i) Assuming that the working gas is an ideal classical gas of N

particles, calculate the mechanical work performed by the engine during T L

one cycle.

(ii) Are the specified conditions sufficient to calculate the

engine’s efficiency? (Justify your answer.)

0

V

V

V

1

2

P  const

1.16. The Diesel cycle (an approximate model of the Diesel P

internal combustion engine’s operation) consists of two adiabatic

2

3 S  const

processes, one isochoric process, and one isobaric process – see the

4

figure on the right. Assuming an ideal working gas with temperature-

V  const

independent C

S  const

V and CP, express the efficiency  of the heat engine using

this cycle via the gas temperature values in its transitional states

1

corresponding to the corners of the cycle diagram.

0

V

47 This name stems from the fact that the cycle is an approximate model of operation of the four-stroke internal combustion engine, which was improved and made practicable (though not invented!) by N. Otto in 1876.

48 The reversed cycle of this type is a reasonable approximation for the operation of the Stirling and Gifford-McMahon (GM) refrigerators, broadly used for cryocooling – for a recent review see, e.g., A. de Waele, J. Low Temp. Phys. 164, 179 (2011).

Chapter 1

Page 24 of 24

Essential Graduate Physics

SM: Statistical Mechanics

Chapter 2. Principles of Physical Statistics

This chapter is the keystone of this course. It starts with a brief discussion of such basic notions of statistical physics as statistical ensembles, probability, and ergodicity. Then the so-called microcanonical distribution postulate is formulated, simultaneously with the statistical definition of the entropy. This allows a derivation of the famous Gibbs (“canonical”) distribution – the most frequently used tool of statistical physics. Then we will discuss one more, “grand canonical” distribution, which is more convenient for some tasks. In particular, it is immediately used for the derivation of the most important Boltzmann, Fermi-Dirac, and Bose-Einstein statistics of independent particles, which will be repeatedly utilized in the following chapters.

2.1. Statistical ensembles and probability

As has been already discussed in Sec. 1.1, statistical physics deals with situations when either unknown initial conditions, or system’s complexity, or the laws of its motion (as in the case of quantum mechanics) do not allow a definite prediction of measurement results. The main formalism for the analysis of such systems is the probability theory, so let me start with a very brief review of its basic concepts, using an informal “physical” language – less rigorous but (hopefully) more transparent than standard mathematical treatments,1 and quite sufficient for our purposes.

Consider N >> 1 independent similar experiments carried out with apparently similar systems (i.e. systems with identical macroscopic parameters such as volume, pressure, etc.), but still giving, by any of the reasons listed above, different results of measurements. Such a collection of experiments, together with a fixed method of result processing, is a good example of a statistical ensemble. Let us start from the case when the experiments may have M different discrete outcomes, and the number of experiments giving the corresponding different results is N 1, N 2,…, NM, so that M

N N.

(2.1)

m

m1

The probability of each outcome, for the given statistical ensemble, is then defined as N

Probability

W  lim

m .

(2.2)

m

N 

N

Though this definition is so close to our everyday experience that it is almost self-evident, a few remarks may still be relevant.

First, the probabilities Wm depend on the exact statistical ensemble they are defined for, notably including the method of result processing. As the simplest example, consider throwing the standard cubic-shaped dice many times. For the ensemble of all thrown and counted dice, the probability of each outcome (say, “1”) is 1/6. However, nothing prevents us from defining another statistical ensemble of dice-throwing experiments in which all outcomes “1” are discounted. Evidently, the probability of 1 For the reader interested in a more rigorous approach, I can recommend, for example, Chapter 18 of the handbook by G. Korn and T. Korn – see MA Sec. 16(ii).

© K. Likharev

Essential Graduate Physics

SM: Statistical Mechanics

finding outcomes “1” in this modified (but legitimate) ensemble is 0, while for all other five outcomes (“2” to “6”), it is 1/5 rather than 1/6.

Second, a statistical ensemble does not necessarily require N similar physical systems, e.g., N

distinct dice. It is intuitively clear that tossing the same die N times constitutes an ensemble with similar statistical properties. More generally, a set of N experiments with the same system gives a statistical ensemble equivalent to the set of experiments with N different systems, provided that the experiments are kept independent, i.e. that outcomes of past experiments do not affect those of the experiments to follow. Moreover, for many physical systems of interest, no special preparation of each experiment is necessary, and N experiments separated by sufficiently long time intervals, form a “good” statistical ensemble – the property called ergodicity.2

Third, the reference to infinite N in Eq. (2) does not strip the notion of probability from its practical relevance. Indeed, it is easy to prove (see Chapter 5) that, at very general conditions, at finite but sufficiently large N, the numbers Nm are approaching their average (or expectation) values 3

N

W N ,

(2.3)

m

m

with the relative deviations decreasing as ~1/ Nm1/2, i.e. as 1/ N 1/2.

Now let me list those properties of probabilities that we will immediately need. First, dividing both sides of Eq. (1) by N and following the limit N  , we get the well-known normalization condition

M

W  1;