Thermodynamics 


The third law of thermodynamics states as follows, regarding the properties of closed systems in thermodynamic equilibrium:
The entropy of a system approaches a constant value as its temperature approaches absolute zero.
This constant value cannot depend on any other parameters characterizing the closed system, such as pressure or applied magnetic field. At absolute zero (zero kelvins) the system must be in a state with the minimum possible energy. Entropy is related to the number of accessible microstates, and there is typically one unique state (called the ground state) with minimum energy.^{[1]} In such a case, the entropy at absolute zero will be exactly zero. If the system does not have a welldefined order (if its order is glassy, for example), then there may remain some finite entropy as the system is brought to very low temperatures, either because the system becomes locked into a configuration with nonminimal energy or because the minimum energy state is nonunique. The constant value is called the residual entropy of the system.^{[2]} The entropy is essentially a statefunction meaning the inherent value of different atoms, molecules, and other configurations of particles including subatomic or atomic material is defined by entropy, which can be discovered near 0 K. The Nernst–Simon statement of the third law of thermodynamics concerns thermodynamic processes at a fixed, low temperature:
The entropy change associated with any condensed system undergoing a reversible isothermal process approaches zero as the temperature at which it is performed approaches 0 K.
Here a condensed system refers to liquids and solids. A classical formulation by Nernst (actually a consequence of the Third Law) is:
It is impossible for any process, no matter how idealized, to reduce the entropy of a system to its absolutezero value in a finite number of operations.^{[3]}
There also exists a formulation of the third law which approaches the subject by postulating a specific energy behavior:
If the composite of two thermodynamic systems constitutes an isolated system, then any energy exchange in any form between those two systems is bounded.^{[4]}
The third law was developed by chemist Walther Nernst during the years 1906–12, and is therefore often referred to as Nernst's theorem or Nernst's postulate. The third law of thermodynamics states that the entropy of a system at absolute zero is a welldefined constant. This is because a system at zero temperature exists in its ground state, so that its entropy is determined only by the degeneracy of the ground state.
In 1912 Nernst stated the law thus: "It is impossible for any procedure to lead to the isotherm T = 0 in a finite number of steps."^{[5]}
An alternative version of the third law of thermodynamics as stated by Gilbert N. Lewis and Merle Randall in 1923:
This version states not only ΔS will reach zero at 0 K, but S itself will also reach zero as long as the crystal has a ground state with only one configuration. Some crystals form defects which cause a residual entropy. This residual entropy disappears when the kinetic barriers to transitioning to one ground state are overcome.^{[6]}
With the development of statistical mechanics, the third law of thermodynamics (like the other laws) changed from a fundamental law (justified by experiments) to a derived law (derived from even more basic laws). The basic law from which it is primarily derived is the statisticalmechanics definition of entropy for a large system:
where is S entropy, k_{B} is the Boltzmann constant, and Ω is the number of microstates consistent with the macroscopic configuration. The counting of states is from the reference state of absolute zero, which corresponds to the entropy of S_{0}.
In simple terms, the third law states that the entropy of a perfect crystal of a pure substance approaches zero as the temperature approaches zero. The alignment of a perfect crystal leaves no ambiguity as to the location and orientation of each part of the crystal. As the energy of the crystal is reduced, the vibrations of the individual atoms are reduced to nothing, and the crystal becomes the same everywhere.
The third law provides an absolute reference point for the determination of entropy at any other temperature. The entropy of a closed system, determined relative to this zero point, is then the absolute entropy of that system. Mathematically, the absolute entropy of any system at zero temperature is the natural log of the number of ground states times Boltzmann's constant k_{B} = 1.38×10^{−23} J K^{−1}.
The entropy of a perfect crystal lattice as defined by Nernst's theorem is zero provided that its ground state is unique, because ln(1) = 0. If the system is composed of onebillion atoms, all alike, and lie within the matrix of a perfect crystal, the number of combinations of onebillion identical things taken onebillion at a time is Ω = 1. Hence:
The difference is zero, hence the initial entropy S_{0} can be any selected value so long as all other such calculations include that as the initial entropy. As a result, the initial entropy value of zero is selected S_{0} = 0 is used for convenience.
Suppose a system consisting of a crystal lattice with volume V of N identical atoms at T = 0 K, and an incoming photon of wavelength λ and energy ε.
Initially, there is only one accessible microstate:
Let's assume the crystal lattice absorbs the incoming photon. There is a unique atom in the lattice that interacts and absorbs this photon. So after absorption, there is N possible microstates accessible by the system, each of the microstates corresponding to one excited atom, and the other atoms remaining at ground state.
The entropy, energy, and temperature of the closed system rises and can be calculated. The entropy change is:
From the second law of thermodynamics:
Hence:
Calculating entropy change:
We assume N = 3 × 10^{22} and λ = 1 cm. The energy change of the system as a result of absorbing the single photon whose energy is ε:
The temperature of the closed system rises by:
This can be interpreted as the average temperature of the system over the range from .^{[7]} A single atom was assumed to absorb the photon but the temperature and entropy change characterizes the entire system.
An example of a system which does not have a unique ground state is one whose net spin is a halfinteger, for which timereversal symmetry gives two degenerate ground states. For such systems, the entropy at zero temperature is at least k_{B} ln(2) (which is negligible on a macroscopic scale). Some crystalline systems exhibit geometrical frustration, where the structure of the crystal lattice prevents the emergence of a unique ground state. Groundstate helium (unless under pressure) remains liquid.
In addition, glasses and solid solutions retain large entropy at 0 K, because they are large collections of nearly degenerate states, in which they become trapped out of equilibrium.^{[citation needed]} Another example of a solid with many nearlydegenerate ground states, trapped out of equilibrium, is ice Ih, which has "proton disorder".
For the entropy at absolute zero to be zero, the magnetic moments of a perfectly ordered crystal must themselves be perfectly ordered; from an entropic perspective, this can be considered to be part of the definition of a "perfect crystal". Only ferromagnetic, antiferromagnetic, and diamagnetic materials can satisfy this condition. However, ferromagnetic materials do not, in fact, have zero entropy at zero temperature, because the spins of the unpaired electrons are all aligned and this gives a groundstate spin degeneracy. Materials that remain paramagnetic at 0 K, by contrast, may have many nearlydegenerate ground states (for example, in a spin glass), or may retain dynamic disorder (a quantum spin liquid).^{[citation needed]}
The third law is equivalent to the statement that
The reason that T = 0 cannot be reached according to the third law is explained as follows: Suppose that the temperature of a substance can be reduced in an isentropic process by changing the parameter X from X_{2} to X_{1}. One can think of a multistage nuclear demagnetization setup where a magnetic field is switched on and off in a controlled way.^{[9]} If there were an entropy difference at absolute zero, T = 0 could be reached in a finite number of steps. However, at T = 0 there is no entropy difference so an infinite number of steps would be needed. The process is illustrated in Fig. 1.
A nonquantitative description of his third law that Nernst gave at the very beginning was simply that the specific heat can always be made zero by cooling the material down far enough.^{[10]} A modern, quantitative analysis follows.
Supposed that the heat capacity of a sample in the low temperature region has the form of a power law C(T,X) = C_{0}T^{α} asymptotically as T → 0, and we wish to find which values of α are compatible with the third law. We have

(11) 
By the discussion of third law (above), this integral must be bounded as T_{0} → 0, which is only possible if α > 0. So the heat capacity must go to zero at absolute zero

(12) 
if it has the form of a power law. The same argument shows that it cannot be bounded below by a positive constant, even if we drop the powerlaw assumption.
On the other hand, the molar specific heat at constant volume of a monatomic classical ideal gas, such as helium at room temperature, is given by C_{V} = (3/2)R with R the molar ideal gas constant. But clearly a constant heat capacity does not satisfy Eq. (12). That is, a gas with a constant heat capacity all the way to absolute zero violates the third law of thermodynamics. We can verify this more fundamentally by substituting C_{V} in Eq. (14), which yields

(13) 
In the limit T_{0} → 0 this expression diverges, again contradicting the third law of thermodynamics.
The conflict is resolved as follows: At a certain temperature the quantum nature of matter starts to dominate the behavior. Fermi particles follow Fermi–Dirac statistics and Bose particles follow Bose–Einstein statistics. In both cases the heat capacity at low temperatures is no longer temperature independent, even for ideal gases. For Fermi gases

(14) 
with the Fermi temperature T_{F} given by

(15) 
Here N_{A} is Avogadro's number, V_{m} the molar volume, and M the molar mass.
For Bose gases

(16) 
with T_{B} given by

(17) 
The specific heats given by Eq. (14) and (16) both satisfy Eq. (12). Indeed, they are power laws with α=1 and α=3/2 respectively.
Even within a purely classical setting, the density of a classical ideal gas at fixed particle number becomes arbitrarily high as T goes to zero, so the interparticle spacing goes to zero. The assumption of noninteracting particles presumably breaks down when they are sufficiently close together, so the value of C_{V} gets modified away from its ideal constant value.
The only liquids near absolute zero are ^{3}He and ^{4}He. Their heat of evaporation has a limiting value given by

(18) 
with L_{0} and C_{p} constant. If we consider a container, partly filled with liquid and partly gas, the entropy of the liquid–gas mixture is

(19) 
where S_{l}(T) is the entropy of the liquid and x is the gas fraction. Clearly the entropy change during the liquid–gas transition (x from 0 to 1) diverges in the limit of T→0. This violates Eq.(8). Nature solves this paradox as follows: at temperatures below about 50 mK the vapor pressure is so low that the gas density is lower than the best vacuum in the universe. In other words: below 50 mK there is simply no gas above the liquid.
The melting curves of ^{3}He and ^{4}He both extend down to absolute zero at finite pressure. At the melting pressure, liquid and solid are in equilibrium. The third law demands that the entropies of the solid and liquid are equal at T = 0. As a result, the latent heat of melting is zero and the slope of the melting curve extrapolates to zero as a result of the Clausius–Clapeyron equation.
The thermal expansion coefficient is defined as

(20) 
With the Maxwell relation

(21) 
and Eq. (8) with X = p it is shown that

(22) 
So the thermal expansion coefficient of all materials must go to zero at zero kelvin.