# 15.6: Entropy and the Second Law of Thermodynamics: Disorder and the Unavailability of Energy

- Page ID
- 1599

**Figure \(\PageIndex{1}\):** The ice in this drink is slowly melting. Eventually the liquid will reach thermal equilibrium, as predicted by the second law of thermodynamics. (credit: Jon Sullivan, PDPhoto.org)

There is yet another way of expressing the second law of thermodynamics. This version relates to a concept called entropy. By examining it, we shall see that the directions associated with the second law—heat transfer from hot to cold, for example—are related to the tendency in nature for systems to become disordered and for less energy to be available for use as work. The entropy of a system can in fact be shown to be a measure of its disorder and of the unavailability of energy to do work.

MAKING CONNECTIONS: ENTROPY, ENERGY, AND WORK

Recall that the simple definition of energy is the ability to do work. Entropy is a measure of how much energy is not available to do work. Although all forms of energy are interconvertible, and all can be used to do work, it is not always possible, even in principle, to convert the entire available energy into work. That unavailable energy is of interest in thermodynamics, because the field of thermodynamics arose from efforts to convert heat to work.

We can see how entropy is defined by recalling our discussion of the Carnot engine. We noted that for a Carnot cycle, and hence for any reversible processes, \(Q_c/Q_h = T_c/T_h\). Rearranging terms yields \[\dfrac{Q_c}{T_c} = \dfrac{Q_h}{T_h}\] for any reversible process. \(Q_c\) and \(Q_h\) are absolute values of the heat transfer at temperatures \(T_c\) and \(T_h\), respectively. This ratio of \(Q/T\) is defined to be the **change in entropy** \(\Delta S\) for a reversible process,

\[\Delta S = \left(\dfrac{Q}{T} \right)_{rev},\]

where \(Q\) is the heat transfer, which is positive for heat transfer into and negative for heat transfer out of, and \(T\) is the absolute temperature at which the reversible process takes place. The SI unit for entropy is joules per kelvin (J/K). If temperature changes during the process, then it is usually a good approximation (for small changes in temperature) to take \(T\) to be the average temperature, avoiding the need to use integral calculus to find \(\Delta S\).

The definition of \(\Delta S\) is strictly valid only for reversible processes, such as used in a Carnot engine. However, we can find \(\Delta S\) precisely even for real, irreversible processes. The reason is that the entropy \(S\) of a system, like internal energy \(U\) depends only on the state of the system and not how it reached that condition. Entropy is a property of state. Thus the change in entropy \(\Delta S\) of a system between state 1 and state 2 is the same no matter how the change occurs. We just need to find or imagine a reversible process that takes us from state 1 to state 2 and calculate \(\Delta S\) for that process. That will be the change in entropy for any process going from state 1 to state 2. (See Figure \(\PageIndex{2}\).)

**Figure \(\PageIndex{2}\):** When a system goes from state 1 to state 2, its entropy changes by the same amount \(\Delta S\) whether a hypothetical reversible path is followed or a real irreversible path is taken.

Now let us take a look at the change in entropy of a Carnot engine and its heat reservoirs for one full cycle. The hot reservoir has a loss of entropy \(\Delta S_h = -Q_h/T_h\), because heat transfer occurs out of it (remember that when heat transfers out, then \(Q\) has a negative sign). The cold reservoir has a gain of entropy \(\Delta S_c = Q_c/T_c\), because heat transfer occurs into it. (We assume the reservoirs are sufficiently large that their temperatures are constant.) So the total change in entropy is

\[\Delta S_{tot} = \Delta S_h + \Delta S_c.\]

Thus, since we know that \(Q_h/T_h = Q_c/T_c\) for a Carnot engine,

\[\Delta S_{tot} = -\dfrac{Q_h}{T_h} + \dfrac{Q_c}{T_c} = 0.\]

This result, which has general validity, means that *the total change in entropy for a system in any reversible process is zero.*

The entropy of various parts of the system may change, but the total change is zero. Furthermore, the system does not affect the entropy of its surroundings, since heat transfer between them does not occur. Thus the reversible process changes neither the total entropy of the system nor the entropy of its surroundings. Sometimes this is stated as follows: *Reversible processes do not affect the total entropy of the universe.* Real processes are not reversible, though, and they do change total entropy. We can, however, use hypothetical reversible processes to determine the value of entropy in real, irreversible processes. The following example illustrates this point.

Example \(\PageIndex{1}\): Entropy Increases in an Irreversible (Real) Process

Spontaneous heat transfer from hot to cold is an irreversible process. Calculate the total change in entropy if 4000 J of heat transfer occurs from a hot reservoir at \(T_h = 600 \, K \, (327^oC) \) to a cold reservoir at \(T_c = 250 \, K \, (-23^oC)\), assuming there is no temperature change in either reservoir. (See Figure \(\PageIndex{3}\).)

**Strategy**

How can we calculate the change in entropy for an irreversible process when \(\Delta S_{tot} = \Delta S_h + \Delta S_c\) is valid only for reversible processes? Remember that the total change in entropy of the hot and cold reservoirs will be the same whether a reversible or irreversible process is involved in heat transfer from hot to cold. So we can calculate the change in entropy of the hot reservoir for a hypothetical reversible process in which 4000 J of heat transfer occurs from it; then we do the same for a hypothetical reversible process in which 4000 J of heat transfer occurs to the cold reservoir. This produces the same changes in the hot and cold reservoirs that would occur if the heat transfer were allowed to occur irreversibly between them, and so it also produces the same changes in entropy.

**Solution**

We now calculate the two changes in entropy using \(\Delta S_{tot} = \Delta S_h + \Delta S_c\). First, for the heat transfer from the hot reservoir, \[\Delta S_h = \dfrac{-Q_h}{T_h} = \dfrac{-4000 \, J}{600 \, K} = -6.67 \, J/K.\] And for the cold reservoir, \[\Delta S_c = \dfrac{Q_c}{T_c} = \dfrac{4000 \, J}{250 \, K} = 16.0 \, J/K.\]

Thus the total is \[\Delta S_{tot} = \Delta S_h + \Delta S_c\] \[= (-6.67 + 16.0) \, J/K\] \[= 9.33 \, J/K.\]

**Discussion**

There is an *increase* in entropy for the system of two heat reservoirs undergoing this irreversible heat transfer. We will see that this means there is a loss of ability to do work with this transferred energy. Entropy has increased, and energy has become unavailable to do work.

**Figure \(\PageIndex{3}\):** (a) Heat transfer from a hot object to a cold one is an irreversible process that produces an overall increase in entropy. (b) The same final state and, thus, the same change in entropy is achieved for the objects if reversible heat transfer processes occur between the two objects whose temperatures are the same as the temperatures of the corresponding objects in the irreversible process.

It is reasonable that entropy increases for heat transfer from hot to cold. Since the change in entropy is \(Q/T\) there is a larger change at lower temperatures. The decrease in entropy of the hot object is therefore less than the increase in entropy of the cold object, producing an overall increase, just as in the previous example. This result is very general:

*There is an increase in entropy for any system undergoing an irreversible process.*

With respect to entropy, there are only two possibilities: entropy is constant for a reversible process, and it increases for an irreversible process. There is a fourth version of** the second law of thermodynamics stated in terms of entropy:**

*The total entropy of a system either increases or remains constant in any process; it never decreases.*

For example, heat transfer cannot occur spontaneously from cold to hot, because entropy would decrease.

Entropy is very different from energy. Entropy is *not* conserved but increases in all real processes. Reversible processes (such as in Carnot engines) are the processes in which the most heat transfer to work takes place and are also the ones that keep entropy constant. Thus we are led to make a connection between entropy and the availability of energy to do work.

# Entropy and the Unavailability of Energy to Do Work

What does a change in entropy mean, and why should we be interested in it? One reason is that entropy is directly related to the fact that not all heat transfer can be converted into work. The next example gives some indication of how an increase in entropy results in less heat transfer into work.

Example \(\PageIndex{2}\): Less Work is Produced by a Given Heat Transfer When Entropy Change is Greater

(a) Calculate the work output of a Carnot engine operating between temperatures of 600 K and 100 K for 4000 J of heat transfer to the engine. (b) Now suppose that the 4000 J of heat transfer occurs first from the 600 K reservoir to a 250 K reservoir (without doing any work, and this produces the increase in entropy calculated above) before transferring into a Carnot engine operating between 250 K and 100 K. What work output is produced? (See Figure \(\PageIndex{4}\).)

**Strategy**

In both parts, we must first calculate the Carnot efficiency and then the work output.

**Solution (a)**

The Carnot efficiency is given by \[Eff_c = 1 - \dfrac{T_c}{T_h}.\]

Substituting the given temperatures yields \[Eff_c = 1 - \dfrac{100 \, K}{600 \, K} = 0.833.\]

Now the work output can be calculated using the definition of efficiency for any heat engine as given by \[Eff = \dfrac{W}{Q_h}.\]

Solving for \(W\) and substituting known terms gives \[W = Eff_cQ_h\]\[= (0.833)(4000 \, J) = 3333 \, J.\]

**Solution (b)**

Similarly, \[Eff_c = 1 - \dfrac{T_c}{T'_c} = 1 - \dfrac{100 \, K}{250 \, K} = 0.600,\] so that \[W = Eff_cQ_h\] \[ = (0.600)(4000 \, J) = 2400 \, J\]

**Discussion**

There is 933 J less work from the same heat transfer in the second process. This result is important. The same heat transfer into two perfect engines produces different work outputs, because the entropy change differs in the two cases. In the second case, entropy is greater and less work is produced. Entropy is associated with the *un*availability of energy to do work.

**Figure \(\PageIndex{4}\):** (a) A Carnot engine working at between 600 K and 100 K has 4000 J of heat transfer and performs 3333 J of work. (b) The 4000 J of heat transfer occurs first irreversibly to a 250 K reservoir and then goes into a Carnot engine. The increase in entropy caused by the heat transfer to a colder reservoir results in a smaller work output of 2400 J. There is a permanent loss of 933 J of energy for the purpose of doing work.

When entropy increases, a certain amount of energy becomes *permanently* unavailable to do work. The energy is not lost, but its character is changed, so that some of it can never be converted to doing work—that is, to an organized force acting through a distance. For instance, in the previous example, 933 J less work was done after an increase in entropy of 9.33 J/K occurred in the 4000 J heat transfer from the 600 K reservoir to the 250 K reservoir. It can be shown that the amount of energy that becomes unavailable for work is

\[W_{unavail} = \Delta S \cdot T_0,\]

where \(T_0\) is the lowest temperature utilized. In the previous example,

\[W_{unavail} = (9.33 \, J/K)(100 \, K) = 933 \, J\]

# Heat Death of the Universe: An Overdose of Entropy

In the early, energetic universe, all matter and energy were easily interchangeable and identical in nature. Gravity played a vital role in the young universe. Although it may have *seemed* disorderly, and therefore, superficially entropic, in fact, there was enormous potential energy available to do work—all the future energy in the universe.

As the universe matured, temperature differences arose, which created more opportunity for work. Stars are hotter than planets, for example, which are warmer than icy asteroids, which are warmer still than the vacuum of the space between them.

Most of these are cooling down from their usually violent births, at which time they were provided with energy of their own—nuclear energy in the case of stars, volcanic energy on Earth and other planets, and so on. Without additional energy input, however, their days are numbered.

As entropy increases, less and less energy in the universe is available to do work. On Earth, we still have great stores of energy such as fossil and nuclear fuels; large-scale temperature differences, which can provide wind energy; geothermal energies due to differences in temperature in Earth’s layers; and tidal energies owing to our abundance of liquid water. As these are used, a certain fraction of the energy they contain can never be converted into doing work. Eventually, all fuels will be exhausted, all temperatures will equalize, and it will be impossible for heat engines to function, or for work to be done.

Entropy increases in a closed system, such as the universe. But in parts of the universe, for instance, in the Solar system, it is not a locally closed system. Energy flows from the Sun to the planets, replenishing Earth’s stores of energy. The Sun will continue to supply us with energy for about another five billion years. We will enjoy direct solar energy, as well as side effects of solar energy, such as wind power and biomass energy from photosynthetic plants. The energy from the Sun will keep our water at the liquid state, and the Moon’s gravitational pull will continue to provide tidal energy. But Earth’s geothermal energy will slowly run down and won’t be replenished.

But in terms of the universe, and the very long-term, very large-scale picture, the entropy of the universe is increasing, and so the availability of energy to do work is constantly decreasing. Eventually, when all stars have died, all forms of potential energy have been utilized, and all temperatures have equalized (depending on the mass of the universe, either at a very high temperature following a universal contraction, or a very low one, just before all activity ceases) there will be no possibility of doing work.

Either way, the universe is destined for thermodynamic equilibrium—maximum entropy. This is often called the *heat death of the universe*, and will mean the end of all activity. However, whether the universe contracts and heats up, or continues to expand and cools down, the end is not near. Calculations of black holes suggest that entropy can easily continue for at least \(10^{100}\) years.

# Order to Disorder

Entropy is related not only to the unavailability of energy to do work—it is also a measure of disorder. This notion was initially postulated by Ludwig Boltzmann in the 1800s. For example, melting a block of ice means taking a highly structured and orderly system of water molecules and converting it into a disorderly liquid in which molecules have no fixed positions. (See Figure \(\PageIndex{5}\).) There is a large increase in entropy in the process, as seen in the following example.

Example \(\PageIndex{3}\): Entropy Associated with Disorder

Find the increase in entropy of 1.00 kg of ice originally at \(0^oC\), that is melted to form water at \(0^oC\).

**Strategy**

As before, the change in entropy can be calculated from the definition of \(\Delta S\) once we find the energy \(Q\) needed to melt the ice.

**Solution**

The change in entropy is defined as: \[\Delta S = \dfrac{Q}{T}.\]

Here \(Q\) is the heat transfer necessary to melt 1.00 kg of ice and is given by \[Q = mL_f,\] where \(m\) is the mass and \(L_f\) is the latent heat of fusion. \(L_f = 334 \, kJ/kg\) for water, so that \[Q = 1.00 \, kg)(334 \, kJ/kg) = 3.34 \times 10^5 \, J.\]

Now the change in entropy is positive, since heat transfer occurs into the ice to cause the phase change; thus, \[\Delta S = \dfrac{Q}{T} = \dfrac{3.34 \times 10^5 \, J}{T}.\] \(T\) is the melting temperature of ice. That is \(T = 0^oC = 273 \, K\). So the change in entropy is \[\Delta S = \dfrac{3.34 \times 10^5 \, J}{273 \, K}\]\[ = 1.22 \times 10^3 \, J/K.\]

**Discussion**

This is a significant increase in entropy accompanying an increase in disorder.

**Figure \(\PageIndex{5}\):** When ice melts, it becomes more disordered and less structured. The systematic arrangement of molecules in a crystal structure is replaced by a more random and less orderly movement of molecules without fixed locations or orientations. Its entropy increases because heat transfer occurs into it. Entropy is a measure of disorder.

In another easily imagined example, suppose we mix equal masses of water originally at two different temperatures, say \(20.0^oC\) and \(40.0^oC\). The result is water at an intermediate temperature of \(30.0^oC\). Three outcomes have resulted: entropy has increased, some energy has become unavailable to do work, and the system has become less orderly. Let us think about each of these results.

First, entropy has increased for the same reason that it did in the example above. Mixing the two bodies of water has the same effect as heat transfer from the hot one and the same heat transfer into the cold one. The mixing decreases the entropy of the hot water but increases the entropy of the cold water by a greater amount, producing an overall increase in entropy.

Second, once the two masses of water are mixed, there is only one temperature—you cannot run a heat engine with them. The energy that could have been used to run a heat engine is now unavailable to do work.

Third, the mixture is less orderly, or to use another term, less structured. Rather than having two masses at different temperatures and with different distributions of molecular speeds, we now have a single mass with a uniform temperature.

These three results—entropy, unavailability of energy, and disorder—are not only related but are in fact essentially equivalent.

# Life, Evolution, and the Second Law of Thermodynamics

Some people misunderstand the second law of thermodynamics, stated in terms of entropy, to say that the process of the evolution of life violates this law. Over time, complex organisms evolved from much simpler ancestors, representing a large decrease in entropy of the Earth’s biosphere. It is a fact that living organisms have evolved to be highly structured, and much lower in entropy than the substances from which they grow. But it is *always* possible for the entropy of one part of the universe to decrease, provided the total change in entropy of the universe increases. In equation form, we can write this as

\[\Delta S_{tot} = \Delta S_{syst} + \Delta S_{envir} > 0.\]

Thus \(\Delta S_{yst}\) can be negative as long as \(\Delta S_{envir}\) is positive and greater in magnitude.

How is it possible for a system to decrease its entropy? Energy transfer is necessary. If I pick up marbles that are scattered about the room and put them into a cup, my work has decreased the entropy of that system. If I gather iron ore from the ground and convert it into steel and build a bridge, my work has decreased the entropy of that system. Energy coming from the Sun can decrease the entropy of local systems on Earth—that is, \(\Delta S_{syst}\) is negative. But the overall entropy of the rest of the universe increases by a greater amount—that is, \(\Delta S_{envir}\) is positive and greater in magnitude. Thus, \(\Delta S_{tot} = \Delta S_{syst} + \Delta S_{envir} > 0 \), and the second law of thermodynamics is *not* violated.

Every time a plant stores some solar energy in the form of chemical potential energy, or an updraft of warm air lifts a soaring bird, the Earth can be viewed as a heat engine operating between a hot reservoir supplied by the Sun and a cold reservoir supplied by dark outer space—a heat engine of high complexity, causing local decreases in entropy as it uses part of the heat transfer from the Sun into deep space. There is a large total increase in entropy resulting from this massive heat transfer. A small part of this heat transfer is stored in structured systems on Earth, producing much smaller local decreases in entropy. (See Figure \(\PageIndex{6}\).)

**Figure \(\PageIndex{6}\):** Earth’s entropy may decrease in the process of intercepting a small part of the heat transfer from the Sun into deep space. Entropy for the entire process increases greatly while Earth becomes more structured with living systems and stored energy in various forms.

PHET EXPLORATIONS: REVERSIBLE REACTIONS

Watch a reaction proceed over time. How does total energy affect a reaction rate? Vary temperature, barrier height, and potential energies. Record concentrations and time in order to extract rate coefficients. Do temperature dependent studies to extract Arrhenius parameters. This simulation is best used with teacher guidance because it presents an analogy of chemical reactions.

**\(\PageIndex{7}\): **Reversible Reaction

# Summary

- Entropy is the loss of energy available to do work.
- Another form of the second law of thermodynamics states that the total entropy of a system either increases or remains constant; it never decreases.
- Entropy is zero in a reversible process; it increases in an irreversible process.
- The ultimate fate of the universe is likely to be thermodynamic equilibrium, where the universal temperature is constant and no energy is available to do work.
- Entropy is also associated with the tendency toward disorder in a closed system.

## Glossary

**entropy**- a measurement of a system's disorder and its inability to do work in a system

**change in entropy**- the ratio of heat transfer to temperature \(Q/T\)

**second law of thermodynamics stated in terms of entropy**- the total entropy of a system either increases or remains constant; it never decrease

## Contributors

Paul Peter Urone (Professor Emeritus at California State University, Sacramento) and Roger Hinrichs (State University of New York, College at Oswego) with Contributing Authors: Kim Dirks (University of Auckland) and Manjula Sharma (University of Sydney). This work is licensed by OpenStax University Physics under a Creative Commons Attribution License (by 4.0).