42

Context

In one of his most popular books Guards! Guards!, Terry Pratchett makes an entropy joke:

Knowledge equals Power, which equals Energy, which equals Mass

Pratchett is a fantasy comedian and every third phrase in his book is a joke, therefore there is no good reason to believe it. Pratchett uses that madness to make up that a huge library has a tremendous gravitational push.

The question

I work with computers and mostly with encryption. My work colleagues believe Terry Pratchett's statement because of entropy. On the other hand, I believe, it is incorrect since entropy of information is a different entropy than the one used in thermodynamics.

Am I correct? And if so, why do use the same name (entropy) to mean two different things?

Also, what would be a good way to explain that these two "entropies" are different things to non-scientists (i.e. people without a chemistry or physics background)?

Qmechanic
  • 201,751
grochmal
  • 539
  • 7
    Maxwell's Demon thought experiment proves you wrong. As shown by Landauer, the fact that reversible computations and interaction free measurements are possible means that you can lower the entropy of a thermodynamic system as the price of increasing the information entropy of the computer's memory by at least the same amount. – Count Iblis Jun 17 '16 at 21:26
  • 2
    Trying to identify in exactly which ways a statement that was never meant to hold true in the scientific sense fails is a fruitless exercise. Go and enjoy Pratchett's wonderful books instead of debating the scientific truth of unscientific statements. – ACuriousMind Jun 17 '16 at 21:26
  • 17
    von Neumann to Shannon: "You should call it entropy, for two reasons. In the first place your uncertainty function has been used in statistical mechanics under that name, so it already has a name. In the second place, and more important, nobody knows what entropy really is, so in a debate you will always have the advantage." – lemon Jun 17 '16 at 21:42
  • 2
    @lemon: You will have the advantage until someone says $dS=dQ_{rev}/T$ and then you either understand thermodynamics for good or you are out of arguments. – CuriousOne Jun 17 '16 at 22:44
  • 9
    The difference depends upon whether you have knowledge of prepared quantum states of a system or not. For classical thermodynamics you have no such knowledge, so entropy is "thermal." If you have knowledge of microstates or quantum states then it is information entropy or entanglement entropy. In the end thermal entropy is information entropy, and all of this is ultimately entanglement entropy. – Lawrence B. Crowell Jun 17 '16 at 23:42
  • 1
    @CountIblis: However, the memory in those systems is a thermodynamic system as well, so you are really just building a thermodynamic/computational Rube-Goldberg machine. – CuriousOne Jun 18 '16 at 00:36
  • 1
    @CountIblis - Reading on Maxwell's Demon thought experiment gave me a new understanding of entropy, thanks. That though experiment is exactly what i needed. On the other hand, I have the impression that I have opened a pandora's box asking this question here. I am now very curious what are the thoughts of different physicists here about information vs. thermal entropy. – grochmal Jun 18 '16 at 18:40
  • 1
    @grochmal It's an interesting question, to be sure. One can put thermodynamic entropy first and call information entropy a subset, since it's always stored in the positions of atoms. Or, one can put information entropy first, and say that atomic positions are just a type of information. I am neutral, but people with more physics background prefer the former while those with more CS background prefer the latter. – knzhou Jun 18 '16 at 23:54
  • 2
    As you saw in the huge comment thread, it turns pretty partisan, I think because there's a point of pride for one's entire field in it -- i.e. people use it to say "all CS/physics is really just physics/CS". – knzhou Jun 18 '16 at 23:56
  • From a neutral standpoint I think there's no big deal. Letting the analogy go either way gives you more ways to think about things, it's a win-win. – knzhou Jun 18 '16 at 23:56
  • " My work colleagues believe Terry Pratchett's statement because of entropy." This is puzzling. Could you elaborate how the concept of entropy (whichever) hints the validity of the quoted statement? – Ján Lalinský Jun 19 '16 at 18:55
  • 1
    @ Ján: that's not puzzling at all. As von Neumann told Shannon when Shannon was asking what to call this quantity "You should call it entropy, for two reasons. In the first place you uncertainty function has been used in statistical mechanics under that name. In the second place, and more importantly, no one knows what entropy really is, so in a debate you will always have the advantage." So with entropy, you can never lose an argument. – Peter Shor Jun 19 '16 at 19:32
  • 3
    Knowledge is power, time is money. Power = Work / time, hence Knowledge = Work / money. The more you work and the less you're paid, the more you'll learn! – Lior Kogan Aug 04 '18 at 15:21

5 Answers5

31

So Pratchett's quote seems to be about energy, rather than entropy. I supposed you could claim otherwise if you assume "entropy is knowledge," but I think that's exactly backwards: I think that knowledge is a special case of low entropy. But your question is still interesting.

The entropy $S$ in thermodynamics is related to the number of indistinguishable states that a system can occupy. If all the indistinguishable states are equally probable, the number of "microstates" associated with a system is $\Omega = \exp( S/k )$, where the constant $k\approx\rm25\,meV/300\,K$ is related to the amount of energy exchanged by thermodynamic systems at different temperatures.

The canonical example is a jar of pennies. Suppose I drop 100 coins on the floor. There are 100 ways that I can have one heads-up and the rest tails-up; there are $100\cdot99/2$ ways to have two heads; there are $10 \cdot99\cdot98/6$ ways to have three heads; there are about $10^{28}$ ways to have forty heads, and $10^{29}$ ways to have fifty heads. If you drop a jar of pennies you're not going to find them 3% heads up, any more than you're going to get struck by lightning while you're dealing yourself a royal flush: there are just too many other alternatives.

The connection to thermodynamics comes when not all of my microstates have the same energy, so that my system can exchange energy with its surroundings by having transitions. For instance, suppose my 100 pennies aren't on the floor of my kitchen, but they're in the floorboard of my pickup truck with the out-of-balance tire. The vibration means that each penny has a chance of flipping over, which will tend to drive the distribution towards 50-50. But if there is some other interaction that makes heads-up more likely than tails-up, then 50-50 isn't where I'll stop. Maybe I have an obsessive passenger who flips over all the tails-up pennies. If the shaking and random flipping over is slow enough that he can flip them all, that's effectively "zero temperature"; if the shaking and random flipping is so vigorous that a penny usually flips itself before he corrects the next one, that's "infinite temperature." (This is actually part of the definition of temperature.)

The Boltzmann entropy I used above, $$ S_B = k_B \ln \Omega, $$ is exactly the same as the Shannon entropy, $$ S_S = k_S \ln \Omega, $$ except that Shannon's constant is $k_S = \frac1{\ln 2}\rm\,bit$, so that a system with ten bits of information entropy can be in any one of $\Omega=2^{10}$ states.

This is a statement with physical consequences. Suppose that I buy a two-terabyte SD card (apparently the standard supports this) and I fill it up with forty hours of video of my guinea pigs turning hay into poop. By reducing the number of possible states of the SD card from $\Omega=2\times2^{40}\times8$ to one, Boltzmann's definition tells me I have reduced the thermodynamic entropy of the card by $\Delta S = 2.6\rm\,meV/K$. That entropy reduction must be balanced by an equal or larger increase in entropy elsewhere in the universe, and if I do this at room temperature that entropy increase must be accompanied by a heat flow of $\Delta Q = T\Delta S = 0.79\rm\,eV = 10^{-19}\,joule$.

And here we come upon practical, experimental evidence for one difference between information and thermodynamic entropy. Power consumption while writing an SD card is milliwatts or watts, and transferring my forty-hour guinea pig movie will not be a brief operation --- that extra $10^{-19}\rm\,J$, enough energy to drive a single infrared atomic transition, that I have to pay for knowing every single bit on the SD card is nothing compared to the other costs for running the device.

The information entropy is part of, but not nearly all of, the total thermodynamic entropy of a system. The thermodynamic entropy includes state information about every atom of every transistor making up every bit, and in any bi-stable system there will be many, many microscopic configurations that correspond to "on" and many, many distinct microscopic configurations that correspond to "off."


CuriousOne asks,

How comes that the Shannon entropy of the text of a Shakespeare folio doesn't change with temperature?

This is because any effective information storage medium must operate at effectively zero temperature --- otherwise bits flip and information is destroyed. For instance, I have a Complete Works of Shakespeare which is about 1 kg of paper and has an information entropy of about maybe a few megabytes.

This means that when the book was printed there was a minimum extra energy expenditure of $10^{-25}\rm\,J = 1\,\mu eV$ associated with putting those words on the page in that order rather than any others. Knowing what's in the book reduces its entropy. Knowing whether the book is sonnets first or plays first reduces its entropy further. Knowing that "Trip away/Make no stay/Meet me all by break of day" is on page 158 reduces its entropy still further, because if your brain is in the low-entropy state where you know Midsummer Night's Dream you know that it must start on page 140 or 150 or so. And me telling you each of these facts and concomitantly reducing your entropy was associated with an extra energy of some fraction of a nano-eV, totally lost in my brain metabolism, the mechanical energy of my fingers, the operation energy of my computer, the operation energy of my internet connection to the disk at the StackExchange data center where this answer is stored, and so on.

If I raise the temperature of this Complete Works from 300 k to 301 K, I raise its entropy by $\Delta S = \Delta Q/T = 1\,\rm kJ/K$, which corresponds to many yottabytes of information; however the book is cleverly arranged so that the information that is disorganized doesn't affect the arrangements of the words on the pages. If, however, I try to store an extra megajoule of energy in this book, then somewhere along its path to a temperature of 1300 kelvin it will transform into a pile of ashes. Ashes are high-entropy: it's impossible to distinguish ashes of "Love's Labours Lost" from ashes of "Timon of Athens."

The information entropy --- which has been removed from a system where information is stored --- is a tiny subset of the thermodynamic entropy, and you can only reliably store information in parts of a system which are effectively at zero temperature.


A monoatomic ideal gas of, say, argon atoms can also be divided into subsystems where the entropy does or does not depend temperature. Argon atoms have at least three independent ways to store energy: translational motion, electronic excitations, and nuclear excitations.

Suppose you have a mole of argon atoms at room temperature. The translational entropy is given by the Sackur-Tetrode equation, and does depend on the temperature. However the Boltzmann factor for the first excited state at 11 eV is $$ \exp\frac{-11\rm\,eV}{k\cdot300\rm\,K} = 10^{-201} $$ and so the number of argon atoms in the first (or higher) excited states is exactly zero and there is zero entropy in the electronic excitation sector. The electronic excitation entropy remains exactly zero until the Boltzmann factors for all of the excited states add up to $10^{-24}$, so that there is on average one excited atom; that happens somewhere around the temperature $$ T = \frac{-11\rm\,eV}{k}\ln 10^{-24} = 2500\rm\,K. $$ So as you raise the temperature of your mole of argon from 300 K to 500 K the number of excited atoms in your mole changes from exactly zero to exactly zero, which is a zero-entropy configuration, independent of the temperature, in a purely thermodynamic process.

Likewise, even at tens of thousands of kelvin, the entropy stored in the nuclear excitations is zero, because the probability of finding a nucleus in the first excited state around 2 MeV is many orders of magnitude smaller than the number of atoms in your sample.

Likewise, the thermodynamic entropy of the information in my Complete Works of Shakespeare is, if not zero, very low: there are a small number of configurations of text which correspond to a Complete Works of Shakespeare rather than a Lord of the Rings or a Ulysses or a Don Quixote made of the same material with equivalent mass. The information entropy ("Shakespeare's Complete Works fill a few megabytes") tells me the minimum thermodynamic entropy which had to be removed from the system in order to organize it into a Shakespeare's Complete Works, and an associated energy cost with transferring that entropy elsewhere; those costs are tiny compared to the total energy and entropy exchanges involved in printing a book.

As long as the temperature of my book stays substantially below 506 kelvin, the probability of any letter in the book spontaneously changing to look like another letter or like an illegible blob is zero, and changes in temperature are reversible.

This argument suggests, by the way, that if you want to store information in a quantum-mechanical system you need to store it in the ground state, which the system will occupy at zero temperature; therefore you need to find a system which has multiple degenerate ground states. A ferromagnet has a degenerate ground state: the atoms in the magnet want to align with their neighbors, but the direction which they choose to align is unconstrained. Once a ferromagnet has "chosen" an orientation, perhaps with the help of an external aligning field, that direction is stable as long as the temperature is substantially below the Curie temperature --- that is, modest changes in temperature do not cause entropy-increasing fluctuations in the orientation of the magnet. You may be familiar with information-storage mechanisms operating on this principle.

rob
  • 89,569
  • How comes that the Shannon entropy of the text of a Shakespeare folio doesn't change with temperature? – CuriousOne Jun 18 '16 at 05:26
  • 2
    @CuriousOne It will if you change the temperature enough; see edit. – rob Jun 18 '16 at 14:28
  • 1
    I am sorry, but that is complete nonsense. Shannon entropy doesn't include the physical terms that you can find in thermodynamics. It is simply, by definition, not sensitive to temperature. If you change the definition to "Shannon entropy = Thermodynamic entropy", it loses all of its utility, not to mention that it invites the question: why are you renaming something that has a name, already? – CuriousOne Jun 18 '16 at 18:35
  • 1
    I believe what I said was that the Shannon entropy is a tiny subset of the information entropy, and I gave a quantitative example of a temperature change which results in an irreversible increase in the information entropy of a storage system. You seem to be bothered that I can change the temperature of a book from, say, 250 K to 300 K without changing its information entropy; but that particular change is reversible, and reversible processes in thermodynamics are by definition isentropic. – rob Jun 18 '16 at 22:55
  • 1
    What I am bothered by is that people can't seem to understand simple definitions. Shannon entropy was and still is clearly defined over symbol sequences. It's an abstract measure of the information content of e.g. a document and as such extremely useful. Thermodynamics entropy is a physical measure of the movement of homogeneous dynamic systems. You can change the temperature of a book and that will change the entropy of the paper, but it won't change the Shannon entropy of the text in the book by a bit. It's really not that hard to grasp. There is no information in homogeneous systems, BTW. – CuriousOne Jun 19 '16 at 00:27
  • 2
    @CuriousOne Shannon entropy is an abstract measure of information content with physical consequences. In particular Shannon entropy is quantized, and your intuition about entropy from continuum thermodynamics may not apply to systems with quantized degrees of freedom. I've edited in another example. I'm learning quite a bit by addressing your comments, so thank you. – rob Jun 19 '16 at 14:40
  • 2
    There are no abstracts in physics. Everything in physics is measured quantities of real systems. Shannon entropy is a mathematical quantity on abstract systems. It makes perfect sense, but it is simply not the same thing as what we call entropy in thermodynamics. The things it refers to are not the same. – CuriousOne Jun 19 '16 at 20:25
  • 2
    And I've told you how to measure it (and acknowledged that it's hard). – rob Jun 19 '16 at 21:46
  • There is nothing hard about calculating Shannon entropy. Here is the code in Python: http://pythonfiddle.com/shannon-entropy-calculation/, http://stackoverflow.com/questions/15450192/fastest-way-to-compute-entropy-in-python (no guarantees that it works) . As for thermodynamics entropy... that requires some serious calorimetric heat measurement. – CuriousOne Jun 19 '16 at 21:50
  • 3
    The above answer is a 'high temperature' explanation of entropy, in the sense that it is verbose and has many loosely related examples. – TLDR Aug 26 '17 at 13:59
19

Formally, the two entropies are the same thing. The Gibbs entropy, in thermodynamics, is $$S = -k_B \sum p_i \ln p_i$$ while the Shannon entropy of information theory is $$H = -\sum p_i \log_2 p_i.$$ These are equal up to some numerical factors. Given a statistical ensemble, you can calculate its (thermodynamic) entropy using the Shannon entropy, then multiplying by constants.


However, there is a sense in which you're right. Often when people talk about Shannon entropy, they only use it to count things that we intuitively perceive as information. For example, one might say the entropy of a transistor, flipped to 'on' or 'off' with equal likelihood, is 1 bit.

But the thermodynamic entropy of the transistor is thousands, if not millions of times higher, because it counts everything, i.e. the configurations of all the atoms making up the transistor. (If you want to explain it to your programmer colleagues, say they're not counting whether each individual atom is "on" or "off".)

In general, the amount of "intuitive" information (like bits, or words in a book) is a totally negligible fraction of the total entropy. The thermodynamic entropy of a library is about the same as that of a warehouse of blank books.

knzhou
  • 101,976
  • Comments are not for extended discussion; this conversation has been moved to chat. – Manishearth Jun 18 '16 at 13:41
  • Take it to that chat room that @Manishearth set up or post another answer, folks. – dmckee --- ex-moderator kitten Jun 19 '16 at 18:11
  • Technically, one cannot compare the magnitude of the thermodynamic and information entropies of a transistor in a direct way unless one specifies the macrostates of the transitor. I agree, however, that one can perhaps not measure more than a few basic quantities for a transistor (perhaps resistance, inductance, etc.), leaving the space of possible microstates still incredibly large. Hence for all purposed we can say that thermodynamic entropy is higher than the information entropy. – PianoEntropy Aug 02 '16 at 09:30
  • 1
    Entropy is an anthropomorphic quantity, defined in terms of the macroscopic observables we choose to specify the system in. See a classic paper by ET Jaynes – PianoEntropy Aug 02 '16 at 09:31
10

To be honest, I believe this question is not really settled, or at least that there is not yet a consensus in the scientific community about what the answer is.

My understanding of the relation is, I think, slightly different than knzhou, rob, or CuriousOne. My understanding is that thermodynamic entropy can be thought of as a particular application of information entropy. In particular, one can apply the principles of information and informational entropy to ask how much one knows about the state of a quantum system, and under certain conditions the thermodynamic Boltzmann entropy seems to be recovered.

As a concrete example, a recent experiment related to this question (1) studies the "entanglement entropy" of an interacting quantum system, which is an application of informational entropy to a quantum state. Under the appropriate circumstances (looking at the single-particle density matrix of a thermalized quantum state), this informational entropy is shown to be identical to the thermodynamic Boltzmann entropy.

From this viewpoint, thermodynamics is "just" a particular application of informational principles. Of course, one can also apply informational principles to entirely different systems such as books and radio communications and so on. As a result, thermodynamic and informational entropies are not the same, but are two particular applications of the same general principle.

However, this opinion is by no means shared by all, and while this correspondence seems to work in some cases like the above experiment, it remains to be explained in a more general setting.

Two somewhat related questions that you might find interesting:

Spontaneous conversion of heat into work at negative temperatures

What are the phenomena responsible for irreversible increase in entropy?


Appendix: Entropy Hierarchy

Here is the hierarchy of entropies I am claiming here (ignoring constants like $k_B$):

  1. Shannon entropy: $S_\textrm{Shannon}=− \sum_i p_i \log p_i$ . Describes, roughly, how much one knows about the state of some system, with $i$ being the possible states. This system could be, for example, a string of binary bits.

  2. Applying this to an unknown quantum state, one gets the Gibbs entropy: $S_\textrm{Gibbs}=− \sum_i p_i \log p_i$, where the $i$ are now specifically the possible quantum states of the system. For this expression to make physical sense, $i$ must be the eigenstates of the system in a basis in which the density matrix is diagonal*. With this stipulation, $S_\textrm{Gibbs}$ is identical to the Von Neumann entropy of a quantum state: $S_\textrm{VN}=− \text{tr}(\rho \log \rho)$, with $\rho$ the density matrix.

  3. The entanglement entropy is simply an application of $S_\textrm{VN}$ to a particular spatial subset of a (usually isolated) system: $S_{EE,A}=− \text{tr}(\rho_A \log \rho_A)$, where $\rho_A$ is the density matrix resulting from the partial trace over the density matrix of a large system, keeping only some local subsystem. In other words, it is the entropy of a particular part of some larger system.

  4. The highly nontrivial claim made in (1) (and elsewhere) is that for a thermalized system, the $S_{EE,A}$ of a small local subsystem $\rho_A$ is equivalent to the Boltzmann thermodynamic entropy, defined as: $S_\textrm{Boltz}=-\sum_i(p_{i,\textrm{th}} \log p_{i,\textrm{th}}) $, with $p_{i,\textrm{th}}=\frac{e^{-E_i/k_B T}}{\sum_i e^{-E_i/k_B T}}$, $i$ as the possible states of $\rho_A$, and $k_B T$ chosen so that the system has the correct average energy. This claim is known, by the way, as the "eigenstate thermalization hypothesis."

*There's nothing too mysterious about this requirement: it is simply because for entropy to have some "nice" properties like additivity the state $i$ must be uncorrelated.

Rococo
  • 7,671
  • 1
  • 23
  • 54
6

Edwin T. Jaynes in The minimum energy production principle E. T. Jaynes points out that there exist different non-equivalent definitions of entropy, with somewhat different properties, which are often confused:

By far the most abused word in science is "entropy". Confusion over the different meanings of this word, already serious 35 years ago, reached disaster proportions with the 1948 advent of Shannon's information theory, which not only appropriated the same word for a new set of meanings; but even worse, proved to be highly relevant to statistical mechanics.

He then proceeds to list six different types of entropy:

The "entropies" with which we shall be concerned here are of a totally different nature. First is the experimental entropy $S_E$ of Clausius, Gibbs and G. N. Lewis, which is by construction a function $S_E(T; P, M,...)$ of the observed macroscopic quantities. For us, as for them, the term "second law" refers to a property of $S_E$ observed in laboratory experiments. It is therefore, by definition, a proposition of macroscopic phenomenology. Whether it might be also a theorem in dynamics was answered negatively already by Gibbs (2) with a very vivid example of gas diffusion.

Second, we use the information entropy $S_I=-\sum p_i\log p_i$, a property of any probability distribution. In quantum theory, the $\{p_i\}$ are eigen-values of a density matrix $\rho$, and $S_I(\rho)=-\text{Tr}(\rho\log\rho)$.

If $S_I$ is maximized subject to certain constraints $\{A_1...A_n\}$, the maximum attained defines a third entropy $S(A_1...A_n)=(S_I)_{\max}$, which is a function of those constraints in many different ways, there are many different quantities $S(A)$, with different meanings. Just as Clausius' $S_E$ is undefined until we specify which macroscopic variables are to be used, one must also indicate in each case which constraints are used - and therefore become the independent variables - in defining $A(A)$. In our applications, the $\{A_i\}$ may be any macroscopic quantities about which we have some information.

To keep the distinctions clear, our $S_E$ is, as in conventional thermodynamics, a numerical multiple of Boltzmann's constant $k$, while $S_I$ and $S(A)$ are dimensionless, following SHannon. Being defined as the maximum in a constrained variational problem, $S(A)$ will have, like $S_E$, a tendency to increase whenever a constraint is removed, thus paralleling in our mathematics what is observed in the laboratory (12).

Many other entropies appear in the literature, among which we note the Boltzmann and Gibbs $S_B$, $S_G$ defined from single-particle and N-particle distribution functions, and the quantity $S_{BEP}=k\log W$ of Boltzmann, Einstein and Planck. The relations between these have been discussed in detail elsewhere (13a, b).

As should be evident, there is no possibility of finding the correct relations for irreversible processes, unless one understands clearly the distinctions in meaning and the different properties of $S_E$, $S_I$, $S(A)$, $S_B$, $S_G$, and $S_{BEP}$. We hardly expect hat the variational principle we seek can hold for all of them. While properties of $S_I$ and $S(A)$ are mathematical theorems, those of $S_E$ are summaries of experimental facts.

The entropies $S_B$ and $S_G$ not detailed above are explicitly spelled in Jaynes' other paper Gibbs vs. Boltzmann entropies in terms of many-particle and single-particle distribution functions $W_N(x_1,p_1; x_2, p_2;...;x_N,p_N;t)$ and $w_1(x_1,p_1;t)=\int W_Nd\tau_{N-1}$: $$ S_G=-\int W_N\log W_N d\tau,\\ S_B=-\int w_1\log w_1 d\tau_1 $$

The statistically defined entropy $S(A)$ is obtained in Jaynes' other paper Information theory and statistical mechanics.

Related posts:
How can Entropy be maximal when it is undefined everywhere else?
How to derive Shannon Entropy from Clausius Theorem?
Lagrange multipliers - isothermal-isobaric ensemble
Density of states misunderstanding in statistical mechanics
Is entropy a measure of how entangled a system is?

Roger V.
  • 58,522
5

So far there have been quite a few insightful answers about statistical mechanical entropy, but so far the only mention of thermodynamic entropy has been made by CuriousOne in the comments, so I thought it would be useful to give a short general reminder about the subtle difference between the notion of entropy in thermodynamics and the formulas that come up from statistical mechanics and coding theory.

One approach to understanding thermodynamic entropy is via the fundamental (or technological) constraints on the maximum attainable efficiency of heat engines. Volume 1 of the Feynman lectures has a section on thermodynamics that eloquently describes how the Carnot efficiency provides a universal temperature scale $T$ (up to an arbitrary choice of units), so that the quantity $\frac{d Q}{T}$ is the differential of a state function $S$ that is called entropy. Since it is essentially defined through the performance of heat engines, thermodynamic entropy is only sensitive to features of a system that are able to absorb heat and relax in ways that allow work to be extracted.

In this sense, information theoretical entropy is a measure of what features you are cognizant of*, while thermodynamic entropy you can think of as a measure of what features at small scales collectively influence systems at larger scales.

*Information theoretical entropy, and statistical mechanical entropy, are (by themselves) essentially just measures of volume for a space of possible outcomes.

TLDR
  • 3,158
  • I appreciate you bringing up the distinction between statistical and thermodynamics notions of entropy. However, this point has a natural response: the connection between these is quite thoroughly explored in standard textbooks. I follow Kittel and Kroemer, for example, in defining thermodynamic entropy through statistical mechanics (as $S_{Gibbs}$ or similar), and defining temperature as $1/ T=\partial S / \partial U$ (equivalent to the definition given in my answer), and using these as the basis for standard thermodynamics. – Rococo Jun 21 '16 at 00:09
  • 1
    I unfortunately do not have Kittel and Kroemer. However, if they define thermodynamic entropy the way you describe, then I would say that the connection between thermodynamics and statistical mechanics is not thoroughly explored in that book. However, Reif and Schroeder (as well as Feynman vol. 1) are careful to establish thermodynamics on solid conceptual footing independently of statistical mechanics. To my ears, defining thermodynamic entropy from $S_{Gibbs}$ sounds circular, and diminishes the brilliance of Boltzmann's equation. – TLDR Jun 21 '16 at 02:06
  • Well, where you see circular reasoning I see a particularly elegant example of reduction to general principles. But I'll have to look as those treatments. – Rococo Jun 21 '16 at 02:54