OK, can someone please just stop the philosophy and give numerical predictions of how entropy helps you predict the future?
The Unexpected Side of Entropy by Daan Frenkel
. Source. 2021.The Biggest Ideas in the Universe | 20. Entropy and Information by Sean Carroll (2020)
Source. In usual Sean Carroll fashion, it glosses over the subject. This one might be worth watching. It mentions 4 possible definitions of entropy: Boltzmann, Gibbs, Shannon (information theory) and John von Neumann (quantum mechanics).- www.quantamagazine.org/what-is-entropy-a-measure-of-just-how-little-we-really-know-20241213/ What Is Entropy? A Measure of Just How Little We Really Know. on Quanta Magazine attempts to make the point that entropy is observer dependant. TODO details on that.
Entropy is a concept that appears in various fields, such as thermodynamics, information theory, and statistical mechanics. Its meaning can vary slightly depending on the context, but generally, it refers to a measure of disorder, uncertainty, or randomness in a system. 1. **Thermodynamics**: In thermodynamics, entropy is a measure of the amount of energy in a physical system that is not available to do work. It is often associated with the degree of disorder or randomness in a system.
New to topics? Read the docs here!