Quantity Catalog - Entropy, Heat capacity
Home Refresh UoMs for Quantitites > Composite Quantity UoMs > Entropy, Heat capacity

Entropy is a measure of the amount of disorder or randomness in a system, indicating the system's tendency towards equilibrium.

Heat capacity is the amount of heat energy required to raise the temperature of a substance by a certain amount. It is a measure of how much heat a substance can absorb before its temperature changes.

  • Entropy helps understand the behavior of systems in thermodynamics
  • Heat capacity is crucial in determining the amount of energy needed for heating or cooling processes
  (entropy base)  m^2*kg/s^2/K

Entropy is a measure of disorder or randomness in a system, while heat capacity is the amount of heat energy required to change the temperature of a substance.

  boltzmann constant  k_b

1 k_B = 1.380649e-23 m^2*kg/s^2/K (#5 of 7 SI Defining constants); The physical unit "k_b" is the Boltzmann constant, a fundamental constant in physics related to entropy and heat capacity. It is one of the seven SI defining constants.


Home Refresh UoMs for Quantitites > Composite Quantity UoMs > Entropy, Heat capacity