Utente:Grasso Luigi/sanbox1/Entropia di miscela

In termodinamica l'entropia di miscela (simbolo: ) è l'incremento dell'entropia totale quando diversi sistemi inizialmente separati con composizioni diverse, ciascuno in uno stato termodinamico di equilibrio interno, sono miscelati senza reazione chimica dall'operazione termodinamica di rimozione delle partizioni impermeabili tra loro, seguito da un tempo per la formazione di un nuovo stato termodinamico di equilibrio interno nel nuovo sistema chiuso non partizionato.

In generale, la miscela può essere vincolata a verificarsi in varie condizioni iniziali. Nelle condizioni iniziali normali, i materiali (sistemi) sono ad una temperatura e pressione identica, e il nuovo sistema può cambiare il suo volume, pur mantenendo costanti la temperatura, la pressione e le masse dei componenti chimici. Il volume disponibile per ciascun materiale aumenta, rispetto a quello del suo volume iniziale quando il materiale è separato, e il sistema raggiunge un volume finale comune. Il volume finale non necessariamente è la somma dei volumi iniziali dei materiali separati, in modo che del lavoro possa essere eseguito dal o sul nuovo sistema chiuso durante il processo di miscela, nonché del calore viene trasferito da o verso i contorni del sistema, a causa del mantenimento di pressione e temperatura costanti.

L'energia interna del nuovo sistema chiuso è uguale alla somma delle energie interne dei sistemi inizialmente separati. I valori di riferimento delle energie interne dovrebbero essere calcolati con tale vincolo, tenendo anche conto che le energie interne sono rispettivamente proporzionali alle masse dei sistemi.[1]

Per conciso, nel presente articolo, col termine miscela ideale s'intende una miscela omogenea nella fase gassosa (gas ideale) o nella fase di una soluzione liquida (soluzione ideale).

Nel caso speciale di miscela ideale, il volume finale comune è infatti la somma dei volumi dei sistemi separati o partizionati. Non c'è trasferimento di calore e nemmeno lavoro fatto. The entropy of mixing is entirely accounted for by the diffusive expansion of each material into a final volume not initially accessible to it.

Nel caso generale di miscela di sostanze non-ideali, however, the total final common volume may be different from the sum of the separate initial volumes, and there may occur transfer of work or heat, to or from the surroundings; also there may be a departure of the entropy of mixing from that of the corresponding ideal case. That departure is the main reason for interest in entropy of mixing. These energy and entropy variables and their temperature dependences provide valuable information about the properties of the materials.

On a molecular level, the entropy of mixing is of interest because it is a macroscopic variable that provides information about constitutive molecular properties. In ideal materials, intermolecular forces are the same between every pair of molecular kinds, so that a molecule feels no difference between other molecules of its own kind and of those of the other kind. In non-ideal materials, there may be differences of intermolecular forces or specific molecular effects between different species, even though they are chemically non-reacting. The entropy of mixing provides information about constitutive differences of intermolecular forces or specific molecular effects in the materials.

Il concetto statistico di casualità viene utilizzato per l'esposizione in meccanica statistica dell'entropia della miscela. La miscela di materiali ideali viene considerato un processo casuale a livello molecolare, e, di conseguenza, la miscela di materiali non ideali può considerarsi un processo non casuale.

Miscela ideale a T, P costanti

modifica

Nelle miscele ideali, le forze intermolecolari sono le stesse tra ogni coppia di tipi di molecole, così che una molecola non "sente" alcuna differenza tra se stessa e le sue vicine. Questo è il caso di riferimento per esaminare la corrispondente miscela non ideale.

Miscela 2-componenti ideale

modifica

Consideriamo il caso semplice di due componenti nella fase gassosa , alla stessa temperatura e pressione, inizialmente separati da una partizione divisoria.

Quando rimuoviamo la partizione che li separa, espandono in un volume finale comune (the sum of the two initial volumes), and the entropy of mixing   is given by

 .

where   is the gas constant,   the total number of moles and   the mole fraction of component  , which initially occupies volume  . After the removal of the partition, the   moles of component   occupa no un volume della miscela somma essendo ideale  , è ne risulta un aumento di entropia per ciascun componente pari a

 

In this case, the increase in entropy is entirely due to the irreversible processes of expansion of the two gases, and involves no heat or work flow between the system and its surroundings.

Miscela c-componenti ideale

modifica

Generalizzando ad un sistema termodinamico a q componenti:

 

Energia libera di Gibbs di miscela

modifica

The Gibbs free energy change   determines whether mixing at constant (absolute) temperature   and pressure   is a spontaneous process. Questa grandezza estensiva è la combinazione di due effetti fisici: l'entalpia di miscela, che misura il cambio di energia per sistemi chiusi a P, T costanti, e l'entropia di miscela.

Per miscele ideali, l'entalpia di miscela è nulla ( ), per cui l'energia libera di Gibbs di miscela viene relazionata al solo termine dell'entropia :

 

For an ideal solution, the Gibbs free energy of mixing is always negative, meaning that mixing of ideal solutions is always spontaneous. The lowest value is when the mole fraction is 0.5 for a mixture of two components, or 1/n for a mixture of n components.

[2]

Soluzioni e dipendenza dalla temperatura della miscibilità

modifica

Soluzioni ideali e regolari

modifica
 
L'entropia di miscela per soluzione ideale di due specie è massima quando la frazione molare di ciascuna specie è 0.5.

The above equation for the entropy of mixing of ideal gases is valid also for certain liquid (or solid) solutions—those formed by completely random mixing so that the components move independently in the total volume. Such random mixing of solutions occurs if the interaction energies between unlike molecules are similar to the average interaction energies between like molecules.[3][4] The value of the entropy corresponds exactly to random mixing for ideal solutions and for regular solutions, and approximately so for many real solutions.[4][5]

For binary mixtures the entropy of random mixing can be considered as a function of the mole fraction of one component.

 

For all possible mixtures,  , so that     and   are both negative and the entropy of mixing   is positive and favors mixing of the pure components.

Also the curvature of   as a function of   is given by the second derivative  

This curvature is negative for all possible mixtures  , so that mixing two solutions to form a solution of intermediate composition also increases the entropy of the system. Random mixing therefore always favors miscibility and opposes phase separation.

For ideal solutions, the enthalpy of mixing is zero so that the components are miscible in all proportions. For regular solutions a positive enthalpy of mixing may cause incomplete miscibility (phase separation for some compositions) at temperatures below the upper critical solution temperature (UCST).[6] This is the minimum temperature at which the   term in the Gibbs energy of mixing is sufficient to produce miscibility in all proportions.

Sistemi con una temperatura della soluzione critica più bassa

modifica

Nonrandom mixing with a lower entropy of mixing can occur when the attractive interactions between unlike molecules are significantly stronger (or weaker) than the mean interactions between like molecules. For some systems this can lead to a lower critical solution temperature (LCST) or lower limiting temperature for phase separation.

For example, triethylamine and water are miscible in all proportions below 19 °C, but above this critical temperature, solutions of certain compositions separate into two phases at equilibrium with each other.[7][8] This means that   is negative for mixing of the two phases below 19 °C and positive above this temperature. Therefore,   is negative for mixing of these two equilibrium phases. This is due to the formation of attractive hydrogen bonds between the two components that prevent random mixing. Triethylamine molecules cannot form hydrogen bonds with each other but only with water molecules, so in solution they remain associated to water molecules with loss of entropy. The mixing that occurs below 19 °C is due not to entropy but to the enthalpy of formation of the hydrogen bonds.

Lower critical solution temperatures also occur in many polymer-solvent mixtures.[9] For polar systems such as polyacrylic acid in 1,4-dioxane, this is often due to the formation of hydrogen bonds between polymer and solvent. For nonpolar systems such as polystyrene in cyclohexane, phase separation has been observed in sealed tubes (at high pressure) at temperatures approaching the liquid-vapor critical point of the solvent. At such temperatures the solvent expands much more rapidly than the polymer, whose segments are covalently linked. Mixing therefore requires contraction of the solvent for compatibility of the polymer, resulting in a loss of entropy.[9]

Termodinamica statistica dell'entropia di miscela di gas ideali

modifica

Since thermodynamic entropy can be related to statistical mechanics or to information theory, it is possible to calculate the entropy of mixing using these two approaches. Here we consider the simple case of mixing ideal gases.

Dimostrazione dalla meccanica statistica

modifica

Assume that the molecules of two different substances are approximately the same size, and regard space as subdivided into a square lattice whose cells are the size of the molecules. (In fact, any lattice would do, including close packing.) This is a crystal-like conceptual model to identify the molecular centers of mass. If the two phases are liquids, there is no spatial uncertainty in each one individually. (This is, of course, an approximation. Liquids have a "free volume". This is why they are (usually) less dense than solids.) Everywhere we look in component 1, there is a molecule present, and likewise for component 2. After the two different substances are intermingled (assuming they are miscible), the liquid is still dense with molecules, but now there is uncertainty about what kind of molecule is in which location. Of course, any idea of identifying molecules in given locations is a thought experiment, not something one could do, but the calculation of the uncertainty is well-defined.

We can use Boltzmann's equation for the entropy change as applied to the mixing process

 

where   is Boltzmann's constant. We then calculate the number of ways   of arranging   molecules of component 1 and   molecules of component 2 on a lattice, where

 

is the total number of molecules, and therefore the number of lattice sites. Calculating the number of permutations of   objects, correcting for the fact that   of them are identical to one another, and likewise for  ,

 

After applying Stirling's approximation for the factorial of a large integer m:

 ,

the result is  

where we have introduced the mole fractions, which are also the probabilities of finding any particular component in a given lattice site.

 

Since the Boltzmann constant  , where   is Avogadro's number, and the number of molecules  , we recover the thermodynamic expression for the mixing of two ideal gases,  

This expression can be generalized to a mixture of   components,  , with  

 

Relazione con la teoria dell'informazione

modifica

The entropy of mixing is also proportional to the Shannon entropy or compositional uncertainty of information theory, which is defined without requiring Stirling's approximation. Claude Shannon introduced this expression for use in information theory, but similar formulas can be found as far back as the work of Ludwig Boltzmann and J. Willard Gibbs. The Shannon uncertainty is not the same as the Heisenberg uncertainty principle in quantum mechanics which is based on variance. The Shannon entropy is defined as:

 

where pi is the probability that an information source will produce the i th symbol from an r-symbol alphabet and is independent of previous symbols. (thus i runs from 1 to r ). H is then a measure of the expected amount of information (log pi ) missing before the symbol is known or measured, or, alternatively, the expected amount of information supplied when the symbol becomes known. The set of messages of length N symbols from the source will then have an entropy of N*H.

The thermodynamic entropy is only due to positional uncertainty, so we may take the "alphabet" to be any of the r different species in the gas, and, at equilibrium, the probability that a given particle is of type i is simply the mole fraction xi for that particle. Since we are dealing with ideal gases, the identity of nearby particles is irrelevant. Multiplying by the number of particles N yields the change in entropy of the entire system from the unmixed case in which all of the pi were either 1 or 0. We again obtain the entropy of mixing on multiplying by the Boltzmann constant  .

 

So thermodynamic entropy with "r" chemical species with a total of N particles has a parallel to an information source that has "r" distinct symbols with messages that are N symbols long.

Applicazione ai gas

modifica

In gases there is a lot more spatial uncertainty because most of their volume is merely empty space. We can regard the mixing process as allowing the contents of the two originally separate contents to expand into the combined volume of the two conjoined containers. The two lattices that allow us to conceptually localize molecular centers of mass also join. The total number of empty cells is the sum of the numbers of empty cells in the two components prior to mixing. Consequently, that part of the spatial uncertainty concerning whether any molecule is present in a lattice cell is the sum of the initial values, and does not increase upon "mixing".

Almost everywhere we look, we find empty lattice cells. Nevertheless, we do find molecules in a few occupied cells. When there is real mixing, for each of those few occupied cells, there is a contingent uncertainty about which kind of molecule it is. When there is no real mixing because the two substances are identical, there is no uncertainty about which kind of molecule it is. Using conditional probabilities, it turns out that the analytical problem for the small subset of occupied cells is exactly the same as for mixed liquids, and the increase in the entropy, or spatial uncertainty, has exactly the same form as obtained previously. Obviously the subset of occupied cells is not the same at different times. But only when there is real mixing and an occupied cell is found do we ask which kind of molecule is there.

See also: Gibbs paradox, in which it would seem that "mixing" two samples of the same gas would produce entropy.

Applicazione alle soluzioni

modifica

If the solute is a crystalline solid, the argument is much the same. A crystal has no spatial uncertainty at all, except for crystallographic defects, and a (perfect) crystal allows us to localize the molecules using the crystal symmetry group. The fact that volumes do not add when dissolving a solid in a liquid is not important for condensed phases. If the solute is not crystalline, we can still use a spatial lattice, as good an approximation for an amorphous solid as it is for a liquid.

The Flory–Huggins solution theory provides the entropy of mixing for polymer solutions, in which the macromolecules are huge compared to the solvent molecules. In this case, the assumption is made that each monomer subunit in the polymer chain occupies a lattice site.

Note that solids in contact with each other also slowly interdiffuse, and solid mixtures of two or more components may be made at will (alloys, semiconductors, etc.). Again, the same equations for the entropy of mixing apply, but only for homogeneous, uniform phases.

Miscela sotto altri vincoli

modifica

Miscela con e senza modifica del volume disponibile

modifica

In the established customary usage, expressed in the lead section of this article, the entropy of mixing comes from two mechanisms, the intermingling and possible interactions of the distinct molecular species, and the change in the volume available for each molecular species, or the change in concentration of each molecular species. For ideal gases, the entropy of mixing at prescribed common temperature and pressure has nothing to do with mixing in the sense of intermingling and interactions of molecular species, but is only to do with expansion into the common volume.[10]

According to Fowler and Guggenheim (1939/1965),[11] the conflating of the just-mentioned two mechanisms for the entropy of mixing is well established in customary terminology, but can be confusing unless it is borne in mind that the independent variables are the common initial and final temperature and total pressure; if the respective partial pressures or the total volume are chosen as independent variables instead of the total pressure, the description is different.

Miscela con ciascun gas mantenuto a volume parziale costante, con variazione del volume totale

modifica

In contrast to the established customary usage, "mixing" might be conducted reversibly at constant volume for each of two fixed masses of gases of equal volume, being mixed by gradually merging their initially separate volumes by use of two ideal semipermeable membranes, each permeable only to one of the respective gases, so that the respective volumes available to each gas remain constant during the merge. Either one of the common temperature or the common pressure is chosen to be independently controlled by the experimenter, the other being allowed to vary so as to maintain constant volume for each mass of gas. In this kind of "mixing", the final common volume is equal to each of the respective separate initial volumes, and each gas finally occupies the same volume as it did initially.[12][13][14][15][16][17]

This constant volume kind of "mixing", in the special case of perfect gases, is referred to in what is sometimes called Gibbs' theorem.[12][15][17] It states that the entropy of such "mixing" of perfect gases is zero.

Miscela a volume totale costante e volume parziale variabile, con P variabile e T costante

modifica

An experimental demonstration may be considered. The two distinct gases, in a cylinder of constant total volume, are at first separated by two contiguous pistons made respectively of two suitably specific ideal semipermeable membranes. Ideally slowly and fictively reversibly, at constant temperature, the gases are allowed to mix in the volume between the separating membranes, forcing them apart, thereby supplying work to an external system. The energy for the work comes from the heat reservoir that keeps the temperature constant. Then, by externally forcing ideally slowly the separating membranes together, back to contiguity, work is done on the mixed gases, fictively reversibly separating them again, so that heat is returned to the heat reservoir at constant temperature. Because the mixing and separation are ideally slow and fictively reversible, the work supplied by the gases as they mix is equal to the work done in separating them again. Passing from fictive reversibility to physical reality, some amount of additional work, that remains external to the gases and the heat reservoir, must be provided from an external source for this cycle, as required by the second law of thermodynamics, because this cycle has only one heat reservoir at constant temperature, and the external provision of work cannot be completely efficient.[13]

Il paradosso di Gibbs: miscele di sostanze identiche e simili

modifica
  Lo stesso argomento in dettaglio: Paradosso di Gibbs.

For entropy of mixing to exist, the putatively mixing molecular species must be chemically or physically detectably distinct. Thus arises the so-called Gibbs paradox, as follows. If molecular species are identical, there is no entropy change on mixing them, because, defined in thermodynamic terms, there is no mass transfer, and thus no thermodynamically recognized process of mixing. Yet the slightest detectable difference in constitutive properties between the two species yields a thermodynamically recognized process of transfer with mixing, and a possibly considerable entropy change, namely the entropy of mixing.

The "paradox" arises because any detectable constitutive distinction, no matter how slight, can lead to a considerably large change in amount of entropy as a result of mixing. Though a continuous change in the properties of the materials that are mixed might make the degree of constitutive difference tend continuously to zero, the entropy change would nonetheless vanish discontinuously when the difference reached zero.[18]

From a general physical viewpoint, this discontinuity is paradoxical. But from a specifically thermodynamic viewpoint, it is not paradoxical, because in that discipline the degree of constitutive difference is not questioned; it is either there or not there. Gibbs himself did not see it as paradoxical. Distinguishability of two materials is a constitutive, not a thermodynamic, difference, for the laws of thermodynamics are the same for every material, while their constitutive characteristics are diverse.[19]

Though one might imagine a continuous decrease of the constitutive difference between any two chemical substances, physically it cannot be continuously decreased till it actually vanishes.[20] It is hard to think of a smaller difference than that between ortho- and para-hydrogen. Yet they differ by a finite amount. The hypothesis, that the distinction might tend continuously to zero, is unphysical. This is neither examined nor explained by thermodynamics. Differences of constitution are explained by quantum mechanics, which postulates discontinuity of physical processes.[21]

For a detectable distinction, some means should be physically available. One theoretical means would be through an ideal semi-permeable membrane.[14] It should allow passage, backwards and forwards, of one species, while passage of the other is prevented entirely. The entirety of prevention should include perfect efficacy over a practically infinite time, in view of the nature of thermodynamic equilibrium. Even the slightest departure from ideality, as assessed over a finite time, would extend to utter non-ideality, as assessed over a practically infinite time. Such quantum phenomena as tunneling ensure that nature does not allow such membrane ideality as would support the theoretically demanded continuous decrease, to zero, of detectable distinction. The decrease to zero detectable distinction must be discontinuous.

For ideal gases, the entropy of mixing does not depend on the degree of difference between the distinct molecular species, but only on the fact that they are distinct; for non-ideal gases, the entropy of mixing can depend on the degree of difference of the distinct molecular species. The suggested or putative "mixing" of identical molecular species is not in thermodynamic terms a mixing at all, because thermodynamics refers to states specified by state variables, and does not permit an imaginary labelling of particles. Only if the molecular species are different is there mixing in the thermodynamic sense.[22][23][24][25][26][27]

  1. ^ (EN) Ilya Prigogine, Introduction to Thermodynamics of Irreversible Processes, 3ª ed., John Wiley & Sons Inc, 1968, ISBN 978-0470699287.
  2. ^ entropia molare  
  3. ^ (EN) Peter Atkins e de Paula Julio, Physical Chemistry, 8ª ed., W.H. Freeman and Company, 2006, p. 149, ISBN 978-0-1987-0072-2.
  4. ^ a b (EN) K. Denbigh, The Principles of Chemical Equilibrium, 3ª ed., CUP, 1971, p. 432, ISBN 978-0521096553.
  5. ^ (EN) P.A. Rock, Chemical Thermodynamics. Principles and Applications, MacMillan, 1969, p. 263, ISBN 978-0024025104.
  6. ^ Atkins & de Paula (2006), page 186.
  7. ^ Atkins & de Paula (2006), page 187.
  8. ^ M.A. White, "Properties of Materials" (Oxford University Press 1999) p.175
  9. ^ a b Cowie, J.M.G. "Polymers: Chemistry and Physics of Modern Materials" (2nd edn, Blackie 1991) p.174-176
  10. ^ Bailyn (1994), page 273.
  11. ^ Fowler, R., Guggenheim, E.A. (1939/1965). Statistical Thermodynamics. A version of Statistical Mechanics for Students of Physics and Chemistry, Cambridge University Press, Cambridge UK, pages 163-164
  12. ^ a b Planck, M. (1897/1903). Treatise on Thermodynamics, translated with the author's sanction by Alexander Ogg, Longmans, Green and Co., London, Sections 235-236.
  13. ^ a b Partington, J.R. (1949), pp. 163–164.
  14. ^ a b Adkins (1968/1983), page 217.
  15. ^ a b Callen, H.B. (1960/1985). Thermodynamics and an Introduction to Thermostatistics, second edition, Wiley, New York, ISBN 981-253-185-8, pages 69-70.
  16. ^ Buchdahl, H.A. (1966). The Concepts of Classical Thermodynamics, Cambridge University Press, London, pages 170-171.
  17. ^ a b Iribarne, J.V., Godson, W.L. (1973/1981), Atmospheric Thermodynamics, second edition, D. Reidel, Kluwer Academic Publishers, Dordrecht, ISBN 90-277-1296-4, pages 48-49.
  18. ^ ter Haar & Wergeland (1966), p. 87.
  19. ^ Truesdell, C. (1969). Rational Thermodynamics: a Course of Lectures on Selected Topics, McGraw-Hill Book Company, New York, p. 6.
  20. ^ Partington, J.R. (1949), p. 164, who cites Larmor, J. (1929), Mathematical and Physical Papers, volume 2, Cambridge University Press, Cambridge UK, p. 99.
  21. ^ Landé, A. (1955). Foundations of Quantum Mechanics: a Study in Continuity and Symmetry, Yale University Press, New Haven, p.10.
  22. ^ Tolman, R.C. (1938). The Principles of Statistical Mechanics, Oxford University Press, Oxford, pages 626-628.
  23. ^ Adkins (1968/1983), pages 217–218.
  24. ^ Landsberg, P.T. (1978). Thermodynamics and Statistical Mechanics, Oxford University Press, Oxford, ISBN 0-19-851142-6, page 74.
  25. ^ Bailyn (1994), pages 274, 516-517.
  26. ^ Grandy, W.T., Jr (2008). Entropy and the Time Evolution of Macroscopic Systems, Oxford University Press, Oxford, ISBN 978-0-19-954617-6, pages 60-62.
  27. ^ Kondepudi, D. (2008). Introduction to Modern Thermodynamics, Wiley, Chichester, ISBN 978-0-470-01598-8, pages 197-199.

Bibliografia

modifica
  • Adkins, C.J. (1968/1983). Equilibrium Thermodynamics, third edition, McGraw-Hill, London, ISBN 0-521-25445-0.
  • Atkins, P.W., de Paula, J. (2006). Atkins' Physical Chemistry, eighth edition, W.H. Freeman, New York, ISBN 978-0-7167-8759-4.
  • Bailyn, M. (1994). A Survey of Thermodynamics, American Institute of Physics, New York, ISBN 0-88318-797-3.
  • Partington, J.R. (1949). An Advanced Treatise on Physical Chemistry, Volume 1, Fundamental Principles. The Properties of Gases, Longmans, Green, and Co., London.
  • ter Haar, D., Wergeland, H. (1966). Elements of Thermodynamics, Addison-Wesley Publishing, Reading MA.

Voci correlate

modifica

Collegamenti esterni

modifica
  NODES
Idea 46
idea 46
INTERN 6
Note 3