Monday, August 5, 2019
Entropy And The Second Law Of Thermodynamics
Entropy And The Second Law Of Thermodynamics The paper examines, explain clearly, rigorously the term entropy, then discuss and evaluate its meaning in the context of the second law of thermodynamics. Also It will give a historical overview of the term entropy and it will give some examples which are taken from the daily life and with these, I will try to explain clearly the term entropy and its intention not only in the context of the second law and also its results in our daily life. 2. Introduction (Appendices A.) The term entropy has some related definitions. The first definition used by the German physicist Rodolf Julius Clausius in the 1850s and 1860s, he did that to state the second law of thermodynamics. The word entropy has been taken from the Greek word Ãâà à ¿Ãâ¬Ã · which means transformation. Also just as the first law of thermodynamics leads to the definition of energy as a property of a system, so the second law, in the form of Clausius inequality, leads to the definition of a new property of fundamental importance. This property is entropy. In the 1870s the term entropy is given by J. Willard Gibbs. The meaning of what he says is that the entropy shows the uncertainly about the state of a system. The latter can be defining from the probability distribution of its micro-states which demostrates, all molecular details about the system such as the position and the velocity of every molecule. If Pi is the possibility of a micro-state i, then the entropy of the system can be ex pressed by S = -k à £ Pi ln Pi Where k is the Boltzmann constant equal to 1.38062 x10^(âËâ23) joule/kelvin. Another definition, is the statistical definition developed by Ludwig Boltzmann in 1870s. This definition, describes the entropy as a measure of the number of possible microscopic configurations of the individual atoms, and molecules of the system; which would give rise to the observed macroscopic state of the system. In statistical thermodynamics, Boltzmanns equation, is a possibility equation relating the Entropy S of an ideal gas to quantity W, which is the number of micro-states corresponding to a given macro-state: S = k log W Where k is Boltzmanns equal to 1.38062 x10^ (âËâ23) joule/kelvin. Boltzmann has proved that the entropy of a given state of thermodynamic al system is connected by a simple relationship to the probability of the state. According to M. Kostic(2004): Entropy is an integral measure of (random) thermal energy redistribution (due to heat transfer or irreversible heat generation) within a system mass and/or space (during system expansion), per absolute temperature level. Entropy is increasing from perfectly-ordered (singular and unique) crystalline structure at zero absolute temperature (zero reference) during reversible heating (entropy transfer) and entropy generation during irreversible energy conversion (lost of work-potential to thermal energy), i.e. energy degradation or random equip-partition within system material structure and space per absolute temperature level. 3. Entropy measures the disorder in a system (Appendices B.) Therefore, metaphorically if a small bookshelf getting disorganized, it will be increasing the entropy of the bookshelf. Because, when the bookshelf is properly organized, finding a book is predictable and easy because all books are in a nice order. As the bookshelf is getting disorganized, the chance of not finding a book increasing, as a result is much higher. So that, when a bookshelf, a room a house are organized and they are moved from being organized to being disorganized, they generate Entropy. Also, liquids have higher entropy than crystals intuitively because their atomic positions are less orderly. Calculating the entropy of mixing illustrates this interpretation. An example is with scrambling eggs because when we mix the yolk and the white we cannot re-separate after. An example from this situation are given in figures 1.1 and 1,2. V V 2V Fig. 1.1 Unmixed atoms. The premixed Fig. 1.2 Mixed atoms. The mixed state: N/2 state: N/2 white atoms on one side, N/2 mixed atoms and N/2 black atoms scattered black atoms on the other. Through the volume, 2V. Fig. 1.1 There are N/2 undistinguished ideal gas white atoms on one side and N/2 undistinguished gas black atoms on the other side. As a result, the entropy of this system: Sunmixed = 2kB log[V N/2/(N/2)] Twice the configurational entropy of N/2 undistinguished atoms in a volume V. We assume that the black and white atoms have the same masses and the same total energy. Now the entropy change when the partition is removed, as a result from the scrambling and the two sets of atoms allowed mixing. Because, the temperatures and pressures from the both sides are equal and when the partition removing does not involve any heat transfer, and the entropy change to the mixing of the white and black atoms. In desegregated state, the entropy has increased to Smixed = 2kB log[(2V )N/2/(N/2)] and it is: ÃâSmixing = Smixed âËâ Sunmixed = =2kB log[[V^N/2/(N/2)] / [(2V)^N/2/(N/2)] = =kB log 2N = NkB log 2 So that, it gain kB log 2 in entropy every time we place an atom into one of the boxes. James P. Sethna (2006) Furthermore, we can give another example which shows us that entropy measures the disorder in a system: Which is more disorder? The glass of ice chips or the glass of water? For a glass of water, the number of molecules is astronomical. The ice chips probable look more disorder when we compare to the glass of water which looks uniform. However, according to thermodynamics the ice chips place limits on the number of ways the molecules can be arranged. The water molecules in the glass can be arranged in many more ways; as a result, they have grater multiplicity and therefore greater entropy. 4. Entropy measures our ignorance in a system The most general is to measure our ignorance about a system. The equilibrium state of a system, maximizes the entropy because, we have lost all information about the initial conditions, as a result, the entropy maximizing immediately maximises and our ignorance about the details of the system. 5. Entropy measures the multiplicity of a system The probability of finding a system in a given state depends upon the multiplicity of that state. As a result it is proportional to the number of ways someone can produce that state. Here, it is a pair of dices, and in throwing this pair, that measurable property is the sum of the number of dots which are facing on the top. The multiplicity for two dots showing is just one because there is only one case of the pair that will give that state. For example, the multiplicity for seven dots is six, because there is six cases of the pair that will show a total of seven dots. Probable one way to define the quantity entropy is to do it in terms of the multiplicity. Multiplicity = W Entropy = k lnW Where K is Boltzmanns constant. For a system, of a large number of particles. We can expect that the system at equilibrium will be found in the state of highest multiplicity since the fluctuations from that the state will usually be extremely small to measure. As a result, as a large system approaches equilibrium, its multiplicity therefore, entropy tends obviously to increase. This is one way of stating the Second Law of Thermodynamic. 6. The Second Law of Thermodynamics (Appendices C.) The second law of thermodynamics states that heat flows always from the warmer to colder bodies and never opposite. This is a common experience which everyone has seen and probably every day we have a case of those. For example, whenever we leave a cup of warm coffee it will become cool in 10 minutes. The special point of this process is that by the end of years can never become backward. It has just one direction as time passes. Indeed, through our everyday experience know that when contacting a hot and a cold body will be transferred heat from the hot to the cold body, so the hot body will be a little cooler and the cold body the opposite will be a little bit hotter. However, it is never possible as the time passes and the two bodies are in contact the cold body to be colder and the hot body to be hotter, for example, if we put an ice-cube into our drink, the drink does not boil. Therefore, it is only one direction in the flow heat which if we displaced it with a line, then this li ne will show everything from the past to now and to future. The second law of thermodynamics states that heat cannot be transferred from a colder to a hotter body within a system net changes occurring in other bodies within that system, in any irreversible process, entropy always increases. In nowadays, it is customary to use the term entropy in conjunction with the second law of thermodynamic. Consequently the entropy indicates the unavailable energy of a system, according to the law the entropy of a closed system can never reduce. Another form of the second law thermodynamic says that the minimum amount of heat which exchange a system during a change, which takes place at constant temperature T, associated with a change which is called entropy, with the equation: dQ=
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment
Note: Only a member of this blog may post a comment.