Once you do that, it's clear that you can't cancel "dx at constant z" with "dx at constant y" etc. And then the remaining logic works out nicely (see thermodynamics for a perfect application of this).
Also integrals with “integrate f(x) dx” where people treat “dx” as some number than can be manipulated, when it’s more just part of the notation “integrate_over_x f(x)”
Sigh. These are sadly some kind of right-of-passage, or mathematical hazing. Sad.
Baez is mixing partial derivatives with different variables treated as constants. Whole different ball game.
Does someone has a good explanation ?
Or you could do it in terms of temperature T and pressure, for instance, to obtain U ~ T (in this case there's no dependence on pressure).
The ideal gas laws let you transform between these choices. But the point is that the same physical quantity, U, has multiple mathematical functions underlying it - depending on which pair you choose to describe it with!
To disambiguate this physicists write stuff like (dU/dP)_T, which means "partial derivative of U wrt P, where we use the expression for U in terms of P and T". Note that this is not the same as (dU/dP)_V, despite the fact that it superficially looks like the same derivative! The former is 0 and the latter is ~V, which you can compute from the expressions I gave above.
The mistake is thinking that U is a single function of many independent variables P, T, S, V, etc. Actually these variables all depend on each other! So there are many possible functions corresponding to U in a formal sense, which is something people gloss over because U is a single physical quantity and it's convenient to use a single letter to denote it.
Maybe it would make more sense to use notation like U(T, P) and U(P, V) to make it clear that these are different functions, if you wanted to be super explicit.
So, in vector space terms, we have different bases for describing U in, but not that many independent variables.
If U is a function of x and y, but x and y are not orthogonal, then I can't treat dU/dx and dU/dy as independent, even for partial derivatives, because x and y aren't really independent.
(The Wikipedia page[1] has nice images of this [2])
The slope of this new 2D function on the x=3 plane at some point y is then the partial derivative ∂z/∂y for constant x at the point (3,y). As we are "fixing" the value of x to a constant, by only considering the intersection of our original function with a plane at x=x_0.
Consider f(x,y,z), let’s say f(x, y, z) = x^2 + 3y^3 - e^(-z). What’s the difference between "the partial derivative of f with respect to x" and "the partial derivative of f with respect to x at constant y" ? The first one is already at constant y !
In standard multivariate calculus, the partial derivative of f with respect to x , as you explained, is always "at constant y and z".
In thermodynamics, you can say things like "partial derivative of pressure with respect to volume" and add "at constant temperature" or "at constant entropy" and get different results. What ? Why ? How ?
In thermodynamics, there often isn't really one "best" choice of two coordinate functions among the many possibilities (pressure, temperature, volume, energy, entropy... these are the must common but you could use arbitrarily many others in principle), and it's natural to switch between these coordinates even within a single problem.
Coming back to the more familiar x, y, r, and θ, you can visualize these 4 coordinate functions by plotting iso-contours for each of them in the plane. Holding one of these coordinate functions constant picks out a curve (its iso-contour) through a given point. Derivatives involving the other coordinates holding that coordinate constant are ratios of changes in the other coordinates along this iso-contour.
For example, you can think of evaluating dr/dx along a curve of constant y or along a curve of constant θ, and these are different.
I first really understood this way of thinking from an unpublished book chapter of Jaynes [1]. Gibbs "Graphical Methods In The Thermodynamics of Fluids" [2] is also a very interesting discussion of different ways of representing thermodynamic processes by diagrams in the plane. His companion paper, "A method of geometrical representation of the thermodynamic properties of substances by means of surfaces" describes an alternative representation as a surface embedded in a larger space, and these two different pictures are complimentary and both very useful.
---
The fundamental issue is physicists use the same symbol for the physical, measurable quantity, and the function relating it to other quantities. To be clear, that isn't a criticism: it's a notational necessity (there are too many quantities to assign distinct symbols for each function). But that makes the semantics muddled.
However, there is also a lack of clarity about the semantics of "quantities". I think it is best to think of quantities as functions over an underlying state space. Functional relationships _between_ the quantities can then be reconstructed from those quantities, subject to uniqueness conditions.
This gives a more natural interpretation for the derivatives. It highlights that an expression like S(U, N, V) doesn't imply S _is_ the function, just that it's associated to it, and that S as a quantity could be associated with other functions.
---
The state space S has the structure of a differential manifold, diffeomorphic to R^n [0].
A quantity -- what in thermodynamics we might call a "state variable" -- is a smooth real-valued function on S.
An diffeomorphism between S and R^n is a co-ordinate system. Its components form the co-ordinates. Intuitively, any collection of quantities X = (X_1, ..., X_n) which uniquely labels all points in S is a co-ordinate system, which is the same thing as saying that it's invertible. [1]
Given such a co-ordinate system, any quantity Y can naturally be associated with a function f_Y : R^n -> R, defined by f_Y(x_1, ..., x_n) := Y(X^-1(x_1, ..., x_n)). In other words, this is the co-ordinate representation of Y. In physics, we would usually write that, as an abuse of notation: Y = Y(X_1, ..., X_n).
This leads to the definition of the partial derivative holding some quantities constant: you map the "held constant" quantities and the quantity in the denominator to the appropriate co-ordinate system, then take the derivative of f_Y, giving you a function which can then be mapped back to a quantity.
In that process, you have to make sure that the held constant quantities and the denominator quantity form a co-ordinate system. A lot of thermodynamic functions are posited to obey monotonicity/convexity properties, and this is why. It might be also possible to find a more permissive definition that uses multi-valued functions, similar to how Riemann surfaces are used in complex analysis.
To do that we'd probably want to be a bit more general and allow for "partial co-ordinate systems", which might also be useful for cases involving composite systems. Any collection of quantities (Y, X_1, ..., X_n) can be naturally associated with a relation [2], where (y, x_1, ..., x_n) is in the relation if there exists a point s in S such that (Y(s), X_1(s), ..., X_n(s)) = (y, x_1, ..., x_n). You can promote that to a function if it satisfies a uniqueness condition.
I think it is also possible to give a metric (Riemannian) structure on the manifold in a way compatible with the Second Law. I remember skimming through some papers on the topic, but didn't look in enough detail.
---
[0] Or half of R^n, or a quadrant maybe.
[1] The "diffeomorphism" definition also adds the condition that the inverse be smooth.
[2] Incidentally, same sense of "relation" that leads to the "relational data model"!
Everytime i manipulate dx i feel like walking on a minefield.
(Seriously though, learn to love the minefield. ~~~~Another physicist)
You don't even need to use "infinity", it starts out as just a variable representing some unknown quantity, then you "round to zero" on output.
I actually collected a bunch of old Infinitesimal calculus math books.
Embrace the minefield, love the minefield!
Signed
a physicistDid you mean "Leibniz's" notation[1]? If so, if you use the esdiff package[2] it's just \diffp{y}{x} for partials or \diff{x}{y} for regular derivatives.
Lagrange's notation is when people do x' = v or x'' = a and Like the Newton's notation you kinda have to know from context that you are differentiating with respect to time unless they write it properly as a function with arguments which people often tend not to (at least I often tend not to I guess).
Sometimes people call the partial derivative notation where you use subscripts "Lagrange's notation" also[3]. So like f_x(x,y) = blah is the partial derivative of f with respect to x.
[1] Actually invented by Euler, or maybe some other guy called Arbogast or something[?sp]
[2] https://ctan.math.illinois.edu/macros/latex/contrib/esdiff/e...
[3] Even though that was also actually invented by Euler apparently.
[0] https://hsm.stackexchange.com/questions/7704/was-english-mat...
"This is ridiculous! We need a better, more intuitive notation that's also easier to do math at."
(And obviously Functional Differential Geometry by the authors of SICM)
My pain was always Hamiltonians and Legendre equations for systems because the lecturer believed in learn by rote rather than explaining something that I'm sure for him was simply intuitive.
At any given point p on an n-dimensional manifold, a 1-form defines an n-dimensional cotangent vector (in the language of bundles[3], a point in the fiber over p).
So how do we define fractions of sections or vectors?
In the article, Baez defines fractions of 2-forms on the plane as the pointwise ratio of coefficients of a basis vector, which he can do because, as he points out, the space of 2-forms at a point on a 2-dimensional manifold is a 1-dimensional vector space (more generally, for k-forms on an n-dimensional manifold, this dimension is n choose k, so only 1 for 0-forms [functions] and n-forms).
[1] https://mathworld.wolfram.com/BundleSection.html
There are many areas of mathematics that spun from this.