Lorentz transformation

On pet physics theories, Scientology, cosmology, relativity and libertarian tendencies.

December 26, 2020 Incoherent ramblings , , , , , , , , , ,

In a recent Brian Keating podcast, he asked people to comment if they had their own theory of physics.

I’ve done a lot of exploration of conventional physics, both on my own, and with some in class studies (non-degree undergrad physics courses at UofT, plus grad QM and QFTI), but I don’t have my own personal theories of physics.  It’s enough of a challenge to figure out the existing theories without making up your own\({}^{1}\).

However, I have had one close encounter with an alternate physics theory, as I had a classmate in undergrad QMI (phy356) that had a personal “Aether” theory of relativity.  He shared that theory with me, but it came in the form of about 50 pages of dense text without equations.  For all intents and purposes, this was a theory that was written in an entirely different language than the rest of physics.  To him, it was all self evident, and he got upset with the suggestion of trying to mathematize it.  A lot of work and thought went into that theory, but it is work that has very little chance of paying off, since it was packaged in a form that made it unpalatable to anybody who is studying conventional physics.  There is also a great deal of work that would be required to “complete” the theory (presuming that could be done), since he would have to show that his theory is not inconsistent with many measurements and experiments, and would not predict nonphysical phenomena.  That was really an impossible task, which he would have found had he attempted to do so.  However, instead of attempting to do that work, he seemed to think that the onus should fall on others to do so.  He had done the work to write what he believed to be a self consistent logical theory that was self evidently true, and shouldn’t have to do anything more.

It is difficult to fully comprehend how he would have arrived at such certainty about his Aether theory, when he did not have the mathematical sophistication to handle the physics found in the theories that he believed his should supplant.  However, in his defence, there are some parts of what I imagine were part of his thought process that I can sympathize with.  The amount of knowledge required to understand the functioning of a even a simple digital watch (not to mention the cell “phone” computers that we now all carry) is absolutely phenomenal.  We are surrounded by so much sophisticated technology that understanding the mechanisms behind it all is practically unknowable.  Much of the world around is us is effectively magic to most people, even those with technical sophistication.  Should there be some sort of catastrophe that wipes out civilization, requiring us to relearn or redevelop everything from first principles, nobody really has the breadth required to reconstruct the world around us.  It is rather humbling to ponder that.

One way of coping with the fact that it is effectively impossible to know how everything works is to not believe in any consensus theories — period.  I think that is the origin of the recent popularization of flat earth models.  I think this was a factor in my classmate’s theory, as he also went on to believe that quantum mechanics was false (or also believed that when I knew him, but never stated it to me.)  People understand that it is impossible to know everything required to build their own satellites, imaging systems, rockets, et-al, (i.e. sophisticated methods of disproving the flat earth theory) and decide to disbelieve everything that they cannot personally prove.  That’s an interesting defence mechanism, but takes things to a rather extreme conclusion.

I have a lot of sympathy for those that do not believe in consensus theories.  Without such disbelief I would not have my current understanding of the world.  It happens that the prevailing consensus theory that I knew growing up was that of Scientology.  Among the many ideas that one finds in Scientology is a statement that relativity is wrong\({}^2\).  It has been too many years for me to accurately state the reasons that Hubbard stated that relativity was incorrect, but I do seem to recall that one of the arguments had to do with the speed of light being non-constant when bent by a prism \({}^3\).  I carried some Scientology derived skepticism of relativity into the undergrad “relativistic electrodynamics“\({}^4\) course that I took back around 2010, but had I not been willing to disregard the consensus beliefs that I had been exposed to up to that point in time, I would not have learned anything from that class.  Throwing away beliefs so that you can form your own conclusions is the key to being able to learn and grow.

I would admit to still carrying around baggage from my early indoctrination, despite not having practised Scientology for 25+ years.  This baggage spans multiple domains.  One example is that I am not subscribed to the religious belief that government and police are my friends.  It is hard to see your father, whom you love and respect, persecuted, and not come away with disrespect for the persecuting institutions.  I now have a rough idea of what Dad back in the Scientology Guardian’s Office did that triggered the ire of the Ontario crown attorneys \({}^5\).  However, that history definitely colored my views and current attitudes.  In particular, I recognize that back history as a key factor that pushed me so strongly in a libertarian direction.  The libertarian characterization of government as an entity that infringes on personal property and rights seems very reasonable, and aligns perfectly with my experience \({}^6\).

A second example of indoctrination based disbelief that I recognize that I carry with me is not subscribing to the current popular cosmological models of physics.  The big bang, and the belief that we know to picosecond granularity how the universe was at it’s beginning seems to me very religious.  That belief requires too much extrapolation, and it does not seem any more convincing to me than the Scientology cosmology.  The Scientology cosmology is somewhat ambiguous, and contains both a multiverse story and a steady state but finite model.  In the steady state aspect of that cosmology, the universe that we inhabit is dated with an age of 76 trillion years, but I don’t recall any sort of origin story for the beginning portion of that 76 trillion.  Given the little bits of things that we can actually measure and observe, I have no inclination to sign up for the big bang testiment any more than any other mythical origin story.  Thankfully, I can study almost anything that has practical application in physics or engineering and no amount of belief or disdain in the big bang or other less popular “physics” cosmologies makes any difference.  All of these, whether they be the big bang, cyclic theories, multiverses (of the quantum, thetan-created \({}^7\), or inflationary varieties), or even the old Scientology 76 trillion years old cosmology of my youth, cannot be measured, proven or disproved.  Just about any cosmology has no impact on anything real.

This throw it all out point of view of cosmology is probably a too cynical and harsh treatment of the subject.  It is certainly not the point of view that most practising physicists would take, but it is imminently practical.  There’s too much that is unknowable, so why waste time on the least knowable aspects of the unknowable when there are so many concrete things that we can learn.

 

Footnotes:

[1] The closest that I have come to my own theory of physics is somewhat zealous advocacy for the use of real Clifford algebras in engineering and physics (aka. geometric algebra.)  However, that is not a new theory, it is just a highly effective way to compactly represent many of the entities that we encounter in more clumsy forms in conventional physics.

[2] Hubbard’s sci-fi writing shows that he had knowledge of special relativistic time-dilation, and length-contraction effects.  I seem to recall that Lorentz transformations were mentioned in passing (on either the Student hat course, or in the “PDC” lectures).  I don’t believe that Hubbard had the mathematical sophistication to describe a Lorentz transformation in a quantitative sense.

[3] The traversal of light through matter is a complex affair, considerably different from light in vacuum, where the relativistic constancy applies.  It would be interesting to have a quantitative understanding of the chances of a photon getting through a piece of glass without interacting (absorption and subsequent time delayed spontaneous remission of new photons when the lattice atoms drop back into low energy states.)  There are probably also interactions of photons with the phonons of the lattice itself, and I don’t know how those would be quantified.  However, in short, I bet there is a large chance that most of the light that exits a transparent piece of matter is not the same light that went in, as it is going to come out as photons with different momentum, polarization, frequency, and so forth.  If we measure characteristics of a pulse of light going into and back out of matter, it’s probably somewhat akin to measuring the characteristics of a clementine orange that is thrown at a piece of heavy chicken wire at fastball speeds.  You’ll get some orange peel, seeds, pulp and other constituent components out the other side of the mesh, but shouldn’t really consider the inputs and the outputs to be equivalent entities.

[4] Relativistic electrodynamics was an extremely redundant course title, but was used to distinguish the class from the 3rd year applied electrodynamics course that had none of the tensor, relativistic, Lagrangian, nor four-vector baggage.

[5] Some information about that court case is publicly available, but it would be interesting to see whether I could use the Canadian or Ontario equivalent to the US freedom of information laws to request records from the Ontario crown and the RCMP about the specifics of Dad’s case.  Dad has passed, and was never terribly approachable about the subject when I could have asked him personally.  I did get his spin on the events as well as the media spin, and suspect that the truth is somewhere in between.

[6] This last year will probably push many people towards libertarian-ism (at least the subset of people that are not happy to be conforming sheep, or are too scared not to conform.)  We’ve had countless examples of watching evil bastards in government positions of power impose dictatorial and oppressive covid lockdowns on the poorest and most unfortunate people that they supposedly represent.  Instead, we see the corruption at full scale, with the little guys clobbered, and this covid lockdown scheme essentially being a way to efficiently channel money into the pockets of the rich.  The little guys loose their savings, livelihoods, and get their businesses shut down by fat corrupt bastards that believe they have the authority to decide whether or not you as an individual are essential.  The fat bastards that have the stamp of government authority do not believe that you should have the right to make up your own mind about what levels of risk are acceptable to you or your family.

[7] In Scientology, a sufficiently capable individual is considered capable of creating their own universes, independent of the 76 trillion year old universe that we currently inhabit.  Thetan is the label for the non-corporal form of that individual (i.e. what would be called the spirit or the soul in other religions.)

 

A couple more reciprocal frame examples.

December 14, 2020 math and physics play , , , , , , , , , , , , ,

[If mathjax doesn’t display properly for you, click here for a PDF of this post]

This post logically follows both of the following:

  1. Curvilinear coordinates and gradient in spacetime, and reciprocal frames, and
  2. Lorentz transformations in Space Time Algebra (STA)

The PDF linked above above contains all the content from this post plus (1.) above [to be edited later into a more logical sequence.]

More examples.

Here are a few additional examples of reciprocal frame calculations.

Problem: Unidirectional arbitrary functional dependence.

Let
\begin{equation}\label{eqn:reciprocal:2540}
x = a f(u),
\end{equation}
where \( a \) is a constant vector and \( f(u)\) is some arbitrary differentiable function with a non-zero derivative in the region of interest.

Answer

Here we have just a single tangent space direction (a line in spacetime) with tangent vector
\begin{equation}\label{eqn:reciprocal:2400}
\Bx_u = a \PD{u}{f} = a f_u,
\end{equation}
so we see that the tangent space vectors are just rescaled values of the direction vector \( a \).
This is a simple enough parameterization that we can compute the reciprocal frame vector explicitly using the gradient. We expect that \( \Bx^u = 1/\Bx_u \), and find
\begin{equation}\label{eqn:reciprocal:2420}
\inv{a} \cdot x = f(u),
\end{equation}
but for constant \( a \), we know that \( \grad a \cdot x = a \), so taking gradients of both sides we find
\begin{equation}\label{eqn:reciprocal:2440}
\inv{a} = \grad f = \PD{u}{f} \grad u,
\end{equation}
so the reciprocal vector is
\begin{equation}\label{eqn:reciprocal:2460}
\Bx^u = \grad u = \inv{a f_u},
\end{equation}
as expected.

Problem: Linear two variable parameterization.

Let \( x = a u + b v \), where \( x \wedge a \wedge b = 0 \) represents spacetime plane (also the tangent space.) Find the curvilinear coordinates and their reciprocals.

Answer

The frame vectors are easy to compute, as they are just
\begin{equation}\label{eqn:reciprocal:1960}
\begin{aligned}
\Bx_u &= \PD{u}{x} = a \\
\Bx_v &= \PD{v}{x} = b.
\end{aligned}
\end{equation}
This is an example of a parametric equation that we can easily invert, as we have
\begin{equation}\label{eqn:reciprocal:1980}
\begin{aligned}
x \wedge a &= – v \lr{ a \wedge b } \\
x \wedge b &= u \lr{ a \wedge b },
\end{aligned}
\end{equation}
so
\begin{equation}\label{eqn:reciprocal:2000}
\begin{aligned}
u
&= \inv{ a \wedge b } \cdot \lr{ x \wedge b } \\
&= \inv{ \lr{a \wedge b}^2 } \lr{ a \wedge b } \cdot \lr{ x \wedge b } \\
&=
\frac{
\lr{b \cdot x} \lr{ a \cdot b }

\lr{a \cdot x} \lr{ b \cdot b }
}{ \lr{a \wedge b}^2 }
\end{aligned}
\end{equation}
\begin{equation}\label{eqn:reciprocal:2020}
\begin{aligned}
v &= -\inv{ a \wedge b } \cdot \lr{ x \wedge a } \\
&= -\inv{ \lr{a \wedge b}^2 } \lr{ a \wedge b } \cdot \lr{ x \wedge a } \\
&=
-\frac{
\lr{b \cdot x} \lr{ a \cdot a }

\lr{a \cdot x} \lr{ a \cdot b }
}{ \lr{a \wedge b}^2 }
\end{aligned}
\end{equation}
Recall that \( \grad \lr{ a \cdot x} = a \), if \( a \) is a constant, so our gradients are just
\begin{equation}\label{eqn:reciprocal:2040}
\begin{aligned}
\grad u
&=
\frac{
b \lr{ a \cdot b }

a
\lr{ b \cdot b }
}{ \lr{a \wedge b}^2 } \\
&=
b \cdot \inv{ a \wedge b },
\end{aligned}
\end{equation}
and
\begin{equation}\label{eqn:reciprocal:2060}
\begin{aligned}
\grad v
&=
-\frac{
b \lr{ a \cdot a }

a \lr{ a \cdot b }
}{ \lr{a \wedge b}^2 } \\
&=
-a \cdot \inv{ a \wedge b }.
\end{aligned}
\end{equation}
Expressed in terms of the frame vectors, this is just
\begin{equation}\label{eqn:reciprocal:2080}
\begin{aligned}
\Bx^u &= \Bx_v \cdot \inv{ \Bx_u \wedge \Bx_v } \\
\Bx^v &= -\Bx_u \cdot \inv{ \Bx_u \wedge \Bx_v },
\end{aligned}
\end{equation}
so we were able to show, for this special two parameter linear case, that the explicit evaluation of the gradients has the exact structure that we intuited that the reciprocals must have, provided they are constrained to the spacetime plane \( a \wedge b \). It is interesting to observe how this structure falls out of the linear system solution so directly. Also note that these reciprocals are not defined at the origin of the \( (u,v) \) parameter space.

Problem: Quadratic two variable parameterization.

Now consider a variation of the previous problem, with \( x = a u^2 + b v^2 \). Find the curvilinear coordinates and their reciprocals.

Answer

\begin{equation}\label{eqn:reciprocal:2100}
\begin{aligned}
\Bx_u &= \PD{u}{x} = 2 u a \\
\Bx_v &= \PD{v}{x} = 2 v b.
\end{aligned}
\end{equation}
Our tangent space is still the \( a \wedge b \) plane (as is the surface itself), but the spacing of the cells starts getting wider in proportion to \( u, v \).
Utilizing the work from the previous problem, we have
\begin{equation}\label{eqn:reciprocal:2120}
\begin{aligned}
2 u \grad u &=
b \cdot \inv{ a \wedge b } \\
2 v \grad v &=
-a \cdot \inv{ a \wedge b }.
\end{aligned}
\end{equation}
A bit of rearrangement can show that this is equivalent to the reciprocal frame identities. This is a second demonstration that the gradient and the algebraic formulations for the reciprocals match, at least for these special cases of linear non-coupled parameterizations.

Problem: Reciprocal frame for generalized cylindrical parameterization.

Let the vector parameterization be \( x(\rho,\theta) = \rho e^{-i\theta/2} x(\rho_0, \theta_0) e^{i \theta} \), where \( i^2 = \pm 1 \) is a unit bivector (\(+1\) for a boost, and \(-1\) for a rotation), and where \(\theta, \rho\) are scalars. Find the tangent space vectors and their reciprocals.

fig. 1. “Cylindrical” boost parameterization.

Note that this is cylindrical parameterization for the rotation case, and traces out hyperbolic regions for the boost case. The boost case is illustrated in fig. 1 where hyperbolas in the light cone are found for boosts of \( \gamma_0\) with various values of \(\rho\), and the spacelike hyperbolas are boosts of \( \gamma_1 \), again for various values of \( \rho \).

Answer

The tangent space vectors are
\begin{equation}\label{eqn:reciprocal:2480}
\Bx_\rho = \frac{x}{\rho},
\end{equation}
and

\begin{equation}\label{eqn:reciprocal:2500}
\begin{aligned}
\Bx_\theta
&= -\frac{i}{2} x + x \frac{i}{2} \\
&= x \cdot i.
\end{aligned}
\end{equation}
Recall that \( x \cdot i \) lies perpendicular to \( x \) (in the plane \( i \)), as illustrated in fig. 2. This means that \( \Bx_\rho \) and \( \Bx_\theta \) are orthogonal, so we can find the reciprocal vectors by just inverting them
\begin{equation}\label{eqn:reciprocal:2520}
\begin{aligned}
\Bx^\rho &= \frac{\rho}{x} \\
\Bx^\theta &= \frac{1}{x \cdot i}.
\end{aligned}
\end{equation}

fig. 2. Projection and rejection geometry.

Parameterization of a general linear transformation.

Given \( N \) parameters \( u^0, u^1, \cdots u^{N-1} \), a general linear transformation from the parameter space to the vector space has the form
\begin{equation}\label{eqn:reciprocal:2160}
x =
{a^\alpha}_\beta \gamma_\alpha u^\beta,
\end{equation}
where \( \beta \in [0, \cdots, N-1] \) and \( \alpha \in [0,3] \).
For such a general transformation, observe that the curvilinear basis vectors are
\begin{equation}\label{eqn:reciprocal:2180}
\begin{aligned}
\Bx_\mu
&= \PD{u^\mu}{x} \\
&= \PD{u^\mu}{}
{a^\alpha}_\beta \gamma_\alpha u^\beta \\
&=
{a^\alpha}_\mu \gamma_\alpha.
\end{aligned}
\end{equation}
We find an interpretation of \( {a^\alpha}_\mu \) by dotting \( \Bx_\mu \) with the reciprocal frame vectors of the standard basis
\begin{equation}\label{eqn:reciprocal:2200}
\begin{aligned}
\Bx_\mu \cdot \gamma^\nu
&=
{a^\alpha}_\mu \lr{ \gamma_\alpha \cdot \gamma^\nu } \\
&=
{a^\nu}_\mu,
\end{aligned}
\end{equation}
so
\begin{equation}\label{eqn:reciprocal:2220}
x = \Bx_\mu u^\mu.
\end{equation}
We are able to reinterpret \ref{eqn:reciprocal:2160} as a contraction of the tangent space vectors with the parameters, scaling and summing these direction vectors to characterize all the points in the tangent plane.

Theorem 1.1: Projecting onto the tangent space.

Let \( T \) represent the tangent space. The projection of a vector onto the tangent space has the form
\begin{equation}\label{eqn:reciprocal:2560}
\textrm{Proj}_{\textrm{T}} y = \lr{ y \cdot \Bx^\mu } \Bx_\mu = \lr{ y \cdot \Bx_\mu } \Bx^\mu.
\end{equation}

Start proof:

Let’s designate \( a \) as the portion of the vector \( y \) that lies outside of the tangent space
\begin{equation}\label{eqn:reciprocal:2260}
y = y^\mu \Bx_\mu + a.
\end{equation}
If we knew the coordinates \( y^\mu \), we would have a recipe for the projection.
Algebraically, requiring that \( a \) lies outside of the tangent space, is equivalent to stating \( a \cdot \Bx_\mu = a \cdot \Bx^\mu = 0 \). We use that fact, and then take dot products
\begin{equation}\label{eqn:reciprocal:2280}
\begin{aligned}
y \cdot \Bx^\nu
&= \lr{ y^\mu \Bx_\mu + a } \cdot \Bx^\nu \\
&= y^\nu,
\end{aligned}
\end{equation}
so
\begin{equation}\label{eqn:reciprocal:2300}
y = \lr{ y \cdot \Bx^\mu } \Bx_\mu + a.
\end{equation}
Similarly, the tangent space projection can be expressed as a linear combination of reciprocal basis elements
\begin{equation}\label{eqn:reciprocal:2320}
y = y_\mu \Bx^\mu + a.
\end{equation}
Dotting with \( \Bx_\mu \), we have
\begin{equation}\label{eqn:reciprocal:2340}
\begin{aligned}
y \cdot \Bx^\mu
&= \lr{ y_\alpha \Bx^\alpha + a } \cdot \Bx_\mu \\
&= y_\mu,
\end{aligned}
\end{equation}
so
\begin{equation}\label{eqn:reciprocal:2360}
y = \lr{ y \cdot \Bx^\mu } \Bx_\mu + a.
\end{equation}
We find the two stated ways of computing the projection.

Observe that, for the special case that all of \( \setlr{ \Bx_\mu } \) are orthogonal, the equivalence of these two projection methods follows directly, since
\begin{equation}\label{eqn:reciprocal:2380}
\begin{aligned}
\lr{ y \cdot \Bx^\mu } \Bx_\mu
&=
\lr{ y \cdot \inv{\Bx_\mu} } \inv{\Bx^\mu} \\
&=
\lr{ y \cdot \frac{\Bx_\mu}{\lr{\Bx_\mu}^2 } } \frac{\Bx^\mu}{\lr{\Bx^\mu}^2} \\
&=
\lr{ y \cdot \Bx_\mu } \Bx^\mu.
\end{aligned}
\end{equation}

End proof.

PHY2403H Quantum Field Theory. Lecture 3: Lorentz transformations and a scalar action. Taught by Prof. Erich Poppitz

September 18, 2018 phy2403 , ,

[Click here for a PDF of this post with nicer formatting]

DISCLAIMER: Very rough notes from class. Some additional side notes, but otherwise barely edited.

These are notes for the UofT course PHY2403H, Quantum Field Theory, taught by Prof. Erich Poppitz.

Determinant of Lorentz transformations

We require that Lorentz transformations leave the dot product invariant, that is \( x \cdot y = x’ \cdot y’ \), or
\begin{equation}\label{eqn:qftLecture3:20}
x^\mu g_{\mu\nu} y^\nu = {x’}^\mu g_{\mu\nu} {y’}^\nu.
\end{equation}
Explicitly, with coordinate transformations
\begin{equation}\label{eqn:qftLecture3:40}
\begin{aligned}
{x’}^\mu &= {\Lambda^\mu}_\rho x^\rho \\
{y’}^\mu &= {\Lambda^\mu}_\rho y^\rho
\end{aligned}
\end{equation}
such a requirement is equivalent to demanding that
\begin{equation}\label{eqn:qftLecture3:500}
\begin{aligned}
x^\mu g_{\mu\nu} y^\nu
&=
{\Lambda^\mu}_\rho x^\rho
g_{\mu\nu}
{\Lambda^\nu}_\kappa y^\kappa \\
&=
x^\mu
{\Lambda^\alpha}_\mu
g_{\alpha\beta}
{\Lambda^\beta}_\nu
y^\nu,
\end{aligned}
\end{equation}
or
\begin{equation}\label{eqn:qftLecture3:60}
g_{\mu\nu}
=
{\Lambda^\alpha}_\mu
g_{\alpha\beta}
{\Lambda^\beta}_\nu
\end{equation}

multiplying by the inverse we find
\begin{equation}\label{eqn:qftLecture3:200}
\begin{aligned}
g_{\mu\nu}
{\lr{\Lambda^{-1}}^\nu}_\lambda
&=
{\Lambda^\alpha}_\mu
g_{\alpha\beta}
{\Lambda^\beta}_\nu
{\lr{\Lambda^{-1}}^\nu}_\lambda \\
&=
{\Lambda^\alpha}_\mu
g_{\alpha\lambda} \\
&=
g_{\lambda\alpha}
{\Lambda^\alpha}_\mu.
\end{aligned}
\end{equation}
This is now amenable to expressing in matrix form
\begin{equation}\label{eqn:qftLecture3:220}
\begin{aligned}
(G \Lambda^{-1})_{\mu\lambda}
&=
(G \Lambda)_{\lambda\mu} \\
&=
((G \Lambda)^\T)_{\mu\lambda} \\
&=
(\Lambda^\T G)_{\mu\lambda},
\end{aligned}
\end{equation}
or
\begin{equation}\label{eqn:qftLecture3:80}
G \Lambda^{-1}
=
(G \Lambda)^\T.
\end{equation}

Taking determinants (using the normal identities for products of determinants, determinants of transposes and inverses), we find
\begin{equation}\label{eqn:qftLecture3:100}
det(G)
det(\Lambda^{-1})
=
det(G) det(\Lambda),
\end{equation}
or
\begin{equation}\label{eqn:qftLecture3:120}
det(\Lambda)^2 = 1,
\end{equation}
or
\( det(\Lambda)^2 = \pm 1 \). We will generally ignore the case of reflections in spacetime that have a negative determinant.

Smart-alec Peeter pointed out after class last time that we can do the same thing easier in matrix notation
\begin{equation}\label{eqn:qftLecture3:140}
\begin{aligned}
x’ &= \Lambda x \\
y’ &= \Lambda y
\end{aligned}
\end{equation}
where
\begin{equation}\label{eqn:qftLecture3:160}
\begin{aligned}
x’ \cdot y’
&=
(x’)^\T G y’ \\
&=
x^\T \Lambda^\T G \Lambda y,
\end{aligned}
\end{equation}
which we require to be \( x \cdot y = x^\T G y \) for all four vectors \( x, y \), that is
\begin{equation}\label{eqn:qftLecture3:180}
\Lambda^\T G \Lambda = G.
\end{equation}
We can find the result \ref{eqn:qftLecture3:120} immediately without having to first translate from index notation to matrices.

Field theory

The electrostatic potential is an example of a scalar field \( \phi(\Bx) \) unchanged by SO(3) rotations
\begin{equation}\label{eqn:qftLecture3:240}
\Bx \rightarrow \Bx’ = O \Bx,
\end{equation}
that is
\begin{equation}\label{eqn:qftLecture3:260}
\phi'(\Bx’) = \phi(\Bx).
\end{equation}
Here \( \phi'(\Bx’) \) is the value of the (electrostatic) scalar potential in a primed frame.

However, the electrostatic field is not invariant under Lorentz transformation.
We postulate that there is some scalar field
\begin{equation}\label{eqn:qftLecture3:280}
\phi'(x’) = \phi(x),
\end{equation}
where \( x’ = \Lambda x \) is an SO(1,3) transformation. There are actually no stable particles (fields that persist at long distances) described by Lorentz scalar fields, although there are some unstable scalar fields such as the Higgs, Pions, and Kaons. However, much of our homework and discussion will be focused on scalar fields, since they are the easiest to start with.

We need to first understand how derivatives \( \partial_\mu \phi(x) \) transform. Using the chain rule
\begin{equation}\label{eqn:qftLecture3:300}
\begin{aligned}
\PD{x^\mu}{\phi(x)}
&=
\PD{x^\mu}{\phi'(x’)} \\
&=
\PD{{x’}^\nu}{\phi'(x’)}
\PD{{x}^\mu}{{x’}^\nu} \\
&=
\PD{{x’}^\nu}{\phi'(x’)}
\partial_\mu \lr{
{\Lambda^\nu}_\rho x^\rho
} \\
&=
\PD{{x’}^\nu}{\phi'(x’)}
{\Lambda^\nu}_\mu \\
&=
\PD{{x’}^\nu}{\phi(x)}
{\Lambda^\nu}_\mu.
\end{aligned}
\end{equation}
Multiplying by the inverse \( {\lr{\Lambda^{-1}}^\mu}_\kappa \) we get
\begin{equation}\label{eqn:qftLecture3:320}
\PD{{x’}^\kappa}{}
=
{\lr{\Lambda^{-1}}^\mu}_\kappa \PD{x^\mu}{}
\end{equation}

This should be familiar to you, and is an analogue of the transformation of the
\begin{equation}\label{eqn:qftLecture3:340}
d\Br \cdot \spacegrad_\Br
=
d\Br’ \cdot \spacegrad_{\Br’}.
\end{equation}

Actions

We will start with a classical action, and quantize to determine a QFT. In mechanics we have the particle position \( q(t) \), which is a classical field in 1+0 time and space dimensions. Our action is
\begin{equation}\label{eqn:qftLecture3:360}
S
= \int dt \LL(t)
= \int dt \lr{
\inv{2} \dot{q}^2 – V(q)
}.
\end{equation}
This action depends on the position of the particle that is local in time. You could imagine that we have a more complex action where the action depends on future or past times
\begin{equation}\label{eqn:qftLecture3:380}
S
= \int dt’ q(t’) K( t’ – t ),
\end{equation}
but we don’t seem to find such actions in classical mechanics.

Principles determining the form of the action.

  • relativity (action is invariant under Lorentz transformation)
  • locality (action depends on fields and the derivatives at given \((t, \Bx)\).
  • Gauge principle (the action should be invariant under gauge transformation). We won’t discuss this in detail right now since we will start with studying scalar fields.
    Recall that for Maxwell’s equations a gauge transformation has the form
    \begin{equation}\label{eqn:qftLecture3:520}
    \phi \rightarrow \phi + \dot{\chi}, \BA \rightarrow \BA – \spacegrad \chi.
    \end{equation}

Suppose we have a real scalar field \( \phi(x) \) where \( x \in \mathbb{R}^{1,d-1} \). We will be integrating over space and time \( \int dt d^{d-1} x \) which we will write as \( \int d^d x \). Our action is
\begin{equation}\label{eqn:qftLecture3:400}
S = \int d^d x \lr{ \text{Some action density to be determined } }
\end{equation}
The analogue of \( \dot{q}^2 \) is
\begin{equation}\label{eqn:qftLecture3:420}
\begin{aligned}
\lr{ \PD{x^\mu}{\phi} }
\lr{ \PD{x^\nu}{\phi} }
g^{\mu\nu}
&=
(\partial_\mu \phi) (\partial_\nu \phi) g^{\mu\nu} \\
&= \partial^\mu \phi \partial_\mu \phi.
\end{aligned}
\end{equation}
This has both time and spatial components, that is
\begin{equation}\label{eqn:qftLecture3:440}
\partial^\mu \phi \partial_\mu \phi =
\dotphi^2 – (\spacegrad \phi)^2,
\end{equation}
so the desired simplest scalar action is
\begin{equation}\label{eqn:qftLecture3:460}
S = \int d^d x \lr{ \dotphi^2 – (\spacegrad \phi)^2 }.
\end{equation}
The measure transforms using a Jacobian, which we have seen is the Lorentz transform matrix, and has unit determinant
\begin{equation}\label{eqn:qftLecture3:480}
d^d x’ = d^d x \Abs{ det( \Lambda^{-1} ) } = d^d x.
\end{equation}

Problems.

Question: Four vector form of the Maxwell gauge transformation.

Show that the transformation
\begin{equation}\label{eqn:qftLecture3:580}
A^\mu \rightarrow A^\mu + \partial^\mu \chi
\end{equation}
is the desired four-vector form of the gauge transformation \ref{eqn:qftLecture3:520}, that is
\begin{equation}\label{eqn:qftLecture3:540}
\begin{aligned}
j^\nu
&= \partial_\mu {F’}^{\mu\nu} \\
&= \partial_\mu F^{\mu\nu}.
\end{aligned}
\end{equation}
Also relate this four-vector gauge transformation to the spacetime split.

Answer

\begin{equation}\label{eqn:qftLecture3:560}
\begin{aligned}
\partial_\mu {F’}^{\mu\nu}
&=
\partial_\mu \lr{ \partial^\mu {A’}^\nu – \partial_\nu {A’}^\mu } \\
&=
\partial_\mu \lr{
\partial^\mu \lr{ A^\nu + \partial^\nu \chi }
– \partial_\nu \lr{ A^\mu + \partial^\mu \chi }
} \\
&=
\partial_\mu {F}^{\mu\nu}
+
\partial_\mu \partial^\mu \partial^\nu \chi

\partial_\mu \partial^\nu \partial^\mu \chi \\
&=
\partial_\mu {F}^{\mu\nu},
\end{aligned}
\end{equation}
by equality of mixed partials. Expanding \ref{eqn:qftLecture3:580} explicitly we find
\begin{equation}\label{eqn:qftLecture3:600}
{A’}^\mu = A^\mu + \partial^\mu \chi,
\end{equation}
which is
\begin{equation}\label{eqn:qftLecture3:620}
\begin{aligned}
\phi’ = {A’}^0 &= A^0 + \partial^0 \chi = \phi + \dot{\chi} \\
\BA’ \cdot \Be_k = {A’}^k &= A^k + \partial^k \chi = \lr{ \BA – \spacegrad \chi } \cdot \Be_k.
\end{aligned}
\end{equation}
The last of which can be written in vector notation as \( \BA’ = \BA – \spacegrad \chi \).

UofT QFT Fall 2018 Lecture 2. Units, scales, and Lorentz transformations. Taught by Prof. Erich Poppitz

September 17, 2018 phy2403 , , ,

[Click here for a PDF of this post with nicer formatting]

DISCLAIMER: Very rough notes from class. Some additional side notes, but otherwise barely edited.

Natural units.

\begin{equation}\label{eqn:qftLecture2:20}
\begin{aligned}
[\Hbar] &= [\text{action}] = M \frac{L^2}{T^2} T = \frac{M L^2}{T} \\
&= [\text{velocity}] = \frac{L}{T} \\
& [\text{energy}] = M \frac{L^2}{T^2}.
\end{aligned}
\end{equation}

Setting \( c = 1 \) means
\begin{equation}\label{eqn:qftLecture2:240}
\frac{L}{T} = 1
\end{equation}
and setting \( \Hbar = 1 \) means
\begin{equation}\label{eqn:qftLecture2:260}
[\Hbar] = [\text{action}] = M L {\frac{L}{T}} = M L
\end{equation}
therefore
\begin{equation}\label{eqn:qftLecture2:280}
[L] = \inv{\text{mass}}
\end{equation}
and
\begin{equation}\label{eqn:qftLecture2:300}
[\text{energy}] = M {\frac{L^2}{T^2}} = \text{mass}\, \text{eV}
\end{equation}

Summary

  • \( \text{energy} \sim \text{eV} \)
  • \( \text{distance} \sim \inv{M} \)
  • \( \text{time} \sim \inv{M} \)

From:
\begin{equation}\label{eqn:qftLecture2:320}
\alpha = \frac{e^2}{4 \pi {\Hbar c}}
\end{equation}
which is dimensionless (\(1/137\)), so electric charge is dimensionless.

Some useful numbers in natural units

\begin{equation}\label{eqn:qftLecture2:40}
\begin{aligned}
m_\txte &\sim 10^{-27} \text{g} \sim 0.5 \text{MeV} \\
m_\txtp &\sim 2000 m_\txte \sim 1 \text{GeV} \\
m_\pi &\sim 140 \text{MeV} \\
m_\mu &\sim 105 \text{MeV} \\
\Hbar c &\sim 200 \text{MeV} \,\text{fm} = 1
\end{aligned}
\end{equation}

Gravity

Interaction energy of two particles

\begin{equation}\label{eqn:qftLecture2:60}
G_\txtN \frac{m_1 m_2}{r}
\end{equation}

\begin{equation}\label{eqn:qftLecture2:80}
[\text{energy}] \sim [G_\txtN] \frac{M^2}{L}
\end{equation}

\begin{equation}\label{eqn:qftLecture2:100}
[G_\txtN]
\sim
[\text{energy}] \frac{L}{M^2}
\end{equation}

but energy x distance is dimensionless (action) in our units

\begin{equation}\label{eqn:qftLecture2:120}
[G_\txtN]
\sim
{\text{dimensionless}}{M^2}
\end{equation}

\begin{equation}\label{eqn:qftLecture2:140}
\frac{G_\txtN}{\Hbar c} \sim \inv{M^2} \sim \frac{1}{10^{20} \text{GeV}}
\end{equation}

Planck mass

\begin{equation}\label{eqn:qftLecture2:160}
M_{\text{Planck}} \sim \sqrt{\frac{\Hbar c}{G_\txtN}}
\sim 10^{-4} g \sim \inv{\lr{10^{20} \text{GeV}}^2}
\end{equation}

We can revisit the scale diagram from last lecture in terms of MeV mass/energy values, as sketched in fig. 1.

fig. 1. Scales, take II.

At the classical electron radius scale, we consider phenomena such as back reaction of radiation, the self energy of electrons. At the Compton wavelength we have to allow for production of multiple particle pairs. At Bohr radius scales we must start using QM instead of classical mechanics.

Cross section.

Verbal discussion of cross section, not captured in these notes. Roughly, the cross section sounds like the number of events per unit time, related to the flux of some source through an area.

We’ll compute the cross section of a number of different systems in this course. The cross section is relevant in scattering such as the electron-electron scattering sketched in fig. 2.

fig. 2. Electron electron scattering.

We assume that QED is highly relativistic. In natural units, our scale factor is basically the square of the electric charge
\begin{equation}\label{eqn:qftLecture2:180}
\alpha \sim e^2,
\end{equation}
so the cross section has the form
\begin{equation}\label{eqn:qftLecture2:200}
\sigma \sim \frac{\alpha^2}{E^2} \lr{ 1 + O(\alpha) + O(\alpha^2) + \cdots }
\end{equation}

In gravity we could consider scattering of electrons, where \( G_\txtN \) takes the place of \( \alpha \). However, \( G_\txtN \) has dimensions.

For electron-electron scattering due to gravitons

\begin{equation}\label{eqn:qftLecture2:220}
\sigma \sim \frac{G_\txtN^2 E^2}{1 + G_\txtN E^2 + \cdots }
\end{equation}

Now the cross section grows with energy. This will cause some problems (violating unitarity: probabilities greater than 1!) when \( O(G_\txtN E^2) = 1 \).

In any quantum field theories when the coupling constant is not-dimensionless we have the same sort of problems at some scale.

The point is that we can get far considering just dimensional analysis.

If the coupling constant has a dimension \((1/\text{mass})^N\,, N > 0\), then unitarity will be violated at high energy. One such theory is the Fermi theory of beta decay (electro-weak theory), which had a coupling constant with dimensions inverse-mass-squared. The relevant scale for beta decay was 4 Fermi, or \( G_\txtF \sim (1/{100 \text{GeV}})^2 \). This was the motivation for introducing the Higgs theory, which was motivated by restoring unitarity.

Lorentz transformations.

The goal, perhaps not for today, is to study the simplest (relativistic) scalar field theory. First studied classically, and then consider such a quantum field theory.

How is relativity implemented when we write the Lagrangian and action?

Our first step must be to consider Lorentz transformations and the Lorentz group.

Spacetime (Minkowski space) is \R{3,1} (or \R{d-1,1}). Our coordinates are

\begin{equation}\label{eqn:qftLecture2:340}
(c t, x^1, x^2, x^3) = (c t, \Br).
\end{equation}

Here, we’ve scaled the time scale by \( c \) so that we measure time and space in the same dimensions. We write this as

\begin{equation}\label{eqn:qftLecture2:360}
x^\mu = (x^0, x^1, x^2, x^3),
\end{equation}

where \( \mu = 0, 1, 2, 3 \), and call this a “4-vector”. These are called the space-time coordinates of an event, which tell us where and when an event occurs.

For two events whose spacetime coordinates differ by \( dx^0, dx^1, dx^2, dx^3 \) we introduce the notion of a space time \underline{interval}

\begin{equation}\label{eqn:qftLecture2:380}
\begin{aligned}
ds^2
&= c^2 dt^2
– (dx^1)^2
– (dx^2)^2
– (dx^3)^2 \\
&=
\sum_{\mu, \nu = 0}^3 g_{\mu\nu} dx^\mu dx^\nu
\end{aligned}
\end{equation}

Here \( g_{\mu\nu} \) is the Minkowski space metric, an object with two indexes that run from 0-3. i.e. this is a diagonal matrix

\begin{equation}\label{eqn:qftLecture2:400}
g_{\mu\nu} \sim
\begin{bmatrix}
1 & 0 & 0 & 0 \\
0 & -1 & 0 & 0 \\
0 & 0 & -1 & 0 \\
0 & 0 & 0 & -1 \\
\end{bmatrix}
\end{equation}

i.e.
\begin{equation}\label{eqn:qftLecture2:420}
\begin{aligned}
g_{00} &= 1 \\
g_{11} &= -1 \\
g_{22} &= -1 \\
g_{33} &= -1 \\
\end{aligned}
\end{equation}

We will use the Einstein summation convention, where any repeated upper and lower indexes are considered summed over. That is \ref{eqn:qftLecture2:380} is written with an implied sum
\begin{equation}\label{eqn:qftLecture2:440}
ds^2 = g_{\mu\nu} dx^\mu dx^\nu.
\end{equation}

Explicit expansion:
\begin{equation}\label{eqn:qftLecture2:460}
\begin{aligned}
ds^2
&= g_{\mu\nu} dx^\mu dx^\nu \\
&=
g_{00} dx^0 dx^0
+g_{11} dx^1 dx^1
+g_{22} dx^2 dx^2
+g_{33} dx^3 dx^3
&=
(1) dx^0 dx^0
+ (-1) dx^1 dx^1
+ (-1) dx^2 dx^2
+ (-1) dx^3 dx^3.
\end{aligned}
\end{equation}

Recall that rotations (with orthogonal matrix representations) are transformations that leave the dot product unchanged, that is

\begin{equation}\label{eqn:qftLecture2:480}
\begin{aligned}
(R \Bx) \cdot (R \By)
&= \Bx^\T R^\T R \By \\
&= \Bx^\T \By \\
&= \Bx \cdot \By,
\end{aligned}
\end{equation}

where \( R \) is a rotation orthogonal 3×3 matrix. The set of such transformations that leave the dot product unchanged have orthonormal matrix representations \( R^\T R = 1 \). We call the set of such transformations that have unit determinant the SO(3) group.

We call a Lorentz transformation, if it is a linear transformation acting on 4 vectors that leaves the spacetime interval (i.e. the inner product of 4 vectors) invariant. That is, a transformation that leaves
\begin{equation}\label{eqn:qftLecture2:500}
x^\mu y^\nu g_{\mu\nu} = x^0 y^0 – x^1 y^1 – x^2 y^2 – x^3 y^3
\end{equation}
unchanged.

Suppose that transformation has a 4×4 matrix form

\begin{equation}\label{eqn:qftLecture2:520}
{x’}^\mu = {\Lambda^\mu}_\nu x^\nu
\end{equation}

For an example of a possible \( \Lambda \), consider the transformation sketched in fig. 3.

fig. 3. Boost transformation.

We know that boost has the form
\begin{equation}\label{eqn:qftLecture2:540}
\begin{aligned}
x &= \frac{x’ + v t’}{\sqrt{1 – v^2/c^2}} \\
y &= y’ \\
z &= z’ \\
t &= \frac{t’ + (v/c^2) x’}{\sqrt{1 – v^2/c^2}} \\
\end{aligned}
\end{equation}
(this is a boost along the x-axis, not y as I’d drawn),
or
\begin{equation}\label{eqn:qftLecture2:560}
\begin{bmatrix}
c t \\
x \\
y \\
z
\end{bmatrix}
=
\begin{bmatrix}
\inv{\sqrt{1 – v^2/c^2}} & \frac{v/c}{\sqrt{1 – v^2/c^2}} & 0 & 0 \\
\frac{v/c}{\sqrt{1 – v^2/c^2}} & \frac{1}{\sqrt{1 – v^2/c^2}} & 0 & 0 \\
0 & 0 & 1 & 0 \\
0 & 0 & 0 & 1 \\
\end{bmatrix}
\begin{bmatrix}
c t’ \\
x’ \\
y’ \\
z’
\end{bmatrix}
\end{equation}

Other examples include rotations (\({\lambda^0}_0 = 1\) zeros in \( {\lambda^0}_k, {\lambda^k}_0 \), and a rotation matrix in the remainder.)

Back to Lorentz transformations (\(\text{SO}(1,3)^+\)), let
\begin{equation}\label{eqn:qftLecture2:600}
\begin{aligned}
{x’}^\mu &= {\Lambda^\mu}_\nu x^\nu \\
{y’}^\kappa &= {\Lambda^\kappa}_\rho y^\rho
\end{aligned}
\end{equation}

The dot product
\begin{equation}\label{eqn:qftLecture2:620}
g_{\mu \kappa}
{x’}^\mu
{y’}^\kappa
=
g_{\mu \kappa}
{\Lambda^\mu}_\nu
{\Lambda^\kappa}_\rho
x^\nu
y^\rho
=
g_{\nu\rho}
x^\nu
y^\rho,
\end{equation}
where the last step introduces the invariance requirement of the transformation. That is

\begin{equation}\label{eqn:qftLecture2:640}
\boxed{
g_{\nu\rho}
=
g_{\mu \kappa}
{\Lambda^\mu}_\nu
{\Lambda^\kappa}_\rho.
}
\end{equation}

Upper and lower indexes

We’ve defined

\begin{equation}\label{eqn:qftLecture2:660}
x^\mu = (t, x^1, x^2, x^3)
\end{equation}

We could also define a four vector with lower indexes
\begin{equation}\label{eqn:qftLecture2:680}
x_\nu = g_{\nu\mu} x^\mu = (t, -x^1, -x^2, -x^3).
\end{equation}
That is
\begin{equation}\label{eqn:qftLecture2:700}
\begin{aligned}
x_0 &= x^0 \\
x_1 &= -x^1 \\
x_2 &= -x^2 \\
x_3 &= -x^3.
\end{aligned}
\end{equation}

which allows us to write the dot product as simply \( x^\mu y_\mu \).

We can also define a metric tensor with upper indexes

\begin{equation}\label{eqn:qftLecture2:401}
g^{\mu\nu} \sim
\begin{bmatrix}
1 & 0 & 0 & 0 \\
0 & -1 & 0 & 0 \\
0 & 0 & -1 & 0 \\
0 & 0 & 0 & -1 \\
\end{bmatrix}
\end{equation}
This is the inverse matrix of \( g_{\mu\nu} \), and it satisfies
\begin{equation}\label{eqn:qftLecture2:720}
g^{\mu \nu} g_{\nu\rho} = {\delta^\mu}_\rho
\end{equation}

Exercise: Check:
\begin{equation}\label{eqn:qftLecture2:740}
\begin{aligned}
g_{\mu\nu} x^\mu y^\nu
&= x_\nu y^\nu \\
&= x^\nu y_\nu \\
&= g^{\mu\nu} x_\mu y_\nu \\
&= {\delta^\mu}_\nu x_\mu y^\nu.
\end{aligned}
\end{equation}

Class ended around this point, but it appeared that we were heading this direction:

Returning to the Lorentz invariant and multiplying both sides of
\ref{eqn:qftLecture2:640} with an inverse Lorentz transformation \( \Lambda^{-1} \), we find
\begin{equation}\label{eqn:qftLecture2:760}
\begin{aligned}
g_{\nu\rho}
{\lr{\Lambda^{-1}}^\rho}_\alpha
&=
g_{\mu \kappa}
{\Lambda^\mu}_\nu
{\Lambda^\kappa}_\rho
{\lr{\Lambda^{-1}}^\rho}_\alpha \\
&=
g_{\mu \kappa}
{\Lambda^\mu}_\nu
{\delta^\kappa}_\alpha \\
&=
g_{\mu \alpha}
{\Lambda^\mu}_\nu,
\end{aligned}
\end{equation}
or
\begin{equation}\label{eqn:qftLecture2:780}
\lr{\Lambda^{-1}}_{\nu \alpha} = \Lambda_{\alpha \nu}.
\end{equation}
This is clearly analogous to \( R^\T = R^{-1} \), although the index notation obscures things considerably. Prof. Poppitz said that next week this would all lead to showing that the determinant of any Lorentz transformation was \( \pm 1 \).

For what it’s worth, it seems to me that this index notation makes life a lot harder than it needs to be, at least for a matrix related question (i.e. determinant of the transformation). In matrix/column-(4)-vector notation, let \(x’ = \Lambda x, y’ = \Lambda y\) be two four vector transformations, then
\begin{equation}\label{eqn:qftLecture2:800}
x’ \cdot y’ = {x’}^T G y’ = (\Lambda x)^T G \Lambda y = x^T ( \Lambda^T G \Lambda) y = x^T G y.
\end{equation}
so
\begin{equation}\label{eqn:qftLecture2:820}
\boxed{
\Lambda^T G \Lambda = G.
}
\end{equation}
Taking determinants of both sides gives \(-(det(\Lambda))^2 = -1\), and thus \(det(\Lambda) = \pm 1\).

The many faces of Maxwell’s equations

March 5, 2018 math and physics play , , , , , , , , , , , , , , , , , , , , , , , ,

[Click here for a PDF of this post with nicer formatting (including equation numbering and references)]

The following is a possible introduction for a report for a UofT ECE2500 project associated with writing a small book: “Geometric Algebra for Electrical Engineers”. Given the space constraints for the report I may have to drop much of this, but some of the history of Maxwell’s equations may be of interest, so I thought I’d share before the knife hits the latex.

Goals of the project.

This project had a few goals

  1. Perform a literature review of applications of geometric algebra to the study of electromagnetism. Geometric algebra will be defined precisely later, along with bivector, trivector, multivector and other geometric algebra generalizations of the vector.
  2. Identify the subset of the literature that had direct relevance to electrical engineering.
  3. Create a complete, and as compact as possible, introduction of the prerequisites required
    for a graduate or advanced undergraduate electrical engineering student to be able to apply
    geometric algebra to problems in electromagnetism.

The many faces of electromagnetism.

There is a long history of attempts to find more elegant, compact and powerful ways of encoding and working with Maxwell’s equations.

Maxwell’s formulation.

Maxwell [12] employs some differential operators, including the gradient \( \spacegrad \) and Laplacian \( \spacegrad^2 \), but the divergence and gradient are always written out in full using coordinates, usually in integral form. Reading the original Treatise highlights how important notation can be, as most modern engineering or physics practitioners would find his original work incomprehensible. A nice translation from Maxwell’s notation to the modern Heaviside-Gibbs notation can be found in [16].

Quaterion representation.

In his second volume [11] the equations of electromagnetism are stated using quaterions (an extension of complex numbers to three dimensions), but quaternions are not used in the work. The modern form of Maxwell’s equations in quaternion form is
\begin{equation}\label{eqn:ece2500report:220}
\begin{aligned}
\inv{2} \antisymmetric{ \frac{d}{dr} }{ \BH } – \inv{2} \symmetric{ \frac{d}{dr} } { c \BD } &= c \rho + \BJ \\
\inv{2} \antisymmetric{ \frac{d}{dr} }{ \BE } + \inv{2} \symmetric{ \frac{d}{dr} }{ c \BB } &= 0,
\end{aligned}
\end{equation}
where \( \ifrac{d}{dr} = (1/c) \PDi{t}{} + \Bi \PDi{x}{} + \Bj \PDi{y}{} + \Bk \PDi{z}{} \) [7] acts bidirectionally, and vectors are expressed in terms of the quaternion basis \( \setlr{ \Bi, \Bj, \Bk } \), subject to the relations \(
\Bi^2 = \Bj^2 = \Bk^2 = -1, \quad
\Bi \Bj = \Bk = -\Bj \Bi, \quad
\Bj \Bk = \Bi = -\Bk \Bj, \quad
\Bk \Bi = \Bj = -\Bi \Bk \).
There is clearly more structure to these equations than the traditional Heaviside-Gibbs representation that we are used to, which says something for the quaternion model. However, this structure requires notation that is arguably non-intuitive. The fact that the quaterion representation was abandoned long ago by most electromagnetism researchers and engineers supports such an argument.

Minkowski tensor representation.

Minkowski introduced the concept of a complex time coordinate \( x_4 = i c t \) for special relativity [3]. Such a four-vector representation can be used for many of the relativistic four-vector pairs of electromagnetism, such as the current \((c\rho, \BJ)\), and the energy-momentum Lorentz force relations, and can also be applied to Maxwell’s equations
\begin{equation}\label{eqn:ece2500report:140}
\sum_{\mu= 1}^4 \PD{x_\mu}{F_{\mu\nu}} = – 4 \pi j_\nu.
\qquad
\sum_{\lambda\rho\mu=1}^4
\epsilon_{\mu\nu\lambda\rho}
\PD{x_\mu}{F_{\lambda\rho}} = 0,
\end{equation}
where
\begin{equation}\label{eqn:ece2500report:160}
F
=
\begin{bmatrix}
0 & B_z & -B_y & -i E_x \\
-B_z & 0 & B_x & -i E_y \\
B_y & -B_x & 0 & -i E_z \\
i E_x & i E_y & i E_z & 0
\end{bmatrix}.
\end{equation}
A rank-2 complex (Hermitian) tensor contains all six of the field components. Transformation of coordinates for this representation of the field may be performed exactly like the transformation for any other four-vector. This formalism is described nicely in [13], where the structure used is motivated by transformational requirements. One of the costs of this tensor representation is that we loose the clear separation of the electric and magnetic fields that we are so comfortable with. Another cost is that we loose the distinction between space and time, as separate space and time coordinates have to be projected out of a larger four vector. Both of these costs have theoretical benefits in some applications, particularly for high energy problems where relativity is important, but for the low velocity problems near and dear to electrical engineers who can freely treat space and time independently, the advantages are not clear.

Modern tensor formalism.

The Minkowski representation fell out of favour in theoretical physics, which settled on a real tensor representation that utilizes an explicit metric tensor \( g_{\mu\nu} = \pm \textrm{diag}(1, -1, -1, -1) \) to represent the complex inner products of special relativity. In this tensor formalism, Maxwell’s equations are also reduced to a set of two tensor relationships ([10], [8], [5]).
\begin{equation}\label{eqn:ece2500report:40}
\begin{aligned}
\partial_\mu F^{\mu \nu} &= \mu_0 J^\nu \\
\epsilon^{\alpha \beta \mu \nu} \partial_\beta F_{\mu \nu} &= 0,
\end{aligned}
\end{equation}
where \( F^{\mu\nu} \) is a \textit{real} rank-2 antisymmetric tensor that contains all six electric and magnetic field components, and \( J^\nu \) is a four-vector current containing both charge density and current density components. \Cref{eqn:ece2500report:40} provides a unified and simpler theoretical framework for electromagnetism, and is used extensively in physics but not engineering.

Differential forms.

It has been argued that a differential forms treatment of electromagnetism provides some of the same theoretical advantages as the tensor formalism, without the disadvantages of introducing a hellish mess of index manipulation into the mix. With differential forms it is also possible to express Maxwell’s equations as two equations. The free-space differential forms equivalent [4] to the tensor equations is
\begin{equation}\label{eqn:ece2500report:60}
\begin{aligned}
d \alpha &= 0 \\
d *\alpha &= 0,
\end{aligned}
\end{equation}
where
\begin{equation}\label{eqn:ece2500report:180}
\alpha = \lr{ E_1 dx^1 + E_2 dx^2 + E_3 dx^3 }(c dt) + H_1 dx^2 dx^3 + H_2 dx^3 dx^1 + H_3 dx^1 dx^2.
\end{equation}
One of the advantages of this representation is that it is valid even for curvilinear coordinate representations, which are handled naturally in differential forms. However, this formalism also comes with a number of costs. One cost (or benefit), like that of the tensor formalism, is that this is implicitly a relativistic approach subject to non-Euclidean orthonormality conditions \( (dx^i, dx^j) = \delta^{ij}, (dx^i, c dt) = 0, (c dt, c dt) = -1 \). Most grievous of the costs is the requirement to use differentials \( dx^1, dx^2, dx^3, c dt \), instead of a more familar set of basis vectors, even for non-curvilinear coordinates. This requirement is easily viewed as unnatural, and likely one of the reasons that electromagnetism with differential forms has never become popular.

Vector formalism.

Euclidean vector algebra, in particular the vector algebra and calculus of \( R^3 \), is the de-facto language of electrical engineering for electromagnetism. Maxwell’s equations in the Heaviside-Gibbs vector formalism are
\begin{equation}\label{eqn:ece2500report:20}
\begin{aligned}
\spacegrad \cross \BE &= – \PD{t}{\BB} \\
\spacegrad \cross \BH &= \BJ + \PD{t}{\BD} \\
\spacegrad \cdot \BD &= \rho \\
\spacegrad \cdot \BB &= 0.
\end{aligned}
\end{equation}
We are all intimately familiar with these equations, with the dot and the cross products, and with gradient, divergence and curl operations that are used to express them.
Given how comfortable we are with this mathematical formalism, there has to be a really good reason to switch to something else.

Space time algebra (geometric algebra).

An alternative to any of the electrodynamics formalisms described above is STA, the Space Time Algebra. STA is a relativistic geometric algebra that allows Maxwell’s equations to be combined into one equation ([2], [6])
\begin{equation}\label{eqn:ece2500report:80}
\grad F = J,
\end{equation}
where
\begin{equation}\label{eqn:ece2500report:200}
F = \BE + I c \BB \qquad (= \BE + I \eta \BH)
\end{equation}
is a bivector field containing both the electric and magnetic field “vectors”, \( \grad = \gamma^\mu \partial_\mu \) is the spacetime gradient, \( J \) is a four vector containing electric charge and current components, and \( I = \gamma_0 \gamma_1 \gamma_2 \gamma_3 \) is the spacetime pseudoscalar, the ordered product of the basis vectors \( \setlr{ \gamma_\mu } \). The STA representation is explicitly relativistic with a non-Euclidean relationships between the basis vectors \( \gamma_0 \cdot \gamma_0 = 1 = -\gamma_k \cdot \gamma_k, \forall k > 0 \). In this formalism “spatial” vectors \( \Bx = \sum_{k>0} \gamma_k \gamma_0 x^k \) are represented as spacetime bivectors, requiring a small slight of hand when switching between STA notation and conventional vector representation. Uncoincidentally \( F \) has exactly the same structure as the 2-form \(\alpha\) above, provided the differential 1-forms \( dx^\mu \) are replaced by the basis vectors \( \gamma_\mu \). However, there is a simple complex structure inherent in the STA form that is not obvious in the 2-form equivalent. The bivector representation of the field \( F \) directly encodes the antisymmetric nature of \( F^{\mu\nu} \) from the tensor formalism, and the tensor equivalents of most STA results can be calcualted easily.

Having a single PDE for all of Maxwell’s equations allows for direct Green’s function solution of the field, and has a number of other advantages. There is extensive literature exploring selected applications of STA to electrodynamics. Many theoretical results have been derived using this formalism that require significantly more complex approaches using conventional vector or tensor analysis. Unfortunately, much of the STA literature is inaccessible to the engineering student, practising engineers, or engineering instructors. To even start reading the literature, one must learn geometric algebra, aspects of special relativity and non-Euclidean geometry, generalized integration theory, and even some tensor analysis.

Paravector formalism (geometric algebra).

In the geometric algebra literature, there are a few authors who have endorsed the use of Euclidean geometric algebras for relativistic applications ([1], [14])
These authors use an Euclidean basis “vector” \( \Be_0 = 1 \) for the timelike direction, along with a standard Euclidean basis \( \setlr{ \Be_i } \) for the spatial directions. A hybrid scalar plus vector representation of four vectors, called paravectors is employed. Maxwell’s equation is written as a multivector equation
\begin{equation}\label{eqn:ece2500report:120}
\lr{ \spacegrad + \inv{c} \PD{t}{} } F = J,
\end{equation}
where \( J \) is a multivector source containing both the electric charge and currents, and \( c \) is the group velocity for the medium (assumed uniform and isometric). \( J \) may optionally include the (fictitious) magnetic charge and currents useful in antenna theory. The paravector formalism uses a the hybrid electromagnetic field representation of STA above, however, \( I = \Be_1 \Be_2 \Be_3 \) is interpreted as the \( R^3 \) pseudoscalar, the ordered product of the basis vectors \( \setlr{ \Be_i } \), and \( F \) represents a multivector with vector and bivector components. Unlike STA where \( \BE \) and \( \BB \) (or \( \BH \)) are interpretted as spacetime bivectors, here they are plain old Euclidian vectors in \( R^3 \), entirely consistent with conventional Heaviyside-Gibbs notation. Like the STA Maxwell’s equation, the paravector form is directly invertible using Green’s function techniques, without requiring the solution of equivalent second order potential problems, nor any requirement to take the derivatives of those potentials to determine the fields.

Lorentz transformation and manipulation of paravectors requires a variety of conjugation, real and imaginary operators, unlike STA where such operations have the same complex exponential structure as any 3D rotation expressed in geometric algebra. The advocates of the paravector representation argue that this provides an effective pedagogical bridge from Euclidean geometry to the Minkowski geometry of special relativity. This author agrees that this form of Maxwell’s equations is the natural choice for an introduction to electromagnetism using geometric algebra, but for relativistic operations, STA is a much more natural and less confusing choice.

Results.

The end product of this project was a fairly small self contained book, titled “Geometric Algebra for Electrical Engineers”. This book includes an introduction to Euclidean geometric algebra focused on \( R^2 \) and \( R^3 \) (64 pages), an introduction to geometric calculus and multivector Green’s functions (64 pages), and applications to electromagnetism (75 pages). This report summarizes results from this book, omitting most derivations, and attempts to provide an overview that may be used as a road map for the book for further exploration. Many of the fundamental results of electromagnetism are derived directly from the geometric algebra form of Maxwell’s equation in a streamlined and compact fashion. This includes some new results, and many of the existing non-relativistic results from the geometric algebra STA and paravector literature. It will be clear to the reader that it is often simpler to have the electric and magnetic on equal footing, and demonstrates this by deriving most results in terms of the total electromagnetic field \( F \). Many examples of how to extract the conventional electric and magnetic fields from the geometric algebra results expressed in terms of \( F \) are given as a bridge between the multivector and vector representations.

The aim of this work was to remove some of the prerequisite conceptual roadblocks that make electromagnetism using geometric algebra inaccessbile. In particular, this project explored non-relativistic applications of geometric algebra to electromagnetism. After derivation from the conventional Heaviside-Gibbs representation of Maxwell’s equations, the paravector representation of Maxwell’s equation is used as the starting point for of all subsequent analysis. However, the paravector literature includes a confusing set of conjugation and real and imaginary selection operations that are tailored for relativisitic applications. These are not neccessary for low velocity applications, and have been avoided completely with the aim of making the subject more accessibility to the engineer.

In the book an attempt has been made to avoid introducing as little new notation as possible. For example, some authors use special notation for the bivector valued magnetic field \( I \BB \), such as \( \boldsymbol{\mathcal{b}} \) or \( \Bcap \). Given the inconsistencies in the literature, \( I \BB \) (or \( I \BH \)) will be used explicitly for the bivector (magnetic) components of the total electromagnetic field \( F \). In the geometric algebra literature, there are conflicting conventions for the operator \( \spacegrad + (1/c) \PDi{t}{} \) which we will call the spacetime gradient after the STA equivalent. For examples of different notations for the spacetime gradient, see [9], [1], and [15]. In the book the spacetime gradient is always written out in full to avoid picking from or explaining some of the subtlties of the competing notations.

Some researchers will find it distasteful that STA and relativity have been avoided completely in this book. Maxwell’s equations are inherently relativistic, and STA expresses the relativistic aspects of electromagnetism in an exceptional and beautiful fashion. However, a student of this book will have learned the geometric algebra and calculus prerequisites of STA. This makes the STA literature much more accessible, especially since most of the results in the book can be trivially translated into STA notation.

References

[1] William Baylis. Electrodynamics: a modern geometric approach, volume 17. Springer Science \& Business Media, 2004.

[2] C. Doran and A.N. Lasenby. Geometric algebra for physicists. Cambridge University Press New York, Cambridge, UK, 1st edition, 2003.

[3] Albert Einstein. Relativity: The special and the general theory, chapter Minkowski’s Four-Dimensional Space. Princeton University Press, 2015. URL http://www.gutenberg.org/ebooks/5001.

[4] H. Flanders. Differential Forms With Applications to the Physical Sciences. Courier Dover Publications, 1989.

[5] David Jeffrey Griffiths and Reed College. Introduction to electrodynamics. Prentice hall Upper Saddle River, NJ, 3rd edition, 1999.

[6] David Hestenes. Space-time algebra, volume 1. Springer, 1966.

[7] Peter Michael Jack. Physical space as a quaternion structure, i: Maxwell equations. a brief note. arXiv preprint math-ph/0307038, 2003. URL https://arxiv.org/abs/math-ph/0307038.

[8] JD Jackson. Classical Electrodynamics. John Wiley and Sons, 2nd edition, 1975.

[9] Bernard Jancewicz. Multivectors and Clifford algebra in electrodynamics. World Scientific, 1988.

[10] L.D. Landau and E.M. Lifshitz. The classical theory of fields. Butterworth-Heinemann, 1980. ISBN 0750627689.

[11] James Clerk Maxwell. A treatise on electricity and magnetism, volume II. Merchant Books, 1881.

[12] James Clerk Maxwell. A treatise on electricity and magnetism, third edition, volume I. Dover publications, 1891.

[13] M. Schwartz. Principles of Electrodynamics. Dover Publications, 1987.

[14] Chappell et al. A simplified approach to electromagnetism using geometric algebra. arXiv preprint arXiv:1010.4947, 2010.

[15] Chappell et al. Geometric algebra for electrical and electronic engineers. 2014.

[16] Chappell et al. Geometric Algebra for Electrical and Electronic Engineers, 2014