5

Background

In many textbooks on electromagnetism, the sequence in which topics are introduced are generally as follows -- electrostatics, magnetostatics, electrodynamics, wave propagation in free space, and wave propagation in confined space, in that order. As an example, consider Griffiths' 3rd edition of Introduction to Electrodynamics. Similarly, consider Orfanidis' Electromagnetic Waves and Antennas.

By using the discussion on wave propagation in free space as a launchpad, the discussion on wave propagation in a confined space usually introduces some additional boundary conditions that the solutions to Maxwell's equations for wave propagation in free space did not have.

If a proposed solution can be plugged into Maxwell's equations and not produce any contradictions, then they are valid. The goal is then to maintain the monochromatic wave solutions that worked for free space wave propagation with some "patches" to account for the fact that the propagating EM energy will be confined by the hollow waveguide. The reason monochromatic (one-wavelength) wave solutions are desirable is that, by superposition, any signal can be expressed as a sum of one or more monochromatic waves, and so any signal's behavior in a hollow waveguide can be understood.

One of these "patches" is the boundary condition that $E_{\parallel}=0$. This is because the net electric field in an ideal conductor, such as the material that makes up the confining hollow waveguide, must be $0$ because any externally applied electric field will influence electrons to rush to counter the externally applied electric field. If there is no net electric field (meaning $\vec{E}=\vec{0}$), then there is no net electric field in the tangential direction ($E_{\parallel}=0$) along and inside the metal.

If there is no net electric field along the metal, then, by Faraday's law, there must be no net changing magnetic field $\frac{\partial \vec{H}}{\partial t}$. This means that each of the components $\frac{\partial H_x}{\partial t}$, $\frac{\partial H_y}{\partial t}$, and $\frac{\partial H_z}{\partial t}$ must all be $0$ too. If there is no initial magnetic field (meaning $\vec{H}=\vec{0}$, then there must be no magnetic field at any later point. Therefore $\vec{H}=\vec{0}$ and $H_{\perp}=0$.

So these are the two additional boundary conditions that must be considered. An implicit assumption is that the material on the interior of the waveguide is isotropic, homogeneous, and linear. This means that $\vec{B}=\mu\vec{H}$ and therefore $\vec{B}=\vec{0}$ and $B_{\perp}=0$ at the boundary of the metal + hollow interior interface as well.

Problem

Frequently, the monochromatic wave solution to Maxwell's equations given the boundary conditions of a hollow waveguide are presented as

$$ \vec{F}\left(x,y,z,t\right)=\left[\text{spatial component}\right]\left[\text{temporal component}\right] $$

For example from Orfanidis,

$$ \vec{E}\left(x,y,z,t\right)=\vec{E}\left(x,y\right)e^{j\omega t - jkz} $$

and from Griffiths,

$$ \vec{E}\left(x,y,z,t\right)=\widetilde{\vec{E}}_0\left(x,y\right)e^{i\left(kz - \omega t\right)} $$

Further, $\vec{E}$ is a 3-dimensional vector with components $E_x\left(x,y\right)$, $E_y\left(x,y\right)$, and $E_z\left(x,y\right)$.

Bottom line: Why is the spatial component only a function of $x$ and $y$? Further, why does spatial component have $x$, $y$, $z$ components [e.g. $E_x$, $E_y$, and $E_z$] (where $E_x$, $E_y$, and $E_z$ are each functions of only $x$ and $y$)?

3 Answers3

4

To begin with, Maxwell's equations are vector partial differential equations, which makes their solution intrinsically more difficult than that of scalar partial differential equations.

To help obviate this difficulty, an ansatz was long ago made (by Hertz himself, perhaps?) who proposed two useful categories of wave solution: transverse electric (TE), and transverse magnetic (TM). The term 'Transverse' here means 'transverse to the waveguide axis', which is conventionally the 'z' direction, and the presumed direction of propagation.

These categories are somewhat self-explanatory. TM waves have no z directed magnetic field component, and TE waves have no z directed electric field component. Each of these is derivable as the curl of an appropriate vector potential, magnetic or electric. If the idea of an electric vector potential is troublesome to you, try the book by Harrington 'Time Harmonic Electromagnetic Fields.' (With no denigration of Griffiths, I advise that you widen the scope of your reading.) In either case the vector potential is assumed to have a single, z-directed, component, comprising a scalar function of the form f(x,y)exp(ikz). This entire scalar function is presumed to satisfy the scalar Helmholtz equation (with propagation constant k), which insures that taking its curl (and the curl of the resultant) will lead to Maxwell's equations.

In essence, TE and TM waves are assumed; these ideas came out of someone's brain. Their justification is that they lead us (as noted) to Maxwell's equations, and from there to innumerable useful solutions and applications. Inasmuch as a waveguide is designed to propagate a monochromatic wave of some chosen frequency, the simplest form of traveling wave solution is exp(ikz), as noted. As far as why one wants a monochromatic wave, one answer would be that (at a guess) >90 % of communication electronics involves the modulation of a carrier wave, with the frequency bandwidth of modulation typically small in comparison to the carrier frequency.

Sorry to go on at such length. I hope this helps.

  • Historical background and long answers are always appreciated. No worries. The book I'm using is actually Sadiku's Elements of Electromagnetics though I have referred to Hayt and Buck, Griffiths, Orfanidis, and various lecture notes online from different universities. I appreciate your suggestion on Harrington (especially since a vector electric potential is a foreign concept to me). My background is in electrical engineering and I am doing this to prepare for my preliminary examinations for my PhD. – Sophisticated Idiot Sep 18 '13 at 02:14
  • I have seen how TE and TM come out of setting $E_z=0$ or $B_z=0$/$H_z=0$ in two separate texts, but I don't get the intuition. It appears as though you are saying the $f(x,y)$ part is just assumed. But why not assume $f(x,y,z)$ instead? Your note about the motivation to study monochromatic waves with regards to communication systems is duly noted. – Sophisticated Idiot Sep 18 '13 at 02:17
  • To limit the complexity there are assumptions you can make about the structure. Afterwards it is common to go back and test for other EM modes to see if the assumption is valid and if it is possible the wave guide will support those modes based on the structure. TE11, TM10 etc. Take a look at http://whites.sdsmt.edu/classes/ee481/notes/481Lecture10.pdf – user6972 Sep 18 '13 at 02:26
  • So as you said, it was an assumption that $f(x,y)$ ($f$ is a function of $x$ and $y$) and the solution that was tested worked. So why not reduce the complexity further and try $f(x)$? or $f(y)$? Would those not work? That sdsmt.edu link and all the lectures leading up to it are very useful. Thanks. The way the sdsmt.edu course covers this waveguide material is similar to the way Sadiku's 3rd edition of Elements of Electromagnetics covers it. Sadiku assumes $\vec{E}\left(x,y,z,t\right)$ and that $E_x$, $E_y$, and $E_z$ are functions of $x$, $y$, and $z$. (consider Eq. 12.5 in Sadiku). – Sophisticated Idiot Sep 18 '13 at 04:34
  • 1
    I'd say that if you reduce it to f(x) (for example) you wouldn't in general be able to solve the boundary conditions at the waveguide walls. As for why not f(x,y,z), the z dependence is subsumed in the assumption of a homogeneous guide supporting a travelling monochromatic wave. – hyperpolarizer Sep 18 '13 at 06:42
  • I'm not sure why omitting one extra variable ($y$) beyond just omitting $z$ from $f(x,y,z)$ would make it so you couldn't solve the boundary conditions at the waveguide walls in general. Can you expand a bit on how $z$ dependence is absorbed into the assumption of a homogeneous guide supporting a traveling monochromatic wave? I see that $z$ is included in the complex exponential. Is that what you mean? So technically $f(x,y)$ is multiplied by $g(z)=\exp\left(\cdots z\cdots\right)$? – Sophisticated Idiot Sep 18 '13 at 21:45
2

About your "bottom line": there is translational invariance along $z$ in the problem. so the electric field can be expanded into a Fourier integral with respect to $z$. So there is dependence on $z$, but it can be expressed as a superposition of exponents.

Some waveguide modes do have nonzero $z$-component of electric field (although this component is zero on the boundaries of the ideally conducting waveguide).

akhmeteli
  • 26,888
  • 2
  • 27
  • 65
  • What does translational invariance along $z$ mean? Translational invariance along $z$, to me, means that some quantity does not change with a change in $z$. Are you saying that the electric field does not change with respect to $z$? If so, why does it not change with respect to $z$? I think I may be unfamiliar with the Fourier integral as well. Is this the same thing as the Fourier transform which takes a function in the time domain and churns out a function in the frequency domain? Can you expand further on the dependence on $z$ and what superposition of exponents are involved please? – Sophisticated Idiot Sep 18 '13 at 01:55
  • Boundary conditions are invariant with respect to translations along $z$, and coefficients of the PDE are translation-invariant, so variable $z$ can be separated (see, e.g., http://math.mit.edu/~stevenj/18.303/separation.pdf ). In this case, Fourier transform takes a function depending on $z$ and transforms it into a function depending on the spatial frequency (a component of the wave vector) $k_z$. – akhmeteli Sep 18 '13 at 04:33
  • In other words, the boundary conditions imposed on the electromagnetic wave as it hits the metal/air interface do not change along the $z$ axis. I am not sure what the coefficients of the PDE are. The PDF goes into a lot of terms that are very much foreign to me. I understand your comment about the Fourier transform taking a function in $z$ (space) and transforming it into a function in $k$ (the spatial frequency variable, which is analogous to the Hertzian frequency $f$ or the angular frequency $\omega$). Can you elaborate on how boundary conditions that are invariant in $z$ affect $E_x$? – Sophisticated Idiot Sep 18 '13 at 21:42
  • @Sophisticated Idiot: PDE means partial differential equation. I did not understand the question about $E_x$: why and what about $E_x$? – akhmeteli Sep 22 '13 at 19:05
2

Akhmeteli is absolutely spot on. Since he/she has thrust to the problem's very heart with his / her keen dagger, just a few more words to clean up the lacerations on the wound!

It IS all about z-axis shift invariance: so this symmetry requires the existence of solutions which are eigenfunctions of $\partial_z$, namely functions of the form $f(x,y) \exp(i\,k_z\,z)$. Note that this form gainsays your assertion that the spatial component is a function of $x$ and $y$ only: there is the $\exp(i\,k_z\,z)$ as well. As you see, the $z$ variation is so simple that you can often overlook it (I have made this very mistake many times myself).

To understand the solution to this problem at a practical level, imagine we are solving this problem numerically and represent the transverse field as some numerical grid: we can rewrite a 2D grid into a 1D column vector $\psi$ with $N = n\times m$ complex components where the transverse 2D grid resolution is $n$ points by $m$ points. Maxwell’s equations are linear so that the action of a length $z$ of waveguide can be represented by a big square $N\times N$ matrix $U(z)$, i.e. $\psi \mapsto U\,\psi$. Now we think about concatenating lengths of waveguide. Owing to our translational invariance, we have:

$$U(z_1)\,U(z_2) = U(z_2)\,U(z_1) = U(z_1 + z_2)$$

for any lengths $z_1,\,z_2\in\mathbb{R}$, and the only continuous solution to this functional equation is $U(z) = \exp(H\,z)$, for some constant square $N\times N$ matrix $H$. Now if we set up our column vector right, its $\ell^2$-length (i.e. its norm) $\sqrt{\psi^\dagger\psi}$ is proportional to the total propagating power through any transverse cross section of the waveguide. If the waveguide is lossless, this norm cannot change, so that $U(z)$ is also unitary (preserves lengths of and angles between vectors) and this can be shown to be equivalent to the condition $H^\dagger = -H$, i.e. $H$ is skew Hermitian. The spectral theory for finite normal operators (i.e. ones that commute with their adjoint, so square skew Hermitian matrices are such operators) is very well studied and easy: such a matrix can always be diagonalised by a unitary matrix of $N$ orthonormal column vectors, and so the matrix exponential $U(z) = \exp(H\,z)$ can be expressed in the form:

$$U(z) = \Omega \, \operatorname{diag}[e^{i\,k_1\,z}, \, e^{i\,k_2\,z}, \, e^{i\,k_3\,z}, \,\cdots] \Omega^\dagger$$

where now the $ e^{i\,k_1\,z}$ are all scalar exponentials and the $k_j$ are all real (this together with the unitary decomposition is equivalent to skew-Hermitianhood of $H$). So now think what the eigenvectors – the columns of $\Omega$ mean, given that this is a numerical problem. They stand for certain special field configurations in the transverse plane, i.e. they are discretised versions of the eigenfunction variation $f_j(x, y)$. So, through this spectral factorization, you can see all numerical solutions to the problem are linear superpositions of the kind:

$$\sum_j\alpha_j\,f_j(x,\,y)\, \exp(i\,k_j\,z)$$

(for the numerical problem, you can think of $x$ and $y$ as discrete grid indices passing to continuous transverse co-ordinates in the limit and the $f_j$ are our "grid" eigenfields) and hopefully you can accept that this numerical representation will approach the real one as the grid spacing gets smaller and smaller, so the above is an explanation for the kinds of solutions you are asking about. The numerical solution of this problem seldom works like this - $N = n\times m$ is huge (think of a problem with a $1024\times1024$ transverse discretization) and there are much more efficient ways of representing the action of propagation through an axial step $\Delta z$ than to work out a matrix $U(\Delta z)$ or $H$ (which would hold $10^{12}$ complex numbers in the $1024\times1024$ grid example!), but you can see the above is an in-principle description: there is a square unitary grid operator and it can always be factorized in the way I describe. So the numerical procedure is always equivalent to the above: one just finds clever algorithms that are equivalent in their action but do not need to build all of the huge $U(\Delta z)$ in memory at once. Finite difference methods, for example, assume, accurately, that over small steps, the field at a point in the updated grid is only influenced by the fields at neighbouring points (for first order PDE problems) or at neighbouring and next-to-neighbouring points (for second order PDE problems, as Maxwell's equations are). So we only need to represent in memory the leading diagonal together with two neighbouring stripes of the matrices $H$ and $U(z)$: this is roughly $5\times 1024\times 1024$ points in my example, or about $16\times 5 = 80$ megabytes on a 64-bit machine representing complex numbers as two one-word (8 bytes) reals, i.e. readily workable in today's technology. The co-efficients in the $H$ matrix come straight from discretized versions of Maxwell's equations and tridiagonal (three stripe) pentadiagonal (five stripe) matrix factorization numerical procedures are very well worked out, tested and numerically sound.

Incidentally, the above discussion is wholly analogous to the theory behind the general Schrödinger picture in quantum mechanics: see my answers to the Physics SE questions A Simple Explanation for the Schrödinger Equation and Model of Atom?, Reference frame involved in the Schrödinger's equation and Why can the Schroedinger equation be used with a time-dependent Hamiltonian?.

In non-numerical simulation words: in a lossless waveguide, the $z$-shift operation is indeed unitary (no energy loss) so the "infinitesimal" shift $\partial_z$ is skew-Hermitian i.e. the eigenvalues $i\,k_z$ are purely imaginary. You need a bit of theory - a vector analogue of Sturm-Liouville theory - to see why such eigenfunctions can be added together to match arbitrary boundary conditions on a transverse cross-section. Actually, given the vector nature and further conditions of solenoidal fields ($\nabla\cdot\vec{E}=\nabla\cdot\vec{H}=0$), there's a bit more to a proof than standard Sturm-Liouville theory. Standard SL theory will work of course for two out of three Cartesian components of the $\vec{E}$ or $\vec{H}$ field (or Cartesian component of the Lorenz-gauged potential four-vector). For a metal waveguide with a compact cross-section, the spectrum of eigenfunctions is wholly discrete: for an optical fibre, there is a continuous component to the spectrum - the radiation field - as well as the discrete, losslessly guided bound eigenfields.

The full story is to be found in Snyder and Love, "Optical Waveguide Theory", particularly chapters 30 and 31 for groundings and earlier chapters for the radiation field of optical waveguides (I understand from John Love that the ANU will make this book freely downloadable sometime) and R. E. Collins, "Field Theory of Guided Waves". However, neither of these books is particularly up front about the mathematical foundations of the "completeness" of the vector eigenfunctions and I don't know of any reference that discusses this idea in simple, modern terms.

  • Can you also expand upon $z$-axis shift invariance for me? My linear algebra is also very weak (something Khan Academy will help me with soon), so I will have to look into eigenfunctions again. I can't say I've ever seen that notation ($\partial_z$) previously. But the conclusion you draw from the stuff you said up until that point -- regarding the existence of solutions of the form $f(x,y)\exp\left(ik_zz\right)$ is what want. I do see the $\exp\left(ik_zz\right)$-like terms in Griffiths and Orfanidis and so there is a $z$-dependence of $E\left(x,y,z,t\right)$. I lost you after "unitary." – Sophisticated Idiot Sep 18 '13 at 02:01
  • @SophisticatedIdiot I have just updated my answer with an explanation in numerical simulation terms. Hopefully this will be clearer to you. BTW you are pretty sophisticated for an idiot to be asking such clear, probing questions - I have no doubt you'll understand all this in time and indeed trying to think in numerical simulation terms like this is a good way to approach linear algebra, rather than the other way around. – Selene Routley Sep 18 '13 at 03:12
  • @SophisticatedIdiot added a little more detail on how numerical procedures actually look and also the similarities with the Schrödinger equation – Selene Routley Sep 18 '13 at 04:49
  • I'll keep this in mind. The information you presented feels very much out of reach for now, mainly because it seems written in a way that is intuitive for a person with a lot more knowledge than me. The link to quantum mechanics is also appreciated as I will be delving into that shortly after I finish off electromagnetism. – Sophisticated Idiot Sep 18 '13 at 21:47
  • @SophisticatedIdiot Then I believe akhmeteli's answer is the one for you. It is all about the symmetry of z-shift invariance. Why this begets the kinds of solutions you see is something you really do need a little linear algebra to explain. But the essential idea is representing a short section of length $\Delta z$ the waveguide as a "black box" that has some transfer matrix $U(\Delta z)$, then a concatenation of a large number $M$ of them is $U(\Delta z)^M$ because you can chop any length of a translationally invariant waveguide up into pieces of any lengths and then paste them back ... – Selene Routley Sep 19 '13 at 01:20
  • @SophisticatedIdiot ....together in any order and the result has to be the same. So now consider chopping a length $L$ into $M$ pieces: the result is $U(\frac{L}{M})^M \left(I + \mathrm{d}_z U \times \frac{L}{M} + O(\left(\frac{L}{M}\right)^2\right)^M$, which approaches $\exp(H,L)$ where $H = \mathrm{d}_z U$ (and $\mathrm{d}_z U$ is constant, by translational invariance), by definition of the exponential (its universally convergent Taylor series means you can work with the matrix argument here as though everything were scalar). From there, it's all linear algebra and the factorisation of ... – Selene Routley Sep 19 '13 at 01:23
  • @SophisticatedIdiot ...matrices (or operators). So the exponential behaviours arise fundamentally from the symmetry. Although all this may seem a bit above you at the moment, if you keep asking probing questions like you are, you may well surprise yourself how quickly you'll grasp it all. On second thoughts, maybe your username isn't too inappropriate, but rather a kind of badge of honour because ironically my experience is that only those people who understand that the best any of us can ever become are "sophisticated idiots" will pierce the mysteries of the world. ... – Selene Routley Sep 19 '13 at 01:31
  • @SophisticatedIdiot ...It really sounds as though learning is going to be a rewarding joy for you! PS: My formula in my other comment should have read $U\left(\frac{L}{M}\right)^M = \left(I + \mathrm{d}_z U \times \frac{L}{M} + O\left(\left(\frac{L}{M}\right)^2\right)\right)^M \to \exp\left(H,L\right)$ as $M\to\infty$. – Selene Routley Sep 19 '13 at 01:35