Showing posts with label Science Jar. Show all posts
Showing posts with label Science Jar. Show all posts

Saturday, March 6, 2021

Electrical Current

This is a question on many student’s minds: what are the electrons actually doing inside a current-carrying wire? 

The answer might seem straightforward at first glance. An electrical current is a flow of charge moving through an electrical conductor or through space. From electronics-notes.com, we have a practical definition of electrical current: it is “the rate of change of flow past a given point in an electric circuit.” From physicsclassroom.com (an excellent teaching site), the definition alludes to a bit of scientific history: An electric current is “by convention the direction in which a positive charge would move.” Electrons “move through the wires in the opposite direction.” Confusing for many students, conventional current in a wire moves from the positive terminal of a battery to the negative terminal. Electrons, being negative charge carriers, actually move in the opposite direction. This might be where many of us stop in our quest to understand current, which is unfortunate because this is a fascinating exploration.

 

The silver lining of this conventional terminology is that it nudges us toward the history, which shows an incredible advancement in solid-state physics. How was electrical current discovered? We probably all know about Benjamin Franklin’s famous kite experiment in 1752. We might not know that he didn’t actually discover electricity with this experiment, nor was he the first to discover that lightning was actually a type of electricity. Electrical forces had been known for hundreds of years by his time. I think, however, we can fairly credit him for contributing to our confusion about current. He studied static electricity  by producing a static charge on the surface of glass, amber and other materials by rubbing them with fur or a dry cloth. This resulted in an exchange of electrons from one material to another. At that time, electrical current was called “electrical fluid,” and as such, Franklin guessed that some materials (such as glass that’s rubbed) contained more of this fluid than others. To his thinking, these charged objects contained excess, or positive, electricity, while others contained a deficiency of the fluid, or negative electricity. Electric batteries were developed soon afterward and it seemed natural to assign the direction of electrical flow from positive to negative (excess to deficient). It was only when electrons, the subatomic particles responsible for static charge, were discovered about one hundred years later, that scientists realized these particles move in the opposite direction. It is an excess of electrons that produces a negative charge, so the flow from excess to deficient must actually be from a negative terminal to a positive terminal. 

 

This “conventional  current” (positive to negative) had staying power. The conventional terminal designation is still used worldwide. It’s not a problem to work with as long as it is consistent, but it can present a problem when we try to understand what is actually happening inside the conductor.

 

What then is going on inside an electrical current-carrying wire, for example? Consider an everyday power cord on a vacuum cleaner. If we could zoom into a cross-section of that cord, we would see wires made of copper through which electrons travel easily and with little resistance, surrounded by material that resists current flow and provides good electrical insulation. We can imagine electrons flowing from the electrical outlet in the wall, through the cord, and into the appliance. But where do these flowing electrons end up? Do they get used up in the process of doing work somehow? When the power is shut off, are we left with an alarming reservoir of electrons somewhere inside the vacuum motor? Where do the electrons in the wall socket originate from? These great questions start us on our journey. 

 

When we think of electrical current as a physical flow of negatively charged particles through a material or through space, we might think of something analogous to water molecules flowing in a stream, and when we do, a number of pressing questions come to mind. Like the out-dated convention of positive to negative terminal current flow, the terms “current” and “flow” themselves lead us away from clear modern evidence that electrical current is a not a physical flow of particles at all. Many physics classrooms begin their discussion of electrical current with a water analogy and it is a good place to go to get a feel for how simple electrical circuits work. But this analogy, while a good start, proves misleading as we deepen our understanding. And there is much more to this fascinating story than this. To learn it we must upgrade our understanding of what electrons are doing at the subatomic level inside a conductor.

 

What Is an Electron?

 

All materials are made of atoms. All atoms consist of a nucleus surrounded by electrons. The nucleus is composed of neutrally charged particles straightforwardly called neutrons and positively charged protons. It therefore has a positive charge. These particles are bound tightly together by a fundamental force called the strong force (strong force). This force is indeed strong. It easily overcomes the repulsive forces between the positively charged protons, but it only acts over an extremely short distance, at the scale of the nucleus itself, and from there, its influence drops off dramatically to zero. Negatively charged particles called electrons surround the nucleus. They are attracted to the nucleus through the attraction of opposite electrical charges. An electrically neutral atom contains equal numbers of electrons as protons. 

 

We might imagine electrons moving around the nucleus in planet-like circular orbits, except that here, the atomic force is electrostatic rather than gravitational. This is the familiar Bohr model (below) introduced by Niels Bohr and Ernest Rutherford in 1913. This model of the hydrogen atom shows three possible energy shells for the electron. At n=1, the electron is at its lowest energy (ground) state. If the atom is in an excited state, the electron will be in a higher energy shell. It will emit a photon of light as it returns to a lower energy state.


The idea that electrons orbit the nucleus in specific stable orbits came from necessity. The researchers knew that atomic electrons can release energy in the form of electromagnetic radiation (light) but they also knew that if an electron loses energy, it should quickly spiral into the nucleus, and in the process it would emit increasingly high frequency radiation (called the ultraviolet catastrophe) No atom would be stable for more than a few trillionths of a second. They also knew, from experiments a few decades prior, that atoms emit light only at specific frequencies. An excited (high energy) hydrogen atom will emit only purple, blue, aqua or red (the Balmer series emission spectrum, shown below), depending on the energy level of its excited-state electron, as it returns to its rest state. 


In an excited pure hydrogen gas, you will see the whole spectrum but hydrogen will never emit green or yellow, for example. These scientists figured out that electrons in atoms must occupy discrete energy levels. They orbit at specific stable distances from the nucleus. The farther an electron is form the nucleus, the higher its energy level. An electron can move up or down only by jumping to/from specific energy levels. Energies are therefore quantized. They come in quanta or packets.  This model is useful for predicting the spectral phenomena of simple atoms with few electrons, like hydrogen, but it cannot explain the spectra of large complex atoms. Nor can it explain the different intensities of spectral lines for any given atom. It is also still used as a simplified model for chemical bonding between atoms, in which atoms share one or more electrons located at their outermost energy levels.

 

The modern model of the atom, the quantum physical model, describes the positions of atomic electrons not as being precisely located within energy shells but as clouds of probability. The hydrogen atom is shown below right. The shapes of the electron orbitals are shown in yellow and blue. Denser regions of colour indicate a higher probability of the electron's location. The energy shells (1, 2 and 3) are arranged in increasing energy from top to bottom. The orbital shapes are s, p and d, shown from left to right.


This model incorporates the fact that we no longer understand electrons to be just point charges, or tiny charge-carrying billiard balls. Under certain circumstances they do act as such. For example, an electron has a measurable momentum and it can take part in elastic collisions. But electrons also act as waves and that wave nature also shows up under certain circumstances, such the creation of wavelike interference patterns. We now think of the electron as both particle and wave, which is not easy to grasp. These qualities seem to be mutually exclusive. Electrons in conductors, however, do display both wave and particle behaviours. To predict such behaviours, we must go quantum and understand electrons as wave functions that exhibit both particle AND wave behaviour. An electron’s position and momentum (or velocity) are now defined as probabilities. Where and how fast an electron is going are assigned probability amplitudes, rather than specific values. Furthermore, thanks to Werner Heisenberg’s uncertainty principle of quantum mechanics, these are complementary variables, which means we can only know one value at the expense of another (this part was actually formulated by Bohr). If we want to know precisely where an electron is, we cannot know its momentum at that same moment, and vice versa. Likewise, the electron’s wave and particle properties are also complementary. A single electron cannot simultaneously exhibit both its full wave-like and particle-like nature, with the exception of the famously fascinating double-slit experiment, in which electrons show some of both behaviours at the same, which I encourage you to look up.

 

This is a difficult conceptual step to make but we must because the easy-to-visualize Bohr model, as useful as it is, leaves something missing. By using Schrodinger’s equation to mathematically describe atomic electron behaviour as a wave function, physicists can predict many of the spectral phenomena that the Bohr model cannot. It’s not easy to do in practise either; the calculations are very complex.

 

What is an Electrical Conductor?

 

The atoms that make up a good electrical conductor have an atomic structure that allows the outermost electrons, called valence electrons, to be loosely bound to the nucleus. Valence electrons are outermost energy shell electrons. These are the electrons that take part in chemical bonding. Now we can describe chemical bonding more precisely, in quantum mechanical terms, as two or more atomic orbitals combining to form a molecular orbital. An atomic orbital is a region of space around the nucleus where an electron is most likely to be. It can be a simple spherical shape or a more complex shape depending on its energy level. The orbital shapes come from solving the Schrodinger equation for electrons bound to their atom through the electric field created by the nucleus. The orbital is part of the electron wavefunction that describes the electron’s location boundary and its wavelike behaviour.

 

Molecular orbitals, on the other hand, come in three different kinds: 1) bonding orbital, which has a lower potential energy than the atomic orbitals it is formed from, 2) antibonding orbital, which has a higher energy than the atomic orbitals, so it opposes chemical bonding and 3) nonbonding orbital, which has the same energy so it has no effect on chemical bonding either way. A chemical bond is a constructive, in-phase, interaction between two valence electrons of two atoms. An antibonding orbital is an interaction between atomic orbitals that is destructive and out of phase. The wavefunction of an antibonding orbital is zero between the two atoms. This means there are no solutions to the Schrodinger equation for this orbital and therefore there is no probability of an electron being available to bond. 

 

Metals tend to be good conductors because their valence electrons are loosely bound to their nuclei. They are delocalized electrons, which means that they are not associated with a particular atom or chemical bond. These molecular orbital electrons extend outward over many atoms. Although the electrons are delocalized, the metal atoms themselves are bound tightly together through metallic bonding, by electrostatic attractions between the positive nuclei and the “sea” of electrons in which they are embedded. The atoms are held tightly together, which means metals tend to have high melting and boiling points. The nuclei in metals act like positive ions in this arrangement, which means that metallic bonding is similar in this sense to ionic bonding. The resulting metal structure is a tight three-dimensional lattice arrangement, similar to the atomic structure of ionic crystals such as sodium chloride (table salt). In contrast to ionic compounds, valence electrons in the metallic lattice form molecular orbitals that extend across the entire metal. The bonding electrons themselves do not orbit the entire metal but their influence extends across the metal. Valence electrons in metals act more like a collective than they do in conventional chemical bonds. 

 

All of the valence electrons in the metal participate in molecular bonding. As vast as this collection of delocalized electrons is, the number of possible delocalized electron energy states is far greater. All metal atoms contain few electrons in their valence energy orbitals. Transition metal atoms, for example, can hold up 18 electrons in the outermost energy shell (which consists of 5 d orbitals, one s orbital and three p orbitals), but they are barely filled. If this is confusing, we will be exploring this in more detail later. The point here is there are many more possible energy states than there are electrons available to fill them. These empty available states, which are all similar in terms of energy, will become important when we look at band theory later on in this article.

 

What Happens When an Electric Potential is Applied?

 

When a metal with high conductivity is placed in an electric field, all the valence electrons tend to move against the direction of the field. An electric field is a vector force field that surrounds an electrical charge and exerts a force on other charges nearby. The electrons themselves each generate an electric field, as do the positively charge nuclei in the metal. When electrons within the metal move, they also generate a magnetic field. We can begin to see electric current as a complex interplay between multiple local electric and magnetic fields.

 

An analogy of a waterfall is often used to describe the movement of charged particles, such as electrons, as moving along or down an electric potential. An electric potential might at first sound the same as an electric field but the electric potential expresses the effect of an electric field at a particular location in the field. We could place a test charge within an electric field and measure its potential energy as a result of being in that field. (A positive test charge is usually used and this is why electrons move against the direction of an electric field.) The potential energy of the electric field will differ, depending on the location within that field. A negative test charge, for example, will have high potential energy close to a negative source charge and lower energy further away from it. In other words, the electron tends to move from where it would have high potential energy toward lower potential energy. Inside a copper wire attached to a battery, for example, the electron is pushed away from the high-energy negative terminal (labelled positive!) and toward the region of lowest electric potential energy, the positive terminal (labeled negative!). This difference in electric potential energy is called voltage. Voltage is defined as the amount of work required per unit of charge to move a test charge between two points. 1 volt = 1 joule of work done per 1 coulomb of charge. For example, if a 12-volt battery is used in a wire circuit to power a light bulb, every coulomb of charge in the circuit gains 12 joules of potential energy it moves through the battery. Every coulomb in turn loses 12 joules of energy into the environment as light (and some heat) as it powers the light bulb.

 

The waterfall analogy, as mentioned earlier, can be misleading. Within the wire, the electrical current does not depend on electrons wriggling free from copper atoms and flowing down the wire, like water molecules would flow down a waterfall. The description of a “delocalized sea of electrons” can naturally lead to such an assumption about current. Instead, each valence electron is held loosely enough to its nucleus to “nudge” a neighbouring electron in a neighbouring atom and so on. A better analogy for electrical current might be that of fans in a football stadium row standing up one after another to do the wave. The people stay in place but the wave moves down the row. If we want to stick with water, we could say that the current is analagous to the wave travelling across a sea. Electric current can be defined as the rate at which charge (not electrons) flows past a point in a circuit. 1 ampere of current = 1 coulomb of charge/second.

 

It’s easy to imagine a flow of electrons under pressure (an analogy for voltage) gliding along inside a wire like water flowing through a pipe, but current is not the physical flow of electrons, but rather the flow of the energy of their movement. We can put this idea more scientifically as a momentum transfer. Put yet another way, it is a transfer of kinetic energy from one electron to another and another and so on. We can think of nudged electrons transferring energy from one to the next like a billiard ball hitting an adjacent billiard ball along a line of billiard balls. This analogy alludes to the particle nature of the electrons.

 

Electrical Conduction: The Drude Model

 

The “billiard ball” description of electrical conduction is called the Drude model, proposed in 1900. This model was a bit before it’s time because it predated even the (Rutherford) atomic model by a few years. The “sea of electrons” embedded within a positive matrix that Rutherford envisioned (also called the raisin bun or plum pudding model) just happens to align pretty well with the Bohr valence electron model for metals, where a sea of valence electrons surrounds each atom. The Drude model is essentially  a description of how kinetic energy is transferred within a conductor by treating electrons like tiny billiard balls. For those science history buffs like me, we can follow how the Drude model evolved into our modern model. We can read into it how our understanding of electron behaviour evolved. The Drude model was not a modern quantum mechanical model of what is happening, but it could predict some electron conduction behaviour by understanding conduction as a momentum transfer between electrons exhibiting particle-like behaviour. In fact, Paul Drude used Maxwell-Boltzmann statistics to derive his model. These statistics describe an average distribution of non-interacting particles at thermal equilibrium. This is a classical description of how an ideal gas behaves, and it was the model available to him at the time. In this view, each atom in a conducting metal such as copper contributes valence electrons to a sea of non-localized, non-interacting electrons. It turns out he was accidentally right. In most cases, at least in the case of copper, we can effectively neglect the interactive forces between the electrons because they are shielded from those forces.  All the relatively stationary (much more massive and more highly charged) atomic nuclei in the copper present an overwhelming influence on them. It is this shielding effect) that allows us to model metal electrons fairly accurately as an ideal gas, a cloud of particles that do not interact with each other. However, the particles, in this case electrons, are far more concentrated together than the atoms in any gas would be. This means that this model doesn’t predict all conduction behaviour. Electrical conduction exhibits behaviours that are too complex to be described through classical ideal gas theory. 

 

Maxwell-Boltzmann statistics were replaced around 1926 by Fermi-Dirac statistics. Rather than focusing on the non-interaction between particles, these statistics describe a distribution of particles over a range of energy states in a group of identical particles that all obey the Pauli exclusion principle. This effectively upgraded the rule for interaction. Rather than saying the electrons don’t interact, we now say they cannot occupy the same quantum state. We now have to treat electrons as quantum particles. Everything about a quantum particle is described by four quantum numbers (spin, magnetic, azimuthal and principle). Electrons belong to a special division of quantum particles called fermions. Fermions can share any three quantum numbers but they can’t overlap in the same location at the same time. A boson such as a photon can, and it is governed by a different set of statistics. We are introducing the Pauli exclusion principle to the gas-like behaviour of valence electrons inside a metal. The Pauli exclusion principle is a critically important rule that electrons in atoms (and all fermions) must obey. It is a quantum mechanical principle. The implications are quite profound. It means that the potential energy of the gas-like electrons must now be taken into account, and by doing so, it limits the numbers of electrons that occupy each orbital in an atom. This rule forbids valence electrons from moving down to occupy already filled lower energy states in the atom. This means there is a lower limit on the  potential energy of the atom’s (lowest energy) ground state. By incorporating quantum mechanical effects, this model significantly improved the predictions we can make about conducting electron behaviour. 

 

Band Theory of Metals

 

The Drude model, as good as it is, is still missing a component we need to take into account with conduction in metals.  We need to consider the effects of how molecular bonding relates to the conductivity of the metal. We’ve been modelling the conduction electrons as particles that have limited occupied energy states but otherwise act like strangers to each other. 

 

Molecular orbitals in metals are much larger than atomic orbitals, which are confined within the atom. In fact, molecular orbitals extend and overlap each other throughout the whole material. Bearing this in mind, the word “orbital” here can seem misleading. There is no orbiting implied here. Better said, the set of molecular orbitals that are created in the metal are a set of interactions that are generated by the valence orbitals of the interacting atoms. This molecular interaction, rather than the electrons themselves, extends throughout the material. 

 

Metallic bonding means that a very high number of atomic valence orbitals interact simultaneously within a metal. This is a critical key to explaining why metals can be such good electrical conductors. Within a free atom, an atom that is not chemically bonded to any other atom, the energy that an electron can possess must fall into one of several possible discrete energy levels. Within a conductive metal, on the other hand, due to the overlapping of a huge number of molecular orbitals, an enormous number of new possible energy states opens up. The available energies of the valence electrons are no longer confined to discrete energies but instead now to a band of available energies that can vary in width. This molecular orbital approach is called band theory. It’s a very useful way to visualize how conduction works, and to understand why some materials are better conductors than others. 

 

Why are the valence energies spread out into a band? First, we must distinguish between an orbital and a shell. Inside any atom, valence electrons are confined to the outermost (highest energy or bonding) valence energy shell. An electron shell is all about the energy. An atomic orbital is about where an electron is most likely to be found. To distinguish these two concepts, we can imagine that we are adding electrons to an atom. First we fill up a sphere-shaped 1s-orbital, which can hold two (opposite-spin) electrons. Then we fill up the next highest energy 2s spherical orbital with two electrons Then we start to fill three 2p orbitals (three dumbbells shapes in three-dimensions) with 2 electrons each, six electrons total, and so on. Every atom of every element possesses the same number of potential orbitals and shells, but they differ in the number of electrons inhabiting them and the orbitals themselves can differ in size. The 1s electrons belong to the lowest energy K shell. The 2s and 2p electrons belong to the next higher energy M shell. The M energy shell, which is just a circle in a Bohr diagram, can now described in three-dimensional detail, as a double-lobbed and spherical orbital. The three 2p orbitals (which extend along the x, y and z axis in three-dimensional space) and 2s orbital hold a total of eight electrons in the M energy shell. A simplified diagram is shown below left.


The same energy level can contain multiple orbitals. This is Hund’s rule

 

Let’s look at the electron configuration of copper as an example. A copper atom, with 29 electrons in total, has an unexpected electron configuration that hints at why it is so conductive. If we simply fill up orbitals in order (1s, 2s, 2p, 3s, 3p, 4s, 3d; a 4-lobed d orbital can contain up to 10 electrons) we will have 1s22s22p63s23p64s23d9. This is actually wrong, based on experimental evidence. For copper and other transition metals, we must write the 3d orbital before the 4s orbital even though 3d is considered to be higher energy than 4s. This is an oddity of the transition elements. In these elements, the 4s orbital behaves like an outermost highest energy orbital. This short 3-minute youtube video explains why:



Following the transition element rule, we should have 1s22s22p63s23p63d94s2, thinking (correctly) that 4s should still fill up before 3d starts to fill. This is still wrong. According to experimental evidence the correct configuration is 1s22s22p63s23p63d104s1. In these metals, the 3d orbital is just slightly larger than the 4s orbital, but it is still higher energy than 4s the same as it is in other atoms. However, it needs only one more electron to be filled (with 10 electrons total). This filled d orbital state is a more stable lower energy configuration, so it pulls up an electron from the slightly lower energy 4s orbital to achieve it. That electron, moving up from 4s to 3d, increases in energy but the increase is very small. That small cost pays off by significantly lowering the atom’s total potential energy. 

 

At first glance you might conclude that copper has one valence electron, just one electron available to delocalize and form extensive molecular orbitals that take part in electrical conduction. You cannot read it from the configuration, but the 3d and 4s orbitals are very similar in energy. That’s why the electron swap is possible. This means that in practice, copper has eleven electrons available to contribute to the valence energy shell, and therefore to molecular orbitals and to electrical conduction.

 

If we go back to our copper wire example, these valence electrons will fall into a range of many possible energy states due to the overlapping of a huge number of molecular orbitals. This range is called the valence band. In conductive metals, there are many empty valence vacancies with energies just above the filled orbitals that electrons, can move into and out. This range of energies just above the valence band is called the conduction band. In fact, in metals, these two bands of energies overlap, so that some electrons are always present in the conduction band.

 

Another energy level that is very useful to understand is the Fermi level. This can be defined as the maximum available energy level for an electron in a material that is at absolute zero. Absolute zero is the coldest possible temperature of a material. It is the lowest possible potential energy state. In a conductor at absolute zero, the valence electrons are all packed into lowest available valence orbitals. Thanks to the Pauli exclusion principle, the electrons must retain a range of base level energies. We can call this range a “Fermi sea” of electron energy states. The Fermi level is like the surface of this sea, where no electron has enough energy to rise above the surface, an analogy I lifted from the Hyperphysics link above. I think this analogy can lead to some confusion. We might assume that the Fermi level is simply the top of the valence band, which is incorrect. From Wikipedia, we can understand the Fermi level more deeply as “a hypothetical energy level of an electron, such that at thermodynamic equilibrium this energy level would have a 50% probability of being occupied at any given time.” It’s critical to note that the Fermi level does not correspond to any actual electron energy level. Instead it is an energy state that lies between the valence and conduction band energies, or where they overlap in the case of metals. In metals, there are an equal number of occupied and unoccupied energy states at the Fermi level. There are many electrons and there are many free states, and therefore, conductivity is maximized. In contrast, in a material where all the energy states are occupied at the Fermi level, there is nowhere for electrons to move to, so it cannot conduct electricity. 

 

At room temperature some copper electrons are already in the conduction band. We might guess that the conductivity of a copper wire would simply continue to increase as the temperature increases above room temperature, but the relationship is more complex. As the temperature goes up, as electrons in the copper atoms gain thermal energy, they also gain non-directional kinetic energy and therefore lattice vibrations in the copper grow. This kind of electron motion is random and “jittery.” Electrical conduction depends on the motion of electrons in one direction. As the temperature increases, electrical conduction becomes increasingly hindered by internal collisions between delocalized electrons, and conductivity therefore decreases. If we cool copper down toward absolute zero, we would expect fewer and fewer electrons with enough energy to conduct electricity as they fall below the Fermi level. At the same time, thermal vibrations within the metal lattice calm down so conducting electrons are less likely to be scattering by other electrons. If it were possible to reach absolute zero, we would expect no electrons at all above the Fermi level and there would be no conduction possible. It seems to me we could never apply an electric potential to test for current without adding the energy of an electric field that would excite some electrons into conduction. 

 

Electric Insulators

 

Every element and material has a valence band and a conduction band. In terms of electrical conductivity, materials fall into one of three categories: conductor, insulator and semiconductor. Even the best insulator is not a perfect insulator. It too has a conduction band that, under the right conditions, can be populated by a few mobile electrons. There is a significant difference between insulators, semiconductors and conductors in terms of how far apart the energies of the valence and conduction bands are. In the diagram below the energy is increasing from bottom to top. 



We can describe the valence electrons in every material as having two possible ranges of energy levels, in other words, two energy bands: a valence band and a conduction band. Between these bands is a gap, a range of energies that no valence electrons can occupy. Its width varies greatly between materials. Each element or material therefore has a unique band structure. In all insulators, there is a large range of energy above the valence band where no electron energy orbitals are available. They have large band gaps. In an insulator, the organization of the atoms does not allow for free electrons. All of the electrons are tightly bound to their nuclei (there is no shielding effect) and they form tight localized bonds between atoms. A great deal of energy must applied to an insulator before its valence electrons have enough energy to jump the band gap into the conduction band. Once enough energy is applied, however, even an insulator will start to conduct electricity. At ground state, valence electrons are tightly bound to the atoms in the insulator material. In a sufficiently high energy environment, they become excited and delocalized into conduction electrons. 

 

I am being careful to write “energy” rather than “temperature” here because a strong electric field is much better than heat for tearing valence electrons away from atoms and turning them into conduction electrons. Heat is a randomly directed influence that generates random jittery electron movements in the material. 

 

To see what happens when an insulator becomes a conductor, consider a resistor hooked up to a circuit. Let’s apply a voltage much higher than it is rated for. The electric field created by the applied voltage will eventually overcome the material’s resistance. We call this resisting property its dielectric strength. When its dielectric strength is overcome, the material breaks down into a conductor. In a solid, the breakdown is physical and permanent. Our little resistor is burned and ruined. We can also observe this breakdown as an electrostatic discharge, a familiar example being lightning. A sudden giant spark is emitted as air, normally a strong insulator, breaks down through ionization. Electrons stripped from the gas atoms become conduction electrons when a sufficient electric potential difference between one area and another builds up during a storm. 

 

Semiconductors

 

Semiconductors are widely used in electronic circuits. A semiconductor is any material that conducts current but only partially. Its conductivity falls between an insulator and a conductor. Most are made of crystals, so they have lattice structures similar to metals. In semiconductors, the band gap is smaller than in insulators, so a smaller amount of energy (a small amount of heat in many cases) is required to bridge the band gap into the conduction band. Some semiconductors can be chemically altered to further enhance their conductivity, through a process called doping. In all materials, the Fermi energy level is somewhere in the middle of the band gap and if there is no band gap, as in metals, it is near the top of the valence band.



In semiconductors, doping can shift the Fermi level either much closer to the conduction band (N type) or much closer to the valence band (P-type). Doping is done by adding a few foreign atoms into the lattice structure of the material. These impurities add extra available energy levels. In N-type semiconductors (shown below left), energy levels are added near the top of the band gap (along with additional free valence electrons that contribute to conduction). 


These additional energy levels just below the conduction band, mean that the electrons occupying them can be easily excited into the conduction band. In P-type conductors (shown below right), extra empty energy levels are created as mobile energy “holes.” 


These are holes that electrons would normally occupy in the top of the valence band of the material. Electrons can jump in and out of them. N-semiconductors are much more conductive than P-type semiconductors because there are many more electrons available in the conduction band in the N-type than there are holes in the valence band in the P-type. 

 

Conductors

 

In conductors, there is no energy gap at all. The valence band overlaps the conduction band. At room temperature, some electrons are energetic enough to be conduction electrons. When an electric potential is applied, current is generated in the material. When we want to know how good an electrical conductor is, we can ask how many electrons are available in the conduction band. Copper, as we discovered, has eleven electrons per atom that populate the valence band. These electrons are delocalized in the lattice as molecular orbitals. That’s a large number of electrons available to generate a current when an electric potential is applied. 

 

If all metals have overlapping valence and conduction bands, why are some better conductors than others? The answer can be quite complicated but we can get some idea by comparing copper with iron, which is much less conductive. Copper has a conductivity of 5.96 x 107 S/m (Siemens per metre) at 20°C. Iron’s is 1.00 x 107 S/m at 20°C, about six times lower. Copper displays an almost perfect Fermi surface which means it acts like a hypothetical free electron sphere would act. Its valence electrons are all lined up close to the Fermi level, the line between occupied and unoccupied energy shells. They move easily in any direction. Iron has two main differences. Iron atoms have a strong magnetic moment; they act like tiny magnets. This splits the band structure into two parts, based on the two magnetic spin states. This splitting separates the electrons, reducing the ways they can move. The second difference is the Fermi surface isn’t smooth for iron. A Fermi surface is an abstract geometric representation of all occupied versus unoccupied electron energy states in a metal at absolute zero. The shape is derived from the symmetry and periodicity of the metal lattice. In iron, it is broken up into many disconnected pockets, rather than a smooth electron sphere. Electrons have to jump from pocket to pocket in order to move. Even though iron has free (delocalized) electrons in molecular orbitals just as copper does, these two factors make it much harder for the electrons to move and generate current.

 

To sum up, when wondering about the conductivity of a material, a crucial question to answer is how close the conduction band is to the valence band. In a conductive material the two bands are very close and in metals they overlap. In a semiconductor at room temperature, there is a small gap in energy between the valence band and the conduction band with the Fermi level within that gap. Doping adjusts the Fermi level and may provide additional conducting electrons. In an insulator at room temperature, the energy difference between valence band and conduction band is so large that at room temperature no electrons in the valence band can absorb enough energy to populate the conduction band. If the energy of an insulator is increased (usually be applying very high voltage), some of the conductor’s valence electrons may absorb enough energy to cross the Fermi level and jump the gap into the conduction energy band. The dielectric strength of the material is overcome and the material breaks down.

 

There is a finer difference between conductors and semiconductors, in terms of the density of energy states crossing  the band energy gap. In a semiconductor, the conduction band is above the Fermi level, so as its energy goes up, the conduction band begins to get populated by electrons as they cross the gap, starting from zero to one electron to two and so on. Conductivity gradually increases starting from zero. In a good conductor, the Fermi level is in the conduction band because it overlaps with the valence band at room temperature. As the energy rises, the number of conduction electrons starts with an already populated conduction band and increases from there. This is additional reason why metal conductors conduct current so readily.

 

In addition to their conductivity, band theory also explains many physical properties of metals. 

Metals conduct heat better than other materials because their delocalized electrons can easily transfer thermal energy from one region to another. Generally, the better a metal is at conducting electricity, the better it is at conducting heat. Metals at room temperature feel cool to the touch because the electrons in the metal readily absorb the thermal energy in your warmer fingertips. This energy absorption represents a very small jump in energy to slightly higher available energy levels. For the same reason, metals under a hot sun tend to feel hotter than other objects nearby. The electrons easily transfer thermal energy to relatively cooler objects such as your fingertips. Metals tend to have a small specific heat capacity, which is a measure of how much heat must be added to a material to raise its temperature. In a similar way, metals are lustrous because numerous similar energy levels available to the valence electrons means they can absorb the energy of various wavelengths of visible light (they absorb various colours simultaneously). When electrons inevitably decay back to lower rest state energy levels, they emit light across the same wide range of wavelengths. As light is continuously absorbed and re-emitted from the surface of the metal, we see it as lustrous and shiny.

 

Three Different Velocities 

 

Now that we know how electrical conduction starts, we can focus on what’s happening when the current flows. Here we can further our quantum mechanical understanding of how the electrons in a conductor behave. A metal, as we know, is organized into a three-dimensional lattice of atoms. Each atom has an outer energy shell of delocalized valence electrons. These electrons are somewhat free of the attractive influence of their nuclei and they interact with each other, forming molecular bonds between atoms within the lattice. This “sea of electrons” is free to move and conduct an electrical current through the metal. For example, when a voltage is applied to a circuit, electrons in a copper wire drift from the negatively charged terminal toward the positive charge. The electrons themselves drift very slowly through the metal, on the order of a just few metres per hour. If we looked inside the  copper wire with the circuit switch turned off, we would see electrons continuously making microscopically short trips in random directions, changing direction as they strike other electrons and bounce off them. The net velocity of the these electrons is zero when no voltage is applied. When a voltage is applied, there is a net flow, a drift, of electrons in the opposite direction (negative to positive) to the electrical field (positive to negative) that is superimposed over the random drift motion. 

 

Wikipedia supplies an interesting mathematical example of electron drift velocity, worked out for a 2-mm diameter copper wire. The drift velocity (which is proportional to the current; here 1 ampere) works out to be 23 um/s. That’s micrometres. For a typical household 60 Hz alternating current, the electrons in the wire drift less than 0.2 um per half cycle. This means that the electrons flowing across the contact point in a switch, for example, never even leave the switch. This doesn’t mean that the electrons are individually lumbering along within the wire. Individual electrons in any material at room temperature have a tremendous amount of kinetic energy, called Fermi energy. The Fermi velocity of electrons is about 1570 km/s! The drift, the net movement of electrons in one direction under an electric potential, however, is extremely slow. The speed of electricity (the speed at which an electrical signal or energy travels) through a copper wire is yet again different. The energy that runs the motor in a vacuum travels as an electromagnetic wave from the socket, through the cord, and to the motor. The wave’s propagation speed is close to but not quite the speed of light. This is why the vacuum starts up immediately once you flip the switch. And this is where our simpler billiard ball explanation of current propagation falls down. If electrons were like tiny balls striking one another down a line, like football fans doing the wave in a stand, delays caused by the inertia of each electron as it is put into directed motion would build up quickly and significantly and slow the current to a stop. The electric signal instead travels as propagating synchronized oscillations  of electric and magnetic fields generated by the oscillations of electrons in conducting atoms.

 

The wire guides, rather than physically carries the wave of electromagnetic energy. This wave of energy, in turn, generates an electromagnetic field that propagates through space, responding to the (just preceding) energy flow. The time required for the field to propagate along the metal means that the electric field can lag slightly behind the electromagnetic wave, an effect that grows with the length of the wire, but the lag  is immeasurably small at the scale of a vacuum cleaner cord. The vacuum is switched on and the wave of energy, which we can think of as an electromagnetic field signal, travels at near light speed down the cord. All the mobile electrons in the copper wires immediately get the signal to start oscillating along the electrical circuit. The electric signal, as an electromagnetic (EM) wave, is composed of oscillating electric and magnetic fields. The electromagnetic energy flows as oscillating electric and magnetic fields that are generating just ahead of it. The electrons, acting as moving charge carriers and magnetic diploes, generate the oscillations of the (EM) signal. 

 

The EM wave loses energy along the cord, as energy is transferred into work done by the charge carriers in the wire. Energy is always lost during the transmission of electricity. Using Ohm’s law, we know that voltage, current and resistance are related to each other. Losses square with the current, which means that a small jump in current through a wire leads to a big jump in loss (through increased electrical resistance). This is why, for example, long-distance transmission lines are so high voltage, to minimize current loss due to resistance. A lot of energy is lost as microscopic friction (electrons bumping into electrons) inside the wire, and it is released as heat. A current-carrying copper wire warms up. Because the atoms are bound tightly together and there are many electrons in close proximity to one another, electrons will experience some resistance even in a good conductor such as copper.

 

Bloch Theorem

 

We haven’t explored yet how the three-dimensional lattice arrangement of atoms in a metal impacts its conductivity. In 1928, Felix Bloch formulated his Bloch theorem to deal with these effects. 

To use this tool to understand how current propagates within a metal lattice, we need to deepen our understanding of quantum mechanics once again. Electrons, like any matter particles, have a dual wave and particle nature. Resistivity can be explained by electron-electron collisions, a particle-like phenomenon. Conductivity, however, is better understood as the transmission of energy waves between electrons. Bloch theorem incorporates both natures simultaneously by treating electrons mathematically as wavefunctions. Put more precisely, this theorem is set up in quantum mechanics as special limitations put on Schrodinger’s famous equation. This equation describes the wave function of a quantum mechanical system. A wave function, in turn, is a mathematical description of an isolated quantum state. The wavefunction is useful because it gives us measurable information about that quantum state. I think it’s fair to say that the wavefunction, a complex function of space and time, pins down the un-pin-able physical properties of the electron. These physical properties are position, momentum, energy and angular momentum, the four quantum numbers mentioned previously. The “pin” is mathematical. If you know these four numbers, you know everything there is to know about any particular individual electron. In a quantum mechanical system the set of possible values for these physical properties cannot be specific values as they would be in a classical system. Here, they are expressed instead as eigenvalues, or as ranges of possible values.

 

You will notice I said “isolated state” when we really want to describe what is happening in a complex system containing many electrons. None of these electrons are acting as perfectly isolated particles, but again we start by simplifying matters.  We make the assumption that the electrons are acting like free particles and we ignore their interactions with each other. Copper, as mentioned earlier, comes pretty close to this hypothetical ideal state. An electron acting like a free electron in a metal such as copper can be treated mathematically as a wavefunction. In this case, we put limits on the wavefunction, so that our solutions to Schrodinger’s wave equation take into account the periodic nature of a three-dimensional lattice. Put more precisely, these solutions form a plane wave that is modulated by a periodic function. A plane wave is set up as a set of two-dimensional wavefronts traveling in forward through three-dimensional space in a perpendicular direction. A periodic function is a mathematical way to describe a system that repeats its values at regular intervals, like the regular intervals of a crystalline arrangement in a metal. These functions are called Bloch functions and we can use them to describe the special wavefunctions, the special quantum states, of electrons in a metallic crystalline solid. The Bloch function itself is not periodic but its probability wavefunction includes the periodicity of the lattice, which tells us that the probability of finding an electron is the same at any equivalent position in the lattice.

 

This mathematical description of electrons underlies the band theory of electron conduction that we discussed. Band gaps, energy states where electrons are forbidden, can now be described mathematically as values for energy, E, where there are no eigenfunctions in the Schrodinger equation. An eigenfunction, also known as a Bloch function here, is the mathematical description of any particular electron, as a wavefunction within a crystalline (metal) solid. We can predict the energy range where electrons are forbidden (the band gap) by working out where there are no eigenfunctions (no wavefunction solutions). The actual calculations are exceedingly complex and I don’t pretend to understand them. I do think, though, that it’s quite astounding that we have a way to precisely predict and describe a very complex behaviour that is critical to understanding how electric conduction works.

 

Emergent Behaviours and Properties

 

It might strike us as surprising that the Bloch theorem actually works so well, considering that we are ignoring many complications that could arise from electron-electron interactions. Complex interactions between subatomic particles in a metal can lead to the emergence of unexpected properties at the larger macro- or everyday scale.

 

The electron, as a elementary particle, has a charge and a mass. But, because it is a quantum particle, its charge and mass are qualities that can act in independent and unexpected ways. This is where our notion of the electron as a tiny physical ball of charge is really challenged. The Bloch theorem works because the electron’s charge moves within a periodic electrical potential as if it were a free electron in a vacuum. It’s mass, in contrast, becomes an effective mass. The mass of an electron at rest is always about 9 x 10-31 kg. Inside a metal, however, an electron can seem to have a different mass based on how it reacts to various forces acted upon it. Effective mass can range considerably, from zero to around 1000 times the electron rest mass, and this can have a big influence on how the metal behaves. For example, “heavy fermion“ metals, those with an effective electron mass in the 1000 range, can exhibit superconducting properties, such as zero electrical resistance below a low critical temperature. Effective mass can even be negative as in the case of P-semiconductor electron holes mentioned earlier. An electron hole is a lack of an electron mass where one should exist in the lattice, and leaving a local net positive charge. Each hole acts like a particle and is referred to a quasiparticle, in this case a positively charged one. It is phenomenon that arises from a complex system. This behaviour plays an important role in current conduction through semiconductors. Excited electrons leave behind holes in their old ground state energy level, and they can move just like electrons do, resulting in an electric current moving in the opposite direction.

 

Conclusion

 

Electric current is perhaps the most basic concept at the heart of the science of electricity. It’s a rare day that goes by when we don’t make use of at least one light bulb or our mobile phone. It’s so familiar to us that we might not give it much thought. Yet, electrical current is mysterious. It cannot be directly seen, heard or felt. It’s not easy to gain an idea of what it actually is. In order to do this we had to dive deeper and deeper into theory, from classical to quantum mechanical, while we updated our mental snapshot of the electron along the way, from the Rutherford haze to the tiny charged billiard ball to a modern quantum cloud of mathematical probabilities.

 

 

 

Sunday, September 29, 2019

Rocket Science

Interstellar Rocket Propulsion: An Exploration of the Options and Challenges

There is considerable scientific interest in designing an interstellar spacecraft, and the technologies required to build one are theoretically feasible. If we ever venture to explore a distant stellar system, one of these technologies might get us there.

Proxima Centauri b: A Destination Case Study

It might go without saying, that we want to look for life elsewhere in the universe. We all look to the heavens and wonder: are we alone? To answer this question, we are going to need to develop new technologies. We now know that the majority of stars in our universe have a complement of orbiting planets. There are several known possibly Earth-like planets we could explore. The problem is that they are so far away. Our nearest star, Proxima Centauri, is 4.25 light years away, shown as a yellow dot below. This might be the first exoplanet system we actually get to.

NASA/Penn State University; Wikipedia
The year when the distance to each star or system was calculated is listed after each name.  Proxima Centauri, as seen through the eye of Hubble Telescope, shines bright white, see below left. Although it is our nearest stellar neighbor, it is invisible to the naked eye. Its luminosity is very low because it is a small star, just 1/8 the mass of our Sun.

ESA/Hubble;Wikipedia
Proxima Centauri is slightly closer to us than its two partner stars, a binary pair called Alpha Centauri AB, or simply Alpha Centauri as shown in the diagram above.

Proxima Centauri is a small low-mass main sequence red dwarf star. It is seen as fairly white through the Hubble lens because even though it is technically a “red” dwarf, its surface temperature is 3040 K (as hot as a light bulb burning warm white).

Proxima Centauri has at least one confirmed planet orbiting it, called Proxima Centauri b  (or Proxima b for short). Below is an artist’s impression of the planet orbiting its star. Based on one of several possible models of formation, it could be an arid, but not water-free, super-Earth (1.3 times Earth’s mass).

Artist’s impression; ESO (European Southern Observatory)/M. Kommesser; Wikipedia
Whether this planet is habitable or not is an open question. Even if it is devoid of life, it is an intriguing planet. Discovered in 2016 by HARPS Spectrometer, this planet orbits in what is called the habitable zone. The habitable zone is the distance from a star where a planet’s surface temperature could support liquid water. The radiant heat from the star is just right, not too cold (a frozen world) and not too hot (a world where any surface water would turn to steam and leave the atmosphere). Such a world could support the kind of water-based biochemistry used by all living creatures on Earth. There could be life based on non-water chemistry, using a solvent other than water, but this is a marker scientists tend to agree on because we have proof that water-based biochemistry works.  Even though it is within the habitable zone, this planet is so close to its dim star, Proxima Centauri, that it is likely to be tidally locked, meaning that the same face of the planet always faces its star. It doesn’t rotate, so there is no night and no day. The star-facing side would likely be too hot to support life as we know it and the night side would be too cold. The strip of area in between, which could be warm enough for liquid surface water, might support life.

A significant challenge faced by any life on Proxima Centauri b is the radiation it would endure from its star. Red dwarfs are more violent and less stable than larger stars such as ours. They are covered in starspots that can dim their output light up to 40% for months on end. At other times the planet’s atmosphere, assuming it has one, would face erosion from frequent and powerful radiation flares. Does it have a magnetosphere that would protect an atmosphere?

Shielded from the direct impact of its star’s unpredictable radiation, the “in between” zone might stand a chance to support life. It’s a bit of a slim shot at alien life, but fortunately it isn’t our only shot. There are numerous other exoplanets, which might be more suitable for life to evolve, and they might all be within our eventual physical reach. The first question, however, is how do we find them?

The Search For Exoplanet Candidates

It is startlingly amazing what we can glean through indirect observation. Proxima Centauri b, our case study planet, was discovered by ESO’s HARPS Spectrometer, located at La Silla observatory in the Chilean desert. Like other planets, it is illuminated by reflected light only and it is far too faint to be been by the naked eye. High-precision HARPS detects minute Doppler shifts in the radiation from its parent star, Proxima Centauri. These tiny shifts in the star’s radial velocity are in reality the tiny regular wobbles it makes as its planet orbits around it. Each time it orbits, the planet shifts the center of gravity of this simple two-body system back and forth. Even though this detection method detects only a fraction of possible planets - those planets whose orbital plane lines up with Earth - in just two decades, more than 4000 exoplanets have been confirmed. We are in the golden age of exoplanet discovery.

HARPS isn’t alone. Sophisticated ground-based telescopes of various kinds (listed here in Wikipedia), have detected exoplanets. New ones are detected as our ability to look improves. HARPS, so far, has discovered over 130 exoplanets. Many additional observatories are located in orbit around Earth, such as the Kepler Space Telescope, for example, with 2421 exoplanets detected and counting. And now TESS has detected 9 exoplanets during its first operational year. The tantalizing question is what kinds of alien life could some of these exoplanets harbour?

These sophisticated observations can answer questions about planetary mass, orbit, the atmosphere, and even possible surface composition regarding planets that are far too distant, too small and too dim for us to observe directly in the night sky. This information is tantalizing. Yet there is no substitute for being there. It is a human need to see with ones own eyes, feel with ones own fingers. Perhaps someday a manned interstellar mission to a planet harbouring alien life will happen. It’s perhaps more likely that we might never surmount the daunting technical obstacles faced with such a long manned spaceflight, one that out of necessity would take many years, perhaps even longer than a single human lifetime to accomplish. Wikipedia  lists our manned travel options, and obstacles, in this regard.

There is hope. A future robotic mission equipped with suites of sensory instruments that could see, hear, physically feel, chemically taste and smell for us might be feasible in the near future. Besides crafting a vehicle that can house these instruments AND withstand years of interstellar travel, the principle challenge facing us is the time it will take to get to any distant solar system. Even traveling at near light speed to our closest exoplanet system, Proxima Centauri, would take at least four years. Assuming we have the propulsion technology to make this happen, imagine how difficult it would be to get funding for a certainly expensive exploratory mission that would take over eight years to receive any data back IF nothing goes wrong.

Spacecraft Propulsion: A Very Brief History

Perhaps the most daunting question is how to propel the spacecraft, and that is really what I’d like to explore in this article. To put this problem in perspective, consider the chemical propulsion technology we currently use. In 2013, Voyager 1 (shown below) entered interstellar space at a velocity of 62,000 km/h (about 17 km/s).

Artist’s rendition; NASA/JPL- Caltech
That’s fast, about 52 times the speed of sound. Yet, at this rate it would take about 70,000 years to reach Alpha Centauri from Earth, just not feasible. As you can imagine, chemical propulsion technology has improved since 1977, when Voyager was launched. Accelerated by Jupiter’s gravity, the Juno probe, launched in 2011, (see image below) briefly clocked in at 266,000 km/h as it locked into orbit around Jupiter. A 35-minute insertion burn decelerated it to a manageable velocity of about 2000 km/h.

Artist’s impression; NASA
The Parker Solar Probe, launched last summer (2018), (shown below) should reach a velocity about three times higher, 692,000 km/h, as it slingshots around the Sun in 2025, setting a new velocity record that is  almost ten times the velocity of Voyager 1.

Artist’s rendition; NASA/John Hopkins APL/Steve Gribben
The fundamental physics behind these speed records is the same for all these craft. They make use of a phenomenon called gravity assist, which can really got things moving.

Current Technique: Gravity Assist

Let’s start at the beginning. In order to successfully launch, a rocket must overcome Earth’s gravitational well. A gravity well? We can think of a mass “denting” the space around it, as depicted below right, with a mass such as a planet, at the bottom of the well. The gravity well is a conceptual model arising from Einstein’s theory of general relativity, which predicts that gravity arises from the curvature of space-time. Technically it is four-dimensional space-time, not three-dimensional physical space, that curves .

AllenMcC;Wikpedia
In rocket science, it helps to conceptualize gravity as a geometry (Einstein) rather than an attractive force (Newton). Both Voyager 1 onboard a Titan 3E rocket and the Parker Probe onboard the current Delta IV Heavy rocket accelerated to velocities of approximately 130,000 km/h (36 km/s) as they left Earth’s atmosphere. Consuming chemical rocket fuel, the rockets achieved enough velocity to climb out of Earth’s gravity well.

An image of Voyager 2 aboard Titan III-Centaur. Both Voyager probes were launched by Titan III rockets, NASA/MSFC
Delta IV Heavy Rocket launch in 2013; U.S. Air Force/Joe Davila. This rocket is currently in use with several launches expected in 2020-2023.
A future interstellar rocket will have to climb out of Earth’s gravity well and then climb out of the massive Sun’s gravity well. Such climbs are steep; chemical fuel is rapidly used up. That will be a challenge for a future interstellar probe, as it was for Voyager 1, the only craft so far to climb out of our solar system gravity well. Earth makes an indent in space that a rocket must climb out of. The Sun makes a far deeper indent, which any spacecraft must overcome in order to leave our solar system. Velocity is key. Consider a black hole, by comparison. Its mass is so great and concentrated in such small volume that its gravitational well is inescapable, even to light. Photons of light, traveling at 300,000 km PER SECOND, are too slow to escape it.

One key problem with chemical rocket propulsion is fuel weight. Voyager 1 was launched on the expendable 4-stage rocket, Titan 3E, which used solid and liquid chemical fuels in its engines, providing enough thrust to lift the fairly heavy (823 kg) Voyager probe. At lift-off, the fully fueled rocket weighed about 633,000 kg, almost all of which was fuel. It needed most of it just to escape Earth’s gravity, with its fuel weight working against it. Even starting out with a respectable launch velocity of 130,000 km/h, Voyager 1 would have quickly petered out to become a dead weight in space were it not for the fortuitous and anticipated line-up of massive planets at that time.

Gravity Assist

Once boosted to a certain velocity, any body will maintain that velocity in a vacuum, and space is an almost perfect vacuum. Nonetheless, gravity acts on it. In Voyager 1’s case, gravity was working against it and for it. It had to climb up the Sun’s gravity well but it had several gravity assists to help it along. It reached its current velocity of 62,000 km/h, by making use of how the planets lined up in 1977. It passed very near Jupiter and Saturn, and then Uranus and Neptune, using each of their gravity wells to slingshot around them. It went with the planet’s spin to gain momentum at the expense of the planet’s momentum in order to boost its velocity. The decrease in spin of a planet would be immeasurably miniscule because its momentum is so much greater than a relatively infinitesimal probe. Voyager 1 was able to enter interstellar space by escaping the Sun’s gravitational well, which we can think of as that same inverted cone with a gradually widening, flattening base. An object traveling near the edge of that cone base, far from the Sun, experiences a much weaker gravitational pull. The solar gravity well isn’t so steep near the edge. Very close to the Sun, where the well is steepest, an object needs to travel at 525 km/s, or almost 2 million km/h, (away from it) to escape its gravity. Escape velocity then drops exponentially with distance. To help visualize how this works, see the graph below. Using data extrapolated from Voyager 2’s similar statistics, Voyager 1’s launch speed was about 36 km/s (about 130,000 km/h) (red line), very close to solar system escape velocity (blue line) at Earth’s orbital distance from the Sun. Its initial thrust had to work against the Earth’s gravity and then against the Sun’s enormous gravity, rapidly decelerating it to about 10 km/s by the time it approached Jupiter. If it kept experiencing that rate of deceleration it would never escape the Sun’s pull. Jupiter’s gravity boosted it back up to about 28km/s, well above solar escape velocity at that distant orbit. Shortly afterward, and with the help of subsequent planetary boosts, Voyager 1 kept getting boosted to keep it well above solar escape velocity. And that’s how Voyager was able to leave our solar system behind for the adventure of interstellar space beyond.
Cmglee;Wikipedia

Velocity and Travel Time

We have solved the issue of leaving our solar system but the issue of achieving a reasonable journey time to an extremely distant solar system remains. There is a limit to what kind of velocity a craft can achieve using chemical fuel for an initial thrust followed by gravity assists. Even if a spacecraft uses the Sun as its gravity assist, like the Parker probe will, reaching a velocity of 692,000 km/h, it would still take about 7000 years to reach Proxima Centauri, 4.25 light years away. The only way to shorten that travel time is to go faster, much faster.

Thrust/Weight Ratio

There is a built-in limit to the usefulness of any chemical rocket fuel, and that is its thrust-to-weight (T/W) ratio. This is a very useful dimensionless ratio calculated by dividing engine thrust (in units of force called newtons (N) by the weight of the fully fueled rocket at sea level (again in newtons), This is an example of weight treated as a force. It is generated by Earth’s gravitational field. Thrust is directly proportional to the acceleration of the rocket (force = mass x acceleration).

A high T/W ratio means high acceleration. Three force vectors work on a rocket as it lifts off: thrust (upward), weight (downward) and drag (downward). Drag is a mechanical force. It is the friction created as a solid object (the rocket) moves through a fluid (air in this case). Drag diminishes as the rocket exits Earth’s atmosphere. Thrust is the force generated by the rocket’s propulsion system.

If the ratio is greater than 1 and the drag is minimal, the rocket can lift off and accelerate upward. The T/W ratio is used as a static measure for a rocket at sea level. In operation, this ratio changes constantly as the rocket uses up fuel, reducing weight, and as drag decreases up through the atmosphere. It also changes as engine efficiency changes. To get an idea of the values involved, we can compare an Airbus A380 fully fueled and loaded at takeoff with a T/W ration of 0.227. It can take off and get airborne but with a T/W less than 1 but it can’t accelerate straight upward. To get a little more nuanced, its weight overcomes its thrust so airspeed always decays during a climb, whether it’s vertical or not. Fighter jets, in contrast usually achieve a T/W of close to 1, which affords them much greater maneuverability and speed. You might find it interesting to compare the W/T ratios of various aircraft here. The Space Shuttle fully fueled at takeoff had a T/W of 1.5 (using three main engines and two solid rocket boosters), plenty of vertical thrust to escape Earth’s gravity and reach orbital speed. To achieve this, the shuttle would launch vertically for a few kilometres while performing a gravity turn. That means the shuttle allows gravity to gradually bend its trajectory from straight up into horizontal to the ground as it accelerates up to the speed and altitude where it can maintain its orbit around Earth once its rockets shut off.

You won’t find a simple T/W ratio for modern rockets at takeoff because this ratio isn’t all that useful. They are all over 1. More important are the individual thrusts of the engines and the overall weight of the fuel, which is constantly changing, and the payload weight. You might guess that a higher T/W ratio is always better for a rocket launch but it isn’t. Instead there is an optimal acceleration rate. A higher T/W ratio means you build up great velocity in the lower atmosphere where the air is thickest. This means the rocket is subject to a lot of drag and therefore heating. These two factors increase wear and tear and engine inefficiency. A lower T/W ratio, on the other hand, means it will take longer to reach orbital speed so more fuel is used up fighting against gravitational forces. There is a sweet spot. Most rockets are designed for a T/W ratio of slightly higher than 1 with a maximum acceleration of about 4 g for a few seconds at the end of the first stage rocket burn, when most of the rocket fuel is burned and the weight is lowest compared to the upward thrust. 4 g means that our bodies feel 4 times heavier than they normally do (at 1 g). That is about the maximum acceleration that is “comfortable” for astronauts. To achieve the very high velocity required for a feasible interstellar mission, such a rocket will need to continue to accelerate after it has achieved orbital speed, and it will also need more boost than a gravity assist from our Sun. Considering that the nearest system is over 4 light years away, it will need to be boosted to a significant fraction of the speed of light. It will also need enough reserve thrust to maneuver into orbit around its target planet, but I am getting ahead of myself.

Rocket Equation

Ultimately no chemical rocket fuel can provide the kind of long-term thrust required to reach a significant fraction of the speed of light, often loosely called relativistic velocity. Why? To answer this question we can start by exploring the ideal rocket equation. This equation describes how a rocket accelerates itself using thrust created by shooting out part of its mass at high velocity in the opposite direction. That’s the Sun-bright blast from the rocket’s engines as it lifts off its pad. The rocket moves according to the common-sense principles of the conservation of momentum, Newton’s second law. But there is a bit more to it than what at first seems quite simple.

The rocket equation steps in where Newton’s second law cannot. Put formally, Newton’s second law states that the acceleration of an object as produced by a net force is directly proportional to the magnitude of the net force, in the same direction as the net force, and inversely proportional to the mass of the object. Or, simply put, it is force = mass x acceleration. In a rocket system, mass is always changing, so Russian scientist Konstantin Tsiolkovsky, in 1903, derived a formula using straightforward calculus, that describes its unique motion. It’s a lovely example of why calculus is good, and it’s a good equation on which to practice your calculus skills. Below is a screenshot of the equation taken from the Wikipedia page to show you what it looks like:


In order to describe the motion of our relativistic interstellar rocket system, we need not only the rocket equation but also to juice it up further to describe a system that is acquiring relativistic mass as it approaches light speed. This kind of transformation is described by Einstein’s special relativity theory. This more complex rocket equation is derived for a rocket accelerating close to light speed. The mass, still changing, is also acquiring significant momentum as it approaches the speed of light, becoming relativistic mass according to the law of special relativity. For those interested in the mathematics of this derivation, this paper offers a side-by-side comparison between the classic rocket equation and its relativistic counterpart.

The rocket equation describes the motion of a rocket system. If you’d like to see how it is derived (in its non-relativistic form), this NASA site goes through it well. For our needs, it’s good enough to simply know the rocket equation relates a vehicle’s maximum possible change in velocity (delta-v) to the engine exhaust velocity, as the vehicle’s mass changes while fuel is consumed. It is from this equation that we get such terms as delta-v and impulse, which we will get into. Using this equation we can explore how the momentum of a rocket system changes as its fuel is consumed. We can compare the efficiencies of various rocket fuels in a rocket system. We can calculate how much fuel is required to change the velocity of a system over time, and what the maximum change in velocity is. Ultimately this equation can tell us which kinds of fuel will get us the acceleration and maximum velocity we need to achieve interstellar travel.

Impulse

If you are sci-fi fan like me, you’ve probably heard or read, “Go to full impulse!” In physics, impulse is another outgrowth of Newton’s second law. It is the integral of a force over a time interval. In other words, we use calculus to describe a force changing over time. As a result, impulse creates a change in momentum of a system. To see how these two concepts are related to each other, check out this physicsclassroom.com link.

Impulse can also be used to figure out the efficiency of the fuel used in the rocket system if the mass of the fuel (also called propellant) is taken into account. It is total impulse per unit of fuel used up. More specifically, we can measure how many seconds a fuel can accelerate its own initial mass at 1 g. The resulting value, in seconds, is called the specific impulse. If we turn this relationship around slightly, we also determine how much of a particular fuel is needed for a given delta-v. There are several different ways to calculate specific impulse. Another method brings us back to the rocket equation. It can be calculated as the generated thrust of a system per unit mass fuel flow rate. This gives us the effective fuel exhaust velocity relative to the rocket. Effective exhaust velocity is proportional to the specific impulse, and both can be used to measure the efficiency of a rocket fuel. To get a feel for how these two values relate to each other, consider a fighter jet’s rocket-like engine running within Earth’s atmosphere. We can calculate the jet’s effective exhaust velocity from the rocket equation. Its actual exhaust velocity, however, will be reduced by atmospheric pressure. This pressure works against the exhaust, and therefore the momentum, and the delta-v. In return, the specific impulse of the engine’s fuel, its efficiency in other words, is also reduced under these real conditions. In jet engines, there is a big difference between effective and actual exhaust velocity. In rockets operating in the vacuum of space, there is none. If you go back to the Delta IV Heavy rocket Wikipedia page, on the right you can compare the specific impulses of its engine stages at sea level (in air) and in vacuum.

Specific impulse can be derived from yet another rocket science term called thrust. Thrust is a mechanical force. It is equal and opposite to the force generated by the mass of gas accelerated and shot out as the rocket’s fuel exhaust. This force, an example of Newton’s third law of motion, is in the forward direction of the rocket. Newton’s third law states that for every action (or force), there is an equal and opposite reaction (force). Specific impulse can be calculated by dividing the thrust by the rate at which the fuel is used up.

We can compare the efficiencies of various rocket fuels by comparing their specific impulse values. Most rocket fuels range from about 175 s to 300 s. “Seconds” might seem like a weird unit to measure how good a fuel is, but it is quite practical, and it makes sense now that you know how it is derived. In practice, thrust and efficiency (which usually increase with the price) tend to trade off one another. Solid rocket fuels tend to have high thrust and relatively low cost. They are usually used in the first stage engines on modern rockets because you need a lot of force to achieve lift off and its okay, even advantageous, to burn off a lot of heavy fuel right away. The more expensive higher impulse fuels (many of these being liquid hydrogen-based) are saved for higher stages. These fuels have a higher specific impulse (more seconds) than solid rocket fuels. If you are curious to compare efficiencies and other features among an extensive list of rocket fuels, see NASA’s list here,

No chemical rocket fuel is efficient enough to accelerate to a relativistic velocity. It is tempting to think that, because there is no air resistance in space and, away from stars and planets, there is no significant gravity to work against, it would be easy to eventually accelerate a spacecraft to near light speed. The problem is that in space, there is also nothing to push against to make the craft go faster except the fuel it brings along with it. The mass of the fuel resists being accelerated. It has its own inertia to overcome. If the fuel is too inefficient, you can’t pack in enough fuel to accelerate the spacecraft long enough to reach relativistic speed and overcome the inertia of the fuel mass itself. (We’ve now covered all three of Newton’s laws.)

In order to achieve a reasonable interstellar velocity, we need a fuel with a much higher efficiency, or specific impulse, than any chemical combustion reaction can provide. I’d like to caution us here because I mentioned exhaust velocity earlier. It would be easy to assume that the final velocity of a rocket is limited by its exhaust velocity. This calls for a close examination of Newton’s third law. The forces must balance out, but not the velocities. The fuel in the engine accelerates from rest to whatever the engine exhaust velocity happens to be. This accelerating mass of fuel creates the equal and opposite force of thrust. The force of thrust will continue to accelerate the rocket as long as the engine keeps working – in a vacuum. A jet flying through the atmosphere, on the other hand, will continue to accelerate until air resistance balances out the thrust of the aircraft. As a little test of this, consider a jet with enough thrust to go supersonic. This doesn’t mean its engine exhaust (which unlike a rocket is almost entirely accelerated air) is supersonic.

Fuel Options for Interstellar Flight

Most of these options are based on the likelihood that our first interstellar exploratory mission to another planetary system will be an unmanned probe. In theory, interstellar (unmanned) exploration is now possible, but much needs to be worked technically out before we can send off our first interstellar probe. We need a promising target planet, one that we will be reasonably certain has liquid water and an atmosphere containing oxygen, the bare essentials for life as we know it. New exoplanets are discovered almost every day, and new details about them are being discovered. We don’t know enough about any one of them to know if it could support life. To get there, a daunting number of technical problems need to be solved, such as propulsion and communications, but also protection from the brutal micro-impacts and radiation of interstellar space. Today’s new projects are at the brain-storming-for-ideas stage. Breakthroughs scientists are making right now are built on from the successes of previous projects, and it’s a very exciting time to follow this progress.

How It All Started: Orion

A silver lining emerged from the aftermath of WWII as some scientists began to wonder if they could harness the awesome power of a nuclear explosion to explore the depths of space. Since then, a number of additional novel propulsion concepts for such as mission have been proposed. To get an idea of them all, see this list in the Wikipedia article linked in the paragraph above. One of the most intriguing possibilities uses either a fission or fusion chain reaction. This concept is a powerful one. It offers a perfect scenario by combining two essential qualities into one: high thrust AND high specific impulse. It can be very efficient and it can also perform the high delta-v maneuvers required to insert itself into orbit or to land onto an exoplanet.

This dream took root in the late 1950’s, at the peak of the atomic age epitomized by Disney’s (original) Tomorrowland and later on by The Jetsons, well before news of the Three Mile Island accident, the Chernobyl meltdown and the Fukushima disaster soured public opinion in the decades to come. The dream was Project Orion, developed by theoretical physicist, Freeman Dyson.

An artist’s conception of an interplanetary Project Orion spacecraft; NASA
The fascinating story of Project Orion is fleshed out in the BBC4 2002 1-hour documentary, “The Secret History of Project Orion: To Mars by A-Bomb.” So well told and painting such a vivid picture of that time, this video is one of my favourites:


In 2002, George Dyson gave an excellent 8-minute TED talk about his late father’s work. With dry wit, he displayed an original chart of wildly oscillating g-forces the passengers would have experienced, exposures of 700 rads/explosion at the crew station, and permanent eye burn for the spectators.


Originally cloaked in government top-secrecy, the Orion concept nonetheless eventually emerged to capture the imagination of the world. Stanley Kubrick’s 1968 movie, “2001: A Space Odyssey,” utilized a fictitious nuclear-powered high-performance Orion III to shuttle passengers from Earth to (also fictitious) Space Station V.

The actual Orion spacecraft would have been propelled by a series of nuclear fission explosions (nuclear pulse propulsion). Of course, one chief challenge was how not to blow the whole ship apart (how to harness and dampen almost instantaneous surges in momentum). This challenge would be revisited and refined decades later. The Partial Test Ban Treaty of 1963 as well as concerns about the radioactive fallout from its propulsion system, really a series of full-scale nuclear bombs, led to its fall from favour in the 1970’s.

But the idea didn’t fall away completely. This concept has been refined over the decades since into smaller better-controlled fission propulsion systems that use tiny pellets of fuel to generate chains of micro-explosions. Advances in techniques to confine and direct the explosions would further increase the efficiency of the system. The size of the original Orion interplanetary prototype vehicle was a huge full-scale manned 90 metre high 4000 tonne rocket. Requiring 800 0.14-kilotonne nuclear bombs detonating in rapid succession just to lift off, the original idea was to use this behemoth to get to Mars. While Orion scientists were focused on Mars, a final nail in the Orion coffin came from NASA’s decision to focus instead on (chemical propulsion) missions to the Moon.

From the grand-plan-anything-goes mindset of the 50’s, a once crazy-sounding idea is gathering new public interest. Building from the Orion days, a possible future micro-fission, or even micro-fusion, version of this system will be far smaller, lighter, and less dangerous, thanks to advances in technology. A nuclear pulse engine could consist of a series of anti-hydrogen pellets as small as 0.1 mm in diameter suspended within 100-gram hollow shells of nuclear fuel. A series of explosive lenses (shaped focusing charges) could break a Penning trap used to suspend antimatter in a vacuum, triggering an annihilation explosion with enough energy to trigger nuclear fission or even fusion, which in the latter case would create a shower of fast neutrons and very little radioactive fallout. Project AIMStar and Project ICAN, both proposed in the 1990’s, depend on versions of this much more compact antimatter-triggered nuclear pulse system. Meanwhile, Project Daedalus and Project Longshot, projects in the1970’s and 1980’s respectively, refined the concept of a new type of fusion energy called inertial confinement fusion for use in a nuclear pulse propulsion system.

1) Nuclear Photonic Propulsion

This theoretical system introduces an interesting twist to nuclear propulsion. Like other nuclear systems, it offers very high efficiency. The maximum specific impulse of any chemical propellant is only about 400 seconds. A variety of theoretical propulsion systems can achieve specific impulses magnitudes higher, with some also potentially achieving relativistic velocities. To compare, an ion propulsion system, which I will describe, can theoretically achieve a specific impulse of about 10,000 s. With this efficiency such a craft could achieve much faster speeds, but probably not relativistic, velocity. A photon rocket, however, could top this, potentially achieving an impulse of 30,000,000 s. This type of propulsion could propel a spacecraft up to 1/10 the speed of light.

A robotic probe trip to Alpha Centauri, for example, could become feasible, at a little over 40 years one-way with an additional 40 years to begin receiving information back on Earth. A nuclear photonic rocket would have an onboard nuclear (fission) reactor, which would create enough heat to emit a broad band of very intense blackbody radiation. Analogous to the radiant heat you feel near red-hot glowing embers, the thermal radiation from nuclear fission works exactly the same but consists of highly energetic gamma and X-ray photons rather than infrared photons. If all the nuclear fuel on board a spacecraft could be converted into photon energy and it was all directed in a perfect beam out the back of the craft, it could provide an enormous impulse. However, because thrust is derived from massless photons, their overall momentum is limited. Therefore, this seemingly ideal scenario, although it is nuclear, has the problem of low thrust.

Low thrust is a problem built into photonic systems. This is actually a classic Newton’s Third Law problem. Chemically fueled rockets shoot out literally tons of mass to achieve high thrust (high acceleration), even though the spent fuel is exhausted at a comparatively low – chemical explosion-level – velocity. A rocket driven by a flow of very low-mass ions or, going further, massless photons, has an extremely low propellant mass flow (low acceleration) but very high exhaust velocity. That means that these systems can eventually achieve very high rocket velocity, but it would take a long time to reach it. A question inherent with such low-thrust systems is how to change velocity quickly enough to fall into orbit around a planet and also possibly land on a planet surface, all things that a probe mission to an exoplanet would likely have to do. It would seem a massive waste to send a craft all that way only to do a fly-by. These space flight orbital maneuvers require steep changes in velocity, which in turn require massive thrust.

On looking up “Energy requirements and comparisons” on the Nuclear Photonic Rocket Wikipedia page linked above, it paints a rather bleak picture for our hopes of a photon propulsion system for interstellar space flight, at least for a manned flight. An example 300,000 kg spacecraft set up with reasonable perimeters, will indeed reach 1/10th speed of light. That’s feasible. However, nuclear energy is not nearly as efficient as we might assume it is. Using even the highest grade of nuclear fuel possible, the fission process itself never converts all of its fuel mass into photons. It converts only about 0.10 % of it. Reaching final velocity would require a conversion of 240 grams of mass into photon energy, which would require an enormous 240,000 kg of nuclear fuel, almost all of which would have to be carried the whole journey. We can assume the full-size reactor itself would contribute significant additional mass. Even with such a high specific impulse, and here the term is estimated because fuel is not consumed in the same way as it is in a chemical reaction, it would take a year to escape Earth’s gravity from a starting position in low orbit and it would then take 80 years to reach its final speed, accelerating at a very anemic rate of 10-5 g.

2) Micro-Probe With a Laser-Driven Sail

We could totally switch up the game by going small and moving away from nuclear energy to a highly focused laser for the photon drive. It would be much easier to focus the photon beam, perhaps into a sail to push the spacecraft and it would obviously do away with the need for a nuclear reactor and all that fissionable material. However, a limitation of lasers is that they are much less efficient at converting energy into light than blackbody radiation emitters. The photon energy will also be many scales smaller than any nuclear system, even with the most powerful laser that is currently technically possible.

Still, it is a tantalizing possibility. There is no need to carry any fuel, so the idea here is to go as small as possible. The 2016 Breakthrough Starshot privately funded initiative describes a series of ultra-light probes (on the scale of a gram) accelerated toward Alpha Centauri by beaming laser photons into their relatively large (a few meters wide) sails from Earth. The mission plans to send back flyby photographs of Proxima Centauri b.

An artists conception of the Starshot solar sail deploying near Earth; Kevin Gill; Wikipedia
To quickly accelerate a sail deployed in orbit around Earth, a massive laser array would be required and it would have to be perfectly focused in order to strike the small sail, a technology not yet realized. Theoretically, these probes could accelerate very quickly because they are so low in mass. Pushed by a laser pulse lasting just a few minutes, they could top out at around 20% light speed, bringing transit time down to about 20 years. Researchers are hopeful they can overcome the obstacles associated with this plan. The probe would have to be robust to handle such rapid acceleration. The sail material would have to be perfectly reflective to the laser beam. If it absorbed even a tiny fraction of such intense photon energy, it would vaporize. Another problem is getting all the analytic tools required to explore an exoplanet down to nano-technology level. Finally there is a problem inherent with any spaceflight that is fast enough to reach another star system in a reasonable time frame. Interstellar space is full of dust (small molecules and ions) and radiation (gamma rays, etc, from various cosmic sources). These particles, though generally extremely sparse, would generate significant drag on a spacecraft travelling close to light speed, slowing it down. As well, even relatively stationary microscopic particles in space would act like high-energy bombs striking a craft going so fast, and any burst of cosmic radiation could destroy the sail. A 2-minute silent animation created by its founder company, Breakthrough Initiatives, shows how the StarShot mission to Alpha Centauri might work.


3) Ion Propulsion

An ion thruster ionizes propellant to produce a beam of ions. The thrust from an ion propulsion engine is almost imperceptible. It is equivalent to the weight of a piece of copy machine paper on your hand (about 90 mN), and this means that even a small light spacecraft would take a very, very long time to accelerate to near light speed. On the upside it is extremely efficient, achieving a specific impulse of about 10,000 seconds.

3a) Electrostatic Ion Thruster

One way to do this is to use a gridded ion thruster design, schematically shown below.

Oona Raisanen;Wikipedia
Basically, this engine uses an electron gun to direct an electron beam at the propellant material. Xenon is a good candidate for the fuel because it ionizes readily and it has the bonus of having a high atomic mass, 131.293 u (remember Newton’s third law) to maximize (albeit still very low) thrust. It’s an inert gas that is easy to store as well. Electrons bombard neutral xenon atoms in a chamber to produce positive ions as well as more electrons (to produce a plasma state). Utilizing the Coulomb force, a series of highly charged electrical grids direct and accelerate the plasma to a very high-velocity ion beam that propels the craft.

Dna-Dennis;Wikipedia
An electric field applies an electrostatic force. The electric fields around like charges repel each other (the two blue positive charges, right) while the fields around two opposite charges (the blue and red charges) supplies an attractive force. This force (denoted F) can be used to accelerate and direct charged ions.

The exhaust velocity of such an engine system can approach about 160,000 km/h. Still, the thrust of such a system would be small. Even using heavy xenon ions, the ion beam would not be able to overcome ordinary air resistance. Such a craft would have to be launched into orbit, and even once in the near vacuum of interstellar space, acceleration would be vey low. On the upside, this technology is very doable, based on test data from a variety of ion thruster engine designs.

In fact, xenon ion thrusters, as part of an NSTAR engine system, were already used very successfully to propel the recent Dawn probe to Vesta and Ceres in the asteroid belt. Arriving at Ceres in 2015, it is currently in an uncontrolled orbit around Ceres, having exhausted all of its fuel as of late 2018.

Artist’s concept of NASA’s Dawn spacecraft above Ceres; NASA/JPL-Caltech. The blue glow comes from excited ions in the engine outflow.
Powered by a 10kW photovoltaic solar array (the two long panel arms, above, powering the thrusters and all the other instruments onboard), and carrying 425 kg of xenon, it was launched in 2007 and its mission was extended until it ran out of fuel. Due to its low thrust, it took four days at full throttle after separation from its launch rocket just to accelerate from zero to 100 km/h.

3b) Electromagnetic Ion Thruster

While an electrostatic thruster uses the Coulomb force from an electrical field to accelerate ions, an electromagnetic thruster uses the Lorentz force in which both electric and magnetic fields exert forces on the ions. An electric field can be produced by a stationary charge, whereas a magnetic field is produced by a moving charge. The electric and magnetic fields, in either of these technologies, are generated using a power source, which can be electric solar panels if used near the Sun. Far from the Sun, another source such as nuclear power could be used instead.

In an electromagnetic thruster, an ion (+q, shown below left) in motion (vector v) is acted upon by two force fields. The magnetic field q(v x B) exerts a perpendicular force (vector B) while an electric field (q E) exerts a force in the same direction as the field (vector E). The Lorentz force is the force responsible for the circular trace patterns you see after particles are bombarded in accelerators.

Maschen;Wikipedia
A magnetoplasmadynamic (MPD) thruster is one type of electromagnetic thruster, the other being the pulsed inductive thruster. I’ll focus on the MPD thruster here.

Like the electrostatic ion thruster, a gas is ionized. It could be xenon once again but lithium so far has been a better performer in tests on this technology. Ionized gas particles accelerated by an electromagnetic field could in theory generate an extremely high specific impulse, achieve an exhaust velocity about three times higher a xenon ion thruster and produce a respectable thrust of 20 N and possibly even up to 200 N, which is far higher than electrostatic ion propulsion. Like an electrostatic ion thruster, an electric field is generated by a power source. The magnetic field can be externally applied to the particles through magnetic rings around the exhaust chamber. Or, the field can be induced by the ion’s electrical current while the ions are accelerated. In this case, a cathode extends down through the middle of the ion chamber. At lower power levels, the magnetic field must be externally applied because the self-induced field is too weak. A CGI rendering of Princeton University’s lithium self-field MPD thruster is shown below.

NASA;Wikpedia
By achieving this thrust in practice, this system, unlike a small light Dawn-type spacecraft, could potentially power a future heavy cargo, or even manned, flight to Mars (or beyond) because it would have enough thrust capacity to perform the kind of quick delta-v maneuver required to lock into orbit around a distant planet and land. Such a propulsion system, if used to travel to Mars, for example, would also have much higher fuel efficiency than any chemical fuel.

A problem with this propulsion system is that it would require far more power to operate than electrostatic ion propulsion, on the order of hundreds of kilowatts (kW) compared to 1-7 kW. As with all ion thrusters, thrust increases with power input. There are no current spacecraft power systems that can provide hundreds of kilowatts of power. For a manned trip to Mars, a Earth-based laser or even a very large solar panel array might do the job. For a much longer interstellar flight, however, nuclear power would be a reasonable choice. NASA’s Project Prometheus’ reactor, dropped in 2005, would have been a small nuclear reactor that could generate electrical energy in this power range to run such an ion engine. (These nuclear power generators should not to be confused with nuclear thermal propulsion, described earlier). This technology would also have been useful for deep space probes in our system that are too far from the Sun to use photovoltaic panels.

Several countries around the world have experimented with MPD thruster technologies. Russia, then the Soviet Union, flew some experimental prototypes on their spacecraft and more recently Japan  successfully operated an MPD thruster in space as part of an experiment. Research on this kind of ion thruster is still being carried out today at the Electric Propulsion and Plasma Dynamics Lab at Princeton University as well as at NASA’s Jet Propulsion Laboratory and Glenn Research Center.

5) Antimatter Propulsion

An antimatter rocket might sound a bit like science fiction but this technology could be one of the most promising for interstellar travel, provided a few practical matters are sorted out. Antimatter has the highest energy density of any proposed rocket fuel. In fact, up to 75% of the mass of the matter/antimatter fuel mixture is directly converted into energy. This is an incredibly high figure but it is not 100% as one might at first guess. Only the annihilation of electrons with positrons yields 100% energy. In practice, other particles and their antimatter twins annihilate each other as well, yielding both energy and a variety of yet more particles, representing mass that is not converted to energy. Condensed antihydrogen fuel might someday be feasible. However, at least with current technology it would be extremely difficult to create any sizeable quantity of antihydrogen. First, antiprotons and positrons (anti-electrons) need to be created. Positrons are relatively very easy to make. Physicists at Lawrence Livermore National Laboratory in California use a special laser to irradiate a plate of gold to create billions of positrons per pulse.

Antiprotons are not easy to make. A few antiprotons are created for each one million particle collisions in a particle collider. These rare particles are separated out from other particle products using a magnetic field. Then, these very fast antiprotons need to be slowed down using electric and magnetic fields so that they can capture positrons in order to make antihydrogen atoms. Then the atoms need to be trapped in an ultracold magnetic “bottle.” So far, a bottle of a few (less than 100) antihydrogen atoms have been stored for almost 17 minutes before they annhiliated. If CERN used its colliders only to make antimatter, it could only make 1 billionth of a gram per year! It takes about a billion times more energy to make antimatter than what is stored in its mass. One (likely very generous) estimate puts the price of 1 gram at about a trillion dollars US. NASA estimates that a trip to Mars would take around 20 milligrams of antimatter. The xenon or lithium required for ion thruster engines, described above, is quite expensive too but nowhere near this scale. This being said, the tiny amounts of antihydrogen mankind might someday feasibly manufacture would provide one hell of a wallop, on the scale of several billion times more than the most efficient chemical rocket fuel.

An antimatter rocket could use the products of annihilation for direct propulsion or it could be used indirectly to power a different drive, for example, to heat a fluid such as liquid hydrogen, which would be expelled, and which would potentially supply more thrust.

The theoretical antimatter AIMStar (Antimatter-Initiated Microfusion) rocket, developed in the 1990’s, would use gamma rays created from matter/antimatter annihilation to trigger a fission nuclear reaction in a deuterium/tritium mixture that would in turn start a fusion burn within that mixture. That superheated plasma would be ejected to propel the rocket using a series of fusion pulses. It would have a specific impulse of about 61,000 s and achieve an exhaust velocity of about 1/3 the speed of light. AIMStar propulsion technology might be feasible in a few decades. Developed by Penn State University, it could be used to visit the distant Oort Cloud, 10,000 A(astronomical units) away. The mission itself would take several decades, accelerating constantly for 22 years to achieve 3/1000th light speed, and require 28.5 micrograms of antimatter, more than we can create with current technology.

Fusion Propulsion

I briefly mentioned fusion as part of a theoretical antimatter engine earlier, but I think the fascinating possibility of nuclear fusion as a practical energy source for propulsion deserves a closer look. One reason that fusion propulsion theory is attractive is it is fuel-efficient. A typical chemical rocket engine has a specific impulse of about 450 s. A fission-based design, such as Orion, would have maximum specific impulse of about 10,000 s. By contrast, a fusion rocket would have a specific impulse over ten times higher, at about 130,000 s.

This option comes with serious challenges, however. One of them is low thrust. The first fission drive concept, Project Orion, demonstrated the potential power of nuclear energy for space travel. It promised high specific impulse AND thrust. Expelled from a series of fission detonations, debris particles, radiation and heavy radioactive nuclei could theoretically achieve an engine output average velocity of around 100,000 km/h and achieve an eventual maximum velocity in interstellar space of an impressive 36 million km/h, just over 3% light speed.. Because the debris would contain lots of large atomic nuclei (lots of mass), it would have high thrust as well. It is one of the few theoretical propulsion systems that could achieve significant thrust and impulse at the same time, a chief advantage of a fission system.

The bulk of the blast from a fusion reaction, in contrast, is made up of massless gamma rays and (low mass) protons. It’s very low in mass but it’s going very fast (on average close to slight speed), delivering an enormous specific impulse but low thrust. There are two general options for nuclear fusion propulsion. Direct propulsion means that the plasma itself is directed out the back of the craft: low thrust. Indirect propulsion means that fusion energy is used to generate electricity, perhaps to power an MPD ion drive. This is a propulsion system that is very efficient but requires at least a few hundred input watts of electricity. Fusion power could supply that, and the efficient MPD drive could achieve enough thrust to push a heavy cargo rocket or even a manned spaceflight over very long distances at significant velocity.

Besides low thrust, the biggest challenge of fusion is achieving ignition temperature. A fusion rocket will, out of necessity, use fusion plasma because that is where a fusion reaction takes place, in extremely hot pressurized plasma (millions of degrees C). In fact, in nature one must look into the interior of a star to find matter in this extreme state. In order to get plasma energetic enough to initiate a fusion reaction, it must be confined under extreme pressure – pressure equivalent to the interior of a star as massive as our Sun – and kept stable. The problems with fusion, in a nutshell, are how to heat gas into a fusion state, how to confine it and how to keep it stable. You must keep nuclei energetic and in close enough proximity to fuse together in an ongoing stable reaction.

Nuclear plasma cannot be confined physically. It is far too hot and under too much pressure for any material in contact to withstand, even momentarily. However, there are other ingenious ways to tame it, and these methods come with the added advantage that they can be very lightweight and compact. Weight is an overarching concern with long-term spaceflight. Everything is designed to minimize the mass that must be hauled across space, so that energy expenditure can be minimized. Fuels can be heavy, adding their own load, and their supply is limited. The main advantage of a fusion propulsion system is that it can be very efficient, requiring a smaller and much lighter fuel supply than a fission system. The fuel would consist of hydrogen isotopes such as deuterium or tritium rather than a heavy fissile isotope such as uranium-235 or plutonium-239. In addition to this advantage, more fuel mass would be turned into energy in a fusion drive than in a fission drive. In a fusion drive, 676 units of energy would be converted from 1 kg fusion fuel. A fission reaction would yield 176 units of energy/kilogram of fuel. See also that hyperphysics website link to see how these values are calculated. Furthermore, because fusion produces less radiation than fission does, less, necessarily heavy, shielding would be required to protect sensitive hardware. A fusion drive will still require a reactor, but it could be built much lighter and smaller than a fission reactor if technologies such as magnetic or inertial plasma confinement are used. Not only can these methods heat the plasma but they also keep it under pressure.

Magnetic confinement works by inducing a circular electrical current in plasma. The current creates a magnetic field that squeezes the plasma into a thin ring. This results into two kinds of heating: Joule heating and adiabatic heating. Joule heating is microscopic friction caused when a current passes through a conductor. During adiabatic heating, the thermal energy (temperature) of a gas or plasma increases as it is compressed, according to the first law of thermodynamics. A technology that currently uses magnetic fusion confinement is the tokamak, a Russian-designed device that confines and stabilizes plasma by winding magnetic field lines in a helix around a toroid (doughnut) shape. Several tokamaks are now working around the globe. To make such a device feasible as a generator, it must produce more energy than is required to maintain the fuel in a fusion state, a tall order. The first fusion tokamak-type full-size reactor, called ITER, just achieves this and is currently under construction in France. The problem with current tokamak technology is that the toroidal reactor is very heavy, severely reducing a potential craft’s weight/thrust ratio.

A different and more feasible approach for a fusion drive would be to go small and introduce a different confinement method. One could focus intense energy onto a very small target, using a powerful laser to heat and compress a tiny fuel pellet of the hydrogen isotopes tritium and deuterium until the pellet ignites a fusion reaction. The fuel is relatively cheap and plentiful, not just on Earth but elsewhere in the universe as well, and the system could be highly efficient, perhaps used for a power source for a secondary thrust technology. An intense focused blast of laser photons heats the outer layer of the fuel pellet exploding it outward in all directions, producing an equal reaction force inward in the form of shockwaves. The shockwaves compress and heat the interior of the pellet enough to ignite fusion. This is an example of Newton’s third law in action.

This type of fusion requires a very powerful and focused laser in the vicinity of megajoules (MJ), which is doable. The National Ignition Facility (NIF), which researches inertial confinement fusion, has such a laser in use. This system also requires a perfect sphere of fuel in order to create a shockwave that is properly focused inward. A way around this issue has been to surround the fuel in a tiny metal cylinder. The laser focuses on the inner surface of the metal, heating it into plasma, which then radiates X-rays. The thermal energy from the X-rays is absorbed by the fuel sphere (more efficiently than by the photons). It then perfectly implodes into fusion plasma.

To directly produce an impulse from this reaction, a magnetic field could confine and direct the intense blast energy of the fusion explosion. The field could also be set up to form a pusher plate. A tremendous amount of energy, in the form of gamma rays, released from a series of fusion pulses could push against a pusher plate, in this case a magnetic field, in order to translate force into acceleration. The plate could also be designed to absorb the shock of the individual pulses to allow for smooth acceleration of the craft. The drawback of this system, as mentioned, is its very low thrust.

Relativistic Travel And Einstein

It would seem that once we work out extremely efficient and continuous fuel consumption, our future interstellar spacecraft, perhaps even manned, should be able to accelerate across the light years of space, perhaps at a comfortable rate of 1 g, the same as Earth’s gravity, to eventually reach light speed. After all, can’t we simply use the formula f = ma? Newton’s second law states that force equals mass times acceleration. Imagine the scenario. We can use the rocket equation to refine our calculations. The rocket equation deals with the fact that we must accelerate both ship and fuel mass and the fuel mass is decreasing as we go. Then imagine that we can reverse the engine thrust and decelerate at 1 g, over several years, to reach planet X. We can assume there is no friction in this space vacuum to work against us. We will assume there are no particles in interstellar space (which is not the case and would cause a serious problem at relativistic velocity). Here, then, is our system in effect: the propellant removes momentum in one direction so the ship can gain momentum in the other direction.

The problem is in the assumption of f = ma. This Newtonian equation assumes that the relationship between mass and acceleration always holds up. And it does, but only until we get to relativistic velocities, a scenario Newton, I suspect, never thought of. At these velocities, the relationship starts to break apart. We must translate our motion equations from Galilean transformations into Lorentz transformations. This sounds complicated, but it means that at relativistic velocities, space-time itself starts to makes itself known within our accelerating spaceship scenario. For Newton, space and time are invariant. You can always count on them to be the same. A ruler is always one foot long and time always flows at the same rate. A second lasts a second. For Einstein, space and time are no longer invariant and we can no longer think of them as separate from each other either. Now they must be treated as part of a 4-dimensional single continuum, space-time. Two things to consider: space-time can stretch and warp and, second, we now must think of any situation in terms of frame of reference. The ONLY constant that doesn’t change with frame of reference is the speed of light.

As our ship revs up to, say 30% light speed, we would begin to notice changes. Time dilates – clocks on the ship slow down compared to those on Earth. Conversely, measured from the ship, Earth’s clocks speed up. Length contracts in the direction of motion. This means that even though length contracts, from Earth the ship would look fine. From a dock in which the ship whizzed past, however, you would see the ship look shorter or squished. If the ship could travel exactly at light speed, it would have no length at all in this reference point. Relative mass increases. The ship, as measured from Earth, would be growing more massive.

What does all this mean for Newtonian dynamics? At lower velocities, Newtonian physics holds up. As velocity becomes relativistic, the ship’s momentum/velocity/acceleration begins to lose what looked like a linear relationship to one another. This happens, according to Einstein, because the time coordinates between the ship and the observer (such as us on Earth) are no longer the same. Time, distance, and mass are now relative, that is, their values depend on your frame of reference. Only the speed of light and the basic laws of physics remain constant from every reference point in the universe.

This isn’t intuitive. The ship’s changes we would measure from a different reference point such as Earth - its mass, its physical dimensions and its time dilation - are not optical illusions. They are real valid measurements. Yet, their values depend entirely upon where we measure them. So, if you were inside the ship, your clock would be running normally, your mass would be normal and the ship would be of the same dimensions as always. Looking out the front window, however, you would notice that your frame of reference is now drastically different from the relatively stationary space around you. The stars whizzing past the ship would begin to bunch up closer, the distances between them shrinking, while the starlight reaching the front window of the ship shifts to blue (Doppler effect). In your reference, they are approaching relativistic velocity toward you.

The spaceship can never exceed light speed. Witnessed from Earth, the reason is clear. As the ship approaches light speed, more and more of the force of 1 g acceleration is converted into RELATIVISTIC mass rather than velocity, until, as the theory goes, the relativistic mass approaches infinity, an impossibility, at light speed. This doesn’t mean that, at any and all reference points, your ship is now so massive it would take infinite force to continue to accelerate it (although that’s what we would correctly observe from Earth). At the same time, all processes on the ship would appear from Earth to slow down to a stop. Frozen there, there would be no way to advance further.

Time, mass and distance in space-time change as your reference changes. Inside the ship, now approaching light speed, you would see your star field bunching up, turning blue and then falling entirely beyond your visual range into an almost perfectly black invisible surface in front of your window. Unfortunately you would soon need to install a powerful blast shield because the photons of starlight will eventually transform into X-rays and then deadly gamma rays as the Doppler shift continues to shorten the wavelength of this electromagnetic radiation. Is there a limit to the Doppler shift of a photon? The wavelength of a photon should theoretically approach zero as the ship approaches light speed but there is no theoretical description that makes sense for a shorter-than-zero photon wavelength. The photon energy would approach infinity. The speed of the photons, however, will always be the same, because light speed is invariant.

Time would seem to speed up for all the processes going in space in front of your ship, say an exploding supernova somewhere. At even 90% light speed, you might still be able to perceive the explosion by instruments. Once your ship passes the supernova you might see it – a Doppler-shifted slow motion explosion – in you rearview mirror. You will have to trust Einstein to continue accelerating to a final velocity of, say 99% light speed, and then decelerate when you approach your destination. As you do so, the star field will deepen, come back into view, as the reel of time once again becomes your time. Remember, everything inside your ship is normal for the whole journey. It’s the objects and their motions outside it, moving near light speed relative to your ship that will seem weird.

A tantalizing feature of near light-speed travel comes from time dilation. As time relatively slows down inside the spacecraft, including for any occupants, time relatively speeds up outside of it, including on Earth. A trip to a star system 100 light years away could be shortened into, perhaps a few years FOR the occupants. This seems to break the rules of faster-than-light travel, but in the reference frame of the ship and its occupants, space contracts and stars appear much closer to each other. What, for Earth would be at least a 100-year trip, would be experienced as much shorter by the occupants because their clocks are running very slowly compared to the universe (and normally for them). To them, the distance to the star system would be just a few light years away, perhaps less, depending on how close to light speed the ship traveled and how long it would take to accelerate and decelerate. Consider that if a ship could travel at exactly light speed (which it cannot), the universe would be completely length contracted (there would be no distance between stars or galaxies at all in the direction that it is moving) and time would not exist, as time in the universe surrounding the ship would run infinitely fast, an infinite nonsense answer to the equation that describes relativistic motion. To a photon, then, there is no time or distance between its emission and its eventual absorption. That is its universe, a mind-bending thought for the next time you are in the shower. Sadly for any ultrafast interstellar travellers, Earth will have run like a movie on extreme fast forward, a minimum of over 200 years would have passed by the time the travellers returned, a few years later.

Unfortunately, this journey is unfeasible. It is impossibly perilous. If your ship strikes even a cold gas cloud in space, it will blast into smithereens. Every gas atom will be a significantly massive relativistic missile striking your ship hull. Unless the ship has a powerful magnetic field to deflect particles away perhaps . . .