F-22A Raptor, FB-22, F-22E, F-22N and Variants Index Page [Click for more ...] People's Liberation Army Air Power Index Page  [Click for more ...]
Military Ethics, Culture, Education and Training Index Page [Click for more ...]
Russian / Soviet Weapon Systems Index Page [Click for more ...]






Last Updated: Mon Jan 27 11:18:09 UTC 2014








Microwave and Millimetric Wave Propagation

Originally published  May, 2000
by Carlo Kopp
© 2000, 2005 Carlo Kopp

The impending death of the multi-billion dollar Iridium constellation is without any doubt a great setback for the satellite communications industry, and microwave communications industry as a whole. What was to be a bonanza, a new frontier of truly global mobile communications, has collapsed into a mire of insolvency and massive commercial losses.

Recent mass media reports suggest that the constellation will very soon be destroyed, by slowing the satellites into descending orbits, to burn them up in the upper atmosphere.

The demise of Iridium is a sad outcome for the world's first Low Earth Orbit (LEO) communications system. The important question which we can ask is that of what persuaded so many engineers and investors to commit so much to a risky venture of this nature. What was the perceived advantage which led them to this eventual commercial quagmire and final disaster ?

The summary answer is microwave propagation. However, to appreciate the reasons fully, we must delve a little deeper into the problem. This indeed will be the subject of this month's feature.

Basic Issues in Propagation

The problem of getting a radio frequency signal, in the decimetric, centimetric or millimetric bands, between two geographically well separated antennas, is in principle simple but in practice messy.

In a genuine free space environment, typified by a crosslink between a pair of satellites in higher orbits, or a pair of spacecraft in interplanetary space, the problem is as simple as pointing the antennas at each other and launching the wave into space. Since the wave obeys the well known 1/r^2 Friis equation, it will diminish in intensity with the inverse square of distance. The power detected by the receiver will depend upon the respective gains of the two antennas, the transmitted power level, and the distance between the transmitter and the receiver. Providing that there is no powerful radio source, such as the sun, within the mainlobe of the receiver's antenna, the signal to noise ratio, and thus achievable throughput per bandwidth, will depend wholly upon the Friis equation and the noise performance of the receiver.

The bit rate achieved across such a link will thus depend on the distance, antenna specs, power output and quality of the receiver. Indeed, every undergraduate comms engineer will have solved this problem in a third or fourth year University assignment or exam paper.

Alas, the "free space" model of radio propagation is very much a "best case scenario", and one which is not frequently encountered in daily operation, with the exception of trivial cases like microwave links between adjacent buildings.

Mother nature, having little respect for the expectations of engineering undergraduates, likes to complicate things a little, and introduces a number of interesting, albeit often difficult to solve, obstacles in this game. These are the effects of refraction, lensing, scattering, and absorption in the natural environment.

The Atmosphere as a Propagation Environment

Virtually all of the difficulties encountered in microwave transmission, be it between terrestrial transceivers, or satellite ground stations, stem from the physical properties of the atmosphere.

The atmosphere is the gaseous shroud covering mother earth, comprising mostly nitrogen, with a decent fraction of oxygen and carbon dioxide, and at lower altitudes also water vapour or droplets. The atmosphere obeys the pull of gravity, indeed the earth's gravity well is what keeps it attached to the planet. Planets with weaker gravity wells cannot maintain an atmosphere, and the gaseous shroud is blown into space over time.

The effects of gravity are pronounced, and manifested in decreasing atmospheric density and pressure with increasing altitude. At the earth's surface, the temperature, density and pressure are highest, as a result of which the propagation impairments are most pronounced.

The atmosphere is divided into layers. The lowest of these is the troposphere, which extends up to about 11 kilometres, depending upon the geographical latitude. In the tropics it may extend up to 15 kilometres, due to the updrafts produced by massive tropical rainbearing cumulonimbus clouds. The troposphere is the warmest, most dense and wettest layer of the atmosphere. I like to call it the "tropospheric soup" since it has, in propagation terms, the attributes of a rich broth full of ingredients.

Above the troposphere is the stratosphere, the domain of Concordes and supersonic military jets. Devoid of dense cloud, much colder and thinner than the troposphere, the stratosphere is a much more benign environment for microwave transmissions.

Whether we are considering a terrestrial link or a satellite link, the effects of the atmosphere frequently dominate losses in propagation. Designers of microwave links ignore it at their peril.

Refraction Effects

Refraction is a physical phenomenon observed in any medium which has a varying refractive index, and produces the effect of bending a light ray or microwave beam. The atmosphere is exactly such an environment, since its density and thus refractive index varies significantly with changing altitude and weather conditions.

At a first glance one might think this is only an issue for satellite transmission, but the curvature of the earth makes it an issue even for terrestrial links overs tens of kilometres of distance, should the beam be particularly narrow.

Refraction usually produces desirable effects, insofar as it can allow a pair of stations to communicate over the horizon, since the beam is effectively bent. It can also allow a satellite ground station at high latitudes to see a geostationary satellite over the equator, from positions which would appear infeasible geometrically.

However, refraction is a double-edged sword. This is for two reasons. The first is that refraction can also allow signals to interfere with other links by propagating over the horizon. This problem is exacerbated by a peculiar effect called "ducting", which can arise when a meteorological effect called an inversion occurs. An inversion happens when the layer of air closer to the earth's surface is colder than the air above it. When this happens, usually under still and warm conditions, the refractive index change in the atmosphere with altitude is such, that a layer of the atmosphere behaves like a waveguide. Microwaves which enter this layer, termed a duct, cannot escape. Like a light ray trapped in an optical fibre, they will propagate far beyond the horizon. I recall the odd experience, many years ago on a European trip, of watching a TV broadcast from North Africa which had ducted itself into central Europe !

Refraction can be an issue for many satellite links, especially GEO links at extreme latitudes. A large temperature change can cause link dropouts or interference.

In practical terms, refraction is a hindrance mostly to GEO satellite links, due to the lower depression angle over the horizon, with increasing latitude. Refraction effects are however minimised with the use of an LEO constellation, since the satellite within a cell is always at a steep elevation angle above the ground station.

Lensing effects arise from a combination of refraction and the earth's curvature, and usually are not significant.












Absorption and Scattering

Absorption and scattering are the dominant sources of microwave band, and especially millimetric wave band transmission losses in the lower atmosphere.

The primary source of absorption losses at all altitudes is an effect called gaseous absorption, which results from the quantum physical behaviour of atmospheric gas molecules. A gas molecule of any species, such as nitrogen, oxygen, carbon dioxide or water vapour experiences resonance effects, not unlike a tuning fork. This analogy is a very good one, insofar as each of these molecules at an atomic level is made up of several nuclei, held together by electromagnetic forces, and therefore behaves in a manner not unlike balls held together by springs. If you perturb the molecule, it will vibrate in a number of possible modes, be they rotational, flexural or longitudinal, depending on the shape of the molecule and the manner it was excited.

What is important for the microwave engineer is that these resonances produce electromagnetic effects. If such a molecule is placed inside a microwave beam, and if the frequency of the beam is close enough to the resonance frequency of the molecule, the molecule will draw energy from the beam as it is excited. If you put enough molecules into the beam, an appreciable amount of power will be lost.

If we are trying to send a microwave signal between a ground station and a satellite, the beam must pass through hundreds of kilometres of atmospheric gasses. Therefore many deciBels of power can be lost.

How much is lost depends upon the frequency of the carrier wave, the pathlength through the atmosphere and local atmospheric conditions.

In times past this was not a major issue, since at frequencies below 10 GHz gaseous losses are almost irrelevant. However, the water vapour molecule has a strong resonance at 22.235 GHz, and moving into the millimetric band, the oxygen molecule has a cluster of resonances around 60 GHz. Moving beyond 60 GHz, further water vapour resonances arise above 100 GHz.

The collective effect of these resonances is depicted in Figures 1, 2 and 3. Figure 1 shows the increasing effect with decreasing altitude, Figure 2 shows the frequency dependency of these losses, for various altitudes, and finally Figure 3 shows the detailed loss behaviour around 60 GHz, all in dB/km loss rates.

The practical consequences of this behaviour are that many frequencies in the millimetric band verge upon the unusable, especially for link distances in excess of several kilometres. This is of importance in local multi-point distribution schemes, where a signal is fed through a fibre to a local microwave antenna head, but it is also critically important for satellite links. Since the frequency bands below 15 GHz are virtually saturated in OECD countries, there is much pressure to start using frequencies above 15 GHz, especially for satellite links. The 28 GHz sub-band has been very popular, since it sits in a "trough" in the loss behaviour curve. Even so, the dB/km loss is almost tenfold that at 10 GHz.

While this matters at shorter distances, it becomes a "go/no-go" factor for satellite links. Also this is the reason why LEO systems have significant propagation advantages over Medium Earth Orbit (MEO) and GEO systems. In both latter instances, the elevation angle of the beam can be very shallow, which introduces a significant increase in the atmospheric pathlength the beam must propagate through. An LEO system can be designed to ensure that the beam elevation angle is of the order of 45 degrees or more, thus minimising the pathlength through the atmosphere.

Gaseous losses are unavoidable as long as we situate our antennas on the surface of the earth. They will vary somewhat with increases or decreases in local humidity, but in principle, cannot be escaped. Even under best case conditions of clear sky, they are ever present.

Scattering losses are no less troublesome a problem in microwave propagation, be it point to point links or satellite links. While they vary significantly with local weather conditions, a complexity within itself, they too cannot be avoided in most parts of the world.

A scattering loss will arise when the microwave beam encounters droplets or particles in the atmosphere. If these particles are smaller than a wavelength, an effect called Raleigh scattering occurs, whereby the droplet or particle reflects a small proportion of the impinging energy, not unlike an aeroplane in a radar beam. Indeed the physics involved are fundamentally the same.

How much energy is scattered and never reaches the receiver depends upon the size of the scatterers relative to the wavelength, their density per volume of the atmosphere, the pathlength through the scattering environment, and the dielectric properties of the scatterers.

The most common source of scattering losses is the humble cloud. Made up of microscopic water droplets, clouds vary significantly in moisture content and thus lossiness. Low density clouds like stratus, stratocumulus and puffy little summer cumulus clouds introduce some losses, increasing with frequency, but are almost insignificant in comparison with dense water laden rainclouds, and especially the cumulonimbus thunderstorm cloud. Figure 4 depicts the loss coefficient in dB/km per cloud density, in grams per cubic metre, against frequency and temperature. At 40 GHz a cloud with a density of 10 grams/m^3 will introduce a loss of 8-20 dB/km. If the beam must travel through 3 km of such cloud, the total loss varies between 24 to 60 dB, which is most instances renders such a link unusable.

While cloud related losses are usually not an issue below 15 GHz, they become an increasingly serious issue with increasing frequency, and there are no troughs in the curve whereby an engineer can cheat ! The millimetric bands above 40 GHz are virtually compromised for satellite work, and even the Ka band 20-35 GHz window in gaseous loss behaviour is rather exposed.

Of course the same caveats concerning pathlength also apply, so for a satellite link, the steeper the elevation angle, the better.

Where there is cloud, there is frequently rain, and rain like cloud is a scattering environment. Rain droplets however tend to be much larger in size, compared to cloud droplets and thus behave a little differently. Like cloud, rain will scatter increasing amounts of the microwave signal with increasing droplet density. Figure 5 depicts the popular Olsen and Rogers semi-empirical rain loss model, showing the dB/km loss against rainfall rate in mm/hr. This plot compares a computer model against CCIR empirical data. Like cloud losses, rain losses increase dramatically with frequency. A 6 inch/hr rainfall, decidedly heavy, will introduce tens of dB/km of signal loss at 40 GHz, presenting an almost impenetrable barrier for millimetric wave transmission.

Since rain tends to be transient, and average rain rates vary enormously across the world's geography, the CCIR and NASA have published extensive charts which divide the surface of the earth into zones, each with characteristic average and worst case rainfall behaviour. The intent is to allow engineers to calculate the average availability of a link, as a function of operating frequency and geography.

This model has served us well for links operating below 15 GHz, since at these wavelengths rain losses predominate over cloud losses. The problem which the satellite community now faces with the above 20 GHz is that cloud losses can be significant, and whereas rain tends to come in transient bursts, cloud cover may hang around for days at a time.

The traditional approach to this problem has been to use to the idea of "spatial diversity", whereby multiple satellite ground stations are situated several kilometres apart. This relies upon the fact, that rain showers tend to be transient and localised. Therefore if one or two antennas are blinded, the others can still operate, and the link remains functional.

Nature however is less forgiving, as we push the frequency beyond 30 GHz, and odds are that unless ground stations are tens of kilometres or more apart, they will fall under the same slab of cloud cover.

What is also obvious if we delve into the literature, is that little effort has been expended in gathering statistical data on cloud coverage and density, in the manner done for rainfall behaviour. Since there is no simple relationship between cloud cover and rainfall rates, it is not possible to easily establish the reliability of a satellite link above 20 GHz.

While considerable commercial pressure exists to satellite and terrestrial communications into the upper microwave and millimetric bands, the sad truth is that the propagation behaviour of these bands is far from ideal, and the cumulative base of research, especially in areas like the statistical coverage behaviour of cloud, is inadequate for robust link engineering.

The problems with propagation behaviour are unfortunate, in the sense that the upper bands allow for very compact antennas and very tight beamwidths, both of which are highly desirable from the engineer's perspective. Tight beams make for efficient use of transmitter power, and better security by making it hard to eavesdrop.

Because the pathlength through a layer of cloud or rain varies with the inverse of the sine of the elevation angle, the steeper the angle is, the lower the propagation loss.

This fact of life is an irresistible temptation for a satellite communications engineer, and the reason while LEO systems are so trendy at this time. The pressure to move into the upper bands forces solutions which can best cope with propagation losses, and LEO systems have an unbeatable advantage over GEO systems in this respect.

The weakness of LEO systems is that by definition they must be global in coverage, the GEO systems game of parking a single satellite above the equator in line with the intended footprint simply does not apply. A GEO system can be "efficient" in the sense that its footprint can be concentrated in the best revenue bearing areas of geography, such as the continental US and EU.

What killed Iridium, and may yet prove to be the downfall of Teledesic, is that the LEO system is by default a much larger enterprise than a GEO sat, and most of its footprint covers oceans and Third World nations which by default are unable to produce good revenue. Unless an LEO system can produce enough revenue from the US and EU markets to survive, it is doomed to failure since the rest of the planet simply cannot pay the LEO telephone bill.

Are there are any real pluses in the upper microwave and millimetric band propagation game ? The simple answer is no, the environment is in the simplest of terms "pathological", compared to the established bands below 15 GHz. The engineering demands are much higher, and many wavelengths are unusable. Does this mean that we should abandon the use of these bands ? The answer is that we no longer have that choice, since the lower bands have become virtually saturated with services.

Terrestrial microwave communications have become a second tier player, compared to optical fibres, and this does suggest that over time, many lower band microwave frequencies will be freed up for satellite use. Providing that system designs can cope with rain losses, not a critical problem in many parts of Australia, there may be some potential to exploit the upper bands for terrestrial services.

For Australia's microwave networking community, many of whom may be very excited about the latest generation of 28 GHz LMDS hardware, a note of caution is appropriate: ensure that your weather models and power budgets are well researched, since the propagation environment above 20 GHz is by any measure, unforgiving.









People's Liberation Army Air Power Index Page [Click for more ...]
Military Ethics, Culture, Education and Training Index Page [Click for more ...]
Russian / Soviet Weapon Systems Index Page [Click for more ...]





Artwork, graphic design, layout and text © 2004 - 2014 Carlo Kopp; Text © 2004 - 2014 Peter Goon; All rights reserved. Recommended browsers. Contact webmaster. Site navigation hints. Current hot topics.

Site Update Status: $Revision: 1.753 $ Site History: Notices and Updates / NLA Pandora Archive