Three massive-MIMO-related highlights from IEEE ICC in Kansas City, MO, USA, this week:
J. H. Thompson from Qualcomm gave a keynote on 5G, relaying several important insights. He stressed the fundamental role of Massive MIMO, utilizing reciprocity (which in turn, of course, implies TDD). This is a message we have been preaching for years now, and it is reassuring to hear a main industry leader echo it at such an important event. He pointed to distributed Massive MIMO (that we know of as “cell-free massive MIMO“) as a forthcoming technology, not only because of the macro-diversity but also because of the improved channel rank it offers to multiple-antenna terminals. This new technology may enable AR/VR/XR, wireless connectivity in factories and much more… where conventional massive MIMO might not be sufficient.
In the exhibition hall Nokia showcased a 64×2=128 Massive MIMO array, with fully digital transceiver chains, small dual-polarized path antennas, operating at 2.5 GHz and utilizing reciprocity – though it wasn’t clear exactly what algorithmic technology that went inside. (See photographs below.) Sprint already has deployed this product commercially, if I understood well, with an LTE TDD protocol. Ericsson had a similar product, but it was not opened, so difficult to tell exactly what the actual array looked like. The Nokia base station was only slightly larger, physically, than the flat-screen-base-station vision I have been talking about for many years now, and along the lines that T. Marzetta from Bell Labs had already back in 2006. Now that cellular Massive MIMO is a commercial reality… what should the research community do? Granted there are still lots of algorithmic innovation possible (and needed), but …. Cell-free massive MIMO with RF over fiber is the probably the obvious next step.
T. Marzetta from NYU gave an industry distinguished talk, speculating about the future of wireless beyond Massive MIMO. What, if anything at all, could give us another 10x or 100x gain? A key point of the talk was that we have to go back to (wave propagation) physics and electromagnetics, a message that I very much subscribe to: the “y=Hx+w” models we typically use in information and communication theory are in many situations rather oversimplified. Speculations included the use of super-directivity, antenna coupling and more… It will be interesting to see where this leads, but at any rate, it is interesting fundamental physics.
There were also lots of other (non-Massive MIMO) interesting things: UAV connectivity, sparsity… and a great deal of questions and discussion on how machine learning could be leveraged, more about that at a later point in time.
No, these are two different but somewhat related concepts, as I will explain in detail below.
Contemporary multiantenna base stations for cellular communications are equipped with 2-8 antennas, which are deployed along a horizontal line. One example is a uniform linear array (ULA), as illustrated in Figure 1 below, where the antenna spacing is uniform. All the antennas in the ULA have the same physical down-tilt, with respect to the ground, and a fixed radiation pattern and directivity.
By sending the same signal from all antennas, but with different phase-shifts, we can steer beams in different angular directions and thereby make the directivity of the radiated signal different from the directivity of the individual antennas. Since the antennas are deployed on a one-dimensional horizontal line in this example, the ULA can only steer beams in the two-dimensional (2D) azimuth plane as illustrated in Figure 1. The elevation angle is the same for all beams, which is why this is called 2D beamforming. The beamwidth in the azimuth domain shrinks the more antennas are deployed. If the array is used for multiuser MIMO, then multiple beams with different azimuth angles are created simultaneously, as illustrated by the colored beams in Figure 1.
If we would rotate the ULA so that the antennas are instead deployed at different heights above the ground, then the array can instead steer beams in different elevation angles. This is illustrated in Figure 2. Note that this is still a form of 2D beamforming since every beam will have the same directivity with respect to the azimuth plane. This antenna array can be used to steer beams towards users at different floors of a building. It is also useful to serve flying objects, such as UAVs, jointly with ground users. The beamwidth in the elevation domain shrinks the more antennas are deployed.
If we instead deploy multiple ULAs on top of each other, it is possible to control both the azimuth and elevation angle of a beam. This is called 3D beamforming and is illustrated in Figure 3 using a planar array with a “massive” number of antennas. This gives the flexibility to not only steer beams towards different buildings but also towards different floors of these buildings, to provide a beamforming gain wherever the user is in the coverage area. It is not necessary to have many antennas to perform 3D beamforming – it is basically enough to have three antennas deployed in a triangle. However, as more antennas are added, the beams become narrower and easier to jointly steer in specific azimuth-elevation directions. This increases the array gain and reduces the interference between beams directed to different users, as illustrated by the colors in Figure 3.
The detailed answer to the question “3D Beamforming, is that Massive MIMO?” is as follows. Massive MIMO and 3D beamforming are two different concepts. 3D beamforming can be performed with few antennas and Massive MIMO can be deployed to only perform 2D beamforming. However, Massive MIMO and 3D beamforming is a great combination in many applications; for example, to spatially multiplex many users in a city with high-rise buildings. One should also bear in mind that, in general, only a fraction of the users are located in line-of-sight so the formation of angular beams (as shown above) might be of limited importance. The ability to control the array’s radiation pattern in 3D is nonetheless helpful to control the multipath environment such that the many signal components add constructively at the location of the intended receiver.
Our 2014 massive MIMO tutorial paper won the IEEE ComSoc best tutorial paper award this year. The idea when writing that paper was to summarize the state of the technology, and to point out research directions that were relevant (at that time). It is of course, reassuring to see that many of those research directions evolved into entire sub-fields themselves in our community. Naturally, in the envisioning of these directions I also made some speculations.
It looks to me now that two of these speculations were wrong:
First, “Massive MIMO increases the robustness against both unintended man-made interference and intentional jamming.” This is only true with some qualifiers, or possibly not true at all. (Actually I don’t really know, and I don’t think it is known for sure. It seems that this question remains a rather pertinent research direction for anyone interested in physical layer security and MIMO.) Subsequent research by others showed that Massive MIMO can be extraordinarily susceptible to attacks on the pilot channels, revealing an important, fundamental vulnerability at least if standard pilot-based channel estimation is used and no excess dimensions are “wasted” on interference suppression or detection. Basically this pilot channel attack exploits the so-called pilot contamination phenomenon, “hijacking” the reciprocity-based beamforming mechanism.
Second, “In a way, massive MIMO relies on the law of large numbers to make sure that noise, fading, and hardware imperfections average out when signals from a large number of antennas are combined in the air.” This is not generally true, except for in-band distortion and with many simultaneously multiplexed users and frequency selective Rayleigh fading. In general the distortion that results from hardware imperfections is correlated among the antennas. In the special case of line-of-sight with a single terminal, an important basic reference case, the distortion is identical (up to a phase shift) at all antennas, hence resulting in a rank-one transmission: the distortion is beamformed in the same direction as the signal of interest and hardware imperfections do not “average out” at all.
This is particularly serious for out-band effects. Readers interested in a thorough mathematical treatment may consult my student’s recent Ph.D. dissertation.
Have you found any more? Let me know. The knowledge in the field continues to evolve.
This is supposedly a simple question to answer; an antenna is a device that emits radio waves. However, it is easy to get confused when comparing wireless communication systems with different number of transmit antennas, because these systems might use antennas with different physical sizes and properties. In fact, you can seldom find fair comparisons between contemporary single-antenna systems and Massive MIMO in the research literature.
Each antenna type has a predefined radiation pattern, which describes its inherent directivity; that is, how the gain of the emitted signal differs in different angular directions. An ideal isotropic antenna has no directivity, but a practical antenna always has a certain directivity, measured in dBi. For example, a half-wavelength dipole antenna has 2.15 dBi, which means that there is one angular direction in which the emitted signal is 2.15 dB stronger than it would be with a corresponding isotropic antenna. On the other hand, there are other angular directions in which the emitted signal is weaker. This is not a problem as long as there will not be any receivers in those directions.
In cellular communications, we are used to deploying large vertical antenna panels that cover a 120 degree horizontal sector and have a strong directivity of 15 dBi or more. Such a panel is made up of many small radiating elements, each having a directivity of a few dBi. By feeding them with the same input signal, a higher dBi is achieved for the panel. For example, if the panel consists of 8 patch antenna elements, each having 7 dBi, then you get a 7+10·log10(8) = 16 dBi antenna.
The picture above shows a real LTE site that I found in Nanjing, China, a couple of years ago. Looking at it from above, the site is structured as illustrated to the right. The site consists of three sectors, each containing a base station with four vertical panels. If you would look inside one of the panels, you will (probably) find 8 cross-polarized vertically stacked radiating elements, as illustrated in Figure 1. There are two RF input signals per panel, one per polarization, thus each panel acts as two antennas. This is how LTE with 8TX-sectors is deployed: 4 panels with dual polarization per base station.
At the exemplified LTE site, there is a total of 8·8·3 =192 radiating elements, but only 8·3 = 24 antennas. This disparity can lead to a lot of confusion. The Massive MIMO version of the exemplified LTE site may have the same form factor, but instead of 24 antennas with 16 dBi, you would have 192 antennas with 7 dBi. More precisely, you would connect each of the existing radiating elements to a separate RF input signal to create a larger number of antennas. Therefore, I suggest to use the following antenna definition from the book Massive MIMO Networks:
Definition: An antenna consists of one or more radiating elements (e.g., dipoles) which are fed by the same RF signal. An antenna array is composed of multiple antennas with individual RF chains.
Note that, with this definition, an array that uses analog beamforming (e.g., a phased array) only constitutes one antenna. It is usually called an adaptive antenna since the radiation pattern can be changed over time, but it is nevertheless a single antenna. Massive MIMO for sub-6 GHz frequencies is all about adding RF chains (also known as antenna ports), while not necessarily adding more radiating elements than in a contemporary system.
What is the purpose of having more RF chains?
With more RF chains, you have more degrees of freedom to modify the radiation pattern of the transmitted signal based on where the receiver is located. When transmitting a precoded signal to a single user, you adjust the phases of the RF input signals to make them all combine constructively at the intended receiver.
The maximum antenna/array gain is the same when using one 16 dBi antenna and when using 8 antennas with 7 dBi. In the first case, the radiation pattern is usually static and thus only a line-of-sight user located in the center of the cell sector will obtain this gain. However, if the antenna is adaptive (i.e., supports analog beamforming), the main lobe of the radiation pattern can be also steered towards line-of-sight users located in other angular directions. This feature might be sufficient for supporting the intended single-user use-cases of mm-wave technology (see Figure 4 in this paper).
In contrast, in the second case, we can adjust the radiation pattern by 8-antenna precoding to deliver the maximum gain to any user in the sector. This feature is particularly important for non-line-of-sight users (e.g., indoor use-cases), for which the signals from the different radiating elements will likely be received with “random” phase shifts and therefore add non-constructively, unless we compensate for the phases by digital precoding.
Note that most papers on Massive MIMO keep the antenna gain constant when comparing systems with different number of antennas. There is nothing wrong with doing that, but one cannot interpret the single-antenna case in such a study as a contemporary system.
Another, perhaps more important, feature of having multiple RF chains is that we can spatially multiplex several users when having multiple antennas. For this you need at least as many RF inputs as there are users. Each of them can get the full array gain and the digital precoding can be also used to avoid inter-user interference.
Last year, I wrote a post about channel hardening. To recap, the achievable data rate of a conventional single-antenna channel varies rapidly over time due to the random small-scale fading realizations, and also over frequency due to frequency-selective fading. However, when you have many antennas at the base station and use them for coherent precoding/combining, the fluctuations in data rate average out; we then say that the channel hardens. One follow-up question that I’ve got several times is:
Can we utilize the channel hardening to estimate the channels less frequently?
Unfortunately, the answer is no. Whenever you move approximately half a wavelength, the multi-path propagation will change each element of the channel vector. The time it takes to move such a distance is called a coherence time. This time is the same irrespectively of how many antennas the base station has and, therefore, you still need to estimate the channel once per coherence time. The same applies to the frequency domain, where the coherence bandwidth is determined by the propagation environment and not the number of antennas.
The following flow-chart shows what need to happen in every channel coherence time:
When you get a new realization (at the top of the flow-chart), you compute an estimate (e.g., based on uplink pilots), then you use the estimate to compute a new receive combining vector and transmit precoding vector. It is when you have applied these vectors to the channel that the hardening phenomena appears; that is, the randomness averages out. If you use maximum ratio (MR) processing, then the random realization h1 of the channel vector turns into an almost deterministic scalar channel ||h1||2. You can communicate over the hardened channel with gain ||h1||2 until the end of the coherence time. You then start over again by estimating the new channel realization h2, applying MR precoding/combining again, and then you get ||h2||2 ≈ ||h1||2.
In conclusion, channel hardening appears after coherent combining/precoding has been applied. To maintain a hardened channel over time (and frequency), you need to estimate and update the combining/precoding as often as you would do for a single-antenna channel. If you don’t do that, you will gradually lose the array gain until the point where the channel and the combining/precoding are practically uncorrelated, so there is no array gain left. Hence, there is more to lose from estimating channels too infrequently in Massive MIMO systems than in conventional systems. This is shown in Fig. 10 in a recent measurement paper from Lund University, where you see how the array gain vanishes with time. However, the Massive MIMO system will never be worse than the corresponding single-antenna system.
The signal-to-noise ratio (SNR) generally depends on the transmit power, channel gain, and noise power:
Since the spectral efficiency (bit/s/Hz) and many other performance metrics of interest depend on the SNR, and not the individual values of the three parameters, it is a common practice to normalize one or two of the parameters to unity. This habit makes it easier to interpret performance expressions, to select reasonable SNR ranges, and to avoid mistakes in analytical derivations.
There are, however, situations when the absolute value of the transmitted/received signal power matters, and not the relative value with respect to the noise power, as measured by the SNR. In these situations, it is easy to make mistakes if you use normalized parameters. I see this type of errors far too often, both as a reviewer and in published papers. I will give some specific examples below, but I won’t tell you who has made these mistakes, to not point the finger at anyone specifically.
Wireless energy transfer
Electromagnetic radiation can be used to transfer energy to wireless receivers. In such wireless energy transfer, it is the received signal energy that is harvested by the receiver, not the SNR. Since the noise power is extremely small, the SNR is (at least) a billion times larger than the received signal power. Hence, a normalization error can lead to crazy conclusions, such as being able to transfer energy at a rate of 1 W instead of 1 nW. The former is enough to keep a wireless transceiver on continuously, while the latter requires you to harvest energy for a long time period before you can turn the transceiver on for a brief moment.
The energy efficiency (EE) of a wireless transmission is measured in bit/Joule. The EE is computed as the ratio between the data rate (bit/s) and the power consumption (Watt=Joule/s). While the data rate depends on the SNR, the power consumption does not. The same SNR value can be achieved over a long propagation distance by using high transmit power or over a short distance by using a low transmit power. The EE will be widely different in these cases. If a “normalized transmit power” is used instead of the actual transmit power when computing the EE, one can get EEs that are one million times smaller than they should be. As a rule-of-thumb, if you compute things correctly, you will get EE numbers in the range of 10 kbit/Joule to 10 Mbit/Joule.
Noise power depends on the bandwidth
The noise power is proportional to the communication bandwidth. When working with a normalized noise power, it is easy to forget that a given SNR value only applies for one particular value of the bandwidth.
Some papers normalize the noise variance and channel gain, but then make the SNR equal to the unnormalized transmit power (measured in W). This may greatly overestimate the SNR, but the achievable rates might still be in the reasonable range if you operate the system in an interference-limited regime.
Some papers contain an alternative EE definition where the spectral efficiency (bit/s/Hz) is divided by the power consumption (Joule/s). This leads to the alternative EE unit bit/Joule/Hz. This definition is not formally wrong, but gives the misleading impression that one can multiply the EE value with any choice of bandwidth to get the desired number of bit/Joule. That is not the case since the SNR only holds for one particular value of the bandwidth.
Knowing when to normalize
In summary, even if it is convenient to normalize system parameters in wireless communications, you should only do it if you understand when normalization is possible and when it is not. Otherwise, you can make embarrassing mistakes, such as submitting a paper where the results are six orders of magnitude wrong. And, unfortunately, there are several such papers that have been published and these create a bad circle by tricking others into making the same mistakes.