Interestingly, some radio resource allocation problems that appear to have exponential complexity can be relaxed to a form that is much easier to solve – this is what I call “relax and conquer”. In optimization theory, relaxation means that you widen the set of permissible solutions to the problem, which in this context means that the discrete optimization variables are replaced with continuous optimization variables. In many cases, it is easier to solve optimization problems with variables that take values in continuous sets than problems with a mix of continuous and discrete variables.
A basic example of this principle arises when communicating over a single-user MIMO channel. To maximize the achievable rate, you first need to select how many data streams to spatially multiplex and then determine the precoding and power allocation for these data streams. This appears to be a mixed-integer optimization problem, but Telatar showed in his seminal paper that it can be solved by the water-filling algorithm. More precisely, you relax the problem by assuming that the maximum number of data streams are transmitted and then you let the solution to a convex optimization problem determine how many of the data streams that are assigned non-zero power; this is the optimal number of data streams. Despite the relaxation, the global optimum to the original problem is obtained.
There are other, less known examples of the “relax and conquer” method. Some years ago, I came across the paper “Jointly optimal downlink beamforming and base station assignment“, which has received much less attention than it deserves. The UE-BS association problem, considered in this paper, is non-trivial since some BSs might have many more UEs in their vicinity than other BSs. Nevertheless, the paper shows that one can solve the problem by first relaxing it so that all BSs transmit to all the UEs. The author formulates a relaxed optimization problem where the beamforming vectors (including power allocation) are selected to satisfy each UEs’ SINR constraint, while minimizing the total transmit power. This problem is solved by convex optimization and, importantly, the optimal solution is always such that each UE only receives a non-zero signal power from one of the BSs. Hence, the seemingly difficult combinatorial UE-BS association problem is relaxed to a convex optimization problem, which provides the optimal solution to the original problem!
I have reused this idea in several papers. The first example is “Massive MIMO and Small Cells: Improving Energy Efficiency by Optimal Soft-cell Coordination“, which considers a similar setup but with a maximum transmit power per BS. The consequence of including this practical constraint is that it might happen that some UEs are served by multiple BSs at the optimal solution. These BSs send different messages to the UE, which decode them by successive interference cancelation, thus the solution is still practically achievable.
One practical weakness with the two aforementioned papers is that they take small-scale fading realizations into account in the optimization, thus the problem must be solved once per coherence interval, requiring extremely high computational power. More recently, in the paper “Joint Power Allocation and User Association Optimization for Massive MIMO Systems“, we applied the same “relax and conquer” method to Massive MIMO, but targeting lower bounds on the downlink ergodic capacity. Since the capacity bounds are valid as long as the channel statistics are fixed (and the same UEs are active), our optimized BS-UE association can be utilized for a relatively long time period. This makes the proposed algorithm practically relevant, in contrast to the prior works that are more of academic interest.
Another example of the “relax and conquer” method is found in the paper “Joint Pilot Design and Uplink Power Allocation in Multi-Cell Massive MIMO Systems”. We consider the assignment of orthogonal pilot sequences to users, which appears to be a combinatorial problem. Instead of assigning a pilot sequence to each UE and then allocate power, we relax the problem by allowing each user to design its own pilot sequence, which is a linear combination of the original orthogonal sequences. Hence, a pair of UEs might have partially overlapping sequences, instead of either identical or orthogonal sequences (as in the original problem). The relaxed problem even allows for pilot contamination within a cell. The sequences are then optimized to maximize the max-min performance. The resulting problem is non-convex, but the combinatorial structure has been relaxed so that there are only optimization variables from continuous sets. A local optimum to the joint pilot assignment and power control problem is found with polynomial complexity, using standard methods from the optimization literature. The optimization might not lead to a set of orthogonal pilot sequences, but the solution is practically implementable and gives better performance.
]]>In most cases, the receiver only has imperfect CSI and then it is harder to measure the performance. In fact, it took me years to understand this properly. To explain the complications, consider the uplink of a single-cell Massive MIMO system with single-antenna users and antennas at the base station. The received -dimensional signal is
where is the unit-power information signal from user , is the fading channel from this user, and is unit-power additive Gaussian noise. In general, the base station will only have access to an imperfect estimate of , for
Suppose the base station uses to select a receive combining vector for user . The base station then multiplies it with to form a scalar that is supposed to resemble the information signal :
From this expression, a common mistake is to directly say that the SINR is
which is obtained by computing the power of each of the terms (averaged over the signal and noise), and then claim that is an achievable rate (where the expectation is with respect to the random channels). You can find this type of arguments in many papers, without proof of the information-theoretic achievability of this rate value. Clearly, is an SINR, in the sense that the numerator contains the total signal power and the denominator contains the interference power plus noise power. However, this doesn’t mean that you can plug into “Shannon’s capacity formula” and get something sensible. This will only yield a correct result when the receiver has perfect CSI.
A basic (but non-conclusive) test of the correctness of a rate expression is to check that the receiver can compute the expression based on its available information (i.e., estimates of random variables and deterministic quantities). Any expression containing fails this basic test since you need to know the exact channel realizations to compute it, although the receiver only has access to the estimates.
What is the right approach?
Remember that the SINR is not important by itself, but we should start from the performance metric of interest and then we might eventually interpret a part of the expression as an effective SINR. In Massive MIMO, we are usually interested in the ergodic capacity. Since the exact capacity is unknown, we look for rigorous lower bounds on the capacity. There are several bounding techniques to choose between, whereof I will describe the two most common ones.
The first uplink bound can be applied when the channels are Gaussian distributed and are the MMSE estimates with the corresponding estimation error covariance matrices . The ergodic capacity of user is then lower bounded by
Note that this expression can be computed at the receiver using only the available channel estimates (and deterministic quantities). The ratio inside the logarithm can be interpreted as an effective SINR, in the sense that the rate is equivalent to that of a fading channel where the receiver has perfect CSI and an SNR equal to this effective SINR. A key difference from is that only the part of the desired signal that is received along the estimated channel appears in the numerator of the SINR, while the rest of the desired signal appears as in the denominator. This is the price to pay for having imperfect CSI at the receiver, according to this capacity bound, which has been used by Hoydis et al. and Ngo et al., among others.
The second uplink bound is
which can be applied for any channel fading distribution. This bound provides a value close to when there is substantial channel hardening in the system, while will greatly underestimate the capacity when varies a lot between channel realizations. The reason is that to obtain this bound, the receiver detects the signal as if it is received over a non-fading channel with gain (which is deterministic and thus known in theory, and easy to measure in practice), but there are no approximations involved so is always a valid bound.
Since all the terms in are deterministic, the receiver can clearly compute it using its available information. The main merit of is that the expectations in the numerator and denominator can sometimes be computed in closed form; for example, when using maximum-ratio and zero-forcing combining with i.i.d. Rayleigh fading channels or maximum-ratio combining with correlated Rayleigh fading. Two early works that used this bound are by Marzetta and by Jose et al..
The two uplink rate expressions can be proved using capacity bounding techniques that have been floating around in the literature for more than a decade; the main principle for computing capacity bounds for the case when the receiver has imperfect CSI is found in a paper by Medard from 2000. The first concise description of both bounds (including all the necessary conditions for using them) is found in Fundamentals of Massive MIMO. The expressions that are presented above can be found in Section 4 of the new book Massive MIMO Networks. In these two books, you can also find the right ways to compute rigorous lower bounds on the downlink capacity in Massive MIMO.
In conclusion, to avoid mistakes, always start with rigorously computing the performance metric of interest. If you are interested in the ergodic capacity, then you start from one of the canonical capacity bounds in the above-mentioned books and verify that all the required conditions are satisfied. Then you may interpret part of the expression as an SINR.
]]>The full title of my webinar is Massive MIMO for 5G below 6 GHz: Achieving Spectral Efficiency, Link Reliability, and Low-Power Operation. I will cover the basics of Massive MIMO and explain how the technology is not only great for enhancing the broadband access, but also for delivering the link reliability and low-power operation required by the internet of things. I have made sure that the overlap with my previous webinar is small.
If you watch the webinar live, you will have the chance to ask questions. Otherwise, you can view the recording of the webinar afterward. All the webinars in the IEEE 5G Webinar Series are available for anyone to view.
As a final note, I wrote a guest blog post at IEEE ComSoc Technology News in late December. It follows up and my previous blog post about GLOBECOM and is called: The Birth of 5G: What to do next?
]]>
While the previous blog post focused on Massive MIMO aspects of UAV communications, other theoretical research findings are reviewed in this tutorial by Walid Saad and Mehdi Bennis:
You can also check out this tutorial by Rui Zhang.
Furthermore, the team of the ERC Advanced PERFUME project, lead by Prof. David Gesbert, has recently demonstrated what appears to be the world’s first autonomous flying base station relays. This exciting achievement is demonstrated in the following video:
]]>I attended GLOBECOM in Singapore earlier this week. Since more and more preprints are posted online before conferences, one of the unique features of conferences is to meet other researchers and attend the invited talks and interactive panel discussions. This year I attended the panel “Massive MIMO – Challenges on the Path to Deployment”, which was organized by Ian Wong (National Instruments). The panelists were Amitava Ghosh (Nokia), Erik G. Larsson (Linköping University), Ali Yazdan (Facebook), Raghu Rao (Xilinx), and Shugong Xu (Shanghai University).
No common definition
The first discussion item was the definition of Massive MIMO. While everyone agreed that the main characteristic is that the number of controllable antenna elements is much larger than the number of spatially multiplexed users, the panelists put forward different additional requirements. The industry prefers to call everything with at least 32 antennas for Massive MIMO, irrespective of whether the beamforming is constructed from codebook-based feedback, grid-of-beams, or by exploiting uplink pilots and TDD reciprocity. This demonstrates that Massive MIMO is becoming a marketing term, rather than a well-defined technology. In contrast, academic researchers often have more restrictive definitions; Larsson suggested to specifically include the TDD reciprocity approach in the definition. This is because it is the robust and overhead-efficient way to acquire channel state information (CSI), particularly for non-line-of-sight users; see Myth 3 in our magazine paper. This narrow definition clearly rules out FDD operation, as pointed out by a member of the audience. Personally, I think that any multi-user MIMO implementation that provides performance similar to the TDD-reciprocity-based approach deserves the Massive MIMO branding, but we should not let marketing people use the name for any implementation just because it has many antennas.
Important use cases
The primary use cases for Massive MIMO in sub-6 GHz bands are to improve coverage and spectral efficiency, according to the panel. Great improvements in spectral efficiency have been demonstrated by prototyping, but the panelist agreed that these should be seen as upper bounds. We should not expect to see more than 4x improvements over LTE in the first deployments, according to Ghosh. Larger gains are expected in later releases, but there will continue to be a substantial gap between the average spectral efficiency observed in real cellular networks and the peak spectral efficiency demonstrated by prototypes. Since Massive MIMO achieves its main spectral efficiency gains by multiplexing of users, we might not need a full-blown Massive MIMO implementation today, when there are only one or two simultaneously active users in most cells. However, the networks need to evolve over time as the number of active users per cell grows.
In mmWave bands, the panel agreed that Massive MIMO is mainly for extending coverage. The first large-scale deployments of Massive MIMO will likely aim at delivering fixed wireless broadband access and this must be done in the mmWave bands; there is too little bandwidth in sub-6 GHz bands to deliver data rates that can compete with wired DSL technology.
Initial cost considerations
The deployment cost is a key factor that will limit the first generations of Massive MIMO networks. Despite all the theoretic research that has demonstrated that each antenna branch can be built using low-resolution hardware, when there are many antennas, one should not forget the higher out-of-band radiation that it can lead to. We need to comply with the spectral emission masks – spectrum is incredibly expensive so a licensee cannot accept interference from adjacent bands. For this reason, several panelists from the industry expressed the view that we need to use similar hardware components in Massive MIMO as in contemporary base stations and, therefore, the hardware cost grows linearly with the number of antennas. On the other hand, Larsson pointed out that the futuristic devices that you could see in James Bond movies 10 years ago can now be bought for $100 in any electronic store; hence, when the technology evolves and the economy of scale kicks in, the cost per antenna should not be more than in a smartphone.
A related debate is the one between analog and digital beamforming. Several panelists said that analog and hybrid approaches will be used to cut cost in the first deployments. To rely on analog technology is somewhat weird in an age when everything is becoming digital, but Yazdan pointed out that it is only a temporary solution. The long-term vision is to do fully digital beamforming, even in mmWave bands.
Another implementation challenge that was discussed is the acquisition of CSI for mobile users. This is often brought up as a showstopper since hybrid beamforming methods have such difficulties – it is like looking at a running person in a binocular and trying to follow the movement. This is a challenging issue for any radio technology, but if you rely on uplink pilots for CSI acquisition, it will not be harder than in a system of today. This has also been demonstrated by measurements.
Open problems
The panel was asked to describe the most important open problems in the Massive MIMO area, from a deployment perspective. One obvious issue, which we called the “grand question” in a previous paper, is to provide better support for Massive MIMO in FDD.
The control plane and MAC layer deserve more attention, according to Larsson. Basic functionalities such as ACK/NACK feedback is often ignored by academia, but incredibly important in practice.
The design of “cell-free” densely distributed Massive MIMO systems also deserve further attention. Connecting all existing antennas together to perform joint transmission seems to be the ultimate approach to wireless networks. Although there is no practical implementation yet, Yazdan stressed that deploying such networks might actually be more practical than it seems, given the growing interest in C-RAN technology.
10 years from now
I asked the panel what will be the status of Massive MIMO in 10 years from now. Rao predicted that we will have Massive MIMO everywhere, just as all access point supports small-scale MIMO today. Yazdan believed that the different radio technology (e.g., WiFi, LTE, NR) will converge into one interconnected system, which also allows operators to share hardware. Larsson thinks that over the next decade many more people will have understood the fundamental benefits of utilizing TDD and channel reciprocity, which will have a profound impact on the regulations and spectrum allocation.
]]>The book has now been published:
Emil Björnson, Jakob Hoydis and Luca Sanguinetti (2017), “Massive MIMO Networks: Spectral, Energy, and Hardware Efficiency”, Foundations and Trends® in Signal Processing: Vol. 11, No. 3-4, pp 154–655. DOI: 10.1561/2000000093.
What is new with this book?
Marzetta et al. published Fundamentals of Massive MIMO last year. It provides an excellent, accessible introduction to the topic. By considering spatially uncorrelated channels and two particular processing schemes (MR and ZF), the authors derive closed-form capacity bounds, which convey many practical insights and also allow for closed-form power control.
In the new book, we consider spatially correlated channels and demonstrate how such correlation (which always appears in practice) affects Massive MIMO networks. This modeling uncovers new fundamental behaviors that are important for practical system design. We go deep into the signal processing aspects by covering several types of channel estimators and deriving advanced receive combining and transmit precoding schemes.
In later chapters of the book, we cover the basics of energy efficiency, transceiver hardware impairments, and various practical aspects; for example, spatial resource allocation, channel modeling, and antenna array deployment.
The book is self-contained and written for graduate students, PhD students, and senior researchers that would like to learn Massive MIMO, either in depth or at an overview level. All the analytical proofs, and the basic results on which they build, are provided in the appendices.
On the website massivemimobook.com, you will find Matlab code that reproduces all the simulation figures in the book. You can also download exercises and other supplementary material.
Limited-time offer: Get a free copy of the book
Next week, we are giving a tutorial at the Globecom conference. In support of this, the publisher is currently providing free digital copies of the book on their website. This offer is available until December 7.
If you like the book, you can also buy a printed copy from the publisher’s website for the special price of $40! Use the discount code 552568, which is valid until December 31, 2017.
]]>Unfortunately, there was not enough time for me to answer all the questions that I received, so I had to answer many of them afterwards. I have gathered ten questions and my answers below. I can also announce that I will give another Massive MIMO webinar in January 2018 and it will also be followed by a Q/A session.
1. What are the differences between 4G and 5G that will affect how Massive MIMO can be implemented?
The channel estimation must be implemented in the right way (i.e., exploiting uplink pilots and channel reciprocity) to obtain sufficiently accurate channel state information (CSI) to perform spatial multiplexing of many users, otherwise the inter-user interference will eliminate most of the gains. Accurate CSI is hard to achieve within the 4G standard, although there are several Massive MIMO field trials for TDD LTE that show promising results. However, if 5G is designed properly, it will support Massive MIMO from scratch, while in 4G it will always be an add-on that must to adhere to the existing air interface.
2. How easy it is to deploy MIMO antennas on the current infrastructure?
Generally speaking, we can reuse the current infrastructure when deploying Massive MIMO, which is why operators show much interest in the technology. You upgrade the radio base stations but keep the same backhaul infrastructure and core network. However, since Massive MIMO supports much higher data rates, some of the backhaul connections might also need to be upgraded to deliver these rates.
3. What are the most suitable channel models for Massive MIMO?
I recommend the channel model that was developed in the MAMMOET project. It is a refinement of the COST 2100 model that takes particular phenomena of having large antenna arrays into account. Check out Deliverable D1.2 from that project.
4. For planar arrays, what is the height to width ratio that gives the highest performance?
You typically need more antennas in the horizontal direction (width) than in the vertical direction (height), because the angular variations between users is larger in the horizontal domain. For example, the array might cover a horizontal sector of 120-180 degrees, while the users’ elevation angles might only differ by a few tens of degrees. This is the reason that 8-antenna LTE base stations use linear arrays in the horizontal direction.
There is no optimal answer to the question. It depends on the deployment scenario. If you have high-rise buildings, users at different floors can have rather different elevation angles (it can differ up to 90 degrees) and you can benefit more from having many antennas in the vertical direction. If all users have almost the same elevation angle, it is preferable to have many antennas in the horizontal direction. These things are further discussed in Sections 7.3 and 7.4 in my new book.
5. What are the difficulties we face in deploying Massive MIMO in FDD systems?
The difficulty is to acquire channel state information at the base station for the frequency band used in the downlink, since it is very resource-demanding to send downlink pilots from a large array; particularly, if you want to spatially multiplex many users. This is an important but challenging problem that researchers have been working on since the 1990s. You can read more about it in Myth 3 and the grand question in the paper Massive MIMO: ten myths and one grand question.
6. Do you believe that there is a value in coordinated resource allocation schemes for Massive MIMO?
Yes, but the resource allocation in Massive MIMO is different from conventional systems. Scheduling might not be so important, since you can multiplex many users spatially, but pilot assignment and power allocation are important aspects that must be addressed. I refer to these things as spatial resource allocation. You can read more about this in Sections 7.1 and 7.2 in my new book, but as you can see from those sections, there are many open problems to be solved.
7. What is channel hardening and what implications does it have on the frequency allocation (in OFDMA networks, for example)?
Channel hardening means that the effective channel after beamforming is almost constant so that the communication link behaves as if there is no small-scale fading. A consequence is that all frequency subcarriers provide almost the same channel quality to a user. Regarding channel assignment, since you can multiplex many tens of users spatially in Massive MIMO, you can assign the entire bandwidth (all subcarriers) to every user; there is no need to use OFDMA to allocate orthogonal frequency resources to the users.
8. Is it practical to estimate the channel for each subcarrier in an OFDM system?
To limit the pilot overhead, you typically place pilots only on a small subset of the subcarriers. The distance between the pilots in the frequency domain can be selected based on how frequency-selective the channels are; if a user has L strong channel taps, it is sufficient to send pilots on L subcarriers, even if you many more subcarriers than that. Based on the received pilot signals, one can either estimate the channels on every subcarrier or estimate the channels on some of them and interpolate to get estimates on the remaining subcarriers.
9. How sensitive are the Massive MIMO spectral efficiency gains to TDD frame synchronization?
If you consider an OFDM system, then timing synchronization mismatches that are smaller than the cyclic prefix can basically be ignored. This is the case in TDD LTE systems and will not change when considering Massive MIMO systems that are implemented using OFDM. However, the synchronization across cells will not be perfect. The implications are investigated in a recent paper.
10. How does the higher computational complexity and delay in Massive MIMO processing affect the system performance?
I used to think that the computational complexity would be a bottleneck, but it turns out that it is not a big deal since all of the operations are standard (i.e., matrix multiplications and matrix inversions). For example, the circuit that was developed at Lund University shows that MIMO detection and precoding for a 20 MHz channel can be implemented very efficiently and only consumes a few mW.
]]>Other Massive MIMO videos can be found on our Youtube channel.
]]>This article is an interview with Prof. Liesbet Van der Perre who was the scientific leader of the project.
In 2012, when you began to draft the project proposal, Massive MIMO was not a popular topic. Why did you initiate the work?
– Theoretically and conceptually it seemed so interesting that it would be a pity not to work on it. The main goal of the MAMMOET project was to make conceptual progress towards a spectrally and energy efficient system and to raise the confidence level by demonstrating a practical hardware implementation. We also wanted to make channel measurements to see if they would confirm what has been seen in theory.
It seems the project partners had a clear vision from the beginning?
– It was actually very easy to write this proposal because everyone was on the same wavelength and knew what we wanted to achieve. We were all eager to start the project and learn from each other. This is quite unique and explains why the project delivered much more than promised. The fact that the team got along very well has also laid the fundament for further research collaborations.
What were the main outcomes of the project?
– We learned a lot on how things change when going from small to large arrays. New channel models are required to capture the new behaviors. We are used to that high-precision hardware is needed, but all the sudden this is not true when drastically increasing the number of antennas. You can then use low-resolution hardware and simple processing, which is very different from conventional MIMO implementation.
Some of the big conceptual differences in massive MIMO turned out to be easier to solve than expected, while some things were more problematic than foreseen. For example, it is difficult to connect all the signals together. You need to do part of the processing distributive to avoid this problem. Synchronization also turned out to be a bottleneck. If we would have known that from the start, we could have designed the testbed differently, but we thought that the channel estimation and MIMO processing would be the challenging part.
What was the most rewarding aspect of leading this project?
– The cross-fertilization of people was unique. We brought people with different background and expertise together in a room to identify the crucial problems in massive MIMO and find new solutions. For example, we realized early that interference will be a main problem and that zero-forcing processing is needed, although matched filtering was popular at the time. By carefully analyzing the zero-forcing complexity, we could show that it was almost negligible compared to other necessary processing and we later demonstrated zero-forcing in real-time at the testbed. This was surprising for many people who thought that massive MIMO would be impossible to implement since 8×8 MIMO systems are terribly complex, but many things can be simplified in massive MIMO. Looking back, it might seem that the outcomes were obvious, but these are things you don’t know until you have gone through the process.
What are the big challenges that remains?
– An important challenge is how to integrate massive MIMO into a network. We assumed that there are many users and we can all give them the same time-frequency resources, but the channels and traffic are not always suitable for that. How should we decide which users to put together? We used an LTE-like frame structure, but it is important to design a frame structure that is well-suited for massive MIMO and real traffic.
There are many tradeoffs and degrees-of-freedom when designing massive MIMO systems. Would you use the technology to provide very good cell coverage or to boost small-cell capacity? Instead of delivering fiber to homes, we could use massive MIMO with very many antennas for spatial multiplexing of fixed wireless connections. Alternatively, in a mobile situation, we might not multiplex so many users. Optimizing massive MIMO for different scenarios is something that remains.
We made a lot of progress on the digital processing side in MAMMOET, while on the analog side we mainly came up with the specifications. We also did not work on the antenna design since, theoretically, it does not matter which antennas you use, but in practice it does.
All the deliverables and publications in the MAMMOET project can be accessed online: https://mammoet-project.eu
The deliverables contain a lot information related to use cases, requirements, channel modeling, signal processing algorithms, algorithmic implementation, and hardware implementation. Some of the results can found in the research literature, but far from everything.
Note: The author of this article worked in the MAMMOET project, but did not take part in the drafting of the proposal.
]]>I have been think that it can go either way – it is in the hands of marketing people. Advanced Wifi routers have been marketed with MIMO functionality for some years, but the impact is limited since most people get their routers as part of their internet subscriptions instead of buying them separately. Hence, the main question is: will handset manufactures and telecom operators start using the MIMO term when marketing products to end customers?
Maybe we have the answer because Sprint, an American telecom operator, is currently marketing their 2018 deployment of new LTE technology by talking publicly about “Massive MIMO”. As I wrote back in March, Sprint and Ericsson were to conduct field tests in the second half of 2017. Results from the tests conducted in Seattle, Washington and Plano, Texas, have now been described in a press release. The tests were carried at a carrier frequency in the 2.5 GHz band using TDD mode and an Ericsson base station with 64 transmit/receive antennas. It is fair to call this Massive MIMO, although 64 antennas is in the lower end of the interval that I would call “massive”.
The press release describes “peak speeds of more than 300 Mbps using a single 20 MHz channel”, which corresponds to a spectral efficiency of 15 bit/s/Hz. That is certainly higher than you can get in legacy LTE networks, but it is less than some previous field tests.
Hence, when the Sprint COO of Technology, Guenther Ottendorfer, describes their Massive MIMO deployment with the words “You ain’t seen nothing yet”, I hope that this means that we will see network deployments with substantially higher spectral efficiencies than 15 bit/s/Hz in the years to come.
Several videos about the field test in Seattle have recently appeared. The first one demonstrates that 100 people can simultaneously download a video, which is not possible in legacy networks. Since the base station has 64 antennas, the 100 users are probably served by a combination of spatial multiplexing and conventional orthogonal time-frequency multiplexing.
The second video provides some more technical details about the setup used in the field test.
]]>In November, the upcoming Massive MIMO webinars are:
Massive MIMO for 5G: How Big Can it Get? by Emil Björnson (Linköping University), Thursday, 9 November 2017, 3:00 PM EST, 12:00 PM PST, 20:00 GMT.
Real-time Prototyping of Massive MIMO: From Theory to Reality by Douglas Kim (NI) and Fredrik Tufvesson (Lund University), Wednesday, 15 November 2017, 12:00 PM EST, 9:00 AM PST, 17:00 GMT.
]]>I sometimes get the question “Isn’t Massive MIMO just MU-MIMO with more antennas?” My answer is no, because the key benefit of Massive MIMO over conventional MU-MIMO is not only about the number of antennas. Marzetta’s Massive MIMO concept is the way to deliver the theoretical gains of MU-MIMO under practical circumstances. To achieve this goal, we need to acquire accurate channel state information, which in general can only be done by exploiting uplink pilots and channel reciprocity in TDD mode. Thanks to the channel hardening and favorable propagation phenomena, one can also simplify the system operation in Massive MIMO.
Six key differences between conventional MU-MIMO and Massive MIMO are provided below.
Conventional MU-MIMO | Massive MIMO | |
Relation between number of BS antennas (M) and users (K) | M ≈ K and both are small (e.g., below 10) | M ≫ K and both can be large (e.g., M=100 and K=20). |
Duplexing mode | Designed to work with both TDD and FDD operation | Designed for TDD operation to exploit channel reciprocity |
Channel acquisition | Mainly based on codebooks with set of predefined angular beams | Based on sending uplink pilots and exploiting channel reciprocity |
Link quality after precoding/combining | Varies over time and frequency, due to frequency-selective and small-scale fading | Almost no variations over time and frequency, thanks to channel hardening |
Resource allocation | The allocation must change rapidly to account for channel quality variations | The allocation can be planned in advance since the channel quality varies slowly |
Cell-edge performance | Only good if the BSs cooperate | Cell-edge SNR increases proportionally to the number of antennas, without causing more inter-cell interference |
Footnote: TDD stands for time-division duplex and FDD stands for frequency-division duplex.
]]>One answer is that beamforming and precoding are two words for exactly the same thing, namely to use an antenna array to transmit one or multiple spatially directive signals.
Another answer is that beamforming can be divided into two categories: analog and digital beamforming. In the former category, the same signal is fed to each antenna and then analog phase-shifters are used to steer the signal emitted by the array. This is what a phased array would do. In the latter category, different signals are designed for each antenna in the digital domain. This allows for greater flexibility since one can assign different powers and phases to different antennas and also to different parts of the frequency bands (e.g., subcarriers). This makes digital beamforming particularly desirable for spatial multiplexing, where we want to transmit a superposition of signals, each with a separate directivity. It is also beneficial when having a wide bandwidth because with fixed phases the signal will get a different directivity in different parts of the band. The second answer to the question is that precoding is equivalent to digital beamforming. Some people only mean analog beamforming when they say beamforming, while others use the terminology for both categories.
A third answer is that beamforming refers to a single-user transmission with one data stream, such that the transmitted signal consists of one main-lobe and some undesired side-lobes. In contrast, precoding refers to the superposition of multiple beams for spatial multiplexing of several data streams.
A fourth answer is that beamforming refers to the formation of a beam in a particular angular direction, while precoding refers to any type of transmission from an antenna array. This definition essentially limits the use of beamforming to line-of-sight (LoS) communications, because when transmitting to a non-line-of-sight (NLoS) user, the transmitted signal might not have a clear angular directivity. The emitted signal is instead matched to the multipath propagation so that the multipath components that reach the user add constructively.
A fifth answer is that precoding consists of two parts: choosing the directivity (beamforming) and choosing the transmit power (power allocation).
I used to use the word beamforming in its widest meaning (i.e., the first answer), as can be seen in my first book on the topic. However, I have since noticed that some people have a more narrow or specific interpretation of beamforming. Therefore, I nowadays prefer only talking about precoding. In Massive MIMO, I think that precoding is the right word to use since what I advocate is a fully digital implementation, where the phases and powers can be jointly designed to achieve high capacity through spatial multiplexing of many users, in both NLoS and LOS scenarios.
]]>The sub-6 GHz spectrum is particularly useful to provide network coverage, since the pathloss and channel coherence time are relatively favorable at such frequencies (recall that the coherence time is inversely proportional to the carrier frequency). Massive MIMO at sub-6 GHz spectrum can increase the efficiency of highly loaded cells, by upgrading the technology at existing base stations. In contrast, the huge available bandwidths in mmWave bands can be utilized for high-capacity services, but only over short distances due to the severe pathloss and high noise power (which is proportional to the bandwidth). Massive MIMO in mmWave bands can thus be used to improve the link budget.
Six key differences between sub-6 GHz and mmWave operation are provided below:
Sub-6 GHz | mmWave | |
Deployment scenario | Macro cells with support for high user mobility | Small cells with low user mobility |
Number of simultaneous users per cell | Up to tens of users, due to the large coverage area | One or a few users, due to the small coverage area |
Main benefit from having many antennas | Spatial multiplexing of tens of users, since the array gain and ability to separate users spatially lead to great spectral efficiency | Beamforming to a single user, which greatly improves the link budget and thereby extends coverage |
Channel characteristics | Rich multipath propagation | Only a few propagation paths |
Spectral efficiency and bandwidth | High spectral efficiency due to the spatial multiplexing, but small bandwidth | Low spectral efficiency due to few users, large pathloss, and large noise power, but large bandwidth |
Transceiver hardware | Fully digital transceiver implementations are feasible and have been prototyped | Hybrid analog-digital transceiver implementations are needed, at least in the first products |
Since Massive MIMO was initially proposed by Tom Marzetta for sub-6 GHz applications, I personally recommend to use the “Massive MIMO” name only for that use case. One can instead say “mmWave Massive MIMO” or just “mmWave” when referring to multi-antenna technologies for mmWave bands.
]]>Prof. Erik. G. Larsson gave a 2.5 hour tutorial on the fundamentals of Massive MIMO, which is highly recommended for anyone learning this topic. You can then follow up by reading his book with the same topic.
When you have viewed Erik’s introduction, you can learn more about the state-of-the-art signal processing schemes for Massive MIMO from another talk at the summer school. Dr. Emil Björnson gave a 3 hour tutorial on this topic:
]]>One option is to let the signal power become times larger than in a single-antenna reference scenario. The increase in SNR will then lead to higher data rates for the users. The gain can be anything from bit/s/Hz to almost negligible, depending on how interference-limited the system is. Another option is to utilize the array gain to reduce the transmit power, to maintain the same SNR as in the reference scenario. The corresponding power saving can be very helpful to improve the energy efficiency of the system.
In the uplink, with single-antenna user terminals, we can choose between these options. However, in the downlink, we might not have a choice. There are strict regulations on the permitted level of out-of-band radiation in practical systems. Since Massive MIMO uses downlink precoding, the transmitted signals from the base station have a stronger directivity than in the single-antenna reference scenario. The signal components that leak into the bands adjacent to the intended frequency band will then also be more directive.
For example, consider a line-of-sight scenario where the precoding creates an angular beam towards the intended user (as illustrated in the figure below). The out-of-band radiation will then get a similar angular directivity and lead to larger interference to systems operating in adjacent bands, if their receivers are close to the user (as the victim in the figure below). To counteract this effect, our only choice might be to reduce the downlink transmit power to keep the worst-case out-of-band radiation constant.
Another alternative is that the regulations are made more flexible with respect to precoded transmissions. The probability that a receiver in an adjacent band is hit by an interfering out-of-band beam, such that the interference becomes times larger than in the reference scenario, reduces with an increasing number of antennas since the beams are narrower. Hence, if one can allow for beamformed out-of-band interference if it occurs with sufficiently low probability, the array gain in Massive MIMO can still be utilized to increase the SNRs. A third option will then be to (partially) reduce the transmit power to also allow for relaxed linearity requirements of the hardware.
These considerations are nicely discussed in an overview article that appeared on ArXiv earlier this year. There are also two papers that analyze the impact of out-of-bound radiation in Massive MIMO: Paper 1 and Paper 2.
]]>First, four carefully selected articles are offered free of charge, see the screenshot below and click here for details.
More precisely, IEEE offers free access to the published versions of these articles, while the accepted versions were already openly available: Paper 1, Paper 2, Paper 3, and Paper 4.
Second, a live webinar entitled “5G Massive MIMO: Achieving Spectrum Efficiency” is organized by IEEE ComSoc on August 24. The speaker is Professor Liesbet Van der Perre from KU Leuven. She was the scientific leader of the MAMMOET project, which is famous for demonstrating that Massive MIMO works in practice. You can expect a unique mix of theoretical concepts and practical implementation insights from this webinar.
]]>Many researchers have analyzed pilot contamination over the six years that have passed since Marzetta uncovered its importance in Massive MIMO systems. We now have a quite good understanding of how to mitigate pilot contamination. There is a plethora of different approaches, whereof many have complementary benefits. If pilot contamination is not mitigated, it will both reduce the array gain and create coherent interference. Some approaches mitigate the pilot interference in the channel estimation phase, while some approaches combat the coherent interference caused by pilot contamination. In this post, I will try to categorize the approaches and point to some key references.
Interference-rejecting precoding and combining
Pilot contamination makes the estimate of a desired channel correlated with the channel from pilot-sharing users in other cells. When these channel estimates are used for receive combining or transmit precoding, coherent interference typically arise. This is particularly the case if maximum ratio processing is used, because it ignores the interference. If multi-cell MMSE processing is used instead, the coherent interference is rejected in the spatial domain. In particular, recent work from Björnson et al. (see also this related paper) have shown that there is no asymptotic rate limit when using this approach, if there is just a tiny amount of spatial correlation in the channels.
Data-aided channel estimation
Another approach is to “decontaminate” the channel estimates from pilot contamination, by using the pilot sequence and the uplink data for joint channel estimation. This have the potential of both improving the estimation quality (leading to a stronger desired signal) and reducing the coherent interference. Ideally, if the data is known, data-aided channel estimation increase the length of the pilot sequences to the length of the uplink transmission block. Since the data is unknown to the receiver, semi-blind estimation techniques are needed to obtain the channel estimates. Ngo et al. and Müller et al. did early works on pilot decontamination for Massive MIMO. Recent work has proved that one can fully decontaminate the estimates, as the length of the uplink block grows large, but it remains to find the most efficient semi-blind decontamination approach for practical block lengths.
Pilot assignment and dimensioning
Which subset of users that share a pilot sequence makes a large difference, since users with large pathloss differences and different spatial channel correlation cause less contamination to each other. Recall that higher estimation quality both increases the gain of the desired signal and reduces the coherent interference. Increasing the number of orthogonal pilot sequences is a straightforward way to decrease the contamination, since each pilot can be assigned to fewer users in the network. The price to pay is a larger pilot overhead, but it seems that a reuse factor of 3 or 4 is often suitable from a sum rate perspective in cellular networks. The joint spatial division and multiplexing (JSDM) provides a basic methodology to take spatial correlation into account in the pilot reuse patterns.
Alternatively, pilot sequences can be superimposed on the data sequences, which gives as many orthogonal pilot sequences as the length of the uplink block and thereby reduces the pilot contamination. This approach also removes the pilot overhead, but it comes at the cost of causing interference between pilot and data transmissions. It is therefore important to assign the right fraction of power to pilots and data. A hybrid pilot solution, where some users have superimposed pilots and some have conventional pilots, may bring the best of both worlds.
If two cells use the same subset of pilots, the exact pilot-user assignment can make a large difference. Cell-center users are generally less sensitive to pilot contamination than cell-edge users, but finding the best assignment is a hard combinatorial problem. There are heuristic algorithms that can be used and also an optimization framework that can be used to evaluate such algorithms.
Multi-cell cooperation
A combination of network MIMO and macro diversity can be utilized to turn the coherent interference into desired signals. This approach is called pilot contamination precoding by Ashikhmin et al. and can be applied in both uplink and downlink. In the uplink, the base stations receive different linear combinations of the user signals. After maximum ratio combining, the coefficients in the linear combinations approach deterministic numbers as the number of antennas grow large. These numbers are only non-zero for the pilot-sharing users. Since the macro diversity naturally creates different linear combinations, the base stations can jointly solve a linear system of equations to obtain the transmitted signals. In the downlink, all signals are sent from all base stations and are precoded in such a way that the coherent interference sent from different base stations cancel out. While this is a beautiful approach for mitigating the coherent interference, it relies heavily on channel hardening, favorable propagation, and i.i.d. Rayleigh fading. It remains to be shown if the approach can provide performance gains under more practical conditions.
]]>Have you reflected over what the purpose of asymptotic analysis is? The goal is not that we should design and deploy wireless networks with a nearly infinite number of antennas. Firstly, it is physically impossible to do that in a finite-sized world, irrespective of whether you let the array aperture grow or pack the antennas more densely. Secondly, the conventional channel models break down, since you will eventually receive more power than you transmitted. Thirdly, the technology will neither be cost nor energy efficient, since the cost/energy grows linearly with , while the delivered system performance either approaches a finite limit or grows logarithmically with .
It is important not to overemphasize the implications of asymptotic results. Consider the popular power-scaling law which says that one can use the array gain of Massive MIMO to reduce the transmit power as and still approach a non-zero asymptotic rate limit. This type of scaling law has been derived for many different scenarios in different papers. The practical implication is that you can reduce the transmit power as you add more antennas, but the asymptotic scaling law does not prescribe how much you should reduce the power when going from, say, 40 to 400 antennas. It all depends on which rates you want to deliver to your users.
The figure below shows the transmit power in a scenario where we start with 1 W for a single-antenna transmitter and then follow the asymptotic power-scaling law as the number of antennas increases. With antennas, the transmit power per antenna is just 1 mW, which is unnecessarily low given the fact that the circuits in the corresponding transceiver chain will consume much more power. By using higher transmit power than 1 mW per antenna, we can deliver higher rates to the users, while barely effecting the total power of the base station.
Similarly, there is a hardware-scaling law which says that one can increase the error vector magnitude (EVM) proportionally to and approach a non-zero asymptotic rate limit. The practical implication is that Massive MIMO systems can use simpler hardware components (that cause more distortion) than conventional systems, since there is a lower sensitivity to distortion. This is the foundation on which the recent works on low-bit ADC resolutions builds (see this paper and references therein).
Even the importance of the coherent interference, caused by pilot contamination, is easily overemphasized if one only considers the asymptotic behavior. For example, the finite rate limit that appears when communicating over i.i.d. Rayleigh fading channels with maximum ratio or zero-forcing processing is only approached in practice if one has around one million antennas.
In my opinion, the purpose of asymptotic analysis is not to understand the asymptotic behaviors themselves, but what the asymptotics can tell us about the performance at practical number of antennas. Here are some usages that I think are particularly sound:
Some form of Massive MIMO will appear in 5G, but to get a well-designed system we need to focus more on demonstrating and optimizing the performance in practical scenarios (e.g., the key 5G use cases) and less on pure asymptotic analysis.
]]>With i.i.d. Rayleigh fading, the channel gain has an Erlang-distribution (this is a scaled distribution) and the channel direction is uniformly distributed over the unit sphere in . The channel gain and the channel direction are also independent random variables, which is why this is a spatially uncorrelated channel model.
One of the key benefits of i.i.d. Rayleigh fading is that one can compute closed-form rate expressions, at least when using maximum ratio or zero-forcing processing; see Fundamentals of Massive MIMO for details. These expressions have an intuitive interpretation, but should be treated with care because practical channels are not spatially uncorrelated. Firstly, due to the propagation environment, the channel vector is more probable to point in some directions than in others. Secondly, the antennas have spatially dependent antenna patterns. Both factors contribute to the fact that spatial channel correlation always appears in practice.
One of the basic properties of spatial channel correlation is that the base station array receives different average signal power from different spatial directions. This is illustrated in Figure 1 below for a uniform linear array with 100 antennas, where the angle of arrival is measured from the boresight of the array.
As seen from Figure 1, with i.i.d. Rayleigh fading the average received power is equally large from all directions, while with spatially correlated fading it varies depending on in which direction the base station applies its receive beamforming. Note that this is a numerical example that was generated by letting the signal come from four scattering clusters located in different angular directions. Channel measurements from Lund University (see Figure 4 in this paper) show how the spatial correlation behaves in practical scenarios.
Correlated Rayleigh fading is a tractable way to model a spatially correlation channel vector: , where the covariance matrix is also the correlation matrix. It is only when is a scaled identity matrix that we have spatially uncorrelated fading. The eigenvalue distribution determines how strongly spatially correlated the channel is. If all eigenvalues are identical, then is a scaled identity matrix and there is no spatial correlation. If there are a few strong eigenvalues that contain most of the power, then there is very strong spatial correlation and the channel vector is very likely to be (approximately) spanned by the corresponding eigenvectors. This is illustrated in Figure 2 below, for the same scenario as in the previous figure. In the considered correlated fading case, there are 20 eigenvalues that are larger than in the i.i.d. fading case. These eigenvalues contain 94% of the power, while the next 20 eigenvalues contain 5% and the smallest 60 eigenvalues only contain 1%. Hence, most of the power is concentrated to a subspace of dimension . The fraction of strong eigenvalues is related to the fraction of the angular interval from which strong signals are received. This relation can be made explicit in special cases.
One example of spatially correlated fading is when the correlation matrix has equal diagonal elements and non-zero off-diagonal elements, which describe the correlation between the channel coefficients of different antennas. This is a reasonable model when deploying a compact base station array in tower. Another example is a diagonal correlation matrix with different diagonal elements. This is a reasonable model when deploying distributed antennas, as in the case of cell-free Massive MIMO.
Finally, a more general channel model is correlated Rician fading: , where the mean value represents the deterministic line-of-sight channel and the covariance matrix determines the properties of the fading. The correlation matrix can still be used to determine the spatial correlation of the received signal power. However, from a system performance perspective, the fraction between the power of the line-of-sight path and the scattered paths can have a large impact on the performance as well. A nearly deterministic channel with a large -factor provide more reliable communication, in particular since under correlated fading it is only the large eigenvalues of that contributes to the channel hardening (which otherwise provides reliability in Massive MIMO).
]]>The first step towards reproducibility is to describe the simulation procedure in such detail that another researcher can repeat the simulation, but a major effort is typically needed to reimplement everything. The second step is to make the simulation code publicly available, so that any scientist can review it and easily reproduce the results. While the first step is mandatory for publishing a scientific study, there is a movement towards open science that would make also the second step a common practice.
I understand that some researchers are skeptical towards sharing their simulation code, in fear of losing their competitive advantage towards other research groups. My personal principle is to not share any code until the research study is finished and the results have been accepted for publication in a full-length journal. After that, I think that the society benefits the most if other researcher can focus on improving my and others’ research, instead of spending excessive amount of time on reimplementing known algorithms. I also believe that the primary competitive advantage in research is the know-how and technical insights, while the simulation code is of secondary importance.
On my GitHub page, I have published Matlab code packages that reproduces the simulation results in one book, one book chapter, and more than 15 peer-reviewed articles. Most of these publications are related to MIMO or Massive MIMO. I see many benefits from doing this:
1) It increases the credibility of my research group’s work;
2) I write better code when I know that other people will read it;
3) Other researchers can dedicate their time into developing new improved algorithms and compare them with my baseline implementations;
4) Young scientists may learn how to implement a basic simulation environment by reading the code.
I hope that other Massive MIMO researchers will also make their simulation code publicly available. Maybe you have already done that? In that case, please feel free to write a comment to this post with a link to your code.
]]>The author argues that, under these circumstances, the MNOs have little to gain from investing in 5G technology. Most customers are not asking for any of the envisaged 5G services and will not be inclined to pay extra for them. Webb even compares the situation with the prisoner’s dilemma: the MNOs would benefit the most from not investing in 5G, but they will anyway make investments to avoid a situation where customers switch to a competitor that has invested in 5G. The picture that Webb paints of 5G is rather pessimistic compared to a recent McKinsey report, where the more cost-efficient network operation is described as a key reason for MNOs to invest in 5G.
The author provides a refreshing description of the market for cellular communications, which is important in a time when the research community focuses more on broad 5G visions than on the customers’ actual needs. The book is thus a recommended read for 5G researchers, since we should all ask ourselves if we are developing a technology that tackles the right unsolved problems.
Webb does not only criticize the economic incentives for 5G deployment, but also the 5G visions and technologies in general. The claims are in many cases reasonable; for example, Webb accurately points out that most of the 5G performance goals are overly optimistic and probably only required by a tiny fraction of the user base. He also accurately points out that some “5G applications” already have a wireless solution (e.g., indoor IoT devices connected over WiFi) or should preferably be wired (e.g., ultra-reliable low-latency applications such as remote surgery).
However, it is also in this part of the book that the argumentation sometimes falls short. For example, Webb extrapolates a recent drop in traffic growth to claim that the global traffic volume will reach a plateau in 2027. It is plausible that the traffic growth rate will reduce as a larger and larger fraction of the global population gets access to wireless high-speed connections. But one should bear in mind that we have witnessed an exponential growth in wireless communication traffic for the past century (known as Cooper’s law), so this trend can just as well continue for a few more decades, potentially at a lower growth rate than in the past decade.
Webb also provides a misleading description of multiuser MIMO by claiming that 1) the antenna arrays would be unreasonable large at cellular frequencies and 2) the beamforming requires complicated angular beam-steering. These are two of the myths that we dispelled in the paper “Massive MIMO: Ten myths and one grand question” last year. In fact, testbeds have demonstrated that massive multiuser MIMO is feasible in lower frequency bands, and particularly useful to improve the spectral efficiency through coherent beamforming and spatial multiplexing of users. Reciprocity-based beamforming is a solution for mobile and cell-edge users, for which angular beam-steering indeed is inefficient.
The book is not as pessimistic about the future as it might seem from this review. Webb provides an alternative vision for future wireless communications, where consistent connectivity rather than higher peak rates is the main focus. This coincides with one of the 5G performance goals (i.e., 50 Mbit/s everywhere), but Webb advocates an extensive government-supported deployment of WiFi instead of 5G technology. The use WiFi is not a bad idea; I personally consume relatively little cellular data since WiFi is available at home, at work, and at many public locations in Sweden. However, the cellular services are necessary to realize the dream of consistent connectivity, particularly outdoors and when in motion. This is where a 5G cellular technology that delivers better coverage and higher data rates at the cell edge is highly desirable. Reciprocity-based Massive MIMO seems to be the solution that can deliver this, thus Webb would have had a stronger case if this technology was properly integrated into his vision.
In summary, the combination of 5G Massive MIMO for wide-area coverage and WiFi for local-area coverage might be the way to truly deliver consistent connectivity.
]]>The article details the teaching principles and experiences that the teachers and students had from the 2015 edition of the CDIO-project. This was also described in a previous blog post. In the following video, the students describe and demonstrate the end-result of the 2016 edition of the project. The acoustic testbed is now truly massive, since 64 loudspeakers were used.
]]>The Bristol team has now worked with British Telecom and conducted trials at their site in Adastral Park, Suffolk, in more demanding user scenarios. In the indoor exhibition hall trial, 24 user streams were multiplexed over a 20 MHz bandwidth, resulting in a sum rate of 2 Gbit/s or a spectral efficiency of 100 bit/s/Hz/cell.
Several outdoor experiments were also conducted, which included user mobility. We are looking forward to see more details on these experiments, but in the meantime one can have a look at the following video:
Update: We have corrected the bandwidth number in this post.
]]>Huawei and Optus carried out an infield trial on February 26, where a sum rate of 655 Mbit/s was obtained over a 20 MHz channel by spatial multiplexing of 16 users. This corresponds to 33 bit/s/Hz or 2 bit/s/Hz/user, which are typical spectral efficiencies to expect from Massive MIMO. The base station was equipped with 128 antenna ports, but the press release provides no details on whether uplink or downlink transmission was considered.
ZTE demonstrated their TDD Massive MIMO solution, which we have described earlier on the blog. The company claimed to set a new record for single-site peak sum rate at their MWC demonstration. Spatial multiplexing of 16 data streams was considered with 256-QAM and the sum rate was 2.1 Gbit/s. Further details are found in their press release.
Nokia and Sprint demonstrated TDD-based Massive MIMO technology for LTE networks, using 64 antenna ports at the base station. Spatial multiplexing of eight commercial LTE terminals was considered. Communication theory predicts that the sum rate should grow proportionally to the number of terminals, which is consistent with the 8x improvement in uplink rates and 5x improvement in downlink rates that were reported. Further details are found in their press release or in the following video:
Ericsson and Sprint are also planning Massive MIMO tests in LTE TDD in the second half of 2017, according to another press release.
Did we miss any Massive MIMO related announcement from MWC? Please tell us in the comment field below!
]]>It is the physics that make it difficult to provide good coverage. The transmitted signals spread out and only a tiny fraction of the transmitted power reaches the receive antenna (e.g., one part of a billion parts). In cellular networks, the received signal power reduces roughly as the propagation distance to the power of four. This results in the following data rate coverage behavior:
This figure considers an area covered by nine base stations, which are located at the middle of the nine peaks. Users that are close to one of the base stations receive the maximum downlink data rate, which in this case is 60 Mbit/s (e.g., spectral efficiency 6 bit/s/Hz over a 10 MHz channel). As a user moves away from a base station, the data rate drops rapidly. At the cell edge, where the user is equally distant from multiple base stations, the rate is nearly zero in this simulation. This is because the received signal power is low as compared to the receiver noise.
What can be done to improve the coverage?
One possibility is to increase the transmit power. This is mathematically equivalent to densifying the network, so that the area covered by each base station is smaller. The figure below shows what happens if we use 100 times more transmit power:
There are some visible differences as compared to Figure 1. First, the region around the base station that gives 60 Mbit/s is larger. Second, the data rates at the cell edge are slightly improved, but there are still large variations within the area. However, it is no longer the noise that limits the cell-edge rates—it is the interference from other base stations.
The inter-cell interference remains even if we would further increase the transmit power. The reason is that the desired signal power as well as the interfering signal power grow in the same manner at the cell edge. Similar things happen if we densify the network by adding more base stations, as nicely explained in a recent paper by Andrews et al.
Ideally, we would like to increase only the power of the desired signals, while keeping the interference power fixed. This is what transmit precoding from a multi-antenna array can achieve; the transmitted signals from the multiple antennas at the base station add constructively only at the spatial location of the desired user. More precisely, the signal power is proportional to M (the number of antennas), while the interference power caused to other users is independent of M. The following figure shows the data rates when we go from 1 to 100 antennas:
Figure 3 shows that the data rates are increased for all users, but particularly for those at the cell edge. In this simulation, everyone is now guaranteed a minimum data rate of 30 Mbit/s, while 60 Mbit/s is delivered in a large fraction of the coverage area.
In practice, the propagation losses are not only distant-dependent, but also affected by other large-scale effects, such as shadowing. The properties described above remain nevertheless. Coherent precoding from a base station with many antennas can greatly improve the data rates for the cell edge users, since only the desired signal power (and not the interference power) is increased. Higher transmit power or smaller cells will only lead to an interference-limited regime where the cell-edge performance remains to be poor. A practical challenge with coherent precoding is that the base station needs to learn the user channels, but reciprocity-based Massive MIMO provides a scalable solution to that. That is why Massive MIMO is the key technology for delivering ubiquitous connectivity in 5G.
]]>There are no specific details of the experimental setup or implementation in any of these press releases, so we cannot tell how well the systems perform compared to a baseline TDD Massive MIMO setup. Maybe this is just a rebranding of the FDD multiuser MIMO functionality in LTE, evolved with a few extra antenna ports. It is nonetheless exciting to see that several major telecom companies want to associate themselves with the Massive MIMO technology and hopefully it will result in something revolutionary in the years to come.
Efficient FDD implementation of multiuser MIMO is a longstanding challenge. The reason is the difficulty in estimating channels and feeding back accurate channel state information (CSI) in a resource-efficient manner. Many researchers have proposed methods to exploit channel parameterizations, such as angles and spatial correlation, to simplify the CSI acquisition. This might be sufficient to achieve an array gain, but the ability to also mitigate interuser interference is less certain and remains to be demonstrated experimentally. Since 85% of the LTE networks use FDD, we have previously claimed that making Massive MIMO work well in FDD is critical for the practical success and adoption of the technology.
We hope to see more field trials of Massive MIMO in FDD, along with details of the measurement setups and evaluations of which channel acquisition schemes that are suitable in practice. Will FDD Massive MIMO be exclusive for static users, whose channels are easily estimated, or can anyone benefit from it in 5G?
Update: Blue Danube Systems has released a press release that is also describing trials of FDD Massive MIMO as well. Many companies apparently want to be “first” with this technology for LTE.
]]>To look into this, consider a communication system operating over a bandwidth of Hz. By assuming an additive white Gaussian noise channel, the capacity becomes
where W is the transmit power, is the channel gain, and W/Hz is the power spectral density of the noise. The term inside the logarithm is referred to as the signal-to-noise ratio (SNR).
Since the bandwidth appears in front of the logarithm, it might seem that the capacity grows linearly with the bandwidth. This is not the case since also the noise term in the SNR also grows linearly with the bandwidth. This fact is illustrated by Figure 1 below, where we consider a system that achieves an SNR of 0 dB at a reference bandwidth of 20 MHz. As we increase the bandwidth towards 2 GHz, the capacity grows only modestly. Despite the 100 times more bandwidth, the capacity only improves by , which is far from the that a linear increase would give.
The reason for this modest capacity growth is the fact that the SNR reduces inversely proportional to the bandwidth. One can show that
The convergence to this limit is seen in Figure 1 and is relatively fast since for .
To achieve a linear capacity growth, we need to keep the SNR fixed as the bandwidth increases. This can be achieved by increasing the transmit power proportionally to the bandwidth, which entails using more power when operating over a wider bandwidth. This might not be desirable in practice, at least not for battery-powered devices.
An alternative is to use beamforming to improve the channel gain. In a Massive MIMO system, the effective channel gain is , where is the number of antennas and is the gain of a single-antenna channel. Hence, we can increase the number of antennas proportionally to the bandwidth to keep the SNR fixed.
Figure 2 considers the same setup as in Figure 1, but now we also let either the transmit power or the number of antennas grow proportionally to the bandwidth. In both cases, we achieve a capacity that grows proportionally to the bandwidth, as we initially hoped for.
In conclusion, to make efficient use of more bandwidth we require more transmit power or more antennas at the transmitter and/or receiver. It is worth noting that these requirements are purely due to the increase in bandwidth. In addition, for any given bandwidth, the operation at millimeter-wave frequencies requires much more transmit power and/or more antennas (e.g., additional constant-gain antennas or one constant-aperture antenna) just to achieve the same SNR as in a system operating at conventional frequencies below 5 GHz.
]]>The Massive MIMO blog has talked with Javier Lorca Hernando at Telefónica to get further details. The trials were carried out at the Telefónica headquarters in Madrid. A base station with 128 antenna ports was deployed at the rooftop of one of their buildings and the users were located in one floor of the central building, approximately 100 m from the base station. The users basically had cell-edge conditions, due to the metallized glass and multiple metallic constructions surrounding them.
The uplink and downlink data transmissions were carried out in the 2.6 GHz band. Typical Massive MIMO time-division duplex (TDD) operation was considered, where the uplink detection and downlink precoding is based on uplink pilots and channel reciprocity. The existing LTE sounding reference signals (SRSs) were used as uplink pilots. The reciprocity-based precoding was implemented by using LTE’s transmission mode 8 (TM8), which supports any type of precoding. Downlink pilots were used for link adaptation and demodulation purposes.
It is great to see that Massive MIMO can be also implemented in LTE systems. In this trial, the users were static and relatively few, but it will be exciting to see if the existing LTE reference signals will also enable Massive MIMO communications for a multitude of mobile users!
Update: ZTE has carried out similar experiments in cooperation with Smartfren in Indonesia. Additional field trials are mentioned in the comments to this post.
]]>The diversity achieved by sending a signal over multiple channels with independent realizations is key to combating small-scale fading. Spatial diversity is particularly attractive, since it can be obtained by simply having multiple antennas at the transmitter or the receiver. Suppose the probability of a bad channel gain realization is p. If we have M antennas with independent channel gains, then the risk that all of them are bad is p^{M}. For example, with p=0.1, there is a 10% risk of getting a bad channel in a single-antenna system and a 0.000001% risk in an 8-antenna system. This shows that just a few antennas can be sufficient to greatly improve reliability.
In Massive MIMO systems, with a “massive” number of antennas at the base station, the spatial diversity also leads to something called “channel hardening”. This terminology was used already in a paper from 2004:
M. Hochwald, T. L. Marzetta, and V. Tarokh, “Multiple-antenna channel hardening and its implications for rate feedback and scheduling,” IEEE Transactions on Information Theory, vol. 50, no. 9, pp. 1893–1909, 2004.
In short, channel hardening means that a fading channel behaves as if it was a non-fading channel. The randomness is still there but its impact on the communication is negligible. In the 2004 paper, the hardening is measured by dividing the instantaneous supported data rate with the fading-averaged data rate. If the relative fluctuations are small, then the channel has hardened.
Since Massive MIMO systems contain random interference, it is usually the hardening of the channel that the desired signal propagates over that is studied. If the channel is described by a random M-dimensional vector h, then the ratio ||h||^{2}/E{||h||^{2}} between the instantaneous channel gain and its average is considered. If the fluctuations of the ratio are small, then there is channel hardening. With an independent Rayleigh fading channel, the variance of the ratio reduces with the number of antennas as 1/M. The intuition is that the channel fluctuations average out over the antennas. A detailed analysis is available in a recent paper.
The figure above shows how the variance of ||h||^{2}/E{||h||^{2}} decays with the number of antennas. The convergence towards zero is gradual and so is the channel hardening effect. I personally think that you need at least M=50 to truly benefit from channel hardening.
Channel hardening has several practical implications. One is the improved reliability of having a nearly deterministic channel, which results in lower latency. Another is the lack of scheduling diversity; that is, one cannot schedule users when their ||h||^{2} are unusually large, since the fluctuations are small. There is also little to gain from estimating the current realization of ||h||^{2}, since it is relatively close to its average value. This can alleviate the need for downlink pilots in Massive MIMO.
]]>The base station wants to know the channel responses of its user terminals and these are estimated in the uplink by sending pilot signals. Each pilot signal is corrupted by inter-cell interference and noise when received at the base station. For example, consider the scenario illustrated below where two terminals are transmitting simultaneously, so that the base station receives a superposition of their signals—that is, the desired pilot signal is contaminated.
When estimating the channel from the desired terminal, the base station cannot easily separate the signals from the two terminals. This has two key implications:
First, the interfering signal acts as colored noise that reduces the channel estimation accuracy.
Second, the base station unintentionally estimates a superposition of the channel from the desired terminal and from the interferer. Later, the desired terminal sends payload data and the base station wishes to coherently combine the received signal, using the channel estimate. It will then unintentionally and coherently combine part of the interfering signal as well. This is particularly poisonous when the base station has M antennas, since the array gain from the receive combining increases both the signal power and the interference power proportionally to M. Similarly, when the base station transmits a beamformed downlink signal towards its terminal, it will unintentionally direct some of the signal towards to interferer. This is illustrated below.
In the academic literature, pilot contamination is often studied under the assumption that the interfering terminal sends the same pilot signal as the desired terminal, but in practice any non-orthogonal interfering signal will cause the two effects described above.
]]>For millimeter wave, the huge bandwidth was identified as the key benefit. Rappaport predicted that 30 GHz of bandwidth would be available in 5 years time, while other panelists made a more conservative prediction of 15-20 GHz in 10 years time. With such a huge bandwidth, a spectral efficiency of 1 bit/s/Hz is sufficient for an access point to deliver tens of Gbit/s to a single user. The panelists agreed that much work remains on millimeter wave channel modeling and the design of circuits for that can deliver the theoretical performance without huge losses. The lack of robustness towards blockage and similar propagation phenomena is also a major challenge.
For Massive MIMO, the straightforward support of user mobility, multiplexing of many users, and wide-area coverage were mentioned as key benefits. A 10x-20x gain in per-cell spectral efficiency, with performance guarantees for every user, was another major factor. Since these gains come from spatial multiplexing of users, rather than increasing the spectral efficiency per user, a large number of users are required to achieve these gains in practice. With a small number of users, the Massive MIMO gains are modest, so it might not be a technology to deploy everywhere. Another drawback is the limited amount of spectrum in the range below 5 GHz, which limits the peak data rates that can be achieved per user. The technology can deliver tens of Mbit/s, but maybe not any Gbit/s per user.
Although the purpose of the panel was to debate the two 5G candidate technologies, I believe that the panelists agree that these technologies have complementary benefits. Today, you connect to WiFi when it is available and switch to cellular when the WiFi network cannot support you. Similarly, I imagine a future where you will enjoy the great data rates offered by millimeter wave, when you are covered by such an access point. Your device will then switch seamlessly to a Massive MIMO network, operating below 5 GHz, to guarantee ubiquitous connectivity when you are in motion or not covered by any millimeter wave access points.
]]>The METIS research project has identified twelve test cases for 5G connectivity. One of these is the “Dense urban information society”, which is
“…concerned with the connectivity required at any place and at any time by humans in dense urban environments. We here consider both the traffic between humans and the cloud, and also direct information exchange between humans or with their environment. The particular challenge lies in the fact that users expect the same quality of experience no matter whether they are at their workplace, enjoying leisure activities such as shopping, or being on the move on foot or in a vehicle.”
Source: METIS, deliverable D1.1 “Scenarios, requirements and KPIs for 5G mobile and wireless system”
Hence, the challenge is to provide ubiquitous connectivity in urban areas, where there will be massive user loads in the future: up to 200,000 devices per km^{2} is predicted by METIS. In their test case, each device requests one data packet per minute, which should be transferred within one second. Hence, there is on average up to 200,000/60 = 3,333 users active per km^{2} at any given time.
This large number of users is a challenge that Massive MIMO is particularly well-suited for. One of the key benefits of the Massive MIMO technology is the high spectral efficiency that it achieves by spatial multiplexing of tens of user per cell. Suppose, for example, that the cells are deployed in a hexagonal pattern with a base station in each cell center, as illustrated in the figure. How many simultaneously active users will there be per cell in the dense urban information society? That depends on the area of a cell. An inter-site distance (ISD) of 0.25 km is common in contemporary urban deployments. In this case, one can show that the area covered by each cell is √3×ISD^{2}/2 = 0.05 km^{2}.
The number of active users per cell is then obtained by multiplying the cell area with the user density. Three examples are provided in the table below:
10^{3} users/km^{2} | 10^{4} users/km^{2} | 10^{5} users/km^{2} | |
Total number of users per cell | 54 | 540 | 5400 |
Average active users per cell | 0.9 | 9 | 90 |
Recall that 1/60 of the total number of users are active simultaneously, in the urban information society test case. This gives the numbers in the second row of the table.
From this table, notice that there will be tens of simultaneously active users per cell, when the user density is above 10,000 per km^{2}. This is a number substantially smaller than the 200,000 per km^{2} predicted by the METIS project. Hence, there will likely be many future urban deployment scenarios with sufficiently many users to benefit from Massive MIMO.
A fraction of these users can (and probably will) be offloaded to WiFi-like networks, maybe operating at mmWave frequencies. But since local-area networks provide only patchy coverage, it is inevitable that many users and devices will rely on the cellular networks to achieve ubiquitous connectivity, with the uniform quality-of-service everywhere.
In summary, Massive MIMO is what we need to realize the dream of ubiquitous connectivity in the dense urban information society.
The multi-user MIMO concept, then called space-division multiple access (SDMA), was picked up by the industry in the nineties. For example, Ericsson made field-trials with antenna arrays in GSM systems, which were reported in “Adaptive antennas for GSM and TDMA systems” from 1999. ArrayComm filed an SDMA patent in 1991 and made trials in the nineties. In cooperation with the manufacturer Kyocera, this resulted in commercial deployment of SDMA as an overlay to the TDD-based Personal Handy-phone System (PHS).
Given this history, why isn’t multi-user MIMO a key ingredient in current cellular networks? I think there are several answers to this question:
Why is multi-user MIMO considered a key 5G technology? Basically because the three issues described above have now changed substantially. There is a renewed interest in TDD, with successful cellular deployments in Asia and WiFi being used everywhere. Massive MIMO is the refined form of multi-user MIMO, where the TDD operation enables channel estimation in any propagation environment, the many antennas allow for low-complexity signal processing, and the scalable protocols are suitable for large-scale deployments. The technology can nowadays be implemented using power-efficient off-the-shelf radio-frequency transceivers, as demonstrated by testbeds. Massive MIMO builds upon a solid ground of information theory, which shows how to communicate efficiently under practical impairments such as interference and imperfect channel knowledge.
Maybe most importantly, spatial multiplexing is needed to manage the future data traffic growth. This is because deploying many more base stations or obtaining much more spectrum are not viable options if we want to maintain network coverage—small cells at the street-level are easily shadowed by buildings and mm-wave frequency signals do not propagate well though walls. In 5G networks, a typical cellular base station might have tens of active users at a time, which is a sufficient number to benefit from the great spectral efficiency offered by Massive MIMO.
]]>My colleagues Erik G. Larsson and Hien Quoc Ngo have written a book entitled “Fundamentals of Massive MIMO” together with Thomas L. Marzetta and Hong Yang at Bell Labs, Nokia. The book is published this October/November by Cambridge University Press.
I have read the book and I think it serves as an excellent introduction to the topic. The text is suitable for graduate students, practicing engineers, professors, and doctoral students who would like to learn the basic Massive MIMO concept, results and properties. It also provides a clean introduction to the theoretical tools that are suitable for analyzing the Massive MIMO performance.
I personally intend to use this book as course material for a Master level course on Multiple-antenna communications next year. I recommend other teachers to also consider this possibility!
A preview of the book can be found on Google Books:
]]>With spectral efficiency, we usually mean the sum spectral efficiency of the transmissions in a cell of a cellular network. It is measured in bit/s/Hz. If you multiply it with the bandwidth, you will get the cell throughput measured in bit/s. Since the bandwidth is a scarce resource, particularly at the frequencies below 5 GHz that are suitable for network coverage, it is highly desirable to improve the cell throughput by increasing the spectral efficiency rather than increasing the bandwidth.
A great way to improve the spectral efficiency is to simultaneously serve many user terminals in the cell, over the same bandwidth, by means of space division multiple access. This is where Massive MIMO is king. There is no doubt that this technology can improve the spectral efficiency. The question is rather “how much?”
Earlier this year, the joint experimental effort by the universities in Bristol and Lund demonstrated an impressive spectral efficiency of 145.6 bit/s/Hz, over a 20 MHz bandwidth in the 3.5 GHz band. The experiment was carried out in a single-cell indoor environment. Their huge spectral efficiency can be compared with 3 bit/s/Hz, which is the IMT Advanced requirement for 4G. The remarkable Massive MIMO gain was achieved by spatial multiplexing of data signals to 22 users using 256-QAM. The raw spectral efficiency is 176 bit/s/Hz, but 17% was lost for practical reasons. You can read more about this measurement campaign here:
http://www.bristol.ac.uk/news/2016/may/5g-wireless-spectrum-efficiency.html
256-QAM is generally not an option in cellular networks, due to the inter-cell interference and unfavorable cell edge conditions. Numerical simulations can, however, predict the practically achievable spectral efficiency. The figure below shows the uplink spectral efficiency for a base station with 200 antennas that serves a varying number of users. Interference from many tiers of neighboring cells is considered. Zero-forcing detection, pilot-based channel estimation, and power control that gives every user 0 dB SNR are assumed. Different curves are shown for different values of τ_{c}, which is the number of symbols per channel coherence interval. The curves have several peaks, since the results are optimized over different pilot reuse factors.
From this simulation figure we observe that the spectral efficiency grows linearly with the number of users, for the first 30-40 users. For larger user numbers, the spectral efficiency saturates due to interference and limited channel coherence. The top value of each curve is in the range from 60 to 110 bit/s/Hz, which are remarkable improvements over the 3 bit/s/Hz of IMT Advanced.
In conclusion, 20x-40x improvements in spectral efficiency over IMT Advanced are what to expect from Massive MIMO.
]]>