Category Archives: Beyond 5G

Cell-Free Massive MIMO and Cloud-RAN

One question that I often receive from reviewers is: What is the difference between Cell-free Massive MIMO and the C-RAN technology? In this blog post, I will provide my usual answer to this question and then elaborate on why reviewers are not always satisfied with this answer.

In a nutshell, a base station consists of an antenna, a radio unit, and a baseband unit (or multiple antennas and radios, in the case of MIMO). These are traditionally co-located as follows: the antenna is deployed in a tower, the baseband unit is placed underneath, and the radio is located at one of these places.

Cloud Radio Access Network (C-RAN)

C-RAN is an alternative deployment architecture where the baseband units are deployed at other locations, “in the cloud” so to say. The C in C-RAN can also stand for “centralized”, which might be a preferable terminology since the word “cloud” is often associated with the use of general-purpose compute hardware. In C-RAN, the baseband processing tasks of many neighboring base stations are carried out in a single central processing unit (CPU), which might use specialized or general-purpose hardware. For latency reasons, the physical distance between a base station site and its CPU shouldn’t be more than a few kilometers. By pooling the processing resources in this way, it is possible to reduce the total hardware expenditure, particularly if some base stations often have low traffic load so the total hardware capability can be reduced by sharing. In other words, one can build a network that can handle high traffic anywhere, as long as it doesn’t happen everywhere at the same time. Another important benefit is that the neighboring base stations can more easily cooperate when their processing is anyway carried out at the same CPU.

In summary, C-RAN is a deployment architecture. Different air interfaces (e.g., 4G, 5G) can be implemented on top of the C-RAN architecture, making use of different physical layer techniques (e.g., Massive MIMO, coordinated multipoint).

Cell-free Massive MIMO

This is a new physical layer technology where the neighboring base stations (called access points in the related literature) are jointly serving the users in their vicinity. By carrying out coherent signal processing, the signal power can be boosted, the interference can be mitigated, and the cell boundaries are alleviated. We are essentially creating a wide-area network that is free from cells. There are different forms of Cell-Free Massive MIMO, characterized by where the baseband processing is carried out. It can either be fully centralized at the CPU or distributed as far as possible to pre-processing units located at each access point. The following video elaborates on these different implementation levels:

My answer

The simple answer to the question posed in the first paragraph is that C-RAN is a network architecture and Cell-free Massive MIMO is a physical-layer technology that can be deployed using either C-RAN or some other architecture. It is not a matter of selecting one or the other, but both can coexist and benefit from each other. My group is presenting a paper at ICC 2022 that exemplifies how to optimize the operation of Cell-free Massive MIMO when it is implemented in C-RAN.

The weak spot with my answer is that C-RAN was proposed already in 2011, based on the needs of the industry to consolidate their network assets, and a large amount of academic research has been carried out since then. Some of the papers on C-RAN have considered physical-layer techniques that resemble Cell-free Massive MIMO but without using that terminology. Some people might rightfully associate C-RAN with cell-free-like processing schemes, because they fit so well together. After all, Cell-free Massive MIMO is a revamp of Network MIMO that makes it more practical.

Making Cell-Free Massive MIMO Competitive

The paper “Making Cell-Free Massive MIMO Competitive with MMSE Processing and Centralized Implementation” that I’ve authored together with Luca Sanguinetti has been awarded the 2022 IEEE Marconi Prize Paper Award in Wireless Communications. This is a great honor that places the paper on the same list as many seminal papers published in the IEEE Transactions on Wireless Communications.

I will take this opportunity to elaborate on five things that I learned while writing the paper. The basic premise is that we analyze the uplink of a system with many distributed access points (APs) that serve a collection of user devices at the same time and frequency. We compared the data rates that can be achieved depending on how deeply the APs are collaborating, from Level 1 (cellular network with no cooperation) to Level 4 (cell-free network with centralized computations based on complete signal knowledge). We also compared maximum ratio (MR) processing of the received signals with local and centralized forms of minimum mean-squared error (MMSE) processing.

I learned the following five things:

  1. MMSE processing always outperforms MR processing. This might seem obvious, since the former scheme can suppress interference, but the really surprising thing was that the performance difference is large even for single-antenna APs that operate distributively. The reason is that MMSE processing provides much more channel hardening.
  2. Distributed MR processing is the worst-case scenario. Many of the early works on cell-free massive MIMO assumed distributed MR processing and focused on developing advanced power control algorithms. We demonstrated that one can achieve better performance with MMSE processing and rudimentary power control; thus, when designing a cell-free system, the choice of processing scheme is of primary importance, while the choice of power control is secondary.
  3. Linear uplink processing is nearly optimal. In a fully centralized implementation, it is possible to implement non-linear processing schemes for signal detection; in particular, successive interference cancellation could be used. We showed that this approach only increases the sum rate by a few percent, which isn’t enough to motivate the increased complexity. The reason is that we seldom have any strong interfering signals, just many weakly interfering signals.
  4. Distributed processing increases fronthaul signaling. Since the received signals are distributed over the APs, it might seem logical that one can reduce the fronthaul signaling by also doing parts of the processing distributively. This is not the case in the intended operating regime of cell-free massive MIMO, where each AP serves more or equally many users than it has antennas. In this case, fewer parameters need to be sent over the fronthaul when making a centralized implementation!
  5. Max-min fairness is a terrible performance goal. While a key motivation behind cell-free massive MIMO is to even out the performance variations in the system, compared to cellular networks, we shouldn’t strive for exact uniformity. To put it differently, the user with the worst channel conditions in the country shouldn’t dictate the performance of everyone else! Several early works on the topic focused on max-min fairness optimization and showed very promising simulation results, but when I attempted to reproduce these results, I noticed that they were obtained by terminating the optimization algorithms long before the max-min fairness solution was found. This indicates that we need a performance goal based on relative fairness (proportional fairness?) instead of the overly strict max-min fairness goal.

Since the paper was written in 2019, I have treated centralized MMSE processing as the golden standard for cell-free massive MIMO. I have continued looking for ways to reduce the fronthaul signaling while making use of distributed computational resources (that likely will be available in practice). I will mention two recent papers in this direction. The first is “MMSE-Optimal Sequential Processing for Cell-Free Massive MIMO With Radio Stripes“, which shows that one can implement centralized MMSE processing in a distributed/sequential manner, if the fronthaul is sequential. The paper “Team MMSE Precoding with Applications to Cell-free Massive MIMO” develops a methodology for dealing with the corresponding downlink problem, which is more challenging due to power and causality constraints.

Finally, let me thank IEEE ComSoc for not only giving us the Marconi Prize Paper Award but also producing the following nice video about the paper:

Episode 28: Ultra-Reliable Low-Latency Communication (With Petar Popovski)

We have now released the 28th episode of the podcast Wireless Future. It has the following abstract:

The reliability of an application is determined by its weakest link, which often is the wireless link. Channel coding and retransmissions are traditionally used to enhance reliability but at the cost of extra latency. 5G promises to enhance both reliability and latency in a new operational mode called ultra-reliable low-latency communication (URLLC). In this episode, Erik G. Larsson and Emil Björnson discuss URLLC with Petar Popovski, Professor at Aalborg University, Denmark. The conversation pinpoints the physical reasons for latency and unreliability, and viable solutions related to network deployment, diversity, digital vs. analog communications, non-orthogonal network slicing, and machine learning. Further details can be found in the article “Wireless Access in Ultra-Reliable Low-Latency Communication (URLLC)” and its companion video

You can watch the video podcast on YouTube:

You can listen to the audio-only podcast at the following places:

Episode 24: Q&A With 5G and 6G Predictions

We have now released the 24nd episode of the podcast Wireless Future, which is a New Year’s special! It has the following abstract:

In this episode, Emil Björnson and Erik G. Larsson answer ten questions from the listeners. The common theme is predictions of how 5G will evolve and which technologies will be important in 6G. The specific questions: Will Moore’s law or Edholm’s law break down first? How important will integrated communication and sensing become? When will private 5G networks start to appear? Will reconfigurable intelligent surfaces be a key enabler of 6G? How can we manage the computational complexity in large-aperture Massive MIMO? Will machine learning be the game-changer in 6G? What is 5G Dynamic Spectrum Sharing? What does the convergence of the Shannon and Maxwell theories imply? What happened to device-to-device communications, is it an upcoming 5G feature? Will full-duplex radios be adopted in the future? If you have a question or idea for a future topic, please share it as a comment to the YouTube version of this episode.

You can watch the video podcast on YouTube:

You can listen to the audio-only podcast at the following places:

Episode 23: Wireless Localization and Sensing (With Henk Wymeersch)

We have now released the 23nd episode of the podcast Wireless Future! It has the following abstract:

For each wireless generation, we are using more bandwidth and more antennas. While the primary reason is to increase the communication capacity, it also increases the network’s ability to localize objects and sense changes in the wireless environment. The localization and sensing applications impose entirely different requirements on the desired signal and channel properties than communications. To learn more about this, Emil Björnson and Erik G. Larsson have invited Henk Wymeersch, Professor at Chalmers University of Technology, Sweden. The conversation covers the fundamentals of wireless localization, the historical evolution, and future developments that might involve machine learning, terahertz bands, and reconfigurable intelligent surfaces. Further details can be found in the articles “Collaborative sensor network localization” and “Integration of communication and sensing in 6G”.

You can watch the video podcast on YouTube:

You can listen to the audio-only podcast at the following places:

Episode 22: Being Near or Far in Wireless

We have now released the 22nd episode of the podcast Wireless Future! It has the following abstract:

Wireless signals look different when observed near to versus far from the transmitter. The notions of near and far also depend on the physical size of the transmitter and receiver, as well as on the wavelength. In this episode, Erik G. Larsson and Emil Björnson discuss these fundamental phenomena and how they can be utilized when designing future communication systems. Concept such as near-field communications, finite-depth beamforming, mutual coupling, and new spatial multiplexing methods such as orbital angular momentum (OAM) are covered. To get more technical details, you can read the paper “A Primer on Near-Field Beamforming for Arrays and Reconfigurable Intelligent Surfaces”.

You can watch the video podcast on YouTube:

You can listen to the audio-only podcast at the following places:

Episode 21: Wireless Coverage Without Beamforming

We have now released the 21st episode of the podcast Wireless Future! It has the following abstract:

The latest wireless technologies rely heavily on beamformed data transmissions, implemented using antenna arrays. Since the signals are spatially directed towards the location of the receiver, the transmitter needs to know where to point the beam. Before the wireless link has been established, the transmitter will not have such knowledge. Hence, the geographical coverage of a network is determined by how we can transmit in the absence of beamforming gains. In this episode, Emil Björnson and Erik G. Larsson discuss how to achieve wide-area coverage in wireless networks without beamforming. The conversation covers deployment fundamentals, pathloss characteristics, beam sweeping, spatial diversity, and space-time codes. To learn more, you can read the textbook “Space-Time Block Coding for Wireless Communications”.

You can watch the video podcast on YouTube:

You can listen to the audio-only podcast at the following places: