Open access peer-reviewed chapter

Complexity Analysis in Channel Estimation Massive MIMO Compared with LMU and GRU

Written By

Hany Helmy, Sherif El Diasty and Hazem Shatila

Reviewed: 15 September 2023 Published: 10 November 2023

DOI: 10.5772/intechopen.113217

From the Annual Volume

Machine Learning and Data Mining Annual Volume 2023

Edited by Marco Antonio Aceves-Fernández

Chapter metrics overview

59 Chapter Downloads

View Full Metrics

Abstract

MIMO: Multiple-input multiple-output technology uses multiple antennas to use reflected signals to provide channel robustness and throughput gains. It is advantageous in several applications like cellular systems, users are distributed over a wide coverage area in various applications such as mobile systems, improving channel state information (CSI) processing efficiency in massive MIMO systems. This chapter proposes two channel-based deep learning methods gated recurrent unit and a Legendre memory unit to enhance the performance in a massive MIMO system and compares the complexity analysis to the previous methods, The complexity analysis is based on the channel state information network combined with gated recurrent units and Legendre memory units compared to indicator parameters which show the difference between literature-based techniques.

Keywords

  • massive MIMO
  • FDD
  • compressed sensing
  • deep learning
  • conventional neural network
  • complexity analysis
  • gated recurrent unit
  • Legendre memory unit

1. Introduction

Complexity analysis in deep learning is a fascinating topic indeed when delving into the realm of deep learning, we often encounter intricate models with numerous layers and countless parameters [1]. It is essential for such models to truly comprehend their inner workings. Now, complexity analysis in deep learning involves assessing the computational requirements and efficiency of these models. One commonly used metric is time complexity, which examines how the computational costs increase with an increase in the input size [2]. It allows us to estimate the computational resources needed for training and inference tasks.

Another aspect to consider is the space complexity, which examines the memory requirements of deep learning models. As the number of layers and parameters grows, so does the memory needed to store them [3]. Understanding this helps us optimize memory usage and select suitable hardware for our neural networks. However, it’s worth mentioning that deep learning is an ever-evolving field, and complexity analysis is just a small part of the puzzle.

There are always new challenges, surprises, and complexities waiting to be unraveled. It’s a journey that continues to intrigue and captivate scientists and enthusiasts alike. The history of complexity analysis in deep learning. It is indeed an intriguing subject. The path to understanding the complexity of deep learning has been paved with both triumphs and challenges. Back in the early days, when I first ventured into the realm of theoretical physics, we were only scratching the surface of what would later become known as deep learning [4]. The concept of neural networks has been around for quite some time, but it was the advent of powerful, computational capabilities that truly allowed us to explore their potential. As we delved deeper, we began to realize that the complexity of deep learning models was no trivial matter.

The number of parameters, the intricate interconnections, and the sheer depth of these networks made it a fascinating puzzle to unravel. The primary challenge was to develop methods for analyzing and understanding the complexity of these models. We needed tools that could help us comprehend the behavior of deep learning systems, predict their performance, and discern their limitations. Over time, researchers developed various approaches to complexity analysis in deep learning [5]. They ranged from straightforward measures like counting the number of parameters or layers, to more sophisticated techniques such as analyzing the network’s computational and memory requirements. Additionally, advancements were made in characterizing the computational complexity of training deep learning models. The notion of time and resource complexity became crucial in comprehending the training process and predicting how long it might take, given a certain dataset and architecture [6].

Advertisement

2. Complexity analysis

2.1 Performance and complexity trade-off of LMU and GRU

The Legendre Memory Unit (LMU) is mathematically derived to orthogonalize its continuous-time history – doing so by solving d coupled ordinary differential equations (ODEs), whose phase space linearly maps onto sliding windows of time via the Legendre polynomials up to degree d−1.

The Gated Recurrent Unit (GRU) is a type of Recurrent Neural Network (RNN) that, in certain cases, has advantages over long short-term memory (LSTM) (Figure 1). GRU uses less memory and is faster than LSTM, however, LSTM is more accurate when using datasets with longer sequences (Figure 2).

Figure 1.

The block diagram of (GRU) [7].

Figure 2.

Time-unrolled LMU layer [8].

The RNN layers for CSI compression and reconstruction of massive MIMO systems might have a remarkable number of parameters. The recurrent neural network module, for example, can dd up to 108 more parameters, raising storage and computation difficulties. To reduce the computational complexity and required memory, a fully connected layer-based autoencoder has been developed for channel state information feedback in time-varying channels [9].

While other research has investigated smaller recurrent neural networks RNNs, the least computationally expensive of these models still needs 107 parameters per snapshot. When the compression ratio is small, the networks also suffer from a significant drop in feedback performance because they must keep the same compression ratio in succeeding time slots and cannot obtain accurate previous information in the first time slot [10]. Despite the reported performance of “stacked” GRUs, the minimum depth of recurrent layers required for CSI recovery accuracy has yet to be determined [11]. Despite the substantial number of RNN parameters, the performance gain should be significant enough to justify the memory overhead.

The practical constraint of how often such feedback can be transmitted, as well as how CSI of fading channels will vary due to the Doppler effect, should be considered in DNN-based CSI feedback and recovery [12]. To reduce computational complexity and model size, we seek to systematically exploit channel characteristics such as forward CSI temporal coherence. We directly leverage known channel coherence temporally by developing a simple but effective Legendre memory unit combined with gated recurrent unit channel state information feedback LMU to improve CSI recovery accuracy.

Advertisement

3. Model size and computational complexity

3.1 The structure

Table 1 shows the number of parameters and floating-point operations (FLOPs) to show the impact of the model size reduction. This information compares the storage size and computational complexity of using the FC-layer along with the CNN-layer in the CsiNet compression module and the related decompression module.

Number of parametersFLOPs
FC-basedCNN-layerFC-basedCNN-layer
CR = 1/42.1 M14.4 K4.2 M0.9 M
CR = 1/161.1 M7.2 K2.1 M0.5 M
CR = 1/320.5 M3.7 K1.0 M0.2 M
CR = 1/640.2 M2.5 K0.5 M0.1 M

Table 1.

Number of parameters and FLOPs comparison for FC-based and CNN-based dimension compression and decompression module.

M, million; K, thousand.

The CNN-layer-based dimension compression and decompression module reduces the number of parameters by more than 100 times and the number of FLOPs by at least four times. The comparison results show that our CNN architecture for CSI compression and decompression is a significant step forward in expanding the spectrum of practical applications for deep learning-based channel state information encoding, feedback, and reconstruction in massive MIMO wireless systems.

Despite the higher cost, CsiNet-GRU and LMU-GRUs can reduce the number of parameters by order of magnitude and save over (5/6) FLOPs over CsiNet. Table 2 shows the complexity analysis of the proposed LMU-GRUs in comparison to various state-of-the-art CSI feedback systems, where the number of parameters and MACCs1 stand for space and temporal complexity, respectively [14]. When compared to CsiNet, our frameworks add fewer parameters while considerably improving CSI recovery quality, as seen in Table 2.

γCsiNet [12]Conv-LSTM-CsiNet [13]CsiNet-GRU [7]LMU-GRUs
Params1/42,103,90428,326,9042,107,6842,108,992
1/161,055,07222,296,3121,058,8521,060,160
1/32530,65619,477,624534,436535,744
1/64268,44818,117,432272,228273,536
MACCs1/44,366,336121,708,54412,424,44812,497,120
1/163,842,04897,591,29610,375,87211,448,544
1/323,579,90486,319,1049,851,54810,924,256
1/643,448,83280,879,6169,589,44010,662,112

Table 2.

The number of parameters and MACCS.

Convolution layers are mainly responsible for the increase in MACCs. When CR is low, convolutional layers require more computation than dense layers. LMU-GRUs have much lower model parameters and MACCs than CsiNet and Conv-LSTM-CsiNet, which improve reconstruction accuracy at the expense of huge space and temporal complexity due to the dense layers in GRU cells [15].

Each network’s average number of floating-point operations (FLOPs) connected with a single timeslot is also shown in Table 3. When compared to the Conv-LSTM-CsiNet in each compression ratio, CsiNet-GRU, and LMU-GRUs can save more than (8/9) and 9/10) FLOPs, respectively. Even at low compression ratios, the amount of computation for Conv-LSTM-CsiNet does not decrease considerably [16].

γCsiNet [12]Conv-LSTM CsiNet [13]CsiNet-GRU [7]LMU-GRUs
Params1/42.1 M2.1 M132.7 M2.1 M
1/161.0 M542.9 K118.5 M1.0 M
1/32530.6 K280.7 K116.1 M535.7 K
1/64268.4 K149.6 K115.0 M273.5 K
FLOPs1/4412.9 M44.5 M41.2 M39.2 M
1/16409.8 M41.3 M40.5 M39.3 M
1/32409.2 M40.8 M40.4 M40.1 M
1/64409.0 M40.5 M40.3 M40.0 M

Table 3.

Model size and computational complexity of tested networks.

M, million; K, thousand.

Without loss of performance, we show that latent convolutional layers require much fewer parameters than FC layers. Table 3 compares the model size and computational complexity of CsiNet, Conv-LSTM-CsiNet, CsiNet-GRU, and LMU-GRUs of a single timeslot, respectively.

LMU-GRUs utilize 60 times fewer parameters than Conv-LSTM-CsiNet across all compression ratios. More crucially, LMU-GRUs can achieve greater CSI recovery accuracy while using 1/3000 of the number of parameters required by Conv-LSTM-CsiNet. A 16-fold reduction in compression ratio (from 1/4 to 1/64), for example, only saves 1% of FLOPs. CsiNet-GRU and CsiNet-LMU-GRUs, on the other hand, require significantly less computing complexity in proportion to lower compression ratios.

A 16-fold CR reduction (from 1/4 to 1/64), for example, reduces the number of FLOPs in CsiNet-GRU and LMU-GRUs networks by 9 and 2%, respectively.

When using CsiNet-GRU and LMU-GRUs as a cooperative learning method at UE, we found that 50 percent more parameters and FLOPs are required than during the training phase. This is because the trained decoder must be repeated on the UE side to generate decoded CSI for the encoder’s previous time slot.

Table 3 summarizes the number of parameters and floating-point operations (FLOPs) to show the impact of the proposed model size reduction. This data compares the storage space and computational difficulty of using the FC layer against the suggested CNN layer in the CSI compression module and the related decompression module.

The proposed LMU-GRUs dimension compression and decompression module, as shown in Table 3, reduces the number of parameters by over 100 times and the number of FLOPs by at least 4 times [17]. The comparison results show that our new LMU-GRUs architecture for CSI compression and decompression is a significant step forward in expanding the range of practical applications for deep learning-based CSI encoding, feedback, and reconstruction in massive MIMO wireless systems.

Advertisement

4. Conclusion

This chapter compares a channel state information (CSI) feedback network by extending the DL-based CsiNet technique to incorporate GRUs and LMU over GRU layers. The proposed LMU-GRUs technique achieved the number of parameters and MACCS compared to other CS-based and CSI-based methods. The work is motivated by the active use of the recurrent convolutional neural network (RCNN) model in model size and computational complexity of tested networks using a channel state information network. The basic concept is to compare the complexity analysis in deep learning techniques using the COST 2100 model to get results related to indicators parameters in time-varying MIMO channels and acquire training samples.

Advertisement

5. Recommendations for future work

Quantum-assisted Deep Learning (QDL) is receiving significant attention towards enhancing various performance metrics of communication networks. The classical DL faces various challenges; where a substantial challenge is to figure out the training method for complex topologies of neural networks (NNs) (which are of similar complexity to that of the natural structure of the human brain).

Advertisement

Acknowledgments

First of all, I thank ALLAH for giving me the will to achieve this work.

It is a great honor for me to take this opportunity to express my deep gratitude to Dr. Sherif El Dyasti, Assistant Professor, Electronics and Communication Department, College of Engineering and Technology, Arab Academy for Science, Technology and Maritime Transport (AAST), for his excellent cooperation, his expert help, continuous encouragement and valuable effort for completion of this work.

My special thanks and appreciation to Prof. Hazem Shatila, Virginia Polytechnic Institute and State University, Professor of Artificial Intelligence & Markovdata, CEO, thanks for spending his precious time and for his continuous encouragement that was behind the completion of this work.

Advertisement

Conflict of interest

“The authors declare no conflict of interest.”

Advertisement

Appendices and nomenclature

AIartificial intelligence
AMPapproximate message-passing
AoAanalysis of alternatives
AEautoencoder
BERbit error rate
CEchannel estimation
CNNconvolutional neural network
CScompressive sensing
CRcompression ratio
CSIchannel state information
CsiNetchannel state information network
CsiNet-GRUchannel state information network-gated recurrent unit
DLdeep learning
DNNdeep neural network
FDDfrequency division duplex
GRUgated recurrent unit
LMULegendre memory unit
LASSOleast absolute shrinkage and selection operator
LSTMlong short-term memory
MIMOmultiple-input multiple-output
mmWavemillimeter wave
MSEmean squared error
NMSEnormalized mean square error
RELUrectified linear unit
RNNrecurrent neural network
SNRsignal-to-noise-ratio

References

  1. 1. Liu G, Jiang D. 5G: Vision and requirements for mobile communication system towards year 2020. Chinese Journal of Engineering. 2016;2016:1-8. DOI: 10.1155/2016/5974586
  2. 2. Rao X, Lau VKN. Distributed compressive CSIT estimation and feedback for FDD multi-user massive MIMO systems. IEEE Transactions on Signal Processing. 2014;62(12):3261-3271. DOI: 10.1109/TSP.2014.2324991
  3. 3. Liu L et al. The COST 2100 MIMO channel model. IEEE Wireless Communications. 2012;19(6):92-99. DOI: 10.1109/MWC.2012.6393523
  4. 4. Jiang Z, Chen S, Molisch AF, Vannithamby R, Zhou S, Niu Z. Exploiting wireless channel state information structures beyond linear correlations: A deep learning approach. IEEE Communications Magazine. 2019;57(3):28-34. DOI: 10.1109/MCOM.2019.1800581
  5. 5. Xie H, Gao F, Jin S. An overview of low-rank channel estimation for massive MIMO systems. IEEE Access. 2016;4:7313-7321. DOI: 10.1109/ACCESS.2016.2623772
  6. 6. Cayamcela MEM, Lim W. Artificial intelligence in 5G technology: A survey, 2018. In: 2018 International Conference on Information and Communication Technology Convergence (ICTC), Jeju, Korea (South). pp. 860-865. DOI: 10.1109/ICTC.2018.8539642
  7. 7. Helmy HMN, Daysti SE, Shatila H, Aboul-Dahab M. Performance enhancement of massive MIMO using deep learning-based channel estimation. IOP Conference Series: Materials Science and Engineering. 2021;1051(1):012029. DOI: 10.1088/1757-899X/1051/1/012029
  8. 8. Diógenes GK, De Sousa Vitória AR, Silva DFC, Pagotto DDP, Sousa RT, Filho ARG. Live births prediction using legendre memory unit: A case study for the health regions of Goiás. In: 2023 IEEE 36th International Symposium on Computer-Based Medical Systems (CBMS); L’Aquila, Italy. 2023. pp. 329-334. DOI: 10.1109/CBMS58004.2023.00239
  9. 9. Almamori A, Mohan S. Improved MMSE channel estimation in massive MIMO system with a method for the prediction of channel correlation matrix. In: 2018 IEEE 8th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA. 2018. pp. 670-672. DOI: 10.1109/CCWC.2018.8301699
  10. 10. Perken ET. Spares channel estimation with regularization methods in massive MIMO systems. In: International Foundation Telemetring Conference Proceedings. 2018
  11. 11. Donoho DL, Maleki A, Montanari A. Message passing algorithms for compressed sensing: I. motivation and construction. In: IEEE Information Theory Workshop on Information Theory (ITW 2010, Cairo), Cairo, Egypt. 2010. pp. 1-5. DOI: 10.1109/ITWKSPS.2010.5503193
  12. 12. Wen C-K, Shih W-T, Jin S. Deep learning for massive MIMO CSI feedback. IEEE Wireless Communications Letters. 2018;7(5):748-751. DOI: 10.1109/LWC.2018.2818160
  13. 13. Li X, Wu H. Spatio-temporal representation with deep neural recurrent network in MIMO CSI feedback. IEEE Wireless Communications Letters. 2020;9(5):653-657. DOI: 10.1109/LWC.2020.2964550
  14. 14. Dong P, Zhang H, Li GY. Machine learning prediction based CSI acquisition for FDD massive MIMO downlink. In: 2018 IEEE Global Communications Conference (GLOBECOM), Abu Dhabi, United Arab Emirates. 2018. pp. 1-6. DOI: 10.1109/GLOCOM.2018.8647328
  15. 15. Su X, Yu S, Zeng J, Kuang Y, Fang N, Li Z. Hierarchical codebook design for massive MIMO. In: 2013 8th International Conference on Communications and Networking in China (CHINACOM), Guilin, China. 2013. pp. 178-182. DOI: 10.1109/ChinaCom.2013.6694587
  16. 16. Schwarz S, Heath RW, Rupp M. Single-user MIMO versus multi-user MIMO in distributed antenna systems with limited feedback. EURASIP Journal of Advanced Signalling Process. 2013;2013:54. DOI: 10.1186/1687-6180-2013-54
  17. 17. Choi J, Chance Z, Love DJ, Madhow U. Noncoherent Trellis coded quantization: A practical limited feedback technique for massive MIMO systems. IEEE Transactions on Communications. 2013;61(12):5016-5029. DOI: 10.1109/TCOMM.2013.111413.130379

Written By

Hany Helmy, Sherif El Diasty and Hazem Shatila

Reviewed: 15 September 2023 Published: 10 November 2023