The field of machine learning (ML) has a long and extremely successful history. For example, the idea of using neural networks (NN) for intelligent machines dates back to as early as 1942 when a simple one-layer model was used to simulate the status of a single neuron. ML has shown its overwhelming advantages in many areas, including computer vision, robotics, and natural language processing, where it is normally difficult to find a concrete mathematical model for feature representation. In those areas, ML has proved to be a powerful tool as it does not require a comprehensive specification of the model. Different from the aforementioned ML applications, the development of communications has vastly relied on theories and models, from information theory to channel modelling. These traditional approaches are showing serious limitations, especially in view of the increased complexity of communication networks. Therefore, research on ML applied to communications, especially to wireless communications, is currently experiencing an incredible boom.
This collection of Best Readings focuses on ML in the physical and medium access control (MAC) layer of communication networks. ML can be used to improve each individual (traditional) component of a communication system, or to jointly optimize the entire transmitter or receiver. Therefore, after introducing some popular textbooks, tutorials, and special issues in this collection, we divide the technical papers into the following six areas:
- Signal detection
- Channel encoding and decoding
- Channel estimation, prediction, and compression
- End-to-end communications
- Resource allocation
- Selected topics
Even if ML in communications is still in its infancy, we believe that a growing number of researchers will be dedicated to the related studies and ML will greatly change the way of communication system design in the near future.
Issued March 2019
Geoffrey Ye Li
Georgia Institute of Technology
Atlanta, Georgia, USA
Member Technical Staff
Nokia Bell Labs
Elisabeth de Carvalho
École Polytechnique Fédérale de Lausanne
Queen Mary University of London
Matthew C. Valenti
Editor-in-Chief, ComSoc Best Readings
West Virginia University
Morgantown, WV, USA
O. Simeone, A Brief Introduction to Machine Learning for Engineers, Foundations and Trends in Signal Processing, 12(3-4), 200-431, 2018.
Targeted specifically at engineers, this book provides a short introduction into key concepts and methods in machine learning (ML). Starting from first principles, it covers a wide range of topics, such as probabilistic models, supervised and unsupervised learning, graphical models, as well as approximate inference. Numerous reproducible numerical examples are provided to help understand the key ideas, while the well-selected and up-to-date list of references provides good entry points for readers willing to deepen their knowledge in a specific area. Overall, the book is an excellent starting point for engineers to familiarize themselves with the broad area of ML.
C. M. Bishop, Pattern Recognition and Machine Learning, Springer, 2006.
This book not only presents developments in the area of machine learning but also provides a comprehensive introduction to the field. No previous knowledge of pattern recognition or machine learning is assumed, and readers only need to be familiar with multivariate calculus, basic linear algebra, and basic probability theory. It is aimed at graduate students, researchers, and practitioners in the area of machine learning, statistics, computer science, and signal processing.
I. Goodfellow, Y. Bengio, and A. Courville, Deep Learning, MIT Press, 2016.
This is a book on deep learning from some of the pioneers of the field. The book starts with background notions on linear algebra and probability theory. The second part discusses a range of neural network architectures that are most commonly used to solve practical problems and gives guidelines on how to use these architectures through practical examples. Finally, in the third part of the book, the authors discuss a wide range of research-related topics in neural networks.
J. Watt, R. Borthani, and A. K. Katsaggelos, Machine Learning Refined: Foundation, Algorithms, and Applications, Cambridge University Press, 2016.
Written by experts in signal processing and communications, this book contains both a lucid explanation of mathematical foundations in machine learning (ML) as well as the practical real-world applications, such as natural language processing and computer vision. It is a perfect resource and an ideal reference for students and researchers. It is also a useful self-study guide for practitioners working in ML, computer science, and signal processing.
R. S. Sutton and A. G. Barto, Reinforcement Learning: An Introduction, MIT Press, 1998.
As an introductory book to reinforcement learning (RL), it is one of the main references in the field. It provides a clear and intuitive explanation of the core principles and algorithms in RL, with very useful examples. The recent reedition of the book (August 2018) includes the most recent developments in RL.
- Overviews and Tutorials
C. Jiang, H. Zhang, Y. Ren, Z. Han, K.-C. Chen, and L. Hanzo, “Machine Learning Paradigms for Next-Generation Wireless Networks,” IEEE Wireless Communications, vol. 24, no. 2, pp. 98-105, April 2017.
The article proposes to use machine learning (ML) paradigms to address challenges in the fifth generation (5G) wireless networks. The article first briefly reviews the rudimentary concepts of ML and introduces their compelling applications in 5G networks. With the help of ML, future smart 5G mobile terminals will autonomously access the most meritorious spectral bands with the aid of sophisticated spectral efficiency learning. The transmission protocols in 5G networks can be adaptively adjusted with the aid of quality of service learning/inference. The article assists the readers in refining the motivation, problem formulation, and methodology of powerful ML algorithms in the context of future wireless networks.
T. O'Shea and J. Hoydis, “An Introduction to Deep Learning for the Physical Layer,” IEEE Transactions on Cognitive Communications and Networking, vol. 3, no. 4, pp. 563-575, December 2017.
This paper presents the idea of learning full physical-layer implementations of communication systems with the help of neural network-based autoencoders. The technique is evaluated through simulations in several simple scenarios, such as the AWGN, Rayleigh fading, and two-user interference channels, where state-of-the-art performance is achieved. The paper discusses several situations when and reasons why deep learning can lead to gains with respect to classical model-based approaches, presents some examples showing how expert knowledge can be injected into the neural network architecture, and outlines a list of future research challenges.
L. Liang, H. Ye, and G. Y. Li, “Toward Intelligent Vehicular Networks: A Machine Learning Framework,” IEEE Internet of Things Journal, vol. 6, no. 1, pp. 124-15, February 2019.
This article provides an extensive overview on how to use machine learning to address the pressing challenges of high-mobility vehicular networks. Through learning the underlying dynamics of a vehicular network, better decisions can be made to optimize network performance. In particular, the article discusses employing reinforcement learning to manage the network resources as a promising alternative to prevalent optimization approaches.
M. Ibnkahla, “Applications of Neural Networks to Digital Communications - A Survey,” Elsevier Signal Processing, no. 80, pp. 1185-1215, July 2000.
This classical survey paper provides an excellent overview of research, mostly carried out in the 1990s, on various applications of neural networks to communication systems. It is a good link between past research and future trends in machine learning in communications.
K. Arulkumaran, M. P. Deisenroth, M. Brundage, and A. A. Bharath, “Deep Reinforcement Learning: A Brief Survey,” IEEE Signal Processing Magazine, vol. 34, no. 6, pp. 26-38, November 2017.
This survey first introduces the principles of deep reinforcement learning (RL) and then presents the main streams of value-based and policy-based methods. It covers all important algorithms in deep RL, including the deep Q-network, trust region policy optimization, and asynchronous advantage actor critic. At the end of the article, several current research areas in the field of deep RL are introduced.
- Special Issues
“Machine Learning for Cognition in Radio Communications and Radar,” IEEE Journal of Selected Topics in Signal Processing, vol. 12, no. 1, pp. 3-247, February 2018.
“Machine Learning and Data Analytics for Optical Communications and Networking,” IEEE/OSA Journal of Optical Communications and Networking, vol. 10, no. 10 October 2018.
- Topic: Signal Detection
H. Ye, G. Y. Li, and B.-H. Juang, “Power of Deep Learning for Channel Estimation and Signal Detection in OFDM Systems,” IEEE Wireless Communications Letters, vol. 7, no. 1, pp. 114-117, February 2018.
This paper proposes a deep learning based joint channel estimation and signal detection approach. A deep neural network is trained to recover the transmit data by feeding the received signals corresponding to transmit data and pilots. This method outperforms the minimum mean-squared error method for a system without adequate pilots or cyclic prefix and with nonlinear distortions.
N. Samuel, T. Diskin, and A. Wiesel, “Deep MIMO Detection,” in Proc. IEEE International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Sapporo, Japan, July 2017.
The paper uses deep learning for massive multi-input multi-output (MIMO) detection by unfolding a projected gradient descent method, and applies the approach to time-invariant and time-varying channels. The deep learning algorithm provides lower complexity than approximate message passing and semidefinite relaxation with the same accuracy and enhanced robustness.
N. Farsad and A. Goldsmith, “Neural Network Detection of Data Sequences in Communication Systems,” IEEE Transactions on Signal Processing, vol. 66, no. 21, pp. 5663-5678, November 2018.
This paper describes a bidirectional recurrent neural network for sequence detection in channels with memory. The method does not require knowledge of the channel model. Alternatively, if the channel model is known, it does not require knowledge of the channel state information (CSI). Simulation and experimental results show that the developed method works well and can outperform Viterbi detection in certain scenarios.
- Topic: Channel Encoding and Decoding
N. Farsad, M. Rao, and A. Goldsmith, “Deep Learning for Joint Source-Channel Coding of Text,” in Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, Canada, April 2018.
This paper addresses joint source and channel coding of structured data, such as natural language, over a noisy channel. The typical approach to this problem is optimal in terms of minimizing end-to-end distortion only when both the source and channel have arbitrarily long block lengths, which is not necessarily optimal for finite-length documents or encoders. This paper demonstrates that, in this scenario, a deep learning based encoder and decoder can achieve lower word-error rates.
T. Gruber, S. Cammerer, J. Hoydis, and S. ten Brink, “On Deep Learning-based Channel Decoding,” in Proc. Information Sciences and Systems (CISS), Baltimore, USA, March 2017.
The authors of this paper use neural networks to learn decoders for random and structured codes, such as polar codes. The key observations are that (i) optimal bit-error rate performance for both code families and short codeword lengths can be achieved, (ii) structured codes are easier to learn, and (iii) the neural network is able to generalize codewords that it has never seen during training for the structured codes, but not for the random codes. Scaling to long codewords is identified as the main challenge for neural network-based decoding due to the curse of dimensionality.
E. Nachmani, E. Marciano, L. Lugosch, W. J. Gross, D. Burshtein, and Y. Be’ery, “Deep Learning Methods for Improved Decoding of Linear Codes,” IEEE Journal of Selected Topics in Signal Processing, vol. 12, no. 1, pp. 119-131, February 2018.
The paper applies deep learning to the decoding of linear block codes with short to moderate block length based on a recurrent neural network architecture. The methods show advantages in complexity and performance in the belief propagation and min-sum algorithms.
- Topic: Channel Estimation, Prediction, and Compression
H. He, C.-K. Wen, S. Jin, and G. Y. Li, “Deep Learning-Based Channel Estimation for Beamspace mmWave Massive MIMO Systems,” IEEE Wireless Communications Letters, vol. 7, no. 5, pp. 852-855, October 2018.
This paper develops a deep learning-based channel estimation network for beam-space millimeter-wave massive multi-input multi-output (MIMO) systems. A neural network is used to learn the channel structure and estimate the channel from a large amount of training data. It provides an analytical framework on the asymptotic performance of the channel estimator. Results demonstrate that that the neural network significantly outperforms state-of-the-art compressed sensing-based algorithms even when the receiver is equipped with a small number of RF chains.
C.-K. Wen, W.-T. Shih, and S. Jin, “Deep Learning for Massive MIMO CSI Feedback,” IEEE Wireless Communications Letters, vol. 7, no. 5, pp. 748-751, October 2018.
This article develops a novel channel state information (CSI) sensing and recovery mechanism using deep learning. The new approach learns to exploit channel structure effectively from training samples and transforms CSI to a near-optimal number of representations/codewords. The proposed approach can recover CSI with significantly improved reconstruction performance compared to the existing compressive sensing (CS)-based methods, even at excessively low compression regions where the traditional CS-based methods fail.
Y. Wang, M. Narasimha, and R. W. Heath, Jr., “MmWave Beam Prediction with Situational Awareness: A Machine Learning Approach,” in Proc. IEEE International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Kalamata, Greece, June 2018.
This article combines machine learning tools and situational awareness to learn the beam information, e.g., received power and the optimal beam index, in millimeter-wave communication systems. It uses the vehicle locations as features to predict the received power of any beam in the beam codebook and shows that situational awareness can largely improve the prediction accuracy. The method requires almost no overhead and can achieve high throughput with only a small performance degradation.
D. Neumann, T. Wiese, and W. Utschick, “Learning the MMSE Channel Estimator,” IEEE Transactions on Signal Processing, vol. 11, no. 66, pp. 2905-2917, June 2018.
This paper addresses the problem of estimating Gaussian random vectors with random covariance matrices. The authors develop a neural network architecture inspired from expert knowledge about the covariance matrix structure, which achieves state-of-the-art performance with an order-of-magnitude less complexity. This is a good example of how expert knowledge can be combined with machine learning methods to outperform purely model-based approaches.
- Topic: End-to-End Communications
S. Dörner, S. Cammerer, J. Hoydis, and S. ten Brink, “Deep Learning-Based Communication over the Air,” IEEE Journal Selected Topics in Signal Processing, vol. 12, no. 1, pp. 132-143, February 2018.
This paper reports the world’s first implementation of a fully neural network-based communication system using software-defined radios. The authors identify the “missing channel gradient” as the biggest obstacle in training such systems over actual channels and propose a workaround based on model-based training in simulations followed by receiver-finetuning on measured data. Their implementation comes close to, but does not outperform, a well-designed baseline. A special neural network structure for the task of synchronization to single-carrier waveforms is introduced.
H. Ye, G. Y. Li, B.-H. Juang, and K. Sivanesan, “Channel Agnostic End-to-End Learning Based Communication Systems with Conditional GAN,” in Proc. IEEE Global Communication Conference (GLOBECOM) Workshops, Abu Dhabi, UAE, December 2018.
This paper employs a conditional generative adversarial net (GAN) to build an end-to-end communication system without a channel model, where deep neural networks (DNNs) represent both the transmitter and the receiver. The conditional GAN learns to generate the channel effects and acts as a bridge for the gradients to pass through in order to jointly train and optimize both the transmitter and the receiver DNNs.
F. Ait Aoudia and J. Hoydis, “End-to-End Learning of Communications Systems without a Channel Model,” in Proc. IEEE Asilomar Conference on Signal, System, Computers, Pacific Grove, USA, October 2018.
The authors provide a solution to the problem of training autoencoder-based communication systems over actual channels without any channel model. The key idea is to estimate the channel gradient using the technique of policy gradients from the field of reinforcement learning. They show through simulations that this approach works as well as model-based learning for the AWGN and Rayleigh channels.
B. Karanov, M. Chagnon, F. Thouin, T. A. Eriksson, H. Bülow, D. Lavery, P. Bayvel, and L. Schmalen, “End-to-End Deep Learning of Optical Fiber Communications,” Journal of Lightwave Technology, vol. 36, no. 20, pp. 4843-4855, October 2018.
This paper uses an autoencoder to learn transmitter and receiver neural networks for use in optical communications. The experimental results show that the autoencoder can effectively learn to deal with nonlinearities and that it provides good performance for different link dispersions.
- Topic: Resource Allocation
H. Sun, X. Chen, Q. Shi, M. Hong, X. Fu, and N. D. Sidiropoulos, “Learning to Optimize: Training Deep Neural Networks for Interference Management,” IEEE Transactions on Signal Processing, vol. 66, no. 20, pp. 5438-5453, October 2018.
This paper exploits deep neural networks (DNNs) to address optimization and interference management issues. The input and output of a signal-processing (SP) algorithm is treated as an unknown nonlinear mapping and is approximated by a DNN. It demonstrates that SP tasks can be performed effectively for those optimization problems that can be learned accurately by a DNN of moderate size. The paper also identifies a class of optimization algorithms that can be solved by a moderate size DNN and then uses interference management algorithm as an example to demonstrate the effectiveness of the proposed approach
V. Va, J. Choi, T. Shimizu, G. Bansal, and R. W. Heath, Jr., “Inverse Multipath Fingerprinting for Millimeter Wave V2I Beam Alignment,” IEEE Transactions on Vehicular Technology, vol. 67, no. 5, pp. 4042-4058, May 2018.
This paper uses multipath fingerprinting to address the beam alignment problem in millimeter wave vehicle-to-infrastructure communications. Based on the vehicle's position (e.g., available via GPS), the multipath fingerprint/signature is first obtained from a database and provides prior knowledge of potential pointing directions for reliable beam alignment, which can be regarded as the inverse of fingerprinting localization. From the extensive simulation results, the proposed approach provides increasing rates with larger antenna arrays while IEEE 802.11ad has decreasing rates due to the higher beam training overhead.
U. Challita, L. Dong, and W. Saad, “Proactive Resource Management for LTE in Unlicensed Spectrum: A Deep Learning Perspective,” IEEE Transactions on Wireless Communications, vol. 17, no. 7, pp. 4674-4689, July 2018.
This paper develops a deep learning-based resource allocation framework for coexistence of long term evolution (LTE) networks with licensed assisted access (LTE-LAA) and WiFi in the unlicensed spectrum. With long short-term memory, each small-cell base station is able to decide on its spectrum allocation autonomously by requiring only limited information on the network state.
R. Daniels and R. W. Heath, Jr., “An Online Learning Framework for Link Adaptation in Wireless Networks,” in Proc. Information Theory and Applications Workshop (ITA), San Diego, USA, February 2009.
This paper is one of the earliest works on machine learning (ML) for link adaptation. The motivation for using ML lies in the difficulty to model the impairments in wireless communications (nonlinearities, interference). It uses real-time measurements to build and continuously adapt a classification procedure based on k-nearest neighbor. Follow-up work relies on support vector machines (SVMs).
- Topic: Selected Topics
T. J. O’Shea, T. Roy, and T. C. Clancy, “Over-the-Air Deep Learning Based Radio Signal Classification,” IEEE Journal of Selected Topics in Signal Processing, vol. 12, no. 1, pp. 168-179, February 2018.
In this paper, the widely studied problem of modulation classification is revisited using a neural network operating on raw IQ samples. The authors demonstrate that neural networks can outperform the best-known alternative methods based on expert features for several realistic datasets obtained from over-the-air measurements and simulations.
X. Wang, L. Gao, S. Mao, and S. Pandey, “CSI-Based Fingerprinting for Indoor Localization: A Deep Learning Approach,” IEEE Transactions on Vehicular Technology, vol. 66, no. 1, pp. 763-776, January 2017.
The paper presents an indoor localization method based on deep learning (DL). The DL algorithm exploits channel state information in the frequency domain (amplitude and phase) from three distant antennas for an indoor OFDM system. Location is mapped to fingerprints that are the optimal weights of the deep learning network. Training and testing are performed with experimental data.
M. Kim, N.-I. Kim, W. Lee, and D.-H. Cho, “Deep Learning-Aided SCMA,” IEEE Communications Letters, vol. 22, no. 4, pp. 720-723, April 2018.
This article uses machine learning to design sparse code multiple access (SCMA) schemes. Deep neural networks (DNNs) are used to adaptively construct a codebook and learn a decoding strategy to minimize bit-error rate for SCMA. The proposed deep learning based SCMA can provide improved spectral efficiency and massive connectivity, which is a promising technique for 5G wireless communication systems.
C. Studer, S. Medjkouh, E. Gönültaş, T. Goldstein, and O. Tirkkonen, “Channel Charting: Locating Users within the Radio Environment Using Channel State Information,” IEEE Access, vol. 6. pp. 47682-47698, August 2018.
This paper uses passively collected channel state information (CSI) in conjunction with autoencoder neural networks and other machine learning methods in order to perform relative localization of users within a cell. Extensive simulation results demonstrate that the various proposed methods are able to successfully preserve the local geometry of users and that the autoencoder performs particularly well.
J. Vieira, E. Leitinger, M. Sarajlic, X. Li, and F. Tufvesson, “Deep Convolutional Neural Networks for Massive MIMO Fingerprint-Based Positioning,” in Proc. IEEE International Symposium on Personal, Indoor, and Mobile Radio Communications (PIMRC), Montreal, Canada, October 2017.
The paper exploits the sparsity of the channel in the angular domain in a massive multi-input multi-output (MIMO) system to build a map between user position and channel angular pattern. To learn the mapping, the paper uses a convolutional neural network that is trained using measured and simulated channels.
A. Balatsoukas-Stimming, “Non-Linear Digital Self-Interference Cancellation for In-Band Full-Duplex Radios using Neural Networks,” in Proc. IEEE International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Kalamata, Greece, June 2018.
This paper uses neural networks in order to perform nonlinear self-interference cancellation in in-band full-duplex communications. Experimental results show that the neural network can achieve similar performance with a state-of-the-art memory polynomial cancellation method, but with significantly lower computational complexity.
C. Häger and H. D. Pfister, “Nonlinear Interference Mitigation via Deep Neural Networks,” in Proc. Optical Fiber Communications Conference and Exposition (OFC), San Diego, USA, March 2018.
This paper uses a deep neural network to mitigate nonlinear effects in optical communications. Inspired by an existing nonlinearity compensation algorithm, the developed neural network structure gives simple and natural choices for the neural network hyperparameters.
J. Liu, R. Deng, S. Zhou, and Z. Niu, "Seeing the Unobservable: Channel Learning for Wireless Communication Networks," in Proc. IEEE Global Communications Conference (GLOBECOM), San Diego, USA, December 2015.
This is an initial work on channel learning. It develops a novel framework to infer unobservable CSI from the observable ones. In particular, it proposes a neural-network-based algorithm for the cell selection in multi-tier networks. Simulations show that the average cell-selection accuracy of the proposed algorithm, which requires no genuine location information, has a 3.9% loss compared to the location-aided algorithm, which requires genuine location.
S. Chen, Z. Jiang, J. Liu, R. Vannithamby, S. Zhou, Z. Niu, and Y. Wu, “Remote Channel Inference for Beamforming in Ultra-Dense Hyper Cellular Network,” in Proc. IEEE Global Communications Conference (GLOBECOM), Singapore, December 2017.
This paper proposes a learning-based channel estimation method for coordinated beamforming in ultra-dense networks. It first shows that the channel state information of geographically separated base stations (BSs) exhibits strong non-linear correlations. Then, an artificial neural network is used to remotely infer the quality of different beamforming patterns at a dense-layer BS. Moreover, by involving more candidate beam patterns, a joint learning scheme for multiple BSs is developed to further improve the performance.