I Introduction
In a traditional wireless communication system shown in Fig. 1.a), the data transmission entails multiple signal processing blocks in the transmitter and the receiver. While the technologies in this system are quite mature, individual blocks therein are separately designed and optimized, often with different assumptions and objectives, making it difficult, if not impossible, to ascertain global optimality of the system. In addition, the channel propagation is expressed as an assumed mathematical model embedded in the design. The assumed model may not correctly or accurately reflect the actual transmission scenario, thereby compromising the system performance.
On the contrary, the learning based datadriven methods provide a new way for handling the imperfection of the assumed channel models [1]. Recently, deep learning has been applied to refine the traditional blockstructure communication systems, including the multipleinput and multipleoutput (MIMO) detection [2, 3], channel decoding [4, 5, 6, 7, 8], and channel estimation [9, 10]. In addition, deep learning based methods have also shown impressive improvement by jointly optimizing the processing blocks, including joint channel estimation and detection [11], joint channel encoding and source encoding [12].
Besides enhancing the traditional communication blocks, deep learning provides a new paradigm for future communication systems. As a pure datadriven method, the features and the parameters of a deep learning model can be learned directly from the data, without handcraft or adhoc designs, by optimizing an endtoend loss function. Inspired by this methodology, endtoend learning based communication systems have been investigated in several prior works
[13, 14, 15, 16, 17], where both the transmitter and the receiver are represented by deep neural networks (DNNs) and can be interpreted as an autoencoder and an autodecoder, respectively, as shown in Fig 1.b).From Fig. 1b), the transmitter learns to encode the transmitted symbols into encoded data, , which is then sent to the channel， while the receiver learns to recover the transmitted symbols based on the received signal,
, from the channel. As a result, the traditional communication modules at the transmitter, such as the encoding and modulation, are replaced by a DNN while the modules in the receiver, such as the decoding and the demodulation, are replaced by another DNN. The weights/parameters of the two DNNs are trained in a supervised learning manner to optimize the endtoend recovery accuracy. The robustness to noise can be obtained by adding noise to the hidden layers to simulate the effect of the wireless channels.
However, several critical challenges in the learning based endtoend communication system need to be addressed in order to apply this framework to various wireless channels. As is well known, the weights of the DNN are usually updated using stochastic gradient descent (SGD) with the computed error gradients propagated from the output layer back to the input layer. When the channel transfer function,
, is not available, the backpropagation of the gradients from the receiver DNN to the transmitter DNN is blocked, preventing the overall learning of the endtoend system. The channel transfer function may be assumed, but any such assumption would bias the learned weights, repeating the pitfalls caused by the likely discrepancy between the assumed model and the actual channel. In addition, in real communication systems, an accurate instantaneous CSI is hard to obtain in advance due to the various inherent uncertainties of wireless channels, such as channel noise and being timevarying. These uncertainties are often unknown or cannot be expressed analytically.Another key challenge of the endtoend paradigm is the curse of dimensionality during the training when the transmitted symbol sequence is long. The code block size in a communication system needs to be long enough to ensure a sufficient coding gain. However, as the size of possible codewords grows exponentially with the code block size, the portion of the unseen codewords during training will significantly increase accordingly. Previous works on the learning based decoding [6] show that the decoding performance of the DNN on the unseen codewords is still poor even if nearly of the codewords have been included in the training of the DNN. Therefore, almost all the previous works on the endtoend paradigm are concentrating on examples with a small block size, such as the Hamming codes (7,4) [13, 18]. As a result, it is desirable to develop a channel agnostic endtoend communication system based on deep learning, where different types of channel effects can be automatically learned without knowing the specific channel transfer function and the blocklength remains long enough to be practical.
In this article, we develop a channel agnostic endtoend communication system to address the challenges, where the distributions of channel output are learned through a conditional generative adversarial net (GAN) [19], as shown in Fig 1.c). The conditioning information for the GAN to generate samples is the encoded signals from the transmitter along with the received pilot information used for estimating the channel. By iteratively training the conditional GAN, the transmitter, and the receiver, the endtoend loss can be optimized in a supervised way.
This channel agnostic endtoend system provides a new way to optimize communication systems and is applicable to a wide range of wireless channels. In addition, the convolutional neural network (CNN) is used to overcome the curse of dimensionality and the blocklength can be extended from several bits to a couple of hundred bits. Our main contributions in this article are fourfold.

We are the first to exploit the conditional GAN to model the channel conditional distribution, , so that the channel effects can be learned based on the data instead of expert knowledge.

By adding the pilot symbol as a part of the conditioning information for the timevarying channels, the conditional GAN can generate more specific samples for the current channel.

Based on the learned channel conditional distribution, an endtoend learning based communication system is developed, where the gradients of the endtoend loss can be propagated to the transmitter DNN through the conditional GAN.

CNN is employed for alleviating curse of dimensionality. From the experimental results, the transmitter DNN with convolutional layers can learn to encode the information bits into a high dimensional embedding vector and the code can be efficiently decoded by the receiver DNN.
Part of the work has been published in [18]. Compared with the previous work, we have made two significant improvements. First, we introduce a convolutional layers so that this approach can be extended from several bits to a couple of hundred bits. Second, we apply our framework to more practical wireless channels, such as frequencyselective channels where there exists intersymbol interference (ISI).
The rest of the paper is organized as follows. The related works are discussed in Section II. In Section III, the conditional GAN based channel modeling approach is introduced. In Section IV, the training for the endtoend system is presented in detail. In Section V, the simulation results are presented and the conclusions are drawn in Section VI.
Ii Related Works
Our proposed method is closely related to GANs, endtoend learning based communication systems, and learning based decoders. In this section, previous works in the related topics are briefly reviewed.
Iia GANs and Conditional GANs
GAN has been proposed in [21]
as a generative framework, where a generator and a discriminator are competing with each other in the training stage. By the feedback of the discriminator, the generator improves its ability to generate samples that are similar to the real samples. GAN is most widely used in computer vision. Much of the recent GAN research is focusing on improving the quality of the generated images
[22].In order to generate samples with a specific property, a conditional GAN is proposed based on the GAN framework, where the context information is added to the generator and the discriminator. Originally, the condition added is the label information so that the generator can generate sample data given a particular category. Nowadays, conditional GAN is widely used in changing the style and the content of the input [23, 27]. For instance, GAN has been utilized to generate highresolute images from lowresolution images [23].
Apart from application in computer vision, recently GAN has been exploited to model the channel effects of additive white Gaussian noise (AWGN) channels [20], similar to our work. However, our approach can be applied to more realistic timevarying channels by using conditional GAN, which employs the received pilot information as a part of the condition information when generating the channel outputs.
IiB DNN based EndtoEnd Communications
An endtoend learning system has been proposed in [13] and has been shown to have a similar performance as the traditional approaches with block structures under the AWGN condition. In [14], the endtoend method has been extended to handle various hardware imperfection. In [15], an endtoend learning method is adopted within the orthogonal frequencydivision multiplexing (OFDM) system. In [24], CNN is employed for modulation and demodulation, where improved results have been shown for very high order modulation. In addition, source coding can also be considered as a part of the endtoend communication system for transmitting text or image [12].
Training the endtoend communication system without channel models has been investigated recently. A reinforcement learning based framework has been employed in
[16] to optimize the endtoend communication system without requiring the channel transfer function or CSI, where the channel and the receiver are considered as the environment when training the transmitter. The recovery performance at the receiver is considered as the reward, which guides the training of the transmitter. In [17], a modelfree endtoend learning method has been developed based on stochastic perturbation methods. However, both works focus on small block length. For example, blocks of eight information bits are used in [16]. How to extend a large block size and how to model the unknown channel using a datadriven approach are still open problems, which are addressed in our proposed approach.IiC Learning based Decoders
Our proposed method is also closely related to learning based encoding and decoding. Learning based approach has been utilized in improving decoding performance for a long time. It dates back to 1990s when several attempts had been made to decode the codes with the recurrent neural network (RNN)
[25].With the widely used deep learning approaches, DNN has been utilized in decoding with a wide range of applications. In [6], a fully connected neural network is trained for decoding short Polar codes. The performance of decoding is similar to maximum likelihood decoding. But they find it difficult to break through the curse of dimensionality. In order to train long codewords, a partition method has been employed in [6]. Moreover, there have been several trials to incorporate some prior information in the decoding process. RNN is used in [5] for decoding the convolutional and turbo codes. In [4], the traditional beliefpropagation decoding algorithm is extended as deep learning layers to decode linear codes.
One of the interesting findings from our experiment is that the curse of dimensionality can be mitigated when the CNN is used to learn the encoding and decoding modules simultaneously rather than just learning to decode humandesigned codewords.
Iii Modeling Channel with Conditional GAN
An endtoend communication system learns to optimize DNNs for the transmitter and the receiver. However, the backpropagation, which is used to train the weights of DNNs, is blocked by the unknown CSI, preventing the overall learning of the endtoend system. To address the issue, we use a conditional GAN to learn the channel effects and to act as a bridge for the gradients to pass through. By the conditional GAN, the output distribution of the channel can be learned in a datadriven manner and therefore many complicated effects of the channel can be addressed. In this section, we introduce the conditional GAN and discuss how to use it to model the channel effects.
Iiia Conditional GAN
GAN [21] is a new class of generative methods for distribution learning, where the objective is to learn a model that can produce samples close to some target distribution, . In our system, a GAN is applied to model the distribution of the channel output and the learned model is then used as a surrogate of the real channel when training the transmitter so that the gradients can pass through to the transmitter.
As shown in Fig. 2, a minmax two players game is introduced in GAN between a generator, , and a discriminator, , both represented by DNNs. The discriminator, , learns to distinguish between the data generated by the generator and the data from the real dataset while the generator, , learns to generate samples to fool the discriminator into making mistakes.
During the training, the generator maps an input noise, , with prior distributions, , to a sample. Then the samples from the real data and those generated from are collected to train
, to maximize the ability to distinguish between the two categories. If the discriminator can successfully classify the samples of the two sources, then its success will be used to generate feedback to
, so that the generator can learn to produce samples more similar to the real samples. The training procedure will end upon reaching an equilibrium, where the discriminator, , can do no better than random guessing to distinguish the real samples and the generated fake samples.Denote the parameter sets of the generator, , and the discriminator, , as and , respectively, and the objective for optimization is
(1) 
The objective of the discriminator, , is to give a high value when the input belongs to the real dataset and a low one when the input is generated by the generator, , while the objective of generator, , is to maximize the output of the discriminator, , given the generated samples, .
The GAN can be extended to a conditional model if both the generator, , and the discriminator, , are conditioned on some extra information, , as in Fig. 2. We only need to feed the conditioning information, , into both the generator and discriminator as the additional input. Therefore, the output of the will be and the output of will be . The minmax optimization objective becomes
(2) 
The conditional GAN is employed in our endtoend communication system to model the channel output distribution with the given conditioning information on the encoded signal and the received pilot data.
IiiB Modeling Channels
Since the channel output, , for given input, , is determined by the conditional distribution, , a conditional GAN can be employed for learning the output distribution of a channel by taking as the condition information. The generator will try to produce the samples similar to the output of the real channel while the discriminator will try to distinguish data coming from the real channel and the data coming from the generator.
The instantaneous CSI, , is regarded as a sample from a large channel set and is also vital for coherent detection of the transmit symbols at the receiver. In order to obtain the CSI, a common practice is to send some pilot information to the receiver so that the channel information is inferred based on the received signal corresponding to the pilot symbols, . In our proposed method, the received signal corresponding to the pilot symbols, , is added as a part of the conditioning information so that the output samples follow the distribution of given the input and the received pilot data, .
IiiC Convolutional Layers based Channel GAN
The convolutional layer has been introduced to efficiently extract features for images based on their sharedweight architecture and translational invariance characteristics [29]
. In a fully connected layer, each neuron is connected to all neurons in the previous layer. In contrast, in a convolutional layer, each neuron is only connected to a few nearby neurons in the previous layer, which is called the receptive field of this neuron, and the same set of weights is shared for all neurons in a layer. Inspired by the convolutional codes, where the encoding process can be represented by a convolutional transform, we use hierarchical onedimensional convolutional layers in the channel GAN as well as the DNNs used in the transmitter and the receiver.
Denote as the output of the th neuron in the th layer of a DNN. For a fully connected layer, the output of the th neuron in the th layer is
where
is an activation function and
is the weight connected the th neuron in the th layer and the th neuron in the th layer. is different for different , , or . Therefore, if there are neurons in the th layer, there will be weights in total to fullly connect the th layer to the th layer.On the other hand, for a convolutional layer, the output of the th neuron in the th layer will be
where is the coefficient for the convolution and is same for different ’s in the th layer. There are weights in total ( is usually much smaller then or ) to implement connections between the th and the th layers. In brief, compared with a fully connected DNN, the convolutional nerual network has fewer connections between adjacent layes and much fewer weights to train, which will reduce the complexity and significantly improve the convergence speed of training.
Apart from being easier to train, CNN has two additioinal merits in the endtoend communication system. First, the curse of dimensionality can be alleviated by the usage of convolutional layers. When both the transmitter and the receiver are represented by CNNs, the codes learned by a CNN are more easily recovered at the receiver than the conventional hand designed codes. Second, it is appropriate to employ convolutional layers to deal with the ISI channels since the effect of the channel can be expressed by the convolutional operation in the ISI channel.
Iv EndtoEnd Communication System
As stated in the introduction, the endtoend communication paradigm can be interpreted as a deep autoencoder framework. With the conditional GAN, the gradients can be backpropagated to the transmitter even if channels are known. In this section, the proposed framework is first introduced and the training procedures for each module are presented in detail.
Iva System Overview
As in Fig. 1.b) the autoencoder learns to map information bits, into a fixed length embedding of length , , and sends the embedding to the channel while the autodecoder learns to recover the original information according to the received signal from the channel. The distance between the original information bits, , and the recovered information, , will be calculated. Here, the binary crossentropy loss is used to measure the distance, which can be expressed as
(3) 
where and represent the th elements of and , respectively.
The training and testing of the proposed endtoend communication system are shown in Fig. 3. To obtain training data set, the information bits, , are randomly generated and the instantaneous CSI is sampled randomly from the channel set. Due to different objectives in modules, the transmitter, the receiver, and the channel generator in the conditional GAN can be trained iteratively based on the training data. When training one component, the parameters of the others remain fixed. The object is to minimize the endtoend loss when training the receiver and the transmitter. It is to minimize the minmax optimization objective when training the conditional GAN for generating the channel. In the testing stage, the endtoend reconstruction performance is evaluated on the learned transmitter and receiver with real channels.
IvB Training Receiver
At the receiver, a DNN model is trained for recovering the transmitted signal , where the input is the received signals corresponding to the transmitted data, , while the output is the estimation . By comparing the and , the loss function can be calculated based on (3). The receiver can be trained easily since the loss function is computed at the receiver and thus the gradients of the loss can be easily obtained. For the timevarying channels, by directly put the received signal, , and the receive pilot data, , together as the input, the receiver can automatically infer the channel condition and perform the channel estimation and detection simultaneously without explicitly estimating the channel, as we have discussed in [11].
IvC Training Transmitter
With the channel generator as a surrogate channel, the training of the transmitter will be similar to that of the receiver. During the training, the transmitter, the generator, and the receiver can be viewed as a whole DNN. The output of the transmitter is the values of the last hidden layer in the transmitter. The endtoend crossentropy loss is computed at the receiver as in (3), and the gradients are propagated back to the transmitter through the conditional GAN. The weights of the transmitter will be updated based on SGD while the weights of the conditional GAN and the receiver remain fixed. The transmitter can learn the constellation of the embedding, , so that the received signal can be easily detected at the receiver.
IvD Training Channel GAN
The conditional GAN is trained by iteratively training the generator and discriminator so that the minmax point can be found. The parameters of one model will be fixed while training the other. The channel generator is trained with the discriminator together. With the learned transmitter, the real data can be obtained with the encoded signal from the transmitter going through the real channel while the fake data is obtained from the encoded data going through the channel generator. The objective function for optimization is as shown in (2).
V Experiments
In this section, the implementation details of the endtoend learning based approach are provided and the simulation results are presented. For several types of most commonly used channels, the channel GAN has shown the ability to model the channel effects in a datadriven way. In addition, the endtoend communication system, which is built on the channel GAN, can achieve similar or better results even when the channel information is unknown when training and optimizing the transmitter and the receiver.
Va Experimental Settings
VA1 Implementation Details
Two types of DNN models are designed in our experiments. One is fully connected networks (FCN) and the other is the CNN. The FCN is used for a small block size and the CNN is used in the a large block size to avoid the curse of dimensionality. The parameters of the FCN and CNN are shown in Table I and Table II, respectively. The weights of both models are updated by Adam [26] and the batch size for training is 320.
Parameters  Values 

Transmitter hidden layers  32, 32 
Learning rate  0.001 
Receiver hidden layers  32, 32 
Learning rate  0.001 
Generator hidden layers  128, 128, 128 
Discriminator hidden layers  32, 32, 32 
Learning rate  0.0001 
Type of layer  Kernel size/Annotation  Output size 

Transmitter  
Input  Input layer  
Conv+Relu 
5  
Conv+Relu  3  
Conv+Relu  3  
Conv  3  
Normalization  Power normalization  
Receiver  
Conv+Relu  5  
Conv+Relu  5  
Conv+Relu  5  
Conv+Relu  5  
Conv+Relu  5  
Conv+Relu  5  
Conv+Relu  5  
Conv+Sigmoid  3  
Generator  
Conv+Relu  5  
Conv+Relu  3  
Conv+Relu  3  
Conv  3  
Discrimator  
Conv+Relu  5  
Conv+Relu  3  
Conv+Relu  3  
Conv+Relu  3  
FC +Relu  100  100 
FC+Sigmoid  1  1 
VA2 Channel Types
Three types of channels are considered in our experiments, i.e., AWGN channels, Rayleigh channels, and frequencyselective multipath channels. In an AWGN channel, the output of the channel, , is the summation of the input signal, , and Gaussian noise, , that is, . Rayleigh fading is a reasonable model for narrowband wireless channels when many objects in the environment scatter the radio signal before arriving at the receiver. In a Rayleigh channel, the channel output is determined by , where . The channel coefficient is timevarying and is unknown when design transceivers. Therefore, channel estimation is required to get the instantaneous CSI, for receiver to detect the transmitter information.
With frequencyselective channels, radio signal propagate via multiple paths, which differ in amplitudes, phases, and delay times, and cause undesired frequencyselective fading and time dispersion of the received signal. The baseband complex channel impulse response can be expressed as
where there are pathes in all, , , and represent the path gain, the phase shift, and the time delay of the th path, respectively, and is the shaping pulse in the communication system. In our simulation, a threetap channel with equal average power is considered, that is, , and with as the symbol duration.
VA3 Baselines
The endtoend learning based communication system is compared with the conventional communication system, which is composed of multiple signal processing modules and each module is designed based on the prior knowledge on the channel. The biterror rate (BER) and blockerror rate (BLER) are compared under each type of channel. In our baseline system, 4 QAM is used as the modulation and the Hamming code or convolutional codes are used. For the convolutional codes, the Viterbi algorithm [30]
is used for the maximum a posteriori probability (MAP) decoding. A commonly used example of a convolutional code, rate1/2 recursive systematic convolutional (RSC) code, is adopted. OFDM is utilized to deal with the ISI in the frequencyselective multipath channel.
VB Modeling the Channel Effects
We use FCN to model the effects of Rayleigh fading channels. Since Rayleigh fading channels are timevarying, additional conditional information is added to the channel generator and the receiver. Besides the encoded signal, the received pilot data, , is used as the additional conditional information. We test the effectiveness of the conditional GAN in learning the distribution of the channel with standard 16 QAM as the encoded symbols. Fig. 4 shows generated samples of a Rayleigh fading channel with different values added to the conditioning information. From the figure, the conditional GAN is able to produce the samples with different channel gains and phase rotations according to conditioning information.
VC Endtoend Communication System
Based on the channel GAN, a channel agnostic endtoend communication system is built on three types of channels, i.e., the AWGN channel, the Rayleigh fading channel, and the frequencyselective multipath channel. We compare our channel agnostic endtoend learning based approach with the traditional methods, which are designed based on the channel transfer functions.
VC1 AWGN Channel
We first use FCN for a small block size. The endtoend recovering performance on the AWGN channel is shown in Fig. 5. At each time, four information bits are transmitted and the length of the transmitter output is set to be seven. From the figure, the BER and BLER of learning based approach is similar to Hamming (7,4) code with maximumlikelihood decoding (MLD).
In order to train models with a large block size, CNN is then used to mitigate the curse of dimensionality. We first train the CNN under the AWGN channel where the noise is added to the hidden layer directly, as used in [13]. The network is trained at dB fixed signaltonoise ratio (SNR) and tested by different SNRs. Fig. 6 shows the the BER and BLER curves of proposed endtoend method with the length of transmit information sequence 64 bits and 100 bits, respectively, which are denoted by “E2E64” and “E2E100”, respectively. From the figure, the performance of the proposed method is similar to RSC in the low SNR area and significantly outperforms RSC in the high SNR area.
VC2 Rayleigh Fading Channel
CNN is employed in transmission with a large block size and the channel encoding is included. We compare the endtoend approach with a baseline method, where QAM is used as the modulation and the RSC of coderate are used as the coding. In each block, information bits will be transmitted, thus the input size of the endtoend approach is 64. From Fig. 7, the endtoend approach shows similar performance to the traditional methods in terms of BER and BLER, where in the baseline system, QAM is used as the modulation and the RSC is used as the coding.
VC3 FrequencySelective Fading Channel
Under the frequencyselective channel, coded and uncoded endtoend communication systems are developed with CNN and the OFDM system is used as the baseline. There are 64 subcarriers in the OFDM system and the length cycleprefix is set as 16 and 4 QAM is used for the modulation. In the coded system, the RSC coding is adopted. In order to have a fair comparison, we set the block size of the endtoend system as 64 bits and pad 16 zeros between every two blocks.
Fig. 8 shows the performance of the propose endtoend approach The input size of endtoend approach 64 bits. From Fig. 8, the proposed endtoend system significantly outperforms the OFDM system when the SNR is over 7 dB. In addition, the mismatch of training and testing is considered. In the training stage, an exponential decay PDP, , is used to generate the channels while in the testing stage the equal strength PDP is used for evaluation. From the figure, even though the PDPs used in the training and testing are different to a large degree, the performance of the endtoend model trained with mismatch is comparable to the model without mismatch, still much better than the OFDM system, which proves the robustness of the proposed method to the discrepancy between training and testing.
Vi Conclusions and Discussions
In this article, we investigate the endtoend learning of a communication system without prior information of the channel. We show that the conditional distribution of the channel can be modeled by a conditional GAN. In addition, by adding the pilot information into the condition information, the conditional GAN can generate data corresponding to the specific instantaneous channel.
The endtoend pipeline consists of DNNs for the transmitter, the channel GAN, and the receiver. By iteratively training these networks, the endtoend loss can be optimized in a supervised way. The simulation results on the AWGN channles, Rayleigh fading channels, and frequencyselective channels confirm the effectiveness of the proposed method, by showing similar or better performance compared with the traditional approaches based on expert knowledge and channel models. Our research opens a new door for building the pure datadriven communication systems.
One of the future directions is to test the proposed method in real data. As we have indicated in the introduction, in the real communication scenario, many imperfections will make the real channel difficult to express, which is very suitable for modeling these effects in a datadriven manner.
References
References
 [1] Z. Qin, H. Ye, G. Y. Li, and B.H. F. Juang, “Deep learning in physical layer communications,” IEEE Wireless Communi., vol, 26, no. 2, Apr. 2019.
 [2] N. Samuel, T. Diskin, and A. Wiesel, “Deep MIMO detection,” in Proc. IEEE Int. Workshop Signal Process. Adv. Wireless Commun (SPAWC), Jul. 2017, pp. 690–694
 [3] H. He, C.K. Wen, S. Jin, and G. Y. Li, “A modeldriven deep learning network for MIMO detection,” in Proc. IEEE Global Conf. Signal Inf. Process., Anaheim, CA, Nov. 2018, pp. 1–5
 [4] E. Nachmani, Y. Be’ery, and D. Burshtein, “Learning to decode linear codes using deep learning,” in Proc. IEEE Annu. Allerton Conf. Commun. Control Comput. (Allerton), Sept. 2016, pp. 341–346.
 [5] H. Kim, Y. Jiang, R. Rana, S. Kannan, S. Oh, and P. Viswanath, “Communication Algorithms via Deep Learning,” arXiv preprint arXiv:1805.09317, 2018.
 [6] T. Gruber, S. Cammerer, J. Hoydis, and S. T. Brink, “On deep learningbased channel decoding,” in Proc. IEEE 51st Annu. Conf. Inf. Sci. Syst. (CISS), Baltimore, MD, USA, 2017, pp. 1–6.
 [7] S. Cammerer, T. Gruber, J. Hoydis, and S. t. Brink, “Scaling deep learningbased decoding of polar codes via partitioning,” arXiv preprint arXiv:1702.06901, 2017.
 [8] W. Xu, Z. Wu, Y.L. Ueng, X. You, and C. Zhang, “Improved polar decoder based on deep learning,” in IEEE Int. Workshop on Signal Process. Syst., Nov. 2017, pp. 1–6.
 [9] H. He, C.K. Wen, S. Jin, and G. Y. Li,“Deep learningbased channel estimation for beamspace mmWave massive MIMO systems,” IEEE Wireless Commun. Lett., vol. 7, no. 5, pp. 852–855, Oct. 2018.
 [10] X. Gao, C.K. Wen, S. Jin, and G. Y. Li, “ConvNet: Combination of deep learning and expert knowledge in OFDM receivers,” IEEE Commun. Lett., Jul. 2018. pp. 2627–2630
 [11] H. Ye, G. Y. Li, and B.H. F. Juang, “Power of deep learning for channel estimation and signal detection in OFDM systems,” IEEE Wireless Commun. Lett., no. 7, pp.114–117, Feb. 2018.
 [12] N. Farsad, M. Rao, and A. Goldsmith, “Deep learning for joint sourcechannel coding of text,” in Proc. ICASSP, Apr. 2018, pp. 1–5.
 [13] T. O’Shea and J. Hoydis, “An introduction to deep learning for the physical layer,” IEEE Trans. on Cogn. Commun. Netw., vol. 3, no. 4, pp. 563–575, Dec. 2017.
 [14] S. Dörner, S. Cammerer, J. Hoydis, and S. ten Brink, “Deep learningbased communication over the air,” IEEE J. Select. Topics Signal Process., vol.12, no. 1, pp. 132–143, Feb. 2018.

[15]
A. Felix, S. Cammerer, S. Dörner, J. Hoydis, and S. ten Brink, “OFDM autoencoder for endtoend learning of communications systems,” in
Proc. IEEE Int. Workshop Signal Proc. Adv. Wireless Commun.(SPAWC), Jun. 2018.  [16] F. Aoudia and J. Hoydis, “Endtoend learning of communications systems without a channel model,” arXiv preprint arXiv:1804.02276, 2018.

[17]
V. Raj and S. Kalyani, “Backpropagating through the air: Deep learning at physical layer without channel models,”
IEEE Commun. Lett,vol. 22, no. 11, pp. 2278–2281, Nov. 2018.  [18] H. Ye, G. Y. Li, B.H. Juang, and K. Sivanesan, “Channel agnostic endtoend learning based communication systems with conditional GAN,” in Proc. IEEE Global Commun. Conf., Abu Dhabi, UAE, Dec. 2018, pp. 1–5.
 [19] M. Mirza and S. Osindero, “Conditional generative adversarial nets,” arXiv preprint arXiv:1411.1784, 2014.
 [20] T. O’Shea, T. Roy, and N. West, “Approximating the Void: learning stochastic channel models from observation with variational generative adversarial networks,” arXiv preprint arXiv:1805.06350, 2018.
 [21] I. Goodfellow et al., “Generative adversarial nets,” in Proc. NIPS, Dec. 2014, pp. 2672–2680
 [22] A. Brock, J. Donahue, and K. Simonyan, “Large scale GAN training for high fidelity natural image synthesis,” arXiv preprint arXiv:1809.11096., 2018.

[23]
C. Ledig, L. Theis, F. Huszár, J. Caballero, A. Cunningham, A. Acosta, A. Aitken, A. Tejani, J. Totz, Z. Wang, and W. Shi, “Photorealistic single image superresolution using a generative adversarial network,” in
Proc. IEEE CVPR, Jul. 2017， pp. 4681–4690.  [24] B. Zhu, J. Wang, L. He, and J. Song, “Joint transceiver optimization for wireless communication PHY with convolutional neural network,” arXiv preprintarXiv:1808.03242.,
 [25] A. Hamalainen and J. Henriksson, “A recurrent neural decoder for convolutional codes,” in Proc. IEEE ICC, vol. 2, Jun. 1999, pp. 1305–1309.
 [26] D. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.

[27]
J.Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired imagetoimage translation using cycleconsistent adversarial networks,” in
Proc. IEEE ICCV, Oct. 2017, pp. 2223–2232. 
[28]
P. Vincent, H. Larochelle, Y. Bengio, and P. A. Manzagol, “Extracting and composing robust features with denoising autoencoders,” in
Proc. ICML Jul. 2008, pp. 1096–1103.  [29] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, and L. D. Jackel, “Backpropagation applied to handwritten zip code recognition,” Neural Computation, vol. 1, no. 4, pp. 541–551, Dec. 1989.
 [30] A. Viterbi, “Error bounds for convolutional codes and an asymptotically optimum decoding algorithm,” IEEE Trans. Inf. Theory, vol. 13, no. 2, pp. 260269, Apr. 1967. 2018.
Comments
There are no comments yet.