Spectrum Sharing in Vehicular Networks Based on Multi-Agent Reinforcement Learning

05/08/2019 ∙ by Le Liang, et al. ∙ Georgia Institute of Technology 0

This paper investigates the spectrum sharing problem in vehicular networks based on multi-agent reinforcement learning, where multiple vehicle-to-vehicle (V2V) links reuse the frequency spectrum preoccupied by vehicle-to-infrastructure (V2I) links. Fast channel variations in high mobility vehicular environments preclude the possibility of collecting accurate instantaneous channel state information at the base station for centralized resource management. In response, we model the resource sharing as a multi-agent reinforcement learning problem, which is then solved using a fingerprint-based deep Q-network method and amenable to a distributed implementation. The V2V links, each acting as an agent, collectively interact with the communication environment, receive distinctive observations yet a common reward, and learn to improve spectrum and power allocation through updating Q-networks using the gained experiences. We demonstrate that with a proper reward design and training mechanism, the multiple V2V agents successfully learn to cooperate in a distributed way to simultaneously improve the sum capacity of V2I links and success probability of payload delivery for V2V links.

READ FULL TEXT VIEW PDF
POST COMMENT

Comments

There are no comments yet.

Authors

page 1

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

Vehicular communication, commonly referred to as vehicle-to-everything (V2X) communication, is envisioned to transform connected vehicles and intelligent transportation services in various aspects, such as road safety, traffic efficiency, and ubiquitous Internet access [1, 2]. More recently, the 3rd Generation Partnership Project (3GPP) has been looking to support V2X services in long-term evolution (LTE) and future 5G cellular networks [3, 4, 5]. Cross-industry consortium, such as the 5G automotive association (5GAA), has been founded by telecommunication and automotive industries to push development, testing, and deployment of cellular V2X technologies.

I-a Problem Statement and Motivation

Fig. 1: An illustrative structure of vehicular networks, where V2I and V2V links are indexed by and (or ), respectively. Each of the V2I links is preassigned an orthogonal spectrum sub-band and hence the sub-band is also indexed by .

This paper considers spectrum access design in vehicular networks, which in general comprise both vehicle-to-infrastructure (V2I) and vehicle-to-vehicle (V2V) connectivity. As illustrated in Fig. 1, the V2I links connect each vehicle to the base station (BS) or BS-type road side unit (RSU) while V2V links provide direct communications among neighboring vehicles. We focus on the cellular based V2X architecture discussed within the 3GPP [3], where V2I and V2V connections are supported through cellular (Uu) and sidelink (PC5) radio interfaces, respectively. A wide array of new use cases and requirements have been proposed and analyzed for 5G V2X enhancements in Release 15 [4, 5]. For example, the 5G cellular V2X networks are required to provide simultaneous support for mobile high data rate entertainment and advanced driving in 5G cellular V2X networks. The entertainment applications require high bandwidth V2I connection to the BS (and further the Internet) for, e.g., video streaming. Meanwhile, the advanced driving service needs to periodically disseminate safety messages among neighboring vehicles (e.g., 10, 20, 50 packets per second depending on vehicle mobility [5]) through V2V communications, with high reliability. The safety messages usually include such information as vehicle position, speed, heading, etc. to increase “co-operative awareness” of the local driving environment for all vehicles.

This work is based on Mode 4 defined in the 3GPP cellular V2X architecture, where the vehicles have a pool of radio resources that they can autonomously select from for V2V communications [5]. To fully use available resources, we propose that such sidelink V2V connections share spectrum with Uu (V2I) links with necessary interference management design. We make some simplification on the V2I communications in that they have preoccupied the spectrum in an orthogonal way with fixed transmission power. Hence, the resource optimization is left for the design of V2V connections that need to devise effective strategies of spectrum sharing with V2I links, including the selection of spectrum sub-band and proper control of transmission power, to meet the diverse service requirements of both V2I and V2V links. Such an architecture provides more opportunities for the coexistence of V2I and V2V connections on limited frequency spectrum, but also complicates interference design in the network and hence motivates this work.

While there exists a rich body of literature applying conventional optimization methods to solve similarly formulated V2X resource allocation problems, they actually find difficulty to fully address them in several aspects. On one hand, fast changing channel conditions in vehicular environments causes substantial uncertainty for resource allocation, e.g., in terms of performance loss induced by inaccuracy of acquired channel state information (CSI). On the other hand, increasingly diverse service requirements are being brought up to support new V2X applications, such as simultaneously maximizing throughput and reliability for a mix of V2X traffic, as discussed earlier in the motivational example. Such requirements are sometimes hard to be modeled in a mathematically exact way, not to mention a systematic approach to find optimal solutions. Fortunately, reinforcement learning (RL) has been shown effective in addressing decision making under uncertainty [6]. In particular, recent success of deep RL in human-level video game play [7] and AlphaGo [8] has sparked a flurry of interest in applying RL techniques to solve problems from a wide variety of areas and remarkable progress has been made ever since [9, 10, 11]. It provides a robust and principled way to treat environment dynamics and perform sequential decision making under uncertainty, thus representing a promising method to handle the unique and challenging V2X dynamics. In addition, the hard-to-optimize objective issues can also be nicely addressed in a RL framework through designing training rewards such that they correlate with the final objective. The learning algorithm can then figure out a clever strategy to approach the ultimate goal by itself. Another potential advantage of using RL for resource allocation is that distributed algorithms are made possible, as demonstrated in [12], which treats each V2V link as an agent that learns to refine its resource sharing strategy through interacting with the unknown vehicular environment. As a result, we investigate the use of multi-agent RL tools to solve the V2X spectrum access problem in this work.

I-B Related Work

To address the challenges caused by fleeting channel conditions in vehicular environments, a heuristic spatial spectrum reuse scheme has been developed in

[13] for device-to-device (D2D) based vehicular networks, relieving requirements on full CSI. In [14], V2X resource allocation, which maximizes throughput of V2I links, adapts to slowly-varying large-scale channel fading and hence reduces network signaling overhead. Further in [15], similar strategies have been adopted while spectrum sharing is allowed not only between V2I and V2V links but also among peer V2V links. A proximity and QoS-aware resource allocation scheme for V2V communications has been developed in [16] that minimizes the total transmission power of all V2V links while satisfying latency and reliability requirements using a Lyapunov-based stochastic optimization framework. Sum ergodic capacity of V2I links has been maximized with V2V reliability guarantee using large-scale fading channel information in [17] or CSI from periodic feedback in [18]. A novel graph-based approach has been further developed in [19] to deal with a generic V2X resource allocation problem.

Apart from the traditional optimization methods, RL based approaches have been developed in several recent works to address resource allocation in V2X networks [20, 21]. In [22], RL algorithms have been applied to address the resource provisioning problem in vehicular clouds such that dynamic resource demands and stringent quality of service requirements of various entities in the clouds are met with minimal overhead. The radio resource management problem for transmission delay minimization in software-defined vehicular networks has been studied in [23]

, which is formulated as an infinite-horizon partially observed Markov decision process (MDP) and solved with an online distributed learning algorithm based on an equivalent Bellman equation and stochastic approximation. In

[24], a deep RL based method has been proposed to jointly manage the networking, caching, and computing resources in virtualized vehicular networks with information-centric networking and mobile edge computing capabilities. The developed deep RL based approach efficiently solves the highly complex joint optimization problem and improves total revenues for the virtual network operators. In [25], the downlink scheduling has been optimized for battery-charged roadside units in vehicular networks using RL methods to maximize the number of fulfilled service requests during a discharge period, where Q learning is employed to obtain the highest long-term returns. The framework has been further extended in [26], where a deep RL based scheme has been proposed to learn a scheduling policy with high dimensional continuous inputs using end-to-end learning. A distributed user association approach based on RL has been developed in [27] for vehicular networks with heterogeneous BSs. The proposed method leverages the -armed bandit model to learn initial association for network load balancing and thereafter updates the solution directly using historical association patterns accumulated at each BS. A similar handoff control problem in heterogeneous vehicular networks has been considered in [28], where a fuzzy Q-learning based approach has been proposed to always connect users to the best network without requirement of prior knowledge on handoff behavior.

This work differentiates itself from existing studies in at least two aspects. First, we explicitly model and solve the problem of improving the V2V payload delivery rate, i.e., the success probability of delivering packets of size within a time budget of . This directly translates to reliability guarantee for periodic message sharing of V2V links, which is essentially a sequential decision making problem across many time steps within the message generation period. Second, we propose a multi-agent RL based approach in this work to encourage and exploit V2V link cooperation to improve network level performance even when all V2V links make distributed spectrum access decisions based on local information.

I-C Contribution

In this paper, we consider the spectrum sharing problem in high mobility vehicular networks, where multiple V2V links attempt to share the frequency spectrum preoccupied by V2I links. To support diverse service requirements in vehicular networks, we design V2V spectrum and power allocation schemes that simultaneously maximize the capacity of V2I links for high bandwidth content delivery and meanwhile improve the payload delivery reliability of V2V links for periodic safety-critical message sharing. The major contributions of this work are summarized as follows.

  • We model the spectrum access of the multiple V2V links as a multi-agent problem and exploit recent progress of multi-agent RL [29, 30] to develop a distributed spectrum and power allocation algorithm that simultaneously improves performance of both V2I and V2V links.

  • We provide a direct treatment of reliability guarantee for periodic safety message sharing of V2V links that adjusts V2V spectrum sub-band selection and power control in response to small-scale channel fading within the message generation period.

  • We show that through a proper reward design and training mechanism, the V2V transmitters can learn from interactions with the communication environment and figure out a clever strategy of working cooperatively with each other in a distributed way to optimize system level performance based on local information.

I-D Paper Organization

The rest of the paper is organized as follows. The system model is described in Section II. We present the proposed multi-agent RL based V2X resource sharing design in Section III. Section IV provides our experimental results and concluding remarks are finally made in Section V.

Ii System Model

We consider a cellular based vehicular communication network in Fig. 1 with V2I and V2V links that provides simultaneous support for mobile high data rate entertainment and reliable periodic safety message sharing for advanced driving service, as discussed in 3GPP Release 15 for cellular V2X enhancement [4]. The V2I links leverage cellular (Uu) interfaces to connect vehicles to the BS for high data rate services while the V2V links disseminate periodically generated safety messages via sidelink (PC5) interfaces with localized D2D communications. We assume all transceivers use a single antenna and the set of V2I links and V2V links in the studied vehicular network are denoted by and , respectively.

We focus on Mode 4 defined in the cellular V2X architecture, where vehicles have a pool of radio resources that they can autonomously select for V2V communications [5]. Such resource pools can overlap with that of the cellular V2I interfaces for better spectrum utilization provided necessary interference management design is in place, which is investigated in this work. We further assume that the V2I links (uplink considered) have been preassigned orthogonal spectrum sub-bands with fixed transmission power, i.e., the th V2I link occupies the th sub-band. As a result, the major challenge is to design an efficient spectrum sharing scheme for V2V links such that both V2I and V2V links achieve their respective goals with minimal signaling overhead given the strong dynamics underlying high mobility vehicular environments.

Orthogonal frequency division multiplexing (OFDM) is exploited to convert the frequency selective wireless channels into multiple parallel flat channels over different subcarriers. Several consecutive subcarriers are grouped to form a spectrum sub-band and we assume channel fading is approximately the same within one sub-band and independent across different sub-bands. During one coherence time period, the channel power gain, , of the th V2V link over the th sub-band (occupied by the th V2I link) follows

(1)

where

is the frequency dependent small-scale fading power component and assumed to be exponentially distributed with unit mean, and

captures the large-scale fading effect, including path loss and shadowing, assumed to be frequency independent. The interfering channel from the th V2V transmitter to the th V2V receiver over the th sub-band, , the interfering channel from the th V2V transmitter to the BS over the th sub-band, , the channel from the th V2I transmitter to the BS over the th sub-band, , and the interfering channel from the th V2I transmitter to the th V2V receiver over the th sub-band, , are similarly defined.

The received signal-to-interference-plus-noise ratios (SINRs) of the th V2I link and the th V2V link over the th sub-band are expressed as

(2)

and

(3)

respectively, where and denote transmit powers of the th V2I transmitter and the th V2V transmitter over the th sub-band, respectively, is the noise power, and

(4)

denotes the interference power. is the binary spectrum allocation indicator with implying the th V2V link uses the th sub-band and otherwise. We assume each V2V link only accesses one sub-band, i.e., .

Capacities of the th V2I link and the th V2V link over the th sub-band are then obtained as

(5)

and

(6)

where is the bandwidth of each spectrum sub-band.

As described earlier, the V2I links are designed to support mobile high data rate entertainment services and hence an appropriate design objective is to maximize their sum capacity, defined as , for smooth mobile broadband access. In the meantime, the V2V links are mainly responsible for reliable dissemination of safety-critical messages that are generated periodically with varying frequencies depending on vehicle mobility for advanced driving services. We mathematically model such a requirement as the delivery rate of packets of size within a time budget as

(7)

where denotes the size of the periodically generated V2V payload in bits, is channel coherence time, and the index is added in to indicate the capacity of the th V2V link at different coherence time slots.

To this end, the resource allocation problem investigated in this work is formally stated as: To design the V2V spectrum allocation, expressed through binary variables

for all , and the V2V transmission power, for all , to simultaneously maximize the sum capacity of all V2I links and the packet delivery rate of V2V links defined in (7).

High mobility in a vehicular environment precludes collection of accurate full CSI at a central controller, hence making distributed V2V resource allocation more preferable. Then how to coordinate actions of multiple V2V links such that they do not act selfishly in their own interests to compromise performance of the system as a whole remains challenging. In addition, the packet delivery rate for V2V links, defined in (7), involves sequential decision making across multiple coherence time slots within the time constraint and causes difficulties for conventional optimization methods due to exponential dimension increase. To address these issues, we will exploit latest findings from multi-agent RL to develop a distributed algorithm for V2V spectrum access in the next section.

Iii Multi-Agent RL Based Resource Allocation

In the resource sharing scenario illustrated in Fig. 1, multiple V2V links attempt to access limited spectrum occupied by V2I links, which can be modeled as a multi-agent RL problem. Each V2V link acts as an agent and interacts with the unknown communication environment to gain experiences, which are then used to direct its own policy design. Multiple V2V agents collectively explore the environment and refine spectrum allocation and power control strategies based on their own observations of the environment state. While the resource sharing problem may appear a competitive game, we turn it into a fully cooperative one through using the same reward for all agents, in the interest of global network performance.

The proposed multi-agent RL based approach is divided into two phases, i.e., the learning (training) and the implementation phases. We focus on settings with centralized learning and distributed implementation. This means in the learning phase, the system performance-oriented reward is readily accessible to each individual V2V agent, which then adjusts its actions toward an optimal policy through updating its deep Q-network (DQN). In the implementation phase, each V2V agent receives local observations of the environment and then selects an action according to its trained DQN on a time scale on par with the small-scale channel fading. Key elements of the multi-agent RL based resource sharing design are described below in detail.

Iii-a State and Observation Space

Fig. 2: The agent-environment interaction in multi-agent RL formulation of the investigated resource sharing in vehicular networks.

In the multi-agent RL formulation of the resource sharing problem, each V2V link acts as an agent, concurrently exploring the unknown environment [29, 30]. Mathematically, the problem can be modeled as an MDP. As shown in Fig. 2, at each coherence time step , given the current environment state , each V2V agent receives an observation of the environment, determined by the observation function as , and then takes an action , forming a joint action . Thereafter, the agent receives a reward and the environment evolves to the next state with probability . The new observations are then received by each agent. Please note that all V2V agents share the same reward in the system such that cooperative behavior among them is encouraged.

The true environment state, , which could include global channel conditions and all agents’ behaviors, is unknown to each individual V2V agent. Each V2V agent can only acquire knowledge of the underlying environment through the lens of an observation function. The observation space of an individual V2V agent contains local channel information, including its own signal channel, , for all , interference channels from other V2V transmitters, , for all and , the interference channel from its own transmitter to the BS, , for all , and the interference channel from all V2I transmitters, , for all . Such channel information, except

, can be accurately estimated by the receiver of the

th V2V link at the beginning of each time slot and we assume it is also available instantaneously at the transmitter through delay-free feedback [31]. The channel is estimated at the BS in each time slot and then broadcast to all vehicles in its coverage, which incurs small signaling overhead. The received interference power over all bands, , for all , expressed in (4), can be measured at the V2V receiver and is also introduced in the local observation. In addition, the local observation space includes the remaining V2V payload, , and the remaining time budget, , to better capture the queuing states of each V2V link. As a result, the observation function for an agent is summarized as

(8)

with .

Independent Q-learning [32] is among the most popular methods to solve multi-agent RL problems, where each agent learns a decentralized policy based on its own action and observation, treating other agents as part of the environment. However, naively combining DQN with independent Q-learning is problematic since each agent would face a nonstationary environment while other agents are also learning to adjust their behaviors. The issue grows even more severe with experience replay, which is the key to the success of DQN, in that sampled experiences no longer reflect current dynamics and thus destabilize learning. To address this issue, we adopt the fingerprint-based method developed in [30]. The idea is that while the action-value function of an agent is nonstationary with other agents changing their behaviors over time, it can be made stationary conditioned on other agents’ policies. This means we can augment each agent’s observation space with an estimate of other agents’ policies to avoid nonstationarity, which is the essential idea of hyper Q-learning [33]

. However, it is undesirable for the action-value function to include as input all parameters of other agents’ neural networks,

, since the policy of each agent consists of a high dimensional DQN. Instead, it is proposed in [30] to simply include a low-dimensional fingerprint that tracks the trajectory of the policy change of other agents. This method works since nonstationarity of the action-value function results from changes of other agents’ policies over time, as opposed to the policies themselves. Further analysis reveals that each agent’s policy change is highly correlated with the training iteration number as well as its rate of exploration, e.g., the probability of random action selection, , in the -greedy policy widely used in Q-learning. Therefore, we include both of them in the observation for an agent , expressed as

(9)

Iii-B Action Space

The resource sharing design of vehicular links comes down to the spectrum sub-band selection and transmission power control for V2V links. While the spectrum naturally breaks into disjoint sub-bands, each preoccupied by one V2I link, the V2V transmission power typically takes continuous value in most existing power control literature. In this paper, however, we limit the power control options to four levels, i.e., dBm, for the sake of both ease of learning and practical circuit restriction. It is noted that the choice of dBm effectively means zero V2V transmission power. As a result, the dimension of the action space is , with each action corresponding to one particular combination of a spectrum sub-band and power selection.

Iii-C Reward Design

What makes RL particularly appealing for solving problems with hard-to-optimize objectives is the flexibility in its reward design. The system performance can be improved when the designed reward signal at each step correlates with the desired objective. In the investigated V2X spectrum sharing problem described in Section II, our objectives are twofold: Maximizing the sum V2I capacity while increasing the success probability of V2V payload delivery within a certain time constraint .

In response to the first objective, we simply include the instantaneous sum capacity of all V2I links, as defined in (5), in the reward at each time step . To achieve the second objective, for each agent , we set the reward equal to the effective V2V transmission rate until the payload is delivered, after which the reward is set to a constant number, , that is greater than the largest possible V2V transmission rate. As such, the V2V-related reward at each time step is set as

(10)

The goal of learning is to find an optimal policy (a mapping from states in to probabilities of selecting each action in ) that maximizes the expected return from any initial state , where the return, denoted by , is defined as the cumulative discounted rewards with a discount rate , i.e.,

(11)

We observe that if setting the discount rate to 1, greater cumulative rewards translate to a larger amount of transmitted data for V2V links until the payload delivery is finished. Hence maximizing the expected cumulative rewards encourages more data to be delivered for V2V links when the remaining payload is still nonzero, i.e., . In addition, the learning process also attempts to achieve as many rewards of as possible, effectively leading to higher possibility of successful delivery of V2V payload.

In practice,

is a hyperparameter that needs to be tuned empirically.

In our training, is tuned such that it is greater than the largest V2V transmission rate that is obtained from running a few steps of random resource allocation, but should not be “too big” and ideally less than twice of the largest value from our tuning experience. The design of represents our thinking about the tradeoff between designing the reward purely toward the ultimate goal and learning efficiency. For pure goal-directed consideration, we just set the rewards at each step to until the V2V payload is delivered, beyond which point the reward is set to . However, our tuning experience suggests such a design will hinder the learning process since the agent can hardly learn anything useful at the beginning of each episode as it always receives a reward of for this period. We then impart some prior knowledge into the reward, i.e., higher V2V transmission rates should be helpful in improving V2V payload delivery rate. Hence, we come up with the reward design described in (10) to blend this two extremes of reward designs.

To this end, we set the reward at each time step as

(12)

where and are positive weights to balance V2I and V2V objectives.

Iii-D Learning Algorithm

1:  Start environment simulator, generating vehicles and links
2:  Initialize Q-networks for all agents randomly
3:  for each episode do
4:     Update vehicle locations and large-scale fading
5:     Reset and , for all
6:     for each step  do
7:        for each V2V agent  do
8:           Observe
9:           Choose action from according to -greedy policy
10:        end for
11:        All agents take actions and receive reward
12:        Update channel small-scale fading
13:        for each V2V agent  do
14:           Observe
15:           Store in the replay memory
16:        end for
17:     end for
18:     for each V2V agent  do
19:        Uniformly sample mini-batches from
20:        Optimize error between Q-network and learning targets, defined in (14

), using variant of stochastic gradient descent

21:     end for
22:  end for
Algorithm 1 Resource Sharing with Multi-Agent RL

We focus on an episodic setting with each episode spanning the V2V payload delivery time constraint . Each episode starts with a randomly initialized environment state (determined by the initial transmission powers of all vehicular links, channel states, etc.) and a full V2V payload of size for transmission, and lasts until the end of . The change of small-scale channel fading triggers a transition of the environment state and causes each individual V2V agent to adjust its action.

Iii-D1 Training Procedure

We leverage deep Q-learning with experience replay [7] to train the multiple V2V agents for effective learning of spectrum access policies. Q-Learning [34] is based on the concept of action-value function, , for policy , which is defined as the expected return starting from the state , taking the action , and thereafter following the policy , formally expressed as

(13)

where is defined in (11). It is easy to determine the optimal policy once its action-value function, , is obtained. It has been shown in [6] that with a variant of the stochastic approximation condition on the learning rate and the assumption that all state-action pairs continue to be updated, the learned action-value function in Q-learning converges with probability to the optimal . In deep Q-learning [7], a deep neural network parameterized by , called DQN, is used to represent the action-value function.

Each V2V agent has a dedicated DQN that takes as input the current observation and outputs the value functions corresponding to all actions. We train the Q-networks through running multiple episodes and, at each training step, all V2V agents explore the state-action space with some soft policies, e.g., -greedy, meaning that the action with maximal estimated value is chosen with probability while a random action is instead selected with probability . Following the environment transition due to channel evolution and actions taken by all V2V agents, each agent collects and stores the transition tuple, , in a replay memory. At each episode, a mini-batch of experiences are uniformly sampled from the memory for updating with variants of stochastic gradient-descent methods, hence the name experience replay, to minimize the sum-squared error:

(14)

where is the parameter set of a target Q-network, which are duplicated from the training Q-network parameter set periodically and fixed for a couple of updates. Experience replay improves sample efficiency through repeatedly sampling stored experiences and breaks correlation in successive updates, thus also stabilizing learning. The training procedure is summarized in Algorithm 1.

Iii-D2 Distributed Implementation

During the implementation phase, at each time step , each V2V agent estimates local channels and compiles a local observation, , of the environment based on (9) with and set to the values from the very last training step. Then it selects an action, , with the maximum action value according to its trained Q-network. Afterwards, all V2V links start transmission with the power level and frequency spectrum sub-band determined by their selected actions.

Note that the computation intensive training procedure in Algorithm 1 can be performed offline for many episodes over different channel conditions and network topology changes while the inexpensive implementation procedure is executed online for network deployment. The trained DQNs for all agents only need to be updated when the environment characteristics have experienced significant changes, say, once a week or even a month, depending on environment dynamics and network performance requirements.

Iv Simulation Results

Parameter Value
Number of V2I links 4
Number of V2V links 4
Carrier frequency 2 GHz
Bandwidth 4 MHz
BS antenna height 25 m
BS antenna gain 8 dBi
BS receiver noise figure 5 dB
Vehicle antenna height 1.5 m
Vehicle antenna gain 3 dBi
Vehicle receiver noise figure 9 dB
Absolute vehicle speed 36 km/h
Vehicle drop and mobility model Urban case of A.1.2 in [3]*
V2I transmit power 23 dBm
V2V transmit power [23,10,5,-100] dBm
Noise power -114 dBm
Time constraint of V2V payload transmission 100 ms
V2V payload size 1060 bytes
  • We shrink the height and width of the simulation area by a factor of 2.

TABLE I: Simulation Parameters [3, 35]
Parameter V2I Link V2V Link
Path loss model 128.1 + 37.6, in km LOS in WINNER + B1 Manhattan [36]
Shadowing distribution Log-normal Log-normal

Shadowing standard deviation

8 dB 3 dB
Decorrelation distance 50 m 10 m
Path loss and shadowing update A.1.4 in [3] every 100 ms A.1.4 in [3] every 100 ms
Fast fading Rayleigh fading Rayleigh fading
Fast fading update Every 1 ms Every 1 ms
TABLE II: Channel Models for V2I and V2V Links [3]

In this section, simulation results are presented to validate the proposed multi-agent RL based resource sharing scheme for vehicular networks. We custom built our simulator following the evaluation methodology for the urban case defined in Annex A of 3GPP TR 36.885 [3], which describes in detail vehicle drop models, densities, speeds, direction of movement, vehicular channels, V2V data traffic, etc.. The V2I links are started by vehicles and the V2V links are formed between each vehicle with its surrounding neighbors. Major simulation parameters are listed in Table I and the channel models for V2I and V2V links are described in Table II. Note that all parameters are set to the values specified in Tables I and II by default, whereas the settings in each figure take precedence wherever applicable.

The DQN for each V2V agent consists of fully connected hidden layers, containing , and neurons, respectively. The rectified linear unit (ReLU),

, is used as the activation function and RMSProp optimizer

[37] is used to update network parameters with a learning rate of . We train each agent’s Q-network for a total of episodes and the exploration rate is linearly annealed from to over the beginning episodes and remains constant afterwards. It is noted that we fix the large-scale fading for a couple of training episodes and let the small-scale fading change over each step such that the learning algorithm can better acquire the underlying fading dynamics, thus helping stabilise training. In addition, we fix the V2V payload size in the training stage to be of bytes, but vary the sizes in the testing stage to verify robustness of the proposed method.

We compare in Figs. 3 and 4 the proposed multi-agent RL based resource sharing scheme, termed MARL, against the following two baseline methods that are executed in a distributed manner.

  1. The single-agent RL based algorithm in [12]

    , termed SARL, where at each moment only one V2V agent updates its action, i.e., spectrum sub-band selection and power control, based on locally acquired information and a trained DQN while others agents’ actions remain unchanged. A single DQN is shared across all V2V agents.

  2. The random baseline, which chooses the spectrum sub-band and transmission power for each V2V link in a random fashion at each time step.

We further benchmark the proposed MARL method in Algorithm 1 against the theoretical performance upper bounds of the V2I and V2V links, derived from the following two idealistic (and extreme) schemes.

  1. We disable the transmission of all V2V links to obtain the upper bound of V2I performance, hence the name upper bound without V2V. In this case, the packet delivery rates for all V2V links are exactly zero, thus not shown in Fig. 4.

  2. We exclusively focus on improving V2V performance while ignoring the requirement of V2I links. Such an assumption breaks the sequential decision making of delivering bytes over multiple steps within the time constraint into separate optimization of sum V2V rates over each step. Then, we exhaustively search the action space of all V2V agents in each step to maximize sum V2V rates. Apart from the complexity due to exhaustive search, this scheme needs to be performed in a centralized way with accurate global CSI available, hence the name centralized maxV2V.

We remark that although these two schemes are way too idealistic and cannot be implemented in practice, they provide meaningful performance upper bounds for V2I and V2V links that illustrate how closely the proposed method can approach the limit.

Fig. 3: Sum capacity performance of V2I links with varying V2V payload sizes .
Fig. 4: V2V payload transmission success probability with varying payload sizes .

Fig. 3 shows the V2I performance with respect to increasing V2V payload sizes for different resource sharing designs. From the figure, the performance drops for all schemes (except the upper bound) with growing V2V payload sizes. An increase of V2V payload leads to longer V2V transmission duration and possibly higher V2V transmit power in order to improve V2V payload transmission success probability. This will inevitably cause stronger interference to V2I links for a longer period and thus jeopardize their capacity performance. We observe that the proposed MARL method in Algorithm 1 achieves better performance than the other two baseline schemes across different V2V payload sizes although it is trained with a fixed size of bytes, demonstrating its robustness against V2V payload variation. It performs measurably close to the V2I performance upper bound, within degradation even in the worst case of bytes of payload. We also note that the centralized maxV2V scheme attains remarkable performance in terms of V2I performance. This could be due to the packet delivery rates of V2V links have been substantially enhanced with centralized maxV2V and the V2V links incur no interference to V2I links once their payload delivery has finished. This is an interesting observation that warrants further investigation into the performance tradeoff between V2I and V2V links. That said, the proposed distributed MARL method tightly follows the idealistic centralized maxV2V scheme, further demonstrating its effectiveness.

Fig. 4 shows the success probability of V2V payload delivery against growing payload sizes under different spectrum sharing schemes. From the figure, as the V2V payload size grows larger, the transmission success probabilities drop for all three distributed algorithms, including the proposed MARL, while the centralized maxV2V can achieve packet delivery throughout the tested cases. The proposed MARL method achieves significantly better performance than the two baseline distributed methods and stays very close to the centralized maxV2V scheme. Remarkably, the proposed method attains V2V payload delivery probability for and bytes and achieves close to perfect performance for and bytes.

We also observe from Fig. 4 that the proposed MARL method achieves highly desirable V2V performance for the low payload cases and suffers from noticeable degradation when the payload size grows beyond bytes. In conjunction with the observations from Fig. 3, we conclude that the robustness of the proposed multi-agent RL based method against V2V payload variation should be taken with a grain of salt: Within a reasonable region of payload size change, the trained DQN is good, which, however, needs to be updated if the change grows beyond the acceptable margin. However, the exact range of such acceptable margin is difficult to be determined in general, which would depend on the actual system parameter settings. For the current setting, we can conclude that no noticeable performance loss is spotted when the packet size is no greater than bytes and to maintain a V2V delivery rate above , the packet size needs to be no larger than bytes. Again, such observations are based on the particular setting for the simulation and extra caution is needed when generalizing them. That said, we can still validate the advantage of the proposed spectrum access design since it outperforms the other two distributed baselines even in the untrained scenarios.

Fig. 5: Return for each training episode with increasing iterations. The V2V payload size bytes.

We show in Fig. 5 the cumulative rewards per training episode with increasing training iterations to study the convergence behavior of the proposed multi-agent RL method. From the figure, the cumulative rewards per episode improve as training continues, demonstrating the effectiveness of the proposed training algorithm. When the training episode approximately reaches , the performance gradually converges despite some fluctuations due to mobility-induced channel fading in vehicular environments. Based on such an observation, we train each agent’s Q-network for episodes when evaluating the performance of V2I and V2V links in Figs. 3 and 4, which should provide a safe convergence guarantee.

(a) The remaining payload of MARL.
(b) The remaining payload of the random baseline.
Fig. 6: The change of the remaining V2V payload of the proposed MARL and the random baseline resource sharing schemes within the time constraint ms. The initial payload size byte.

To understand why the proposed multi-agent RL based method achieves better performance compared with the random baseline, we select an episode in which the proposed method enables all V2V links to successfully deliver the payload of bytes while the random baseline fails. We plot in Fig. 6 the change of the remaining V2V payload within the time constraint, i.e., ms, for all V2V links. From Fig. 6(a), the V2V Link 4 finishes payload delivery early in the episode while the other three links end transmission roughly at the same time for the proposed multi-agent RL based method. For the random baseline, Fig. 6(b) shows that V2V Links 1 and 4 successfully deliver all payload early in the episode. V2V Link 3 also finishes payload transmission albeit much later in the episode while V2V Link 2 fails to deliver the required payload.

(a) V2V transmission rates of MARL.
(b) V2V transmission rates of the random baseline.
Fig. 7: V2V transmission rates of the proposed MARL and the random baseline resource allocation schemes within the same episode as Fig. 6. Only the results of the beginning ms are plotted for better presentation. The initial payload size bytes.

In Fig. 7, we further show the instantaneous rates of all V2V links under the two different resource allocation schemes at each step in the same episode as Fig. 6. Several valuable observations can be made from comparing Figs. 7(a) and (b) that demonstrate the effectiveness of the proposed method in encouraging cooperation among multiple V2V agents. From Fig. 7(a), with the proposed method, V2V Link 4 gets very high transmission rates at the beginning to finish transmission early such that the good channel condition of this link is fully exploited and no interference will be generated toward other links at later stages of the episode. V2V Link 1 keeps low transmission rates at first such that the vulnerable V2V Links 2 and 3 can get relatively good transmission rates to deliver payload, and then jumps to high data rates to deliver its own data when Links 2 and 3 almost finish transmission. Moreover, a closer examination of the rates of Links 2 and 3 reveals that the two links figure out a clever strategy to take turns to transmit such that both of their payloads can be delivered quickly. To summarize, the proposed multi-agent RL based method learns to leverage good channels of some V2V links and meanwhile provides protection for those with bad channel conditions. The success probability of V2V payload transmission is thus significantly improved. In contrast, Fig. 7(b) shows that the random baseline method fails to provide such protection for vulnerable V2V links, leading to high probability of failed payload delivery for them.

V Conclusion

In this paper, we have developed a distributed resource sharing scheme based on multi-agent RL for vehicular networks with multiple V2V links reusing the spectrum of V2I links. A fingerprint-based method has been exploited to address nonstationary issues of independent Q-learning for multi-agent RL problems when combined with DQN with experience replay. The proposed multi-agent RL based method is divided into a centralized training stage and a distributed implementation stage. We demonstrate that through such a mechanism, the proposed resource sharing scheme is effective in encouraging cooperation among V2V links to improve system level performance although decision making is performed locally at each V2V transmitter. Future work will include an in-depth analysis and comparison of the robustness of both single-agent and multi-agent RL based algorithms to gain better understanding on when the trained Q-networks need to be updated and how to efficiently perform such updates. Extension of the proposed multi-agent RL based resource allocation method to the multiple-input multiple-output (MIMO) and the millimeter MIMO scenarios for vehicular communications is also an interesting direction worth further investigation.

References

  • [1] L. Liang, H. Peng, G. Y. Li, and X. Shen, “Vehicular communications: A physical layer perspective,” IEEE Trans. Veh. Technol., vol. 66, no. 12, pp. 10 647–10 659, Dec. 2017.
  • [2] H. Peng, L. Liang, X. Shen, and G. Y. Li, “Vehicular communications: A network layer perspective,” IEEE Trans. Veh. Technol., vol. 2, no. 68, pp. 1064–1078, Feb. 2019.
  • [3] 3rd Generation Partnership Project; Technical Specification Group Radio Access Network; Study on LTE-based V2X Services; (Release 14), 3GPP TR 36.885 V14.0.0, Jun. 2016.
  • [4] 3rd Generation Partnership Project; Technical Specification Group Radio Access Network; Study on enhancement of 3GPP Support for 5G V2X Services; (Release 15), 3GPP TR 22.886 V15.1.0, Mar. 2017.
  • [5] R. Molina-Masegosa and J. Gozalvez, “LTE-V for sidelink 5G V2X vehicular communications: A new 5G technology for short-range vehicle-to-everything communications,” IEEE Veh. Technol. Mag., vol. 12, no. 4, pp. 30–39, Dec. 2017.
  • [6] R. S. Sutton, A. G. Barto et al., Reinforcement learning: An introduction.   MIT press, 1998.
  • [7] V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G. Bellemare, A. Graves, M. Riedmiller, A. K. Fidjeland, G. Ostrovski et al., “Human-level control through deep reinforcement learning,” Nature, vol. 518, no. 7540, pp. 529–533, Feb. 2015.
  • [8] D. Silver, A. Huang, C. J. Maddison, A. Guez, L. Sifre, G. Van Den Driessche, J. Schrittwieser, I. Antonoglou, V. Panneershelvam, M. Lanctot et al., “Mastering the game of go with deep neural networks and tree search,” Nature, vol. 529, no. 7587, pp. 484–489, Jan. 2016.
  • [9] Y. He, Z. Zhang, F. R. Yu, N. Zhao, H. Yin, V. C. M. Leung, and Y. Zhang, “Deep-reinforcement-learning-based optimization for cache-enabled opportunistic interference alignment wireless networks,” IEEE Trans. Veh. Technol., vol. 66, no. 11, pp. 10 433–10 445, Nov. 2017.
  • [10] Y. He, F. R. Yu, N. Zhao, and H. Yin, “Secure social networks in 5G systems with mobile edge computing, caching, and device-to-device communications,” IEEE Wireless Commun., vol. 25, no. 3, pp. 103–109, Jun. 2018.
  • [11] H. Mao, M. Alizadeh, I. Menache, and S. Kandula, “Resource management with deep reinforcement learning,” in Proc. ACM Workshop Hot Topics Netw. (HotNets), 2016, pp. 50–56.
  • [12] H. Ye, G. Y. Li, and B.-H. Juang, “Deep reinforcement learning based resource allocation for V2V communications,” IEEE Trans. Veh. Technol., vol. 68, no. 4, pp. 3163––3173, Apr. 2019.
  • [13] M. Botsov, M. Klügel, W. Kellerer, and P. Fertl, “Location dependent resource allocation for mobile device-to-device communications,” in Proc. IEEE WCNC, Apr. 2014, pp. 1679–1684.
  • [14] W. Sun, E. G. Ström, F. Brännström, K. Sou, and Y. Sui, “Radio resource management for D2D-based V2V communication,” IEEE Trans. Veh. Technol., vol. 65, no. 8, pp. 6636–6650, Aug. 2016.
  • [15] W. Sun, D. Yuan, E. G. Ström, and F. Brännström, “Cluster-based radio resource management for D2D-supported safety-critical V2X communications,” IEEE Trans. Wireless Commun., vol. 15, no. 4, pp. 2756–2769, Apr. 2016.
  • [16] M. I. Ashraf, C.-F. Liu, M. Bennis, and W. Saad, “Dynamic resource allocation for optimized latency and reliability in vehicular networks,” IEEE Access, vol. 6, pp. 63 843–63 858, Oct. 2018.
  • [17] L. Liang, G. Y. Li, and W. Xu, “Resource allocation for D2D-enabled vehicular communications,” IEEE Trans. Commun., vol. 65, no. 7, pp. 3186–3197, Jul. 2017.
  • [18] L. Liang, J. Kim, S. C. Jha, K. Sivanesan, and G. Y. Li, “Spectrum and power allocation for vehicular communications with delayed CSI feedback,” IEEE Wireless Comun. Lett., vol. 6, no. 4, pp. 458–461, Aug. 2017.
  • [19] L. Liang, S. Xie, G. Y. Li, Z. Ding, and X. Yu, “Graph-based resource sharing in vehicular communication,” IEEE Trans. Wireless Commun., vol. 17, no. 7, pp. 4579–4592, Jul. 2018.
  • [20]

    H. Ye, L. Liang, G. Y. Li, J. Kim, L. Lu, and M. Wu, “Machine learning for vehicular networks: Recent advances and application examples,”

    IEEE Veh. Technol. Mag., vol. 13, no. 2, pp. 94–101, Jun. 2018.
  • [21] L. Liang, H. Ye, and G. Y. Li, “Toward intelligent vehicular networks: A Machine Learning Framework,” IEEE Internet Things J., vol. 6, no. 1, pp. 124–135, Feb. 2019.
  • [22] M. A. Salahuddin, A. Al-Fuqaha, and M. Guizani, “Reinforcement learning for resource provisioning in the vehicular cloud,” IEEE Wireless Commun., vol. 23, no. 4, pp. 128–135, Jun. 2016.
  • [23] Q. Zheng, K. Zheng, H. Zhang, and V. C. M. Leung, “Delay-optimal virtualized radio resource scheduling in software-defined vehicular networks via stochastic learning,” IEEE Trans. Veh. Technol., vol. 65, no. 10, pp. 7857–7867, Oct. 2016.
  • [24] Y. He, N. Zhao, and H. Yin, “Integrated networking, caching and computing for connected vehicles: A deep reinforcement learning approach,” IEEE Trans. Veh. Technol., vol. 67, no. 1, pp. 44 – 55, Jan. 2018.
  • [25] R. Atallah, C. Assi, and J. Y. Yu, “A reinforcement learning technique for optimizing downlink scheduling in an energy-limited vehicular network,” IEEE Trans. Veh. Technol., vol. 66, no. 6, pp. 4592–4601, Jun. 2017.
  • [26] R. Atallah, C. Assi, and M. Khabbaz, “Deep reinforcement learning-based scheduling for roadside communication networks,” in Proc. IEEE WiOpt, May 2017, pp. 1–8.
  • [27] Z. Li, C. Wang, and C.-J. Jiang, “User association for load balancing in vehicular networks: An online reinforcement learning approach,” IEEE Trans. Intell. Transp. Syst., vol. 18, no. 8, pp. 2217–2228, Aug. 2017.
  • [28] Y. Xu, L. Li, B.-H. Soong, and C. Li, “Fuzzy Q-learning based vertical handoff control for vehicular heterogeneous wireless network,” in Proc. IEEE ICC, Jun. 2014, pp. 5653–5658.
  • [29] S. Omidshafiei, J. Pazis, C. Amato, J. P. How, and J. Vian, “Deep decentralized multi-task multi-agent reinforcement learning under partial observability,” in Proc. Int. Conf. Mach. Learning (ICML), 2017, pp. 2681–2690.
  • [30] J. Foerster, N. Nardelli, G. Farquhar, T. Afouras, P. H. S. T. 1, P. Kohli, and S. Whiteson, “Stabilising experience replay for deep multi-agent reinforcement learning,” in Proc. Int. Conf. Mach. Learning (ICML), 2017, pp. 1146–1155.
  • [31] Y. S. Nasir and D. Guo, “Deep reinforcement learning for distributed dynamic power allocation in wireless networks,” arXiv preprint arXiv:1808.00490, 2018.
  • [32] M. Tan, “Multi-agent reinforcement learning: Independent vs. cooperative agents,” in Proc. Int. Conf. Mach. Learning (ICML), 1993, pp. 330–337.
  • [33] G. Tesauro, “Extending Q-learning to general adaptive multi-agent systems,” in Proc. Advances Neural Inf. Process. Syst. (NIPS), 2004, pp. 871–878.
  • [34] C. J. Watkins and P. Dayan, “Q-learning,” Machine Learning, vol. 8, no. 3-4, pp. 279–292, May 1992.
  • [35] R1-165704, WF on SLS evaluation assumptions for eV2X, 3GPP TSG RAN WG1 Meeting #85, May 2016.
  • [36] WINNER II Channel Models, IST-4-027756 WINNER II D1.1.2 V1.2, Sep. 2007. [Online]. Available: http://projects.celtic-initiative.org/winner+/WINNER2-Deliverables/D1.1%.2v1.1.pdf.
  • [37] S. Ruder, “An overview of gradient descent optimization algorithms,” arXiv preprint arXiv:1609.04747, 2016.