Channel Allocation in Cellular Telephone. Systems. Lab. for Info. and Decision Sciences. Cambridge, MA 02139. bertsekas@lids.mit.edu.



Similar documents
Call Admission Control and Routing in Integrated Service Networks Using Reinforcement Learning

TD(0) Leads to Better Policies than Approximate Value Iteration

Introduction to Logistic Regression

System Design in Wireless Communication. Ali Khawaja

constraint. Let us penalize ourselves for making the constraint too big. We end up with a

Ulrich A. Muller UAM June 28, 1995

How To Maximize Admission Control On A Network

Scheduling using Optimization Decomposition in Wireless Network with Time Performance Analysis

Options with exceptions

Neuro-Dynamic Programming An Overview

A Sarsa based Autonomous Stock Trading Agent

Adaptive Radio Resource Management for GSM/GPRS Networks

Competitive Analysis of On line Randomized Call Control in Cellular Networks

NEURAL NETWORKS AND REINFORCEMENT LEARNING. Abhijit Gosavi

CURRENT wireless personal communication systems are

Revenue Management for Transportation Problems

Motivation. Motivation. Can a software agent learn to play Backgammon by itself? Machine Learning. Reinforcement Learning

Architecture bits. (Chromosome) (Evolved chromosome) Downloading. Downloading PLD. GA operation Architecture bits

8. Cellular Systems. 1. Bell System Technical Journal, Vol. 58, no. 1, Jan R. Steele, Mobile Communications, Pentech House, 1992.

Module 5. Broadcast Communication Networks. Version 2 CSE IIT, Kharagpur

On Frequency Assignment in Cellular Networks

Memory Allocation Technique for Segregated Free List Based on Genetic Algorithm


A Tool for Generating Partition Schedules of Multiprocessor Systems

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION

Eligibility Traces. Suggested reading: Contents: Chapter 7 in R. S. Sutton, A. G. Barto: Reinforcement Learning: An Introduction MIT Press, 1998.

An Ecient Dynamic Load Balancing using the Dimension Exchange. Ju-wook Jang. of balancing load among processors, most of the realworld

Beyond Reward: The Problem of Knowledge and Data

Oscillations of the Sending Window in Compound TCP

Bidding Dynamics of Rational Advertisers in Sponsored Search Auctions on the Web

Network Model. University of Tsukuba. of the system. Load balancing policies are often. used for balancing the workload of distributed systems.

Partial Path Protection for WDM Networks: End-to-End Recovery Using Local Failure Information

An Environment Model for N onstationary Reinforcement Learning

Making Sense of the Mayhem: Machine Learning and March Madness

Channel assignment for GSM half-rate and full-rate traffic

Packet Queueing Delay in Wireless Networks with Multiple Base Stations and Cellular Frequency Reuse

Relative Analysis of Human Behavior and Network Marketing

1 Location Management in Cellular

Compact Representations and Approximations for Compuation in Games

Inductive QoS Packet Scheduling for Adaptive Dynamic Networks

Dynamic Load Balance Algorithm (DLBA) for IEEE Wireless LAN

Learning Tetris Using the Noisy Cross-Entropy Method

Temporal Difference Learning in the Tetris Game

6.231 Dynamic Programming and Stochastic Control Fall 2008

Liski and Montero As a Successful Solution to the Recession

1 Example of Time Series Analysis by SSA 1

Congestion-Dependent Pricing of Network Services

PERFORMANCE AND EFFICIENCY EVALUATION OF CHANNEL ALLOCATION SCHEMES FOR HSCSD IN GSM

arxiv: v1 [cs.ni] 2 Dec 2013

Simple Channel-Change Games for Spectrum- Agile Wireless Networks

Complexity Theory. IE 661: Scheduling Theory Fall 2003 Satyaki Ghosh Dastidar

A Novel Decentralized Time Slot Allocation Algorithm in Dynamic TDD System

Lecture 1. Introduction to Wireless Communications 1


Performance of TD-CDMA systems during crossed slots

Dynamic Reconfiguration & Efficient Resource Allocation for Indoor Broadband Wireless Networks

Load Balancing in Cellular Networks with User-in-the-loop: A Spatial Traffic Shaping Approach

Visibility optimization for data visualization: A Survey of Issues and Techniques

Introduction to Natural Computation. Lecture 15. Fruitflies for Frequency Assignment. Alberto Moraglio

Clustering and scheduling maintenance tasks over time

Weiping Zhu C.F. Steketee. processes into account. to the potential performance gain from this service.

Experiments on the local load balancing algorithms; part 1

Enhancement of QoS in Mobile Network through Channel Allocation using Software Agents

Approximation Algorithms

On closed-form solutions of a resource allocation problem in parallel funding of R&D projects

Flexible mobility management strategy in cellular networks

A Brief Study of the Nurse Scheduling Problem (NSP)

How performance metrics depend on the traffic demand in large cellular networks

Finite Mathematics Using Microsoft Excel

Jan F. Prins. Work-efficient Techniques for the Parallel Execution of Sparse Grid-based Computations TR91-042

The Time Value of Money

VOICE OVER WI-FI CAPACITY PLANNING

Basic Network Design

Load balancing in a heterogeneous computer system by self-organizing Kohonen network

Capacity planning and.

Workload Generation for ns. Simulations of Wide Area Networks

Dynamic programming formulation

Attenuation (amplitude of the wave loses strength thereby the signal power) Refraction Reflection Shadowing Scattering Diffraction

IN THIS PAPER, we study the delay and capacity trade-offs

EE4367 Telecom. Switching & Transmission. Prof. Murat Torlak

On the Traffic Capacity of Cellular Data Networks. 1 Introduction. T. Bonald 1,2, A. Proutière 1,2

A Randomized Saturation Degree Heuristic for Channel Assignment in Cellular Radio Networks

Subscriber Maximization in CDMA Cellular Networks

On-Line Learning Control by Association and Reinforcement

Intelligent Agents Serving Based On The Society Information

ENSC 427: Communication Networks. Analysis of Voice over IP performance on Wi-Fi networks


Path Selection Methods for Localized Quality of Service Routing

RESEARCH PAPER International Journal of Recent Trends in Engineering, Vol 1, No. 1, May 2009

Scheduling Video Stream Transmissions for Distributed Playback over Mobile Cellular Networks

Jae Won idee. School of Computer Science and Engineering Sungshin Women's University Seoul, , South Korea

Support Vector Machines with Clustering for Training with Very Large Datasets

Continuous Fastest Path Planning in Road Networks by Mining Real-Time Traffic Event Information

Winning the Kaggle Algorithmic Trading Challenge with the Composition of Many Models and Feature Engineering

AN ADAPTIVE DISTRIBUTED LOAD BALANCING TECHNIQUE FOR CLOUD COMPUTING


Grid Computing Approach for Dynamic Load Balancing

A Learning Based Method for Super-Resolution of Low Resolution Images

4-5 Limiting the Holding Time in Mobile Cellular Phone Systems During Disasters

In Proceedings of the First IEEE Workshop on Real-Time Applications, New York, NY, May 1993.

Transcription:

Reinforcement Learning for Dynamic Channel Allocation in Cellular Telephone Systems Satinder Singh Department of Computer Science University of Colorado Boulder, CO 80309-0430 baveja@cs.colorado.edu Dimitri Bertsekas Lab. for Info. and Decision Sciences MIT Cambridge, MA 02139 bertsekas@lids.mit.edu Abstract In cellular telephone systems, an important problem is to dynamically allocate the communication resource (channels) so as to maximize service in a stochastic caller environment. This problem is naturally formulated as a dynamic programming problem and we use a reinforcement learning () method to nd dynamic channel allocation policies that are better than previous heuristic solutions. The policies obtained perform well for a broad variety of call traf- c patterns. We present results on a large cellular system with approximately 49 49 states. In cellular communication systems, an important problem is to allocate the communication resource (bandwidth) so as to maximize the service provided to a set of mobile callers whose demand for service changes stochastically. A given geographical area is divided into mutually disjoint cells, and each cell serves the calls that are within its boundaries (see Figure 1a). The total system bandwidth is divided into channels, with each channel centered around a frequency. Each channel can be used simultaneously at dierent cells, provided these cells are suciently separated spatially, so that there is no interference between them. The minimum separation distance between simultaneous reuse of the same channel is called the channel reuse constraint. When a call requests service in a given cell either a free channel (one that does not violate the channel reuse constraint) may be assigned to the call, or else the call is blocked from the system; this will happen if no free channel can be found. Also, when a mobile caller crosses from one cell to another, the call is \handed o" to the cell of entry; that is, a new free channel is provided to the call at the new cell. If no such channel is available, the call must be dropped/disconnected from the system.

One objective of a channel allocation policy is to allocate the available channels to calls so that the number of blocked calls is minimized. An additional objective is to minimize the number of calls that are dropped when they are handed o to a busy cell. These two objectives must be weighted appropriately to reect their relative importance, since dropping existing calls is generally more undesirable than blocking new calls. To illustrate the qualitative nature of the channel assignment decisions, suppose that there are only two channels and three cells arranged in a line. Assume a channel reuse constraint of 2, i.e., a channel may be used simultaneously in cells 1 and 3, but may not be used in channel 2 if it is already used in cell 1 or in cell 3. Suppose that the system is serving one call in cell 1 and another call in cell 3. Then serving both calls on the same channel results in a better channel usage pattern than serving them on dierent channels, since in the former case the other channel is free to be used in cell 2. The purpose of the channel assignment and channel rearrangement strategy is, roughly speaking, to create such favorable usage patterns that minimize the likelihood of calls being blocked. We formulate the channel assignment problem as a dynamic programming problem, which, however, is too complex to be solved exactly. We introduce approximations based on the methodology of reinforcement learning () (e.g., Barto, Bradtke and Singh, 1995, or the recent textbook by Bertsekas and Tsitsiklis, 1996). Our method learns channel allocation policies that outperform not only the most commonly used policy in cellular systems, but also the best heuristic policy we could nd in the literature. 1 CHANNEL ASSIGNMENT POLICIES Many cellular systems are based on a xed assignment () channel allocation; that is, the set of channels is partitioned, and the partitions are permanently assigned to cells so that all cells can use all the channels assigned to them simultaneously without interference (see Figure 1a). When a call arrives in a cell, if any preassigned channel is unused; it is assigned, else the call is blocked. No rearrangement is done when a call terminates. Such a policy is static and cannot take advantage of temporary stochastic variations in demand for service. More ecient are dynamic channel allocation policies, which assign channels to dierent cells, so that every channel is available to every cell on a need basis, unless the channel is used in a nearby cell and the reuse constraint is violated. The best existing dynamic channel allocation policy we found in the literature is Borrowing with Directional Channel Locking () of Zhang & Yum (1989). It numbers the channels from 1 to N, partitions and assigns them to cells as in. The channels assigned to a cell are its nominal channels. If a nominal channel is available when a call arrives in a cell, the smallest numbered such channel is assigned to the call. If no nominal channel is available, then the largest numbered free channel is borrowed from the neighbour with the most free channels. When a channel is borrowed, careful accounting of the directional eect of which cells can no longer use that channel because of interference is done. The call is blocked if there are no free channels at all. When a call terminates in a cell and the channel so freed is a nominal channel, say numbered i, of that cell, then if there is a call in that cell on a borrowed channel, the call on the smallest numbered borrowed channel is reassigned to i and the borrowed channel is returned to the appropriate cell. If there is no call on a borrowed channel, then if there is a call on a nominal channel numbered larger than i, the call on the highest numbered nominal channel is reassigned to i. If the call just terminated was itself on a borrowed channel, the

call on the smallest numbered borrowed channel is reassigned to it and that channel is returned to the cell from which it was borrowed. Notice that when a borrowed channel is returned to its original cell, a nominal channel becomes free in that cell and triggers a reassignment. Thus, in the worst case a call termination in one cell can sequentially cause reassignments in arbitrarily far away cells making somewhat impractical. is quite sophisticated and combines the notions of channel-ordering, nominal channels, and channel borrowing. Zhang and Yum (1989) show that is superior to its competitors, including. Generally, has continued to be highly regarded in the literature as a powerful heuristic (Enrico et.al., 1996). In this paper, we compare the performance of dynamic channel allocation policies learned by with both and. 1.1 DYNAMIC PROGRAMMING FORMULATION We can formulate the dynamic channel allocation problem using dynamic programming (e.g., Bertsekas, 1995). State transitions occur when channels become free due to call departures, or when a call arrives at a given cell and wishes to be assigned achannel, or when there is a hando, which can be viewed as a simultaneous call departure from one cell and a call arrival at another cell. The state at each time consists of two components: (1) The list of occupied and unoccupied channels at each cell. We call this the conguration of the cellular system. It is exponential in the number of cells. (2) The event that causes the state transition (arrival, departure, or hando). This component of the state is uncontrollable. The decision/control applied at the time of a call departure is the rearrangement of the channels in use with the aim of creating a more favorable channel packing pattern among the cells (one that will leave more channels free for future assignments). Unlike, our solution will restrict this rearrangement to the cell with the current call departure. The control exercised at the time of a call arrival is the assignment of a free channel, or the blocking of the call if no free channel is currently available. In general, it may also be useful to do admission control, i.e., to allow the possibility of not accepting a new call even when there exists a free channel to minimize the dropping of ongoing calls during hando in the future. We address admission control in a separate paper and here restrict ourselves to always accepting a call if a free channel is available. The objective is to learn a policy that assigns decisions (assignment or rearrangement depending on event) to each state so as to maximize J = EZ 1 0 e,t c(t)dt where Efg is the expectation operator, c(t) is the number of ongoing calls at time t, and is a discount factor that makes immediate prot more valuable than future prot. Maximizing J is equivalent to minimizing the expected (discounted) number of blocked calls over an innite horizon. 2 REINFORCEMENT LEARNING SOLUTION methods solve optimal control (or dynamic programming) problems by learning good approximations to the optimal value function, J, given by the solution to ;

the Bellman optimality equation which takes the following form for the dynamic channel allocation problem: J(x) = E e max a2a(x;e) [E t fc(x; a; t)+(t)j(y)g] ; (1) where x is a conguration, e is the random event (a call arrival or departure), A(x; e) is the set of actions available in the current state (x; e), t is the random time until the next event, c(x; a; t) is the eective immediate payo with the discounting, and (t) is the eective discount for the next conguration y. learns approximations to J using Sutton's (1988) temporal dierence (TD(0)) algorithm. A xed feature extractor is used to form an approximate compact representation of the exponential conguration of the cellular array. This approximate representation forms the input to a function approximator (see Figure 1) that learns/stores estimates of J. No partitioning of channels is done; all channels are available in each cell. On each event, the estimates of J are used both to make decisions and to update the estimates themselves as follows: Call Arrival: When a call arrives, evaluate the next conguration for each free channel and assign the channel that leads to the conguration with the largest estimated value. If there is no free channel at all, no decision has to be made. Call Termination: When a call terminates, one by one each ongoing call in that cell is considered for reassignment to the just freed channel; the resulting congurations are evaluated and compared to the value of not doing any reassignment at all. The action that leads to the highest value conguration is then executed. On call arrival, as long as there is a free channel, the number of ongoing calls and the time to next event do not depend on the free channel assigned. Similarly, the number of ongoing calls and the time to next event do not depend on the rearrangement done on call departure. Therefore, both the sample immediate payo which depends on the number of ongoing calls and the time to next event, and the eective discount factor which depends only on the time to next event are independent of the choice of action. Thus one can choose the current best action by simply considering the estimated values of the next congurations. The next conguration for each action is deterministic and trivial to compute. When the next random event occurs, the sample payo and the discount factor become available and are used to update the value function as follows: on a transition from conguration x to y on action a in time t, J new (~x) = (1, )J old (~x)+ (c(x; a; t)+(t)j old (~y)) (2) where ~x is used to indicate the approximate feature-based representation of x. The parameters of the function approximator are then updated to best represent J new (~x) using gradient descent in mean-squared error (J new (~x), J old (~x)) 2. 3 SIMULATION RESULTS Call arrivals are modeled as Poisson processes with a separate mean for each cell, and call durations are modeled with an exponential distribution. The rst set of results are on the 7 by 7 cellular array of Figure??a with 70 channels (roughly 70 49 congurations) and a channel reuse constraint of 3 (this problem is borrowed from Zhang and Yum's (1989) paper on an empirical comparison of and its competitors). Figures 2a, b & c are for uniform call arrival rates of 150, 200, and 350 calls/hr respectively in each cell. The mean call duration for all the experiments

reported here is 3 minutes. Figure 2d is for non-uniform call arrival rates. Each curve plots the cumulative empirical blocking probability as a function of simulated time. Each data point is therefore the percentage of system-wide calls that were blocked up until that point in time. All simulations start with no ongoing calls. Base Station Handoffs a) 6 b) 2 1 1 5 3 5 3 4 7 4 7 6 7 6 2 1 2 1 5 1 5 3 4 3 4 7 4 7 6 2 6 2 1 2 1 5 3 5 3 4 4 7 Configuration Feature Extractor Features (Availability and Packing ) TD(0) Training Function Value Approximator Figure 1: a) Cellular Array. The market area is divided up into cells, shown here as hexagons. The available bandwidth is divided into channels. Each cell has a base station responsible for calls within its area. Calls arrive randomly, have random durations and callers may move around in the market area creating handos. The channel reuse constraint requires that there be a minimum distance between simultaneous reuse of the same channel. In a xed assignment channel allocation policy (assuming a channel reuse constraint of 3), the channels are partitioned into 7 lots labeled 1 to 7 and assigned to the cells in the compact pattern shown here. Note that the minimum distance between cells with the same number is at least three. b) A block diagram of the system. The exponential conguration is mapped into a feature-based approximate representation which forms the input to a function approximation system that learns values for congurations. The parameters of the function approximator are trained using gradient descent on the squared TD(0) error in value function estimates (c.f. Equation 2). The system uses a linear neural network and two sets of features as input: one availability feature for each cell and one packing feature for each cell-channel pair. The availability feature for a cell is the number of free channels in that cell, while the packing feature for a cell-channel pair is the number of times that channel is used in a 4 cell radius. Other packing features were tried but are not reported because they were insignicant. The curves in Figure 2 show the empirical blocking probability whilst learning. Note that learning is quite rapid. As the mean call arrival rate is increased the relative dierence between the 3 algorithms decreases. In fact, can be shown to be optimal in the limit of innite call arrival rates (see McEliece and Sivarajan, 1994). With so many customers in every cell there are no short-term uctuations to exploit. However, as demonstrated in Figure 2, for practical trac rates consistently gives a big win over and a smaller win over. One dierence between and is that while the policy is independent of call trac, adapts its policy to the particulars of the call trac it is trained on and should therefore be less sensitive to dierent patterns of non-uniformity of call trac across cells. Figure 3b presents multiple sets of bar-graphs of asymptotic blocking probabilities for the three algorithms on a 20 by 1 cellular array with 24 channels and a channel reuse constraint of 3. For each set, the average per-cell call arrival rate is the same (120 calls/hr; mean duration of 3 minutes), but the pattern of call arrival rates across the 20 cells is varied. The patterns are shown on the left of the bar-graphs and are explained in the caption of Figure 3b. From Figure 3b it is apparent that is much less sensitive to varied patterns of non-uniformity than both and. We have showed that with a linear function approximator is able to nd better dynamic channel allocation policies than the and policies. Using nonlinear neural networks as function approximators for did in some cases improve

150 calls/hr a) b) 0.05 0.20 200 calls/hr c) d) 0.50 350 calls/hr Non Uniform 0.40 0.30 0.20 0.20 Figure 2: a), b), c) & d) These gures compare performance of,, and on the 7 by 7 cellular array of Figure 1a. The means of the call arrival (Poisson) processes are shown in the graph titles. Each curve presents the cumulative empirical blocking probability as a function of time elapsed in minutes. All simulations start with no ongoing calls and therefore the blocking probabilities are low in the early minutes of the performance curves. The curves presented here are for a linear function approximator and show performance while learning. Note that learning is quite rapid. performance over linear networks by a small amount but at the cost of a big increase in training time. We chose to present results for linear networks because they have the advantage that even though training is centralized, the policy so learned is decentralized because the features are local and therefore just the weights from the local features in the trained linear network can be used to choose actions in each cell. For large cellular arrays, training itself could be decentralized because the choice of action in a particular cell has a minor eect on far away cells. We will explore the eect of decentralized training in future work. 4 CONCLUSION The dynamic channel allocation problem is naturally formulated as an optimal control or dynamic programming problem, albeit one with very large state spaces. Traditional dynamic programming techniques are computationally infeasible for such large-scale problems. Therefore, knowledge-intensive heuristic solutions that ignore the optimal control framework have been developed. Recent approximations to dynamic programming introduced in the reinforcement learning () community make it possible to go back to the channel assignment problem and solve it as an optimal control problem, in the process nding better solutions than previously available. We presented such a solution using Sutton's (1988) TD(0) with a feature-based linear network and demonstrated its superiority on a problem with approximately 70 49 states. Other recent examples of similar successes are the game

a) b) Patterns (mmm...m) (lhlhlhlhlhlhlhlhlhlh) (llhllhllhllhllhllhll) (hhlhhlhhlhhlhhlhhlhh) (lhllhllhllhllhllhllh) (hlhhlhhlhhlhhlhhlhhl) 10% 20% 30% 40% Empirical Blocking Probability Figure 3: a) Screen dump of a Java Demonstration available publicly at http:// www.cs.colorado.edu/ebaveja/demo.html b) Sensitivity ofchannel assignment methods to non-uniform trac patterns. This gure plots asymptotic empirical blocking probability for,, and for a linear array of cells with dierent patterns (shown at left) of mean call arrival rates chosen so that the average per cell call arrival rate is the same across patterns. The symbol l is for low, m for medium, and h for high. The numeric values of l, h, and m are chosen separately for each pattern to ensure that the average per cell rate of arrival is 120 calls/hr. The results show that is able to adapt its allocation strategy and thereby is better able to exploit the non-uniform call arrival rates. of backgammon (Tesauro, 1992), elevator-scheduling (Crites & Barto, 1995), and job-shop scheduling (Zhang & Dietterich, 1995). The neuro-dynamic programming textbook (Bertsekas and Tsitsiklis, 1996) presents a variety of related case studies. References Barto, A.G., Bradtke, S.J. & Singh, S. (1995) Learning to act using real-time dynamic programming. Articial Intelligence, 72:81{138. Bertsekas, D.P. (1995) Dynamic Programming and Optimal Control: Vols 1 and 2. Athena- Scientic, Belmont, MA. Bertsekas, D.P. & Tsitsiklis, J. (1996) Neuro-Dynamic Programming Athena-Scientic, Belmont, MA. Crites, R.H. & Barto, A.G. (1996) Improving elevator performance using reinforcement learning. In Advances is Neural Information Processing Systems 8. Del Re, W., Fantacci, R. & Ronga, L. (1996) A dynamic channel allocation technique based on Hopeld Neural Networks. IEEE Transactions on Vehicular Technology, 45:1. McEliece, R.J. & Sivarajan, K.N. (1994), Performance limits for channelized cellular telephone systems. IEEE Trans. Inform. Theory, pp. 21{34, Jan. Sutton, R.S. (1988) Learning to predict by the methods of temporal dierences. Machine Learning, 3:9{44. Tesauro, G.J. (1992) Practical issues in temporal dierence learning. Machine Learning, 8(3/4):257{277. Zhang, M. & Yum, T.P. (1989) Comparisons of Channel-Assignment Strategies in Cellular Mobile Telephone Systems. IEEE Transactions on Vehicular Technology Vol. 38, No. 4. Zhang, W. & Dietterich, T.G. (1996) High-performance job-shop scheduling with a timedelay TD(lambda) network. In Advances is Neural Information Processing Systems 8.