A Reinforcement Learning Framework For Dynamic Power Management of a Portable, Multi-Camera Traffic Monitoring System
|
|
|
- Monica Hines
- 10 years ago
- Views:
Transcription
1 A Reinforcement Learning Framework For Dynamic Power Management of a Portable, Multi-Camera Traffic Monitoring System Umair Ali Khan, Bernhard Rinner Institute of Networked and Embedded Systems Alpen-Adria Universitat, 9020-Klagenfurt, Austria [email protected], [email protected] Abstract Dynamic Power Management (DPM) refers to a set of strategies that achieves efficient power consumption by selectively turning off (or reducing the performance of) a system components when they are idle or are serving light workloads. This paper presents a Reinforcement Learning (RL) based DPM technique for a portable, multi-camera traffic monitoring system. We target the computing hardware of the sensing platform which is the major contributor to the entire power consumption. The RL technique used for the DPM of the sensing platform uses a model-free learning algorithm that does not require a priori model of the system. In addition, a robust workload estimator based on an online, Multi- Layer Artificial Neural Network (ML-ANN) is incorporated to the learning algorithm to provide partial information about the workload and to take better decisions according to the changing workload. Based on the estimated workload and a selected power-latency tradeoff parameter, the algorithm learns to use optimal time-out values in sleep and idle modes of the computing hardware. Our results show that the learning algorithm learns an optimal DPM policy for the non-stationary workload, while significantly reducing the power consumption and keeping the system response to a desired level. Keywords-Dynamic Power Management; Reinforcement Learning; Traffic Monitoring; I. INTRODUCTION Most of the existing traffic monitoring systems use no or limited image processing capabilities and exploit sensors such as induction loops, laser sensors and radars, etc. Additionally, these systems are based on stationary installations where the sensor nodes are permanently mounted at gantries. These fixed-mounted installations are usually expensive to set-up and decrease the flexibility of the monitoring system. During the installation, the road needs to be closed and a substantial calibration effort is also required. MobiTrick is a portable and compact traffic monitoring system that is aimed at temporary installations. The desired operation of the sensing platform is intended for short-terms (e.g., a few hours, days) and it can be deployed more flexibly for various monitoring tasks, e.g., law enforcement and construction site monitoring. Instead of using large sensors, MobiTrick uses only image processing operations for traffic monitoring. Since the system is portable and compact, its embedded design restricts to use intricate cooling systems and large batteries. Therefore, it has a strict limitation on power consumption and apart from its low-power design, it must use an online power management strategy to minimize the power consumption during operation. This motivates the search for an effective DPM technique for MobiTrick sensing platform. Relevant literature about DPM demonstrates various approaches. In the simplest approach, the greedy policy [9][8], a device transitions to the sleep state as soon as it is idle. Here, the term device refers to any electronic equipment that serves a particular purpose and has more than one modes of power consumption (or performance). The greedy policy can give the best power optimization as long as the requests arrive at long time intervals. A request represents a task generated by an application that needs processing. Another simple heuristic policy is the time-out policy [10][1] where a device is shut down after it has been idle for a certain threshold of time period. The time-out policy can be static or adaptive [5][20] which adjusts the time-out threshold based on the previous idle period history. The main shortcoming of time-out policies is the power wasted during the time-out period, specially when the workload (arrival rate of requests) is lower. This problem is better dealt by the predictive policies [2][14][4] which work on a system model that is learned from the history information in order to best adjust themselves to the dynamic system of a device. The basic idea in predictive policies is to predict the length of idle periods and shut down the device when the predicted idle period is longer than a certain threshold time period. Nevertheless, predictive policies do share a few limitations. First, they do not consider the response time of the device. Second, they do not deal with general system models where multiple incoming requests can be queued before processing. Third, they cannot perform well with non-stationary requests where the workload model is unknown. Some of these limitations (queuing, power-performance trade-off) are addressed by stochastic policies [9][11][16][13]. These approaches make probabilistic assumptions about the usage patterns of a device and exploit the nature of the probability distribution to formulate an optimization problem, the solution of which derives an optimal DPM strategy. The device states and
2 queues in stochastic policies are generally modeled as Markov chains. These policies do provide a flexible way to control the trade-off between power consumption and device response depending on the optimization constraints. However, Markov model is generally assumed to be stationary and known in advance. Therefore, these policies no longer remain optimal as workload becomes nonstationary. From the above discussion, it is evident that the performance of any selected policy heavily depends on the workload. Real workloads are usually non-stationary and compose a strict limitation on the success of any single policy. A model-free, machine learning approach can cope with this issue by interacting with the environment, implementing certain actions, evaluating the effects of the implemented actions and adjusting itself according to the environment. Compared with the existing machine learning DPM approaches, the RL based DPM approaches can deal with the non-stationary workloads in a much better way and can explore the trade-off between a system s power consumption and response time. The model-free, RL based approaches presented in [7][19] use online learning algorithms that dynamically select the best DPM policies from a set of pre-selected candidate policies. These algorithms do lead to optimal DPM policies, but they heavily rely on and are limited to the pre-selected candidate policies. In [21], authors propose an enhanced RL algorithm for system-level DPM. It is also a model-free approach that does not require prior knowledge of the state-transition probabilities. However, the number of state-action pairs in this system is quite large, which may result in increased computational complexity and slow convergence. Another similar work [22] uses the merits of [21] and proposes a RL based DPM algorithm with workload prediction for the power management of a WLAN card. The workload prediction in this work is performed with a Bayes classifier which performs well for this setting due to the regular traffic. Nevertheless, the same workload prediction can not render acceptable accuracy with a vehicle traffic data where the vehicles arrival rates follow a non-stationary pattern. In a recent work [18], a model-free, RL based DPM approach was used for non-stationary workload. The learning agent in this approach receives partial information about the workload from a workload estimator using a ML-ANN with backpropagation algorithm. Based on the estimated workload, this approach evaluates certain time-out values in idle state and waits for certain number of requests to be accumulated in the service queue when the system is in sleep state. Workload estimation using a ML-ANN achieves higher accuracy with the traffic data and the results show that the algorithm is capable of exploring the trade-off in the power-performance design space and converging to an optimal policy. However, since the algorithm waits for certain number of requests in the queue, a drawback of this approach is the high latency in requests processing when the workload drops abruptly. Following the merits of [18], we propose a novel RL based DPM algorithm in this paper for the power management of our sensing platform. The proposed algorithm uses time-out values both in sleep and idle states with workload estimation from a ML-ANN. Apart from this, we use multiple-states update in both sleep and idle modes and use a better exploration-exploitation policy to help algorithm converge fast and explore the design space deeper. As compared to the algorithm in [18], our results show that the algorithm proposed in this paper can find a better pareto-optimal trade-off curve of power-performance and results in a much lower latency while keeping the power consumption at an acceptable level. The remainder of this paper is organized as follows. Section 2 provides a brief overview of the MobiTrick sensing platform. Section 3 gives some background on reinforcement learning. Section 4 explains the implementation of RL based DPM for MobiTrick. Section 5 presents the overall learning algorithm and the results. We conclude our work with some discussion on future work in Section 6. II. MOBITRICK SENSING PLATFORM MobiTrick has a heterogeneous setup with different types of sensors, each having different capabilities. The three main components of the sensing platform comprise visual sensors (RGB, grayscale, infrared, D/N), non-visual sensors (inertial measurement units, GPS receiver) and a computing board for image processing. The sensing platform is intended to perform typical traffic monitoring tasks such as vehicle detection and classification, license plate detection, overheight estimation, incident detection. The heterogeneous setup serves many purposes. First, the tasks can be distributed among different sensors (e.g., license plate detection with an infrared camera and context image from a color camera). Second, low-level operations can be performed with less capable and more power-efficient sensors. Third, the redundant information from multiple sensors helps increase reliability. Figure 1 provides a high-level overview of the MobiTrick sensing platform. The sensing platform has a multi-tier architecture where the sensors reside at different levels based on their energy consumption and capabilities. A smart, low-power, color camera that can run on-board algorithms operates at the lowest level and triggers other cameras at the higher levels at the detection of an event. When triggered, the higher-level cameras send images to the smart camera. The queued images from all the cameras are then sent periodically to the computing board for processing. The contribution of an efficient DPM policy in this scenario
3 Tier 3 Tier 2 Images sent to the smart cam Tier 1 Input Stream Triggers for CAM 2 & CAM 3 CAM1 (smart color CAM) Event Detection Images Queue Figure 1. CAM3 (D/N) Triggered Mode CAM2 (Infrared) Triggered Mode Synchronized images IMUs & GPS Position, tilt MobiTrick sensing Platform. COMPUTING PLATFORM is to determine when it is appropriate to wake-up the board and process the images stored in the smart camera. In addition, it must also decide when it is efficient to shutdown the board, while minimizing the power consumption and keeping the system performance at an acceptable level. III. REINFORCEMENT LEARNING RL is a machine learning approach that is concerned with mapping situations to actions, in order to minimize a numerical penalty (or cost). As opposed to other machine learning approaches, for example supervised learning which is based on learning from examples provided by an external supervisor, RL is not dictated which actions to take. Instead, it must interact with the environment and discover the actions which yield the most reward (minimum penalty) by trying them. During the learning process, the agent observes the environment and issues appropriate actions based on the system state. As a result, the system changes state and the new state assigns the agent a penalty (or reward) which indicates the value (appropriateness) of the state transition. The overall goal of the learning process is to maximize the scalar reward (or minimize the penalty) in each state. RL assumes that the system dynamics follow Markov property, i.e., the next state s S and immediate reward r depend only on the current state s S and action a A, as given by equation 1. P r {s t+1 = s,r t+1 = r s t,a t } (1) Where P r is the probability of reaching state s and getting reward r at time t+1. A policy, π, is a mapping from each state, s S, and action, a A(s) to the probability of taking action a when in state s. Informally, the value of a state s under a policy π, denoted by V π (s), is the expected reward when starting in state s and following the policy π thereafter. We can define V π (s) as follows [12]: V π (s) = E π {R t s t = s} { } = E π γ k r t+k+1 s t = s k=0 Where E π {.} denotes the expected value given that the agent follows policy π, and t is any time step, γ (0,1) is a discount factor. Similarly, we define the value of taking action a in state s under a policy π, denoted by Q π (s,a), as the expected reward starting from s, taking action a, and thereafter following policy π. k=0 (2) Q π (s,a) = E π {R t s t = s,a t = a} { } = E π γ k r t+k+1 s t = s,a t = a (3) In a typical Markovian environment, we use a value-iteration algorithm with state transition probabilities to take an action in some state s. However, in a model-free learning, the agent has no prior information about the state transition probabilities. Therefore, we need an estimate of the value function described in equation 3. A variant of RL, the Q-learning [3] is the simplest form of RL that can directly approximate the value function V π (s) independent of the policy being followed. The Q-learning principle is given in equation 4. (s,a) S A : Q(s t,a t ) = Q(s t,a t ) +α t (s t,a t ){r t+1 +γmax a Q(s t+1,a) Q(s t,a t )} (4) Where α t (s t,a t ) (0,1) is the learning rate. Q π (s,a) for each state-action pair represents the expected long-term reward if the system starts from state s, takes action a, and thereafter follows policy π. Based on this value function, the agent decides which action should be taken in current state to achieve the maximum long-term reward, without knowing the state-transition probabilities. IV. RL BASED DPM FOR MOBITRICK In the MobiTrick sensing platform, we target the DPM of the computing board which is the main source of power consumption. It is an Intel Atom based computing platform [17] whose power and state-transition delay characteristics are given in Table I. Table I POWER AND DELAYS CHARACTERISTICS OF THE COMPUTING BOARD P sleep P idle P busy P trans t s2i t i2s 3W 25W 32W 15W 6s 4s Where P sleep, P idle, P busy and P trans represent the power consumption in sleep, idle, busy and transition state. t s2i and t i2s represent the time taken to switch from sleep to
4 idle state and vice versa, respectively. Since the model of the system is pre-characterized, we know how many power modes the system has and how it switches its power mode given a power command. Additionally, we also have partial information about the workload (ML-ANN based estimation). This information is adequate to design a Q- learning algorithm to find an optimal DPM policy, based on a selected power-performance parameter. In our setup, the RL environment consists of a WorkLoad estimator (WL), a Service Queue (SQ) to buffer the requests before processing, the computing board which works as a Service Provider (SP) and a Power Manager (PM) - the learning agent that issues appropriate power commands to SP. The workload estimator, SQ and PM reside on the smart camera. In this way, the smart camera works as a controller in the sensing platform that issues control signals (triggers, power commands, request receive/send) to other components of the system. We consider three power states of SP, i.e., sp = {sleep,idle,busy}. Therefore, the available power commands (actions) selected by PM include a = {go sleep, go idle, go busy}. At each decision epoch, PM receives an observation of the system that include the current state of WL, SQ and SP. Based on this composite state, S = {WL,SQ,SP}, PM issues a command to SP from the action set. Figure 2 depicts the high-level view of the power management setup. Img. Transfer Img. Transfer Figure 2. Img. Queue Smart CAM Power Manager Img. Capture Wakeup/Sleep Ack Observations Workload Estimator Event Detection Triggers Images Computing Board Handshaking Img. Receive Img. Processing Infrared CAM High. Res. CAM Depiction of the system under power management The learning algorithm is described below. M represents the transition matrix which keeps track of the visited states, actions, corresponding cost and other parameters. In each decision epoch, the system finds itself either in sleep state or in idle state. In both the states, the PM selects a time-out value (Equation 11) and relinquishes the control until the time-out period expires (or if some requests arrive during the time-out period in idle state). At the end of the time-out period (or when the time-out period is forced to terminate by Algorithm 1 RL based time-out policy Require: Power-performance parameter λ (0, 1) 1. Initialize Q, M and probability matrix p r arbitrarily. while Policy not good enough do 2. Obtain the current workload estimation (Sec. V.B) 3. Get the current observation: (s, a) 4. Calculate action probabilities: p k r(s,a k ) (Sec. V.E) 5. Select an action, a, with probability p r (Sec. V.E) 6. Execute the selected action 7. Calculate cost of the last action: c t+1 (s,a) (Sec. V.A) 8. Update the learning rate: α t (s,a) (Sec. V.D) 9. Update M with new state-action pair 10. Update Q-value: Q t+1 (s,a) (Sec. V.C) end while 11. Generate policy: π = min a Q(s,a), s S,a A some requests), the PM regains the control and evaluates the last action by assigning it a cost (Equation 5) and updating the Q-value of the last state-action pair (Equation 9). The PM, then, selects another action in the new state and issues appropriate signal to the SP. V. THE LEARNING ELEMENTS This section describes various elements of the learning, including the cost function, workload estimation, state-action pair updating principle, and the action-selection policy. A. Cost Function In the learning algorithm, we use cost instead of reward which can be treated in the similar way. The cost assigned to an action is a weighted combination of the average power consumption incurred due to the action and the performance penalty. We consider the average latency per request as the performance measure which is equal to the average queuing time plus the average execution time. The cost function is given in equation 5. t k+1 1 c t (s,a,λ) = λ P j +(1 λ)l t (s,a) (5) (t k+1 t k ) j=t k In the above expression, t k+1 t k is the time that the SP remains in state s, and λ (0,1) is power-performance trade-off parameter. For λ 0, the learning algorithm gives more importance to latency, thus resulting in a higher power consumption. On the other hand, when λ 1, the learning algorithm turns to aggressive power savings, resulting in higher latency. The value of λ can be varied slowly from 0 to 1 to obtain the pareto-optimal trade-off curve. B. Workload Estimation In this work, workload (or the request rate) comprises the rate of images captured by all the cameras at the detection of
5 events (vehicle detection). Therefore, the workload reflects the arrival rate of the vehicles. We used a real workload for the learning algorithm by taking a test recordings on a highway, where we recorded the inter-arrival times of vehicles by a vehicle detection algorithm [6] and acquired requests (events) that served as the data size for the learning algorithm. The mean inter-arrival times per hour from the 22-hours (2:00 A.M to A.M) recordings are depicted in Figure 3. Sliding window on previous inter-arrival periods x 0 x x k x k 1 x n 1 x n... v ij... y m Input Layer Hidden Layer Output Layer y 1 y 2 y m 1 w i z Next arrival period Figure 4. ML-ANN based workload estimator Where f is the sigmoid function. The weights are adjusted according to the following rule [15]. Figure 3. Mean request inter-arrival times per hour It is evident that the recorded workload exhibits a nonstationary pattern. For workload estimation, we use a fixsized moving window on the history of previous interarrival periods and input these inter-arrival periods to the ML-ANN. The ML-ANN estimates the length of the next inter-arrival period. If the length of the next inter-arrival period is estimated to be longer than a certain threshold T thr, the workload is classified as low (or high otherwise). Figure 4 shows the 3-layer ANN based workload estimator where v ij denotes the weight of the connection between j th neuron from the input layer and the i th neuron of the hidden layer. Whereas, w i represents the weight of the connection between i th neuron from the hidden layer and the neuron of the output layer. The outputs of the input layer, hidden layer and output layer, represented by I i, h and z are calculated as follows: I i = n v ij x j, i = 1,...,m j=1 (6a) y i = f(i i ) (6b) m h = w i y i (6c) i=1 z = f(s) (6d) v ij (t) = η(u z)w i x j f (h)f (I i )+µ v ij (t 1) (7a) w i (t) = η(u z)y i f (h)+µ w i (t 1) f = f(1 f) (7b) (7c) Whereuis the observed inter-arrival period,η is the learning rate, and µ is a positive constant which is determined experimentally. The experimental size of the moving window was selected to be n = 8. The prediction accuracy of the online ML-ANN workload estimator, which is defined as the ratio of correctly predicted intervals to the total number of predicted intervals, is 81.24%. C. The Learning Principle In our case, each system state has a composite form, i.e., S = {WL,SQ,SP}, where WL = {0,1} and SP = {sleep, busy, idle}. We use state aggregation for limiting the values of SQ to a state having no requests in the queue and the one having some requests, i.e., sq SQ,sq = {0,N N N}. The action set A for each state comprises time-out values t k out which depend on T thr and are defined as: t i out = ǫ i T thr, ǫ i R + (8) Where ǫ i is a positive weight. The learning algorithm in our case looks for minimizing long-term cost, rather than maximizing long-term reward. Hence, for the composite state S, action set A = { t 1 out,t 2 out,...,t n out}, and (s,a) S A, equation 4 can be modified as follows: Q (t+1) (s t,a t ) =Q t (s t,a t ) +α t (s t,a t ){(1 e βtout )c (t+1) (s,a,λ) +e βtout min a Q (t+1) (s (t+1),a) Q t (s t,a t )} (9)
6 Where Q (t+1) (s t,a t ) is the estimated Q-value of state s after taking the action a; Q t (s t,a t ) is the Q-value of state s before taking the action a; and (1 e βtout )c (t+1) (s,a,λ) is the sample discounted cost received by executing a time-out period, where β (0,1). The term min a Q (t+1) (s (t+1),a) refers to the best (minimum) value in next state s (t+1) corresponding to an action a. The learning algorithm looks for minimizing the difference between the current estimate of a state-action pair and the best discounted value of the next state. D. Updating Learning Rate The learning rate α t (s,a) is decreased slowly in such a way that it reflects the degree to which a state-action pair has been chosen in the recent past. It is calculated as: α t (s,a) = ξ visited(s, a) (10) Where ξ is a positive constant. Every time a state-action pair (s, a) is visited with this learning rate, the difference between its estimated Q-value Q (t+1) (s,a) and the current Q-value Q t (s,a) approaches to zero. Hence, for all stateaction pairs, the algorithm converges to an optimal policy. E. Exploration-Exploitation Policy It is necessary for a learning algorithm not only exploring the state-space deeper, but also exploiting the experience gained so far. In order to acquire sufficient knowledge about the dynamics of the system, every state-action pair must be visited adequate number of times. We use a semi-greedy policy in this work. This policy starts out with selecting random actions (exploration), each with equal probabilities. As the algorithm proceeds and acquires more knowledge about the system, the probability of actions with minimum cost begins to increase. Eventually, the policy tends to become greedy by selecting actions with heighest probability and minimum cost (exploitation). The action probabilities in a state is given by Equation. p k r(s,a) = e Qt (s,a k )/T n, a A, n = A (11) k=1 eqt (s,a k )/T Where T is called a temperature coefficient. It is initialized with a high value which gives equal weights (probabilities) to all the actions and hence the policy tends to explore the state-space, giving less importance to exploitation. T is then decayed over time and as the learning progresses the distribution becomes more tight on low-cost actions (equivalent to greedy decisions), and the exploration tendency of the agent reduces. F. Multiple States Update Selecting time-out values as actions enables multiple updates in both sleep and idle states in order to accelerate the convergence speed of the algorithm. Consider a case where the SP is in sleep state and the PM selects an action corresponding to a specific time-out value. If no request comes before the time-out expires, all the actions (timeout values) corresponding to smaller time-out values can be evalualted using Equation 9. This is because all the actions with smaller time-out values, if taken, would result in the same cost and the same next state Q-value estimate. Likewise, if the PM selects a time-out value in idle state and a request comes before the time-out expires, all the actions with higher time-out values can be evaluated. VI. EXPERIMENTAL RESULTS We varied the power-performance trade-off parameter, λ, between 0 and 1 to produce a pareto-optimal trade-off curve shown in Figure 5. We compared our modified RL based DPM algorithm (Timeout/Timeout model) with the algorithm presented in [18] which uses time-out values in idle state and waits for N number of requests in sleep state (Timeout/N model). Figure 5. Power-performance pareto optimal trade-off curve Figure 5 shows that RL based Timeout/Timeout model achieves aggresive power saving and produces a much deeper trade-off curve, while significantly reducing the average latency per request as compared to Timeout/N model. Selecting time-out values in sleep state has a direct impact on the time-out values in idle state. The optimal selection of time-out values in sleep state helps selecting optimal time-out values in idle state as well (and vice versa), thus avoiding the algorithm to waste energy by executing higher time-out values in idle state when the workload is lower. In Timeout/Timeout model, we can ensure the range of latency for the requests, which is given as follows: [l min, l max ] = [tp, (t i2s,t sleep,t s2i,t p )] (12) Where the minimum latency, l min, can be as small as the processing time, t p, of the request. On the other hand,
7 suppose that PM issues a sleep command to the SP and a request arrives at the same time. The maximum latency, l max, in this case can be as high as the sum of transition time from idle to sleep, t i2s, time-out period in sleep state, t sleep, transition time from sleep to idle, t s2i, and the processing time of the request. The probability distribution of the actions (time-out values) in each state is shown in Figure 6. The two modes SP = {0, 1} represent the sleep and idle state respectively. The actions sleep busy and busy idle are assumed to be autonomous and hence can be excluded from the action set. Figure 6 shows the probability distribution of action set for small value of the power-performance parameter (λ = 0.01), where the learning algorithm is very sensitive to the average latency per request and hence the probability of smaller timeout values (based on the estimated workload) in sleep state is high. In contrast, the probabilities of larger time-out values (based on the estimated workload) in idle state are high. The grayed cells shows the best actions with the highest probablities in each state that will be consistently selected when the policy becomes greedy. Time-out values (Increasing order) States S=(WL, SQ, SP) (0,0,0) (0,N,0) (1,0,0) (1,N,0) (0,0,1) (1,0,1) This also ensures the reliability of the workload estimator. However, it is still safe to keep the state (1,0,0) in statespace, lest it may be visited due to some random action or in a different workload dataset. We also analyzed the impact of using different threshold periods T thr on the learning algorithm. This also changes the action set of time-out values which are function of T thr. Figure 7 shows a comparison of power-performance curves for different values of T thr. This comparison shows that the power-performance curve does not change significantly by changing the threshold period T thr and hence the learning algorithm can adopt to different threshold periods. This is because selecting a smaller value of T thr does allow the algorithm choosing (relatively) smaller time-out values in the sleep state resulting in decreasing the average latency. However, it also compels the algorithm to select (relatively) smaller time-out values in idle state as well in order to compensate the (increased) power consumption. As a result, the average latency and the power consumption do not vary drastically for different values of T thr. Instead of using a static value of T thr, it may be adjusted over time by the mean of last n inter-arrival periods. Figure 6. Probability distribution of actions in each state From the above figure, it is also evident that the state(1,0,0) is never visited. This can be explained as follows. Suppose that the workload is estimated to be high and the SP has just entered idle state after processing all the requests in the queue. This represents a composite state s = (1,0,1). The PM now selects and execute a time-out period based on the workload estimation. Since the workload is estimated to be high and the time-out period is selected accordingly, it is highly likely that some requests will arrive before the timeout period expires. Even if no requests arrive during the time-out period and the PM issues sleep command to SP, it is highly likely that some requests will arrive during the transition period, leaving the SP either in state s = (1,N,0) or in state s = (0,N,0). Likewise, if the SP is in state s = (0,0,0) and the workload becomes high, the queue can no longer be empty, again leading to state s = (1,N,0). Figure 7. Comparison of power-performance curves for different threshold periods It is worth mentioning that the value of T thr must be bounded for a given set of weights ǫ i to avoid divergence of the algorithm or to prevent a local minima. Moreover, selecting very small values of weights-threshold combination results in smaller time-out values in action set and hence increases the average power consumption due to frequent transitions among the states. Likewise, selecting very high values of weights-threshold combination can also mitigate the power saving achieved in sleep state (with higher time-out values) by spending more time in idle state, executing higher time-out values and hence wasting energy.
8 For a given set of weights ǫ i, the boundary values for T thr can be determined experimentally based on a specific workload. For our experimental workload, the acceptable range of T thr is between 5 and 20 for the set of weights ǫ i = {0.1,0.2,...,1,1.1,...}. VII. CONCLUSION AND FUTURE WORK In this paper, we presented a RL based DPM approach for our stereo-vision based, multi-camera traffic monitoring system. The presented approach does not require a priori model of the system (the transition probabilities of individual components), as against to the existing stochastic DPM approaches. The algorithm learns the dynamics of the system by interacting with it, implementing certain actions during learning and evaluating their effects. Additionally, our proposed approach augments the good features of the traditional DPM approaches, i.e., time-out, greedy, predictive and stochastic, into a single learning scheme. The algorithm finds the optimal time-out values both in sleep and idle mode and tries to shutdown the computing board as soon as possible (greedy). Likewise, the algorithm also uses a predictive approach for workload prediction. Besides, the learning algorithm uses the same elements as used in stochastic approaches (queuing model, etc). Our results show that the proposed Timeout/Timeout algorithm provides a decent and much deeper power-latency trade-off curve as compared to Timeout/N algorithm. In future, we aim to formulate the DPM optimization problem as a dual problem, i.e., minimizing power consumption (or latency) for a given performance (or power) constraint. The RL based learning algorithm can then be implemented to find the right value of the power-performance constraint (λ) that exactly meets the power (or performance) constraint. ACKNOWLEDGEMENT This work has been sponsored in part by the Austrian Research Promotion Agency under grant REFERENCES [1] A. Karlin, M. Manasse, L. McGeoch and S. Owickim. Competitive randomized algorithms for non-uniform problems. Algorithmica, 11(6): , [2] C. H. Hwang, A. C. Wu. A predictive system shutdown method for energy saving of event-driven computation. In International Conerence on Computer Aided Design, [3] C. Watkins. Learning From Delayed Rewards. PhD thesis, Cambridge University, Cambridge, [4] E. -Y. Chung, L. Benini, G. De Micheli. Dynamic power management using adaptive learning tree. In International Conference on Computer Aided Design, pages , [5] F. Douglis, P. Krishnan, B. Bershad. Adaptive disk spin-down policies for mobile computers. Computing Systems, 8: , [6] F. Pletzer, R. Tusch, B. Rinner, L. Böszörmenyi. Robust traffic state estimation for smart cameras. In Proceedings of 9th IEEE International Conference on Advanced Video and Signal-Based Surveillance (AVSS), Beijing, China, [7] G. Dhiman, T. Rosing. Dynamic power management using machine learning. In IEEE/ACM International Conference on Computer-Aided Design, [8] J. M. Pedram. Power Aware Design Methodologies. Kluwer Academic, [9] L. Benini, A. Bogliolo, G. A. Paleolog, G. de Micheli. Policy optimization for dynamic power management. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 18: , [10] P. Greenawalt. Modeling power management for hard disks. In International Workshop on Modeling, Analysis, and Simulation for Computer and Telecomm. Systems, pages 62 65, [11] Q. Qiu, M. Pedram. Dynamic power management based on continuous-time markov decision process. In Design Automation Conference, pages , [12] R. S. Sutton, A. G. Barto. Reinforcement Learning: An Introduction. MIT Press, [13] S. Shukla, R. Gupta. A model checking approach to evaluating system level dynamic power management policies for embedded systems. In High-Level Design Validation and Test Workshop, pages 53 57, [14] Srivasta et al. Predictive system shutdown and other architecture techniques for energy efficient programmable computation. IEEE Transactions on VLSI Systems, 4:42 55, [15] T. Phit, K. Abe. Packet inter-arrival time estimation using neural network models. In Internet Conference, [16] T. Simunic, L. Benini, G. De Micheli. Event-driven power management of portable systems. In International Symp. on System Synthesis, pages 18 23, [17] U. A. Khan, M. Quaritsch, B. Rinner. Design of a heterogeneous, energy-aware, stereo-vision based sensing platform for traffic surveillance. In 9th Workshop on Intelligent Solutions in Embedded Systems, [18] U. Khan, B. Rinner. Dynamic power management for portable, multi-camera traffic monitoring. In 18th IEEE Real- Time and Embedded Technology and Applications Symposium, [19] V. Lakshmi, E. Monie. Hardware architecture of reinforcement learning scheme for dynamic power management in embedded systems. EURASIP Journal on Embedded Systems, 2007: , [20] Y. -H. Lu, G. De Micheli. Adaptive hard disk power management on personal computers. In Great Lakes Symp. VLSI, pages 50 53, [21] Y. Tan, W. Liu, Q. Qiu. Adaptive power management using reinforcement learning. In EEE/ACM International Conference on Computer-Aided Design, [22] Y. Wang, Q. Xie, A. Ammari, M. Pedram. Deriving a near-optimal power management policy using model-free reinforcement learning and bayesian classification. In 48th Design Automation Conference, pages 41 46, 2011.
POWER MANAGEMENT FOR DESKTOP COMPUTER: A REVIEW
POWER MANAGEMENT FOR DESKTOP COMPUTER: A REVIEW Ria Candrawati 1, Nor Laily Hashim 2, and Massudi Mahmuddin 3 1,2,3 Universiti Utara Malaysia, Malaysia, [email protected], [email protected], [email protected]
Inductive QoS Packet Scheduling for Adaptive Dynamic Networks
Inductive QoS Packet Scheduling for Adaptive Dynamic Networks Malika BOURENANE Dept of Computer Science University of Es-Senia Algeria [email protected] Abdelhamid MELLOUK LISSI Laboratory University
A Sarsa based Autonomous Stock Trading Agent
A Sarsa based Autonomous Stock Trading Agent Achal Augustine The University of Texas at Austin Department of Computer Science Austin, TX 78712 USA [email protected] Abstract This paper describes an autonomous
An Environment Model for N onstationary Reinforcement Learning
An Environment Model for N onstationary Reinforcement Learning Samuel P. M. Choi Dit-Yan Yeung Nevin L. Zhang pmchoi~cs.ust.hk dyyeung~cs.ust.hk lzhang~cs.ust.hk Department of Computer Science, Hong Kong
A Survey of Design Techniques for System-Level Dynamic Power Management
IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, VOL. 8, NO. 3, JUNE 2000 299 A Survey of Design Techniques for System-Level Dynamic Power Management Luca Benini, Member, IEEE, Alessandro
Dynamic Resource Allocation in Software Defined and Virtual Networks: A Comparative Analysis
Dynamic Resource Allocation in Software Defined and Virtual Networks: A Comparative Analysis Felipe Augusto Nunes de Oliveira - GRR20112021 João Victor Tozatti Risso - GRR20120726 Abstract. The increasing
Load Balancing and Switch Scheduling
EE384Y Project Final Report Load Balancing and Switch Scheduling Xiangheng Liu Department of Electrical Engineering Stanford University, Stanford CA 94305 Email: [email protected] Abstract Load
Prediction of DDoS Attack Scheme
Chapter 5 Prediction of DDoS Attack Scheme Distributed denial of service attack can be launched by malicious nodes participating in the attack, exploit the lack of entry point in a wireless network, and
A Power Efficient QoS Provisioning Architecture for Wireless Ad Hoc Networks
A Power Efficient QoS Provisioning Architecture for Wireless Ad Hoc Networks Didem Gozupek 1,Symeon Papavassiliou 2, Nirwan Ansari 1, and Jie Yang 1 1 Department of Electrical and Computer Engineering
Eligibility Traces. Suggested reading: Contents: Chapter 7 in R. S. Sutton, A. G. Barto: Reinforcement Learning: An Introduction MIT Press, 1998.
Eligibility Traces 0 Eligibility Traces Suggested reading: Chapter 7 in R. S. Sutton, A. G. Barto: Reinforcement Learning: An Introduction MIT Press, 1998. Eligibility Traces Eligibility Traces 1 Contents:
Comparison of K-means and Backpropagation Data Mining Algorithms
Comparison of K-means and Backpropagation Data Mining Algorithms Nitu Mathuriya, Dr. Ashish Bansal Abstract Data mining has got more and more mature as a field of basic research in computer science and
Introduction to Machine Learning and Data Mining. Prof. Dr. Igor Trajkovski [email protected]
Introduction to Machine Learning and Data Mining Prof. Dr. Igor Trakovski [email protected] Neural Networks 2 Neural Networks Analogy to biological neural systems, the most robust learning systems
A Slow-sTart Exponential and Linear Algorithm for Energy Saving in Wireless Networks
1 A Slow-sTart Exponential and Linear Algorithm for Energy Saving in Wireless Networks Yang Song, Bogdan Ciubotaru, Member, IEEE, and Gabriel-Miro Muntean, Member, IEEE Abstract Limited battery capacity
Energy Constrained Resource Scheduling for Cloud Environment
Energy Constrained Resource Scheduling for Cloud Environment 1 R.Selvi, 2 S.Russia, 3 V.K.Anitha 1 2 nd Year M.E.(Software Engineering), 2 Assistant Professor Department of IT KSR Institute for Engineering
Profit Maximization and Power Management of Green Data Centers Supporting Multiple SLAs
Profit Maximization and Power Management of Green Data Centers Supporting Multiple SLAs Mahdi Ghamkhari and Hamed Mohsenian-Rad Department of Electrical Engineering University of California at Riverside,
A Load Balancing Algorithm based on the Variation Trend of Entropy in Homogeneous Cluster
, pp.11-20 http://dx.doi.org/10.14257/ ijgdc.2014.7.2.02 A Load Balancing Algorithm based on the Variation Trend of Entropy in Homogeneous Cluster Kehe Wu 1, Long Chen 2, Shichao Ye 2 and Yi Li 2 1 Beijing
Chapter 4: Artificial Neural Networks
Chapter 4: Artificial Neural Networks CS 536: Machine Learning Littman (Wu, TA) Administration icml-03: instructional Conference on Machine Learning http://www.cs.rutgers.edu/~mlittman/courses/ml03/icml03/
Motivation. Motivation. Can a software agent learn to play Backgammon by itself? Machine Learning. Reinforcement Learning
Motivation Machine Learning Can a software agent learn to play Backgammon by itself? Reinforcement Learning Prof. Dr. Martin Riedmiller AG Maschinelles Lernen und Natürlichsprachliche Systeme Institut
ARTIFICIAL NEURAL NETWORKS FOR ADAPTIVE MANAGEMENT TRAFFIC LIGHT OBJECTS AT THE INTERSECTION
The 10 th International Conference RELIABILITY and STATISTICS in TRANSPORTATION and COMMUNICATION - 2010 Proceedings of the 10th International Conference Reliability and Statistics in Transportation and
Smart Queue Scheduling for QoS Spring 2001 Final Report
ENSC 833-3: NETWORK PROTOCOLS AND PERFORMANCE CMPT 885-3: SPECIAL TOPICS: HIGH-PERFORMANCE NETWORKS Smart Queue Scheduling for QoS Spring 2001 Final Report By Haijing Fang([email protected]) & Liu Tang([email protected])
Examining Self-Similarity Network Traffic intervals
Examining Self-Similarity Network Traffic intervals Hengky Susanto Byung-Guk Kim Computer Science Department University of Massachusetts at Lowell {hsusanto, kim}@cs.uml.edu Abstract Many studies have
T-REDSPEED White paper
T-REDSPEED White paper Index Index...2 Introduction...3 Specifications...4 Innovation...6 Technology added values...7 Introduction T-REDSPEED is an international patent pending technology for traffic violation
Urban Traffic Control Based on Learning Agents
Urban Traffic Control Based on Learning Agents Pierre-Luc Grégoire, Charles Desjardins, Julien Laumônier and Brahim Chaib-draa DAMAS Laboratory, Computer Science and Software Engineering Department, Laval
Journal of Theoretical and Applied Information Technology 20 th July 2015. Vol.77. No.2 2005-2015 JATIT & LLS. All rights reserved.
EFFICIENT LOAD BALANCING USING ANT COLONY OPTIMIZATION MOHAMMAD H. NADIMI-SHAHRAKI, ELNAZ SHAFIGH FARD, FARAMARZ SAFI Department of Computer Engineering, Najafabad branch, Islamic Azad University, Najafabad,
SINGLE-STAGE MULTI-PRODUCT PRODUCTION AND INVENTORY SYSTEMS: AN ITERATIVE ALGORITHM BASED ON DYNAMIC SCHEDULING AND FIXED PITCH PRODUCTION
SIGLE-STAGE MULTI-PRODUCT PRODUCTIO AD IVETORY SYSTEMS: A ITERATIVE ALGORITHM BASED O DYAMIC SCHEDULIG AD FIXED PITCH PRODUCTIO Euclydes da Cunha eto ational Institute of Technology Rio de Janeiro, RJ
Introduction to Engineering System Dynamics
CHAPTER 0 Introduction to Engineering System Dynamics 0.1 INTRODUCTION The objective of an engineering analysis of a dynamic system is prediction of its behaviour or performance. Real dynamic systems are
2004 Networks UK Publishers. Reprinted with permission.
Riikka Susitaival and Samuli Aalto. Adaptive load balancing with OSPF. In Proceedings of the Second International Working Conference on Performance Modelling and Evaluation of Heterogeneous Networks (HET
PART III. OPS-based wide area networks
PART III OPS-based wide area networks Chapter 7 Introduction to the OPS-based wide area network 7.1 State-of-the-art In this thesis, we consider the general switch architecture with full connectivity
Optimization models for congestion control with multipath routing in TCP/IP networks
Optimization models for congestion control with multipath routing in TCP/IP networks Roberto Cominetti Cristóbal Guzmán Departamento de Ingeniería Industrial Universidad de Chile Workshop on Optimization,
Shareability and Locality Aware Scheduling Algorithm in Hadoop for Mobile Cloud Computing
Shareability and Locality Aware Scheduling Algorithm in Hadoop for Mobile Cloud Computing Hsin-Wen Wei 1,2, Che-Wei Hsu 2, Tin-Yu Wu 3, Wei-Tsong Lee 1 1 Department of Electrical Engineering, Tamkang University
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
An Active Packet can be classified as
Mobile Agents for Active Network Management By Rumeel Kazi and Patricia Morreale Stevens Institute of Technology Contact: rkazi,[email protected] Abstract-Traditionally, network management systems
A CP Scheduler for High-Performance Computers
A CP Scheduler for High-Performance Computers Thomas Bridi, Michele Lombardi, Andrea Bartolini, Luca Benini, and Michela Milano {thomas.bridi,michele.lombardi2,a.bartolini,luca.benini,michela.milano}@
Per-Flow Queuing Allot's Approach to Bandwidth Management
White Paper Per-Flow Queuing Allot's Approach to Bandwidth Management Allot Communications, July 2006. All Rights Reserved. Table of Contents Executive Overview... 3 Understanding TCP/IP... 4 What is Bandwidth
Scheduling using Optimization Decomposition in Wireless Network with Time Performance Analysis
Scheduling using Optimization Decomposition in Wireless Network with Time Performance Analysis Aparna.C 1, Kavitha.V.kakade 2 M.E Student, Department of Computer Science and Engineering, Sri Shakthi Institute
Multisensor Data Fusion and Applications
Multisensor Data Fusion and Applications Pramod K. Varshney Department of Electrical Engineering and Computer Science Syracuse University 121 Link Hall Syracuse, New York 13244 USA E-mail: [email protected]
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
Transport Layer Protocols
Transport Layer Protocols Version. Transport layer performs two main tasks for the application layer by using the network layer. It provides end to end communication between two applications, and implements
CHAPTER 1 INTRODUCTION
1 CHAPTER 1 INTRODUCTION 1.1 MOTIVATION OF RESEARCH Multicore processors have two or more execution cores (processors) implemented on a single chip having their own set of execution and architectural recourses.
Appendix: Simple Methods for Shift Scheduling in Multi-Skill Call Centers
MSOM.1070.0172 Appendix: Simple Methods for Shift Scheduling in Multi-Skill Call Centers In Bhulai et al. (2006) we presented a method for computing optimal schedules, separately, after the optimal staffing
Chapter 3 ATM and Multimedia Traffic
In the middle of the 1980, the telecommunications world started the design of a network technology that could act as a great unifier to support all digital services, including low-speed telephony and very
Multi-service Load Balancing in a Heterogeneous Network with Vertical Handover
1 Multi-service Load Balancing in a Heterogeneous Network with Vertical Handover Jie Xu, Member, IEEE, Yuming Jiang, Member, IEEE, and Andrew Perkis, Member, IEEE Abstract In this paper we investigate
Jae Won idee. School of Computer Science and Engineering Sungshin Women's University Seoul, 136-742, South Korea
STOCK PRICE PREDICTION USING REINFORCEMENT LEARNING Jae Won idee School of Computer Science and Engineering Sungshin Women's University Seoul, 136-742, South Korea ABSTRACT Recently, numerous investigations
Reformulating the monitor placement problem: Optimal Network-wide wide Sampling
Reformulating the monitor placement problem: Optimal Network-wide wide Sampling Gianluca Iannaccone Intel Research @ Cambridge Joint work with: G. Cantieni,, P. Thiran (EPFL) C. Barakat (INRIA), C. Diot
Design call center management system of e-commerce based on BP neural network and multifractal
Available online www.jocpr.com Journal of Chemical and Pharmaceutical Research, 2014, 6(6):951-956 Research Article ISSN : 0975-7384 CODEN(USA) : JCPRC5 Design call center management system of e-commerce
Chapter 2 The Research on Fault Diagnosis of Building Electrical System Based on RBF Neural Network
Chapter 2 The Research on Fault Diagnosis of Building Electrical System Based on RBF Neural Network Qian Wu, Yahui Wang, Long Zhang and Li Shen Abstract Building electrical system fault diagnosis is the
Scheduling Allowance Adaptability in Load Balancing technique for Distributed Systems
Scheduling Allowance Adaptability in Load Balancing technique for Distributed Systems G.Rajina #1, P.Nagaraju #2 #1 M.Tech, Computer Science Engineering, TallaPadmavathi Engineering College, Warangal,
A Routing Algorithm Designed for Wireless Sensor Networks: Balanced Load-Latency Convergecast Tree with Dynamic Modification
A Routing Algorithm Designed for Wireless Sensor Networks: Balanced Load-Latency Convergecast Tree with Dynamic Modification Sheng-Cong Hu [email protected] Jen-Hou Liu [email protected] Min-Sheng
Building an Advanced Invariant Real-Time Human Tracking System
UDC 004.41 Building an Advanced Invariant Real-Time Human Tracking System Fayez Idris 1, Mazen Abu_Zaher 2, Rashad J. Rasras 3, and Ibrahiem M. M. El Emary 4 1 School of Informatics and Computing, German-Jordanian
AUTOMATION OF ENERGY DEMAND FORECASTING. Sanzad Siddique, B.S.
AUTOMATION OF ENERGY DEMAND FORECASTING by Sanzad Siddique, B.S. A Thesis submitted to the Faculty of the Graduate School, Marquette University, in Partial Fulfillment of the Requirements for the Degree
6.231 Dynamic Programming and Stochastic Control Fall 2008
MIT OpenCourseWare http://ocw.mit.edu 6.231 Dynamic Programming and Stochastic Control Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 6.231
Routing in packet-switching networks
Routing in packet-switching networks Circuit switching vs. Packet switching Most of WANs based on circuit or packet switching Circuit switching designed for voice Resources dedicated to a particular call
APPENDIX 1 USER LEVEL IMPLEMENTATION OF PPATPAN IN LINUX SYSTEM
152 APPENDIX 1 USER LEVEL IMPLEMENTATION OF PPATPAN IN LINUX SYSTEM A1.1 INTRODUCTION PPATPAN is implemented in a test bed with five Linux system arranged in a multihop topology. The system is implemented
An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network
Proceedings of the 8th WSEAS Int. Conf. on ARTIFICIAL INTELLIGENCE, KNOWLEDGE ENGINEERING & DATA BASES (AIKED '9) ISSN: 179-519 435 ISBN: 978-96-474-51-2 An Energy-Based Vehicle Tracking System using Principal
Reconfigurable Architecture Requirements for Co-Designed Virtual Machines
Reconfigurable Architecture Requirements for Co-Designed Virtual Machines Kenneth B. Kent University of New Brunswick Faculty of Computer Science Fredericton, New Brunswick, Canada [email protected] Micaela Serra
Design of an FX trading system using Adaptive Reinforcement Learning
University Finance Seminar 17 March 2006 Design of an FX trading system using Adaptive Reinforcement Learning M A H Dempster Centre for Financial Research Judge Institute of Management University of &
Distributed Load Balancing for FREEDM system
Distributed Load Balancing for FREEDM system Ravi Akella, Fanjun Meng, Derek Ditch, Bruce McMillin, and Mariesa Crow Department of Electrical Engineering Department of Computer Science Missouri University
PEDAMACS: Power efficient and delay aware medium access protocol for sensor networks
PEDAMACS: Power efficient and delay aware medium access protocol for sensor networks Sinem Coleri and Pravin Varaiya Department of Electrical Engineering and Computer Science University of California,
OPTIMIZATION OF VENTILATION SYSTEMS IN OFFICE ENVIRONMENT, PART II: RESULTS AND DISCUSSIONS
OPTIMIZATION OF VENTILATION SYSTEMS IN OFFICE ENVIRONMENT, PART II: RESULTS AND DISCUSSIONS Liang Zhou, and Fariborz Haghighat Department of Building, Civil and Environmental Engineering Concordia University,
Chapter 1. Introduction
Chapter 1 Introduction 1.1. Motivation Network performance analysis, and the underlying queueing theory, was born at the beginning of the 20th Century when two Scandinavian engineers, Erlang 1 and Engset
Multiobjective Multicast Routing Algorithm
Multiobjective Multicast Routing Algorithm Jorge Crichigno, Benjamín Barán P. O. Box 9 - National University of Asunción Asunción Paraguay. Tel/Fax: (+9-) 89 {jcrichigno, bbaran}@cnc.una.py http://www.una.py
Offline sorting buffers on Line
Offline sorting buffers on Line Rohit Khandekar 1 and Vinayaka Pandit 2 1 University of Waterloo, ON, Canada. email: [email protected] 2 IBM India Research Lab, New Delhi. email: [email protected]
Data Mining Algorithms Part 1. Dejan Sarka
Data Mining Algorithms Part 1 Dejan Sarka Join the conversation on Twitter: @DevWeek #DW2015 Instructor Bio Dejan Sarka ([email protected]) 30 years of experience SQL Server MVP, MCT, 13 books 7+ courses
A Network Flow Approach in Cloud Computing
1 A Network Flow Approach in Cloud Computing Soheil Feizi, Amy Zhang, Muriel Médard RLE at MIT Abstract In this paper, by using network flow principles, we propose algorithms to address various challenges
Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm
1 Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm Hani Mehrpouyan, Student Member, IEEE, Department of Electrical and Computer Engineering Queen s University, Kingston, Ontario,
On the Interaction and Competition among Internet Service Providers
On the Interaction and Competition among Internet Service Providers Sam C.M. Lee John C.S. Lui + Abstract The current Internet architecture comprises of different privately owned Internet service providers
Intelligent Heuristic Construction with Active Learning
Intelligent Heuristic Construction with Active Learning William F. Ogilvie, Pavlos Petoumenos, Zheng Wang, Hugh Leather E H U N I V E R S I T Y T O H F G R E D I N B U Space is BIG! Hubble Ultra-Deep Field
6.2.8 Neural networks for data mining
6.2.8 Neural networks for data mining Walter Kosters 1 In many application areas neural networks are known to be valuable tools. This also holds for data mining. In this chapter we discuss the use of neural
Simulated Annealing Based Hierarchical Q-Routing: a Dynamic Routing Protocol
Simulated Annealing Based Hierarchical Q-Routing: a Dynamic Routing Protocol Antonio Mira Lopez Power Costs Inc. Norman, OK 73072 [email protected] Douglas R. Heisterkamp Computer Science Department
Advanced Load Balancing Mechanism on Mixed Batch and Transactional Workloads
Advanced Load Balancing Mechanism on Mixed Batch and Transactional Workloads G. Suganthi (Member, IEEE), K. N. Vimal Shankar, Department of Computer Science and Engineering, V.S.B. Engineering College,
Sla Aware Load Balancing Algorithm Using Join-Idle Queue for Virtual Machines in Cloud Computing
Sla Aware Load Balancing Using Join-Idle Queue for Virtual Machines in Cloud Computing Mehak Choudhary M.Tech Student [CSE], Dept. of CSE, SKIET, Kurukshetra University, Haryana, India ABSTRACT: Cloud
CHAPTER 1 INTRODUCTION
21 CHAPTER 1 INTRODUCTION 1.1 PREAMBLE Wireless ad-hoc network is an autonomous system of wireless nodes connected by wireless links. Wireless ad-hoc network provides a communication over the shared wireless
1. The subnet must prevent additional packets from entering the congested region until those already present can be processed.
Congestion Control When one part of the subnet (e.g. one or more routers in an area) becomes overloaded, congestion results. Because routers are receiving packets faster than they can forward them, one
THE BENEFITS OF SIGNAL GROUP ORIENTED CONTROL
THE BENEFITS OF SIGNAL GROUP ORIENTED CONTROL Authors: Robbin Blokpoel 1 and Siebe Turksma 2 1: Traffic Engineering Researcher at Peek Traffic, [email protected] 2: Product manager research
TD(0) Leads to Better Policies than Approximate Value Iteration
TD(0) Leads to Better Policies than Approximate Value Iteration Benjamin Van Roy Management Science and Engineering and Electrical Engineering Stanford University Stanford, CA 94305 [email protected] Abstract
OpenFlow Based Load Balancing
OpenFlow Based Load Balancing Hardeep Uppal and Dane Brandon University of Washington CSE561: Networking Project Report Abstract: In today s high-traffic internet, it is often desirable to have multiple
Thesis work and research project
Thesis work and research project Hélia Pouyllau, INRIA of Rennes, Campus Beaulieu 35042 Rennes, [email protected] July 16, 2007 1 Thesis work on Distributed algorithms for endto-end QoS contract
Network management and QoS provisioning - QoS in the Internet
QoS in the Internet Inernet approach is based on datagram service (best effort), so provide QoS was not a purpose for developers. Mainly problems are:. recognizing flows;. manage the issue that packets
Energy Efficient Load Balancing among Heterogeneous Nodes of Wireless Sensor Network
Energy Efficient Load Balancing among Heterogeneous Nodes of Wireless Sensor Network Chandrakant N Bangalore, India [email protected] Abstract Energy efficient load balancing in a Wireless Sensor
International Journal of Computer Science Trends and Technology (IJCST) Volume 3 Issue 3, May-June 2015
RESEARCH ARTICLE OPEN ACCESS Ensuring Reliability and High Availability in Cloud by Employing a Fault Tolerance Enabled Load Balancing Algorithm G.Gayathri [1], N.Prabakaran [2] Department of Computer
CHAPTER 1 INTRODUCTION
CHAPTER 1 INTRODUCTION 1.1 Background The command over cloud computing infrastructure is increasing with the growing demands of IT infrastructure during the changed business scenario of the 21 st Century.
Comparison between Vertical Handoff Decision Algorithms for Heterogeneous Wireless Networks
Comparison between Vertical Handoff Decision Algorithms for Heterogeneous Wireless Networks Enrique Stevens-Navarro and Vincent W.S. Wong Department of Electrical and Computer Engineering The University
American International Journal of Research in Science, Technology, Engineering & Mathematics
American International Journal of Research in Science, Technology, Engineering & Mathematics Available online at http://www.iasir.net ISSN (Print): 2328-349, ISSN (Online): 2328-3580, ISSN (CD-ROM): 2328-3629
