Check out the new USENIX Web site. next up previous
Next: Energy Efficient Disk I/O Up: Currentcy: A Unifying Abstraction Previous: Coordinated Scheduling of Multiple


Low Variance in Response Time Through Pacing Currentcy Expenditures

Given a case in which power consumption must be constrained, our epoch-based allocation has the potential to produce bursty behavior if tasks consume currentcy as quickly as they can at the beginning of an epoch and then go idle after consuming their budget. One approach to smoothing consumption rates (and as a side-effect, response times), is to shorten the epoch.

Another approach to managing the rate of consumption is self-pacing in our EC scheduler. The idea is to delay a task if its consumption of currentcy is ahead of schedule during an epoch. Progress is defined as the amount of currentcy spent thus far in the current epoch divided by the task's budget for the epoch. If this progress is greater than the ratio of elapsed time in this epoch over epoch length, then the task is delayed and the processor may go idle for a short interval of time. This approach exploits the ability of currentcy to reflect an application's rate of progress. This approach to stretching execution is appropriate for a non-Dynamic Voltage Scaled processor. If available, DVS would be a preferred alternative to consider.

To compare these two approaches, we first look at the overhead of the first approach because it increases with the shortened epoch length and could become a performance bottleneck. However, experiments show the overhead for currentcy allocation is very small. Even if we perform allocation every 10ms (a timer interrupt occurs every 10ms, while the CPU scheduling quantum is 60ms in our system), the overhead is only 206$\mu$s for 18 resource containers and 40$\mu$s for 3 containers.

To explore the effects on response time of our two approaches for reducing bursty performance, we run Netscape and continuously load our department's web page. This page contains a banner image and some simple text. The autoload is implemented using a javascript and this also allows us to measure the page load latency. This latency is composed of several http requests, displaying the content and updating the disk cache. We set the think time between successive page loads at 2 seconds. Executing without any throttling requires about 2197mW.

We evaluate both an epoch-based approach that uses 0.01 second epochs, and the self-pacing approach with 10 second epochs. We allocate currentcy equivalent to an average of 1200mW to Netscape and measure 54 consecutive page loads for the self-paced test and 41 page loads for the epoch based approach. Differences are apparent in Table 3 when we examine the delay for a page load. Although the average delay is similar for the two policies, the self-paced scheduler has much lower variation in the delay. This can translate into a user perceived difference in performance as the self-pacing policy can provide a visibly smoother display of the web page. We note that similar visible differences occur when executing other applications, such as RealPlayer, Acrobat, and StarOffice.


Table 3: Response Time Variation
Power Consumption (mW) Delay (seconds)
Scheduling CPU Disk Network Total Min Max Average Std. Dev
Unthrottled 1047 1013 136 2197 0.27 0.68 0.43 0.11
Epoch 351 812 48 1212 1.0 33.8 3. 8 5.8
Self-Pacing 313 842 43 1199 3.3 5.6 4.0 0.6



next up previous
Next: Energy Efficient Disk I/O Up: Currentcy: A Unifying Abstraction Previous: Coordinated Scheduling of Multiple
Heng Zeng 2003-04-07