Sie sind auf Seite 1von 4

Real-Time Parallel Disk Scheduling for VBR Video Servers

Özgür Ertug Mahesh Kallahalla Peter J. Varman


Department of Electrical and Computer Engineering
Rice University
Houston TX 77251

Abstract (CDL) storage schemes are preferred over con-


stant time length (CTL) schemes [2]. In this pa-
We introduce a framework for real-time I/O per, we consider VOD systems that store VBR
scheduling for multiple-disk parallel I/O systems. video data in CDL format on multiple disks.
The framework is used to model a video server Given a fixed set of resources such as disk band-
delivering VBR encoded video data with real- width, storage volume, and main memory buffer,
time requirements. The video streams are as- the video server can deliver a limited number of
sumed to be stored in CDL format and dis- streams while providing the specified QoS guar-
tributed across multiple disks. antee. The primary design goal is to maximize
We present a novel algorithm RT-OPT for op- the number of clients that can be concurrently
timally prefetching blocks into the server buffer. served with the guaranteed QoS.
We show that for any fixed disk-head scheduling The storage subsystem is in general the bot-
policy, RT-OPT is guaranteed to produce a feasi- tleneck for real-time delivery of video due to the
ble schedule, in which all blocks meet their dead- high I/O latencies incurred by secondary storage
lines, if one exists. Simulations show that the devices. Parallel I/O systems incorporating mul-
number of clients supported by RT-OPT is su- tiple disks have the potential to alleviate this
perior to intuitive algorithms like GREED-EDF problem by providing a higher disk bandwidth.
that aggressively keep the disks busy fetching in However harnessing the raw disk bandwidth to
order of deadlines. decrease the latency seen by individual blocks
requires sophisticated prefetching and caching
techniques [4]. The problem is compounded by
1 Introduction the necessity for real-time guarantees in a VOD
server.
Video-on-Demand (VOD) systems are distributed
server-network-client architectures that provide In a parallel I/O system, idle disks can be used
real-time video service to multiple clients with to prefetch video blocks concurrently with de-
specified quality of service (QoS) guarantees. mand I/Os from other disks. These prefetched
Video data is conventionally stored in com- blocks are held in the buffer until their dead-
pressed form on parallel disk arrays. The main lines, when they can be transferred to the client.
goal of a VOD server is to deliver the requested However, deciding which blocks to prefetch is a
video clips, stored on secondary storage, to the non-trivial task. Blocks that are prefetched long
clients in a timely manner honoring QoS guar- before their deadline occupy buffer space waste-
antees such as delay, delay jitter and maximum fully, reducing the effective buffer size, and de-
frame-loss probability. creasing the number of clients that the server can
Video data compressed in MPEG format inher- handle; on the other hand delaying a prefetch po-
ently results in a variable bit rate (VBR) stream. tentially wastes disk bandwidth, causing blocks
To simplify storage allocation, avoid fragmen- downstream to miss their deadlines and violat-
tation, and ease buffer management such com- ing QoS guarantees.
pressed video streams are stored in constant- We present a framework for real-time paral-
sized data blocks. Such constant data length lel I/O. In Section 2, we present our parallel I/O
 model. A real-time, parallel-disk scheduling al-
Supported in part by the National Science Foundation gorithm RT-OPT that maximizes the number of
under grant CCR-9704562, a grant from the Schlumberger
clients that can be supported, is presented in Sec-
Foundation, and a Graduate Research Fellowship from Texas
Instruments. tion 3. For any fixed disk-head scheduling pol-
To appear in Proceedings of CSI’99 icy [6], we show that RT-OPT is optimal in that
if in its schedule, any block misses its deadline a deadline on the latest time an I/O for a block
then there does not exist any schedule in which can be initiated by the disk system.
all blocks can meet their deadlines. In Section 4 From the server’s viewpoint the model can
we present simulation results for RT-OPT using be simplified by incorporating the network de-
traces of MPEG video streams, by empirically de- lay and client-side decompression and process-
termining the number of clients that can be sup- ing into the playback interval. Further to stress
ported for different QoS and server parameters. the server we assume minimal client features: a
There have been a number of different ap- client only buffers the current block. This makes
proaches for storage and retrieval of real-time the client-buffer residence time zero.
video data. Disk-head scheduling and single-disk The input to the scheduling algorithm is a
data placement were addressed in [6, 8]. For string of blocks  , each with its associated disk
the SPIFFI video-server, [3] proposed a real-time access time  (IO End - IO Start), and deadline
priority-based disk scheduling algorithm that  , the latest time at which the block can com-
partitions requests into priority classes based on plete its I/O. A block for which IO End is less than
the nearness of their deadlines. In [1, 2] ad-  stays in the server buffer till  at which time
mission and resource allocation algorithms for it is transferred to the client. The scheduling al-
VBR-CDL video streams were studied in a single- gorithm then determines the time at which each
disk system based on the concept of fetching in block should begin its I/O.
rounds. While simplifying the admission con- For illustration consider an intuitive schedul-
trol algorithm, these techniques under-utilize the ing algorithm GREED-EDF. GREED-EDF, ag-
buffer and consequently can handle fewer clients. gressively keeps the disks busy, fetching in the
In [5], caching methods to reuse blocks of pop- order of deadlines. The blocks that cannot meet
ular movies were proposed and analyzed. None their deadline are dropped. Consider an I/O sys-
of these works attempts to optimize the use of tem with 3 disks and a buffer of 2 blocks. Let
buffer in a multiple-disk situation among multi- blocks  ,  , and  with deadlines 1, 3, and 4 re-
ple disks. spectively, be requested from disks 1, 2, and 3 re-
spectively. Let the I/O time for blocks  and  be
1 each while that of  be 4. Since the deadline
2 Real-time Model of  (3) is earlier than that of  (4), GREED-EDF
fetches blocks  and  initially. Thus an I/O for
We consider a server-network-client architecture  cannot start before time 1, which causes it to
of a video server. The server consists of an array miss its deadline. On the other hand we could
 blocks.
of parallel disks and a main memory of size initiate an I/O for  at time 0 and start fetching 
A block is the unit of access from the only at time 1 which can allow all deadlines to be
disks. Video streams are VBR encoded and stored met. This schedule is constructed by RT-OPT.
in CDL format so that blocks are of fixed size,
and therefore the number of frames per block is
variable. Our framework places no restriction on 3 Algorithm RT-OPT
the placement of blocks on disks. The clients are
assumed to have buffers to hold the current block Algorithm RT-OPT generates a schedule for a set
being played back, and to mask network delays. of blocks, given their disk access times and dead-
The temporal evolution of each block is mod- lines. The access times are obtained by assum-
eled by a timeline, and indicates the state of the ing a particular disk scheduling policy for each
block between the start of its I/O and the end disk. Intuitively, RT-OPT operates as follows.
of its playback. The timeline consists of disk The blocks are first tentatively scheduled so that
I/O time, time spent in the server buffer, net- each block’s IO End time coincides with its dead-
work transmission delay, time spent in the client line. Next RT-OPT iterates through the set of
buffer, processing delay at client, and playback. blocks in order to resolve conflicts. The first type
The playback deadline for each block, is fixed of conflict occurs due to overlapping disk access
by the frame rate and encoding of the previous intervals on a single disk. Since only one I/O can
blocks. The decompression time is also fixed and be in progress from a single disk at any time, the
depends on the encoding of that block. The net- disk access interval of one of these blocks must be
work transmission time is assumed to be known moved earlier in time to remove the overlap. The
for every client. The disk I/O time is determined chosen order of access from the disk determines
by the disk-head scheduling policy and disk char- which will stay and which will be prefetched. A
acteristics. All these components together place block is pushed back the minimal amount, so that
Algorithm RT-OPT

for d=1 to Number of disks /* Initialization */


for b = 1 to last block on disk d
Assign     
       and       
 
call        
At any time 
Fetch   blocks from   giving preference to blocks with earlier I/O Start time

Procedure     (disk d, block b)


if  
  and          then
Assign           and              
call         

Figure 1: Algorithm RT-OPT

its IO End coincides with the IO Start of the next pressed video traces described in [7]. The frame
block to be fetched from that disk. Finally RT- rate for these traces is 24 fps and each trace con-
OPT checks for buffer violations in the sched- tains 40000 frames. Using data on the coded
ule. At any time the total number of blocks be- frame sizes for each of the video sequences, we
ing fetched or which are in the server buffer can- created CDL blocks of 64 Kbytes from 5 rep-
not exceed , the size of the server buffer. The resentative streams: Bond, Terminator, Lambs,
blocks are scanned in increasing order of time, Dino and Starwars. The blocks were then striped
and blocks whose I/O initiation would violate the across the disks in a round-robin fashion. To
buffer constraint are dropped. model the probabilistic access structure of clients
The pseudo-code for algorithm RT-OPT is we used the Zipf distribution, with a skewness
shown in Figure 1. In the description, StartPB(b) factor of 0.271. Clients were assumed to access
denotes the playback start time PB Start for the chosen clip from a random starting frame;
block b; similarly StartIO(b) and EndIO(b) de- wrapping around when the end of a stream is
note IO Start and IO End times for block b. At reached. The results reported correspond to
any time  let !  represent the set of idle disks, accesses to 50000 blocks. We assumed EDF
and      "#   $!  $%, where #  is the scheduling at individual disks, and I/O times uni-
number of free blocks in the server buffer. Let formly distributed between 16 and 44ms. In all
  be a set of blocks, from idle disks, with the cases averaged results over several Monte Carlo
earliest I/O start time among all blocks from that trials are reported.
disk satisfying  &      ' 
 . We first consider the case in which all blocks
RT-OPT’s policy of fetching blocks as late as are required to meet their deadlines and measure
possible minimizes the buffer occupancy of each the number of clients that can be supported. Fig-
block and consequently the probability of buffer ure 2 shows the the number of clients serviced
overflow for a given buffer size. Using this we versus buffer size for 64 disks. RT-OPT does sub-
can formally prove that RT-OPT is optimal in stantially better than GREED-EDF over a wode
the sense that if it drops some blocks then every range of buffer sizes. Similar trends were ob-
other algorithm must also drop at least one block served with smaller number of disks.
for the same data placement, buffer size and sin- The scalability of the algorithms RT-OPT and
gle disk-head scheduling policy. As an admission GREED-EDF with number of disks is illustrated
control method RT-OPT can determine whether a in Figure 3. The scalability of RT-OPT is bet-
client can be admitted by checking if a schedule ter than that of GREED-EDF, and close to lin-
exists in which each block meets its deadline. ear at all buffer sizes: M = 32, 64, 128 and 256.
For brevity only results with M =256 are shown
4 Simulations in the figure. The performance of GREED-EDF
saturates quickly; that is, GREED-EDF requires
The performance of RT-OPT was evaluated us- a much larger amount of buffer to exploit the
ing a trace-driven simulator based on our real- available disk bandwidth. In contrast RT-OPT
time model. We used public-domain MPEG com- supports a substantially larger number of clients
3500

Greedy
RT−OPT
D=64

5 Conclusions
3000

2500
We introduced a framework for incorporating
real-time constraints on block accesses in a par-
Number of Users Allowed allel I/O system. The framework was used to de-
2000

1500
sign prefetching and scheduling algorithms for
1000
real-time playback of multiple video streams in
500
a multiple-disk, parallel-I/O-based video server.
We introduced a scheduling algorithm RT-OPT
0
0 50 100 150
Buffer Size
200 250 300 for prefetching blocks into the server buffer. RT-
OPT is shown to be optimal, in the sense that
Figure 2: Clients Serviced vs. Buffer Size if it drops a block then there exists no schedule
in which all blocks meet their deadline. Using
traces from MPEG compressed video clips we em-
pirically observed the performance of RT-OPT. By
with modest buffer requirements. dynamically multiplexing the buffer among dif-
M=256
ferent clients and disks optimally, RT-OPT was
3500

Greedy
shown to be highly scalable with the number of
3000
RT−OPT
disks with only modest buffer requirements. The
number of clients supported was shown to be su-
2500
perior to intuitive but suboptimal algorithms like
Number of users allowed

2000 GREED-EDF that attempt to keep the disks busy


1500
and fetch in order of deadlines.
1000

500
References
0
0 10 20 30 40 50 60 70
[1]E. W. Biersack, F. Thiesse, and C. Bernhardt. Con-
Number of disks
stant Data Length Retrieval for Video Servers with
Variable Bit Rate Streams. In Proc. of Conf. on
Figure 3: Scalability with number of disks Multimedia, pp 151–155, 1996.
[2]Ed Chang and A. Zakhor. Admission Control and
Finally in Figure 4 the two algorithms are com- Data Placement for VBR Video Servers. In Proc. of
pared by determining the number of clients that Intl. Conf. on Image Processing, pp 316–329, 1994.
can be supported with a given block dropping
[3]C. .S Freedman and D. .J DeWitt. The SPIFFI Scal-
probability. The maximum drop rate permitted
able Video-on-Demand System. In Proc. of SIG-
was fixed at 2%, and M was fixed at 256. For MOD’95, pp 352–363, 1995.
brevity only results for 64 disks are presented,
[4]M. Kallahalla and P. Varman. Optimal Read-Once
the trend is similar for 8, 16 or 32 disks also.
Parallel Disk Scheduling. In Proc. of ACM Wkshp.
RT-OPT can support substantially more number
on I/O in Parallel and Distributed Systems, 1999.
of clients at a given drop probability. The im-
provement was noticed to be greatest with larger [5]B. Ozden R. Rastogi and A. Silberschatz. Demand
Paging for Video-on-Demand Servers. In Intl. Conf.
number of disks when the need for efficient buffer
on Multimedia Computing and Systems, pp 264–
management becomes more important.
272, 1995.

3500
D = 64 [6]A. L. N. Reddy and J. C. Wyllie. I/O issues in a
Multimedia System. Computer, 27(3):69–74, 1994.
3000 [7]O. Rose. Statistical Properties of MPEG Video
Traffic and their Impact on Traffic Modelling in
2500
ATM Systems. Research Report Series 101, Uni-
Number of users

versity of Wurzburg, Institute of Computer Sci-


2000
ence, February 1995. ftp://ftp-info3.informatik.uni-
wuerzburg.de.
1500

Greedy
[8]P. S. Yu, M-S. Chen, and D. D. Kandlur. Grouped
1000
RT−OPT
Sweeping Scheduling for DASD-based Multime-
0 0.002 0.004 0.006 0.008 0.01 0.012 0.014 0.016 0.018 0.02
Drop probability
dia Storage Management. Multimedia Systems,
1(1):99–109, 1993.
Figure 4: Clients Serviced vs. Drop Probability

Das könnte Ihnen auch gefallen