The NRT-VBR Service ATM SERVICE CATEGORIES

CALL ADMISSION CONTROL CAC 95 This type of queueing structure is known as the ATM multiplexer. It represents a number of ATM sources feeding a finite-capacity queue, which is served by a server, i.e., the output port. The service time is constant and is equal to the time it takes to transmit an ATM cell. Let us assume that the QoS, expressed in cell loss rate, of the existing connections is satisfied. The question that arises is whether the cell loss rate will still be maintained if the new connection is accepted. This can be answered by solving the ATM multiplexer queueing model with the existing connections and the new connection. However, the solution to this problem is CPU intensive and it cannot be done in real-time. In view of this, a variety of different CAC algorithms have been proposed which do not require the solution of such a queueing model. Most of the CAC algorithms that have been proposed are based solely on the cell loss rate QoS parameter. That is, the decision to accept or reject a new connection is based on whether the switch can provide the new connection with the requested cell loss rate without affecting the cell loss rate of the existing connections. No other QoS parameters, such as peak-to-peak cell delay variation and the max CTD, are considered by these algorithms. A very popular example of this type of algorithm is the equivalent bandwidth, described below. CAC algorithms based on the cell transfer delay have also been proposed. In these algorithms, the decision to accept or reject a new connection is based on a calculated absolute upper bound of the end-to-end delay of a cell. These algorithms are closely associated with specific scheduling mechanisms, such as static priorities, early deadline first, and weighted fair queueing. Given that the same scheduling algorithm runs on all of the switches in the path of a connection, it is possible to construct an upper bound of the end-to-end delay. If this is less than the requested end-to-end delay, then the new connection is accepted. Below, we examine the equivalent bandwidth scheme and then we present the ATM block transfer ABT scheme used for bursty sources. In this scheme, bandwidth is allo- cated on demand and only for the duration of a burst. Finally, we present a scheme for controlling the amount of traffic in an ATM network based on virtual path connec- tions VPC.

4.6.2 Equivalent Bandwidth

Let us consider a finite capacity queue served by a server at the rate of µ. This queue can be seen as representing an output port and its buffer in a non-blocking switch with output buffering. Assume that this queue is fed by a single source, and let us calculate its equivalent bandwidth. If we set µ equal to the source’s peak bit rate, then we will observe no accumulation of cells in the buffer. This is because the cells arrive as fast as they are transmitted out. If we slightly reduce the service rate µ, then we will see that cells are beginning to accumulate in the buffer. If we reduce the service rate still a little bit more, then the buffer occupancy will increase. If we keep repeating this experiment each time slightly lowering the service rate, then we will see that the cell loss rate begins to increase. The equivalent bandwidth of the source is defined as the service rate e at which the queue is served that corresponds to a cell loss rate of ε. The equivalent bandwidth of a source falls somewhere between its average bit rate and its peak bit rate. If the source is very bursty, it is closer to its peak bit rate; otherwise, it is closer to its average bit rate. Note that the equivalent bandwidth of a source is not related the source’s SCR.