Network throughput (or just throughput, when in context) refers to the rate of message delivery over a communication channel in a communication network, such as Ethernet or packet radio. The data that these messages contain may be delivered over physical or logical links, or through network nodes. Throughput is usually measured in bits per second (, sometimes abbreviated bps), and sometimes in packets per second ( or pps) or data packets per time slot.
The aggregate throughput is the sum of the data rates that are delivered over all channels in a network.Guowang Miao, Jens Zander, K-W Sung, and Ben Slimane, Fundamentals of Mobile Data Networks, Cambridge University Press, , 2016. Throughput represents digital bandwidth consumption.
The throughput of a communication system may be affected by various factors, including the limitations of the underlying physical medium, available processing power of the system components, end-user behavior, etc. When taking various protocol overheads into account, the useful rate of the data transfer can be significantly lower than the maximum achievable throughput; the useful part is usually referred to as goodput.
Four different values are relevant in the context of maximum throughput are used in comparing the upper limit conceptual performance of multiple systems. They are maximum theoretical throughput, maximum achievable throughput, peak measured throughput, and maximum sustained throughput. These values represent different qualities, and care must be taken that the same definitions are used when comparing different maximum throughput values.
Each bit must carry the same amount of information if throughput values are to be compared. Data compression can significantly alter throughput calculations, including generating values exceeding 100% in some cases.
If the communication is mediated by several links in series with different bit rates, the maximum throughput of the overall link is lower than or equal to the lowest bit rate. The lowest value link in the series is referred to as the bottleneck.
Asymptotic throughput is usually estimated by sending or simulating a very large message (sequence of data packets) through the network, using a greedy source and no flow control mechanism (i.e., UDP rather than TCP), and measuring the volume of data received at the destination node. Traffic load between other sources may reduce this maximum network path throughput. Alternatively, a large number of sources and sinks may be modeled, with or without flow control, and the aggregate maximum network throughput measured (the sum of traffic reaching its destinations). In a network simulation model with infinitely large packet queues, the asymptotic throughput occurs when the network latency (due to packet queuing time) goes to infinity, while if the packet queues are limited, or the network is a multi-drop network with many sources, and collisions may occur, the packet-dropping rate approaches 100%.
A well-known application of asymptotic throughput is in modeling point-to-point communication where message latency is modeled as a function of message length as where is the asymptotic bandwidth and is the half-peak length. Recent Advances in Parallel Virtual Machine and Message Passing Interface by Jack Dongarra, Emilio Luque and Tomas Margalef 1999 page 134
As well as its use in general network modeling, asymptotic throughput is used in modeling performance on massively parallel computer systems, where system operation is highly dependent on communication overhead, as well as processor performance.M. Resch et al. A comparison of MPI performance on different MPPsin Recent Advances in Parallel Virtual Machine and Message Passing Interface, Lecture Notes in Computer Science, 1997, Volume 1332/1997, 25-32 In these applications, asymptotic throughput is used modeling which includes the number of processors, so that both the latency and the asymptotic throughput are functions of the number of processors. High-Performance Computing and Networking edited by Angelo Mañas, Bernardo Tafalla and Rou Rey Jay Pallones 1998 page 935
The channel efficiency, also known as bandwidth utilization efficiency, is the percentage of the net bit rate (in ) of a digital communication channel that goes to the achieved throughput. For example, if the throughput is over a Ethernet connection, the channel efficiency is 70%.
Channel utilization includes both the data bits and the transmission overhead in the channel. The transmission overhead consists of preamble sequences, frame headers and acknowledgment packets. In a simplistic approach, channel efficiency can be equal to channel utilization assuming that acknowledge packets are zero-length and that the communications provider will not see any bandwidth relative to retransmissions or headers. Therefore, certain texts mark a difference between channel utilization and protocol efficiency.
In a point-to-point or point-to-multipoint communication link, where only one terminal is transmitting, the maximum throughput is often equivalent to or very near the physical data rate (the channel capacity), since the channel utilization can be almost 100% in such a network, except for a small inter-frame gap.
For example, the maximum frame size in Ethernet is 1526 bytes: up to 1500 bytes for the payload, eight bytes for the preamble, 14 bytes for the header, and 4 bytes for the trailer. An additional minimum interframe gap corresponding to 12 bytes is inserted after each frame. This corresponds to a maximum channel utilization of 1526 / (1526 + 12) × 100% = 99.22%, or a maximum channel use of inclusive of Ethernet datalink layer protocol overhead over a Ethernet connection. The maximum throughput or channel efficiency is then 1500 / (1526 + 12) = 97.5%, exclusive of the Ethernet protocol overhead.
Despite the conceptual simplicity of digital information, all electrical signals traveling over wires are analog. The analog limitations of wires or wireless systems inevitably provide an upper bound on the amount of information that can be sent. The dominant equation here is the Shannon–Hartley theorem, and analog limitations of this type can be understood as factors that affect either the analog bandwidth of a signal or as factors that affect the signal-to-noise ratio. The bandwidth of twisted pair cabling used by Ethernet is limited to approximately 1 GHz, and PCB traces are limited by a similar amount.
Digital systems refer to the knee frequency,Johnson, 1993, 2-5 the amount of time for the digital voltage to rise from 10% of a nominal digital 0 to a nominal digital 1 or vice versa. The knee frequency is related to the bandwidth of a channel, and can be related to the 3 db bandwidth of a system by the equation:Johnson, 1993, 9 Where Tr is the 10% to 90% rise time, and K is a constant of proportionality related to the pulse shape, equal to 0.35 for an exponential rise, and 0.338 for a Gaussian rise.
Other analog factors include:
CSMA/CD and CSMA/CA backoff waiting time and frame retransmissions after detected collisions slows transmissions. This may occur in Ethernet bus networks and hub networks, as well as in wireless networks.
Flow control, for example, in the Transmission Control Protocol (TCP) protocol, affects the throughput if the bandwidth-delay product is larger than the TCP window. In that case, the sending computer must wait for acknowledgement of the data packets before it can send more packets.
TCP congestion avoidance controls the data rate. A so-called slow start occurs in the beginning of a file transfer, and after packet drops caused by router congestion or bit errors in, for example, wireless links.
Network schedulers in routers and switches ultimately determine how bandwidth is shared. If fair queuing is not provided, users that send large packets will get higher bandwidth. Some users may be prioritized in a weighted fair queuing (WFQ) algorithm if differentiated or guaranteed quality of service (QoS) is provided.
In some communications systems, such as satellite networks, only a finite number of channels may be available to a given user at a given time. Channels are assigned either through preassignment or through Demand Assigned Multiple Access (DAMA).Roddy, 2001, 370 - 371 In these cases, throughput is quantized per channel, and unused capacity on partially utilized channels is lost.
In schemes that include forward error correction codes, the redundant error code is normally excluded from the throughput. An example in modem communication, where the throughput is typically measured in the interface between the Point-to-Point Protocol (PPP) and the circuit-switched modem connection. In this case, the maximum throughput is often called net bit rate or useful bit rate.
To determine the actual data rate of a network or connection, the goodput measurement definition may be used. For example, in file transmission, the goodput corresponds to the file size (in bits) divided by the file transmission time. The goodput is the amount of useful information that is delivered per second to the application layer protocol. Dropped packets or packet retransmissions, as well as protocol overhead, are excluded. Because of that, the goodput is lower than the maximum throughput.
|
|