Burstable billing

Last updated

Burstable billing is a method of measuring bandwidth based on peak use. It allows usage to exceed a specified threshold for brief periods of time without the financial penalty of purchasing a higher committed information rate (CIR, or commitment) from an Internet service provider (ISP).

Contents

Most ISPs use a five-minute sampling and 95% usage when calculating usage.

95th percentile

95th percentile measurement on a regular bandwidth pattern Bandwidth 95th percentile.png
95th percentile measurement on a regular bandwidth pattern

The 95th percentile is a widely used mathematical calculation to evaluate the regular and sustained use of a network connection. The 95th percentile method more closely reflects the needed capacity of the link in question than tracking by other methods such as mean or maximum rate. The bytes that make up the packets themselves do not actually cost money, but the link and the infrastructure on either end of the link cost money to set up and support. This method of billing is commonly used in peering arrangements between corporate networks; it is not often used by ISPs because such entities need committed information rates (CIRs) for planning purposes.

Since most networks are overprovisioned, there is often some room for some bursting without advanced planning (hence burstable billing). Ignoring the top 5% of the samples is a reasonable compromise in most cases (hence 95th percentile).

Many sites have the majority of their traffic on Mondays, so the Monday traffic determines the rate for the whole month. Some providers offer billing on the 90th percentile as an incentive to attract customers with irregular bandwidth patterns. [1]

The 95th percentile allows a customer to have a short (less than 36 hours, given a monthly billing period) burst in traffic without overage charges. The 95th percentile says that 95% of the time, the usage is at or below this amount. Conversely, 5% of the samples may be bursting above this rate.

The sampling interval, or how often samples (or data points ) are taken, is an important factor in percentile calculation. A percentile is calculated on some set of data points. Every data point represents the average bandwidth used during the sampling interval (e.g., five minutes) and is calculated as the number of bits transferred throughout the interval divided by the duration of the interval (e.g., 300 seconds). The resulting value represents the average use rate for a single sampling interval and is expressed as bits per second (see data transfer rate).

Burstable rate calculation

Mrtg95.png

Bandwidth is measured (or sampled) from the switch or router and recorded in a log file. In most cases, this is done every 5 minutes. At the end of the month, the samples are sorted from highest to lowest, and the top 5% (which equal to approximately 36 hours of a 30-day billing cycle) of data is thrown away. The next highest measurement becomes the billable use for the entire month.

Based on this model, the top 36 hours (top 5% of 720 hours) of peak traffic is not taken into account when billed for an entire month. Bandwidth could be used at a higher rate for up to 72 min a day with no financial penalty. Conversely, if peak traffic only appears for a brief instant and no additional traffic is generated the billing amount can be substantially higher than Average usage billing.

Special consideration

Inbound and outbound traffic is usually counted separately, as connections are full duplex allowing traffic in-bound and out-bound simultaneously. Some common algorithms are:

Critics of the 95th percentile billing method usually advocate the use of a flat rate system or using the average throughput rather than the 95th percentile. Both those methods favour heavy users (who have interest in advocating for changes to billing method). Other critics call for billing per byte of data transferred, which is considered most accurate and fair.

See also

Related Research Articles

Network throughput refers to the rate of message delivery over a communication channel, such as Ethernet or packet radio, in a communication network. The data that these messages contain may be delivered over physical or logical links, or through network nodes. Throughput is usually measured in bits per second, and sometimes in data packets per second or data packets per time slot.

<span class="mw-page-title-main">Frame Relay</span> Wide area network technology

Frame Relay is a standardized wide area network (WAN) technology that specifies the physical and data link layers of digital telecommunications channels using a packet switching methodology. Originally designed for transport across Integrated Services Digital Network (ISDN) infrastructure, it may be used today in the context of many other network interfaces.

In telecommunications and computing, bit rate is the number of bits that are conveyed or processed per unit of time.

In statistics, a k-thpercentile, also known as percentile score or centile, is a score below which a given percentage k of scores in its frequency distribution falls or a score at or below which a given percentage falls. Percentiles are expressed in the same unit of measurement as the input scores, not in percent; for example, if the scores refer to human weight, the corresponding percentiles will be expressed in kilograms or pounds. In the limit of an infinite sample size, the percentile approximates the percentile function, the inverse of the cumulative distribution function.

<span class="mw-page-title-main">Leaky bucket</span> Network traffic shaping and policing algorithm

The leaky bucket is an algorithm based on an analogy of how a bucket with a constant leak will overflow if either the average rate at which water is poured in exceeds the rate at which the bucket leaks or if more water than the capacity of the bucket is poured in all at once. It can be used to determine whether some sequence of discrete events conforms to defined limits on their average and peak rates or frequencies, e.g. to limit the actions associated with these events to these rates or delay them until they do conform to the rates. It may also be used to check conformance or limit to an average rate alone, i.e. remove any variation from the average.

The token bucket is an algorithm used in packet-switched and telecommunications networks. It can be used to check that data transmissions, in the form of packets, conform to defined limits on bandwidth and burstiness. It can also be used as a scheduling algorithm to determine the timing of transmissions that will comply with the limits set for the bandwidth and burstiness: see network scheduler.

If a network service wishes to use a broadband network to transport a particular kind of traffic, it must first inform the network about what kind of traffic is to be transported, and the performance requirements of that traffic. The application presents this information to the network in the form of a traffic contract.

The generic cell rate algorithm (GCRA) is a leaky bucket-type scheduling algorithm for the network scheduler that is used in Asynchronous Transfer Mode (ATM) networks. It is used to measure the timing of cells on virtual channels (VCs) and or Virtual Paths (VPs) against bandwidth and jitter limits contained in a traffic contract for the VC or VP to which the cells belong. Cells that do not conform to the limits given by the traffic contract may then be re-timed (delayed) in traffic shaping, or may be dropped (discarded) or reduced in priority (demoted) in traffic policing. Nonconforming cells that are reduced in priority may then be dropped, in preference to higher priority cells, by downstream components in the network that are experiencing congestion. Alternatively they may reach their destination if there is enough capacity for them, despite them being excess cells as far as the contract is concerned: see priority control.

Bandwidth throttling consists in the intentional limitation of the communication speed, of the ingoing (received) or outgoing (sent) data in a network node or in a network device.

A dedicated hosting service, dedicated server, or managed hosting service is a type of Internet hosting in which the client leases an entire server not shared with anyone else. This is more flexible than shared hosting, as organizations have full control over the server(s), including choice of operating system, hardware, etc.

Memory bandwidth is the rate at which data can be read from or stored into a semiconductor memory by a processor. Memory bandwidth is usually expressed in units of bytes/second, though this can vary for systems with natural data sizes that are not a multiple of the commonly used 8-bit bytes.

Average Per-Bit Delivery Cost, or APBDC, is the cost accounting method by which Internet Service Providers (ISPs) calculate their cost of goods sold.

The Media Delivery Index (MDI) is a set of measures that can be used to monitor both the quality of a delivered video stream as well as to show system margin for IPTV systems by providing an accurate measurement of jitter and delay at network level (Internet Protocol, IP), which are the main causes for quality loss. Identifying and quantizing such problems in this kind of networks is key to maintaining high quality video delivery and providing indications that warn system operators with enough advance notice to allow corrective action.

In computing, bandwidth is the maximum rate of data transfer across a given path. Bandwidth may be characterized as network bandwidth, data bandwidth, or digital bandwidth.

User-in-the-Loop (UIL) refers to the notion that a technology can improve a performance objective by engaging its human users. The idea can be applied in various technological fields. UIL assumes that human users of a network are among the smartest but also most unpredictable units of that network. Furthermore, human users often have a certain set of (input) values that they sense. Both elements of smart decision-making and observed values can help towards improving the bigger objective.

Usage Parameter Control (UPC) and Network Parameter Control (NPC) are functions that may be performed in a computer network. UPC may be performed at the input to a network "to protect network resources from malicious as well as unintentional misbehaviour". NPC is the same and done for the same reasons as UPC, but at the interface between two networks.

CoDel is an active queue management (AQM) algorithm in network routing, developed by Van Jacobson and Kathleen Nichols and published as RFC8289. It is designed to overcome bufferbloat in networking hardware, such as routers, by setting limits on the delay network packets experience as they pass through buffers in this equipment. CoDel aims to improve on the overall performance of the random early detection (RED) algorithm by addressing some of its fundamental misconceptions, as perceived by Jacobson, and by being easier to manage.

Internet rush hour is the time period when the majority of Internet users are online at the same time. Typically, in the UK the peak hours are between 7 and 11 pm. During this time frame, users commonly experience slowness while browsing or downloading content. The congestion experienced during the rush hour is similar to transportation rush hour, where demand for resources outweighs capacity.

Time-Sensitive Networking (TSN) is a set of standards under development by the Time-Sensitive Networking task group of the IEEE 802.1 working group. The TSN task group was formed in November 2012 by renaming the existing Audio Video Bridging Task Group and continuing its work. The name changed as a result of the extension of the working area of the standardization group. The standards define mechanisms for the time-sensitive transmission of data over deterministic Ethernet networks.

The Zettabyte Era or Zettabyte Zone is a period of human and computer science history that started in the mid-2010s. The precise starting date depends on whether it is defined as when the global IP traffic first exceeded one zettabyte, which happened in 2016, or when the amount of digital data in the world first exceeded a zettabyte, which happened in 2012. A zettabyte is a multiple of the unit byte that measures digital storage, and it is equivalent to 1,000,000,000,000,000,000,000 (1021) bytes.

References

  1. Goldman, Alex. "Cogent's Latest Price Drop" Archived July 20, 2008, at the Wayback Machine , ISP PLANET, April 3, 2006. Accessed April 24, 2008.