One document matched: draft-baker-aqm-recommendation-01.xml


<?xml version="1.0" encoding="US-ASCII"?>
<!DOCTYPE rfc SYSTEM "rfc2629.dtd">
<?xml-stylesheet type='text/xsl' href='rfc2629.xslt' ?>
<!-- Some of the more generally applicable PIs that most I-Ds might want to use -->
<!-- Try to enforce the ID-nits conventions and DTD validity -->
<?rfc strict="yes" ?>
<!-- Items used when reviewing the document -->
<?rfc comments="no" ?>
<!-- Controls display of <cref> elements -->
<?rfc inline="no" ?>
<!-- When no, put comments at end in comments section,
                                 otherwise, put inline -->
<?rfc editing="no" ?>
<!-- When yes, insert editing marks: editing marks consist of a 
                                 string such as <29> printed in the blank line at the 
                                 beginning of each paragraph of text. -->
<!-- Create Table of Contents (ToC) and set some options for it.  
         Note the ToC may be omitted for very short documents,but idnits insists on a ToC 
         if the document has more than 15 pages. -->
<?rfc toc="yes"?>
<?rfc tocompact="yes"?>
<!-- If "yes" eliminates blank lines before main section entries. -->
<?rfc tocdepth="3"?>
<!-- Sets the number of levels of sections/subsections... in ToC -->
<!-- Choose the options for the references. 
         Some like symbolic tags in the references (and citations) and others prefer 
         numbers. The RFC Editor always uses symbolic tags.
         The tags used are the anchor attributes of the references. -->
<?rfc symrefs="yes"?>
<?rfc sortrefs="yes" ?>
<!-- If "yes", causes the references to be sorted in order of tags.
                                 This doesn't have any effect unless symrefs is "yes" also. -->
<!-- These two save paper: Just setting compact to "yes" makes savings by not starting each 
         main section on a new page but does not omit the blank lines between list items. 
         If subcompact is also "yes" the blank lines between list items are also omitted. -->
<?rfc compact="yes" ?>
<?rfc subcompact="no" ?>
<!-- end of list of popular I-D processing instructions -->
<!-- end of list of processing instructions -->
<rfc category="bcp" docName="draft-baker-aqm-recommendation-01"
     ipr="trust200902" obsoletes="2309">
  <front>
    <title abbrev="Active Queue Management Recommendations">IETF
    Recommendations Regarding Active Queue Management</title>

    <author fullname="Fred Baker" initials="F." role="editor" surname="Baker">
      <organization>Cisco Systems</organization>

      <address>
        <postal>
          <street></street>

          <city>Santa Barbara</city>

          <code>93117</code>

          <region>California</region>

          <country>USA</country>
        </postal>

        <email>fred@cisco.com</email>
      </address>
    </author>

    <date year="2013" />

    <area>Internet Engineering Task Force</area>

    <workgroup></workgroup>

    <abstract>
      <t>This memo presents recommendations to the Internet community
      concerning measures to improve and preserve Internet performance. It
      presents a strong recommendation for testing, standardization, and
      widespread deployment of active queue management in routers, to improve
      the performance of today's Internet. It also urges a concerted effort of
      research, measurement, and ultimate deployment of router mechanisms to
      protect the Internet from flows that are not sufficiently responsive to
      congestion notification.</t>
<t>
The note largely repeats the recommendations of RFC 2309, updated after fifteen years of experience and new research.
</t>
    </abstract>

    <!--
    <note title="Note to IESG">
      <t></t>
      <t></t>
    </note>
  -->

    <!--
      <texttable anchor="table_example" title="A Very Simple Table">
      <preamble>Tables use ttcol to define column headers and widths.
      Every cell then has a "c" element for its content.</preamble>
          <ttcol align="center">ttcol #1</ttcol>
                                    <ttcol align="center">ttcol #2</ttcol>
                      <c>c #1</c>		<c>c #2</c>
                      <c>c #3</c>		<c>c #4</c>
                      <c>c #5</c>		<c>c #6</c>
      <postamble>which is a very simple example.</postamble>
      </texttable>
    -->
  </front>

  <middle>
    <!--		
      <t>There are multiple list styles: "symbols", "letters", "numbers",
"hanging", "format", etc.</t>
      <t>
	<list style="symbols">
	    <t>First bullet</t>
	    <t>Second bullet</t>
	</list>
     </t>
-->

    <!--
<figure anchor="reference" title="Figure">
<artwork align="center">
<![CDATA[
	ASCII artwork goes here... 
]]>
</artwork>
</figure>
-->

    <section title="Introduction">
      <t>The Internet protocol architecture is based on a connectionless end-
      to-end packet service using the Internet Protocol, whether <xref
      target="RFC0791">IPv4</xref> or <xref target="RFC2460">IPv6</xref>. The
      advantages of its connectionless design, flexibility and robustness,
      have been amply demonstrated. However, these advantages are not without
      cost: careful design is required to provide good service under heavy
      load. In fact, lack of attention to the dynamics of packet forwarding
      can result in severe service degradation or "Internet meltdown". This
      phenomenon was first observed during the early growth phase of the
      Internet of the mid 1980s <xref target="RFC0896"></xref><xref
      target="RFC0970"></xref>, and is technically called "congestive
      collapse".</t>

      <t>The original fix for Internet meltdown was provided by Van Jacobsen.
      Beginning in 1986, Jacobsen developed the congestion avoidance
      mechanisms that are now required in TCP implementations <xref
      target="Jacobson88"></xref> <xref target="RFC1122"></xref>. These
      mechanisms operate in the hosts to cause TCP connections to "back off"
      during congestion. We say that TCP flows are "responsive" to congestion
      signals (i.e., marked or dropped packets) from the network. It is
      primarily these TCP congestion avoidance algorithms that prevent the
      congestive collapse of today's Internet.</t>

      <t>However, that is not the end of the story. Considerable research has
      been done on Internet dynamics since 1988, and the Internet has grown.
      It has become clear that the <xref target="RFC5681">TCP congestion
      avoidance mechanisms</xref>, while necessary and powerful, are not
      sufficient to provide good service in all circumstances. Basically,
      there is a limit to how much control can be accomplished from the edges
      of the network. Some mechanisms are needed in the routers to complement
      the endpoint congestion avoidance mechanisms.</t>

      <t>It is useful to distinguish between two classes of router algorithms
      related to congestion control: "queue management" versus "scheduling"
      algorithms. To a rough approximation, queue management algorithms manage
      the length of packet queues by marking or dropping packets when
      necessary or appropriate, while scheduling algorithms determine which
      packet to send next and are used primarily to manage the allocation of
      bandwidth among flows. While these two router mechanisms are closely
      related, they address rather different performance issues.</t>

      <t>This memo highlights two performance issues. The first issue is the
      need for an advanced form of queue management that we call "active queue
      management." <xref target="Section2"></xref> summarizes the benefits
      that active queue management can bring. A number of Active Queue
      Management procedures are described in the literature, with different
      characteristics. This document does not recommend any of them in
      particular, but does make recommendations that ideally would affect the
      choice of procedure used in a given implementation.</t>

      <t>The second issue, discussed in <xref target="Section4"></xref> of
      this memo, is the potential for future congestive collapse of the
      Internet due to flows that are unresponsive, or not sufficiently
      responsive, to congestion indications. Unfortunately, there is no
      consensus solution to controlling congestion caused by such aggressive
      flows; significant research and engineering will be required before any
      solution will be available. It is imperative that this work be
      energetically pursued, to ensure the future stability of the
      Internet.</t>

      <t><xref target="Section5"></xref> concludes the memo with a set of
      recommendations to the Internet community concerning these topics.</t>

      <t>The discussion in this memo applies to "best-effort" traffic, which
      is to say, traffic generated by applications that accept the occasional
      loss, duplication, or reordering of traffic in flight. It is most
      effective, on time scales of a single RTT or a small number of RTTs, for
      <xref target="RFC1633">elastic traffic</xref>, but also impacts real
      time traffic generated by adaptive applications.</t>

      <t><xref target="RFC2309"></xref> resulted from past discussions of
      end-to-end performance, Internet congestion, and RED in the End-to-End
      Research Group of the Internet Research Task Force (IRTF). This update
      results from experience with that and other algorithms, and the Active
      Queue Management discussion within the IETF.</t>

      <section title="Requirements Language">
        <t>The key words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT",
        "SHOULD", "SHOULD NOT", "RECOMMENDED", "MAY", and "OPTIONAL" in this
        document are to be interpreted as described in <xref
        target="RFC2119"></xref>.</t>
      </section>
    </section>

    <section anchor="Section2" title="The Need For Active Queue Management">
      <t>The traditional technique for managing router queue lengths is to set
      a maximum length (in terms of packets) for each queue, accept packets
      for the queue until the maximum length is reached, then reject (drop)
      subsequent incoming packets until the queue decreases because a packet
      from the queue has been transmitted. This technique is known as "tail
      drop", since the packet that arrived most recently (i.e., the one on the
      tail of the queue) is dropped when the queue is full. This method has
      served the Internet well for years, but it has two important drawbacks.
      <list style="numbers">
          <t>Lock-Out <vspace blankLines="1" /> In some situations tail drop
          allows a single connection or a few flows to monopolize queue space,
          preventing other connections from getting room in the queue. This
          "lock-out" phenomenon is often the result of synchronization or
          other timing effects.</t>

          <t>Full Queues <vspace blankLines="1" /> The tail drop discipline
          allows queues to maintain a full (or, almost full) status for long
          periods of time, since tail drop signals congestion (via a packet
          drop) only when the queue has become full. It is important to reduce
          the steady-state queue size, and this is perhaps queue management's
          most important goal. <vspace blankLines="1" /> The naive assumption
          might be that there is a simple tradeoff between delay and
          throughput, and that the recommendation that queues be maintained in
          a "non-full" state essentially translates to a recommendation that
          low end-to-end delay is more important than high throughput.
          However, this does not take into account the critical role that
          packet bursts play in Internet performance. Even though TCP
          constrains a flow's window size, packets often arrive at routers in
          bursts <xref target="Leland94"></xref>. If the queue is full or
          almost full, an arriving burst will cause multiple packets to be
          dropped. This can result in a global synchronization of flows
          throttling back, followed by a sustained period of lowered link
          utilization, reducing overall throughput. <vspace blankLines="1" />
          The point of buffering in the network is to absorb data bursts and
          to transmit them during the (hopefully) ensuing bursts of silence.
          This is essential to permit the transmission of bursty data. It
          should be clear why we would like to have normally- small queues in
          routers: we want to have queue capacity to absorb the bursts. The
          counter-intuitive result is that maintaining normally-small queues
          can result in higher throughput as well as lower end-to-end delay.
          In short, queue limits should not reflect the steady state queues we
          want maintained in the network; instead, they should reflect the
          size of bursts we need to absorb.</t>
        </list></t>

      <t>Besides tail drop, two alternative queue disciplines that can be
      applied when the queue becomes full are "random drop on full" or "drop
      front on full". Under the random drop on full discipline, a router drops
      a randomly selected packet from the queue (which can be an expensive
      operation, since it naively requires an O(N) walk through the packet
      queue) when the queue is full and a new packet arrives. Under the "drop
      front on full" discipline <xref target="Lakshman96"></xref>, the router
      drops the packet at the front of the queue when the queue is full and a
      new packet arrives. Both of these solve the lock-out problem, but
      neither solves the full-queues problem described above.</t>

      <t>We know in general how to solve the full-queues problem for
      "responsive" flows, i.e., those flows that throttle back in response to
      congestion notification. In the current Internet, dropped packets serve
      as a critical mechanism of congestion notification to end nodes. The
      solution to the full-queues problem is for routers to drop packets
      before a queue becomes full, so that end nodes can respond to congestion
      before buffers overflow. We call such a proactive approach "active queue
      management". By dropping packets before buffers overflow, active queue
      management allows routers to control when and how many packets to
      drop.</t>

      <t>In summary, an active queue management mechanism can provide the
      following advantages for responsive flows. <list style="numbers">
          <t>Reduce number of packets dropped in routers <vspace
          blankLines="1" /> Packet bursts are an unavoidable aspect of packet
          networks <xref target="Willinger95"></xref>. If all the queue space
          in a router is already committed to "steady state" traffic or if the
          buffer space is inadequate, then the router will have no ability to
          buffer bursts. By keeping the average queue size small, active queue
          management will provide greater capacity to absorb naturally-
          occurring bursts without dropping packets. <vspace blankLines="1" />
          Furthermore, without active queue management, more packets will be
          dropped when a queue does overflow. This is undesirable for several
          reasons. First, with a shared queue and the tail drop discipline, an
          unnecessary global synchronization of flows cutting back can result
          in lowered average link utilization, and hence lowered network
          throughput. Second, TCP recovers with more difficulty from a burst
          of packet drops than from a single packet drop. Third, unnecessary
          packet drops represent a possible waste of bandwidth on the way to
          the drop point. <vspace blankLines="1" /> We note that while Active
          Queue Management can manage queue lengths and reduce end- to-end
          latency even in the absence of end-to-end congestion control, Active
          Queue Management will be able to reduce packet dropping only in an
          environment that continues to be dominated by end-to-end congestion
          control.</t>

          <t>Provide lower-delay interactive service <vspace blankLines="1" />
          By keeping the average queue size small, queue management will
          reduce the delays seen by flows. This is particularly important for
          interactive applications such as short Web transfers, Telnet
          traffic, or interactive audio-video sessions, whose subjective (and
          objective) performance is better when the end-to-end delay is
          low.</t>

          <t>Avoid lock-out behavior <vspace blankLines="1" /> Active queue
          management can prevent lock-out behavior by ensuring that there will
          almost always be a buffer available for an incoming packet. For the
          same reason, active queue management can prevent a router bias
          against low bandwidth but highly bursty flows. <vspace
          blankLines="1" /> It is clear that lock-out is undesirable because
          it constitutes a gross unfairness among groups of flows. However, we
          stop short of calling this benefit "increased fairness", because
          general fairness among flows requires per-flow state, which is not
          provided by queue management. For example, in a router using queue
          management but only FIFO scheduling, two TCP flows may receive very
          different bandwidths simply because they have different round-trip
          times <xref target="Floyd91"></xref>, and a flow that does not use
          congestion control may receive more bandwidth than a flow that does.
          Per-flow state to achieve general fairness might be maintained by a
          per-flow scheduling algorithm such as Fair Queueing (FQ) <xref
          target="Demers90"></xref>, or a class-based scheduling algorithm
          such as CBQ <xref target="Floyd95"></xref>, for example. <vspace
          blankLines="1" /> On the other hand, active queue management is
          needed even for routers that use per-flow scheduling algorithms such
          as FQ or class-based scheduling algorithms such as CBQ. This is
          because per-flow scheduling algorithms by themselves do nothing to
          control the overall queue size or the size of individual queues.
          Active queue management is needed to control the overall average
          queue sizes, so that arriving bursts can be accommodated without
          dropping packets. In addition, active queue management should be
          used to control the queue size for each individual flow or class, so
          that they do not experience unnecessarily high delays. Therefore,
          active queue management should be applied across the classes or
          flows as well as within each class or flow. <vspace
          blankLines="1" /> In short, scheduling algorithms and queue
          management should be seen as complementary, not as replacements for
          each other.</t>
        </list></t>
    </section>

    <section anchor="Section4" title="Managing Aggressive Flows">
      <t>One of the keys to the success of the Internet has been the
      congestion avoidance mechanisms of TCP. Because TCP "backs off" during
      congestion, a large number of TCP connections can share a single,
      congested link in such a way that bandwidth is shared reasonably
      equitably among similarly situated flows. The equitable sharing of
      bandwidth among flows depends on the fact that all flows are running
      basically the same congestion avoidance algorithms, conformant with the
      current TCP specification <xref target="RFC1122"></xref>.</t>

      <t>Flows that behaves under congestion like a flow produced by a
      conformant TCP have come to be called <xref target="RFC5348">"TCP
      Friendly"</xref>. A TCP Friendly flow is responsive to congestion
      notification, and in steady-state it uses no more bandwidth than a
      conformant TCP running under comparable conditions (drop rate, RTT, MTU,
      etc.)</t>

      <t>It is convenient to divide flows into three classes: (1) TCP Friendly
      flows, (2) unresponsive flows, i.e., flows that do not slow down when
      congestion occurs, and (3) flows that are responsive but are not TCP
      Friendly. The last two classes contain more aggressive flows that pose
      significant threats to Internet performance, as we will now discuss.
      <list style="symbols">
          <t>Non-Responsive Flows <vspace blankLines="1" /> There is a growing
          set of UDP-based applications whose congestion avoidance algorithms
          are inadequate or nonexistent (i.e, the flow does not throttle back
          upon receipt of congestion notification). Such UDP applications
          include streaming applications like packet voice and video, and also
          multicast bulk data transport <xref target="SRM96"></xref>. If no
          action is taken, such unresponsive flows could lead to a new
          congestive collapse. <vspace blankLines="1" /> In general, all
          UDP-based streaming applications should incorporate effective
          congestion avoidance mechanisms. For example, recent research has
          shown the possibility of incorporating congestion avoidance
          mechanisms such as Receiver- driven Layered Multicast (RLM) within
          UDP-based streaming applications such as packet video <xref
          target="McCanne96"></xref> <xref target="Bolot94"></xref>. Further
          research and development on ways to accomplish congestion avoidance
          for streaming applications will be very important. <vspace
          blankLines="1" /> However, it will also be important for the network
          to be able to protect itself against unresponsive flows, and
          mechanisms to accomplish this must be developed and deployed.
          Deployment of such mechanisms would provide incentive for every
          streaming application to become responsive by incorporating its own
          congestion control.</t>

          <t>Non-TCP-Friendly Transport Protocols <vspace blankLines="1" />
          The second threat is posed by transport protocol implementations
          that are responsive to congestion notification but, either
          deliberately or through faulty implementations, are not TCP
          Friendly. Such applications can grab an unfair share of the network
          bandwidth. <vspace blankLines="1" /> For example, the popularity of
          the Internet has caused a proliferation in the number of TCP
          implementations. Some of these may fail to implement the TCP
          congestion avoidance mechanisms correctly because of poor
          implementation. Others may deliberately be implemented with
          congestion avoidance algorithms that are more aggressive in their
          use of bandwidth than other TCP implementations; this would allow a
          vendor to claim to have a "faster TCP". The logical consequence of
          such implementations would be a spiral of increasingly aggressive
          TCP implementations, leading back to the point where there is
          effectively no congestion avoidance and the Internet is chronically
          congested. <vspace blankLines="1" /> Another example of such flows
          is RTP/UDP video data flows in which the application uses an
          adaptive codec. Such data flows are not responsive to congestion
          signals in a time frame comparable to a small number of end-to-end
          transmission delays. However, over a longer timescale, perhaps
          seconds in duration, they will moderate their speed, or will
          increase their speed if they determine bandwidth to be available.
          <vspace blankLines="1" /> Note that there is a well-known way to
          achieve more aggressive TCP performance without even changing TCP:
          open multiple connections to the same place, as has been done in
          multiple Web browsers and in Peer-to-Peer applications such as
          BitTorrent.</t>
        </list></t>

      <t>The projected increase in more aggressive flows of both these
      classes, as a fraction of total Internet traffic, clearly poses a threat
      to the future Internet. There is an urgent need for measurements of
      current conditions and for further research into the various ways of
      managing such flows. There are many difficult issues in identifying and
      isolating unresponsive or Non-TCP-Friendly flows at an acceptable router
      overhead cost. Finally, there is little measurement or simulation
      evidence available about the rate at which these threats are likely to
      be realized, or about the expected benefit of router algorithms for
      managing such flows.</t>

      <t>There is an issue about the appropriate granularity of a "flow".
      There are a few "natural" answers: 1) a TCP or UDP connection (source
      address/port, destination address/port); 2) a source/destination host
      pair; 3) a given source host or a given destination host. We would guess
      that the source/destination host pair gives the most appropriate
      granularity in many circumstances. However, it is possible that
      different vendors/providers could set different granularities for
      defining a flow (as a way of "distinguishing" themselves from one
      another), or that different granularities could be chosen for different
      places in the network. It may be the case that the granularity is less
      important than the fact that we are dealing with more unresponsive flows
      at *some* granularity. The granularity of flows for congestion
      management is, at least in part, a policy question that needs to be
      addressed in the wider IETF community.</t>
    </section>

    <section anchor="Section5" title="Conclusions and Recommendations">
      <t>The IRTF, in developing <xref target="RFC2309"></xref>, and the IETF
      in subsequent discussion, has developed a set of specific
      recommendations regarding the implementation and operational use of
      Active Queue Management procedures. These include: <list style="numbers">
          <t>Internet routers SHOULD implement some active queue management
          mechanism to manage queue lengths, reduce end-to-end latency, reduce
          packet dropping, and avoid lock-out phenomena within the
          Internet.</t>

          <t>Deployed Active Queue Management SHOULD use ECN as well as loss
          in signaling congestion to endpoints.</t>

          <t>Active Queue Management algorithms deployed SHOULD NOT require
          operational (especially manual) configuration or tuning.</t>

          <t>Active Queue Management algorithms deployed SHOULD be effective
          on all common Internet traffic, including traffic that uses TCP,
          SCTP, UDP, and DCCP as transports.</t>

          <t>TCP and SCTP congestion control algorithms SHOULD maximize their
          use of available bandwidth without incurring loss or undue round
          trip delay when possible.</t>

          <t>It is urgent to continue research, engineering, and measurement
          efforts contributing to the design of mechanisms to deal with flows
          that are unresponsive to congestion notification or are responsive
          but more aggressive than TCP.</t>
        </list></t>

      <t>These recommendations are expressed using the word "SHOULD". This is
      in recognition that there may be use cases unenvisaged in this document
      in which the recommendation does not apply. However, care should be
      taken in concluding that one's use case falls in that category; during
      the life of the Internet, such use cases have been rarely if ever
      observed and reported on. To the contrary, available <xref
      target="Papagiannaki"> research </xref> says that even high speed links
      in network cores that are normally very stable in depth and behavior
      experience occasional issues that need moderation.</t>

      <section anchor="useAQM"
               title="Operational deployments SHOULD implement Active Queue Management procedures">
        <t>In short, Active Queue Management procedures are designed to
        minimize delay induced in the network by queues which have filled as a
        result of host behavior. Marking and loss behaviors signal to the
        senders of data that network buffers are becoming unnecessarily full,
        and they would do well to moderate their behavior.</t>
      </section>

      <section anchor="signaling"
               title="Signaling to the endpoints of a session">
        <t>Means of signaling to an endpoint regarding its effect on the
        network and how it might consider adapting include, at least: <list
            style="symbols">
            <t>Delaying data segments in flight, such as in a queue, which
            affects Ack Clocking and as a result the transmission of new
            data.</t>

            <t>Marking traffic, such as using Explicit Congestion Control<xref
            target="RFC3168"></xref> <xref target="RFC4301"></xref> <xref
            target="RFC4774"></xref> <xref target="RFC6040"></xref> <xref
            target="RFC6679"></xref>.</t>

            <t>Dropping traffic in transit.</t>
          </list></t>

        <t>The use of advanced scheduling mechanisms, such as priority
        queuing, classful queuing, and fair queuing, is often effective in
        networks to help a network to serve the needs of an application. It
        can be used to manage traffic passing a choke point. This is discussed
        in <xref target="RFC2474"></xref> and <xref target="RFC2475"></xref>.
        They are used operationally when an operator considers it important to
        do so.</t>

        <t>Loss has two effects. It protects the network, which is the primary
        reason the network imposes it. Its use as a signal to TCP or SCTP is a
        pragmatic heuristic; "when the network discards a message in flight,
        it may imply the presence of faulty equipment or media in a path, and
        it may imply the presence of congestion. Presume the latter." However,
        it also has an effect on the efficiency of the data flow. The data in
        question must be retransmitted, or its absence must otherwise be
        adapted to by the application in question, which implies at least
        inefficient use of available bandwidth and may affect other data
        flows. Hence, loss is not entirely positive; it is a necessary
        evil.</t>

        <t>Explicit Congestion Control, however, communicates information
        about network congestion that is assuredly about congestion, and
        avoids the unintended consequences of loss.</t>

        <t>Hence, network communication to the host regarding the moderation
        of its traffic flow SHOULD use an AQM algorithm to determine which
        packets it should affect, and then implement that effect by marking
        ECN-capable traffic "Congestion Experienced (CE)" or dropping
        non-ECN-capable traffic.</t>

        <t>Due to the possibility of abuse, the queue must also impose an
        upper bound, so that even ECN-capable traffic experiences tail-drop if
        necessary; this possibility, while equipment must design for the end
        case, should in theory be very uncommon.</t>
      </section>

      <section anchor="autotuning"
               title="Active Queue Management algorithms deployed SHOULD NOT require operational tuning">
        <t>A number of algorithms have been proposed. Many require some form
        of tuning or initial condition, which makes them difficult to use
        operationally. Hence, self-tuning algorithms are to be preferred.</t>
      </section>

      <section anchor="alltraffic"
               title="Active Queue Management algorithms deployed SHOULD be effective on all common Internet traffic">
        <t>Active Queue Management algorithms often target <xref
        target="RFC0793">TCP</xref>, as it is by far the predominant transport
        in the Internet today. However, we have significant use of <xref
        target="RFC0768">UDP</xref> in voice and video services, and find
        utility in <xref target="RFC4960">SCTP</xref> and <xref
        target="RFC4340"> DCCP </xref>. Hence, Active Queue Management
        algorithms that are effective with all of those transports and the
        applications that use them are to be preferred.</t>
      </section>

      <section anchor="tcpcc"
               title="TCP and SCTP congestion control algorithms SHOULD maximize their use of available bandwidth without incurring loss or undue round trip delay">
        <t>The terms "knee" and "cliff" area defined by <xref
        target="Jain94"></xref>. They respectively refer to the minimum and
        maximum values of the effective window that have the effect of
        maximizing transmission rate in a congestion control algorithm such as
        is used by TCP or SCTP. For the sender of data, exceeding the cliff is
        ineffective, as it (by definition) induces loss; operating at a point
        close to the cliff has a negative impact on other traffic and
        applications, triggering operator activities such as discussed in
        <xref target="RFC6057"></xref>.</t>

        <t>Operating below the knee is also ineffective, as it fails to use
        available network capacity. If the objective is to deliver data from
        its source to its recipient in the least possible time, as a result,
        the behavior of any TCP/SCTP congestion control algorithm SHOULD be to
        seek and use effective window values at or above the knee and well
        below the cliff.</t>
      </section>

      <section anchor="research" title="The need for further research">
        <t><xref target="RFC2309"></xref> called for, as its second
        recommendation, further research in the interaction between network
        queues and host applications, and the means of signaling between them.
        This research occurred, and we as a community have learned a lot.
        However, we are not done. An obvious example in 2013 is in the use of
        Map/Reduce applications in data centers; do we need to extend our
        taxonomy of TCP/SCTP sessions to include not only "mice" and
        "elephants", but "lemmings"? "Lemmings" are flash crowds of "mice"
        that the network inadvertently tries to signal to as if they were
        elephant flows, resulting in head of line blocking in data center
        applications.</t>

        <t>Hence, this document reiterates the call: we need continuing
        research as applications develop.</t>
      </section>
    </section>

    <section anchor="IANA" title="IANA Considerations">
      <t>This memo asks the IANA for no new parameters.</t>
    </section>

    <section anchor="Security" title="Security Considerations">
      <t>While security is a very important issue, it is largely orthogonal to
      the performance issues discussed in this memo. We note, however, that
      denial-of-service attacks may create unresponsive traffic flows that are
      indistinguishable from flows from normal high-bandwidth isochronous
      applications, and the mechanism suggested in The recommendation in
      support of ongoing research will be equally applicable to such
      attacks.</t>
    </section>

    <section anchor="Privacy" title="Privacy Considerations">
      <t>This document, by itself, presents no new privacy issues.</t>
    </section>

    <section anchor="Acknowledgements" title="Acknowledgements">
      <t>The original recommendation in <xref target="RFC2309"></xref> was
      written by the End-to-End Research Group, which is to say Bob Braden,
      Dave Clark, Jon Crowcroft, Bruce Davie, Steve Deering, Deborah Estrin,
      Sally Floyd, Van Jacobson, Greg Minshall, Craig Partridge, Larry
      Peterson, KK Ramakrishnan, Scott Shenker, John Wroclawski, and Lixia
      Zhang. This is an edited version of that document, with much of its text
      and arguments unchanged.</t>

      <t>The need for an updated document was agreed to in the tsvarea meeting
      at IETF 86. This document was reviewed on the aqm@ietf.org list. Comments
      came from Colin Perkins, Richard Scheffenegger, and Dave Taht.</t>
    </section>
  </middle>

  <back>
    <!-- references split to informative and normative -->

    <references title="Normative References">
      <?rfc include="reference.RFC.2119"?>

      <?rfc include="reference.RFC.3168" ?>

      <?rfc include="reference.RFC.6679" ?>

      <?rfc include="reference.RFC.4301" ?>

      <?rfc include="reference.RFC.4774" ?>

      <?rfc include="reference.RFC.6040" ?>
    </references>

    <references title="Informative References">
      <?rfc include="reference.RFC.0768" ?>

      <?rfc include="reference.RFC.0791" ?>

      <?rfc include="reference.RFC.0793" ?>

      <?rfc include="reference.RFC.0896" ?>

      <?rfc include="reference.RFC.0970" ?>

      <?rfc include="reference.RFC.1122" ?>

      <?rfc include='reference.RFC.1633'?>

      <?rfc include="reference.RFC.2309"?>

      <?rfc include="reference.RFC.2460" ?>

      <?rfc include="reference.RFC.2474" ?>

      <?rfc include="reference.RFC.2475" ?>

      <?rfc include="reference.RFC.4340" ?>

      <?rfc include="reference.RFC.4960" ?>

      <?rfc include='reference.RFC.5348'?>

      <?rfc include="reference.RFC.5681" ?>

      <?rfc include="reference.RFC.6057" ?>

      <reference anchor="Floyd91">
        <front>
          <title>Connections with Multiple Congested Gateways in
          Packet-Switched Networks Part 1: One-way Traffic.</title>

          <author fullname="S. Floyd" initials="S" surname="Floyd">
            <organization></organization>
          </author>

          <date month="October" year="1991" />
        </front>

        <seriesInfo name="Computer Communications Review" value="" />
      </reference>

      <reference anchor="Floyd95">
        <front>
          <title>Link-sharing and Resource Management Models for Packet
          Networks</title>

          <author fullname="S. Floyd" initials="S" surname="Floyd">
            <organization></organization>
          </author>

          <author fullname="Van Jacobson" initials="V" surname="Jacobson">
            <organization></organization>
          </author>

          <date month="August" year="1995" />
        </front>

        <seriesInfo name="IEEE/ACM Transactions on Networking" value="" />
      </reference>

      <reference anchor="Demers90">
        <front>
          <title>Analysis and Simulation of a Fair Queueing Algorithm,
          Internetworking: Research and Experience</title>

          <author fullname="A. Demers" initials="A" surname="Demers">
            <organization></organization>
          </author>

          <author fullname="S. Keshav" initials="S" surname="Keshav">
            <organization></organization>
          </author>

          <author fullname="S. Shenker" initials="S" surname="Shenker">
            <organization></organization>
          </author>

          <date year="1990" />
        </front>

        <seriesInfo name="SIGCOMM Symposium proceedings on Communications architectures and protocols"
                    value="" />
      </reference>

      <reference anchor="SRM96">
        <front>
          <title>A Reliable Multicast Framework for Light-weight Sessions and
          Application Level Framing</title>

          <author fullname="S. Floyd" initials="S" surname="Floyd">
            <organization></organization>
          </author>

          <author fullname="Van Jacobson" initials="V" surname="Jacobson">
            <organization></organization>
          </author>

          <author fullname="S. McCanne" initials="S" surname="McCanne">
            <organization></organization>
          </author>

          <author fullname="R. Liu" initials="C" surname="Liu">
            <organization></organization>
          </author>

          <author fullname="L. Zhang" initials="L" surname="Zhang">
            <organization></organization>
          </author>

          <date year="1996" />
        </front>

        <seriesInfo name="SIGCOMM Symposium proceedings on Communications architectures and protocols"
                    value="" />
      </reference>

      <reference anchor="McCanne96">
        <front>
          <title>Receiver-driven Layered Multicast</title>

          <author fullname="W. Willinger" initials="S" surname="McCanne">
            <organization></organization>
          </author>

          <author fullname="Van Jacobson" initials="V" surname="Jacobson">
            <organization></organization>
          </author>

          <author fullname="M. Vetterli" initials="M" surname="Vetterli">
            <organization></organization>
          </author>

          <date month="August" year="1996" />
        </front>

        <seriesInfo name="SIGCOMM Symposium proceedings on Communications architectures and protocols"
                    value="" />
      </reference>

      <reference anchor="Bolot94">
        <front>
          <title>Scalable Feedback Control for Multicast Video Distribution in
          the Internet</title>

          <author fullname="W. Bolot" initials="JC" surname="Bolot">
            <organization></organization>
          </author>

          <author fullname="M. Turletti" initials="T" surname="Turletti">
            <organization></organization>
          </author>

          <author fullname="R. Wakeman" initials="T" surname="Wakeman">
            <organization></organization>
          </author>

          <date month="August" year="1994" />
        </front>

        <seriesInfo name="SIGCOMM Symposium proceedings on Communications architectures and protocols"
                    value="" />
      </reference>

      <reference anchor="Willinger95">
        <front>
          <title>Self-Similarity Through High-Variability: Statistical
          Analysis of Ethernet LAN Traffic at the Source Level</title>

          <author fullname="W. Willinger" initials="W" surname="Willinger">
            <organization></organization>
          </author>

          <author fullname="M. Taqqu" initials="M" surname="Taqqu">
            <organization></organization>
          </author>

          <author fullname="R." initials="R" surname="Sherman">
            <organization></organization>
          </author>

          <author fullname="D.V. Wilson" initials="D" surname="Wilson">
            <organization></organization>
          </author>

          <author fullname="Van Jacobson" initials="V" surname="Jacobson">
            <organization></organization>
          </author>

          <date month="August" year="1995" />
        </front>

        <seriesInfo name="SIGCOMM Symposium proceedings on Communications architectures and protocols"
                    value="" />
      </reference>

      <reference anchor="Jacobson88">
        <front>
          <title>Congestion Avoidance and Control</title>

          <author fullname="Van Jacobson" initials="V" surname="Jacobson">
            <organization>Lawrence Berkeley Network Labs</organization>
          </author>

          <date month="August" year="1988" />
        </front>

        <seriesInfo name="SIGCOMM Symposium proceedings on Communications architectures and protocols"
                    value="" />
      </reference>

      <reference anchor="Lakshman96">
        <front>
          <title>The Drop From Front Strategy in TCP Over ATM and Its
          Interworking with Other Control Features</title>

          <author fullname="T. V. Lakshman" initials="TV" surname="Lakshman">
            <organization></organization>
          </author>

          <author fullname="Arnie Neidhardt" initials="A" surname="Neidhardt">
            <organization></organization>
          </author>

          <author fullname="Teunis Ott" initials="T" surname="Ott">
            <organization></organization>
          </author>

          <date year="1996" />
        </front>

        <seriesInfo name="IEEE Infocomm" value="" />
      </reference>

      <reference anchor="Leland94">
        <front>
          <title>On the Self-Similar Nature of Ethernet Traffic (Extended
          Version)</title>

          <author fullname="W. Leland" initials="W" surname="Leland">
            <organization></organization>
          </author>

          <author fullname="M. Taqqu" initials="M" surname="Taqqu">
            <organization></organization>
          </author>

          <author fullname="W. Willinger" initials="W" surname="Willinger">
            <organization></organization>
          </author>

          <author fullname="D. Wilson" initials="D" surname="Wilson">
            <organization></organization>
          </author>

          <date month="February" year="1994" />
        </front>

        <seriesInfo name="IEEE/ACM Transactions on Networking" value="" />
      </reference>

      <reference anchor="Jain94">
        <front>
          <title>Congestion avoidance scheme for computer networks</title>

          <author fullname="Rajendra K. Jain" initials="Raj" surname="Jain">
            <organization>Digital Equipment Corporation</organization>
          </author>

          <author fullname="KK Ramakrishnan" initials="KK"
                  surname="Ramakrishnan">
            <organization>Digital Equipment Corporation</organization>
          </author>

          <author fullname="Chiu Dah-Ming" initials="Chiu" surname="Dah-Ming">
            <organization>Digital Equipment Corporation</organization>
          </author>

          <date day="27" month="December" year="1994" />
        </front>

        <seriesInfo name="US Patent Office" value="5377327" />
      </reference>

      <reference anchor="Papagiannaki"
                 target="http://www.ieee-infocom.org/2004/Papers/37_4.PDF">
        <front>
          <title>Analysis of Point-To-Point Packet Delay In an Operational
          Network</title>

          <author fullname="B-Y. Choi">
            <organization>Sprint ATL</organization>
          </author>

          <author fullname="S. Moon">
            <organization>KAIST</organization>
          </author>

          <author fullname="Z-L. Zhang">
            <organization>University of Minnesota</organization>
          </author>

          <author fullname="K. Papagiannaki">
            <organization>Sprint ATL</organization>
          </author>

          <author fullname="C. Diot">
            <organization>Intel Research</organization>
          </author>

          <date month="March" year="2004" />
        </front>

        <seriesInfo name="IEEE Infocom" value="2004" />
      </reference>
    </references>

    <section anchor="log" title="Change Log">
      <t><list style="hanging">
          <t hangText="Initial Version:">March 2013</t>

          <t hangText="Minor update:">April 2013</t>
        </list></t>
    </section>
  </back>
</rfc>

PAFTECH AB 2003-20262026-04-24 01:19:39