Extensible Prioritization Scheme for HTTPFastlykazuhooku@gmail.comCloudflarelucaspardue.24.7@gmail.com
Applications and Real-Time
HTTPResponse priorityStream multiplexingReprioritizationServer schedulingThis document describes a scheme that allows an HTTP client to communicate its
preferences for how the upstream server prioritizes responses to its requests,
and also allows a server to hint to a downstream intermediary how its responses
should be prioritized when they are forwarded. This document defines the
Priority header field for communicating the initial priority in an HTTP
version-independent manner, as well as HTTP/2 and HTTP/3 frames for
reprioritizing responses. These share a common format structure that is designed
to provide future extensibility.Status of This Memo
This is an Internet Standards Track document.
This document is a product of the Internet Engineering Task Force
(IETF). It represents the consensus of the IETF community. It has
received public review and has been approved for publication by
the Internet Engineering Steering Group (IESG). Further
information on Internet Standards is available in Section 2 of
RFC 7841.
Information about the current status of this document, any
errata, and how to provide feedback on it may be obtained at
.
Copyright Notice
Copyright (c) 2022 IETF Trust and the persons identified as the
document authors. All rights reserved.
This document is subject to BCP 78 and the IETF Trust's Legal
Provisions Relating to IETF Documents
() in effect on the date of
publication of this document. Please review these documents
carefully, as they describe your rights and restrictions with
respect to this document. Code Components extracted from this
document must include Revised BSD License text as described in
Section 4.e of the Trust Legal Provisions and are provided without
warranty as described in the Revised BSD License.
Table of Contents
. Introduction
. Notational Conventions
. Motivation for Replacing RFC 7540 Stream Priorities
. Disabling RFC 7540 Stream Priorities
. Advice when Using Extensible Priorities as the Alternative
. Applicability of the Extensible Priority Scheme
. Priority Parameters
. Urgency
. Incremental
. Defining New Priority Parameters
. Registration
. The Priority HTTP Header Field
. Reprioritization
. The PRIORITY_UPDATE Frame
. HTTP/2 PRIORITY_UPDATE Frame
. HTTP/3 PRIORITY_UPDATE Frame
. Merging Client- and Server-Driven Priority Parameters
. Client Scheduling
. Server Scheduling
. Intermediaries with Multiple Backend Connections
. Scheduling and the CONNECT Method
. Retransmission Scheduling
. Fairness
. Coalescing Intermediaries
. HTTP/1.x Back Ends
. Intentional Introduction of Unfairness
. Why Use an End-to-End Header Field?
. Security Considerations
. IANA Considerations
. References
. Normative References
. Informative References
Acknowledgements
Authors' Addresses
IntroductionIt is common for representations of an HTTP
resource to have relationships to one or more other resources. Clients will
often discover these relationships while processing a retrieved representation,
which may lead to further retrieval requests. Meanwhile, the nature of the
relationships determines whether a client is blocked from continuing to process
locally available resources. An example of this is the visual rendering of an HTML
document, which could be blocked by the retrieval of a Cascading Style Sheets (CSS) file that the
document refers to. In contrast, inline images do not block rendering and get drawn
incrementally as the chunks of the images arrive.HTTP/2 and HTTP/3
support multiplexing of requests and responses in
a single connection. An important feature of any implementation of a protocol
that provides multiplexing is the ability to prioritize the sending of
information. For example, to provide meaningful presentation of an HTML document
at the earliest moment, it is important for an HTTP server to prioritize the
HTTP responses, or the chunks of those HTTP responses, that it sends to a
client.HTTP/2 and HTTP/3 servers can schedule transmission of concurrent response data
by any means they choose. Servers can ignore client priority signals and still
successfully serve HTTP responses. However, servers that operate in ignorance
of how clients issue requests and consume responses can cause suboptimal client
application performance. Priority signals allow clients to communicate their
view of request priority. Servers have their own needs that are independent of
client needs, so they often combine priority signals with other available
information in order to inform scheduling of response data.RFC 7540 stream priority allowed a client to send a series of
priority signals that communicate to the server a "priority tree"; the structure
of this tree represents the client's preferred relative ordering and weighted
distribution of the bandwidth among HTTP responses. Servers could use these
priority signals as input into prioritization decisions.The design and implementation of RFC 7540 stream priority were observed to have
shortcomings, as explained in . HTTP/2
has consequently deprecated the use of
these stream priority signals. The prioritization scheme and priority signals
defined herein can act as a substitute for RFC 7540 stream priority.This document describes an extensible scheme for prioritizing HTTP responses
that uses absolute values. defines priority parameters, which are
a standardized and extensible format of priority information.
defines the Priority HTTP header field, which is an
end-to-end priority signal that is independent of protocol version. Clients can send this header field to signal their
view of how responses should be prioritized. Similarly, servers behind an
intermediary can use it to signal priority to the intermediary. After sending a
request, a client can change their view of response priority (see
) by sending HTTP-version-specific frames as defined in
Sections and .Header field and frame priority signals are input to a server's response
prioritization process. They are only a suggestion and do not guarantee any
particular processing or transmission order for one response relative to any
other response. Sections and provide
considerations and guidance about how servers might act upon signals.Notational ConventionsThe key words "MUST", "MUST NOT",
"REQUIRED", "SHALL",
"SHALL NOT", "SHOULD",
"SHOULD NOT",
"RECOMMENDED", "NOT RECOMMENDED",
"MAY", and "OPTIONAL" in this document
are to be interpreted as described in BCP 14
when, and only
when, they appear in all capitals, as shown here.This document uses the following terminology from to specify syntax and parsing: "Boolean", "Dictionary", and "Integer".Example HTTP requests and responses use the HTTP/2-style formatting from
.This document uses the variable-length integer encoding from
.The term "control stream" is used to describe both the HTTP/2 stream with
identifier 0x0 and the HTTP/3 control stream; see .The term "HTTP/2 priority signal" is used to describe the priority information
sent from clients to servers in HTTP/2 frames; see .Motivation for Replacing RFC 7540 Stream PrioritiesRFC 7540 stream priority (see ) is a complex system
where clients signal stream dependencies and weights to describe an unbalanced
tree. It suffered from limited deployment and interoperability and has been deprecated
in a revision of HTTP/2 . HTTP/2 retains these protocol elements in
order to maintain wire compatibility (see ), which
means that they might still be used even in the presence of alternative signaling,
such as the scheme this document describes.Many RFC 7540 server implementations do not act on HTTP/2 priority
signals.Prioritization can use information that servers have about resources or
the order in which requests are generated. For example, a server, with knowledge
of an HTML document structure, might want to prioritize the delivery of images
that are critical to user experience above other images. With RFC 7540, it is
difficult for servers to interpret signals from clients for prioritization, as
the same conditions could result in very different signaling from different
clients. This document describes signaling that is simpler and more constrained,
requiring less interpretation and allowing less variation.RFC 7540 does not define a method that can be used by a server to provide a
priority signal for intermediaries.RFC 7540 stream priority is expressed relative to other requests sharing the same
connection at the same time. It is difficult to incorporate such a design into
applications that generate requests without knowledge of how other requests
might share a connection, or into protocols that do not have strong ordering
guarantees across streams, like HTTP/3 .Experiments from independent research have shown
that simpler schemes can reach at least equivalent performance characteristics
compared to the more complex RFC 7540 setups seen in practice, at least for the
Web use case.Disabling RFC 7540 Stream PrioritiesThe problems and insights set out above provided the motivation for an
alternative to RFC 7540 stream priority (see ).The SETTINGS_NO_RFC7540_PRIORITIES HTTP/2 setting is defined by this document in
order to allow endpoints to omit or ignore HTTP/2 priority signals (see
), as described below. The value of
SETTINGS_NO_RFC7540_PRIORITIES MUST be 0 or 1. Any value other than 0 or 1 MUST
be treated as a connection error (see ) of type
PROTOCOL_ERROR. The initial value is 0.If endpoints use SETTINGS_NO_RFC7540_PRIORITIES, they MUST send it in the first
SETTINGS frame. Senders MUST NOT change the SETTINGS_NO_RFC7540_PRIORITIES value
after the first SETTINGS frame. Receivers that detect a change MAY treat it as a
connection error of type PROTOCOL_ERROR.Clients can send SETTINGS_NO_RFC7540_PRIORITIES with a value of 1 to indicate
that they are not using HTTP/2 priority signals. The SETTINGS frame precedes any
HTTP/2 priority signal sent from clients, so servers can determine whether they
need to allocate any resources to signal handling before signals arrive. A
server that receives SETTINGS_NO_RFC7540_PRIORITIES with a value of 1 MUST
ignore HTTP/2 priority signals.Servers can send SETTINGS_NO_RFC7540_PRIORITIES with a value of 1 to indicate
that they will ignore HTTP/2 priority signals sent by clients.Endpoints that send SETTINGS_NO_RFC7540_PRIORITIES are encouraged to use
alternative priority signals (for example, see or
), but there is no requirement to use a specific signal type.Advice when Using Extensible Priorities as the AlternativeBefore receiving a SETTINGS frame from a server, a client does not know if the server
is ignoring HTTP/2 priority signals. Therefore, until the client receives the
SETTINGS frame from the server, the client SHOULD send both the HTTP/2
priority signals and the signals of this prioritization scheme (see
Sections and ).Once the client receives the first SETTINGS frame that contains the
SETTINGS_NO_RFC7540_PRIORITIES parameter with a value of 1, it SHOULD stop sending
the HTTP/2 priority signals. This avoids sending redundant signals that are
known to be ignored.Similarly, if the client receives SETTINGS_NO_RFC7540_PRIORITIES with a value of 0
or if the settings parameter was absent, it SHOULD stop sending PRIORITY_UPDATE
frames (), since those frames are likely to be ignored.
However, the client MAY continue sending the Priority header field
(), as it is an end-to-end signal that might be useful to nodes
behind the server that the client is directly connected to.Applicability of the Extensible Priority SchemeThe priority scheme defined by this document is primarily focused on the
prioritization of HTTP response messages (see ). It
defines new priority parameters () and a means of conveying those
parameters (Sections and ), which is intended to communicate
the priority of responses to a server that is responsible for prioritizing
them. provides considerations for servers about acting on
those signals in combination with other inputs and factors.The CONNECT method (see ) can be used to establish
tunnels. Signaling applies similarly to tunnels; additional considerations for
server prioritization are given in . describes how clients can optionally apply elements of
this scheme locally to the request messages that they generate.Some forms of HTTP extensions might change HTTP/2 or HTTP/3 stream behavior or
define new data carriage mechanisms. Such extensions can themselves define
how this priority scheme is to be applied.Priority ParametersThe priority information is a sequence of key-value pairs, providing room for
future extensions. Each key-value pair represents a priority parameter.The Priority HTTP header field () is an end-to-end way to
transmit this set of priority parameters when a request or a response is issued.
After sending a request, a client can change their view of response priority
() by sending HTTP-version-specific PRIORITY_UPDATE frames as
defined in Sections and . Frames transmit priority
parameters on a single hop only.Intermediaries can consume and produce priority signals in a PRIORITY_UPDATE
frame or Priority header field. An intermediary that passes only the Priority
request header field to the next hop preserves the original end-to-end signal
from the client; see .
An intermediary could pass the Priority header field and additionally send a PRIORITY_UPDATE frame. This would have the effect of preserving the original client end-to-end signal, while instructing the next hop to use a different priority, per the guidance in . An intermediary that replaces or adds a Priority request header field overrides the original client end-to-end signal, which can affect prioritization for all subsequent recipients of the request.For both the Priority header field and the PRIORITY_UPDATE frame, the set of
priority parameters is encoded as a Dictionary (see
).This document defines the urgency (u) and incremental (i) priority parameters.
When receiving an HTTP request that does not carry these priority parameters, a
server SHOULD act as if their default values were specified.An intermediary can combine signals from requests and responses that it forwards.
Note that omission of priority parameters in responses is handled differently from
omission in requests; see .Receivers parse the Dictionary as described in . Where the Dictionary is successfully parsed, this document
places the additional requirement that unknown priority parameters, priority
parameters with out-of-range values, or values of unexpected types MUST be
ignored.UrgencyThe urgency (u) parameter value is Integer (see ), between 0 and 7 inclusive, in descending order of priority. The default is 3.Endpoints use this parameter to communicate their view of the precedence of
HTTP responses. The chosen value of urgency can be based on the expectation that
servers might use this information to transmit HTTP responses in the order of
their urgency. The smaller the value, the higher the precedence.The following example shows a request for a CSS file with the urgency set to
0:
:method = GET
:scheme = https
:authority = example.net
:path = /style.css
priority = u=0
A client that fetches a document that likely consists of multiple HTTP resources
(e.g., HTML) SHOULD assign the default urgency level to the main resource. This
convention allows servers to refine the urgency using
knowledge specific to the website (see ).The lowest urgency level (7) is reserved for background tasks such as delivery
of software updates. This urgency level SHOULD NOT be used for fetching
responses that have any impact on user interaction.IncrementalThe incremental (i) parameter value is Boolean (see ). It indicates
if an HTTP response can be processed incrementally, i.e., provide some
meaningful output as chunks of the response arrive.The default value of the incremental parameter is false (0).If a client makes concurrent requests with the incremental parameter set to
false, there is no benefit in serving responses with the same urgency concurrently
because the client is not going to process those responses incrementally.
Serving non-incremental responses with the same urgency one by one, in the order in which those
requests were generated, is considered to be the best strategy.If a client makes concurrent requests with the incremental parameter set to
true, serving requests with the same urgency concurrently might be beneficial.
Doing this distributes the connection bandwidth, meaning that responses take
longer to complete. Incremental delivery is most useful where multiple
partial responses might provide some value to clients ahead of a
complete response being available.The following example shows a request for a JPEG file with the urgency parameter
set to 5 and the incremental parameter set to true.
:method = GET
:scheme = https
:authority = example.net
:path = /image.jpg
priority = u=5, i
Defining New Priority ParametersWhen attempting to define new priority parameters, care must be taken so that
they do not adversely interfere with prioritization performed by existing
endpoints or intermediaries that do not understand the newly defined priority
parameters. Since unknown priority parameters are ignored, new priority
parameters should not change the interpretation of, or modify, the urgency (see
) or incremental (see ) priority parameters in a way
that is not backwards compatible or fallback safe.For example, if there is a need to provide more granularity than eight urgency
levels, it would be possible to subdivide the range using an additional priority
parameter. Implementations that do not recognize the parameter can safely
continue to use the less granular eight levels.Alternatively, the urgency can be augmented. For example, a graphical user agent
could send a visible priority parameter to indicate if the resource being requested is
within the viewport.Generic priority parameters are preferred over vendor-specific,
application-specific, or deployment-specific values. If a generic value cannot be
agreed upon in the community, the parameter's name should be correspondingly
specific (e.g., with a prefix that identifies the vendor, application, or
deployment).RegistrationNew priority parameters can be defined by registering them in the "HTTP Priority"
registry. This registry governs the keys (short textual strings) used
in the Dictionary (see ).
Since each HTTP request can have associated priority signals, there is value
in having short key lengths, especially single-character strings. In order to
encourage extensions while avoiding unintended conflict among attractive key
values, the "HTTP Priority" registry operates two registration policies,
depending on key length.
Registration requests for priority parameters with a key length of one use the
Specification Required policy, per .
Registration requests for priority parameters with a key length greater than
one use the Expert Review policy, per . A
specification document is appreciated but not required.
When reviewing registration requests, the designated expert(s) can consider the
additional guidance provided in but cannot use it as a basis
for rejection.Registration requests should use the following template:
Name:
[a name for the priority parameter that matches the parameter key]
Description:
[a description of the priority parameter semantics and value]
Reference:
[to a specification defining this priority parameter]
See the registry at for details on
where to send registration requests.The Priority HTTP Header FieldThe Priority HTTP header field is a Dictionary that carries priority parameters (see ).
It can appear in requests and responses. It is an end-to-end signal that
indicates the endpoint's view of how HTTP responses should be prioritized.
describes how intermediaries can combine the priority information
sent from clients and servers. Clients cannot interpret the appearance or
omission of a Priority response header field as acknowledgement that any
prioritization has occurred. Guidance for how endpoints can act on Priority
header values is given in Sections and .An HTTP request with a Priority header field might be cached and reused for
subsequent requests; see . When an origin
server generates the Priority response header field based on properties of an
HTTP request it receives, the server is expected to control the cacheability or
the applicability of the cached response by using header fields that control
the caching behavior (e.g., Cache-Control, Vary).ReprioritizationAfter a client sends a request, it may be beneficial to change the priority of
the response. As an example, a web browser might issue a prefetch request for a
JavaScript file with the urgency parameter of the Priority request header field
set to u=7 (background). Then, when the user navigates to a page that
references the new JavaScript file, while the prefetch is in progress, the
browser would send a reprioritization signal with the Priority Field Value set
to u=0. The PRIORITY_UPDATE frame () can be used for such
reprioritization.The PRIORITY_UPDATE FrameThis document specifies a new PRIORITY_UPDATE frame for HTTP/2
and HTTP/3 . It carries priority parameters and
references the target of the prioritization based on a version-specific
identifier. In HTTP/2, this identifier is the stream ID; in HTTP/3, the
identifier is either the stream ID or push ID. Unlike the Priority header field,
the PRIORITY_UPDATE frame is a hop-by-hop signal.PRIORITY_UPDATE frames are sent by clients on the control stream, allowing them
to be sent independently of the stream that carries the response. This means
they can be used to reprioritize a response or a push stream, or to signal the
initial priority of a response instead of the Priority header field.A PRIORITY_UPDATE frame communicates a complete set of all priority parameters
in the Priority Field Value field. Omitting a priority parameter is a signal to
use its default value. Failure to parse the Priority Field Value MAY be treated
as a connection error. In HTTP/2, the error is of type PROTOCOL_ERROR; in HTTP/3,
the error is of type H3_GENERAL_PROTOCOL_ERROR.A client MAY send a PRIORITY_UPDATE frame before the stream that it references
is open (except for HTTP/2 push streams; see ). Furthermore,
HTTP/3 offers no guaranteed ordering across streams, which could cause the frame
to be received earlier than intended. Either case leads to a race condition
where a server receives a PRIORITY_UPDATE frame that references a request stream
that is yet to be opened. To solve this condition, for the purposes of
scheduling, the most recently received PRIORITY_UPDATE frame can be considered
as the most up-to-date information that overrides any other signal. Servers
SHOULD buffer the most recently received PRIORITY_UPDATE frame and apply it once
the referenced stream is opened. Holding PRIORITY_UPDATE frames for each stream
requires server resources, which can be bounded by local implementation policy.
Although there is no limit to the number of PRIORITY_UPDATE frames that can be
sent, storing only the most recently received frame limits resource commitment.HTTP/2 PRIORITY_UPDATE FrameThe HTTP/2 PRIORITY_UPDATE frame (type=0x10) is used by clients to signal the
initial priority of a response, or to reprioritize a response or push stream. It
carries the stream ID of the response and the priority in ASCII text, using the
same representation as the Priority header field value.The Stream Identifier field (see ) in the
PRIORITY_UPDATE frame header MUST be zero (0x0). Receiving a PRIORITY_UPDATE
frame with a field of any other value MUST be treated as a connection error of
type PROTOCOL_ERROR.The Length, Type, Unused Flag(s), Reserved, and Stream Identifier fields are
described in . The PRIORITY_UPDATE frame payload
contains the following additional fields:
Prioritized Stream ID:
A 31-bit stream identifier for the stream that is the target of the priority
update.
Priority Field Value:
The priority update value in ASCII text, encoded using Structured Fields. This
is the same representation as the Priority header field value.
When the PRIORITY_UPDATE frame applies to a request stream, clients SHOULD
provide a prioritized stream ID that refers to a stream in the "open",
"half-closed (local)", or "idle" state (i.e., streams where data might still be received). Servers can discard frames where the
prioritized stream ID refers to a stream in the "half-closed (local)" or
"closed" state (i.e., streams where no further data will be sent).
The number of streams that have been prioritized but remain in
the "idle" state plus the number of active streams (those in the "open" state or
in either of the "half-closed" states; see ) MUST NOT exceed
the value of the SETTINGS_MAX_CONCURRENT_STREAMS parameter. Servers that receive
such a PRIORITY_UPDATE MUST respond with a connection error of type
PROTOCOL_ERROR.When the PRIORITY_UPDATE frame applies to a push stream, clients SHOULD provide
a prioritized stream ID that refers to a stream in the "reserved (remote)" or
"half-closed (local)" state. Servers can discard frames where the prioritized
stream ID refers to a stream in the "closed" state. Clients MUST NOT provide a
prioritized stream ID that refers to a push stream in the "idle" state. Servers
that receive a PRIORITY_UPDATE for a push stream in the "idle" state MUST
respond with a connection error of type PROTOCOL_ERROR.If a PRIORITY_UPDATE frame is received with a prioritized stream ID of 0x0, the
recipient MUST respond with a connection error of type PROTOCOL_ERROR.Servers MUST NOT send PRIORITY_UPDATE frames. If a client receives a
PRIORITY_UPDATE frame, it MUST respond with a connection error of type
PROTOCOL_ERROR.HTTP/3 PRIORITY_UPDATE FrameThe HTTP/3 PRIORITY_UPDATE frame (type=0xF0700 or 0xF0701) is used by clients to
signal the initial priority of a response, or to reprioritize a response or push
stream. It carries the identifier of the element that is being prioritized and
the updated priority in ASCII text that uses the same representation as that of
the Priority header field value. PRIORITY_UPDATE with a frame type of 0xF0700 is
used for request streams, while PRIORITY_UPDATE with a frame type of 0xF0701 is
used for push streams.The PRIORITY_UPDATE frame MUST be sent on the client control stream
(see ). Receiving a PRIORITY_UPDATE frame on a
stream other than the client control stream MUST be treated as a connection
error of type H3_FRAME_UNEXPECTED.The PRIORITY_UPDATE frame payload has the following fields:
Prioritized Element ID:
The stream ID or push ID that is the target of the priority update.
Priority Field Value:
The priority update value in ASCII text, encoded using Structured Fields. This
is the same representation as the Priority header field value.
The request-stream variant of PRIORITY_UPDATE (type=0xF0700) MUST reference a
request stream. If a server receives a PRIORITY_UPDATE (type=0xF0700) for a
stream ID that is not a request stream, this MUST be treated as a connection
error of type H3_ID_ERROR. The stream ID MUST be within the client-initiated
bidirectional stream limit. If a server receives a PRIORITY_UPDATE
(type=0xF0700) with a stream ID that is beyond the stream limits, this SHOULD be
treated as a connection error of type H3_ID_ERROR. Generating an error is not
mandatory because HTTP/3 implementations might have practical barriers to
determining the active stream concurrency limit that is applied by the QUIC
layer.The push-stream variant of PRIORITY_UPDATE (type=0xF0701) MUST reference a promised
push stream. If a server receives a PRIORITY_UPDATE (type=0xF0701) with a push ID
that is greater than the maximum push ID or that has not yet been promised, this
MUST be treated as a connection error of type H3_ID_ERROR.Servers MUST NOT send PRIORITY_UPDATE frames of either type. If a client
receives a PRIORITY_UPDATE frame, this MUST be treated as a connection error of
type H3_FRAME_UNEXPECTED.Merging Client- and Server-Driven Priority ParametersIt is not always the case that the client has the best understanding of how the
HTTP responses deserve to be prioritized. The server might have additional
information that can be combined with the client's indicated priority in order
to improve the prioritization of the response. For example, use of an HTML
document might depend heavily on one of the inline images; the existence of such
dependencies is typically best known to the server. Or, a server that receives
requests for a font and images with the same urgency might give
higher precedence to the font, so that a visual client can render textual
information at an early moment.An origin can use the Priority response header field to indicate its view on how
an HTTP response should be prioritized. An intermediary that forwards an HTTP
response can use the priority parameters found in the Priority response header
field, in combination with the client Priority request header field, as input to
its prioritization process. No guidance is provided for merging priorities; this
is left as an implementation decision.The absence of a priority parameter in an HTTP response indicates the server's
disinterest in changing the client-provided value. This is different from the
request header field, in which omission of a priority parameter implies the use of its default value (see ).As a non-normative example, when the client sends an HTTP request with the
urgency parameter set to 5 and the incremental parameter set to true
:method = GET
:scheme = https
:authority = example.net
:path = /menu.png
priority = u=5, i
and the origin responds with
:status = 200
content-type = image/png
priority = u=1
the intermediary might alter its understanding of the urgency from 5 to 1,
because it prefers the server-provided value over the client's. The incremental
value continues to be true, i.e., the value specified by the client, as the server did
not specify the incremental (i) parameter.Client SchedulingA client MAY use priority values to make local processing or scheduling choices
about the requests it initiates.Server SchedulingIt is generally beneficial for an HTTP server to send all responses as early as
possible. However, when serving multiple requests on a single connection, there
could be competition between the requests for resources such as connection
bandwidth. This section describes considerations regarding how servers can
schedule the order in which the competing responses will be sent when such
competition exists.Server scheduling is a prioritization process based on many inputs, with
priority signals being only one form of input. Factors such as implementation
choices or deployment environment also play a role. Any given connection is
likely to have many dynamic permutations. For these reasons, it is not possible
to describe a universal scheduling algorithm. This document provides some basic,
non-exhaustive recommendations for how servers might act on priority
parameters. It does not describe in detail how servers might combine priority
signals with other factors. Endpoints cannot depend on particular treatment
based on priority signals. Expressing priority is only a suggestion.It is RECOMMENDED that, when possible, servers respect the urgency parameter
(), sending higher-urgency responses before lower-urgency responses.The incremental parameter indicates how a client processes response bytes as
they arrive. It is RECOMMENDED that, when possible, servers respect the
incremental parameter ().Non-incremental responses of the same urgency SHOULD be served by prioritizing
bandwidth allocation in ascending order of the stream ID, which corresponds to
the order in which clients make requests. Doing so ensures that clients can use
request ordering to influence response order.Incremental responses of the same urgency SHOULD be served by sharing bandwidth
among them. The message content of incremental responses is used as parts, or chunks,
are received. A client might benefit more from receiving a portion of all
these resources rather than the entirety of a single resource. How large a
portion of the resource is needed to be useful in improving performance varies.
Some resource types place critical elements early; others can use information
progressively. This scheme provides no explicit mandate about how a server
should use size, type, or any other input to decide how to prioritize.There can be scenarios where a server will need to schedule multiple incremental
and non-incremental responses at the same urgency level. Strictly abiding by the
scheduling guidance based on urgency and request generation order might lead
to suboptimal results at the client, as early non-incremental responses might
prevent the serving of incremental responses issued later. The following are
examples of such challenges:
At the same urgency level, a non-incremental request for a large resource
followed by an incremental request for a small resource.
At the same urgency level, an incremental request of indeterminate length
followed by a non-incremental large resource.
It is RECOMMENDED that servers avoid such starvation where possible. The method
for doing so is an implementation decision. For example, a server might
preemptively send responses of a particular incremental type based on other
information such as content size.Optimal scheduling of server push is difficult, especially when pushed resources
contend with active concurrent requests. Servers can consider many factors when
scheduling, such as the type or size of resource being pushed, the priority of
the request that triggered the push, the count of active concurrent responses,
the priority of other active concurrent responses, etc. There is no general
guidance on the best way to apply these. A server that is too simple could
easily push at too high a priority and block client requests, or push at too low
a priority and delay the response, negating intended goals of server push.Priority signals are a factor for server push scheduling. The concept of
parameter value defaults applies slightly differently because there is no
explicit client-signaled initial priority. A server can apply priority signals
provided in an origin response; see the merging guidance given in .
In the absence of origin signals, applying default parameter values could be
suboptimal. By whatever means a server decides to schedule a pushed response, it
can signal the intended priority to the client by including the Priority field
in a PUSH_PROMISE or HEADERS frame.Intermediaries with Multiple Backend ConnectionsAn intermediary serving an HTTP connection might split requests over multiple
backend connections. When it applies prioritization rules strictly, low-priority
requests cannot make progress while requests with higher priorities are in
flight. This blocking can propagate to backend connections, which the peer might
interpret as a connection stall. Endpoints often implement protections against
stalls, such as abruptly closing connections after a certain time period. To
reduce the possibility of this occurring, intermediaries can avoid strictly
following prioritization and instead allocate small amounts of bandwidth for all
the requests that they are forwarding, so that every request can make some
progress over time.Similarly, servers SHOULD allocate some amount of bandwidths to streams acting
as tunnels.Scheduling and the CONNECT MethodWhen a stream carries a CONNECT request, the scheduling guidance in
this document applies to the frames on the stream. A client that issues multiple
CONNECT requests can set the incremental parameter to true. Servers that
implement the recommendations for handling of the incremental parameter () are likely to schedule these fairly, preventing one
CONNECT stream from blocking others.Retransmission SchedulingTransport protocols such as TCP and QUIC provide reliability by detecting packet
losses and retransmitting lost information. In addition to the considerations in
, scheduling of retransmission data could compete with new
data. The remainder of this section discusses considerations when using QUIC. states the following: "Endpoints SHOULD prioritize
retransmission of data over sending new data, unless priorities specified by the
application indicate otherwise". When an HTTP/3 application uses the priority
scheme defined in this document and the QUIC transport implementation supports
application-indicated stream priority, a transport that considers the relative
priority of streams when scheduling both new data and retransmission data might
better match the expectations of the application. However, there are no
requirements on how a transport chooses to schedule based on this information
because the decision depends on several factors and trade-offs. It could
prioritize new data for a higher-urgency stream over retransmission data for a
lower-priority stream, or it could prioritize retransmission data over new data
irrespective of urgencies. also highlights considerations regarding
application priorities when sending probe packets after Probe Timeout timer
expiration. A QUIC implementation supporting application-indicated priorities
might use the relative priority of streams when choosing probe data.FairnessTypically, HTTP implementations depend on the underlying transport to maintain
fairness between connections competing for bandwidth. When an intermediary receives HTTP requests on client connections, it forwards them to backend connections. Depending on how the intermediary coalesces or splits requests across different backend connections, different clients might experience dissimilar performance. This dissimilarity might expand if the intermediary also uses priority signals when
forwarding requests. Sections and discuss
mitigations of this expansion of unfairness.Conversely, discusses how servers might intentionally
allocate unequal bandwidth to some connections, depending on the priority
signals.Coalescing IntermediariesWhen an intermediary coalesces HTTP requests coming from multiple clients into
one HTTP/2 or HTTP/3 connection going to the backend server, requests that
originate from one client might carry signals indicating higher priority than
those coming from others.It is sometimes beneficial for the server running behind an intermediary to obey
Priority header field values. As an example, a resource-constrained
server might defer the transmission of software update files that have the
background urgency level (7). However, in the worst case, the asymmetry
between the priority declared by multiple clients might cause all responses going to
one user agent to be delayed until all responses going to another user agent have
been sent.In order to mitigate this fairness problem, a server could use knowledge about
the intermediary as another input in its prioritization decisions. For
instance, if a server knows the intermediary is coalescing requests, then it
could avoid serving the responses in their entirety and instead distribute
bandwidth (for example, in a round-robin manner). This can work if the
constrained resource is network capacity between the intermediary and the user
agent, as the intermediary buffers responses and forwards the chunks based on
the prioritization scheme it implements.A server can determine if a request came from an intermediary through
configuration or can check to see if the request contains one of the following
header fields:
Forwarded , X-Forwarded-For
Via (see )
HTTP/1.x Back EndsIt is common for Content Delivery Network (CDN) infrastructure to support different HTTP versions on the
front end and back end. For instance, the client-facing edge might support
HTTP/2 and HTTP/3 while communication to backend servers is done using
HTTP/1.1. Unlike connection coalescing, the CDN will "demux" requests into
discrete connections to the back end. Response multiplexing in a single connection is not supported by HTTP/1.1 (or older), so there is not a fairness problem.
However, backend servers MAY still use client headers for request scheduling.
Backend servers SHOULD only schedule based on client priority information where
that information can be scoped to individual end clients. Authentication and
other session information might provide this linkability.Intentional Introduction of UnfairnessIt is sometimes beneficial to deprioritize the transmission of one connection
over others, knowing that doing so introduces a certain amount of unfairness
between the connections and therefore between the requests served on those
connections.For example, a server might use a scavenging congestion controller on
connections that only convey background priority responses such as software
update images. Doing so improves responsiveness of other connections at the cost
of delaying the delivery of updates.Why Use an End-to-End Header Field?In contrast to the prioritization scheme of HTTP/2, which uses a hop-by-hop frame,
the Priority header field is defined as "end-to-end".The way that a client processes a response is a property associated with the
client generating that request, not that of an intermediary. Therefore, it is
an end-to-end property. How these end-to-end properties carried by the Priority
header field affect the prioritization between the responses that share a
connection is a hop-by-hop issue.Having the Priority header field defined as end-to-end is important for caching
intermediaries. Such intermediaries can cache the value of the Priority header
field along with the response and utilize the value of the cached header field
when serving the cached response, only because the header field is defined as
end-to-end rather than hop-by-hop.Security Considerations describes considerations for server buffering of PRIORITY_UPDATE
frames. presents examples where servers that prioritize responses
in a certain way might be starved of the ability to transmit responses.The security considerations from apply to the processing of
priority parameters defined in .IANA ConsiderationsThis specification registers the following entry in the "Hypertext Transfer
Protocol (HTTP) Field Name Registry" defined in :
Field Name:
Priority
Status:
permanent
Reference:
This document
This specification registers the following entry in the "HTTP/2 Settings" registry
defined in :
Code:
0x9
Name:
SETTINGS_NO_RFC7540_PRIORITIES
Initial Value:
0
Reference:
This document
This specification registers the following entry in the "HTTP/2 Frame Type"
registry defined in :
Code:
0x10
Frame Type:
PRIORITY_UPDATE
Reference:
This document
This specification registers the following entry in the "HTTP/3 Frame Types"
registry established by :
Value:
0xF0700-0xF0701
Frame Type:
PRIORITY_UPDATE
Status:
permanent
Reference:
This document
Change Controller:
IETF
Contact:
ietf-http-wg@w3.org
IANA has created the "Hypertext Transfer Protocol (HTTP) Priority" registry at
and has populated it with the entries in
; see for its associated procedures.
Initial Priority Parameters
Name
Description
Reference
u
The urgency of an HTTP response.
i
Whether an HTTP response can be processed incrementally.
ReferencesNormative ReferencesHTTP SemanticsHTTP/2HTTP/3QUIC: A UDP-Based Multiplexed and Secure TransportKey words for use in RFCs to Indicate Requirement LevelsIn many standards track documents several words are used to signify the requirements in the specification. These words are often capitalized. This document defines these words as they should be interpreted in IETF documents. This document specifies an Internet Best Current Practices for the Internet Community, and requests discussion and suggestions for improvements.Guidelines for Writing an IANA Considerations Section in RFCsMany protocols make use of points of extensibility that use constants to identify various protocol parameters. To ensure that the values in these fields do not have conflicting uses and to promote interoperability, their allocations are often coordinated by a central record keeper. For IETF protocols, that role is filled by the Internet Assigned Numbers Authority (IANA).To make assignments in a given registry prudently, guidance describing the conditions under which new values should be assigned, as well as when and how modifications to existing values can be made, is needed. This document defines a framework for the documentation of these guidelines by specification authors, in order to assure that the provided guidance for the IANA Considerations is clear and addresses the various issues that are likely in the operation of a registry.This is the third edition of this document; it obsoletes RFC 5226.Ambiguity of Uppercase vs Lowercase in RFC 2119 Key WordsRFC 2119 specifies common key words that may be used in protocol specifications. This document aims to reduce the ambiguity by clarifying that only UPPERCASE usage of the key words have the defined special meanings.Structured Field Values for HTTPInformative ReferencesHTTP CachingForwarded HTTP ExtensionOf the Utmost Importance: Resource Prioritization in HTTP/3 over QUICSCITEPRESS Proceedings of the 15th International Conference on Web Information Systems and Technologies (pages 130-143)Declaring Support for HTTP/2 PrioritiesGoogleCloudflare HTTP/2 provides a prioritization scheme but experience has shown that
implementation support varies. This document defines an HTTP/2
setting that endpoints can use as an affirmative signal to indicate
their support for HTTP/2 Priorities.
Work in ProgressQUIC Loss Detection and Congestion ControlHypertext Transfer Protocol Version 2 (HTTP/2)This specification describes an optimized expression of the semantics of the Hypertext Transfer Protocol (HTTP), referred to as HTTP version 2 (HTTP/2). HTTP/2 enables a more efficient use of network resources and a reduced perception of latency by introducing header field compression and allowing multiple concurrent exchanges on the same connection. It also introduces unsolicited push of representations from servers to clients.This specification is an alternative to, but does not obsolete, the HTTP/1.1 message syntax. HTTP's existing semantics remain unchanged.The "font" Top-Level Media TypeThis memo serves to register and document the "font" top-level media type, under which subtypes for representation formats for fonts may be registered. This document also serves as a registration application for a set of intended subtypes, which are representative of some existing subtypes already in use, and currently registered under the "application" tree by their separate registrations.Acknowledgements
presented the idea of using a header field for representing
priorities in .
In ,
advocated for representing the priorities using a tuple of urgency and
concurrency. The ability to disable HTTP/2 prioritization is inspired by
, authored by and , with
modifications based on feedback that was not incorporated into an update to that
document.The motivation for defining an alternative to HTTP/2 priorities is drawn from
discussion within the broad HTTP community. Special thanks to ,
, and Netflix for text that was incorporated explicitly in this
document.In addition to the people above, this document owes a lot to the extensive
discussion in the HTTP priority design team, consisting of ,
, , , ,
, , , , and the authors
of this document. contributed the section on retransmission scheduling.Authors' AddressesFastlykazuhooku@gmail.comCloudflarelucaspardue.24.7@gmail.com