One document matched: draft-alvestrand-constraints-resolution-03.xml
<?xml version="1.0" encoding="US-ASCII"?>
<!DOCTYPE rfc SYSTEM "rfc2629.dtd">
<?rfc toc="yes"?>
<?rfc tocompact="yes"?>
<?rfc tocdepth="3"?>
<?rfc tocindent="yes"?>
<?rfc symrefs="yes"?>
<?rfc sortrefs="yes"?>
<?rfc comments="yes"?>
<?rfc inline="yes"?>
<?rfc compact="yes"?>
<?rfc subcompact="no"?>
<rfc category="std" docName="draft-alvestrand-constraints-resolution-03"
ipr="trust200902">
<front>
<title abbrev="Resolution Constraints">Resolution Constraints in Web Real
Time Communications</title>
<author fullname="Harald Alvestrand" initials="H. T." surname="Alvestrand">
<organization>Google</organization>
<address>
<email>harald@alvestrand.no</email>
</address>
</author>
<date day="26" month="August" year="2013" />
<abstract>
<t>This document specifies the constraints necessary for a Javascript
application to successfully indicate to a browser that supports WebRTC
what resolutions it desires on a video stream.</t>
<t>It also discusses the possible use of SDP to carry that information
between browsers.</t>
</abstract>
<note title="Requirements Language">
<t>The key words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT",
"SHOULD", "SHOULD NOT", "RECOMMENDED", "MAY", and "OPTIONAL" in this
document are to be interpreted as described in <xref
target="RFC2119">RFC 2119</xref>.</t>
</note>
</front>
<middle>
<section title="Introduction">
<t>There are a number of scenarios where it's useful for a WebRTC
application to indicate to the WebRTC implementation in the supported
browser what the desired characteristics of a video stream are. These
include, but are not limited to:</t>
<t><list style="symbols">
<t>Specifying a minimum desired resolution for a given application,
in order to control the user experience or resource tradeoffs made
by the browser to favour a particular stream</t>
<t>Specifying a maximum desired resolution for a given stream, in
order to save some resource (bandwidth, CPU....), possibly outside
of the browser where the browser can't tell that it's exceeding a
constraint</t>
<t>Specifying resolutions that are a reasonable fit for the current
usage of the video stream, for instance fitting with the number of
pixels available on the part of a device's display surface that is
devoted to displaying this video stream</t>
<t>Specifying the shape of a video stream, in order to fit the video
onto a display surface without the need for black bars or image
distortion</t>
</list>Similar considerations apply for framerate.</t>
<t></t>
<section title="Disposition of this text">
<t>This draft is written in order to get something specific out to
refer to during spec-writing and implementation. Some text may
eventually get merged into the JSEP specification, <xref
target="I-D.ietf-rtcweb-jsep"></xref>.</t>
</section>
</section>
<section title="Usage Scenarios">
<t>Consider the following (simplified) model of a video stream through a
WebRTC application:</t>
<figure>
<artwork><![CDATA[
|<-------------- Browser A -------------------->|
Camera ---> MediaStream A ---> Peerconnection A ------+
|<------- Application A ---------->| |
v ^ v
Signalling channel Internet (media)
v ^ |
|<------- Application B ---------->| |
<video> tag <-- MediaStream B <--- Peerconnection B --+
|<-------------Browser B ----------------------->|
]]></artwork>
</figure>
<t></t>
<t>Both applications are running in browsers, with Application A
connected to a camera that is able to deliver video streams up to HD
quality (1280x720).</t>
<t></t>
<t></t>
<section title="Scenario: Resolution change">
<t>At one particular moment in time, the <video> tag in
Application B is rendered as a thumbnail, and video is flowing to it
in a 160x100 resolution; there is no need to send any more data, since
no more pixels are available for its display anyway.</t>
<t>Then the user of Application B hits the "full-screen" button. There
are now 1600x1200 pixels available for display.</t>
<t>Initially, Application B will splay the 160x100 image across the
larger surface, because there is no other choice, but it will desire
to have as many pixels as possible available to provide a high quality
image.</t>
</section>
<section title="Scenario: Constrained bandwidth">
<t>At one particular moment in time, the camera is generating
1280x720, resulting in a 2 Mbits/second data flow from A to B.
Congestion control signals that this data rate is no longer available;
rather than letting the browser reduce the bandwidth of some flow of
its choice, Application A decides that the high definition video is
the feature that is least valuable. It can then apply a new constraint
to Mediastream A, specifying that resolution should be at most
640x360; browser A is then responsible for making sure this decision
is communicated to browser B (if it needs to be).</t>
</section>
<section title="Scenario: Limited processing capacity">
<t>If application B is running on a slow machine (2000-class PC or
2010-class mobile phone), the maximum capacity of the video decoder
may be 320x200 - Application B may then wish to indicate that
application A should limit the stream sent across the network to that
resolution - sending more bits isn't useful, because the receiver
doesn't have enough capacity to decode and downscale the video
stream.</t>
</section>
</section>
<section title="Models for resolution manipulation">
<t>As specified in the "Media Capture And Streams" document <xref
target="W3C.WD-mediacapture-streams-20130516"></xref> the consumer of a
video track in a MediaStream will have a "native resolution", which
indicates what size video it's useful to push to it. The application can
also set (and change) constraints on the video MediaStreamTrack,
indicating which range of properties it sees useful for the purposes of
the application.</t>
<t>In SDP, the "a=imageattr" attribute is available to provide
information on the resolution of video streams described by an SDP
m-line.</t>
<t>If both mechanisms are available, the choices available to the writer
of application B in the "increase screen area" above are:</t>
<t><list style="numbers">
<t>Signal (by non-standard means) to Application A that more pixels
are needed. Application A will then modify the constraints on
Mediastream A to say that the desired (not mandatory) min resolution
is 1600x1200; Browser A will then reconfigure the camera to generate
the closest available resolution, which is 1280x720.</t>
<t>Apply a new constraint set to Mediastream B's video track, saying
that the desired resolution is now 1600x1200. Browser B will then
have to figure out that this is an incoming track via Peerconnection
B, and that the resolution needs to be signalled; it will then fire
a NegotationNeeded event at Application B, which will then
renegotiate the desired resolutions using an SDP exchange with
Browser A; Browser A will then figure out from the SDP that it's
useful to generate a higher resolution video stream, and reconfigure
the camera as above.</t>
<t>Execute a renegotiation with Application A, adding attributes as
described in <xref target="s-examples"></xref> by modifying the SDP
generated by CreateOffer, and triggering the behaviour in the
previous alternative inside Browser A. API-wise, this is perhaps the
most complex method.</t>
</list></t>
<t>The advantage of the first method is that it does not require any SDP
parsing or generation.</t>
<t>The advantage of the second method is that it will work when
appliation A and application B are different applications; there is no
need for them to have any private agreement on how to set bitrate. It
does require both the implementation of constraints and that browser B
has the ability to generate the proper constraints in the SDP.</t>
<t>The third method requires SDP parsing in browser A, but not SDP
generation in browser B. It does require SDP manipulation in Javascript
at application B.</t>
<t></t>
<section title="Sender-side constraint manipulation">
<t>The following Javascript code (somewhat pseudocode) will achieve
the "increase screen area" according to method 1 above.</t>
<t>The examples use APIs from "Media Capture And Streams as well as
from WebRTC <xref target="W3C.WD-webrtc-20120821"></xref>.</t>
<figure>
<artwork><![CDATA[
// B side
function needResolutionToChange(newWidth, newHeight) {
message = makeMessage("resolution", newWidth, newHeight);
remote.send(message)
}
// A side
function handleMessage(message) {
if (message.verb === "resolution") {
constraints = video.constraints();
// The function below can be a polyfill.
constraints.replaceOrAddOptConstraint("width", message.arg1);
constraints.replaceOrAddOptConstraint("height", message.arg2);
}
}
]]></artwork>
</figure>
<t></t>
</section>
<section title="Receiver-side constraint manipulation">
<t>This implements version 2 of the constraint manipulation above.
Note that the handing of "onnegotiationneeded" is the same as for any
other renegotiation.</t>
<figure>
<artwork><![CDATA[// B side
function needResolutionToChange(newWidth, newHeight) {
constraints = video.constraints();
constraints.replaceOrAddOptConstraint("width",
{ "max": message.arg1 });
constraints.replaceOrAddOptConstraint("height",
{ "max": message.arg2 });
video.applyCoinstraints(constraints);
}
pc.onnegotiationneeded = function() {
offer = pc.CreateOffer();
message = makeMessage("offer", offer);
remote.send(message);
}
// Functions to handle answer from A side omitted.
// A side
function handleMessage(message) {
if (message.verb === "offer") {
pc.SetRemoteDescription(message.arg1, success, failure)
}
}
// Functions to return answer to B side omitted.
]]></artwork>
</figure>
<t></t>
<t></t>
</section>
</section>
<section title="Constraints for specifying resolution">
<t>All constraints needed are registered in the IANA registry by <xref
target="W3C.WD-mediacapture-streams-20130516"></xref>. In summary, they
are:</t>
<t><list style="symbols">
<t>width - unsigned long or MinMaxConstraint - ie width: { min: 640,
max: 1024 }</t>
<t>height - unsigned long or MinMaxConstraint</t>
<t>frameRate - float or MinMaxConstraint</t>
</list></t>
</section>
<section anchor="s-examples" title="Syntax and Mapping Examples">
<t>See <xref target="IANA"></xref> for the actual definition of the
constraints used here.</t>
<section title="Examples with GetUserMedia">
<t>A constraint saying that we absolutely must have a minimum
resolution of 1024x768:</t>
<figure>
<artwork><![CDATA[
getUserMedia({
video: { mandatory: { width: { min: 1024 }, height: { min: 768 }}}
}, successCallback, errorCallback);
]]></artwork>
</figure>
<t>A constraint saying that we'd prefer 60 frames per second, if
available, and if we can get that, we'd like to lock the resolution to
640x480, but in all cases, the screen must be clamped to a 4:3 aspect
ratio - 16:9 or odd aspect ratios are not acceptable to this
application:</t>
<figure>
<artwork><![CDATA[
getUserMedia({
video: {
mandatory: { aspectRatio: { min: 1.333, max: 1.334 } },
optional [
{ frameRate: 60 },
{ width: 640 },
{ heigth: 480 }
]
}
}, successCallback, errorCallback);
]]></artwork>
</figure>
<t></t>
</section>
<section title="SDP mappings">
<t></t>
<t>The examples below are based on <xref
target="I-D.roach-mmusic-unified-plan"></xref> and <xref
target="I-D.ietf-mmusic-msid"></xref>.</t>
<t>An optional constraint has been applied to an incoming stream where
both upper and lower are constrained to 320x200. The stream has been
assigned to a hardware video decoder that can decode most resolutions
up to 1024x768, in any aspect ratio, but only if all divisions are
divisible by 16. The incoming stream has MediaStream ID aaaa, and
MediaStreamTrack id bbbb.</t>
<t>Escaped line breaks are added for readability.</t>
<figure>
<artwork><![CDATA[
m=video
a=imageattr:* [x=320,y=200,q=1.0] \
[x=[120:16:1024],y=[100:16:768],q=0.2]
a=msid: aaaa bbbb
]]></artwork>
</figure>
<t></t>
</section>
</section>
<section anchor="IANA" title="IANA Considerations">
<t>This document makes no requests of IANA:</t>
<t>Note to RFC Editor: This section can be deleted before publication as
an RFC.</t>
</section>
<section anchor="Security" title="Security Considerations">
<t>No security considerations particular to these specific constraints
have so far been identified.</t>
</section>
<section anchor="Acknowledgements" title="Acknowledgements">
<t>Special thanks are given to Dan Burnett, Cullen Jennings, the IETF
RTCWEB WG and the W3C WEBRTC WG for strongly influencing this memo, and
to Per Kjellander for being the first to implement the constraints in
getUserMedia.</t>
</section>
</middle>
<back>
<references title="Normative References">
<?rfc include="reference.RFC.2119"?>
</references>
<references title="Informative References">
<?rfc include='reference.I-D.ietf-rtcweb-jsep'?>
<?rfc include='reference.I-D.ietf-mmusic-msid'?>
<?rfc include='reference.I-D.roach-mmusic-unified-plan'?>
<?rfc include='reference.W3C.WD-mediacapture-streams-20130516'?>
<?rfc include='reference.W3C.WD-webrtc-20120821'?>
</references>
<section title="Changes from -00 to -01">
<t>Added the "Usage Scenarios" chapter.</t>
<t>Repointed the eventual target to be incorporation in the JSEP
draft.</t>
<t>Made sure the constraints are consistently spelled in camelCase, with
a small initial letter.</t>
</section>
<section title="Changes from -01 to -02">
<t>Moved a bit of the text around between sections, and referred to the
"settings API" proposal from the Media Capture task force.</t>
</section>
<section title="Changes from -02 to -03">
<t>Retargeted document to be a "here's how you can do it" draft.</t>
<t>Updated constraints format to be as per the May 16 W3C draft of
"media capture and streams".</t>
</section>
</back>
</rfc>
| PAFTECH AB 2003-2026 | 2026-04-24 02:57:25 |