1. Introduction
The [WEBRTC-NV-USE-CASES] document describes several functions that can only be achieved by access to media (requirements N20-N22), including, but not limited to:
-
Funny Hats
-
Machine Learning
-
Virtual Reality Gaming
These use cases further require that processing can be done in worker threads (requirement N23-N24).
Furthermore, the "trusted JavaScript cloud conferencing" use case requires such processing to be done on encoded media, not just the raw media.
This specification gives an interface inspired by [WEB-CODECS] to provide access to such functionality while retaining the setup flow of RTCPeerConnection.
This iteration of the specification provides access to encoded media, which is the output of the encoder part of a codec and the input to the decoder part of a codec.
2. Terminology
3. Specification
The Streams definition doesn’t use WebIDL much, but the WebRTC spec does. This specification shows the IDL extensions for WebRTC.
It uses an extension to RTCConfiguration in order to notify the RTCPeerConnection
that insertable streams will be used, and uses
an additional API on RTCRtpSender
and RTCRtpReceiver
to
insert the processing into the pipeline.
// New dictionarydictionary {
RTCInsertableStreams ReadableStream ;
readable WritableStream ; }; // New fields in RTCConfiguration
writable partial dictionary RTCConfiguration {boolean =
encodedInsertableStreams false ; };typedef (SFrameTransform or RTCRtpScriptTransform ); // New methods for RTCRtpSender and RTCRtpReceiver
RTCRtpTransform partial interface RTCRtpSender {RTCInsertableStreams createEncodedStreams ();attribute RTCRtpTransform ?transform ; };partial interface RTCRtpReceiver {RTCInsertableStreams ();
createEncodedStreams attribute RTCRtpTransform ?transform ; };
3.1. Extension operation
At the time when a codec is initialized as part of the encoder, and the
corresponding flag is set in the RTCPeerConnection
's RTCConfiguration
argument, ensure that the codec is disabled and produces no output.
3.1.1. Stream creation
At construction of each RTCRtpSender
or RTCRtpReceiver
, run the following steps:
-
Initialize this.
[[Streams]]
to null. -
Initialize this.
[[transform]]
to null. -
Initialize this.
[[readable]]
to a newReadableStream
. -
Set up this.
[[readable]]
. this.[[readable]]
is provided frames using the readEncodedData algorithm given this as parameter. -
Set this.
[[readable]]
.[[owner]]
to this. -
Initialize this.
[[writable]]
to a newWritableStream
. -
Set up this.
[[writable]]
with its writeAlgorithm set to writeEncodedData given this as parameter. -
Set this.
[[writable]]
.[[owner]]
to this. -
Initialize this.
[[pipeToController]]
to null. -
Initialize this.
[[lastReceivedFrameTimestamp]]
to zero. -
If the
RTCPeerConnection
's configuration does not haveencodedInsertableStreams
set to "true", queue a task to run the following steps:
The createEncodedStreams()
method steps are:
-
If the
RTCPeerConnection
's configuration does not haveencodedInsertableStreams
set to "true", throw an "InvalidAccessError
"DOMException
and abort these steps. -
If the data source does not permit access, throw an "
InvalidAccessError
"DOMException
and abort these steps. -
If this.
[[Streams]]
is not null, throw an "InvalidAccessError
"DOMException
. -
If this.
[[pipeToController]]
is not null, throw an "InvalidAccessError
"DOMException
. -
Set this.
[[Streams]]
to anRTCInsertableStreams
object. -
Enable the encoded data source.
-
Return this.
[[Streams]]
.
3.1.2. Stream processing
The readEncodedData algorithm is given a rtcObject as parameter. It is defined by running the following steps:
-
Wait for a frame to be produced by rtcObject’s encoder if it is a
RTCRtpSender
or rtcObject’s packetizer if it is aRTCRtpReceiver
. -
Let frame be the newly produced frame.
-
Set frame.
[[owner]]
to rtcObject. -
Enqueue frame in rtcObject.
[[readable]]
.
The writeEncodedData algorithm is given a rtcObject as parameter and a frame as input. It is defined by running the following steps:
-
If frame.
[[owner]]
is not equal to rtcObject, abort these steps and return a promise resolved with undefined. A processor cannot create frames, or move frames between streams. -
If the frame’s
timestamp
is equal to or larger than rtcObject.[[lastReceivedFrameTimestamp]]
, abort these steps and return a promise resolved with undefined. A processor cannot reorder frames, although it may delay them or drop them. -
Set rtcObject.
[[lastReceivedFrameTimestamp]]
to the frame’stimestamp
. -
Enqueue the frame for processing as if it came directly from the encoded data source, by running one of the following steps:
-
If rtcObject is a
RTCRtpSender
, enqueue it to rtcObject’s packetizer, to be processed in parallel. -
If rtcObject is a
RTCRtpReceiver
, enqueue it to rtcObject’s decoder, to be processed in parallel.
-
-
Return a promise resolved with undefined.
3.2. Extension attribute
A RTCRtpTransform has two private slots called [[readable]]
and [[writable]]
.
The transform
getter steps are:
-
Return this.
[[transform]]
.
The transform
setter steps are:
-
Let transform be the argument to the setter.
-
Let checkedTransform set to transform if it is not null or to an identity transform stream otherwise.
-
Let reader be the result of getting a reader for checkedTransform.
[[readable]]
. -
Let writer be the result of getting a writer for checkedTransform.
[[writable]]
. -
Initialize newPipeToController to a new
AbortController
. -
If this.
[[pipeToController]]
is not null, run the following steps:-
Add the chain transform algorithm to this.
[[pipeToController]]
.signal. -
signal abort this.
[[pipeToController]]
.signal.
-
-
Else, run the chain transform algorithm steps.
-
Set this.
[[pipeToController]]
to newPipeToController. -
Set this.
[[transform]]
to transform.
The chain transform algorithm steps are defined as:
-
If newPipeToController’s aborted flag is true, abort these steps.
-
Release reader.
-
Release writer.
-
Assert that newPipeToController is the same object as rtcObject.
[[pipeToController]]
. -
Call pipeTo with rtcObject.
[[readable]]
, checkedTransform.[[writable]]
, preventClose equal to false, preventAbort equal to false, preventCancel equal to true and newPipeToController.signal. -
Call pipeTo with checkedTransform.
[[readable]]
, rtcObject.[[writable]]
, preventClose equal to true, preventAbort equal to true, preventCancel equal to false and newPipeToController.signal.
This algorithm is defined so that transforms can be updated dynamically. There is no guarantee on which frame will happen the switch from the previous transform to the new transform.
If a web application sets the transform synchronously at creation of the RTCRtpSender
(for instance when calling addTrack), the transform will receive the first frame generated by the RTCRtpSender
's encoder.
Similarly, if a web application sets the transform synchronously at creation of the RTCRtpReceiver
(for instance when calling addTrack, or at track event handler), the transform will receive the first full frame generated by the RTCRtpReceiver
's packetizer.
4. SFrameTransform
enum {
SFrameTransformRole ,
"encrypt" };
"decrypt" dictionary {
SFrameTransformOptions SFrameTransformRole = "encrypt"; };
role typedef [EnforceRange ]unsigned long long ;
SmallCryptoKeyID typedef (SmallCryptoKeyID or bigint ); [
CryptoKeyID Exposed =(Window ,DedicatedWorker )]interface {
SFrameTransform constructor (optional SFrameTransformOptions = {});
options Promise <undefined >setEncryptionKey (CryptoKey ,
key optional CryptoKeyID ); };
keyID SFrameTransform includes GenericTransformStream ;
The new SFrameTransform(options)
constructor steps are:
-
Let transformAlgorithm be an algorithm which takes a frame as input and runs the SFrame transform algorithm with this and frame.
-
Set this.
[[transform]]
to a newTransformStream
. -
Set up this.
[[transform]]
with transformAlgorithm set to transformAlgorithm. -
Let options be the method’s first argument.
-
Set this.
[[role]]
to options["role
"]. -
Set this.
[[readable]]
to this.[[transform]]
.[[readable]]
. -
Set this.
[[writable]]
to this.[[transform]]
.[[writable]]
.
4.1. Algorithm
The SFrame transform algorithm, given sframe as a SFrameTransform object and frame, runs these steps:
-
Let role be sframe.
[[role]]
. -
If frame.
[[owner]]
is aRTCRtpSender
, set role to 'encrypt'. -
If frame.
[[owner]]
is aRTCRtpReceiver
, set role to 'decrypt'. -
Let data be undefined.
-
If frame is a
BufferSource
, set data to frame. -
If frame is a
RTCEncodedAudioFrame
, set data to frame.data
-
If frame is a
RTCEncodedVideoFrame
, set data to frame.data
-
If data is undefined, abort these steps.
-
Let buffer be the result of running the SFrame algorithm with data and role as parameters. This algorithm is defined by the SFrame specification and returns an
ArrayBuffer
. -
If frame is a
BufferSource
, set frame to buffer. -
If frame is a
RTCEncodedAudioFrame
, set frame.data
to buffer. -
If frame is a
RTCEncodedVideoFrame
, set frame.data
to buffer. -
Enqueue frame in sframe.
[[transform]]
.
4.2. Methods
ThesetEncryptionKey(key, keyID)
method steps are:
-
Let promise be a new promise.
-
If keyID is a
bigint
which cannot be represented as a integer between 0 and 264-1 inclusive, reject promise with aRangeError
exception. -
Otherwise, in parallel, run the following steps:
-
Set key with its optional keyID as key material to use for the SFrame transform algorithm, as defined by the SFrame specification.
-
If setting the key material fails, reject promise with an
InvalidModificationError
exception and abort these steps. -
Resolve promise with undefined.
-
-
Return promise.
5. RTCRtpScriptTransform
interface mixin {
EncodedMediaChunk readonly attribute unsigned long long ; // microseconds
timestamp readonly attribute ArrayBuffer ; }; // WebCodecs definitions with introduction of EncodedMediaChunk to more easily refer to timestamp and data. // They should be removed from this spec at some point. [
data Exposed =(Window ,DedicatedWorker )]interface {
EncodedVideoChunk (
constructor EncodedVideoChunkInit );
init readonly attribute EncodedVideoChunkType ;
type readonly attribute unsigned long long ?; // microseconds };
duration EncodedVideoChunk includes EncodedMediaChunk ;dictionary {
EncodedVideoChunkInit required EncodedVideoChunkType ;
type required unsigned long long ;
timestamp unsigned long long ;
duration required BufferSource ; };
data enum {
EncodedVideoChunkType ,
"key" , }; [
"delta" Exposed =(Window ,DedicatedWorker )]interface {
EncodedAudioChunk (
constructor EncodedAudioChunkInit );
init readonly attribute EncodedAudioChunkType ; };
type EncodedAudioChunk includes EncodedMediaChunk ;dictionary {
EncodedAudioChunkInit required EncodedAudioChunkType ;
type required unsigned long long ;
timestamp required BufferSource ; };
data enum {
EncodedAudioChunkType ,
"key" , };
"delta"
dictionary {
RTCEncodedVideoFrameMetadata long long ;
frameId sequence <long long >;
dependencies unsigned short ;
width unsigned short ;
height long ;
spatialIndex long ;
temporalIndex long ;
synchronizationSource sequence <long >; }; [
contributingSources Exposed =(Window ,DedicatedWorker )]interface :
RTCEncodedVideoFrame EncodedVideoChunk {RTCEncodedVideoFrameMetadata (); };
getMetadata dictionary {
RTCEncodedAudioFrameMetadata long ;
synchronizationSource sequence <long >; }; [
contributingSources Exposed =(Window ,DedicatedWorker )]interface :
RTCEncodedAudioFrame EncodedAudioChunk {RTCEncodedAudioFrameMetadata (); }; // New interfaces to expose JavaScript-based transforms. [
getMetadata Exposed =DedicatedWorker ]interface :
RTCTransformEvent Event {readonly attribute RTCRtpScriptTransformer ; };
transformer partial interface DedicatedWorkerGlobalScope {attribute EventHandler ; }; [
onrtctransform Exposed =DedicatedWorker ]interface {
RTCRtpScriptTransformer readonly attribute ReadableStream readable ;readonly attribute WritableStream writable ;readonly attribute any options ; }; [Exposed =Window ]interface {
RTCRtpScriptTransform constructor (Worker ,
worker optional any ,
options optional sequence <object >); };
transfer
5.1. Operations
The new RTCRtpScriptTransform(worker, options, transfer)
constructor steps are:
-
Set t1 to an identity transform stream.
-
Set t2 to an identity transform stream.
-
Set this.
[[writable]]
to t1.[[writable]]
. -
Set this.
[[readable]]
to t2.[[readable]]
. -
Let serializedOptions be the result of StructuredSerializeWithTransfer(options, transfer).
-
Let serializedReadable be the result of StructuredSerializeWithTransfer(t1.
[[readable]]
, « t1.[[readable]]
»). -
Let serializedWritable be the result of StructuredSerializeWithTransfer(t2.
[[writable]]
, « t2.[[writable]]
»). -
Queue a task on the DOM manipulation task source worker’s global scope to run the following steps:
-
Let transformerOptions be the result of StructuredDeserialize(serializedOptions, the current Realm).
-
Let readable be the result of StructuredDeserialize(serializedReadable, the current Realm).
-
Let writable be the result of StructuredDeserialize(serializedWritable, the current Realm).
-
Let transformer be a new
RTCRtpScriptTransformer
. -
Set transformer.
[[options]]
to transformerOptions. -
Set transformer.
[[readable]]
to readable. -
Set transformer.
[[writable]]
to writable. -
Let event be the result of creating an event with
RTCTransformEvent
. -
Set event.type attribute to "rtctransform".
-
Set event.transformer to transformer.
-
Dispatch event on worker’s global scope.
-
// FIXME: Describe error handling (worker closing flag true at RTCRtpScriptTransform creation time. And worker being terminated while transform is processing data).
5.2. Attributes
A RTCRtpScriptTransformer has three private slots called [[options]]
, [[readable]]
and [[writable]]
.
The options
getter steps are:
-
Return this.
[[options]]
.
The readable
getter steps are:
-
Return this.
[[readable]]
.
The writable
getter steps are:
-
Return this.
[[writable]]
.
6. Privacy and security considerations
This API gives Javascript access to the content of media streams. This is also available from other sources, such as Canvas and WebAudio.
However, streams that are isolated (as specified in [WEBRTC-IDENTITY]) or tainted with another origin, cannot be accessed using this API, since that would break the isolation rule.
The API will allow access to some aspects of timing information that are otherwise unavailable, which allows some fingerprinting surface.
7. Examples
See the explainer document.