The getMetadata()
method of the RTCEncodedAudioFrame
interface returns an object containing the metadata associated with the frame.
This includes information about the frame, including the audio encoding used, the synchronization source and contributing sources, and the sequence number (for incoming frames).
An object with the following properties:
synchronizationSource
-
A positive integer value indicating synchronization source ("ssrc") of the stream of RTP packets that are described by this frame. A source might be something like a microphone, or a mixer application that combines multiple sources. All packets from the same source share the same time source and sequence space, and so can be ordered relative to each other. Note that two frames with the same value refer to the same source.
payloadType
-
A positive integer value in the range from 0 to 127 that describes the format of the RTP payload. The mappings of values to formats is defined in RFC3550, and more specifically Section 6: Payload Type Definitions of RFC3551.
contributingSources
-
An Array
of sources (ssrc) that have contributed to the frame. Consider the case of a conferencing application that combines audio from multiple users. The synchronizationSource
would include the ssrc of the application, while contributingSources
would include the ssrc values of all the individual audio sources.
sequenceNumber
-
The sequence number of an incoming audio frame (not used for outgoing frames) that can be used for reconstructing the original send-order of frames. This is number between 0 and 32767 . Note that while numbers are allocated sequentially when sent, they will overflow at 32767 and restart back at 0. Therefore to compare two frame sequence numbers, in order to determine whether one is assumed to be after another, you must use serial number arithmetic.
This example WebRTC encoded transform implementation shows how you might get the frame metadata in a transform()
function and log it.
addEventListener("rtctransform", (event) => {
const async transform = new TransformStream({
async transform(encodedFrame, controller) {
const frameMetaData = encodedFrame.getMetadata();
console.log(frameMetaData)
controller.enqueue(encodedFrame);
},
});
event.transformer.readable
.pipeThrough(transform)
.pipeTo(event.transformer.writable);
});
The resulting object from a local microphone might look like the one shown below. Note that there are no contributing sources because there is just one source, and no sequenceNumber
because this is an outgoing frame.
{
"payloadType": 109,
"synchronizationSource": 1876443470
}