Source code
Revision control
Copy as Markdown
Other Tools
Test Info: Warnings
- This test has a WPT meta file that expects 4 subtest issues.
- This WPT test may be referenced by the following Test IDs:
- /webrtc-encoded-transform/tentative/RTCEncodedAudioFrame-metadata.https.html - WPT Dashboard Interop Dashboard
<!DOCTYPE html>
<meta charset="utf-8">
<title>RTCEncodedAudioFrame can be cloned and distributed</title>
<script src="/resources/testharness.js"></script>
<script src="/resources/testharnessreport.js"></script>
<script src=/resources/testdriver.js></script>
<script src=/resources/testdriver-vendor.js></script>
<script src='../../mediacapture-streams/permission-helper.js'></script>
<script src="../../webrtc/RTCPeerConnection-helper.js"></script>
<script src="../../service-workers/service-worker/resources/test-helpers.sub.js"></script>
<script src='RTCEncodedFrame-timestamps-helper.js'></script>
<script>
"use strict";
promise_test(async t => {
const caller1 = new RTCPeerConnection();
t.add_cleanup(() => caller1.close());
const callee1 = new RTCPeerConnection({encodedInsertableStreams:true});
t.add_cleanup(() => callee1.close());
await setMediaPermission("granted", ["microphone"]);
const inputStream = await navigator.mediaDevices.getUserMedia({audio:true});
const inputTrack = inputStream.getAudioTracks()[0];
t.add_cleanup(() => inputTrack.stop());
caller1.addTrack(inputTrack)
exchangeIceCandidates(caller1, callee1);
const caller2 = new RTCPeerConnection({encodedInsertableStreams:true});
t.add_cleanup(() => caller2.close());
const sender2 = caller2.addTransceiver("audio").sender;
const writer2 = sender2.createEncodedStreams().writable.getWriter();
sender2.replaceTrack(new MediaStreamTrackGenerator({ kind: 'audio' }));
const framesReceivedCorrectly = new Promise((resolve, reject) => {
callee1.ontrack = async e => {
const receiverStreams = e.receiver.createEncodedStreams();
const receiverReader = receiverStreams.readable.getReader();
const result = await receiverReader.read();
const original = result.value;
let newFrame = new RTCEncodedAudioFrame(original);
assert_true(original.getMetadata().hasOwnProperty('receiveTime'));
assert_true(original.getMetadata().receiveTime > 0);
assert_equals(original.getMetadata().rtpTimestamp, newFrame.getMetadata().rtpTimestamp);
assert_equals(original.getMetadata().captureTime, newFrame.getMetadata().captureTime);
assert_equals(original.getMetadata().receiveTime, newFrame.getMetadata().receiveTime);
assert_array_equals(Array.from(original.data), Array.from(newFrame.data));
await writer2.write(newFrame);
resolve();
}
});
await exchangeOfferAnswer(caller1, callee1);
return framesReceivedCorrectly;
}, "Constructing audio frame before sending works");
promise_test(async t => {
const caller1 = new RTCPeerConnection();
t.add_cleanup(() => caller1.close());
const callee1 = new RTCPeerConnection({encodedInsertableStreams:true});
t.add_cleanup(() => callee1.close());
await setMediaPermission("granted", ["microphone"]);
const inputStream = await navigator.mediaDevices.getUserMedia({audio:true});
const inputTrack = inputStream.getAudioTracks()[0];
t.add_cleanup(() => inputTrack.stop());
caller1.addTrack(inputTrack)
exchangeIceCandidates(caller1, callee1);
const caller2 = new RTCPeerConnection({encodedInsertableStreams:true});
t.add_cleanup(() => caller2.close());
const sender2 = caller2.addTransceiver("audio").sender;
const writer2 = sender2.createEncodedStreams().writable.getWriter();
sender2.replaceTrack(new MediaStreamTrackGenerator({ kind: 'audio' }));
const framesReceivedCorrectly = new Promise((resolve, reject) => {
callee1.ontrack = async e => {
const receiverStreams = e.receiver.createEncodedStreams();
const receiverReader = receiverStreams.readable.getReader();
const result = await receiverReader.read();
const original = result.value;
let newMetadata = original.getMetadata();
newMetadata.rtpTimestamp = newMetadata.rtpTimestamp + 1;
let newFrame = new RTCEncodedAudioFrame(original, {metadata: newMetadata});
assert_not_equals(original.getMetadata().rtpTimestamp, newFrame.getMetadata().rtpTimestamp);
assert_equals(newMetadata.rtpTimestamp, newFrame.getMetadata().rtpTimestamp);
assert_equals(original.getMetadata().receiveTime, newFrame.getMetadata().receiveTime);
assert_equals(original.getMetadata().captureTime, newFrame.getMetadata().captureTime);
assert_array_equals(Array.from(original.data), Array.from(newFrame.data));
await writer2.write(newFrame);
resolve();
}
});
await exchangeOfferAnswer(caller1, callee1);
return framesReceivedCorrectly;
}, "Constructing audio frame with metadata argument before sending works");
promise_test(async t => {
const caller1 = new RTCPeerConnection();
t.add_cleanup(() => caller1.close());
const callee1 = new RTCPeerConnection({encodedInsertableStreams:true});
t.add_cleanup(() => callee1.close());
await setMediaPermission("granted", ["microphone"]);
const inputStream = await navigator.mediaDevices.getUserMedia({audio:true});
const inputTrack = inputStream.getAudioTracks()[0];
t.add_cleanup(() => inputTrack.stop());
caller1.addTrack(inputTrack)
exchangeIceCandidates(caller1, callee1);
const caller2 = new RTCPeerConnection({encodedInsertableStreams:true});
t.add_cleanup(() => caller2.close());
const sender2 = caller2.addTransceiver("audio").sender;
const writer2 = sender2.createEncodedStreams().writable.getWriter();
sender2.replaceTrack(new MediaStreamTrackGenerator({ kind: 'audio' }));
const framesReceivedCorrectly = new Promise((resolve, reject) => {
callee1.ontrack = async e => {
const receiverStreams = e.receiver.createEncodedStreams();
const receiverReader = receiverStreams.readable.getReader();
const result = await receiverReader.read();
const original = result.value;
let newMetadata = original.getMetadata();
newMetadata.synchronizationSource = newMetadata.synchronizationSource + 1;
assert_throws_dom("InvalidModificationError", () => new RTCEncodedAudioFrame(original, {metadata: newMetadata}));
resolve();
}
});
await exchangeOfferAnswer(caller1, callee1);
return framesReceivedCorrectly;
}, "Constructing audio frame with bad metadata argument before sending does not work");
promise_test(async t => {
const kCaptureTime = 12345;
const pc1 = new RTCPeerConnection({encodedInsertableStreams:true});
t.add_cleanup(() => pc1.close());
const pc2 = new RTCPeerConnection({encodedInsertableStreams:true});
t.add_cleanup(() => pc2.close());
exchangeIceCandidates(pc1, pc2);
let numFrames = 0;
let audioLevelRead = new Promise((resolve, reject) => {
pc2.ontrack = t.step_func(e => {
const receiverTransformer = new TransformStream({
async transform(encodedFrame, controller) {
const metadata = encodedFrame.getMetadata();
if (metadata.captureTime < kCaptureTime - 1 || metadata.captureTime > kCaptureTime + 1) {
reject("Unexpected captureTime");
}
controller.enqueue(encodedFrame);
if (++numFrames == 10)
resolve();
}
});
const receiverStreams = e.receiver.createEncodedStreams();
receiverStreams.readable
.pipeThrough(receiverTransformer)
.pipeTo(receiverStreams.writable);
});
});
const stream = await navigator.mediaDevices.getUserMedia({audio:true});
t.add_cleanup(() => stream.getTracks().forEach(track => track.stop()));
const sender = pc1.addTrack(stream.getAudioTracks()[0]);
const senderStreams = sender.createEncodedStreams();
const senderTransformer = new TransformStream({
async transform(encodedFrame, controller) {
let metadata = encodedFrame.getMetadata();
metadata.captureTime = kCaptureTime;
controller.enqueue(new RTCEncodedAudioFrame(encodedFrame, {metadata}));
}
});
senderStreams.readable
.pipeThrough(senderTransformer)
.pipeTo(senderStreams.writable);
await addAbsCaptureTimeAndExchangeOffer(pc1, pc2);
await checkAbsCaptureTimeAndExchangeAnswer(pc1, pc2, true);
await audioLevelRead;
}, 'Basic simulcast setup with three spatial layers');
</script>