blob: 719464a6d61e30cf27acc227e0474bcaaedf469e [file] [log] [blame]
<!doctype html><!-- webkit-test-runner [ WebRTCRemoteVideoFrameEnabled=true ] -->
<html>
<head>
<meta charset="utf-8">
<title>A recorded muted audio track should produce silence</title>
<script src="/resources/testharness.js"></script>
<script src="/resources/testharnessreport.js"></script>
</head>
<body>
<video id="video1" controls></video>
<video id="video2" controls></video>
<video id="video3" controls></video>
<canvas id="canvas3" width="320" height="240"></canvas>
<script>
function setupAudioContext(context, streamOrVideo)
{
var sourceNode = streamOrVideo instanceof MediaStream ? context.createMediaStreamSource(streamOrVideo) : context.createMediaElementSource(streamOrVideo);
var analyser = context.createAnalyser();
var gain = context.createGain();
analyser.fftSize = 2048;
analyser.smoothingTimeConstant = 0;
analyser.minDecibels = -100;
analyser.maxDecibels = 0;
gain.gain.value = 0;
sourceNode.connect(analyser);
analyser.connect(gain);
gain.connect(context.destination);
return analyser;
}
function analyseAudio(analyser, duration, context)
{
return new Promise((resolve, reject) => {
var results = { heardHum: false, heardBip: false, heardBop: false, heardNoise: false };
function analyse() {
var freqDomain = new Uint8Array(analyser.frequencyBinCount);
analyser.getByteFrequencyData(freqDomain);
var hasFrequency = expectedFrequency => {
var bin = Math.floor(expectedFrequency * analyser.fftSize / context.sampleRate);
return bin < freqDomain.length && freqDomain[bin] >= 150;
};
if (!results.heardHum)
results.heardHum = hasFrequency(150);
if (!results.heardBip)
results.heardBip = hasFrequency(1500);
if (!results.heardBop)
results.heardBop = hasFrequency(500);
if (!results.heardNoise)
results.heardNoise = hasFrequency(3000);
if (results.heardHum && results.heardBip && results.heardBop && results.heardNoise)
done();
};
function done() {
clearTimeout(timeout);
clearInterval(interval);
resolve(results);
}
var timeout = setTimeout(done, 3 * duration);
var interval = setInterval(analyse, duration / 30);
analyse();
});
}
function waitFor(duration)
{
return new Promise((resolve) => setTimeout(resolve, duration));
}
async function doHumAnalysis(streamOrVideo, expected)
{
let context = new AudioContext();
let analyser = setupAudioContext(context, streamOrVideo);
for (var cptr = 0; cptr < 20; cptr++) {
const results = await analyseAudio(analyser, 200, context);
if (results.heardHum === expected) {
await context.close();
return true;
}
await waitFor(50);
}
await context.close();
return false;
}
function isVideoBlack(canvas, video, startX, startY, grabbedWidth, grabbedHeight)
{
canvas.width = video.videoWidth;
canvas.height = video.videoHeight;
if (!grabbedHeight) {
startX = 10;
startY = 10;
grabbedWidth = canvas.width - 20;
grabbedHeight = canvas.height - 20;
}
canvas.getContext('2d').drawImage(video, 0, 0, canvas.width, canvas.height);
imageData = canvas.getContext('2d').getImageData(startX, startY, grabbedWidth, grabbedHeight);
data = imageData.data;
for (var cptr = 0; cptr < grabbedWidth * grabbedHeight; ++cptr) {
// Approximatively black pixels.
if (data[4 * cptr] > 30 || data[4 * cptr + 1] > 30 || data[4 * cptr + 2] > 30)
return false;
}
return true;
}
async function checkVideoBlack(expected, canvas, video, errorMessage, counter)
{
if (isVideoBlack(canvas, video) === expected)
return Promise.resolve();
if (counter === undefined)
counter = 0;
if (counter > 100) {
if (!errorMessage)
errorMessage = "checkVideoBlack timed out expecting " + expected;
return Promise.reject(errorMessage);
}
await waitFor(50);
return checkVideoBlack(expected, canvas, video, errorMessage, ++counter);
}
promise_test(async (test) => {
const stream = await navigator.mediaDevices.getUserMedia({ audio: true, video: false });
const localTrack = stream.getAudioTracks()[0];
localTrack.enabled = false;
setTimeout(() => localTrack.enabled = false, 0);
const recorder = new MediaRecorder(stream);
const dataPromise = new Promise(resolve => recorder.ondataavailable = (e) => resolve(e.data));
recorder.start();
await waitFor(1000);
recorder.stop();
const blob = await dataPromise;
const url = URL.createObjectURL(blob);
video1.src = url;
await video1.play();
assert_greater_than(video1.duration, 0.5);
const results = await doHumAnalysis(video1, false);
assert_true(results, "Should not hear hum");
URL.revokeObjectURL(url);
}, "Recording a muted audio track should produce silence");
promise_test(async (test) => {
const stream = await navigator.mediaDevices.getUserMedia({ audio: true, video: true });
const localTrack = stream.getAudioTracks()[0];
setTimeout(() => localTrack.enabled = false, 50);
const recorder = new MediaRecorder(stream);
const dataPromise = new Promise(resolve => recorder.ondataavailable = (e) => resolve(e.data));
recorder.start();
await waitFor(1000);
recorder.stop();
const blob = await dataPromise;
const url = URL.createObjectURL(blob);
video2.src = url;
await video2.play();
assert_greater_than(video2.duration, 0.5);
const results = await doHumAnalysis(video2, false);
assert_true(results, "Should not hear hum");
URL.revokeObjectURL(url);
}, "Muting an audio track should record silence");
promise_test(async (test) => {
const stream = await navigator.mediaDevices.getUserMedia({ audio: true, video: true });
const localTrack = stream.getVideoTracks()[0];
setTimeout(() => localTrack.enabled = false, 50);
const recorder = new MediaRecorder(stream);
const dataPromise = new Promise(resolve => recorder.ondataavailable = (e) => resolve(e.data));
recorder.start();
await waitFor(1000);
recorder.stop();
const blob = await dataPromise;
const url = URL.createObjectURL(blob);
video3.src = url;
await video3.play();
assert_greater_than(video3.duration, 0.5);
const result = await checkVideoBlack(true, canvas3, video3, "muted video should produce black frames");
URL.revokeObjectURL(url);
}, "Muting a video track should produce black frames");
</script>
</body>
</html>