<!DOCTYPE HTML>
<html>
<head>
  <title>Test ScriptProcessorNode</title>
  <script type="text/javascript" src="/tests/SimpleTest/SimpleTest.js"></script>
  <script type="text/javascript" src="webaudio.js"></script>
  <link rel="stylesheet" type="text/css" href="/tests/SimpleTest/test.css" />
</head>
<body>
<pre id="test">
<script class="testbody" type="text/javascript">

// We do not use our generic graph test framework here because
// the testing logic here is sort of complicated, and would
// not be easy to map to OfflineAudioContext, as ScriptProcessorNodes
// can experience delays.

SimpleTest.waitForExplicitFinish();
addLoadEvent(function() {
  var context = new AudioContext();
  var buffer = null;

  var sourceSP = context.createScriptProcessor(2048);
  sourceSP.addEventListener("audioprocess", function(e) {
    // generate the audio
    for (var i = 0; i < 2048; ++i) {
      // Make sure our first sample won't be zero
      e.outputBuffer.getChannelData(0)[i] = Math.sin(440 * 2 * Math.PI * (i + 1) / context.sampleRate);
      e.outputBuffer.getChannelData(1)[i] = Math.sin(880 * 2 * Math.PI * (i + 1) / context.sampleRate);
    }
    // Remember our generated audio
    buffer = e.outputBuffer;

    sourceSP.removeEventListener("audioprocess", arguments.callee);
  }, false);

  expectException(function() {
    context.createScriptProcessor(1);
  }, DOMException.INDEX_SIZE_ERR);
  expectException(function() {
    context.createScriptProcessor(2);
  }, DOMException.INDEX_SIZE_ERR);
  expectException(function() {
    context.createScriptProcessor(128);
  }, DOMException.INDEX_SIZE_ERR);
  expectException(function() {
    context.createScriptProcessor(255);
  }, DOMException.INDEX_SIZE_ERR);

  is(sourceSP.channelCount, 2, "script processor node has 2 input channels by default");
  is(sourceSP.channelCountMode, "explicit", "Correct channelCountMode for the script processor node");
  is(sourceSP.channelInterpretation, "speakers", "Correct channelCountInterpretation for the script processor node");

  function findFirstNonZeroSample(buffer) {
    for (var i = 0; i < buffer.length; ++i) {
      if (buffer.getChannelData(0)[i] != 0) {
        return i;
      }
    }
    return buffer.length;
  }

  var sp = context.createScriptProcessor(2048);
  sourceSP.connect(sp);
  sp.connect(context.destination);
  var lastPlaybackTime = 0;

  var emptyBuffer = context.createBuffer(1, 2048, context.sampleRate);

  function checkAudioProcessingEvent(e) {
    is(e.target, sp, "Correct event target");
    ok(e.playbackTime > lastPlaybackTime, "playbackTime correctly set");
    lastPlaybackTime = e.playbackTime;
    is(e.inputBuffer.numberOfChannels, 2, "Correct number of channels for the input buffer");
    is(e.inputBuffer.length, 2048, "Correct length for the input buffer");
    is(e.inputBuffer.sampleRate, context.sampleRate, "Correct sample rate for the input buffer");
    is(e.outputBuffer.numberOfChannels, 2, "Correct number of channels for the output buffer");
    is(e.outputBuffer.length, 2048, "Correct length for the output buffer");
    is(e.outputBuffer.sampleRate, context.sampleRate, "Correct sample rate for the output buffer");

    compareChannels(e.outputBuffer.getChannelData(0), emptyBuffer.getChannelData(0));
    compareChannels(e.outputBuffer.getChannelData(1), emptyBuffer.getChannelData(0));
  }

  sp.onaudioprocess = function(e) {
    isnot(buffer, null, "The audioprocess handler for sourceSP must be run at this point");
    checkAudioProcessingEvent(e);

    // Because of the initial latency added by the second script processor node,
    // we will never see any generated audio frames in the first callback.
    compareChannels(e.inputBuffer.getChannelData(0), emptyBuffer.getChannelData(0));
    compareChannels(e.inputBuffer.getChannelData(1), emptyBuffer.getChannelData(0));

    sp.onaudioprocess = function(e) {
      checkAudioProcessingEvent(e);

      var firstNonZero = findFirstNonZeroSample(e.inputBuffer);
      ok(firstNonZero <= 2048, "First non-zero sample within range");

      compareChannels(e.inputBuffer.getChannelData(0), emptyBuffer.getChannelData(0), firstNonZero);
      compareChannels(e.inputBuffer.getChannelData(1), emptyBuffer.getChannelData(0), firstNonZero);
      compareChannels(e.inputBuffer.getChannelData(0), buffer.getChannelData(0), 2048 - firstNonZero, firstNonZero, 0);
      compareChannels(e.inputBuffer.getChannelData(1), buffer.getChannelData(1), 2048 - firstNonZero, firstNonZero, 0);

      if (firstNonZero == 0) {
        // If we did not experience any delays, the test is done!
        sp.onaudioprocess = null;

        SimpleTest.finish();
      } else if (firstNonZero != 2048) {
        // In case we just saw a zero buffer this time, wait one more round
        sp.onaudioprocess = function(e) {
          checkAudioProcessingEvent(e);

          compareChannels(e.inputBuffer.getChannelData(0), buffer.getChannelData(0), firstNonZero, 0, 2048 - firstNonZero);
          compareChannels(e.inputBuffer.getChannelData(1), buffer.getChannelData(1), firstNonZero, 0, 2048 - firstNonZero);
          compareChannels(e.inputBuffer.getChannelData(0), emptyBuffer.getChannelData(0), undefined, firstNonZero);
          compareChannels(e.inputBuffer.getChannelData(1), emptyBuffer.getChannelData(0), undefined, firstNonZero);

          sp.onaudioprocess = null;

          SimpleTest.finish();
        };
      }
    };
  };
});

</script>
</pre>
</body>
</html>