Telegram Web, preconfigured for usage in I2P. http://web.telegram.i2p/
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
 
 

9663 lines
277 KiB

(function webpackUniversalModuleDefinition(root, factory) {
if(typeof exports === 'object' && typeof module === 'object')
module.exports = factory();
else if(typeof define === 'function' && define.amd)
define([], factory);
else if(typeof exports === 'object')
exports["ogvjs"] = factory();
else
root["ogvjs"] = factory();
})(this, function() {
return /******/ (function(modules) { // webpackBootstrap
/******/ // The module cache
/******/ var installedModules = {};
/******/ // The require function
/******/ function __webpack_require__(moduleId) {
/******/ // Check if module is in cache
/******/ if(installedModules[moduleId])
/******/ return installedModules[moduleId].exports;
/******/ // Create a new module (and put it into the cache)
/******/ var module = installedModules[moduleId] = {
/******/ exports: {},
/******/ id: moduleId,
/******/ loaded: false
/******/ };
/******/ // Execute the module function
/******/ modules[moduleId].call(module.exports, module, module.exports, __webpack_require__);
/******/ // Flag the module as loaded
/******/ module.loaded = true;
/******/ // Return the exports of the module
/******/ return module.exports;
/******/ }
/******/ // expose the modules object (__webpack_modules__)
/******/ __webpack_require__.m = modules;
/******/ // expose the module cache
/******/ __webpack_require__.c = installedModules;
/******/ // __webpack_public_path__
/******/ __webpack_require__.p = "";
/******/ // Load entry module and return exports
/******/ return __webpack_require__(0);
/******/ })
/************************************************************************/
/******/ ([
/* 0 */
/***/ (function(module, exports, __webpack_require__) {
//
// -- ogv.js
// https://github.com/brion/ogv.js
// Copyright (c) 2013-2016 Brion Vibber
//
var OGVCompat = __webpack_require__(1),
OGVLoader = __webpack_require__(3),
OGVMediaType = __webpack_require__(7),
OGVPlayer = __webpack_require__(8),
OGVVersion = ("1.4.2-20170425024925-504d7197");
// Version 1.0's web-facing and test-facing interfaces
if (window) {
window.OGVCompat = OGVCompat;
window.OGVLoader = OGVLoader;
window.OGVMediaType = OGVMediaType;
window.OGVPlayer = OGVPlayer;
window.OGVVersion = OGVVersion;
}
module.exports = {
OGVCompat: OGVCompat,
OGVLoader: OGVLoader,
OGVMediaType: OGVMediaType,
OGVPlayer: OGVPlayer,
OGVVersion: OGVVersion
};
/***/ }),
/* 1 */
/***/ (function(module, exports, __webpack_require__) {
var BogoSlow = __webpack_require__(2);
var OGVCompat = {
benchmark: new BogoSlow(),
hasTypedArrays: function() {
// emscripten-compiled code requires typed arrays
return !!window.Uint32Array;
},
hasWebAudio: function() {
return !!(window.AudioContext || window.webkitAudioContext);
},
hasFlash: function() {
if (navigator.userAgent.indexOf('Trident') !== -1) {
// We only do the ActiveX test because we only need Flash in
// Internet Explorer 10/11. Other browsers use Web Audio directly
// (Edge, Safari) or native playback, so there's no need to test
// other ways of loading Flash.
try {
var obj = new ActiveXObject('ShockwaveFlash.ShockwaveFlash');
return true;
} catch(e) {
return false;
}
}
return false;
},
hasAudio: function() {
return this.hasWebAudio() || this.hasFlash();
},
isBlacklisted: function(userAgent) {
// JIT bugs in old Safari versions
var blacklist = [
/\(i.* OS [67]_.* like Mac OS X\).* Mobile\/.* Safari\//,
/\(Macintosh.* Version\/6\..* Safari\/\d/
];
var blacklisted = false;
blacklist.forEach(function(regex) {
if (userAgent.match(regex)) {
blacklisted = true;
}
});
return blacklisted;
},
isSlow: function() {
return this.benchmark.slow;
},
isTooSlow: function() {
return this.benchmark.tooSlow;
},
supported: function(component) {
if (component === 'OGVDecoder') {
return (this.hasTypedArrays() && !this.isBlacklisted(navigator.userAgent));
}
if (component === 'OGVPlayer') {
return (this.supported('OGVDecoder') && this.hasAudio() && !this.isTooSlow());
}
return false;
}
};
module.exports = OGVCompat;
/***/ }),
/* 2 */
/***/ (function(module, exports) {
/**
* A quick CPU/JS engine benchmark to guesstimate whether we're
* fast enough to handle 360p video in JavaScript.
*/
function BogoSlow() {
var self = this;
var timer;
// FIXME: avoid to use window scope here
if (window.performance && window.performance.now) {
timer = function() {
return window.performance.now();
};
} else {
timer = function() {
return Date.now();
};
}
var savedSpeed = null;
function run() {
var ops = 0;
var fibonacci = function(n) {
ops++;
if (n < 2) {
return n;
} else {
return fibonacci(n - 2) + fibonacci(n - 1);
}
};
var start = timer();
fibonacci(30);
var delta = timer() - start;
savedSpeed = (ops / delta);
}
/**
* Return a scale value of operations/sec from the benchmark.
* If the benchmark has already been run, uses a memoized result.
*
* @property {number}
*/
Object.defineProperty(self, 'speed', {
get: function() {
if (savedSpeed === null) {
run();
}
return savedSpeed;
}
});
/**
* Return the defined cutoff speed value for 'slow' devices,
* based on results measured from some test devices.
*
* @property {number}
*/
Object.defineProperty(self, 'slowCutoff', {
get: function() {
// 2012 Retina MacBook Pro (Safari 7) ~150,000
// 2009 Dell T5500 (IE 11) ~100,000
// iPad Air (iOS 7) ~65,000
// 2010 MBP / OS X 10.9 (Safari 7) ~62,500
// 2010 MBP / Win7 VM (IE 11) ~50,000+-
// ^ these play 360p ok
// ----------- line of moderate doom ----------
return 50000;
// v these play 160p ok
// iPad Mini non-Retina (iOS 8 beta) ~25,000
// Dell Inspiron Duo (IE 11) ~25,000
// Surface RT (IE 11) ~18,000
// iPod Touch 5th-gen (iOS 8 beta) ~16,000
}
});
/**
* Return the defined cutoff speed value for 'too slow' devices,
* based on results measured from some test devices.
*
* No longer used.
*
* @property {number}
* @deprecated
*/
Object.defineProperty(self, 'tooSlowCutoff', {
get: function() {
return 0;
}
});
/**
* 'Slow' devices can play audio and should sorta play our
* extra-tiny Wikimedia 160p15 transcodes
*
* @property {boolean}
*/
Object.defineProperty(self, 'slow', {
get: function() {
return (self.speed < self.slowCutoff);
}
});
/**
* 'Too slow' devices aren't reliable at all
*
* @property {boolean}
*/
Object.defineProperty(self, 'tooSlow', {
get: function() {
return (self.speed < self.tooSlowCutoff);
}
});
}
module.exports = BogoSlow;
/***/ }),
/* 3 */
/***/ (function(module, exports, __webpack_require__) {
var OGVVersion = ("1.4.2-20170425024925-504d7197");
(function() {
var global = this;
var scriptMap = {
OGVDemuxerOgg: 'ogv-demuxer-ogg.js',
OGVDemuxerOggW: 'ogv-demuxer-ogg-wasm.js',
OGVDemuxerWebM: 'ogv-demuxer-webm.js',
OGVDemuxerWebMW: 'ogv-demuxer-webm-wasm.js',
OGVDecoderAudioOpus: 'ogv-decoder-audio-opus.js',
OGVDecoderAudioOpusW: 'ogv-decoder-audio-opus-wasm.js',
OGVDecoderAudioVorbis: 'ogv-decoder-audio-vorbis.js',
OGVDecoderAudioVorbisW: 'ogv-decoder-audio-vorbis-wasm.js',
OGVDecoderVideoTheora: 'ogv-decoder-video-theora.js',
OGVDecoderVideoTheoraW: 'ogv-decoder-video-theora-wasm.js',
OGVDecoderVideoVP8: 'ogv-decoder-video-vp8.js',
OGVDecoderVideoVP8W: 'ogv-decoder-video-vp8-wasm.js',
OGVDecoderVideoVP8MT: 'ogv-decoder-video-vp8-mt.js',
OGVDecoderVideoVP9: 'ogv-decoder-video-vp9.js',
OGVDecoderVideoVP9W: 'ogv-decoder-video-vp9-wasm.js',
OGVDecoderVideoVP9MT: 'ogv-decoder-video-vp9-mt.js'
};
// @fixme make this less awful
var proxyTypes = {
OGVDecoderAudioOpus: 'audio',
OGVDecoderAudioOpusW: 'audio',
OGVDecoderAudioVorbis: 'audio',
OGVDecoderAudioVorbisW: 'audio',
OGVDecoderVideoTheora: 'video',
OGVDecoderVideoTheoraW: 'video',
OGVDecoderVideoVP8: 'video',
OGVDecoderVideoVP8W: 'video',
OGVDecoderVideoVP9: 'video',
OGVDecoderVideoVP9W: 'video'
};
var proxyInfo = {
audio: {
proxy: __webpack_require__(4),
worker: 'ogv-worker-audio.js',
},
video: {
proxy: __webpack_require__(6),
worker: 'ogv-worker-video.js'
}
}
function urlForClass(className) {
var scriptName = scriptMap[className];
if (scriptName) {
return urlForScript(scriptName);
} else {
throw new Error('asked for URL for unknown class ' + className);
}
};
function urlForScript(scriptName) {
if (scriptName) {
var base = OGVLoader.base;
if (base === undefined) {
base = '';
} else {
base += '/';
}
return base + scriptName + '?version=' + encodeURIComponent(OGVVersion);
} else {
throw new Error('asked for URL for unknown script ' + scriptName);
}
};
var scriptStatus = {},
scriptCallbacks = {};
function loadWebScript(src, callback) {
if (scriptStatus[src] == 'done') {
callback();
} else if (scriptStatus[src] == 'loading') {
scriptCallbacks[src].push(callback);
} else {
scriptStatus[src] = 'loading';
scriptCallbacks[src] = [callback];
var scriptNode = document.createElement('script');
function done(event) {
var callbacks = scriptCallbacks[src];
delete scriptCallbacks[src];
scriptStatus[src] = 'done';
callbacks.forEach(function(cb) {
cb();
});
}
scriptNode.addEventListener('load', done);
scriptNode.addEventListener('error', done);
scriptNode.src = src;
document.querySelector('head').appendChild(scriptNode);
}
}
function loadWebAssembly(src, callback) {
if (!src.match(/-wasm\.js/)) {
callback(null);
} else {
var wasmSrc = src.replace(/-wasm\.js/, '-wasm.wasm');
var xhr = new XMLHttpRequest();
xhr.responseType = 'arraybuffer';
xhr.onload = function() {
callback(xhr.response);
};
xhr.onerror = function() {
callback(null);
};
xhr.open('GET', wasmSrc);
xhr.send();
}
}
function defaultBase() {
if (typeof global.window === 'object') {
// for browser, try to autodetect
var scriptNodes = document.querySelectorAll('script'),
regex = /^(?:|(.*)\/)ogv(?:-support)?\.js(?:\?|#|$)/,
path,
matches;
for (var i = 0; i < scriptNodes.length; i++) {
path = scriptNodes[i].getAttribute('src');
if (path) {
matches = path.match(regex);
if (matches) {
return matches[1];
}
}
}
return undefined; // current dir
} else {
// for workers, assume current directory
// if not a worker, too bad.
return undefined;
}
}
var OGVLoader = {
base: defaultBase(),
loadClass: function(className, callback, options) {
options = options || {};
if (options.worker) {
this.workerProxy(className, callback);
return;
}
var url = urlForClass(className);
loadWebAssembly(url, function(wasmBinary) {
function wasmClassWrapper(options) {
options = options || {};
if (wasmBinary !== null) {
options.wasmBinary = wasmBinary;
}
return new global[className](options);
}
if (typeof global[className] === 'function') {
// already loaded!
callback(wasmClassWrapper);
} else if (typeof global.window === 'object') {
loadWebScript(url, function() {
callback(wasmClassWrapper);
});
} else if (typeof global.importScripts === 'function') {
// worker has convenient sync importScripts
global.importScripts(url);
callback(wasmClassWrapper);
}
});
},
workerProxy: function(className, callback) {
var proxyType = proxyTypes[className],
info = proxyInfo[proxyType];
if (!info) {
throw new Error('Requested worker for class with no proxy: ' + className);
}
var proxyClass = info.proxy,
workerScript = info.worker,
codecUrl = urlForScript(scriptMap[className]),
workerUrl = urlForScript(workerScript),
worker;
var construct = function(options) {
return new proxyClass(worker, className, options);
};
if (workerUrl.match(/^https?:|\/\//i)) {
// Can't load workers natively cross-domain, but if CORS
// is set up we can fetch the worker stub and the desired
// class and load them from a blob.
var getCodec,
getWorker,
codecResponse,
workerResponse,
codecLoaded = false,
workerLoaded = false,
blob;
function completionCheck() {
if ((codecLoaded == true) && (workerLoaded == true)) {
try {
blob = new Blob([codecResponse + " " + workerResponse], {type: 'application/javascript'});
} catch (e) { // Backwards-compatibility
window.BlobBuilder = window.BlobBuilder || window.WebKitBlobBuilder || window.MozBlobBuilder;
blob = new BlobBuilder();
blob.append(codecResponse + " " + workerResponse);
blob = blob.getBlob();
}
// Create the web worker
worker = new Worker(URL.createObjectURL(blob));
callback(construct);
}
}
// Load the codec
getCodec = new XMLHttpRequest();
getCodec.open("GET", codecUrl, true);
getCodec.onreadystatechange = function() {
if(getCodec.readyState == 4 && getCodec.status == 200) {
codecResponse = getCodec.responseText;
// Update the codec response loaded flag
codecLoaded = true;
completionCheck();
}
};
getCodec.send();
// Load the worker
getWorker = new XMLHttpRequest();
getWorker.open("GET", workerUrl, true);
getWorker.onreadystatechange = function() {
if(getWorker.readyState == 4 && getWorker.status == 200) {
workerResponse = getWorker.responseText;
// Update the worker response loaded flag
workerLoaded = true;
completionCheck();
}
};
getWorker.send();
} else {
// Local URL; load it directly for simplicity.
worker = new Worker(workerUrl);
callback(construct);
}
}
};
module.exports = OGVLoader;
})();
/***/ }),
/* 4 */
/***/ (function(module, exports, __webpack_require__) {
var OGVProxyClass = __webpack_require__(5);
var OGVDecoderAudioProxy = OGVProxyClass({
loadedMetadata: false,
audioFormat: null,
audioBuffer: null,
cpuTime: 0
}, {
init: function(callback) {
this.proxy('init', [], callback);
},
processHeader: function(data, callback) {
this.proxy('processHeader', [data], callback, [data]);
},
processAudio: function(data, callback) {
this.proxy('processAudio', [data], callback, [data]);
},
close: function() {
this.terminate();
}
});
module.exports = OGVDecoderAudioProxy;
/***/ }),
/* 5 */
/***/ (function(module, exports) {
/**
* Proxy object for web worker interface for codec classes.
*
* Used by the high-level player interface.
*
* @author Brion Vibber <brion@pobox.com>
* @copyright 2015
* @license MIT-style
*/
function OGVProxyClass(initialProps, methods) {
return function(worker, className, options) {
options = options || {};
var self = this;
var transferables = (function() {
var buffer = new ArrayBuffer(1024),
bytes = new Uint8Array(buffer);
try {
worker.postMessage({
action: 'transferTest',
bytes: bytes
}, [buffer]);
if (buffer.byteLength) {
// No transferable support
return false;
} else {
return true;
}
} catch (e) {
return false;
}
})();
// Set up proxied property getters
var props = {};
for (var iPropName in initialProps) {
if (initialProps.hasOwnProperty(iPropName)) {
(function(propName) {
props[propName] = initialProps[propName];
Object.defineProperty(self, propName, {
get: function getProperty() {
return props[propName];
}
});
})(iPropName);
}
}
// Current player wants to avoid async confusion.
var processingQueue = 0;
Object.defineProperty(self, 'processing', {
get: function() {
return (processingQueue > 0);
}
});
// Set up proxied methods
for (var method in methods) {
if (methods.hasOwnProperty(method)) {
self[method] = methods[method];
}
}
// And some infrastructure!
var messageCount = 0,
pendingCallbacks = {};
this.proxy = function(action, args, callback, transfers) {
if (!worker) {
throw 'Tried to call "' + action + '" method on closed proxy object';
}
var callbackId = 'callback-' + (++messageCount) + '-' + action;
if (callback) {
pendingCallbacks[callbackId] = callback;
}
var out = {
'action': action,
'callbackId': callbackId,
'args': args || []
};
processingQueue++;
if (transferables) {
worker.postMessage(out, transfers || []);
} else {
worker.postMessage(out);
}
};
this.terminate = function() {
if (worker) {
worker.terminate();
worker = null;
processingQueue = 0;
pendingCallbacks = {};
}
};
worker.addEventListener('message', function proxyOnMessage(event) {
processingQueue--;
if (event.data.action !== 'callback') {
// ignore
return;
}
var data = event.data,
callbackId = data.callbackId,
args = data.args,
callback = pendingCallbacks[callbackId];
// Save any updated properties returned to us...
if (data.props) {
for (var propName in data.props) {
if (data.props.hasOwnProperty(propName)) {
props[propName] = data.props[propName];
}
}
}
if (callback) {
delete pendingCallbacks[callbackId];
callback.apply(this, args);
}
});
// Tell the proxy to load and initialize the appropriate class
self.proxy('construct', [className, options], function() {});
return self;
};
}
module.exports = OGVProxyClass;
/***/ }),
/* 6 */
/***/ (function(module, exports, __webpack_require__) {
var OGVProxyClass = __webpack_require__(5);
var OGVDecoderVideoProxy = OGVProxyClass({
loadedMetadata: false,
videoFormat: null,
frameBuffer: null,
cpuTime: 0
}, {
init: function(callback) {
this.proxy('init', [], callback);
},
processHeader: function(data, callback) {
this.proxy('processHeader', [data], callback, [data]);
},
processFrame: function(data, callback) {
this.proxy('processFrame', [data], callback, [data]);
},
close: function() {
this.terminate();
}
});
module.exports = OGVDecoderVideoProxy;
/***/ }),
/* 7 */
/***/ (function(module, exports) {
function OGVMediaType(contentType) {
contentType = '' + contentType;
var self = this;
self.major = null;
self.minor = null;
self.codecs = null;
function trim(str) {
return str.replace(/^\s+/, '').replace(/\s+$/, '');
}
function split(str, sep, limit) {
var bits = str.split(sep, limit).map(function(substr) {
return trim(substr);
});
if (typeof limit === 'number') {
while (bits.length < limit) {
bits.push(null);
}
}
return bits;
}
var parts = split(contentType, ';');
if (parts.length) {
var base = parts.shift();
if (base) {
var bits = split(base, '/', 2);
self.major = bits[0];
self.minor = bits[1];
}
parts.forEach(function(str) {
var matches = str.match(/^codecs\s*=\s*"(.*?)"$/);
if (matches) {
self.codecs = split(matches[1], ',');
}
});
}
}
module.exports = OGVMediaType;
/***/ }),
/* 8 */
/***/ (function(module, exports, __webpack_require__) {
__webpack_require__(9).polyfill();
var YUVCanvas = __webpack_require__(12);
// -- OGVLoader.js
var OGVLoader = __webpack_require__(3);
// -- StreamFile.js
var StreamFile = __webpack_require__(19);
// -- AudioFeeder.js
var AudioFeeder = __webpack_require__(30),
dynamicaudio_swf = __webpack_require__(31);
// -- Bisector.js
var Bisector = __webpack_require__(32);
// -- OGVMediaType.js
var OGVMediaType = __webpack_require__(7);
// -- OGVWrapperCodec.js
var OGVWrapperCodec = __webpack_require__(33);
// -- OGVDecoderAudioProxy.js
var OGVDecoderAudioProxy = __webpack_require__(4);
// -- OGVDecoderVideoProxy.js
var OGVDecoderVideoProxy = __webpack_require__(6);
/**
* Constructor for an analogue of the TimeRanges class
* returned by various HTMLMediaElement properties
*
* Pass an array of two-element arrays, each containing a start and end time.
*/
// FIXME: not use window scope if possible here
var OGVTimeRanges = window.OGVTimeRanges = function(ranges) {
Object.defineProperty(this, 'length', {
get: function getLength() {
return ranges.length;
}
});
this.start = function(i) {
return ranges[i][0];
};
this.end = function(i) {
return ranges[i][1];
};
return this;
};
var OGVMediaErrorConstants = {
MEDIA_ERR_ABORTED: 1,
MEDIA_ERR_NETWORK: 2,
MEDIA_ERR_DECODE: 3,
MEDIA_ERR_SRC_NOT_SUPPORTED: 4
};
function extend(dest, src) {
for (prop in src) {
if (src.hasOwnProperty(prop)) {
dest[prop] = src[prop];
}
}
}
/**
* Constructor for analogue of the MediaError class
* returned by HTMLMediaElement.error property
*/
var OGVMediaError = window.OGVMediaError = function(code, message) {
this.code = code;
this.message = message;
};
extend(OGVMediaError, OGVMediaErrorConstants);
extend(OGVMediaError.prototype, OGVMediaErrorConstants);
/**
* Player class -- instantiate one of these to get an 'ogvjs' HTML element
* which has a similar interface to the HTML audio/video elements.
*
* @param options: optional dictionary of options:
* 'base': string; base URL for additional resources, such as Flash audio shim
* 'webGL': bool; pass true to use WebGL acceleration if available
* 'forceWebGL': bool; pass true to require WebGL even if not detected
* 'sync': string; A/V sync plan: one of:
* 'delay-audio' (pre-1.3.2 behavior) stop audio when behind, then play all frames to catch up
* 'skip-frames' (default) to preserve audio, skipping frames until the next sync point (keyframe)
*/
var OGVPlayer = function(options) {
options = options || {};
var instanceId = 'ogvjs' + (++OGVPlayer.instanceCount);
var codecClass = null,
codecType = null;
var canvasOptions = {};
if (options.webGL !== undefined) {
// @fixme confirm format of webGL option
canvasOptions.webGL = options.webGL;
}
if(!!options.forceWebGL) {
canvasOptions.webGL = 'required';
}
// Experimental options
var enableWebM = !!options.enableWebM;
// Running the codec in a worker thread equals happy times!
var enableWorker = !!window.Worker;
if (typeof options.worker !== 'undefined') {
enableWorker = !!options.worker;
}
var enableThreading = !!options.threading;
var enableWASM = !!options.wasm;
if (options.sync === undefined) {
options.sync = 'skip-frames';
}
var State = {
INITIAL: 'INITIAL',
SEEKING_END: 'SEEKING_END',
LOADED: 'LOADED',
PRELOAD: 'PRELOAD',
READY: 'READY',
PLAYING: 'PLAYING',
SEEKING: 'SEEKING',
ENDED: 'ENDED',
ERROR: 'ERROR'
}, state = State.INITIAL;
var SeekState = {
NOT_SEEKING: 'NOT_SEEKING',
BISECT_TO_TARGET: 'BISECT_TO_TARGET',
BISECT_TO_KEYPOINT: 'BISECT_TO_KEYPOINT',
LINEAR_TO_TARGET: 'LINEAR_TO_TARGET'
}, seekState = SeekState.NOT_SEEKING;
var audioOptions = {},
codecOptions = {};
options.base = options.base || OGVLoader.base;
if (typeof options.base === 'string') {
// Pass the resource dir down to AudioFeeder, so it can load the dynamicaudio.swf
audioOptions.base = options.base;
// And to the worker thread, so it can load the codec JS
codecOptions.base = options.base;
}
if (typeof options.audioContext !== 'undefined') {
// Try passing a pre-created audioContext in?
audioOptions.audioContext = options.audioContext;
}
if (typeof options.audioDestination !== 'undefined') {
// Try passing a pre-created audioContext in?
audioOptions.output = options.audioDestination;
}
// Buffer in largeish chunks to survive long CPU spikes on slow CPUs (eg, 32-bit iOS)
audioOptions.bufferSize = 8192;
codecOptions.worker = enableWorker;
codecOptions.threading = enableThreading;
codecOptions.wasm = enableWASM;
if (typeof options.memoryLimit === 'number') {
codecOptions.memoryLimit = options.memoryLimit;
}
var canvas = document.createElement('canvas');
var frameSink;
// Return a magical custom element!
var self = document.createElement('ogvjs');
self.className = instanceId;
extend(self, constants);
canvas.style.position = 'absolute';
canvas.style.top = '0';
canvas.style.left = '0';
canvas.style.width = '100%';
canvas.style.height = '100%';
canvas.style.objectFit = 'contain';
self.appendChild(canvas);
var getTimestamp;
// FIXME: don't use window scope, see BogoSlow.js
if (window.performance === undefined || window.performance.now === undefined) {
getTimestamp = Date.now;
} else {
getTimestamp = window.performance.now.bind(window.performance);
}
function time(cb) {
var start = getTimestamp();
cb();
var delta = getTimestamp() - start;
lastFrameDecodeTime += delta;
return delta;
}
var then = getTimestamp();
function log(msg) {
if (options.debug) {
var now = getTimestamp(),
delta = now - then;
//console.log('+' + delta + 'ms ' + msg);
//then = now;
if (!options.debugFilter || msg.match(options.debugFilter)) {
console.log('[' + (Math.round(delta * 10) / 10) + 'ms] ' + msg);
}
}
}
function fireEvent(eventName, props) {
log('fireEvent ' + eventName);
var event;
props = props || {};
var standard = (typeof window.Event == 'function');
if (standard) {
// standard event creation
event = new CustomEvent(eventName);
} else {
// IE back-compat mode
// https://msdn.microsoft.com/en-us/library/dn905219%28v=vs.85%29.aspx
event = document.createEvent('Event');
event.initEvent(eventName, false, false);
}
for (var prop in props) {
if (props.hasOwnProperty(prop)) {
event[prop] = props[prop];
}
}
var allowDefault = self.dispatchEvent(event);
if (!standard && eventName === 'resize' && self.onresize && allowDefault) {
// resize demands special treatment!
// in IE 11 it doesn't fire through to the .onresize handler
// for some crazy reason
self.onresize.call(self, event);
}
}
function fireEventAsync(eventName, props) {
log('fireEventAsync ' + eventName);
setTimeout(function() {
fireEvent(eventName, props);
}, 0);
}
var codec = null,
videoInfo = null,
audioInfo = null,
actionQueue = [],
audioFeeder = null;
var muted = false,
initialPlaybackPosition = 0.0,
initialPlaybackOffset = 0.0,
prebufferingAudio = false,
stoppedForLateFrame = false,
initialSeekTime = 0.0;
function initAudioFeeder() {
audioFeeder = new AudioFeeder( audioOptions );
audioFeeder.init(audioInfo.channels, audioInfo.rate);
// At our requested 8192 buffer size, bufferDuration should be
// about 185ms at 44.1 kHz or 170ms at 48 kHz output, covering
// 4-6 frames at 24-30fps.
//
// 2 or 3 buffers' worth will be in flight at any given time;
// be warned that durationBuffered includes that in-flight stuff.
//
// Set our bufferThreshold to a full 1s for plenty of extra headroom,
// and make sure we've queued up twice that when we're operating.
audioFeeder.bufferThreshold = 1;
audioFeeder.volume = self.volume;
audioFeeder.muted = self.muted;
// If we're in a background tab, timers may be throttled.
// audioFeeder will call us when buffers need refilling,
// without any throttling.
audioFeeder.onbufferlow = function audioCallback() {
log('onbufferlow');
if ((stream && (stream.buffering || stream.seeking)) || pendingAudio) {
// We're waiting on input or other async processing;
// we'll get triggered later.
} else {
// We're in an async event so it's safe to run the loop:
pingProcessing();
}
};
// If we ran out of audio *completely* schedule some more processing.
// This shouldn't happen if we keep up with onbufferlow, except at
// the very beginning of playback when we haven't buffered any data yet.
// @todo pre-buffer a little data to avoid needing this
audioFeeder.onstarved = function () {
if (dataEnded) {
// Probably end of file.
// Do nothing!
log('onstarved: appear to have reached end of audio');
} else {
log('onstarved: halting audio due to starvation');
stopPlayback();
prebufferingAudio = true;
}
if (isProcessing()) {
// We're waiting on input or other async processing;
// we'll get triggered later.
} else {
// Schedule loop after this synchronous event.
pingProcessing(0);
}
};
}
function startPlayback(offset) {
if (audioFeeder) {
audioFeeder.start();
var state = audioFeeder.getPlaybackState();
initialPlaybackPosition = state.playbackPosition;
} else {
initialPlaybackPosition = getTimestamp() / 1000;
}
if (offset !== undefined) {
initialPlaybackOffset = offset;
}
// Clear the late flag if it was set.
prebufferingAudio = false;
stoppedForLateFrame = false;
log('continuing at ' + initialPlaybackPosition + ', ' + initialPlaybackOffset);
}
function stopPlayback() {
initialPlaybackOffset = getPlaybackTime();
log('pausing at ' + initialPlaybackOffset);
if (audioFeeder) {
audioFeeder.stop();
}
}
/**
* Get audio playback time position in file's units
*
* @return {number} seconds since file start
*/
function getPlaybackTime(state) {
if (prebufferingAudio || stoppedForLateFrame || paused) {
return initialPlaybackOffset;
} else {
var position;
if (audioFeeder) {
state = state || audioFeeder.getPlaybackState();
position = state.playbackPosition;
} else {
// @fixme handle paused/stoped time better
position = getTimestamp() / 1000;
}
return (position - initialPlaybackPosition) + initialPlaybackOffset;
}
}
var currentSrc = '',
stream,
streamEnded = false,
seekCancel = {},
readCancel = {},
mediaError = null,
dataEnded = false,
byteLength = 0,
duration = null,
lastSeenTimestamp = null,
nextProcessingTimer,
nextFrameTimer = null,
loading = false,
started = false,
paused = true,
ended = false,
startedPlaybackInDocument = false;
var framesPlayed = 0;
// Benchmark data, exposed via getPlaybackStats()
var framesProcessed = 0, // frames
targetPerFrameTime = 1000 / 60, // ms
actualPerFrameTime = 0, // ms
totalFrameTime = 0, // ms
totalFrameCount = 0, // frames
playTime = 0, // ms
bufferTime = 0, // ms
drawingTime = 0, // ms
proxyTime = 0, // ms
totalJitter = 0; // sum of ms we're off from expected frame delivery time
// Benchmark data that doesn't clear
var droppedAudio = 0, // number of times we were starved for audio
delayedAudio = 0, // seconds audio processing was delayed by blocked CPU
lateFrames = 0; // number of times a frame was late and we had to halt audio
var poster = '', thumbnail;
// called when stopping old video on load()
function stopVideo() {
log("STOPPING");
// kill the previous video if any
state = State.INITIAL;
seekState = SeekState.NOT_SEEKING;
started = false;
//paused = true; // don't change this?
ended = false;
frameEndTimestamp = 0.0;
audioEndTimestamp = 0.0;
lastFrameDecodeTime = 0.0;
stoppedForLateFrame = false;
prebufferingAudio = false;
// Abort all queued actions
actionQueue.splice(0, actionQueue.length);
if (stream) {
// @todo fire an abort event if still loading
// @todo fire an emptied event if previously had data
stream.abort();
stream = null;
streamEnded = false;
}
if (codec) {
codec.close();
codec = null;
pendingFrame = 0;
pendingAudio = 0;
dataEnded = false;
}
videoInfo = null;
audioInfo = null;
if (audioFeeder) {
audioFeeder.close();
audioFeeder = null;
}
if (nextProcessingTimer) {
clearTimeout(nextProcessingTimer);
nextProcessingTimer = null;
}
if (nextFrameTimer) {
clearTimeout(nextFrameTimer);
nextFrameTimer = null;
}
if (frameSink) {
frameSink.clear();
frameSink = null;
}
if (decodedFrames) {
decodedFrames = [];
}
if (pendingFrames) {
pendingFrames = [];
}
initialSeekTime = 0.0;
// @todo set playback position, may need to fire timeupdate if wasnt previously 0
initialPlaybackPosition = 0;
initialPlaybackOffset = 0;
duration = null; // do not fire durationchange
// timeline offset to 0?
}
var lastFrameTime = getTimestamp(),
frameEndTimestamp = 0.0,
audioEndTimestamp = 0.0,
decodedFrames = [],
pendingFrames = [];
var lastFrameDecodeTime = 0.0,
lastFrameVideoCpuTime = 0,
lastFrameAudioCpuTime = 0,
lastFrameDemuxerCpuTime = 0,
lastFrameDrawingTime = 0,
lastFrameBufferTime = 0,
lastFrameProxyTime = 0,
lastVideoCpuTime = 0,
lastAudioCpuTime = 0,
lastDemuxerCpuTime = 0,
lastBufferTime = 0,
lastProxyTime = 0,
lastDrawingTime = 0;
var lastFrameTimestamp = 0.0,
currentVideoCpuTime = 0.0;
var lastTimeUpdate = 0, // ms
timeUpdateInterval = 250; // ms
function doFrameComplete(data) {
data = data || {};
if (startedPlaybackInDocument && !document.body.contains(self)) {
// We've been de-parented since we last ran
// Stop playback at next opportunity!
setTimeout(function() {
self.stop();
}, 0);
}
var newFrameTimestamp = getTimestamp(),
wallClockTime = newFrameTimestamp - lastFrameTimestamp,
//jitter = wallClockTime - targetPerFrameTime;
jitter = actualPerFrameTime - targetPerFrameTime;
totalJitter += Math.abs(jitter);
playTime += wallClockTime;
var timing = {
cpuTime: lastFrameDecodeTime,
drawingTime: drawingTime - lastFrameDrawingTime,
bufferTime: bufferTime - lastFrameBufferTime,
proxyTime: proxyTime - lastFrameProxyTime,
demuxerTime: 0,
videoTime: 0,
audioTime: 0,
//clockTime: wallClockTime
clockTime: actualPerFrameTime,
late: stoppedForLateFrame || data.dropped,
dropped: data.dropped
};
if (codec) {
timing.demuxerTime = (codec.demuxerCpuTime - lastFrameDemuxerCpuTime);
timing.videoTime += (currentVideoCpuTime - lastFrameVideoCpuTime);
timing.audioTime += (codec.audioCpuTime - lastFrameAudioCpuTime);
}
timing.cpuTime += timing.demuxerTime;
lastFrameDecodeTime = 0;
lastFrameTimestamp = newFrameTimestamp;
if (codec) {
lastFrameVideoCpuTime = currentVideoCpuTime;
lastFrameAudioCpuTime = codec.audioCpuTime;
lastFrameDemuxerCpuTime = codec.demuxerCpuTime;
} else {
lastFrameVideoCpuTime = 0;
lastFrameAudioCpuTime = 0;
lastFrameDemuxerCpuTime = 0;
}
lastFrameDrawingTime = drawingTime;
lastFrameBufferTime = bufferTime;
lastFrameProxyTime = proxyTime;
function n(x) {
return Math.round(x * 10) / 10;
}
log('drew frame ' + data.frameEndTimestamp + ': ' +
'clock time ' + n(wallClockTime) + ' (jitter ' + n(jitter) + ') ' +
'cpu: ' + n(timing.cpuTime) + ' (mux: ' + n(timing.demuxerTime) + ' buf: ' + n(timing.bufferTime) + ' draw: ' + n(timing.drawingTime) + ' proxy: ' + n(timing.proxyTime) + ') ' +
'vid: ' + n(timing.videoTime) + ' aud: ' + n(timing.audioTime));
fireEventAsync('framecallback', timing);
if (!lastTimeUpdate || (newFrameTimestamp - lastTimeUpdate) >= timeUpdateInterval) {
lastTimeUpdate = newFrameTimestamp;
fireEventAsync('timeupdate');
}
}
// -- seek functions
var seekTargetTime = 0.0,
seekTargetKeypoint = 0.0,
bisectTargetTime = 0.0,
seekMode,
lastSeekPosition,
lastFrameSkipped,
seekBisector,
didSeek;
var SeekMode = {
EXACT: "exact",
FAST: "fast"
};
function Cancel(msg) {
Error.call(this, msg);
}
Cancel.prototype = Object.create(Error);
function seekStream(offset) {
if (stream.seeking) {
seekCancel.cancel(new Cancel('cancel for new seek'));
}
if (stream.buffering) {
readCancel.cancel(new Cancel('cancel for new seek'));
}
streamEnded = false;
dataEnded = false;
ended = false;
stream.seek(offset, seekCancel).then(function() {
readBytesAndWait();
}).catch(onStreamError);
}
function startBisection(targetTime) {
// leave room for potentially long Ogg page syncing
var endPoint = Math.max(0, stream.length - 65536);
bisectTargetTime = targetTime;
seekBisector = new Bisector({
start: 0,
end: endPoint,
process: function(start, end, position) {
if (position == lastSeekPosition) {
return false;
} else {
lastSeekPosition = position;
lastFrameSkipped = false;
codec.flush(function() {
seekStream(position);
});
return true;
}
}
});
seekBisector.start();
}
function seek(toTime, mode) {
log('requested seek to ' + toTime + ', mode ' + mode);
if (self.readyState == self.HAVE_NOTHING) {
log('not yet loaded; saving seek position for later');
initialSeekTime = toTime;
return;
}
if (stream && !stream.seekable) {
console.log(stream)
throw new Error('Cannot seek a non-seekable stream');
}
function prepForSeek(callback) {
if (stream && stream.buffering) {
readCancel.cancel(new Cancel('cancel for new seek'));
}
if (stream && stream.seeking) {
seekCancel.cancel(new Cancel('cancel for new seek'));
}
// clear any queued input/seek-start
actionQueue.splice(0, actionQueue.length);
stopPlayback();
prebufferingAudio = false;
stoppedForLateFrame = false;
if (audioFeeder) {
audioFeeder.flush();
}
state = State.SEEKING;
seekTargetTime = toTime;
seekMode = mode;
if (codec) {
codec.flush(callback);
} else {
callback();
}
}
// Abort any previous seek or play suitably
prepForSeek(function() {
if (isProcessing()) {
// already waiting on input
} else {
// start up the new load *after* event loop churn
pingProcessing(0);
}
});
actionQueue.push(function() {
// Just in case another async task stopped us...
prepForSeek(function() {
doSeek(toTime)
});
});
}
function doSeek(toTime) {
streamEnded = false;
dataEnded = false;
ended = false;
state = State.SEEKING;
seekTargetTime = toTime;
seekTargetKeypoint = -1;
lastFrameSkipped = false;
lastSeekPosition = -1;
decodedFrames = [];
pendingFrames = [];
pendingFrame = 0;
pendingAudio = 0;
didSeek = false;
codec.seekToKeypoint(toTime, function(seeking) {
if (seeking) {
seekState = SeekState.LINEAR_TO_TARGET;
fireEventAsync('seeking');
if (didSeek) {
// wait for i/o to trigger readBytesAndWait
return;
} else {
pingProcessing();
return;
}
}
// Use the old interface still implemented on ogg demuxer
codec.getKeypointOffset(toTime, function(offset) {
if (offset > 0) {
// This file has an index!
//
// Start at the keypoint, then decode forward to the desired time.
//
seekState = SeekState.LINEAR_TO_TARGET;
seekStream(offset);
} else {
// No index.
//
// Bisect through the file finding our target time, then we'll
// have to do it again to reach the keypoint, and *then* we'll
// have to decode forward back to the desired time.
//
seekState = SeekState.BISECT_TO_TARGET;
startBisection(seekTargetTime);
}
fireEventAsync('seeking');
});
});
}
function continueSeekedPlayback() {
seekState = SeekState.NOT_SEEKING;
state = State.READY;
frameEndTimestamp = codec.frameTimestamp;
audioEndTimestamp = codec.audioTimestamp;
if (codec.hasAudio) {
seekTargetTime = codec.audioTimestamp;
} else {
seekTargetTime = codec.frameTimestamp;
}
initialPlaybackOffset = seekTargetTime;
function finishedSeeking() {
lastTimeUpdate = seekTargetTime;
fireEventAsync('timeupdate');
fireEventAsync('seeked');
if (isProcessing()) {
// wait for whatever's going on to complete
} else {
pingProcessing();
}
}
// Decode and show first frame immediately
if (codec.hasVideo && codec.frameReady) {
// hack! move this into the main loop when retooling
// to avoid maintaining this double draw
codec.decodeFrame(function(ok) {
if (ok) {
if (thumbnail) {
self.removeChild(thumbnail);
thumbnail = null;
}
frameSink.drawFrame(codec.frameBuffer);
}
finishedSeeking();
} );
return;
} else {
finishedSeeking();
}
}
/**
* @return {boolean} true to continue processing, false to wait for input data
*/
function doProcessLinearSeeking() {
var frameDuration;
if (codec.hasVideo) {
frameDuration = targetPerFrameTime / 1000;
} else {
frameDuration = 1 / 256; // approximate packet audio size, fake!
}
if (codec.hasVideo) {
if (pendingFrame) {
// wait
return;
} else if (!codec.frameReady) {
// Haven't found a frame yet, process more data
codec.process(function(more) {
if (more) {
// need more packets
pingProcessing();
} else if (streamEnded) {
log('stream ended during linear seeking on video');
dataEnded = true;
continueSeekedPlayback();
} else {
readBytesAndWait();
}
});
return;
} else if (seekMode === SeekMode.FAST && codec.keyframeTimestamp == codec.frameTimestamp) {
// Found some frames? Go ahead now!
continueSeekedPlayback();
return;
} else if (codec.frameTimestamp + frameDuration < seekTargetTime) {
// Haven't found a time yet, or haven't reached the target time.
// Decode it in case we're at our keyframe or a following intraframe...
codec.decodeFrame(function() {
pingProcessing();
});
return;
} else {
// Reached or surpassed the target time.
if (codec.hasAudio) {
// Keep processing the audio track
// fall through...
} else {
continueSeekedPlayback();
return;
}
}
}
if (codec.hasAudio) {
if (pendingAudio) {
// wait
return;
} if (!codec.audioReady) {
// Haven't found an audio packet yet, process more data
codec.process(function(more) {
if (more) {
// need more packets
pingProcessing();
} else if (streamEnded) {
log('stream ended during linear seeking on audio');
dataEnded = true;
continueSeekedPlayback();
} else {
readBytesAndWait();
}
});
return;
} else if (codec.audioTimestamp + frameDuration < seekTargetTime) {
// Haven't found a time yet, or haven't reached the target time.
// Decode it so when we reach the target we've got consistent data.
codec.decodeAudio(function() {
pingProcessing();
});
return;
} else {
continueSeekedPlayback();
return;
}
}
}
function doProcessBisectionSeek() {
var frameDuration,
timestamp;
if (codec.hasVideo) {
timestamp = codec.frameTimestamp;
frameDuration = targetPerFrameTime / 1000;
} else if (codec.hasAudio) {
timestamp = codec.audioTimestamp;
frameDuration = 1 / 256; // approximate packet audio size, fake!
} else {
throw new Error('Invalid seek state; no audio or video track available');
}
if (timestamp < 0) {
// Haven't found a packet yet.
codec.process(function(more) {
if (more) {
// need more data
pingProcessing();
} else if (streamEnded) {
log('stream ended during bisection seek');
// We may have to go back further to find packets.
if (seekBisector.right()) {
// wait for new data to come in
} else {
log('failed going back');
throw new Error('not sure what to do');
}
} else {
readBytesAndWait();
}
});
} else if (timestamp - frameDuration / 2 > bisectTargetTime) {
if (seekBisector.left()) {
// wait for new data to come in
} else {
log('close enough (left)');
// We're having trouble finding a new packet position;
// likely it's an audio file with lots of small packets.
// Since we can't find an earlier packet, just continue here.
seekTargetTime = timestamp;
continueSeekedPlayback();
}
} else if (timestamp + frameDuration / 2 < bisectTargetTime) {
if (seekBisector.right()) {
// wait for new data to come in
} else {
log('close enough (right)');
// We're having trouble finding a new packet position;
// likely it's an audio file with lots of small packets.
// Switch to linear mode to find the final target.
seekState = SeekState.LINEAR_TO_TARGET;
pingProcessing();
}
} else {
// Reached the bisection target!
if (seekState == SeekState.BISECT_TO_TARGET && (codec.hasVideo && codec.keyframeTimestamp < codec.frameTimestamp)) {
// We have to go back and find a keyframe. Sigh.
log('finding the keypoint now');
seekState = SeekState.BISECT_TO_KEYPOINT;
startBisection(codec.keyframeTimestamp);
} else {
log('straight seeking now');
// Switch to linear mode to find the final target.
seekState = SeekState.LINEAR_TO_TARGET;
pingProcessing();
}
}
}
function setupVideo() {
if (videoInfo.fps > 0) {
targetPerFrameTime = 1000 / videoInfo.fps;
} else {
targetPerFrameTime = 16.667; // recalc this later
}
canvas.width = videoInfo.displayWidth;
canvas.height = videoInfo.displayHeight;
OGVPlayer.styleManager.appendRule('.' + instanceId, {
width: videoInfo.displayWidth + 'px',
height: videoInfo.displayHeight + 'px'
});
OGVPlayer.updatePositionOnResize();
frameSink = YUVCanvas.attach(canvas, canvasOptions);
}
var depth = 0,
needProcessing = false,
pendingFrame = 0,
pendingAudio = 0,
framePipelineDepth = 3,
audioPipelineDepth = 4;
function doProcessing() {
if (didSeek) {
didSeek = false;
}
nextProcessingTimer = null;
if (isProcessing()) {
// Called async while waiting for something else to complete...
// let it finish, then we'll get called again to continue.
//return;
//throw new Error('REENTRANCY FAIL: doProcessing during processing');
}
if (depth > 0) {
throw new Error('REENTRANCY FAIL: doProcessing recursing unexpectedly');
}
var iters = 0;
do {
needProcessing = false;
depth++;
doProcessingLoop();
depth--;
if (needProcessing && isProcessing()) {
throw new Error('REENTRANCY FAIL: waiting on input or codec but asked to keep processing');
}
if (++iters > 500) {
log('stuck in processing loop; breaking with timer');
needProcessing = 0;
pingProcessing(0);
}
} while (needProcessing);
}
function doProcessingLoop() {
if (actionQueue.length) {
// data or user i/o to process in our serialized event stream
// The function should eventually bring us back here via pingProcessing(),
// directly or via further i/o.
var action = actionQueue.shift();
action();
} else if (state == State.INITIAL) {
if (codec.loadedMetadata) {
// we just fell over from headers into content; call onloadedmetadata etc
if (!codec.hasVideo && !codec.hasAudio) {
throw new Error('No audio or video found, something is wrong');
}
if (codec.hasAudio) {
audioInfo = codec.audioFormat;
}
if (codec.hasVideo) {
videoInfo = codec.videoFormat;
setupVideo();
}
if (!isNaN(codec.duration)) {
duration = codec.duration;
}
if (duration === null) {
if (stream.seekable) {
state = State.SEEKING_END;
lastSeenTimestamp = -1;
codec.flush(function() {
seekStream(Math.max(0, stream.length - 65536 * 2));
});
} else {
// Stream not seekable and no x-content-duration; assuming infinite stream.
state = State.LOADED;
pingProcessing();
}
} else {
// We already know the duration.
state = State.LOADED;
pingProcessing();
}
} else {
codec.process(function processInitial(more) {
if (more) {
// Keep processing headers
pingProcessing();
} else if (streamEnded) {
throw new Error('end of file before headers found');
} else {
// Read more data!
log('reading more cause we are out of data');
readBytesAndWait();
}
});
}
} else if (state == State.SEEKING_END) {
// Look for the last item.
if (codec.frameReady) {
log('saw frame with ' + codec.frameTimestamp);
lastSeenTimestamp = Math.max(lastSeenTimestamp, codec.frameTimestamp);
codec.discardFrame(function() {
pingProcessing();
});
} else if (codec.audioReady) {
log('saw audio with ' + codec.audioTimestamp);
lastSeenTimestamp = Math.max(lastSeenTimestamp, codec.audioTimestamp);
codec.discardAudio(function() {
pingProcessing();
});
} else {
codec.process(function processSeekingEnd(more) {
if (more) {
// Keep processing headers
pingProcessing();
} else {
// Read more data!
if (!stream.eof) {
readBytesAndWait();
} else {
// We are at the end!
log('seek-duration: we are at the end: ' + lastSeenTimestamp);
if (lastSeenTimestamp > 0) {
duration = lastSeenTimestamp;
}
// Ok, seek back to the beginning and resync the streams.
state = State.LOADED;
codec.flush(function() {
streamEnded = false;
dataEnded = false;
seekStream(0);
});
}
}
});
}
} else if (state == State.LOADED) {
state = State.PRELOAD;
fireEventAsync('loadedmetadata');
fireEventAsync('durationchange');
if (codec.hasVideo) {
fireEventAsync('resize');
}
pingProcessing(0);
} else if (state == State.PRELOAD) {
if ((codec.frameReady || !codec.hasVideo) &&
(codec.audioReady || !codec.hasAudio)) {
state = State.READY;
fireEventAsync('loadeddata');
pingProcessing();
} else {
codec.process(function doProcessPreload(more) {
if (more) {
pingProcessing();
} else if (streamEnded) {
// Ran out of data before data available...?
ended = true;
} else {
readBytesAndWait();
}
});
}
} else if (state == State.READY) {
log('initial seek to ' + initialSeekTime);
if (initialSeekTime > 0) {
var target = initialSeekTime;
initialSeekTime = 0;
log('initial seek to ' + target);
doSeek(target);
} else if (paused) {
// Paused? stop here.
log('paused while in ready');
} else {
function finishStartPlaying() {
log('finishStartPlaying');
state = State.PLAYING;
lastFrameTimestamp = getTimestamp();
if (codec.hasAudio && audioFeeder) {
// Pre-queue audio before we start the clock
prebufferingAudio = true;
} else {
startPlayback();
}
pingProcessing(0);
fireEventAsync('play');
fireEventAsync('playing');
}
if (codec.hasAudio && !audioFeeder && !muted) {
initAudioFeeder();
audioFeeder.waitUntilReady(finishStartPlaying);
} else {
finishStartPlaying();
}
}
} else if (state == State.SEEKING) {
if (seekState == SeekState.NOT_SEEKING) {
throw new Error('seeking in invalid state (not seeking?)');
} else if (seekState == SeekState.BISECT_TO_TARGET) {
doProcessBisectionSeek();
} else if (seekState == SeekState.BISECT_TO_KEYPOINT) {
doProcessBisectionSeek();
} else if (seekState == SeekState.LINEAR_TO_TARGET) {
doProcessLinearSeeking();
} else {
throw new Error('Invalid seek state ' + seekState);
}
} else if (state == State.PLAYING) {
function doProcessPlay() {
//console.log(more, codec.audioReady, codec.frameReady, codec.audioTimestamp, codec.frameTimestamp);
if (paused) {
// ok we're done for now!
log('paused during playback; stopping loop');
} else {
if ((!codec.hasAudio || codec.audioReady || pendingAudio || dataEnded) &&
(!codec.hasVideo || codec.frameReady || pendingFrame || decodedFrames.length || dataEnded)
) {
var audioState = null,
playbackPosition = 0,
readyForAudioDecode,
audioEnded = false,
readyForFrameDraw,
frameDelay = 0,
readyForFrameDecode,
// timers never fire earlier than 4ms
timerMinimum = 4,
// ok to draw a couple ms early
fudgeDelta = timerMinimum;
if (codec.hasAudio && audioFeeder) {
// Drive on the audio clock!
audioState = audioFeeder.getPlaybackState();
playbackPosition = getPlaybackTime(audioState);
audioEnded = (dataEnded && audioFeeder.durationBuffered == 0);
if (prebufferingAudio && (audioFeeder.durationBuffered >= audioFeeder.bufferThreshold * 2 || dataEnded)) {
log('prebuffering audio done; buffered to ' + audioFeeder.durationBuffered);
startPlayback(playbackPosition);
prebufferingAudio = false;
}
if (audioState.dropped != droppedAudio) {
log('dropped ' + (audioState.dropped - droppedAudio));
}
if (audioState.delayed != delayedAudio) {
log('delayed ' + (audioState.delayed - delayedAudio));
}
droppedAudio = audioState.dropped;
delayedAudio = audioState.delayed;
readyForAudioDecode = audioFeeder.durationBuffered <=
audioFeeder.bufferThreshold * 2;
if (!readyForAudioDecode) {
// just to skip the remaining items in debug log
} else if (!codec.audioReady) {
// NEED MOAR BUFFERS
readyForAudioDecode = false;
} else if (pendingAudio >= audioPipelineDepth) {
// We'll check in when done decoding
log('audio decode disabled: ' + pendingAudio + ' packets in flight');
readyForAudioDecode = false;
}
} else {
// No audio; drive on the general clock.
// @fixme account for dropped frame times...
playbackPosition = getPlaybackTime();
// If playing muted with no audio output device,
// just keep up with audio in general.
readyForAudioDecode = codec.audioReady && (audioEndTimestamp < playbackPosition);
}
if (codec.hasVideo) {
readyForFrameDraw = decodedFrames.length > 0;
readyForFrameDecode = (pendingFrame + decodedFrames.length < framePipelineDepth) && codec.frameReady;
if (readyForFrameDraw) {
frameDelay = (decodedFrames[0].frameEndTimestamp - playbackPosition) * 1000;
actualPerFrameTime = targetPerFrameTime - frameDelay;
//frameDelay = Math.max(0, frameDelay);
//frameDelay = Math.min(frameDelay, targetPerFrameTime);
}
var audioSyncThreshold = targetPerFrameTime;
if (prebufferingAudio) {
if (readyForFrameDecode) {
log('decoding a frame during prebuffering');
}
readyForFrameDraw = false;
} else if (readyForFrameDraw && dataEnded && audioEnded) {
// If audio timeline reached end, make sure the last frame draws
log('audio timeline ended? ready to draw frame');
} else if (readyForFrameDraw && -frameDelay >= audioSyncThreshold) {
// late frame!
if (options.sync == 'delay-audio') {
// Delay audio while video catches up. Old default.
if (!stoppedForLateFrame) {
log('late frame at ' + playbackPosition + ': ' + (-frameDelay) + ' expected ' + audioSyncThreshold);
lateFrames++;
if (decodedFrames.length > 1) {
log('late frame has a neighbor; skipping to next frame');
var frame = decodedFrames.shift();
frameDelay = (frame.frameEndTimestamp - playbackPosition) * 1000;
actualPerFrameTime = targetPerFrameTime - frameDelay;
framesProcessed++; // pretend!
doFrameComplete({
frameEndTimestamp: frame.frameEndTimestamp,
dropped: true
});
} else {
stopPlayback();
stoppedForLateFrame = true;
}
}
} else if (options.sync == 'skip-frames') {
var skipPast = -1;
for (var i = 0; i < decodedFrames.length - 1; i++) {
if (decodedFrames[i].frameEndTimestamp < playbackPosition) {
skipPast = i - 1;
}
}
if (skipPast >= 0) {
while (skipPast-- >= 0) {
lateFrames++;
var frame = decodedFrames.shift();
log('skipping already-decoded late frame at ' + frame.frameEndTimestamp);
frameDelay = (frame.frameEndTimestamp - playbackPosition) * 1000;
frameEndTimestamp = frame.frameEndTimestamp;
actualPerFrameTime = targetPerFrameTime - frameDelay;
framesProcessed++; // pretend!
frame.dropped = true;
doFrameComplete();
}
}
// Resync at the next keyframe. Default as of 1.3.2.
var nextKeyframe = codec.nextKeyframeTimestamp;
// When resyncing, allow time to decode a couple frames!
var videoSyncPadding = (targetPerFrameTime / 1000) * (framePipelineDepth + pendingFrame);
var timeToResync = nextKeyframe - videoSyncPadding;
if (nextKeyframe >= 0 && nextKeyframe != codec.frameTimestamp && playbackPosition >= timeToResync) {
log('skipping late frame at ' + decodedFrames[0].frameEndTimestamp + ' vs ' + playbackPosition + ', expect to see keyframe at ' + nextKeyframe);
// First skip any already-decoded frames
for (var i = 0; i < decodedFrames.length; i++) {
var frame = decodedFrames[i];
lateFrames++;
framesProcessed++; // pretend!
frameEndTimestamp = frame.frameEndTimestamp;
frameDelay = (frame.frameEndTimestamp - playbackPosition) * 1000;
actualPerFrameTime = targetPerFrameTime - frameDelay;
frame.dropped = true;
doFrameComplete(frame);
}
decodedFrames = [];
for (var i = 0; i < pendingFrames.length; i++) {
var frame = pendingFrames[i];
lateFrames++;
framesProcessed++; // pretend!
frameEndTimestamp = frame.frameEndTimestamp;
frameDelay = (frame.frameEndTimestamp - playbackPosition) * 1000;
actualPerFrameTime = targetPerFrameTime - frameDelay;
frame.dropped = true;
doFrameComplete(frame);
}
pendingFrames = [];
pendingFrame = 0;
// Now discard anything up to the keyframe
while (codec.frameReady && codec.frameTimestamp < nextKeyframe) {
// note: this is a known synchronous operation :)
var frame = {
frameEndTimestamp: codec.frameTimestamp,
dropped: true
};
frameDelay = (frame.frameEndTimestamp - playbackPosition) * 1000;
actualPerFrameTime = targetPerFrameTime - frameDelay;
lateFrames++;
codec.discardFrame(function() {/*fake*/});
framesProcessed++; // pretend!
doFrameComplete(frame);
}
if (isProcessing()) {
// wait
} else {
pingProcessing();
}
return;
}
}
} else if (readyForFrameDraw && stoppedForLateFrame && !readyForFrameDecode && !readyForAudioDecode && frameDelay > fudgeDelta) {
// catching up, ok if we were early
log('late frame recovery reached ' + frameDelay);
startPlayback(playbackPosition);
stoppedForLateFrame = false;
readyForFrameDraw = false; // go back through the loop again
} else if (readyForFrameDraw && stoppedForLateFrame) {
// draw asap
} else if (readyForFrameDraw && frameDelay <= fudgeDelta) {
// on time! draw
} else {
// not yet
readyForFrameDraw = false;
}
}
if (readyForFrameDecode) {
log('play loop: ready to decode frame; thread depth: ' + pendingFrame + ', have buffered: ' + decodedFrames.length);
if (videoInfo.fps == 0 && (codec.frameTimestamp - frameEndTimestamp) > 0) {
// WebM doesn't encode a frame rate
targetPerFrameTime = (codec.frameTimestamp - frameEndTimestamp) * 1000;
}
totalFrameTime += targetPerFrameTime;
totalFrameCount++;
var nextFrameEndTimestamp = frameEndTimestamp = codec.frameTimestamp;
pendingFrame++;
pendingFrames.push({
frameEndTimestamp: nextFrameEndTimestamp
});
var currentPendingFrames = pendingFrames;
var wasAsync = false;
var frameDecodeTime = time(function() {
codec.decodeFrame(function processingDecodeFrame(ok) {
if (currentPendingFrames !== pendingFrames) {
log('play loop callback after flush, discarding');
return;
}
log('play loop callback: decoded frame');
pendingFrame--;
pendingFrames.shift();
if (ok) {
// Save the buffer until it's time to draw
decodedFrames.push({
yCbCrBuffer: codec.frameBuffer,
videoCpuTime: codec.videoCpuTime,
frameEndTimestamp: nextFrameEndTimestamp
});
} else {
// Bad packet or something.
log('Bad video packet or something');
}
codec.process(function() {
if (isProcessing()) {
// wait
} else {
pingProcessing(wasAsync ? undefined : 0);
}
});
});
});
if (pendingFrame) {
wasAsync = true;
proxyTime += frameDecodeTime;
// We can process something else while that's running
pingProcessing();
}
} else if (readyForAudioDecode) {
log('play loop: ready for audio; depth: ' + pendingAudio);
pendingAudio++;
var nextAudioEndTimestamp = codec.audioTimestamp;
var audioDecodeTime = time(function() {
codec.decodeAudio(function processingDecodeAudio(ok) {
pendingAudio--;
log('play loop callback: decoded audio');
audioEndTimestamp = nextAudioEndTimestamp;
if (ok) {
var buffer = codec.audioBuffer;
if (buffer) {
// Keep track of how much time we spend queueing audio as well
// This is slow when using the Flash shim on IE 10/11
bufferTime += time(function() {
if (audioFeeder) {
audioFeeder.bufferData(buffer);
}
});
if (!codec.hasVideo) {
framesProcessed++; // pretend!
var frame = {
frameEndTimestamp: audioEndTimestamp
};
doFrameComplete(frame);
}
}
}
if (isProcessing()) {
// wait
} else {
pingProcessing();
}
});
});
if (pendingAudio) {
proxyTime += audioDecodeTime;
// We can process something else while that's running
if (codec.audioReady) {
pingProcessing();
} else {
// Trigger a demux immediately if we need it;
// audio processing is our mostly-idle time
doProcessPlayDemux();
}
}
} else if (readyForFrameDraw) {
log('play loop: ready to draw frame');
if (nextFrameTimer) {
clearTimeout(nextFrameTimer);
nextFrameTimer = null;
}
// Ready to draw the decoded frame...
if (thumbnail) {
self.removeChild(thumbnail);
thumbnail = null;
}
var frame = decodedFrames.shift();
currentVideoCpuTime = frame.videoCpuTime;
drawingTime += time(function() {
frameSink.drawFrame(frame.yCbCrBuffer);
});
framesProcessed++;
framesPlayed++;
doFrameComplete(frame);
pingProcessing();
} else if (decodedFrames.length && !nextFrameTimer && !prebufferingAudio) {
var targetTimer = frameDelay;
// @todo consider using requestAnimationFrame
log('play loop: setting a timer for drawing ' + targetTimer);
nextFrameTimer = setTimeout(function() {
nextFrameTimer = null;
pingProcessing();
}, targetTimer);
} else if (dataEnded && !(pendingAudio || pendingFrame || decodedFrames.length)) {
log('play loop: playback reached end of data ' + [pendingAudio, pendingFrame, decodedFrames.length]);
var finalDelay = 0;
if (codec.hasAudio && audioFeeder) {
finalDelay = audioFeeder.durationBuffered * 1000;
}
if (finalDelay > 0) {
log('play loop: ending pending ' + finalDelay + ' ms');
pingProcessing(Math.max(0, finalDelay));
} else {
log('play loop: ENDING NOW: playback time ' + getPlaybackTime() + '; frameEndTimestamp: ' + frameEndTimestamp);
stopPlayback();
stoppedForLateFrame = false;
prebufferingAudio = false;
initialPlaybackOffset = Math.max(audioEndTimestamp, frameEndTimestamp);
ended = true;
// @todo implement loop behavior
paused = true;
fireEventAsync('pause');
fireEventAsync('ended');
}
} else {
log('play loop: waiting on async/timers');
}
} else {
log('play loop: demuxing');
doProcessPlayDemux();
}
}
}
function doProcessPlayDemux() {
var wasFrameReady = codec.frameReady,
wasAudioReady = codec.audioReady;
codec.process(function doProcessPlayDemuxHandler(more) {
if ((codec.frameReady && !wasFrameReady) || (codec.audioReady && !wasAudioReady)) {
log('demuxer has packets');
pingProcessing();
} else if (more) {
// Have to process some more pages to find data.
log('demuxer processing to find more packets');
pingProcessing();
} else {
log('demuxer ran out of data');
if (!streamEnded) {
// Ran out of buffered input
log('demuxer loading more data');
readBytesAndWait();
} else {
// Ran out of stream!
log('demuxer reached end of data stream');
dataEnded = true;
pingProcessing();
}
}
});
}
doProcessPlay();
} else if (state == State.ERROR) {
// Nothing to do.
console.log("Reached error state. Sorry bout that.");
} else {
throw new Error('Unexpected OGVPlayer state ' + state);
}
}
/**
* Are we waiting on an async operation that will return later?
*/
function isProcessing() {
return (stream && (stream.buffering || stream.seeking)) || (codec && codec.processing);
}
function readBytesAndWait() {
if (stream.buffering || stream.seeking) {
log('readBytesAndWait during i/o');
return;
}
// keep i/o size small to reduce CPU impact of demuxing on slow machines
// @todo make buffer size larger when packets are larger?
var bufferSize = 32768;
stream.read(bufferSize, readCancel).then(function(data) {
log('got input ' + [data.byteLength]);
if (data.byteLength) {
// Save chunk to pass into the codec's buffer
actionQueue.push(function doReceiveInput() {
codec.receiveInput(data, function() {
pingProcessing();
});
});
}
if (stream.eof) {
// @todo record doneness in networkState
log('stream is at end!');
streamEnded = true;
}
if (isProcessing()) {
// We're waiting on the codec already...
} else {
// Let the read/decode/draw loop know we're out!
pingProcessing();
}
}).catch(onStreamError);
}
function onStreamError(err) {
if (err instanceof Cancel) {
// do nothing
log('i/o promise canceled; ignoring');
} else {
log("i/o error: " + err);
mediaError = new OGVMediaError(OGVMediaError.MEDIA_ERR_NETWORK, '' + err);
state = State.ERROR;
stopPlayback();
}
}
function pingProcessing(delay) {
if (delay === undefined) {
delay = -1;
}
/*
if (isProcessing()) {
// We'll get pinged again when whatever we were doing returns...
log('REENTRANCY FAIL: asked to pingProcessing() while already waiting');
return;
}
*/
if (stream && stream.waiting) {
// wait for this input pls
log('waiting on input');
return;
}
if (nextProcessingTimer) {
log('canceling old processing timer');
clearTimeout(nextProcessingTimer);
nextProcessingTimer = null;
}
var fudge = -1 / 256;
if (delay > fudge) {
//log('pingProcessing delay: ' + delay);
nextProcessingTimer = setTimeout(function() {
// run through pingProcessing again to check
// in case some io started asynchronously in the meantime
pingProcessing();
}, delay);
} else if (depth) {
//log('pingProcessing tail call (' + delay + ')');
needProcessing = true;
} else {
doProcessing();
}
}
var videoInfo,
audioInfo;
function startProcessingVideo() {
if (started || codec) {
return;
}
framesProcessed = 0;
bufferTime = 0;
drawingTime = 0;
proxyTime = 0;
started = true;
ended = false;
codec = new codecClass(codecOptions);
lastVideoCpuTime = 0;
lastAudioCpuTime = 0;
lastDemuxerCpuTime = 0;
lastBufferTime = 0;
lastDrawingTime = 0;
lastProxyTime = 0;
lastFrameVideoCpuTime = 0;
lastFrameAudioCpuTime = 0;
lastFrameDemuxerCpuTime = 0;
lastFrameBufferTime = 0;
lastFrameProxyTime = 0;
lastFrameDrawingTime = 0;
currentVideoCpuTime = 0;
codec.onseek = function(offset) {
didSeek = true;
if (stream) {
seekStream(offset);
}
};
codec.init(function() {
readBytesAndWait();
});
}
function loadCodec(callback) {
// @todo use the demuxer and codec interfaces directly
// @todo fix detection proper
if (enableWebM && currentSrc.match(/\.webm$/i)) {
codecOptions.type = 'video/webm';
} else {
codecOptions.type = 'video/ogg';
}
codecClass = OGVWrapperCodec;
callback();
}
/**
* HTMLMediaElement load method
*
* https://www.w3.org/TR/html5/embedded-content-0.html#concept-media-load-algorithm
*/
self.load = function() {
prepForLoad();
};
function prepForLoad(preload) {
stopVideo();
function doLoad() {
// @todo networkState == NETWORK_LOADING
stream = new StreamFile({
url: self.src,
cacheSize: 16 * 1024 * 1024,
});
stream.load().then(function() {
loading = false;
// @todo handle failure / unrecognized type
currentSrc = self.src;
// Fire off the read/decode/draw loop...
byteLength = stream.seekable ? stream.length : 0;
// If we get X-Content-Duration, that's as good as an explicit hint
var durationHeader = stream.headers['x-content-duration'];
if (typeof durationHeader === 'string') {
duration = parseFloat(durationHeader);
}
loadCodec(startProcessingVideo);
}).catch(onStreamError);
}
// @todo networkState = self.NETWORK_NO_SOURCE;
// @todo show poster
// @todo set 'delay load event flag'
currentSrc = '';
loading = true;
actionQueue.push(function() {
if (preload && self.preload === 'none') {
// Done for now, we'll pick up if someone hits play() or load()
loading = false;
} else {
doLoad();
}
});
pingProcessing(0);
}
/**
* HTMLMediaElement canPlayType method
*/
self.canPlayType = function(contentType) {
var type = new OGVMediaType(contentType);
if (type.minor === 'ogg' &&
(type.major === 'audio' || type.major === 'video' || type.major === 'application')) {
if (type.codecs) {
var supported = ['vorbis', 'opus', 'theora'],
knownCodecs = 0,
unknownCodecs = 0;
type.codecs.forEach(function(codec) {
if (supported.indexOf(codec) >= 0) {
knownCodecs++;
} else {
unknownCodecs++;
}
});
if (knownCodecs === 0) {
return '';
} else if (unknownCodecs > 0) {
return '';
}
// All listed codecs are ones we know. Neat!
return 'probably';
} else {
return 'maybe';
}
} else {
// @todo when webm support is more complete, handle it
return '';
}
};
/**
* HTMLMediaElement play method
*/
self.play = function() {
if (!muted && !audioOptions.audioContext) {
OGVPlayer.initSharedAudioContext();
}
if (paused) {
startedPlaybackInDocument = document.body.contains(self);
paused = false;
if (state == State.SEEKING) {
// Seeking? Just make sure we know to pick up after.
} else if (started && codec && codec.loadedMetadata) {
if (ended && stream && byteLength) {
log('.play() starting over after end');
seek(0);
} else {
log('.play() while already started');
}
state = State.READY;
if (isProcessing()) {
// waiting on the codec already
} else {
pingProcessing();
}
} else if (loading) {
log('.play() while loading');
} else {
log('.play() before started');
// Let playback begin when metadata loading is complete
if (!stream) {
self.load();
}
}
}
};
/**
* custom getPlaybackStats method
*/
self.getPlaybackStats = function() {
return {
targetPerFrameTime: targetPerFrameTime,
framesProcessed: framesProcessed,
playTime: playTime,
demuxingTime: codec ? (codec.demuxerCpuTime - lastDemuxerCpuTime) : 0,
videoDecodingTime: codec ? (codec.videoCpuTime - lastVideoCpuTime) : 0,
audioDecodingTime: codec ? (codec.audioCpuTime - lastAudioCpuTime) : 0,
bufferTime: bufferTime - lastBufferTime,
drawingTime: drawingTime - lastDrawingTime,
proxyTime: proxyTime - lastProxyTime,
droppedAudio: droppedAudio,
delayedAudio: delayedAudio,
jitter: totalJitter / framesProcessed,
lateFrames: lateFrames
};
};
self.resetPlaybackStats = function() {
framesProcessed = 0;
playTime = 0;
if (codec) {
lastDemuxerCpuTime = codec.demuxerCpuTime;
lastVideoCpuTime = codec.videoCpuTime;
lastAudioCpuTime = codec.audioCpuTime;
}
lastBufferTime = bufferTime;
lastDrawingTime = drawingTime;
lastProxyTime = proxyTime;
totalJitter = 0;
totalFrameTime = 0;
totalFrameCount = 0;
};
self.getVideoFrameSink = function() {
return frameSink;
};
self.getCanvas = function() {
return canvas;
};
/**
* HTMLMediaElement pause method
*/
self.pause = function() {
if (!paused) {
clearTimeout(nextProcessingTimer);
nextProcessingTimer = null;
stopPlayback();
prebufferingAudio = false;
stoppedForLateFrame = false;
paused = true;
fireEvent('pause');
}
};
/**
* custom 'stop' method
*/
self.stop = function() {
stopVideo();
paused = true;
};
/**
* @todo implement the fast part of the behavior!
*/
self.fastSeek = function(seekToTime) {
seek(+seekToTime, SeekMode.FAST);
};
/**
* HTMLMediaElement src property
*/
Object.defineProperty(self, "src", {
get: function getSrc() {
return self.getAttribute('src') || '';
},
set: function setSrc(val) {
self.setAttribute('src', val);
loading = false; // just in case?
prepForLoad("interactive");
}
});
/**
* HTMLMediaElement buffered property
*/
Object.defineProperty(self, "buffered", {
get: function getBuffered() {
var ranges;
if (stream && byteLength && duration) {
ranges = stream.getBufferedRanges().map(function(range) {
return range.map(function(offset) {
return (offset / stream.length) * duration;
});
});
} else {
ranges = [[0, 0]];
}
return new OGVTimeRanges(ranges);
}
});
/**
* HTMLMediaElement seekable property
*/
Object.defineProperty(self, "seekable", {
get: function getSeekable() {
if (self.duration < Infinity && stream && stream.seekable && codec && codec.seekable) {
return new OGVTimeRanges([[0, duration]]);
} else {
return new OGVTimeRanges([]);
}
}
});
/**
* HTMLMediaElement currentTime property
*/
Object.defineProperty(self, "currentTime", {
get: function getCurrentTime() {
if (state == State.SEEKING) {
return seekTargetTime;
} else {
if (codec) {
if (state == State.PLAYING && !paused) {
return getPlaybackTime();
} else {
return initialPlaybackOffset;
}
} else {
return initialSeekTime;
}
}
},
set: function setCurrentTime(val) {
seek(val, SeekMode.EXACT);
}
});
/**
* HTMLMediaElement duration property
*/
Object.defineProperty(self, "duration", {
get: function getDuration() {
if (codec && codec.loadedMetadata) {
if (duration !== null) {
return duration;
} else {
return Infinity;
}
} else {
return NaN;
}
}
});
/**
* HTMLMediaElement paused property
*/
Object.defineProperty(self, "paused", {
get: function getPaused() {
return paused;
}
});
/**
* HTMLMediaElement ended property
*/
Object.defineProperty(self, "ended", {
get: function getEnded() {
return ended;
}
});
/**
* HTMLMediaElement ended property
*/
Object.defineProperty(self, "seeking", {
get: function getSeeking() {
return (state == State.SEEKING);
}
});
/**
* HTMLMediaElement muted property
*/
Object.defineProperty(self, "muted", {
get: function getMuted() {
return muted;
},
set: function setMuted(val) {
muted = val;
if (audioFeeder) {
audioFeeder.muted = muted;
} else if (started && !muted && codec && codec.hasAudio) {
log('unmuting: switching from timer to audio clock');
initAudioFeeder();
startPlayback(audioEndTimestamp);
}
fireEventAsync('volumechange');
}
});
Object.defineProperty(self, "poster", {
get: function getPoster() {
return poster;
},
set: function setPoster(val) {
poster = val;
if (!started) {
if (thumbnail) {
self.removeChild(thumbnail);
}
thumbnail = new Image();
thumbnail.src = poster;
thumbnail.className = 'ogvjs-poster';
thumbnail.style.position = 'absolute';
thumbnail.style.top = '0';
thumbnail.style.left = '0';
thumbnail.style.width = '100%';
thumbnail.style.height = '100%';
thumbnail.style.objectFit = 'contain';
thumbnail.style.visibility = 'hidden';
thumbnail.addEventListener('load', function() {
if (thumbnail === this) {
OGVPlayer.styleManager.appendRule('.' + instanceId, {
width: thumbnail.naturalWidth + 'px',
height: thumbnail.naturalHeight + 'px'
});
OGVPlayer.updatePositionOnResize();
thumbnail.style.visibility = 'visible';
}
});
self.appendChild(thumbnail);
}
}
});
// Video metadata properties...
Object.defineProperty(self, "videoWidth", {
get: function getVideoWidth() {
if (videoInfo) {
return videoInfo.displayWidth;
} else {
return 0;
}
}
});
Object.defineProperty(self, "videoHeight", {
get: function getVideoHeight() {
if (videoInfo) {
return videoInfo.displayHeight;
} else {
return 0;
}
}
});
Object.defineProperty(self, "ogvjsVideoFrameRate", {
get: function getOgvJsVideoFrameRate() {
if (videoInfo) {
if (videoInfo.fps == 0) {
return totalFrameCount / (totalFrameTime / 1000);
} else {
return videoInfo.fps;
}
} else {
return 0;
}
}
});
// Audio metadata properties...
Object.defineProperty(self, "ogvjsAudioChannels", {
get: function getOgvJsAudioChannels() {
if (audioInfo) {
return audioInfo.channels;
} else {
return 0;
}
}
});
Object.defineProperty(self, "ogvjsAudioSampleRate", {
get: function getOgvJsAudioChannels() {
if (audioInfo) {
return audioInfo.rate;
} else {
return 0;
}
}
});
// Display size...
var width = 0, height = 0;
/**
* @property width
* @todo reflect to the width attribute?
*/
Object.defineProperty(self, "width", {
get: function getWidth() {
return width;
},
set: function setWidth(val) {
width = parseInt(val, 10);
self.style.width = width + 'px';
OGVPlayer.updatePositionOnResize();
}
});
/**
* @property height
* @todo reflect to the height attribute?
*/
Object.defineProperty(self, "height", {
get: function getHeight() {
return height;
},
set: function setHeight(val) {
height = parseInt(val, 10);
self.style.height = height + 'px';
OGVPlayer.updatePositionOnResize();
}
});
/**
* @property autoplay {boolean} stub prop
* @todo reflect to the autoplay attribute?
* @todo implement actual autoplay behavior
*/
Object.defineProperty(self, "autoplay", {
get: function getAutoplay() {
return false;
},
set: function setAutoplay(val) {
// ignore
}
});
/**
* @property controls {boolean} stub prop
* @todo reflect to the controls attribute?
* @todo implement actual control behavior
*/
Object.defineProperty(self, "controls", {
get: function getControls() {
return false;
},
set: function setControls(val) {
// ignore
}
});
/**
* @property loop {boolean} stub prop
* @todo reflect to the controls attribute?
* @todo implement actual loop behavior
*/
Object.defineProperty(self, "loop", {
get: function getLoop() {
return false;
},
set: function setLoop(val) {
// ignore
}
});
/**
* @property crossOrigin {string|null} stub prop
* @todo reflect to the crossorigin attribute?
* @todo implement actual behavior
*/
Object.defineProperty(self, "crossOrigin", {
get: function getCrossOrigin() {
return null;
},
set: function setCrossOrigin(val) {
// ignore
}
});
/**
* Returns the URL to the currently-playing resource.
* @property currentSrc {string|null}
*/
Object.defineProperty(self, "currentSrc", {
get: function getCurrentSrc() {
// @todo return absolute URL per spec
return currentSrc;
}
});
Object.defineProperty(self, "defaultMuted", {
get: function getDefaultMuted() {
return false;
}
});
Object.defineProperty(self, "defaultPlaybackRate", {
get: function getDefaultPlaybackRate() {
return 1;
}
});
/**
* @property error {string|null}
* @todo implement
*/
Object.defineProperty(self, "error", {
get: function getError() {
if (state === State.ERROR) {
if (mediaError) {
return mediaError;
} else {
return new OGVMediaError("unknown error occurred in media procesing");
}
} else {
return null;
}
}
});
/**
* @property preload {string}
*/
Object.defineProperty(self, "preload", {
get: function getPreload() {
return self.getAttribute('preload') || '';
},
set: function setPreload(val) {
self.setAttribute('preload', val);
}
});
/**
* @property readyState {number}
* @todo return more accurate info about availability of data
*/
Object.defineProperty(self, "readyState", {
get: function getReadyState() {
if (stream && codec && codec.loadedMetadata) {
// for now we don't really calc this stuff
// just pretend we have lots of data coming in already
return OGVPlayer.HAVE_ENOUGH_DATA;
} else {
return OGVPlayer.HAVE_NOTHING;
}
}
});
/**
* @property networkState {number}
* @todo implement
*/
Object.defineProperty(self, "networkState", {
get: function getNetworkState() {
if (stream) {
if (stream.waiting) {
return OGVPlayer.NETWORK_LOADING;
} else {
return OGVPlayer.NETWORK_IDLE;
}
} else {
if (self.readyState == OGVPlayer.HAVE_NOTHING) {
return OGVPlayer.NETWORK_EMPTY;
} else {
return OGVPlayer.NETWORK_NO_SOURCE;
}
}
}
});
/**
* @property playbackRate {number}
* @todo implement
*/
Object.defineProperty(self, "playbackRate", {
get: function getPlaybackRate() {
return 1;
},
set: function setPlaybackRate(val) {
// ignore
}
});
/**
* @property played {OGVTimeRanges}
* @todo implement correctly more or less
*/
Object.defineProperty(self, "played", {
get: function getPlayed() {
return new OGVTimeRanges([[0, self.currentTime]]);
}
});
var _volume = 1;
/**
* @property volume {number}
* @todo implement
*/
Object.defineProperty(self, "volume", {
get: function getVolume() {
return _volume;
},
set: function setVolume(val) {
_volume = +val;
if (audioFeeder) {
audioFeeder.volume = _volume;
}
fireEventAsync('volumechange');
}
});
// Events!
/**
* custom onframecallback, takes frame decode time in ms
*/
self.onframecallback = null;
/**
* Network state events
* @todo implement
*/
self.onloadstate = null;
self.onprogress = null;
self.onsuspend = null;
self.onabort = null;
self.onemptied = null;
self.onstalled = null;
/**
* Called when all metadata is available.
* Note in theory we must know 'duration' at this point.
*/
self.onloadedmetadata = null;
/**
* Called when enough data for first frame is ready.
* @todo implement
*/
self.onloadeddata = null;
/**
* Called when enough data is ready to start some playback.
* @todo implement
*/
self.oncanplay = null;
/**
* Called when enough data is ready to play through to the
* end if no surprises in network download rate.
* @todo implement
*/
self.oncanplaythrough = null;
/**
* Called when playback continues after a stall
* @todo implement
*/
self.onplaying = null;
/**
* Called when playback is delayed waiting on data
* @todo implement
*/
self.onwaiting = null;
/**
* Called when seeking begins
* @todo implement
*/
self.onseeking = null;
/**
* Called when seeking ends
* @todo implement
*/
self.onseeked = null;
/**
* Called when playback ends
*/
self.onended = null;
/**
* Called when duration becomes known
* @todo implement
*/
self.ondurationchange = null;
/**
* Called periodically during playback
*/
self.ontimeupdate = null;
/**
* Called when we start playback
*/
self.onplay = null;
/**
* Called when we get paused
*/
self.onpause = null;
/**
* Called when the playback rate changes
* @todo implement
*/
self.onratechange = null;
/**
* Called when the size of the video stream changes
* @todo implement
*/
self.onresize = null;
/**
* Called when the volume setting changes
* @todo implement
*/
self.onvolumechange = null;
return self;
};
OGVPlayer.initSharedAudioContext = function() {
AudioFeeder.initSharedAudioContext();
};
/**
* Set up constants on the class and instances
*/
var constants = {
/**
* Constants for networkState
*/
NETWORK_EMPTY: 0,
NETWORK_IDLE: 1,
NETWORK_LOADING: 2,
NETWORK_NO_SOURCE: 3,
/**
* Constants for readyState
*/
HAVE_NOTHING: 0,
HAVE_METADATA: 1,
HAVE_CURRENT_DATA: 2,
HAVE_FUTURE_DATA: 3,
HAVE_ENOUGH_DATA: 4
};
extend(OGVPlayer, constants);
OGVPlayer.instanceCount = 0;
function StyleManager() {
var self = this;
var el = document.createElement('style');
el.type = 'text/css';
el.textContent = 'ogvjs { ' +
'display: inline-block; ' +
'position: relative; ' +
'-webkit-user-select: none; ' +
'-webkit-tap-highlight-color: rgba(0,0,0,0); '
'}';
document.head.appendChild(el);
var sheet = el.sheet;
self.appendRule = function(selector, defs) {
var bits = [];
for (var prop in defs) {
if (defs.hasOwnProperty(prop)) {
bits.push(prop + ':' + defs[prop]);
}
}
var rule = selector + '{' + bits.join(';') + '}';
sheet.insertRule(rule, sheet.cssRules.length - 1);
};
}
OGVPlayer.styleManager = new StyleManager();
// IE 10/11 and Edge 12 don't support object-fit.
// Also just for fun, IE 10 doesn't support 'auto' sizing on canvas. o_O
OGVPlayer.supportsObjectFit = (typeof document.createElement('div').style.objectFit === 'string');
if (OGVPlayer.supportsObjectFit && navigator.userAgent.match(/iPhone|iPad|iPod Touch/)) {
// Safari for iOS 8/9 supports it but positions our <canvas> incorrectly when using WebGL >:(
OGVPlayer.supportsObjectFit = false;
}
if (OGVPlayer.supportsObjectFit) {
OGVPlayer.updatePositionOnResize = function() {
// no-op
};
} else {
OGVPlayer.updatePositionOnResize = function() {
function fixup(el, width, height) {
var container = el.offsetParent || el.parentNode,
containerAspect = container.offsetWidth / container.offsetHeight,
intrinsicAspect = width / height;
if (intrinsicAspect > containerAspect) {
var vsize = container.offsetWidth / intrinsicAspect,
vpad = (container.offsetHeight - vsize) / 2;
el.style.width = '100%';
el.style.height = vsize + 'px';
el.style.marginLeft = 0;
el.style.marginRight = 0;
el.style.marginTop = vpad + 'px';
el.style.marginBottom = vpad + 'px';
} else {
var hsize = container.offsetHeight * intrinsicAspect,
hpad = (container.offsetWidth - hsize) / 2;
el.style.width = hsize + 'px';
el.style.height = '100%';
el.style.marginLeft = hpad + 'px';
el.style.marginRight = hpad + 'px';
el.style.marginTop = 0;
el.style.marginBottom = 0;
}
}
function queryOver(selector, callback) {
var nodeList = document.querySelectorAll(selector),
nodeArray = Array.prototype.slice.call(nodeList);
nodeArray.forEach(callback);
}
queryOver('ogvjs > canvas', function(canvas) {
fixup(canvas, canvas.width, canvas.height);
});
queryOver('ogvjs > img', function(poster) {
fixup(poster, poster.naturalWidth, poster.naturalHeight);
});
};
var fullResizeVideo = function() {
// fullscreens may ping us before the resize happens
setTimeout(OGVPlayer.updatePositionOnResize, 0);
};
window.addEventListener('resize', OGVPlayer.updatePositionOnResize);
window.addEventListener('orientationchange', OGVPlayer.updatePositionOnResize);
document.addEventListener('fullscreenchange', fullResizeVideo);
document.addEventListener('mozfullscreenchange', fullResizeVideo);
document.addEventListener('webkitfullscreenchange', fullResizeVideo);
document.addEventListener('MSFullscreenChange', fullResizeVideo);
}
module.exports = OGVPlayer;
/***/ }),
/* 9 */
/***/ (function(module, exports, __webpack_require__) {
var require;/* WEBPACK VAR INJECTION */(function(process, global) {/*!
* @overview es6-promise - a tiny implementation of Promises/A+.
* @copyright Copyright (c) 2014 Yehuda Katz, Tom Dale, Stefan Penner and contributors (Conversion to ES6 API by Jake Archibald)
* @license Licensed under MIT license
* See https://raw.githubusercontent.com/stefanpenner/es6-promise/master/LICENSE
* @version 4.1.0
*/
(function (global, factory) {
true ? module.exports = factory() :
typeof define === 'function' && define.amd ? define(factory) :
(global.ES6Promise = factory());
}(this, (function () { 'use strict';
function objectOrFunction(x) {
return typeof x === 'function' || typeof x === 'object' && x !== null;
}
function isFunction(x) {
return typeof x === 'function';
}
var _isArray = undefined;
if (!Array.isArray) {
_isArray = function (x) {
return Object.prototype.toString.call(x) === '[object Array]';
};
} else {
_isArray = Array.isArray;
}
var isArray = _isArray;
var len = 0;
var vertxNext = undefined;
var customSchedulerFn = undefined;
var asap = function asap(callback, arg) {
queue[len] = callback;
queue[len + 1] = arg;
len += 2;
if (len === 2) {
// If len is 2, that means that we need to schedule an async flush.
// If additional callbacks are queued before the queue is flushed, they
// will be processed by this flush that we are scheduling.
if (customSchedulerFn) {
customSchedulerFn(flush);
} else {
scheduleFlush();
}
}
};
function setScheduler(scheduleFn) {
customSchedulerFn = scheduleFn;
}
function setAsap(asapFn) {
asap = asapFn;
}
var browserWindow = typeof window !== 'undefined' ? window : undefined;
var browserGlobal = browserWindow || {};
var BrowserMutationObserver = browserGlobal.MutationObserver || browserGlobal.WebKitMutationObserver;
var isNode = typeof self === 'undefined' && typeof process !== 'undefined' && ({}).toString.call(process) === '[object process]';
// test for web worker but not in IE10
var isWorker = typeof Uint8ClampedArray !== 'undefined' && typeof importScripts !== 'undefined' && typeof MessageChannel !== 'undefined';
// node
function useNextTick() {
// node version 0.10.x displays a deprecation warning when nextTick is used recursively
// see https://github.com/cujojs/when/issues/410 for details
return function () {
return process.nextTick(flush);
};
}
// vertx
function useVertxTimer() {
if (typeof vertxNext !== 'undefined') {
return function () {
vertxNext(flush);
};
}
return useSetTimeout();
}
function useMutationObserver() {
var iterations = 0;
var observer = new BrowserMutationObserver(flush);
var node = document.createTextNode('');
observer.observe(node, { characterData: true });
return function () {
node.data = iterations = ++iterations % 2;
};
}
// web worker
function useMessageChannel() {
var channel = new MessageChannel();
channel.port1.onmessage = flush;
return function () {
return channel.port2.postMessage(0);
};
}
function useSetTimeout() {
// Store setTimeout reference so es6-promise will be unaffected by
// other code modifying setTimeout (like sinon.useFakeTimers())
var globalSetTimeout = setTimeout;
return function () {
return globalSetTimeout(flush, 1);
};
}
var queue = new Array(1000);
function flush() {
for (var i = 0; i < len; i += 2) {
var callback = queue[i];
var arg = queue[i + 1];
callback(arg);
queue[i] = undefined;
queue[i + 1] = undefined;
}
len = 0;
}
function attemptVertx() {
try {
var r = require;
var vertx = __webpack_require__(11);
vertxNext = vertx.runOnLoop || vertx.runOnContext;
return useVertxTimer();
} catch (e) {
return useSetTimeout();
}
}
var scheduleFlush = undefined;
// Decide what async method to use to triggering processing of queued callbacks:
if (isNode) {
scheduleFlush = useNextTick();
} else if (BrowserMutationObserver) {
scheduleFlush = useMutationObserver();
} else if (isWorker) {
scheduleFlush = useMessageChannel();
} else if (browserWindow === undefined && "function" === 'function') {
scheduleFlush = attemptVertx();
} else {
scheduleFlush = useSetTimeout();
}
function then(onFulfillment, onRejection) {
var _arguments = arguments;
var parent = this;
var child = new this.constructor(noop);
if (child[PROMISE_ID] === undefined) {
makePromise(child);
}
var _state = parent._state;
if (_state) {
(function () {
var callback = _arguments[_state - 1];
asap(function () {
return invokeCallback(_state, child, callback, parent._result);
});
})();
} else {
subscribe(parent, child, onFulfillment, onRejection);
}
return child;
}
/**
`Promise.resolve` returns a promise that will become resolved with the
passed `value`. It is shorthand for the following:
```javascript
let promise = new Promise(function(resolve, reject){
resolve(1);
});
promise.then(function(value){
// value === 1
});
```
Instead of writing the above, your code now simply becomes the following:
```javascript
let promise = Promise.resolve(1);
promise.then(function(value){
// value === 1
});
```
@method resolve
@static
@param {Any} value value that the returned promise will be resolved with
Useful for tooling.
@return {Promise} a promise that will become fulfilled with the given
`value`
*/
function resolve(object) {
/*jshint validthis:true */
var Constructor = this;
if (object && typeof object === 'object' && object.constructor === Constructor) {
return object;
}
var promise = new Constructor(noop);
_resolve(promise, object);
return promise;
}
var PROMISE_ID = Math.random().toString(36).substring(16);
function noop() {}
var PENDING = void 0;
var FULFILLED = 1;
var REJECTED = 2;
var GET_THEN_ERROR = new ErrorObject();
function selfFulfillment() {
return new TypeError("You cannot resolve a promise with itself");
}
function cannotReturnOwn() {
return new TypeError('A promises callback cannot return that same promise.');
}
function getThen(promise) {
try {
return promise.then;
} catch (error) {
GET_THEN_ERROR.error = error;
return GET_THEN_ERROR;
}
}
function tryThen(then, value, fulfillmentHandler, rejectionHandler) {
try {
then.call(value, fulfillmentHandler, rejectionHandler);
} catch (e) {
return e;
}
}
function handleForeignThenable(promise, thenable, then) {
asap(function (promise) {
var sealed = false;
var error = tryThen(then, thenable, function (value) {
if (sealed) {
return;
}
sealed = true;
if (thenable !== value) {
_resolve(promise, value);
} else {
fulfill(promise, value);
}
}, function (reason) {
if (sealed) {
return;
}
sealed = true;
_reject(promise, reason);
}, 'Settle: ' + (promise._label || ' unknown promise'));
if (!sealed && error) {
sealed = true;
_reject(promise, error);
}
}, promise);
}
function handleOwnThenable(promise, thenable) {
if (thenable._state === FULFILLED) {
fulfill(promise, thenable._result);
} else if (thenable._state === REJECTED) {
_reject(promise, thenable._result);
} else {
subscribe(thenable, undefined, function (value) {
return _resolve(promise, value);
}, function (reason) {
return _reject(promise, reason);
});
}
}
function handleMaybeThenable(promise, maybeThenable, then$$) {
if (maybeThenable.constructor === promise.constructor && then$$ === then && maybeThenable.constructor.resolve === resolve) {
handleOwnThenable(promise, maybeThenable);
} else {
if (then$$ === GET_THEN_ERROR) {
_reject(promise, GET_THEN_ERROR.error);
GET_THEN_ERROR.error = null;
} else if (then$$ === undefined) {
fulfill(promise, maybeThenable);
} else if (isFunction(then$$)) {
handleForeignThenable(promise, maybeThenable, then$$);
} else {
fulfill(promise, maybeThenable);
}
}
}
function _resolve(promise, value) {
if (promise === value) {
_reject(promise, selfFulfillment());
} else if (objectOrFunction(value)) {
handleMaybeThenable(promise, value, getThen(value));
} else {
fulfill(promise, value);
}
}
function publishRejection(promise) {
if (promise._onerror) {
promise._onerror(promise._result);
}
publish(promise);
}
function fulfill(promise, value) {
if (promise._state !== PENDING) {
return;
}
promise._result = value;
promise._state = FULFILLED;
if (promise._subscribers.length !== 0) {
asap(publish, promise);
}
}
function _reject(promise, reason) {
if (promise._state !== PENDING) {
return;
}
promise._state = REJECTED;
promise._result = reason;
asap(publishRejection, promise);
}
function subscribe(parent, child, onFulfillment, onRejection) {
var _subscribers = parent._subscribers;
var length = _subscribers.length;
parent._onerror = null;
_subscribers[length] = child;
_subscribers[length + FULFILLED] = onFulfillment;
_subscribers[length + REJECTED] = onRejection;
if (length === 0 && parent._state) {
asap(publish, parent);
}
}
function publish(promise) {
var subscribers = promise._subscribers;
var settled = promise._state;
if (subscribers.length === 0) {
return;
}
var child = undefined,
callback = undefined,
detail = promise._result;
for (var i = 0; i < subscribers.length; i += 3) {
child = subscribers[i];
callback = subscribers[i + settled];
if (child) {
invokeCallback(settled, child, callback, detail);
} else {
callback(detail);
}
}
promise._subscribers.length = 0;
}
function ErrorObject() {
this.error = null;
}
var TRY_CATCH_ERROR = new ErrorObject();
function tryCatch(callback, detail) {
try {
return callback(detail);
} catch (e) {
TRY_CATCH_ERROR.error = e;
return TRY_CATCH_ERROR;
}
}
function invokeCallback(settled, promise, callback, detail) {
var hasCallback = isFunction(callback),
value = undefined,
error = undefined,
succeeded = undefined,
failed = undefined;
if (hasCallback) {
value = tryCatch(callback, detail);
if (value === TRY_CATCH_ERROR) {
failed = true;
error = value.error;
value.error = null;
} else {
succeeded = true;
}
if (promise === value) {
_reject(promise, cannotReturnOwn());
return;
}
} else {
value = detail;
succeeded = true;
}
if (promise._state !== PENDING) {
// noop
} else if (hasCallback && succeeded) {
_resolve(promise, value);
} else if (failed) {
_reject(promise, error);
} else if (settled === FULFILLED) {
fulfill(promise, value);
} else if (settled === REJECTED) {
_reject(promise, value);
}
}
function initializePromise(promise, resolver) {
try {
resolver(function resolvePromise(value) {
_resolve(promise, value);
}, function rejectPromise(reason) {
_reject(promise, reason);
});
} catch (e) {
_reject(promise, e);
}
}
var id = 0;
function nextId() {
return id++;
}
function makePromise(promise) {
promise[PROMISE_ID] = id++;
promise._state = undefined;
promise._result = undefined;
promise._subscribers = [];
}
function Enumerator(Constructor, input) {
this._instanceConstructor = Constructor;
this.promise = new Constructor(noop);
if (!this.promise[PROMISE_ID]) {
makePromise(this.promise);
}
if (isArray(input)) {
this._input = input;
this.length = input.length;
this._remaining = input.length;
this._result = new Array(this.length);
if (this.length === 0) {
fulfill(this.promise, this._result);
} else {
this.length = this.length || 0;
this._enumerate();
if (this._remaining === 0) {
fulfill(this.promise, this._result);
}
}
} else {
_reject(this.promise, validationError());
}
}
function validationError() {
return new Error('Array Methods must be provided an Array');
};
Enumerator.prototype._enumerate = function () {
var length = this.length;
var _input = this._input;
for (var i = 0; this._state === PENDING && i < length; i++) {
this._eachEntry(_input[i], i);
}
};
Enumerator.prototype._eachEntry = function (entry, i) {
var c = this._instanceConstructor;
var resolve$$ = c.resolve;
if (resolve$$ === resolve) {
var _then = getThen(entry);
if (_then === then && entry._state !== PENDING) {
this._settledAt(entry._state, i, entry._result);
} else if (typeof _then !== 'function') {
this._remaining--;
this._result[i] = entry;
} else if (c === Promise) {
var promise = new c(noop);
handleMaybeThenable(promise, entry, _then);
this._willSettleAt(promise, i);
} else {
this._willSettleAt(new c(function (resolve$$) {
return resolve$$(entry);
}), i);
}
} else {
this._willSettleAt(resolve$$(entry), i);
}
};
Enumerator.prototype._settledAt = function (state, i, value) {
var promise = this.promise;
if (promise._state === PENDING) {
this._remaining--;
if (state === REJECTED) {
_reject(promise, value);
} else {
this._result[i] = value;
}
}
if (this._remaining === 0) {
fulfill(promise, this._result);
}
};
Enumerator.prototype._willSettleAt = function (promise, i) {
var enumerator = this;
subscribe(promise, undefined, function (value) {
return enumerator._settledAt(FULFILLED, i, value);
}, function (reason) {
return enumerator._settledAt(REJECTED, i, reason);
});
};
/**
`Promise.all` accepts an array of promises, and returns a new promise which
is fulfilled with an array of fulfillment values for the passed promises, or
rejected with the reason of the first passed promise to be rejected. It casts all
elements of the passed iterable to promises as it runs this algorithm.
Example:
```javascript
let promise1 = resolve(1);
let promise2 = resolve(2);
let promise3 = resolve(3);
let promises = [ promise1, promise2, promise3 ];
Promise.all(promises).then(function(array){
// The array here would be [ 1, 2, 3 ];
});
```
If any of the `promises` given to `all` are rejected, the first promise
that is rejected will be given as an argument to the returned promises's
rejection handler. For example:
Example:
```javascript
let promise1 = resolve(1);
let promise2 = reject(new Error("2"));
let promise3 = reject(new Error("3"));
let promises = [ promise1, promise2, promise3 ];
Promise.all(promises).then(function(array){
// Code here never runs because there are rejected promises!
}, function(error) {
// error.message === "2"
});
```
@method all
@static
@param {Array} entries array of promises
@param {String} label optional string for labeling the promise.
Useful for tooling.
@return {Promise} promise that is fulfilled when all `promises` have been
fulfilled, or rejected if any of them become rejected.
@static
*/
function all(entries) {
return new Enumerator(this, entries).promise;
}
/**
`Promise.race` returns a new promise which is settled in the same way as the
first passed promise to settle.
Example:
```javascript
let promise1 = new Promise(function(resolve, reject){
setTimeout(function(){
resolve('promise 1');
}, 200);
});
let promise2 = new Promise(function(resolve, reject){
setTimeout(function(){
resolve('promise 2');
}, 100);
});
Promise.race([promise1, promise2]).then(function(result){
// result === 'promise 2' because it was resolved before promise1
// was resolved.
});
```
`Promise.race` is deterministic in that only the state of the first
settled promise matters. For example, even if other promises given to the
`promises` array argument are resolved, but the first settled promise has
become rejected before the other promises became fulfilled, the returned
promise will become rejected:
```javascript
let promise1 = new Promise(function(resolve, reject){
setTimeout(function(){
resolve('promise 1');
}, 200);
});
let promise2 = new Promise(function(resolve, reject){
setTimeout(function(){
reject(new Error('promise 2'));
}, 100);
});
Promise.race([promise1, promise2]).then(function(result){
// Code here never runs
}, function(reason){
// reason.message === 'promise 2' because promise 2 became rejected before
// promise 1 became fulfilled
});
```
An example real-world use case is implementing timeouts:
```javascript
Promise.race([ajax('foo.json'), timeout(5000)])
```
@method race
@static
@param {Array} promises array of promises to observe
Useful for tooling.
@return {Promise} a promise which settles in the same way as the first passed
promise to settle.
*/
function race(entries) {
/*jshint validthis:true */
var Constructor = this;
if (!isArray(entries)) {
return new Constructor(function (_, reject) {
return reject(new TypeError('You must pass an array to race.'));
});
} else {
return new Constructor(function (resolve, reject) {
var length = entries.length;
for (var i = 0; i < length; i++) {
Constructor.resolve(entries[i]).then(resolve, reject);
}
});
}
}
/**
`Promise.reject` returns a promise rejected with the passed `reason`.
It is shorthand for the following:
```javascript
let promise = new Promise(function(resolve, reject){
reject(new Error('WHOOPS'));
});
promise.then(function(value){
// Code here doesn't run because the promise is rejected!
}, function(reason){
// reason.message === 'WHOOPS'
});
```
Instead of writing the above, your code now simply becomes the following:
```javascript
let promise = Promise.reject(new Error('WHOOPS'));
promise.then(function(value){
// Code here doesn't run because the promise is rejected!
}, function(reason){
// reason.message === 'WHOOPS'
});
```
@method reject
@static
@param {Any} reason value that the returned promise will be rejected with.
Useful for tooling.
@return {Promise} a promise rejected with the given `reason`.
*/
function reject(reason) {
/*jshint validthis:true */
var Constructor = this;
var promise = new Constructor(noop);
_reject(promise, reason);
return promise;
}
function needsResolver() {
throw new TypeError('You must pass a resolver function as the first argument to the promise constructor');
}
function needsNew() {
throw new TypeError("Failed to construct 'Promise': Please use the 'new' operator, this object constructor cannot be called as a function.");
}
/**
Promise objects represent the eventual result of an asynchronous operation. The
primary way of interacting with a promise is through its `then` method, which
registers callbacks to receive either a promise's eventual value or the reason
why the promise cannot be fulfilled.
Terminology
-----------
- `promise` is an object or function with a `then` method whose behavior conforms to this specification.
- `thenable` is an object or function that defines a `then` method.
- `value` is any legal JavaScript value (including undefined, a thenable, or a promise).
- `exception` is a value that is thrown using the throw statement.
- `reason` is a value that indicates why a promise was rejected.
- `settled` the final resting state of a promise, fulfilled or rejected.
A promise can be in one of three states: pending, fulfilled, or rejected.
Promises that are fulfilled have a fulfillment value and are in the fulfilled
state. Promises that are rejected have a rejection reason and are in the
rejected state. A fulfillment value is never a thenable.
Promises can also be said to *resolve* a value. If this value is also a
promise, then the original promise's settled state will match the value's
settled state. So a promise that *resolves* a promise that rejects will
itself reject, and a promise that *resolves* a promise that fulfills will
itself fulfill.
Basic Usage:
------------
```js
let promise = new Promise(function(resolve, reject) {
// on success
resolve(value);
// on failure
reject(reason);
});
promise.then(function(value) {
// on fulfillment
}, function(reason) {
// on rejection
});
```
Advanced Usage:
---------------
Promises shine when abstracting away asynchronous interactions such as
`XMLHttpRequest`s.
```js
function getJSON(url) {
return new Promise(function(resolve, reject){
let xhr = new XMLHttpRequest();
xhr.open('GET', url);
xhr.onreadystatechange = handler;
xhr.responseType = 'json';
xhr.setRequestHeader('Accept', 'application/json');
xhr.send();
function handler() {
if (this.readyState === this.DONE) {
if (this.status === 200) {
resolve(this.response);
} else {
reject(new Error('getJSON: `' + url + '` failed with status: [' + this.status + ']'));
}
}
};
});
}
getJSON('/posts.json').then(function(json) {
// on fulfillment
}, function(reason) {
// on rejection
});
```
Unlike callbacks, promises are great composable primitives.
```js
Promise.all([
getJSON('/posts'),
getJSON('/comments')
]).then(function(values){
values[0] // => postsJSON
values[1] // => commentsJSON
return values;
});
```
@class Promise
@param {function} resolver
Useful for tooling.
@constructor
*/
function Promise(resolver) {
this[PROMISE_ID] = nextId();
this._result = this._state = undefined;
this._subscribers = [];
if (noop !== resolver) {
typeof resolver !== 'function' && needsResolver();
this instanceof Promise ? initializePromise(this, resolver) : needsNew();
}
}
Promise.all = all;
Promise.race = race;
Promise.resolve = resolve;
Promise.reject = reject;
Promise._setScheduler = setScheduler;
Promise._setAsap = setAsap;
Promise._asap = asap;
Promise.prototype = {
constructor: Promise,
/**
The primary way of interacting with a promise is through its `then` method,
which registers callbacks to receive either a promise's eventual value or the
reason why the promise cannot be fulfilled.
```js
findUser().then(function(user){
// user is available
}, function(reason){
// user is unavailable, and you are given the reason why
});
```
Chaining
--------
The return value of `then` is itself a promise. This second, 'downstream'
promise is resolved with the return value of the first promise's fulfillment
or rejection handler, or rejected if the handler throws an exception.
```js
findUser().then(function (user) {
return user.name;
}, function (reason) {
return 'default name';
}).then(function (userName) {
// If `findUser` fulfilled, `userName` will be the user's name, otherwise it
// will be `'default name'`
});
findUser().then(function (user) {
throw new Error('Found user, but still unhappy');
}, function (reason) {
throw new Error('`findUser` rejected and we're unhappy');
}).then(function (value) {
// never reached
}, function (reason) {
// if `findUser` fulfilled, `reason` will be 'Found user, but still unhappy'.
// If `findUser` rejected, `reason` will be '`findUser` rejected and we're unhappy'.
});
```
If the downstream promise does not specify a rejection handler, rejection reasons will be propagated further downstream.
```js
findUser().then(function (user) {
throw new PedagogicalException('Upstream error');
}).then(function (value) {
// never reached
}).then(function (value) {
// never reached
}, function (reason) {
// The `PedgagocialException` is propagated all the way down to here
});
```
Assimilation
------------
Sometimes the value you want to propagate to a downstream promise can only be
retrieved asynchronously. This can be achieved by returning a promise in the
fulfillment or rejection handler. The downstream promise will then be pending
until the returned promise is settled. This is called *assimilation*.
```js
findUser().then(function (user) {
return findCommentsByAuthor(user);
}).then(function (comments) {
// The user's comments are now available
});
```
If the assimliated promise rejects, then the downstream promise will also reject.
```js
findUser().then(function (user) {
return findCommentsByAuthor(user);
}).then(function (comments) {
// If `findCommentsByAuthor` fulfills, we'll have the value here
}, function (reason) {
// If `findCommentsByAuthor` rejects, we'll have the reason here
});
```
Simple Example
--------------
Synchronous Example
```javascript
let result;
try {
result = findResult();
// success
} catch(reason) {
// failure
}
```
Errback Example
```js
findResult(function(result, err){
if (err) {
// failure
} else {
// success
}
});
```
Promise Example;
```javascript
findResult().then(function(result){
// success
}, function(reason){
// failure
});
```
Advanced Example
--------------
Synchronous Example
```javascript
let author, books;
try {
author = findAuthor();
books = findBooksByAuthor(author);
// success
} catch(reason) {
// failure
}
```
Errback Example
```js
function foundBooks(books) {
}
function failure(reason) {
}
findAuthor(function(author, err){
if (err) {
failure(err);
// failure
} else {
try {
findBoooksByAuthor(author, function(books, err) {
if (err) {
failure(err);
} else {
try {
foundBooks(books);
} catch(reason) {
failure(reason);
}
}
});
} catch(error) {
failure(err);
}
// success
}
});
```
Promise Example;
```javascript
findAuthor().
then(findBooksByAuthor).
then(function(books){
// found books
}).catch(function(reason){
// something went wrong
});
```
@method then
@param {Function} onFulfilled
@param {Function} onRejected
Useful for tooling.
@return {Promise}
*/
then: then,
/**
`catch` is simply sugar for `then(undefined, onRejection)` which makes it the same
as the catch block of a try/catch statement.
```js
function findAuthor(){
throw new Error('couldn't find that author');
}
// synchronous
try {
findAuthor();
} catch(reason) {
// something went wrong
}
// async with promises
findAuthor().catch(function(reason){
// something went wrong
});
```
@method catch
@param {Function} onRejection
Useful for tooling.
@return {Promise}
*/
'catch': function _catch(onRejection) {
return this.then(null, onRejection);
}
};
function polyfill() {
var local = undefined;
if (typeof global !== 'undefined') {
local = global;
} else if (typeof self !== 'undefined') {
local = self;
} else {
try {
local = Function('return this')();
} catch (e) {
throw new Error('polyfill failed because global object is unavailable in this environment');
}
}
var P = local.Promise;
if (P) {
var promiseToString = null;
try {
promiseToString = Object.prototype.toString.call(P.resolve());
} catch (e) {
// silently ignored
}
if (promiseToString === '[object Promise]' && !P.cast) {
return;
}
}
local.Promise = Promise;
}
// Strange compat..
Promise.polyfill = polyfill;
Promise.Promise = Promise;
return Promise;
})));
/* WEBPACK VAR INJECTION */}.call(exports, __webpack_require__(10), (function() { return this; }())))
/***/ }),
/* 10 */
/***/ (function(module, exports) {
// shim for using process in browser
var process = module.exports = {};
// cached from whatever global is present so that test runners that stub it
// don't break things. But we need to wrap it in a try catch in case it is
// wrapped in strict mode code which doesn't define any globals. It's inside a
// function because try/catches deoptimize in certain engines.
var cachedSetTimeout;
var cachedClearTimeout;
function defaultSetTimout() {
throw new Error('setTimeout has not been defined');
}
function defaultClearTimeout () {
throw new Error('clearTimeout has not been defined');
}
(function () {
try {
if (typeof setTimeout === 'function') {
cachedSetTimeout = setTimeout;
} else {
cachedSetTimeout = defaultSetTimout;
}
} catch (e) {
cachedSetTimeout = defaultSetTimout;
}
try {
if (typeof clearTimeout === 'function') {
cachedClearTimeout = clearTimeout;
} else {
cachedClearTimeout = defaultClearTimeout;
}
} catch (e) {
cachedClearTimeout = defaultClearTimeout;
}
} ())
function runTimeout(fun) {
if (cachedSetTimeout === setTimeout) {
//normal enviroments in sane situations
return setTimeout(fun, 0);
}
// if setTimeout wasn't available but was latter defined
if ((cachedSetTimeout === defaultSetTimout || !cachedSetTimeout) && setTimeout) {
cachedSetTimeout = setTimeout;
return setTimeout(fun, 0);
}
try {
// when when somebody has screwed with setTimeout but no I.E. maddness
return cachedSetTimeout(fun, 0);
} catch(e){
try {
// When we are in I.E. but the script has been evaled so I.E. doesn't trust the global object when called normally
return cachedSetTimeout.call(null, fun, 0);
} catch(e){
// same as above but when it's a version of I.E. that must have the global object for 'this', hopfully our context correct otherwise it will throw a global error
return cachedSetTimeout.call(this, fun, 0);
}
}
}
function runClearTimeout(marker) {
if (cachedClearTimeout === clearTimeout) {
//normal enviroments in sane situations
return clearTimeout(marker);
}
// if clearTimeout wasn't available but was latter defined
if ((cachedClearTimeout === defaultClearTimeout || !cachedClearTimeout) && clearTimeout) {
cachedClearTimeout = clearTimeout;
return clearTimeout(marker);
}
try {
// when when somebody has screwed with setTimeout but no I.E. maddness
return cachedClearTimeout(marker);
} catch (e){
try {
// When we are in I.E. but the script has been evaled so I.E. doesn't trust the global object when called normally
return cachedClearTimeout.call(null, marker);
} catch (e){
// same as above but when it's a version of I.E. that must have the global object for 'this', hopfully our context correct otherwise it will throw a global error.
// Some versions of I.E. have different rules for clearTimeout vs setTimeout
return cachedClearTimeout.call(this, marker);
}
}
}
var queue = [];
var draining = false;
var currentQueue;
var queueIndex = -1;
function cleanUpNextTick() {
if (!draining || !currentQueue) {
return;
}
draining = false;
if (currentQueue.length) {
queue = currentQueue.concat(queue);
} else {
queueIndex = -1;
}
if (queue.length) {
drainQueue();
}
}
function drainQueue() {
if (draining) {
return;
}
var timeout = runTimeout(cleanUpNextTick);
draining = true;
var len = queue.length;
while(len) {
currentQueue = queue;
queue = [];
while (++queueIndex < len) {
if (currentQueue) {
currentQueue[queueIndex].run();
}
}
queueIndex = -1;
len = queue.length;
}
currentQueue = null;
draining = false;
runClearTimeout(timeout);
}
process.nextTick = function (fun) {
var args = new Array(arguments.length - 1);
if (arguments.length > 1) {
for (var i = 1; i < arguments.length; i++) {
args[i - 1] = arguments[i];
}
}
queue.push(new Item(fun, args));
if (queue.length === 1 && !draining) {
runTimeout(drainQueue);
}
};
// v8 likes predictible objects
function Item(fun, array) {
this.fun = fun;
this.array = array;
}
Item.prototype.run = function () {
this.fun.apply(null, this.array);
};
process.title = 'browser';
process.browser = true;
process.env = {};
process.argv = [];
process.version = ''; // empty string to avoid regexp issues
process.versions = {};
function noop() {}
process.on = noop;
process.addListener = noop;
process.once = noop;
process.off = noop;
process.removeListener = noop;
process.removeAllListeners = noop;
process.emit = noop;
process.binding = function (name) {
throw new Error('process.binding is not supported');
};
process.cwd = function () { return '/' };
process.chdir = function (dir) {
throw new Error('process.chdir is not supported');
};
process.umask = function() { return 0; };
/***/ }),
/* 11 */
/***/ (function(module, exports) {
/* (ignored) */
/***/ }),
/* 12 */
/***/ (function(module, exports, __webpack_require__) {
/*
Copyright (c) 2014-2016 Brion Vibber <brion@pobox.com>
Permission is hereby granted, free of charge, to any person obtaining a copy of
this software and associated documentation files (the "Software"), to deal in
the Software without restriction, including without limitation the rights to
use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of
the Software, and to permit persons to whom the Software is furnished to do so,
subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
MPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS
FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR
COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER
IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
*/
(function() {
"use strict";
var FrameSink = __webpack_require__(13),
SoftwareFrameSink = __webpack_require__(14),
WebGLFrameSink = __webpack_require__(17);
/**
* @typedef {Object} YUVCanvasOptions
* @property {boolean} webGL - Whether to use WebGL to draw to the canvas and accelerate color space conversion. If left out, defaults to auto-detect.
*/
var YUVCanvas = {
FrameSink: FrameSink,
SoftwareFrameSink: SoftwareFrameSink,
WebGLFrameSink: WebGLFrameSink,
/**
* Attach a suitable FrameSink instance to an HTML5 canvas element.
*
* This will take over the drawing context of the canvas and may turn
* it into a WebGL 3d canvas if possible. Do not attempt to use the
* drawing context directly after this.
*
* @param {HTMLCanvasElement} canvas - HTML canvas element to attach to
* @param {YUVCanvasOptions} options - map of options
* @returns {FrameSink} - instance of suitable subclass.
*/
attach: function(canvas, options) {
options = options || {};
var webGL = ('webGL' in options) ? options.webGL : WebGLFrameSink.isAvailable();
if (webGL) {
return new WebGLFrameSink(canvas, options);
} else {
return new SoftwareFrameSink(canvas, options);
}
}
};
module.exports = YUVCanvas;
})();
/***/ }),
/* 13 */
/***/ (function(module, exports) {
(function() {
"use strict";
/**
* Create a YUVCanvas and attach it to an HTML5 canvas element.
*
* This will take over the drawing context of the canvas and may turn
* it into a WebGL 3d canvas if possible. Do not attempt to use the
* drawing context directly after this.
*
* @param {HTMLCanvasElement} canvas - HTML canvas element to attach to
* @param {YUVCanvasOptions} options - map of options
* @throws exception if WebGL requested but unavailable
* @constructor
* @abstract
*/
function FrameSink(canvas, options) {
throw new Error('abstract');
}
/**
* Draw a single YUV frame on the underlying canvas, converting to RGB.
* If necessary the canvas will be resized to the optimal pixel size
* for the given buffer's format.
*
* @param {YUVBuffer} buffer - the YUV buffer to draw
* @see {@link https://www.npmjs.com/package/yuv-buffer|yuv-buffer} for format
*/
FrameSink.prototype.drawFrame = function(buffer) {
throw new Error('abstract');
};
/**
* Clear the canvas using appropriate underlying 2d or 3d context.
*/
FrameSink.prototype.clear = function() {
throw new Error('abstract');
};
module.exports = FrameSink;
})();
/***/ }),
/* 14 */
/***/ (function(module, exports, __webpack_require__) {
/*
Copyright (c) 2014-2016 Brion Vibber <brion@pobox.com>
Permission is hereby granted, free of charge, to any person obtaining a copy of
this software and associated documentation files (the "Software"), to deal in
the Software without restriction, including without limitation the rights to
use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of
the Software, and to permit persons to whom the Software is furnished to do so,
subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
MPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS
FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR
COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER
IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
*/
(function() {
"use strict";
var FrameSink = __webpack_require__(13),
YCbCr = __webpack_require__(15);
/**
* @param {HTMLCanvasElement} canvas - HTML canvas eledment to attach to
* @constructor
*/
function SoftwareFrameSink(canvas) {
var self = this,
ctx = canvas.getContext('2d'),
imageData = null,
resampleCanvas = null,
resampleContext = null;
function initImageData(width, height) {
imageData = ctx.createImageData(width, height);
// Prefill the alpha to opaque
var data = imageData.data,
pixelCount = width * height * 4;
for (var i = 0; i < pixelCount; i += 4) {
data[i + 3] = 255;
}
}
function initResampleCanvas(cropWidth, cropHeight) {
resampleCanvas = document.createElement('canvas');
resampleCanvas.width = cropWidth;
resampleCanvas.height = cropHeight;
resampleContext = resampleCanvas.getContext('2d');
}
/**
* Actually draw a frame into the canvas.
* @param {YUVFrame} buffer - YUV frame buffer object to draw
*/
self.drawFrame = function drawFrame(buffer) {
var format = buffer.format;
if (canvas.width !== format.displayWidth || canvas.height !== format.displayHeight) {
// Keep the canvas at the right size...
canvas.width = format.displayWidth;
canvas.height = format.displayHeight;
}
if (imageData === null ||
imageData.width != format.width ||
imageData.height != format.height) {
initImageData(format.width, format.height);
}
// YUV -> RGB over the entire encoded frame
YCbCr.convertYCbCr(buffer, imageData.data);
var resample = (format.cropWidth != format.displayWidth || format.cropHeight != format.displayHeight);
var drawContext;
if (resample) {
// hack for non-square aspect-ratio
// putImageData doesn't resample, so we have to draw in two steps.
if (!resampleCanvas) {
initResampleCanvas(format.cropWidth, format.cropHeight);
}
drawContext = resampleContext;
} else {
drawContext = ctx;
}
// Draw cropped frame to either the final or temporary canvas
drawContext.putImageData(imageData,
0, 0,
format.cropLeft, format.cropTop,
format.cropWidth, format.cropHeight);
if (resample) {
ctx.drawImage(resampleCanvas, 0, 0, format.displayWidth, format.displayHeight);
}
};
self.clear = function() {
ctx.clearRect(0, 0, canvas.width, canvas.height);
};
return self;
}
SoftwareFrameSink.prototype = Object.create(FrameSink.prototype);
module.exports = SoftwareFrameSink;
})();
/***/ }),
/* 15 */
/***/ (function(module, exports, __webpack_require__) {
/*
Copyright (c) 2014-2016 Brion Vibber <brion@pobox.com>
Permission is hereby granted, free of charge, to any person obtaining a copy of
this software and associated documentation files (the "Software"), to deal in
the Software without restriction, including without limitation the rights to
use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of
the Software, and to permit persons to whom the Software is furnished to do so,
subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
MPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS
FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR
COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER
IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
*/
(function() {
"use strict";
var depower = __webpack_require__(16);
/**
* Basic YCbCr->RGB conversion
*
* @author Brion Vibber <brion@pobox.com>
* @copyright 2014-2016
* @license MIT-style
*
* @param {YUVFrame} buffer - input frame buffer
* @param {Uint8ClampedArray} output - array to draw RGBA into
* Assumes that the output array already has alpha channel set to opaque.
*/
function convertYCbCr(buffer, output) {
var width = buffer.format.width,
height = buffer.format.height,
hdec = depower(buffer.format.width / buffer.format.chromaWidth),
vdec = depower(buffer.format.height / buffer.format.chromaHeight),
bytesY = buffer.y.bytes,
bytesCb = buffer.u.bytes,
bytesCr = buffer.v.bytes,
strideY = buffer.y.stride,
strideCb = buffer.u.stride,
strideCr = buffer.v.stride,
outStride = 4 * width,
YPtr = 0, Y0Ptr = 0, Y1Ptr = 0,
CbPtr = 0, CrPtr = 0,
outPtr = 0, outPtr0 = 0, outPtr1 = 0,
colorCb = 0, colorCr = 0,
multY = 0, multCrR = 0, multCbCrG = 0, multCbB = 0,
x = 0, y = 0, xdec = 0, ydec = 0;
if (hdec == 1 && vdec == 1) {
// Optimize for 4:2:0, which is most common
outPtr0 = 0;
outPtr1 = outStride;
ydec = 0;
for (y = 0; y < height; y += 2) {
Y0Ptr = y * strideY;
Y1Ptr = Y0Ptr + strideY;
CbPtr = ydec * strideCb;
CrPtr = ydec * strideCr;
for (x = 0; x < width; x += 2) {
colorCb = bytesCb[CbPtr++];
colorCr = bytesCr[CrPtr++];
// Quickie YUV conversion
// https://en.wikipedia.org/wiki/YCbCr#ITU-R_BT.2020_conversion
// multiplied by 256 for integer-friendliness
multCrR = (409 * colorCr) - 57088;
multCbCrG = (100 * colorCb) + (208 * colorCr) - 34816;
multCbB = (516 * colorCb) - 70912;
multY = (298 * bytesY[Y0Ptr++]);
output[outPtr0++] = (multY + multCrR) >> 8;
output[outPtr0++] = (multY - multCbCrG) >> 8;
output[outPtr0++] = (multY + multCbB) >> 8;
outPtr0++;
multY = (298 * bytesY[Y0Ptr++]);
output[outPtr0++] = (multY + multCrR) >> 8;
output[outPtr0++] = (multY - multCbCrG) >> 8;
output[outPtr0++] = (multY + multCbB) >> 8;
outPtr0++;
multY = (298 * bytesY[Y1Ptr++]);
output[outPtr1++] = (multY + multCrR) >> 8;
output[outPtr1++] = (multY - multCbCrG) >> 8;
output[outPtr1++] = (multY + multCbB) >> 8;
outPtr1++;
multY = (298 * bytesY[Y1Ptr++]);
output[outPtr1++] = (multY + multCrR) >> 8;
output[outPtr1++] = (multY - multCbCrG) >> 8;
output[outPtr1++] = (multY + multCbB) >> 8;
outPtr1++;
}
outPtr0 += outStride;
outPtr1 += outStride;
ydec++;
}
} else {
outPtr = 0;
for (y = 0; y < height; y++) {
xdec = 0;
ydec = y >> vdec;
YPtr = y * strideY;
CbPtr = ydec * strideCb;
CrPtr = ydec * strideCr;
for (x = 0; x < width; x++) {
xdec = x >> hdec;
colorCb = bytesCb[CbPtr + xdec];
colorCr = bytesCr[CrPtr + xdec];
// Quickie YUV conversion
// https://en.wikipedia.org/wiki/YCbCr#ITU-R_BT.2020_conversion
// multiplied by 256 for integer-friendliness
multCrR = (409 * colorCr) - 57088;
multCbCrG = (100 * colorCb) + (208 * colorCr) - 34816;
multCbB = (516 * colorCb) - 70912;
multY = 298 * bytesY[YPtr++];
output[outPtr++] = (multY + multCrR) >> 8;
output[outPtr++] = (multY - multCbCrG) >> 8;
output[outPtr++] = (multY + multCbB) >> 8;
outPtr++;
}
}
}
}
module.exports = {
convertYCbCr: convertYCbCr
};
})();
/***/ }),
/* 16 */
/***/ (function(module, exports) {
/*
Copyright (c) 2014-2016 Brion Vibber <brion@pobox.com>
Permission is hereby granted, free of charge, to any person obtaining a copy of
this software and associated documentation files (the "Software"), to deal in
the Software without restriction, including without limitation the rights to
use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of
the Software, and to permit persons to whom the Software is furnished to do so,
subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
MPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS
FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR
COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER
IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
*/
(function() {
"use strict";
/**
* Convert a ratio into a bit-shift count; for instance a ratio of 2
* becomes a bit-shift of 1, while a ratio of 1 is a bit-shift of 0.
*
* @author Brion Vibber <brion@pobox.com>
* @copyright 2016
* @license MIT-style
*
* @param {number} ratio - the integer ratio to convert.
* @returns {number} - number of bits to shift to multiply/divide by the ratio.
* @throws exception if given a non-power-of-two
*/
function depower(ratio) {
var shiftCount = 0,
n = ratio >> 1;
while (n != 0) {
n = n >> 1;
shiftCount++
}
if (ratio !== (1 << shiftCount)) {
throw 'chroma plane dimensions must be power of 2 ratio to luma plane dimensions; got ' + ratio;
}
return shiftCount;
}
module.exports = depower;
})();
/***/ }),
/* 17 */
/***/ (function(module, exports, __webpack_require__) {
/*
Copyright (c) 2014-2016 Brion Vibber <brion@pobox.com>
Permission is hereby granted, free of charge, to any person obtaining a copy of
this software and associated documentation files (the "Software"), to deal in
the Software without restriction, including without limitation the rights to
use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of
the Software, and to permit persons to whom the Software is furnished to do so,
subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
MPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS
FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR
COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER
IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
*/
(function() {
"use strict";
var FrameSink = __webpack_require__(13),
shaders = __webpack_require__(18);
/**
* Warning: canvas must not have been used for 2d drawing prior!
*
* @param {HTMLCanvasElement} canvas - HTML canvas element to attach to
* @constructor
*/
function WebGLFrameSink(canvas) {
var self = this,
gl = canvas.getContext('webgl') || canvas.getContext('experimental-webgl'),
debug = false; // swap this to enable more error checks, which can slow down rendering
if (gl === null) {
throw new Error('WebGL unavailable');
}
// GL!
function checkError() {
if (debug) {
err = gl.getError();
if (err !== 0) {
throw new Error("GL error " + err);
}
}
}
function compileShader(type, source) {
var shader = gl.createShader(type);
gl.shaderSource(shader, source);
gl.compileShader(shader);
if (!gl.getShaderParameter(shader, gl.COMPILE_STATUS)) {
var err = gl.getShaderInfoLog(shader);
gl.deleteShader(shader);
throw new Error('GL shader compilation for ' + type + ' failed: ' + err);
}
return shader;
}
var vertexShader,
fragmentShader,
program,
buf,
err;
// In the world of GL there are no rectangles.
// There are only triangles.
// THERE IS NO SPOON.
var rectangle = new Float32Array([
// First triangle (top left, clockwise)
-1.0, -1.0,
+1.0, -1.0,
-1.0, +1.0,
// Second triangle (bottom right, clockwise)
-1.0, +1.0,
+1.0, -1.0,
+1.0, +1.0
]);
var textures = {};
function attachTexture(name, register, index, width, height, data) {
var texture,
texWidth = WebGLFrameSink.stripe ? (width / 4) : width,
format = WebGLFrameSink.stripe ? gl.RGBA : gl.LUMINANCE,
filter = WebGLFrameSink.stripe ? gl.NEAREST : gl.LINEAR;
if (textures[name]) {
// Reuse & update the existing texture
texture = textures[name];
} else {
textures[name] = texture = gl.createTexture();
checkError();
gl.uniform1i(gl.getUniformLocation(program, name), index);
checkError();
}
gl.activeTexture(register);
checkError();
gl.bindTexture(gl.TEXTURE_2D, texture);
checkError();
gl.texParameteri(gl.TEXTURE_2D, gl.TEXTURE_WRAP_S, gl.CLAMP_TO_EDGE);
checkError();
gl.texParameteri(gl.TEXTURE_2D, gl.TEXTURE_WRAP_T, gl.CLAMP_TO_EDGE);
checkError();
gl.texParameteri(gl.TEXTURE_2D, gl.TEXTURE_MIN_FILTER, filter);
checkError();
gl.texParameteri(gl.TEXTURE_2D, gl.TEXTURE_MAG_FILTER, filter);
checkError();
gl.texImage2D(
gl.TEXTURE_2D,
0, // mip level
format, // internal format
texWidth,
height,
0, // border
format, // format
gl.UNSIGNED_BYTE, //type
data // data!
);
checkError();
return texture;
}
function buildStripe(width, height) {
var len = width * height,
out = new Uint32Array(len);
for (var i = 0; i < len; i += 4) {
out[i ] = 0x000000ff;
out[i + 1] = 0x0000ff00;
out[i + 2] = 0x00ff0000;
out[i + 3] = 0xff000000;
}
return new Uint8Array(out.buffer);
}
function init(buffer) {
vertexShader = compileShader(gl.VERTEX_SHADER, shaders.vertex);
if (WebGLFrameSink.stripe) {
fragmentShader = compileShader(gl.FRAGMENT_SHADER, shaders.fragmentStripe);
} else {
fragmentShader = compileShader(gl.FRAGMENT_SHADER, shaders.fragment);
}
program = gl.createProgram();
gl.attachShader(program, vertexShader);
checkError();
gl.attachShader(program, fragmentShader);
checkError();
gl.linkProgram(program);
if (!gl.getProgramParameter(program, gl.LINK_STATUS)) {
var err = gl.getProgramInfoLog(program);
gl.deleteProgram(program);
throw new Error('GL program linking failed: ' + err);
}
gl.useProgram(program);
checkError();
if (WebGLFrameSink.stripe) {
attachTexture(
'uStripeLuma',
gl.TEXTURE3,
3,
buffer.y.stride * 4,
buffer.format.height,
buildStripe(buffer.y.stride, buffer.format.height)
);
checkError();
attachTexture(
'uStripeChroma',
gl.TEXTURE4,
4,
buffer.u.stride * 4,
buffer.format.chromaHeight,
buildStripe(buffer.u.stride, buffer.format.chromaHeight)
);
checkError();
}
}
/**
* Actually draw a frame.
* @param {YUVFrame} buffer - YUV frame buffer object
*/
self.drawFrame = function(buffer) {
var format = buffer.format;
if (canvas.width !== format.displayWidth || canvas.height !== format.displayHeight) {
// Keep the canvas at the right size...
canvas.width = format.displayWidth;
canvas.height = format.displayHeight;
self.clear();
}
if (!program) {
init(buffer);
}
// Set up the rectangle and draw it
//
// Set up geometry
//
buf = gl.createBuffer();
checkError();
gl.bindBuffer(gl.ARRAY_BUFFER, buf);
checkError();
gl.bufferData(gl.ARRAY_BUFFER, rectangle, gl.STATIC_DRAW);
checkError();
var positionLocation = gl.getAttribLocation(program, 'aPosition');
checkError();
gl.enableVertexAttribArray(positionLocation);
checkError();
gl.vertexAttribPointer(positionLocation, 2, gl.FLOAT, false, 0, 0);
checkError();
// Set up the texture geometry...
function setupTexturePosition(varname, texWidth) {
// Warning: assumes that the stride for Cb and Cr is the same size in output pixels
var textureX0 = format.cropLeft / texWidth;
var textureX1 = (format.cropLeft + format.cropWidth) / texWidth;
var textureY0 = (format.cropTop + format.cropHeight) / format.height;
var textureY1 = format.cropTop / format.height;
var textureRectangle = new Float32Array([
textureX0, textureY0,
textureX1, textureY0,
textureX0, textureY1,
textureX0, textureY1,
textureX1, textureY0,
textureX1, textureY1
]);
var texturePositionBuffer = gl.createBuffer();
gl.bindBuffer(gl.ARRAY_BUFFER, texturePositionBuffer);
checkError();
gl.bufferData(gl.ARRAY_BUFFER, textureRectangle, gl.STATIC_DRAW);
checkError();
var texturePositionLocation = gl.getAttribLocation(program, varname);
checkError();
gl.enableVertexAttribArray(texturePositionLocation);
checkError();
gl.vertexAttribPointer(texturePositionLocation, 2, gl.FLOAT, false, 0, 0);
checkError();
}
setupTexturePosition('aLumaPosition', buffer.y.stride);
setupTexturePosition('aChromaPosition', buffer.u.stride * format.width / format.chromaWidth);
// Create the textures...
var textureY = attachTexture(
'uTextureY',
gl.TEXTURE0,
0,
buffer.y.stride,
format.height,
buffer.y.bytes
);
var textureCb = attachTexture(
'uTextureCb',
gl.TEXTURE1,
1,
buffer.u.stride,
format.chromaHeight,
buffer.u.bytes
);
var textureCr = attachTexture(
'uTextureCr',
gl.TEXTURE2,
2,
buffer.v.stride,
format.chromaHeight,
buffer.v.bytes
);
// Aaaaand draw stuff.
gl.drawArrays(gl.TRIANGLES, 0, rectangle.length / 2);
checkError();
};
self.clear = function() {
gl.viewport(0, 0, canvas.width, canvas.height);
gl.clearColor(0.0, 0.0, 0.0, 0.0);
gl.clear(gl.COLOR_BUFFER_BIT);
};
self.clear();
return self;
}
// For Windows; luminance and alpha textures are ssllooww to upload,
// so we pack into RGBA and unpack in the shaders.
//
// This seems to affect all browsers on Windows, probably due to fun
// mismatches between GL and D3D.
WebGLFrameSink.stripe = (function() {
if (navigator.userAgent.indexOf('Windows') !== -1) {
return true;
}
return false;
})();
/**
* Static function to check if WebGL will be available with appropriate features.
*
* @returns {boolean} - true if available
*/
WebGLFrameSink.isAvailable = function() {
var canvas = document.createElement('canvas'),
gl;
canvas.width = 1;
canvas.height = 1;
var options = {
// Still dithering on whether to use this.
// Recommend avoiding it, as it's overly conservative
//failIfMajorPerformanceCaveat: true
};
try {
gl = canvas.getContext('webgl', options) || canvas.getContext('experimental-webgl', options);
} catch (e) {
return false;
}
if (gl) {
var register = gl.TEXTURE0,
width = 4,
height = 4,
texture = gl.createTexture(),
data = new Uint8Array(width * height),
texWidth = WebGLFrameSink.stripe ? (width / 4) : width,
format = WebGLFrameSink.stripe ? gl.RGBA : gl.LUMINANCE,
filter = WebGLFrameSink.stripe ? gl.NEAREST : gl.LINEAR;
gl.activeTexture(register);
gl.bindTexture(gl.TEXTURE_2D, texture);
gl.texParameteri(gl.TEXTURE_2D, gl.TEXTURE_WRAP_S, gl.CLAMP_TO_EDGE);
gl.texParameteri(gl.TEXTURE_2D, gl.TEXTURE_WRAP_T, gl.CLAMP_TO_EDGE);
gl.texParameteri(gl.TEXTURE_2D, gl.TEXTURE_MIN_FILTER, filter);
gl.texParameteri(gl.TEXTURE_2D, gl.TEXTURE_MAG_FILTER, filter);
gl.texImage2D(
gl.TEXTURE_2D,
0, // mip level
format, // internal format
texWidth,
height,
0, // border
format, // format
gl.UNSIGNED_BYTE, //type
data // data!
);
var err = gl.getError();
if (err) {
// Doesn't support luminance textures?
return false;
} else {
return true;
}
} else {
return false;
}
};
WebGLFrameSink.prototype = Object.create(FrameSink.prototype);
module.exports = WebGLFrameSink;
})();
/***/ }),
/* 18 */
/***/ (function(module, exports) {
module.exports = {
vertex: "attribute vec2 aPosition;\nattribute vec2 aLumaPosition;\nattribute vec2 aChromaPosition;\nvarying vec2 vLumaPosition;\nvarying vec2 vChromaPosition;\nvoid main() {\n gl_Position = vec4(aPosition, 0, 1);\n vLumaPosition = aLumaPosition;\n vChromaPosition = aChromaPosition;\n}\n",
fragment: "// inspired by https://github.com/mbebenita/Broadway/blob/master/Player/canvas.js\n\nprecision mediump float;\nuniform sampler2D uTextureY;\nuniform sampler2D uTextureCb;\nuniform sampler2D uTextureCr;\nvarying vec2 vLumaPosition;\nvarying vec2 vChromaPosition;\nvoid main() {\n // Y, Cb, and Cr planes are uploaded as LUMINANCE textures.\n float fY = texture2D(uTextureY, vLumaPosition).x;\n float fCb = texture2D(uTextureCb, vChromaPosition).x;\n float fCr = texture2D(uTextureCr, vChromaPosition).x;\n\n // Premultipy the Y...\n float fYmul = fY * 1.1643828125;\n\n // And convert that to RGB!\n gl_FragColor = vec4(\n fYmul + 1.59602734375 * fCr - 0.87078515625,\n fYmul - 0.39176171875 * fCb - 0.81296875 * fCr + 0.52959375,\n fYmul + 2.017234375 * fCb - 1.081390625,\n 1\n );\n}\n",
fragmentStripe: "// inspired by https://github.com/mbebenita/Broadway/blob/master/Player/canvas.js\n// extra 'stripe' texture fiddling to work around IE 11's poor performance on gl.LUMINANCE and gl.ALPHA textures\n\nprecision mediump float;\nuniform sampler2D uStripeLuma;\nuniform sampler2D uStripeChroma;\nuniform sampler2D uTextureY;\nuniform sampler2D uTextureCb;\nuniform sampler2D uTextureCr;\nvarying vec2 vLumaPosition;\nvarying vec2 vChromaPosition;\nvoid main() {\n // Y, Cb, and Cr planes are mapped into a pseudo-RGBA texture\n // so we can upload them without expanding the bytes on IE 11\n // which doesn\\'t allow LUMINANCE or ALPHA textures.\n // The stripe textures mark which channel to keep for each pixel.\n vec4 vStripeLuma = texture2D(uStripeLuma, vLumaPosition);\n vec4 vStripeChroma = texture2D(uStripeChroma, vChromaPosition);\n\n // Each texture extraction will contain the relevant value in one\n // channel only.\n vec4 vY = texture2D(uTextureY, vLumaPosition) * vStripeLuma;\n vec4 vCb = texture2D(uTextureCb, vChromaPosition) * vStripeChroma;\n vec4 vCr = texture2D(uTextureCr, vChromaPosition) * vStripeChroma;\n\n // Now assemble that into a YUV vector, and premultipy the Y...\n vec3 YUV = vec3(\n (vY.x + vY.y + vY.z + vY.w) * 1.1643828125,\n (vCb.x + vCb.y + vCb.z + vCb.w),\n (vCr.x + vCr.y + vCr.z + vCr.w)\n );\n // And convert that to RGB!\n gl_FragColor = vec4(\n YUV.x + 1.59602734375 * YUV.z - 0.87078515625,\n YUV.x - 0.39176171875 * YUV.y - 0.81296875 * YUV.z + 0.52959375,\n YUV.x + 2.017234375 * YUV.y - 1.081390625,\n 1\n );\n}\n"
};
/***/ }),
/* 19 */
/***/ (function(module, exports, __webpack_require__) {
"use strict";
var _createClass = function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; }();
function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError("Cannot call a class as a function"); } }
var EventEmitter = __webpack_require__(20);
var CachePool = __webpack_require__(21);
var Backend = __webpack_require__(24);
/**
* @typedef {Object} StreamFileOptions
* @property {string} url - the URL to fetch
* @property {number} chunkSize - max size of each chunked HTTP request / readahead target
* @property {number} cacheSize - max amount of data to keep buffered in memory for seeks
*/
/**
* Utility class for chunked streaming of large files via XMLHttpRequest.
* Provides an abstraction of a seekable input stream, backed by in-memory
* caching, and some convenient promise-based i/o methods.
* @param {StreamFileOptions} options
* @constructor
*/
var StreamFile = function () {
function StreamFile(_ref) {
var _ref$url = _ref.url;
var url = _ref$url === undefined ? '' : _ref$url;
var _ref$chunkSize = _ref.chunkSize;
var chunkSize = _ref$chunkSize === undefined ? 1 * 1024 * 1024 : _ref$chunkSize;
var _ref$cacheSize = _ref.cacheSize;
var cacheSize = _ref$cacheSize === undefined ? 0 : _ref$cacheSize;
_classCallCheck(this, StreamFile);
// InputStream public API
this.length = -1;
this.loaded = false;
this.loading = false;
this.seekable = false;
this.buffering = false;
this.seeking = false;
Object.defineProperties(this, {
/**
* Byte offset of the read head
*/
offset: {
get: function get() {
return this._cache.readOffset;
}
},
/**
* Is the read head at the end of the file?
*/
eof: {
get: function get() {
return this.length === this._cache.readOffset;
}
}
});
// StreamFile public API
this.url = url;
this.headers = {};
// Private
this._cache = new CachePool({
cacheSize: cacheSize
});
this._backend = null;
this._cachever = 0;
this._chunkSize = chunkSize;
}
/**
* Open the file, get metadata, and start buffering some data.
* On success, loaded will become true, headers may be filled out,
* and length may be available.
*
* @returns {Promise}
*/
_createClass(StreamFile, [{
key: 'load',
value: function load() {
var _this = this;
return new Promise(function (resolve, reject) {
if (_this.loading) {
throw new Error('cannot load when loading');
}
if (_this.loaded) {
throw new Error('cannot load when loaded');
}
_this.loading = true;
_this._openBackend().then(function (backend) {
// Save metadata from the first set...
// Beware this._backend may be null already,
// if the first segment was very short!
_this.seekable = backend.seekable;
_this.headers = backend.headers;
_this.length = backend.length;
_this.loaded = true;
_this.loading = false;
resolve();
}).catch(function (err) {
if (err.name !== 'AbortError') {
_this.loading = false;
}
reject(err);
});
});
}
/**
* Create a backend and wait for it to load.
* The returned 'backend' object may be null if there is no data to read.
*
* @returns {Promise}
*/
}, {
key: '_openBackend',
value: function _openBackend() {
var _this2 = this;
return new Promise(function (resolve, reject) {
if (_this2._backend) {
resolve(_this2._backend);
} else if (_this2.eof) {
reject(new Error('cannot open at end of file'));
} else {
var cache = _this2._cache;
var max = _this2._chunkSize;
// Seek forward to the next unread point, up to chunk size
var readable = cache.bytesReadable(max);
var readTail = cache.readOffset + readable;
cache.seekWrite(readTail);
// Did we already cache the entire file?
if (_this2.length >= 0 && readTail >= _this2.length) {
resolve(null);
return;
}
// Do we have space to write within that chunk?
var writable = cache.bytesWritable(max);
if (writable === 0) {
// Nothing to read/write within the current readahead area.
resolve(null);
} else {
(function () {
var backend = _this2._backend = new Backend({
url: _this2.url,
offset: _this2._cache.writeOffset,
length: writable,
cachever: _this2._cachever
});
var oncomplete = null;
var checkOpen = function checkOpen() {
if (backend !== _this2._backend) {
oncomplete();
reject(new Error('invalid state'));
} else {
backend.on('buffer', function (buffer) {
if (backend === _this2._backend) {
_this2._cache.write(buffer);
}
});
backend.on('done', function () {
if (backend === _this2._backend) {
if (_this2.length === -1) {
// save length on those final thingies
_this2.length = _this2._backend.offset + _this2._backend.bytesRead;
}
_this2._backend = null;
}
});
resolve(backend);
}
};
var checkError = function checkError(err) {
if (backend !== _this2._backend) {
reject(new Error('invalid state'));
} else {
_this2._backend = null;
reject(err);
}
};
oncomplete = function oncomplete() {
backend.off('open', checkOpen);
backend.off('error', checkError);
};
backend.on('open', checkOpen);
backend.on('error', checkError);
backend.on('cachever', function () {
_this2._cachever++;
});
backend.load();
})();
}
}
});
}
/**
* If we have empty space within the readahead area and there is not already
* a download backend in place, create one and start it loading in background.
* @returns {Promise}
*/
}, {
key: '_readAhead',
value: function _readAhead() {
var _this3 = this;
return new Promise(function (resolve, reject) {
if (_this3._backend || _this3.eof) {
// do nothing
resolve();
} else {
_this3._openBackend().then(function () {
resolve();
}).catch(function (err) {
reject(err);
});
}
});
}
/**
* Seek the read position to a new location in the file, asynchronously.
* After succesful completion, reads will continue at the new offset.
* May fail due to network problems, invalid input, or bad state.
* @param {number} offset - target byte offset from beginning of file
* @returns {Promise} - resolved when ready to read at the new position
*/
}, {
key: 'seek',
value: function seek(offset) {
var _this4 = this;
return new Promise(function (resolve, reject) {
if (!_this4.loaded || _this4.buffering || _this4.seeking) {
throw new Error('invalid state');
} else if (offset !== (offset | 0) || offset < 0) {
throw new Error('invalid input');
} else if (_this4.length >= 0 && offset > _this4.length) {
throw new Error('seek past end of file');
} else if (!_this4.seekable) {
throw new Error('seek on non-seekable stream');
} else {
if (_this4._backend) {
// @todo if a short seek forward, just keep reading?
_this4.abort();
}
_this4._cache.seekRead(offset);
_this4._cache.seekWrite(offset);
// Fire off a download if necessary.
_this4._readAhead().then(resolve).catch(reject);
}
});
}
/**
* Read up to the requested number of bytes, or until end of file is reached,
* and advance the read head.
*
* May wait on network activity if data is not yet available.
*
* @param {number} nbytes - max number of bytes to read
* @returns {ArrayBuffer} - between 0 and nbytes of data, inclusive
*/
}, {
key: 'read',
value: function read(nbytes) {
var _this5 = this;
return this.buffer(nbytes).then(function (available) {
return _this5.readSync(available);
});
}
/**
* Read up to the requested number of bytes, or however much is available
* in the buffer until the next empty segment, and advance the read head.
*
* Returns immediately.
*
* @param {number} nbytes - max number of bytes to read
* @returns {ArrayBuffer} - between 0 and nbytes of data, inclusive
*/
}, {
key: 'readSync',
value: function readSync(nbytes) {
var available = this.bytesAvailable(nbytes);
var dest = new Uint8Array(available);
var actual = this.readBytes(dest);
if (actual !== available) {
throw new Error('failed to read expected data');
}
return dest.buffer;
}
/**
* Read bytes into destination array until out of buffer or space,
* and advance the read head.
*
* Returns immediately.
*
* @param {dest} Uint8Array - destination byte array
* @returns {number} - count of actual bytes read
*/
}, {
key: 'readBytes',
value: function readBytes(dest) {
if (!this.loaded || this.buffering || this.seeking) {
throw new Error('invalid state');
} else if (!(dest instanceof Uint8Array)) {
throw new Error('invalid input');
}
var nbytes = this._cache.readBytes(dest);
// Trigger readahead if necessary.
this._readAhead();
return nbytes;
}
/**
* Wait until the given number of bytes are available to read, or end of file.
* @param {number} nbytes - max bytes to wait for
* @returns {Promise} - resolved with available byte count when ready
*/
}, {
key: 'buffer',
value: function buffer(nbytes) {
var _this6 = this;
return new Promise(function (resolve, reject) {
if (!_this6.loaded || _this6.buffering || _this6.seeking) {
throw new Error('invalid state');
} else if (nbytes !== (nbytes | 0) || nbytes < 0) {
throw new Error('invalid input');
}
var end = _this6._clampToLength(_this6.offset + nbytes);
var readable = end - _this6.offset;
var canceled = false;
var available = _this6.bytesAvailable(readable);
if (available >= readable) {
// Requested data is immediately available.
resolve(available);
} else {
_this6.buffering = true;
// If we don't already have a backend open, start downloading.
_this6._openBackend().then(function (backend) {
if (backend) {
return backend.bufferToOffset(end).then(function () {
// We might have to roll over to another download,
// so loop back around!
_this6.buffering = false;
return _this6.buffer(nbytes);
});
} else {
// No more data to read.
return Promise.resolve(available);
}
}).then(function (available) {
_this6.buffering = false;
resolve(available);
}).catch(function (err) {
if (err.name !== 'AbortError') {
// was already set synchronously; avoid stomping on old promise
_this6.buffering = false;
}
reject(err);
});
}
});
}
/**
* Number of bytes available to read immediately from the current offset.
* This is the max number of bytes that can be returned from a read() call.
* @returns {boolean}
*/
}, {
key: 'bytesAvailable',
value: function bytesAvailable() {
var max = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : Infinity;
return this._cache.bytesReadable(max);
}
/**
* Abort any currently running downloads and operations.
*/
}, {
key: 'abort',
value: function abort() {
// Clear state synchronously, so can immediately launch new i/o...
if (this.loading) {
this.loading = false;
}
if (this.buffering) {
this.buffering = false;
}
if (this.seeking) {
this.seeking = false;
}
// Abort any active backend request...
if (this._backend) {
this._backend.abort();
this._backend = null;
}
}
/**
* Return an array of byte ranges that are buffered.
* Each range is a two-element array of start and end.
* @returns {Array<Array<number>>}
*/
}, {
key: 'getBufferedRanges',
value: function getBufferedRanges() {
return this._cache.ranges();
}
// ------
// private methods
// ------
}, {
key: '_clampToLength',
value: function _clampToLength(offset) {
if (this.length < 0) {
return offset;
} else {
return Math.min(this.length, offset);
}
}
}]);
return StreamFile;
}();
module.exports = StreamFile;
/***/ }),
/* 20 */
/***/ (function(module, exports) {
"use strict";
// just the bits we need
var _createClass = function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; }();
function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError("Cannot call a class as a function"); } }
var TinyEvents = function () {
function TinyEvents() {
_classCallCheck(this, TinyEvents);
this._e = {};
}
_createClass(TinyEvents, [{
key: "on",
value: function on(name, handler) {
(this._e[name] || (this._e[name] = [])).push(handler);
}
}, {
key: "off",
value: function off(name, handler) {
var l = this._e[name] || [];
var i = l.indexOf(handler);
if (handler >= 0) {
l.splice(i, 1);
}
}
}, {
key: "emit",
value: function emit(name, arg) {
(this._e[name] || []).slice().forEach(function (f) {
return f(arg);
});
}
}]);
return TinyEvents;
}();
module.exports = TinyEvents;
/***/ }),
/* 21 */
/***/ (function(module, exports, __webpack_require__) {
"use strict";
module.exports = __webpack_require__(22);
/***/ }),
/* 22 */
/***/ (function(module, exports, __webpack_require__) {
"use strict";
var _createClass = function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; }();
function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError("Cannot call a class as a function"); } }
var CacheItem = __webpack_require__(23);
/**
* Seekable, readable, writable buffer cache to represent a file.
* @todo add max cache size and LRU cache expiration
*
* Internally, will always contain entries from 0 to some given out point.
* Each item may either contain data, or be empty.
* Empty ranges cannot be copied out via read(), non-empty ranges can.
* Empty ranges can be filled up with write(), non-empty ranges cannot.
*
* Internal invariants:
* - head and tail are always present, may be same for empty
* - tail item is always empty/eof
* - non-empty items are never 0 bytes
* - adjacent list items are always continguous
* - empty items are never adjacent to each other
*/
var CachePool = function () {
function CachePool() {
var _ref = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : {};
var _ref$cacheSize = _ref.cacheSize;
var cacheSize = _ref$cacheSize === undefined ? 0 : _ref$cacheSize;
_classCallCheck(this, CachePool);
var eof = new CacheItem({ eof: true });
this.head = eof;
this.tail = eof;
this.readOffset = 0;
this.readCursor = eof;
this.writeOffset = 0;
this.writeCursor = eof;
this.cacheSize = cacheSize;
}
/**
* Is the read cursor at the end of the file?
*/
_createClass(CachePool, [{
key: 'bytesReadable',
/**
* Count how many bytes are available from the given offset.
* @param {number} max - optional maximum to read
* @returns {number} 0 or more
*/
value: function bytesReadable() {
var max = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : Infinity;
var offset = this.readOffset;
var cursor = this.readCursor;
var last = cursor.last(function (item) {
return !item.empty && item.start <= offset + max;
});
if (last) {
return Math.min(max, last.end - offset);
}
return 0;
}
/**
* Count how many bytes are available to write.
* @param {number} max - optional maximum to write
* @returns {number} 0 or more, or +Infinity
*/
}, {
key: 'bytesWritable',
value: function bytesWritable() {
var max = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : Infinity;
var offset = this.writeOffset;
var cursor = this.writeCursor;
if (cursor.eof) {
return max;
}
var last = cursor.last(function (item) {
return item.empty && item.start <= offset + max;
});
if (last) {
return Math.min(max, last.end - offset);
}
return 0;
}
/**
* Move the read head to a given offset. The read head can move beyond the
* currently known end of the file, but cannot move before 0.
* @param {number} offset - bytes from beginning of virtual file to read from
*/
}, {
key: 'seekRead',
value: function seekRead(offset) {
var target = this.head.first(function (item) {
return item.contains(offset);
});
if (!target) {
throw new Error('read seek out of range');
}
this.readOffset = offset;
this.readCursor = target;
}
/**
* Move the write head to a given offset. The write head can move beyond the
* currently known end of the file, but cannot move before 0.
* @param {number} offset - bytes from beginning of virtual file to write to
*/
}, {
key: 'seekWrite',
value: function seekWrite(offset) {
var target = this.head.first(function (item) {
return item.contains(offset);
});
if (!target) {
throw new Error('write seek out of range');
}
this.writeOffset = offset;
this.writeCursor = target;
}
/**
* Read up to the requested number of bytes, or however much is available
* in the buffer until the next empty segment, and advance the read head.
*
* Returns immediately.
*
* @param {Uint8Array} dest - destination array to read to
* @returns {number} - count of bytes actually read
*/
}, {
key: 'readBytes',
value: function readBytes(dest) {
var nbytes = dest.byteLength;
var len = this.bytesReadable(nbytes);
var start = this.readOffset;
var end = start + len;
var readHead = start;
var writeHead = 0;
for (var item = this.readCursor; item; item = item.next) {
if (item.empty) {
break;
}
if (item.start >= end) {
break;
}
var readTail = Math.min(end, item.end);
var chunk = dest.subarray(readHead - start, readTail - start);
item.readBytes(chunk, readHead, readTail);
readHead = readTail;
}
this.readOffset = readHead;
this.readCursor = this.readCursor.first(function (item) {
return item.contains(readHead);
});
return len;
}
/**
* Write a data buffer at the write head and advance the write head.
* The data must fit in the available empty space in the buffer cache.
* @param {ArrayBuffer|String} buffer
*/
}, {
key: 'write',
value: function write(buffer) {
var item = this.bufferItem(buffer);
var cursor = this.writeCursor;
if (!cursor.empty) {
throw new Error('write cursor not empty');
}
if (!cursor.contains(item.end) && cursor.end !== item.end) {
throw new Error('write cursor too small');
}
if (cursor.start < item.start) {
this.split(cursor, item.start);
cursor = this.writeCursor;
}
if (item.end < cursor.end || cursor.eof) {
this.split(cursor, item.end);
cursor = this.writeCursor;
}
this.splice(cursor, cursor, item, item);
this.writeOffset = item.end;
this.writeCursor = item.next;
this.gc();
}
}, {
key: 'bufferItem',
value: function bufferItem(buffer) {
if (buffer instanceof ArrayBuffer) {
return new CacheItem({
start: this.writeOffset,
end: this.writeOffset + buffer.byteLength,
buffer: buffer
});
} else if (typeof buffer === 'string') {
return new CacheItem({
start: this.writeOffset,
end: this.writeOffset + buffer.length,
string: buffer
});
} else {
throw new Error('invalid input to write');
}
}
}, {
key: 'split',
value: function split(oldItem, offset) {
var items = oldItem.split(offset);
this.splice(oldItem, oldItem, items[0], items[1]);
}
/**
* Return an array of arrays of consolidated cached ranges
*/
}, {
key: 'ranges',
value: function ranges() {
var ranges = [];
for (var item = this.head; item; item = item.next) {
if (item.empty) {
continue;
}
var start = item;
item = item.last(function (i) {
return !i.empty;
});
ranges.push([start.start, item.end]);
}
return ranges;
}
}, {
key: 'gc',
value: function gc() {
// Simple gc: look at anything not between read head and write head,
// and discard the oldest items until we have room
var cachedBytes = 0;
var candidates = [];
for (var item = this.head; item; item = item.next) {
if (!item.empty) {
cachedBytes += item.length;
if (item.end < this.readOffset || item.start > this.readOffset + this.chunkSize) {
// Not in the 'hot' readahead range
candidates.push(item);
}
}
}
if (cachedBytes > this.cacheSize) {
candidates.sort(function (a, b) {
return a.timestamp - b.timestamp;
});
for (var i = 0; i < candidates.length; i++) {
var _item = candidates[i];
if (cachedBytes <= this.cacheSize) {
break;
}
this.remove(_item);
cachedBytes -= _item.length;
}
}
}
}, {
key: 'remove',
value: function remove(item) {
var replacement = new CacheItem({
start: item.start,
end: item.end
});
this.splice(item, item, replacement, replacement);
item = replacement;
// Consolidate adjacent ranges
if (item.prev && item.prev.empty) {
item = this.consolidate(item.prev);
}
if (item.next && item.next.empty && !item.next.eof) {
item = this.consolidate(item);
}
if (item.start === 0) {
this.head = item;
}
}
}, {
key: 'consolidate',
value: function consolidate(first) {
var last = first.last(function (item) {
return item.empty && !item.eof;
});
var replacement = new CacheItem({
start: first.start,
end: last.end
});
this.splice(first, last, replacement, replacement);
return replacement;
}
}, {
key: 'splice',
value: function splice(oldHead, oldTail, newHead, newTail) {
var _this = this;
if (oldHead.start !== newHead.start) {
throw new Error('invalid splice head');
}
if (oldTail.end !== newTail.end) {
if (oldTail.eof && newTail.eof) {
// only eof is expandable
} else {
throw new Error('invalid splice tail');
}
}
var prev = oldHead.prev;
var next = oldTail.next;
oldHead.prev = null;
oldTail.next = null;
if (prev) {
prev.next = newHead;
newHead.prev = prev;
}
if (next) {
next.prev = newTail;
newTail.next = next;
}
if (oldHead === this.head) {
this.head = newHead;
}
if (oldTail === this.tail) {
this.tail = newTail;
}
this.readCursor = this.head.first(function (item) {
return item.contains(_this.readOffset);
});
this.writeCursor = this.head.first(function (item) {
return item.contains(_this.writeOffset);
});
}
}, {
key: 'eof',
get: function get() {
return this.readCursor.eof;
}
}]);
return CachePool;
}();
module.exports = CachePool;
/***/ }),
/* 23 */
/***/ (function(module, exports) {
"use strict";
/**
* Double-linked list cache items
*/
var _createClass = function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; }();
function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError("Cannot call a class as a function"); } }
var CacheItem = function () {
function CacheItem() {
var _ref = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : {};
var _ref$buffer = _ref.buffer;
var buffer = _ref$buffer === undefined ? undefined : _ref$buffer;
var _ref$string = _ref.string;
var string = _ref$string === undefined ? undefined : _ref$string;
var _ref$start = _ref.start;
var start = _ref$start === undefined ? 0 : _ref$start;
var _ref$end = _ref.end;
var end = _ref$end === undefined ? start + (buffer ? buffer.byteLength : string ? string.length : 0) : _ref$end;
var _ref$prev = _ref.prev;
var prev = _ref$prev === undefined ? null : _ref$prev;
var _ref$next = _ref.next;
var next = _ref$next === undefined ? null : _ref$next;
var _ref$eof = _ref.eof;
var eof = _ref$eof === undefined ? false : _ref$eof;
var _ref$empty = _ref.empty;
var empty = _ref$empty === undefined ? !(buffer || string) : _ref$empty;
var _ref$timestamp = _ref.timestamp;
var timestamp = _ref$timestamp === undefined ? Date.now() : _ref$timestamp;
_classCallCheck(this, CacheItem);
this.start = start;
this.end = end;
this.prev = prev;
this.next = next;
this.eof = eof;
this.empty = empty;
this.timestamp = timestamp;
this.buffer = buffer;
this.string = string;
Object.defineProperty(this, 'length', {
get: function get() {
return this.end - this.start;
}
});
}
/**
* True if this cache item contains the given byte offset.
* False if outside.
*/
_createClass(CacheItem, [{
key: 'contains',
value: function contains(offset) {
return offset >= this.start && (offset < this.end || this.eof);
}
}, {
key: 'readBytes',
value: function readBytes(dest, start, end) {
var readHead = start - this.start;
var len = end - start;
if (this.buffer) {
var sourceBytes = new Uint8Array(this.buffer, readHead, len);
dest.set(sourceBytes);
} else if (this.string) {
var chunk = this.string;
for (var i = 0; i < len; i++) {
dest[i] = chunk.charCodeAt(readHead + i);
}
} else {
throw new Error('invalid state');
}
this.timestamp = Date.now();
}
}, {
key: 'split',
value: function split(offset) {
if (!this.empty || !this.contains(offset)) {
throw new Error('invalid split');
}
var a = new CacheItem({
start: this.start,
end: offset
});
var b = new CacheItem({
start: offset,
end: this.eof ? offset : this.end,
eof: this.eof
});
a.next = b;
b.prev = a;
return [a, b];
}
/**
* Iterate forwards, returning the first element matching the callback.
* @param {function} callback - return true for a match on item
* @returns {CacheItem|null} - matching item or null if none found
*/
}, {
key: 'first',
value: function first(callback) {
for (var item = this; item; item = item.next) {
if (callback(item)) {
return item;
}
}
return null;
}
/**
* Iterate forwards, returning the last element matching the callback before
* reaching one that doesn't match or we find the end.
* @param {function} callback - return true for a match on item
* @returns {CacheItem|null} - matching item or null if none found
*/
}, {
key: 'last',
value: function last(callback) {
var last = null;
for (var item = this; item; item = item.next) {
if (!callback(item)) {
break;
}
last = item;
}
return last;
}
}]);
return CacheItem;
}();
module.exports = CacheItem;
/***/ }),
/* 24 */
/***/ (function(module, exports, __webpack_require__) {
"use strict";
var MozChunkedBackend = __webpack_require__(25);
var MSStreamBackend = __webpack_require__(28);
var BinaryStringBackend = __webpack_require__(29);
function autoselect() {
if (MozChunkedBackend.supported()) {
return MozChunkedBackend;
} else if (BinaryStringBackend.supported()) {
return BinaryStringBackend;
} else if (MSStreamBackend.supported()) {
return MSStreamBackend;
} else {
return null;
}
}
var backendClass = null;
function instantiate(options) {
if (!backendClass) {
backendClass = autoselect();
}
if (!backendClass) {
throw new Error('No supported backend class');
}
return new backendClass(options);
}
module.exports = instantiate;
/***/ }),
/* 25 */
/***/ (function(module, exports, __webpack_require__) {
"use strict";
var _createClass = function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; }();
var _get = function get(object, property, receiver) { if (object === null) object = Function.prototype; var desc = Object.getOwnPropertyDescriptor(object, property); if (desc === undefined) { var parent = Object.getPrototypeOf(object); if (parent === null) { return undefined; } else { return get(parent, property, receiver); } } else if ("value" in desc) { return desc.value; } else { var getter = desc.get; if (getter === undefined) { return undefined; } return getter.call(receiver); } };
function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError("Cannot call a class as a function"); } }
function _possibleConstructorReturn(self, call) { if (!self) { throw new ReferenceError("this hasn't been initialised - super() hasn't been called"); } return call && (typeof call === "object" || typeof call === "function") ? call : self; }
function _inherits(subClass, superClass) { if (typeof superClass !== "function" && superClass !== null) { throw new TypeError("Super expression must either be null or a function, not " + typeof superClass); } subClass.prototype = Object.create(superClass && superClass.prototype, { constructor: { value: subClass, enumerable: false, writable: true, configurable: true } }); if (superClass) Object.setPrototypeOf ? Object.setPrototypeOf(subClass, superClass) : subClass.__proto__ = superClass; }
var DownloadBackend = __webpack_require__(26);
var type = 'moz-chunked-arraybuffer';
var MozChunkedBackend = function (_DownloadBackend) {
_inherits(MozChunkedBackend, _DownloadBackend);
function MozChunkedBackend() {
_classCallCheck(this, MozChunkedBackend);
return _possibleConstructorReturn(this, (MozChunkedBackend.__proto__ || Object.getPrototypeOf(MozChunkedBackend)).apply(this, arguments));
}
_createClass(MozChunkedBackend, [{
key: 'initXHR',
value: function initXHR() {
_get(MozChunkedBackend.prototype.__proto__ || Object.getPrototypeOf(MozChunkedBackend.prototype), 'initXHR', this).call(this);
this.xhr.responseType = type;
}
}, {
key: 'onXHRProgress',
value: function onXHRProgress() {
var buffer = this.xhr.response;
this.bytesRead += buffer.byteLength;
this.emit('buffer', buffer);
}
}]);
return MozChunkedBackend;
}(DownloadBackend);
MozChunkedBackend.supported = function () {
try {
var xhr = new XMLHttpRequest();
xhr.responseType = type;
return xhr.responseType === type;
} catch (e) {
return false;
}
};
module.exports = MozChunkedBackend;
/***/ }),
/* 26 */
/***/ (function(module, exports, __webpack_require__) {
"use strict";
var _createClass = function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; }();
var _get = function get(object, property, receiver) { if (object === null) object = Function.prototype; var desc = Object.getOwnPropertyDescriptor(object, property); if (desc === undefined) { var parent = Object.getPrototypeOf(object); if (parent === null) { return undefined; } else { return get(parent, property, receiver); } } else if ("value" in desc) { return desc.value; } else { var getter = desc.get; if (getter === undefined) { return undefined; } return getter.call(receiver); } };
function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError("Cannot call a class as a function"); } }
function _possibleConstructorReturn(self, call) { if (!self) { throw new ReferenceError("this hasn't been initialised - super() hasn't been called"); } return call && (typeof call === "object" || typeof call === "function") ? call : self; }
function _inherits(subClass, superClass) { if (typeof superClass !== "function" && superClass !== null) { throw new TypeError("Super expression must either be null or a function, not " + typeof superClass); } subClass.prototype = Object.create(superClass && superClass.prototype, { constructor: { value: subClass, enumerable: false, writable: true, configurable: true } }); if (superClass) Object.setPrototypeOf ? Object.setPrototypeOf(subClass, superClass) : subClass.__proto__ = superClass; }
var Backend = __webpack_require__(27);
/**
* Backend for progressive downloading.
* Subclasses handle details of strings/buffers.
*/
var DownloadBackend = function (_Backend) {
_inherits(DownloadBackend, _Backend);
function DownloadBackend() {
_classCallCheck(this, DownloadBackend);
return _possibleConstructorReturn(this, (DownloadBackend.__proto__ || Object.getPrototypeOf(DownloadBackend)).apply(this, arguments));
}
_createClass(DownloadBackend, [{
key: 'bufferToOffset',
value: function bufferToOffset(end) {
var _this2 = this;
return new Promise(function (resolve, reject) {
if (_this2.eof || _this2.offset >= end) {
resolve();
} else {
(function () {
var oncomplete = null;
_this2._onAbort = function (err) {
oncomplete();
reject(reason);
};
var checkBuffer = function checkBuffer() {
if (_this2.offset >= end && !_this2.eof) {
oncomplete();
resolve();
}
};
var checkDone = function checkDone() {
oncomplete();
resolve();
};
var checkError = function checkError() {
oncomplete();
reject(new Error('error streaming'));
};
oncomplete = function oncomplete() {
_this2.buffering = false;
_this2.off('buffer', checkBuffer);
_this2.off('done', checkDone);
_this2.off('error', checkError);
_this2._onAbort = null;
};
_this2.buffering = true;
_this2.on('buffer', checkBuffer);
_this2.on('done', checkDone);
_this2.on('error', checkError);
})();
}
});
}
}, {
key: 'initXHR',
value: function initXHR() {
_get(DownloadBackend.prototype.__proto__ || Object.getPrototypeOf(DownloadBackend.prototype), 'initXHR', this).call(this);
}
}, {
key: 'onXHRStart',
value: function onXHRStart() {
var _this3 = this;
// Event handlers to drive output
this.xhr.addEventListener('progress', function () {
return _this3.onXHRProgress();
});
this.xhr.addEventListener('error', function () {
return _this3.onXHRError();
});
this.xhr.addEventListener('load', function () {
return _this3.onXHRLoad();
});
this.emit('open');
}
}, {
key: 'onXHRProgress',
value: function onXHRProgress() {
throw new Error('abstract');
}
}, {
key: 'onXHRError',
value: function onXHRError() {
this.emit('error');
}
}, {
key: 'onXHRLoad',
value: function onXHRLoad() {
this.eof = true;
this.emit('done');
}
}]);
return DownloadBackend;
}(Backend);
module.exports = DownloadBackend;
/***/ }),
/* 27 */
/***/ (function(module, exports, __webpack_require__) {
"use strict";
var _createClass = function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; }();
function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError("Cannot call a class as a function"); } }
function _possibleConstructorReturn(self, call) { if (!self) { throw new ReferenceError("this hasn't been initialised - super() hasn't been called"); } return call && (typeof call === "object" || typeof call === "function") ? call : self; }
function _inherits(subClass, superClass) { if (typeof superClass !== "function" && superClass !== null) { throw new TypeError("Super expression must either be null or a function, not " + typeof superClass); } subClass.prototype = Object.create(superClass && superClass.prototype, { constructor: { value: subClass, enumerable: false, writable: true, configurable: true } }); if (superClass) Object.setPrototypeOf ? Object.setPrototypeOf(subClass, superClass) : subClass.__proto__ = superClass; }
var TinyEvents = __webpack_require__(20);
/**
* Extract the file's total length from the XHR returned headers.
* @returns {number} - byte length or -1
* @access private
*/
function getXHRLength(xhr) {
if (xhr.status == 206) {
return getXHRRangeTotal(xhr);
} else {
var contentLength = xhr.getResponseHeader('Content-Length');
if (contentLength === null || contentLength === '') {
// Unknown file length... maybe streaming live?
return -1;
} else {
return parseInt(contentLength, 10);
}
}
}
/**
* Extract the range chunk info from the XHR returned headers.
* @returns {Array} - byte length or -1
* @access private
*/
function getXHRRangeMatches(xhr) {
// Note Content-Range must be whitelisted for CORS requests
var contentRange = xhr.getResponseHeader('Content-Range');
return contentRange && contentRange.match(/^bytes (\d+)-(\d+)\/(\d+)/);
}
/**
* Extract the chunk start position from the XHR returned headers.
* @returns {number} - byte position or 0
* @access private
*/
function getXHRRangeStart(xhr) {
var matches = getXHRRangeMatches(xhr);
if (matches) {
return parseInt(matches[1], 10);
} else {
return 0;
}
}
/**
* Extract the file's total length from the XHR returned headers.
* @returns {number} - byte length or -1
* @access private
*/
function getXHRRangeTotal(xhr) {
var matches = getXHRRangeMatches(xhr);
if (matches) {
return parseInt(matches[3], 10);
} else {
return -1;
}
}
/**
* Record the HTTP headers from the initial request, in case some are useful.
* @returns {Object} map of headers
* @access private
*/
function getXHRHeaders(xhr) {
var headers = {};
var headerLines = xhr.getAllResponseHeaders().split(/\n/);
headerLines.forEach(function (line) {
var bits = line.split(/:\s*/, 2);
if (bits.length > 1) {
headers[bits[0].toLowerCase()] = bits[1];
}
});
return headers;
}
/**
* Represents a single HTTP request pass through part of a URL.
*
* Subclasses handle details of chunking/strings/streams and provide
* a unified internal API.
*
* Events sent:
* - 'open' - called when file metadata ready
* - 'buffer' - passes a BufferSegment in with some new data
* - 'done' - called at end of file
* - 'error' - called in case of error
* - 'cachever' - triggered when old Safari caching bug found
*/
var Backend = function (_TinyEvents) {
_inherits(Backend, _TinyEvents);
function Backend(_ref) {
var url = _ref.url;
var offset = _ref.offset;
var length = _ref.length;
var _ref$cachever = _ref.cachever;
var cachever = _ref$cachever === undefined ? 0 : _ref$cachever;
_classCallCheck(this, Backend);
var _this = _possibleConstructorReturn(this, (Backend.__proto__ || Object.getPrototypeOf(Backend)).call(this));
_this.url = url;
_this.offset = offset;
_this.length = length;
_this.cachever = cachever;
_this.loaded = false;
_this.seekable = false;
_this.headers = {};
_this.eof = false;
_this.bytesRead = 0;
_this.xhr = new XMLHttpRequest();
return _this;
}
_createClass(Backend, [{
key: 'load',
value: function load() {
var _this2 = this;
return new Promise(function (resolve, reject) {
var oncomplete = null;
_this2._onAbort = function (err) {
oncomplete();
reject(err);
};
var checkOpen = function checkOpen() {
// There doesn't seem to be a good match for readyState 2 on the XHR2 events model.
if (_this2.xhr.readyState == 2) {
if (_this2.xhr.status == 206) {
// Partial content -- we are streamable
var foundPosition = getXHRRangeStart(_this2.xhr);
if (_this2.offset != foundPosition) {
//
// Safari sometimes messes up and gives us the wrong chunk.
// Seems to be a general problem with Safari and cached XHR ranges.
//
// Interestingly, it allows you to request _later_ ranges successfully,
// but when requesting _earlier_ ranges it returns the latest one retrieved.
// So we only need to update the cache-buster when we rewind and actually
// get an incorrect range.
//
// https://bugs.webkit.org/show_bug.cgi?id=82672
//
console.log('Expected start at ' + _this2.offset + ' but got ' + foundPosition + '; working around Safari range caching bug: https://bugs.webkit.org/show_bug.cgi?id=82672');
_this2.cachever++;
_this2.emit('cachever');
_this2.abort();
oncomplete();
_this2.load().then(resolve).catch(reject);
return;
}
_this2.seekable = true;
}
if (_this2.xhr.status >= 200 && _this2.xhr.status < 300) {
_this2.length = getXHRLength(_this2.xhr);
_this2.headers = getXHRHeaders(_this2.xhr);
_this2.onXHRStart();
if (_this2.xhr.responseURL &&
_this2.xhr.responseURL.substr(0, 5) == 'data:') {
_this2.seekable = true;
}
} else {
oncomplete();
reject(new Error('HTTP error ' + _this2.xhr.status));
}
}
};
var checkError = function checkError() {
oncomplete();
reject(new Error('network error'));
};
var checkBackendOpen = function checkBackendOpen() {
oncomplete();
resolve();
};
oncomplete = function oncomplete() {
_this2.xhr.removeEventListener('readystatechange', checkOpen);
_this2.xhr.removeEventListener('error', checkError);
_this2.off('open', checkBackendOpen);
_this2._onAbort = null;
};
_this2.initXHR();
// Events for the open promise
_this2.xhr.addEventListener('readystatechange', checkOpen);
_this2.xhr.addEventListener('error', checkError);
_this2.on('open', checkBackendOpen);
_this2.xhr.send();
});
}
/**
* Wait until we download up to the given offset, reach eof, or error out.
* Actual data will be returned via 'buffer' events in the meantime.
*
* Note that MSStream backend will need this to be called explicitly,
* while the other backends download progressively even without a call.
*/
}, {
key: 'bufferToOffset',
value: function bufferToOffset(end) {
return Promise.reject(new Error('abstract'));
}
}, {
key: 'abort',
value: function abort() {
this.xhr.abort();
if (this._onAbort) {
var onAbort = this._onAbort;
this._onAbort = null;
var err = new Error('Aborted');
err.name = 'AbortError';
onAbort(err);
}
}
// ---------------
// Private methods
// ---------------
}, {
key: 'initXHR',
value: function initXHR() {
var getUrl = this.url;
if (this.cachever) {
//
// Safari sometimes messes up and gives us the wrong chunk.
// Seems to be a general problem with Safari and cached XHR ranges.
//
// Interestingly, it allows you to request _later_ ranges successfully,
// but when requesting _earlier_ ranges it returns the latest one retrieved.
// So we only need to update the cache-buster when we rewind.
//
// https://bugs.webkit.org/show_bug.cgi?id=82672
//
getUrl += '?buggy_cachever=' + this.cachever;
}
this.xhr.open("GET", getUrl);
var range = null;
if (this.offset || this.length) {
range = 'bytes=' + this.offset + '-';
}
if (this.length) {
range += this.offset + this.length - 1;
}
if (range !== null) {
this.xhr.setRequestHeader('Range', range);
}
}
}, {
key: 'onXHRStart',
value: function onXHRStart() {
throw new Error('abstract');
}
}]);
return Backend;
}(TinyEvents);
module.exports = Backend;
/***/ }),
/* 28 */
/***/ (function(module, exports, __webpack_require__) {
"use strict";
var _createClass = function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; }();
var _get = function get(object, property, receiver) { if (object === null) object = Function.prototype; var desc = Object.getOwnPropertyDescriptor(object, property); if (desc === undefined) { var parent = Object.getPrototypeOf(object); if (parent === null) { return undefined; } else { return get(parent, property, receiver); } } else if ("value" in desc) { return desc.value; } else { var getter = desc.get; if (getter === undefined) { return undefined; } return getter.call(receiver); } };
function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError("Cannot call a class as a function"); } }
function _possibleConstructorReturn(self, call) { if (!self) { throw new ReferenceError("this hasn't been initialised - super() hasn't been called"); } return call && (typeof call === "object" || typeof call === "function") ? call : self; }
function _inherits(subClass, superClass) { if (typeof superClass !== "function" && superClass !== null) { throw new TypeError("Super expression must either be null or a function, not " + typeof superClass); } subClass.prototype = Object.create(superClass && superClass.prototype, { constructor: { value: subClass, enumerable: false, writable: true, configurable: true } }); if (superClass) Object.setPrototypeOf ? Object.setPrototypeOf(subClass, superClass) : subClass.__proto__ = superClass; }
var Backend = __webpack_require__(27);
var type = 'ms-stream';
var MSStreamBackend = function (_Backend) {
_inherits(MSStreamBackend, _Backend);
function MSStreamBackend(options) {
_classCallCheck(this, MSStreamBackend);
var _this = _possibleConstructorReturn(this, (MSStreamBackend.__proto__ || Object.getPrototypeOf(MSStreamBackend)).call(this, options));
_this.stream = null;
_this.streamReader = null;
return _this;
}
_createClass(MSStreamBackend, [{
key: 'initXHR',
value: function initXHR() {
_get(MSStreamBackend.prototype.__proto__ || Object.getPrototypeOf(MSStreamBackend.prototype), 'initXHR', this).call(this);
this.xhr.responseType = type;
}
}, {
key: 'onXHRStart',
value: function onXHRStart() {
var _this2 = this;
var checkProgress = function checkProgress() {
if (_this2.xhr.readyState === 3) {
// We don't get the stream until readyState 3, and it's gone after load.
_this2.stream = _this2.xhr.response;
_this2.xhr.removeEventListener('readystatechange', checkProgress);
_this2.emit('open');
}
};
this.xhr.addEventListener('readystatechange', checkProgress);
}
}, {
key: 'waitForStream',
value: function waitForStream() {
var _this3 = this;
return new Promise(function (resolve, reject) {
if (_this3.stream) {
resolve(_this3.stream);
} else {
(function () {
var oncomplete = null;
_this3._onAbort = function (err) {
oncomplete();
reject(err);
};
var checkStart = function checkStart() {
resolve(_this3.stream);
};
oncomplete = function oncomplete() {
_this3.off('open', checkStart);
_this3._onAbort = null;
};
_this3.on('open', checkStart);
})();
}
});
}
/**
* Trigger further download of bytes
* @returns {Promise}
*/
}, {
key: 'bufferToOffset',
value: function bufferToOffset(end) {
var _this4 = this;
return this.waitForStream().then(function (stream) {
return new Promise(function (resolve, reject) {
if (_this4.streamReader) {
throw new Error('cannot trigger read when reading');
}
if (_this4.offset >= end || _this4.eof) {
resolve();
} else {
var nbytes = end - _this4.offset;
_this4.streamReader = new MSStreamReader();
_this4.streamReader.onload = function (event) {
_this4.streamReader = null;
var buffer = event.target.result;
if (buffer.byteLength > 0) {
_this4.bytesRead += buffer.byteLength;
_this4.emit('buffer', buffer);
} else {
// Zero length means end of stream.
_this4.eof = true;
_this4.emit('done');
}
resolve();
};
_this4.streamReader.onerror = function () {
_this4.streamReader = null;
_this4.stream = null;
_this4.emit('error');
reject(new Error('mystery error streaming'));
};
_this4._onAbort = function (err) {
_this4.streamReader.abort();
_this4.streamReader = null;
_this4.stream = null;
_this4.emit('error');
reject(reason);
};
_this4.streamReader.readAsArrayBuffer(stream, nbytes);
}
});
});
}
}, {
key: 'abort',
value: function abort() {
if (this.streamReader) {
this.streamReader.abort();
this.streamReader = null;
}
if (this.stream) {
this.stream.msClose();
this.stream = null;
}
_get(MSStreamBackend.prototype.__proto__ || Object.getPrototypeOf(MSStreamBackend.prototype), 'abort', this).call(this);
}
}]);
return MSStreamBackend;
}(Backend);
MSStreamBackend.supported = function () {
try {
var xhr = new XMLHttpRequest();
// IE demands that open() be called before we can set xhr.responseType
xhr.open("GET", "/robots.txt");
xhr.responseType = type;
return xhr.responseType === type;
} catch (e) {
return false;
}
};
module.exports = MSStreamBackend;
/***/ }),
/* 29 */
/***/ (function(module, exports, __webpack_require__) {
"use strict";
var _createClass = function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; }();
var _get = function get(object, property, receiver) { if (object === null) object = Function.prototype; var desc = Object.getOwnPropertyDescriptor(object, property); if (desc === undefined) { var parent = Object.getPrototypeOf(object); if (parent === null) { return undefined; } else { return get(parent, property, receiver); } } else if ("value" in desc) { return desc.value; } else { var getter = desc.get; if (getter === undefined) { return undefined; } return getter.call(receiver); } };
function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError("Cannot call a class as a function"); } }
function _possibleConstructorReturn(self, call) { if (!self) { throw new ReferenceError("this hasn't been initialised - super() hasn't been called"); } return call && (typeof call === "object" || typeof call === "function") ? call : self; }
function _inherits(subClass, superClass) { if (typeof superClass !== "function" && superClass !== null) { throw new TypeError("Super expression must either be null or a function, not " + typeof superClass); } subClass.prototype = Object.create(superClass && superClass.prototype, { constructor: { value: subClass, enumerable: false, writable: true, configurable: true } }); if (superClass) Object.setPrototypeOf ? Object.setPrototypeOf(subClass, superClass) : subClass.__proto__ = superClass; }
var DownloadBackend = __webpack_require__(26);
var BinaryStringBackend = function (_DownloadBackend) {
_inherits(BinaryStringBackend, _DownloadBackend);
function BinaryStringBackend() {
_classCallCheck(this, BinaryStringBackend);
return _possibleConstructorReturn(this, (BinaryStringBackend.__proto__ || Object.getPrototypeOf(BinaryStringBackend)).apply(this, arguments));
}
_createClass(BinaryStringBackend, [{
key: "initXHR",
value: function initXHR() {
_get(BinaryStringBackend.prototype.__proto__ || Object.getPrototypeOf(BinaryStringBackend.prototype), "initXHR", this).call(this);
this.xhr.responseType = "text";
this.xhr.overrideMimeType('text/plain; charset=x-user-defined');
}
}, {
key: "onXHRProgress",
value: function onXHRProgress() {
var slice = this.xhr.responseText.slice(this.bytesRead);
if (slice.length > 0) {
this.bytesRead += slice.length;
this.emit('buffer', slice);
}
}
}, {
key: "onXHRLoad",
value: function onXHRLoad() {
// We may or may not get that final event
this.onXHRProgress();
_get(BinaryStringBackend.prototype.__proto__ || Object.getPrototypeOf(BinaryStringBackend.prototype), "onXHRLoad", this).call(this);
}
}]);
return BinaryStringBackend;
}(DownloadBackend);
BinaryStringBackend.supported = function () {
try {
var xhr = new XMLHttpRequest();
return !!xhr.overrideMimeType;
} catch (e) {
return false;
}
};
module.exports = BinaryStringBackend;
/***/ }),
/* 30 */
/***/ (function(module, exports, __webpack_require__) {
(function webpackUniversalModuleDefinition(root, factory) {
if(true)
module.exports = factory();
else if(typeof define === 'function' && define.amd)
define([], factory);
else if(typeof exports === 'object')
exports["AudioFeeder"] = factory();
else
root["AudioFeeder"] = factory();
})(this, function() {
return /******/ (function(modules) { // webpackBootstrap
/******/ // The module cache
/******/ var installedModules = {};
/******/ // The require function
/******/ function __webpack_require__(moduleId) {
/******/ // Check if module is in cache
/******/ if(installedModules[moduleId])
/******/ return installedModules[moduleId].exports;
/******/ // Create a new module (and put it into the cache)
/******/ var module = installedModules[moduleId] = {
/******/ exports: {},
/******/ id: moduleId,
/******/ loaded: false
/******/ };
/******/ // Execute the module function
/******/ modules[moduleId].call(module.exports, module, module.exports, __webpack_require__);
/******/ // Flag the module as loaded
/******/ module.loaded = true;
/******/ // Return the exports of the module
/******/ return module.exports;
/******/ }
/******/ // expose the modules object (__webpack_modules__)
/******/ __webpack_require__.m = modules;
/******/ // expose the module cache
/******/ __webpack_require__.c = installedModules;
/******/ // __webpack_public_path__
/******/ __webpack_require__.p = "";
/******/ // Load entry module and return exports
/******/ return __webpack_require__(0);
/******/ })
/************************************************************************/
/******/ ([
/* 0 */
/***/ function(module, exports, __webpack_require__) {
(function() {
var BufferQueue = __webpack_require__(1),
WebAudioBackend = __webpack_require__(2),
FlashBackend = __webpack_require__(4);
/**
* Audio sample buffer format passed to {@link AudioFeeder#bufferData} and its backends.
*
* Buffers are arrays containing one Float32Array of sample data
* per channel. Channel counts must match the expected value, and
* all channels within a buffer must have the same length in samples.
*
* Since input data may be stored for a while before being taken
* back out, be sure that your Float32Arrays for channel data are
* standalone, not backed on an ArrayBuffer that might change!
*
* @typedef {SampleBuffer} SampleBuffer
* @todo consider replacing this with AudioBuffer-like wrapper object
*/
/**
* Object dictionary format used to pass options into {@link AudioFeeder} and its backends.
*
* @typedef {Object} AudioFeederOptions
* @property {string} base - (optional) base URL to find additional resources in,
* such as the Flash audio output shim
* @property {AudioContext} audioContext - (optional) Web Audio API AudioContext
* instance to use inplace of creating a default one
*/
/**
* Object dictionary format returned from {@link AudioFeeder#getPlaybackState} and friends.
*
* @typedef {Object} PlaybackState
* @property {number} playbackPosition - total seconds so far of audio data that have played
* @property {number} samplesQueued - number of samples at target rate that are queued up for playback
* @property {number} dropped - number of underrun events, when we had to play silence due to data starvation
* @property {number} delayedTime - total seconds so far of silent time during playback due to data starvation
* @todo drop 'dropped' in favor of delayedTime
* @todo replace sampledQueued with a time unit?
*/
/**
* Object that we can throw audio data into and have it drain out.
* @class
* @param {AudioFeederOptions} options - dictionary of config settings
*
* @classdesc
* Object that we can throw audio data into and have it drain out.
*/
function AudioFeeder(options) {
this._options = options || {};
this._backend = null; // AudioBackend instance, after init...
};
/**
* Sample rate in Hz, as requested by the caller in {@link AudioFeeder#init}.
*
* If the backend's actual sample rate differs from this requested rate,
* input data will be resampled automatically.
*
* @type {number}
* @readonly
* @see AudioFeeder#targetRate
*/
AudioFeeder.prototype.rate = 0;
/**
* Actual output sample rate in Hz, as provided by the backend.
* This may differ from the rate requested, in which case input data
* will be resampled automatically.
*
* @type {number}
* @readonly
* @see AudioFeeder#rate
*/
AudioFeeder.prototype.targetRate = 0;
/**
* Number of output channels, as requested by the caller in {@link AudioFeeder#init}.
*
* If the backend's actual channel count differs from this requested count,
* input data will be resampled automatically.
*
* Warning: currently more than 2 channels may result in additional channels
* being cropped out, as downmixing has not yet been implemented.
*
* @type {number}
* @readonly
*/
AudioFeeder.prototype.channels = 0;
/**
* Size of output buffers in samples, as a hint for latency/scheduling
* @type {number}
* @readonly
*/
AudioFeeder.prototype.bufferSize = 0;
/**
* Duration of the minimum buffer size, in seconds.
* If the amount of buffered data falls below this,
* caller will receive a synchronous 'starved' event
* with a last chance to buffer more data.
*
* @type {number}
* @readonly
*/
Object.defineProperty(AudioFeeder.prototype, 'bufferDuration', {
get: function getBufferDuration() {
if (this.targetRate) {
return this.bufferSize / this.targetRate;
} else {
return 0;
}
}
});
/**
* Duration of remaining data at which a 'bufferlow' event will be
* triggered, in seconds.
*
* This defaults to twice bufferDuration, but can be overridden.
*
* @type {number}
*/
Object.defineProperty(AudioFeeder.prototype, 'bufferThreshold', {
get: function getBufferThreshold() {
if (this._backend) {
return this._backend.bufferThreshold / this.targetRate;
} else {
return 0;
}
},
set: function setBufferThreshold(val) {
if (this._backend) {
this._backend.bufferThreshold = Math.round(val * this.targetRate);
} else {
throw 'Invalid state: AudioFeeder cannot set bufferThreshold before init';
}
}
});
/**
* Current playback position, in seconds.
* This compensates for drops and underruns, and is suitable for a/v sync.
*
* @type {number}
* @readonly
*/
Object.defineProperty(AudioFeeder.prototype, 'playbackPosition', {
get: function getPlaybackPosition() {
if (this._backend) {
var playbackState = this.getPlaybackState();
return playbackState.playbackPosition;
} else {
return 0;
}
}
});
/**
* Duration of remaining queued data, in seconds.
*
* @type {number}
*/
Object.defineProperty(AudioFeeder.prototype, 'durationBuffered', {
get: function getDurationBuffered() {
if (this._backend) {
var playbackState = this.getPlaybackState();
return playbackState.samplesQueued / this.targetRate;
} else {
return 0;
}
}
});
/**
* Is the feeder currently set to mute output?
* When muted, this overrides the volume property.
*
* @type {boolean}
*/
Object.defineProperty(AudioFeeder.prototype, 'muted', {
get: function getMuted() {
if (this._backend) {
return this._backend.muted;
} else {
throw 'Invalid state: cannot get mute before init';
}
},
set: function setMuted(val) {
if (this._backend) {
this._backend.muted = val;
} else {
throw 'Invalid state: cannot set mute before init';
}
}
});
/**
* @deprecated in favor of muted and volume properties
*/
AudioFeeder.prototype.mute = function() {
this.muted = true;
};
/**
* @deprecated in favor of muted and volume properties
*/
AudioFeeder.prototype.unmute = function() {
this.muted = false;
};
/**
* Volume multiplier, defaults to 1.0.
* @name volume
* @type {number}
*/
Object.defineProperty(AudioFeeder.prototype, 'volume', {
get: function getVolume() {
if (this._backend) {
return this._backend.volume;
} else {
throw 'Invalid state: cannot get volume before init';
}
},
set: function setVolume(val) {
if (this._backend) {
this._backend.volume = val;
} else {
throw 'Invalid state: cannot set volume before init';
}
}
});
/**
* Start setting up for output with the given channel count and sample rate.
* Audio data you provide will be resampled if necessary to whatever the
* backend actually supports.
*
* @param {number} numChannels - requested number of channels (output may differ)
* @param {number} sampleRate - requested sample rate in Hz (output may differ)
*
* @todo merge into constructor?
*/
AudioFeeder.prototype.init = function(numChannels, sampleRate) {
this.channels = numChannels;
this.rate = sampleRate;
if (WebAudioBackend.isSupported()) {
this._backend = new WebAudioBackend(numChannels, sampleRate, this._options);
} else if (FlashBackend.isSupported()) {
this._backend = new FlashBackend(numChannels, sampleRate, this._options);
} else {
throw 'No supported backend';
}
this.targetRate = this._backend.rate;
this.bufferSize = this._backend.bufferSize;
// Pass through the starved event
this._backend.onstarved = (function() {
if (this.onstarved) {
this.onstarved();
}
}).bind(this);
this._backend.onbufferlow = (function() {
if (this.onbufferlow) {
this.onbufferlow();
}
}).bind(this);
};
/**
* Resample a buffer from the input rate/channel count to the output.
*
* This is horribly naive and wrong.
* Replace me with a better algo!
*
* @param {SampleBuffer} sampleData - input data in requested sample rate / channel count
* @returns {SampleBuffer} output data in backend's sample rate/channel count
*/
AudioFeeder.prototype._resample = function(sampleData) {
var rate = this.rate,
channels = this.channels,
targetRate = this._backend.rate,
targetChannels = this._backend.channels;
if (rate == targetRate && channels == targetChannels) {
return sampleData;
} else {
var newSamples = [];
for (var channel = 0; channel < targetChannels; channel++) {
var inputChannel = channel;
if (channel >= channels) {
// Flash forces output to stereo; if input is mono, dupe the first channel
inputChannel = 0;
}
var input = sampleData[inputChannel],
output = new Float32Array(Math.round(input.length * targetRate / rate));
for (var i = 0; i < output.length; i++) {
output[i] = input[(i * rate / targetRate) | 0];
}
newSamples.push(output);
}
return newSamples;
}
};
/**
* Queue up some audio data for playback.
*
* @param {SampleBuffer} sampleData - input data to queue up for playback
*
* @todo throw if data invalid or uneven
*/
AudioFeeder.prototype.bufferData = function(sampleData) {
if (this._backend) {
var samples = this._resample(sampleData);
this._backend.appendBuffer(samples);
} else {
throw 'Invalid state: AudioFeeder cannot bufferData before init';
}
};
/**
* Get an object with information about the current playback state.
*
* @return {PlaybackState} - info about current playback state
*/
AudioFeeder.prototype.getPlaybackState = function() {
if (this._backend) {
return this._backend.getPlaybackState();
} else {
throw 'Invalid state: AudioFeeder cannot getPlaybackState before init';
}
};
/**
* Checks if audio system is ready and calls the callback when ready
* to begin playback.
*
* This will wait for the Flash shim to load on IE 10/11; waiting
* is not required when using native Web Audio but you should use
* this callback to support older browsers.
*
* @param {function} callback - called when ready
*/
AudioFeeder.prototype.waitUntilReady = function(callback) {
if (this._backend) {
this._backend.waitUntilReady(callback);
} else {
throw 'Invalid state: AudioFeeder cannot waitUntilReady before init';
}
};
/**
* Start/continue playback as soon as possible.
*
* You should buffer some audio ahead of time to avoid immediately
* running into starvation.
*/
AudioFeeder.prototype.start = function() {
if (this._backend) {
this._backend.start();
} else {
throw 'Invalid state: AudioFeeder cannot start before init';
}
};
/**
* Stop/pause playback as soon as possible.
*
* Audio that has been buffered but not yet sent to the device will
* remain buffered, and can be continued with another call to start().
*/
AudioFeeder.prototype.stop = function() {
if (this._backend) {
this._backend.stop();
} else {
throw 'Invalid state: AudioFeeder cannot stop before init';
}
};
/**
* Flush any queued data out of the system.
*/
AudioFeeder.prototype.flush = function() {
if (this._backend) {
this._backend.flush();
} else {
throw 'Invalid state: AudioFeeder cannot flush before init';
}
}
/**
* Close out the audio channel. The AudioFeeder instance will no
* longer be usable after closing.
*
* @todo close out the AudioContext if no longer needed
* @todo make the instance respond more consistently once closed
*/
AudioFeeder.prototype.close = function() {
if (this._backend) {
this._backend.close();
this._backend = null;
}
};
/**
* Synchronous callback when we find we're out of buffered data.
*
* @type {function}
*/
AudioFeeder.prototype.onstarved = null;
/**
* Asynchronous callback when we're running low on buffered data.
*
* @type {function}
*/
AudioFeeder.prototype.onbufferlow = null;
/**
* Is the AudioFeeder class supported in this browser?
*
* Note that it's still possible to be supported but not work, for instance
* if there are no audio output devices but the APIs are available.
*
* @returns {boolean} - true if Web Audio API is available
*/
AudioFeeder.isSupported = function() {
return !!Float32Array && (WebAudioBackend.isSupported() || FlashBackend.isSupported());
};
/**
* Force initialization of the default Web Audio API context, if applicable.
*
* Some browsers (such as mobile Safari) disable audio output unless
* first triggered from a UI event handler; call this method as a hint
* that you will be starting up an AudioFeeder soon but won't have data
* for it until a later callback.
*
* @returns {AudioContext|null} - initialized AudioContext instance, if applicable
*/
AudioFeeder.initSharedAudioContext = function() {
if (WebAudioBackend.isSupported()) {
return WebAudioBackend.initSharedAudioContext();
} else {
return null;
}
};
module.exports = AudioFeeder;
})();
/***/ },
/* 1 */
/***/ function(module, exports) {
/**
* @file Abstraction around a queue of audio buffers.
*
* @author Brion Vibber <brion@pobox.com>
* @copyright (c) 2013-2016 Brion Vibber
* @license MIT
*/
/**
* Constructor for BufferQueue class.
* @class
* @param {number} numChannels - channel count to validate against
* @param {number} bufferSize - desired size of pre-chunked output buffers, in samples
*
* @classdesc
* Abstraction around a queue of audio buffers.
*
* Stuff input buffers of any length in via {@link BufferQueue#appendBuffer},
* check how much is queued with {@link BufferQueue#sampleCount}, and pull out
* data in fixed-size chunks from the start with {@link BufferQueue#shift}.
*/
function BufferQueue(numChannels, bufferSize) {
if (numChannels < 1 || numChannels !== Math.round(numChannels)) {
throw 'Invalid channel count for BufferQueue';
}
this.channels = numChannels;
this.bufferSize = bufferSize;
this.flush();
}
/**
* Flush any data out of the queue, resetting to empty state.
*/
BufferQueue.prototype.flush = function() {
this._buffers = [];
this._pendingBuffer = this.createBuffer(this.bufferSize);
this._pendingPos = 0;
};
/**
* Count how many samples are queued up
*
* @returns {number} - total count in samples
*/
BufferQueue.prototype.sampleCount = function() {
var count = 0;
this._buffers.forEach(function(buffer) {
count += buffer[0].length;
});
return count;
};
/**
* Create an empty audio sample buffer with space for the given count of samples.
*
* @param {number} sampleCount - number of samples to reserve in the buffer
* @returns {SampleBuffer} - empty buffer
*/
BufferQueue.prototype.createBuffer = function(sampleCount) {
var output = [];
for (var i = 0; i < this.channels; i++) {
output[i] = new Float32Array(sampleCount);
}
return output;
};
/**
* Validate a buffer for correct object layout
*
* @param {SampleBuffer} buffer - an audio buffer to check
* @returns {boolean} - true if input buffer is valid
*/
BufferQueue.prototype.validate = function(buffer) {
if (buffer.length !== this.channels) {
return false;
}
var sampleCount;
for (var i = 0; i < buffer.length; i++) {
var channelData = buffer[i];
if (!(channelData instanceof Float32Array)) {
return false;
}
if (i == 0) {
sampleCount = channelData.length;
} else if (channelData.length !== sampleCount) {
return false;
}
}
return true;
};
/**
* Append a buffer of input data to the queue...
*
* @param {SampleBuffer} sampleData - an audio buffer to append
* @throws exception on invalid input
*/
BufferQueue.prototype.appendBuffer = function(sampleData) {
if (!this.validate(sampleData)) {
throw "Invalid audio buffer passed to BufferQueue.appendBuffer";
}
var firstChannel = sampleData[0],
sampleCount = firstChannel.length;
// @todo this seems hella inefficient
for (var i = 0; i < sampleCount; i++) {
for (var channel = 0; channel < this.channels; channel++) {
this._pendingBuffer[channel][this._pendingPos] = sampleData[channel][i];
}
if (++this._pendingPos == this.bufferSize) {
this._buffers.push(this._pendingBuffer);
this._pendingPos = 0;
this._pendingBuffer = this.createBuffer(this.bufferSize);
}
}
};
/**
* Unshift the given sample buffer onto the beginning of the buffer queue.
*
* @param {SampleBuffer} sampleData - an audio buffer to prepend
* @throws exception on invalid input
*
* @todo this is currently pretty inefficient as it rechunks all the buffers.
*/
BufferQueue.prototype.prependBuffer = function(sampleData) {
if (!this.validate(sampleData)) {
throw "Invalid audio buffer passed to BufferQueue.prependBuffer";
}
// Since everything is pre-chunked in the queue, we're going to have
// to pull everything out and re-append it.
var buffers = this._buffers.slice(0)
buffers.push(this.trimBuffer(this._pendingBuffer, 0, this._pendingPos));
this.flush();
this.appendBuffer(sampleData);
// Now put back any old buffers, dividing them up into chunks.
for (var i = 0; i < buffers.length; i++) {
this.appendBuffer(buffers[i]);
}
};
/**
* Shift out a buffer from the head of the queue, containing a maximum of
* {@link BufferQueue#bufferSize} samples; if there are not enough samples
* you may get a shorter buffer. Call {@link BufferQueue#sampleCount} to
* check if enough samples are available for your needs.
*
* @returns {SampleBuffer} - an audio buffer with zero or more samples
*/
BufferQueue.prototype.nextBuffer = function() {
if (this._buffers.length) {
return this._buffers.shift();
} else {
var trimmed = this.trimBuffer(this._pendingBuffer, 0, this._pendingPos);
this._pendingBuffer = this.createBuffer(this.bufferSize);
this._pendingPos = 0;
return trimmed;
}
};
/**
* Trim a buffer down to a given maximum sample count.
* Any additional samples will simply be cropped off of the view.
* If no trimming is required, the same buffer will be returned.
*
* @param {SampleBuffer} sampleData - input data
* @param {number} start - sample number to start at
* @param {number} maxSamples - count of samples to crop to
* @returns {SampleBuffer} - output data with at most maxSamples samples
*/
BufferQueue.prototype.trimBuffer = function(sampleData, start, maxSamples) {
var bufferLength = sampleData[0].length,
end = start + Math.min(maxSamples, bufferLength);
if (start == 0 && end >= bufferLength) {
return sampleData;
} else {
var output = [];
for (var i = 0; i < this.channels; i++) {
output[i] = sampleData[i].subarray(start, end);
}
return output;
}
};
module.exports = BufferQueue;
/***/ },
/* 2 */
/***/ function(module, exports, __webpack_require__) {
/**
* @file Web Audio API backend for AudioFeeder
* @author Brion Vibber <brion@pobox.com>
* @copyright (c) 2013-2016 Brion Vibber
* @license MIT
*/
(function() {
var AudioContext = window.AudioContext || window.webkitAudioContext,
BufferQueue = __webpack_require__(1),
nextTick = __webpack_require__(3);
/**
* Constructor for AudioFeeder's Web Audio API backend.
* @class
* @param {number} numChannels - requested count of output channels
* @param {number} sampleRate - requested sample rate for output
* @param {Object} options - pass URL path to directory containing 'dynamicaudio.swf' in 'base' parameter
*
* @classdesc Web Audio API output backend for AudioFeeder.
* Maintains an internal {@link BufferQueue} of audio samples to be output on demand.
*/
function WebAudioBackend(numChannels, sampleRate, options) {
var context = options.audioContext || WebAudioBackend.initSharedAudioContext();
this._context = context;
/*
* Optional audio node can be provided to connect the feeder to
* @type {AudioNode}
*/
this.output = options.output || context.destination
/**
* Actual sample rate supported for output, in Hz
* @type {number}
* @readonly
*/
this.rate = context.sampleRate;
/**
* Actual count of channels supported for output
* @type {number}
* @readonly
*/
this.channels = Math.min(numChannels, 2); // @fixme remove this limit
if (options.bufferSize) {
this.bufferSize = (options.bufferSize | 0);
}
this.bufferThreshold = 2 * this.bufferSize;
this._bufferQueue = new BufferQueue(this.channels, this.bufferSize);
this._playbackTimeAtBufferTail = context.currentTime;
this._queuedTime = 0;
this._delayedTime = 0;
this._dropped = 0;
this._liveBuffer = this._bufferQueue.createBuffer(this.bufferSize);
// @todo support new audio worker mode too
if (context.createScriptProcessor) {
this._node = context.createScriptProcessor(this.bufferSize, 0, this.channels);
} else if (context.createJavaScriptNode) {
// In older Safari versions
this._node = context.createJavaScriptNode(this.bufferSize, 0, this.channels);
} else {
throw new Error("Bad version of web audio API?");
}
}
/**
* Size of output buffers in samples, as a hint for latency/scheduling
* @type {number}
* @readonly
*/
WebAudioBackend.prototype.bufferSize = 4096;
/**
* Remaining sample count at which a 'bufferlow' event will be triggered.
*
* Will be pinged when falling below bufferThreshold or bufferSize,
* whichever is larger.
*
* @type {number}
*/
WebAudioBackend.prototype.bufferThreshold = 8192;
/**
* Internal volume property backing.
* @type {number}
* @access private
*/
WebAudioBackend.prototype._volume = 1;
/**
* Volume multiplier, defaults to 1.0.
* @name volume
* @type {number}
*/
Object.defineProperty(WebAudioBackend.prototype, 'volume', {
get: function getVolume() {
return this._volume;
},
set: function setVolume(val) {
this._volume = +val;
}
});
/**
* Internal muted property backing.
* @type {number}
* @access private
*/
WebAudioBackend.prototype._muted = false;
/**
* Is the backend currently set to mute output?
* When muted, this overrides the volume property.
*
* @type {boolean}
*/
Object.defineProperty(WebAudioBackend.prototype, 'muted', {
get: function getMuted() {
return this._muted;
},
set: function setMuted(val) {
this._muted = !!val;
}
});
/**
* onaudioprocess event handler for the ScriptProcessorNode
* @param {AudioProcessingEvent} event - audio processing event object
* @access private
*/
WebAudioBackend.prototype._audioProcess = function(event) {
var channel, input, output, i, playbackTime;
if (typeof event.playbackTime === 'number') {
playbackTime = event.playbackTime;
} else {
// Safari 6.1 hack
playbackTime = this._context.currentTime + (this.bufferSize / this.rate);
}
var expectedTime = this._playbackTimeAtBufferTail;
if (expectedTime < playbackTime) {
// we may have lost some time while something ran too slow
this._delayedTime += (playbackTime - expectedTime);
}
if (this._bufferQueue.sampleCount() < this.bufferSize) {
// We might be in a throttled background tab; go ping the decoder
// and let it know we need more data now!
// @todo use standard event firing?
if (this.onstarved) {
this.onstarved();
}
}
// If we still haven't got enough data, write a buffer of silence
// to all channels and record an underrun event.
// @todo go ahead and output the data we _do_ have?
if (this._bufferQueue.sampleCount() < this.bufferSize) {
for (channel = 0; channel < this.channels; channel++) {
output = event.outputBuffer.getChannelData(channel);
for (i = 0; i < this.bufferSize; i++) {
output[i] = 0;
}
}
this._dropped++;
return;
}
var volume = (this.muted ? 0 : this.volume);
// Actually get that data and write it out...
var inputBuffer = this._bufferQueue.nextBuffer();
if (inputBuffer[0].length < this.bufferSize) {
// This should not happen, but trust no invariants!
throw 'Audio buffer not expected length.';
}
for (channel = 0; channel < this.channels; channel++) {
input = inputBuffer[channel];
// Save this buffer data for later in case we pause
this._liveBuffer[channel].set(inputBuffer[channel]);
// And play it out with volume applied...
output = event.outputBuffer.getChannelData(channel);
for (i = 0; i < input.length; i++) {
output[i] = input[i] * volume;
}
}
this._queuedTime += (this.bufferSize / this.rate);
this._playbackTimeAtBufferTail = playbackTime + (this.bufferSize / this.rate);
if (this._bufferQueue.sampleCount() < Math.max(this.bufferSize, this.bufferThreshold)) {
// Let the decoder know ahead of time we're running low on data.
// @todo use standard event firing?
if (this.onbufferlow) {
nextTick(this.onbufferlow.bind(this));
}
}
};
/**
* Return a count of samples that have been queued or output but not yet played.
*
* @returns {number} - sample count
* @access private
*/
WebAudioBackend.prototype._samplesQueued = function() {
var bufferedSamples = this._bufferQueue.sampleCount();
var remainingSamples = Math.floor(this._timeAwaitingPlayback() * this.rate);
return bufferedSamples + remainingSamples;
};
/**
* Return time duration between the present and the endpoint of audio
* we have already sent out from our queue to Web Audio.
*
* @returns {number} - seconds
*/
WebAudioBackend.prototype._timeAwaitingPlayback = function() {
return Math.max(0, this._playbackTimeAtBufferTail - this._context.currentTime);
};
/**
* Get info about current playback state.
*
* @return {PlaybackState} - info about current playback state
*/
WebAudioBackend.prototype.getPlaybackState = function() {
return {
playbackPosition: this._queuedTime - this._timeAwaitingPlayback(),
samplesQueued: this._samplesQueued(),
dropped: this._dropped,
delayed: this._delayedTime
};
};
/**
* Wait asynchronously until the backend is ready before continuing.
*
* This will always call immediately for the Web Audio API backend,
* as there is no async setup process.
*
* @param {function} callback - to be called when ready
*/
WebAudioBackend.prototype.waitUntilReady = function(callback) {
callback();
};
/**
* Append a buffer of audio data to the output queue for playback.
*
* Audio data must be at the expected sample rate; resampling is done
* upstream in {@link AudioFeeder}.
*
* @param {SampleBuffer} sampleData - audio data at target sample rate
*/
WebAudioBackend.prototype.appendBuffer = function(sampleData) {
this._bufferQueue.appendBuffer(sampleData);
};
/**
* Start playback.
*
* Audio should have already been queued at this point,
* or starvation may occur immediately.
*/
WebAudioBackend.prototype.start = function() {
this._node.onaudioprocess = this._audioProcess.bind(this);
this._node.connect(this.output);
this._playbackTimeAtBufferTail = this._context.currentTime;
};
/**
* Stop playback, but don't release resources or clear the buffers.
* We'll probably come back soon.
*/
WebAudioBackend.prototype.stop = function() {
if (this._node) {
var timeRemaining = this._timeAwaitingPlayback();
if (timeRemaining > 0) {
// We have some leftover samples that got queued but didn't get played.
// Unshift them back onto the beginning of the buffer.
// @todo make this not a horrible hack
var samplesRemaining = Math.round(timeRemaining * this.rate),
samplesAvailable = this._liveBuffer ? this._liveBuffer[0].length : 0;
if (samplesRemaining > samplesAvailable) {
//console.log('liveBuffer size ' + samplesRemaining + ' vs ' + samplesAvailable);
this._bufferQueue.prependBuffer(this._liveBuffer);
this._bufferQueue.prependBuffer(
this._bufferQueue.createBuffer(samplesRemaining - samplesAvailable));
} else {
this._bufferQueue.prependBuffer(
this._bufferQueue.trimBuffer(this._liveBuffer, samplesAvailable - samplesRemaining, samplesRemaining));
}
this._playbackTimeAtBufferTail -= timeRemaining;
}
this._node.onaudioprocess = null;
this._node.disconnect();
}
};
/**
* Flush any queued data out of the system.
*/
WebAudioBackend.prototype.flush = function() {
this._bufferQueue.flush();
};
/**
* Close out the playback system and release resources.
*
* @todo consider releasing the AudioContext when possible
*/
WebAudioBackend.prototype.close = function() {
this.stop();
this._context = null;
};
/**
* Synchronous callback for when we run out of input data
*
* @type function|null
*/
WebAudioBackend.prototype.onstarved = null;
/**
* Asynchronous callback for when the buffer runs low and
* should be refilled soon.
*
* @type function|null
*/
WebAudioBackend.prototype.onbufferlow = null;
/**
* Check if Web Audio API appears to be supported.
*
* Note this is somewhat optimistic; will return true even if there are no
* audio devices available, as long as the API is present.
*
* @returns {boolean} - true if this browser appears to support Web Audio API
*/
WebAudioBackend.isSupported = function() {
return !!AudioContext;
};
/**
* Holder of audio context to be used/reused by WebAudioBackend.
* @see {WebAudioBackend#initSharedAudioContext}
*
* @type {AudioContext}
*/
WebAudioBackend.sharedAudioContext = null;
/**
* Force initialization of the default Web Audio API context.
*
* Some browsers (such as mobile Safari) disable audio output unless
* first triggered from a UI event handler; call this method as a hint
* that you will be starting up an AudioFeeder soon but won't have data
* for it until a later callback.
*
* @returns {AudioContext|null} - initialized AudioContext instance, if applicable
*/
WebAudioBackend.initSharedAudioContext = function() {
if (!WebAudioBackend.sharedAudioContext) {
if (WebAudioBackend.isSupported()) {
// We're only allowed 4 contexts on many browsers
// and there's no way to discard them (!)...
var context = new AudioContext(),
node;
if (context.createScriptProcessor) {
node = context.createScriptProcessor(1024, 0, 2);
} else if (context.createJavaScriptNode) {
node = context.createJavaScriptNode(1024, 0, 2);
} else {
throw new Error( "Bad version of web audio API?" );
}
// Don't actually run any audio, just start & stop the node
node.connect(context.destination);
node.disconnect();
// So far so good. Keep it around!
WebAudioBackend.sharedAudioContext = context;
}
}
return WebAudioBackend.sharedAudioContext;
};
module.exports = WebAudioBackend;
})();
/***/ },
/* 3 */
/***/ function(module, exports) {
module.exports = (function() {
// Don't try to check for setImmediate directly; webpack implements
// it using setTimeout which will be throttled in background tabs.
// Checking directly on the global window object skips this interference.
if (typeof window.setImmediate !== 'undefined') {
return window.setImmediate;
}
// window.postMessage goes straight to the event loop, no throttling.
if (window && window.postMessage) {
var nextTickQueue = [];
window.addEventListener('message', function(event) {
if (event.source === window) {
var data = event.data;
if (typeof data === 'object' && data.nextTickBrowserPingMessage) {
var callback = nextTickQueue.pop();
if (callback) {
callback();
}
}
}
});
return function(callback) {
nextTickQueue.push(callback);
window.postMessage({
nextTickBrowserPingMessage: true
}, document.location.toString())
};
}
// Timeout fallback may be poor in background tabs
return function(callback) {
setTimeout(callback, 0);
}
})();
/***/ },
/* 4 */
/***/ function(module, exports, __webpack_require__) {
(function() {
/* global ActiveXObject */
var dynamicaudio_swf = __webpack_require__(5);
var nextTick = __webpack_require__(3);
/**
* Constructor for AudioFeeder's Flash audio backend.
* @class
* @param {number} numChannels - requested count of output channels (actual will be fixed at 2)
* @param {number} sampleRate - requested sample rate for output (actual will be fixed at 44.1 kHz)
* @param {Object} options - pass URL path to directory containing 'dynamicaudio.swf' in 'base' parameter
*
* @classdesc Flash audio output backend for AudioFeeder.
* Maintains a local queue of data to be sent down to the Flash shim.
* Resampling to stereo 44.1 kHz is done upstream in AudioFeeder.
*/
var FlashBackend = function(numChannels, sampleRate, options) {
options = options || {};
var flashOptions = {};
if (typeof options.base === 'string') {
// @fixme replace the version string with an auto-updateable one
flashOptions.swf = options.base + '/' + dynamicaudio_swf;
}
if (options.bufferSize) {
this.bufferSize = (options.bufferSize | 0);
}
this._flashaudio = new DynamicAudio(flashOptions);
this._flashBuffer = '';
this._flushTimeout = null;
this._flushInterval = 40; // flush added buffers no more often than every X ms
this._cachedFlashState = null;
this._cachedFlashTime = 0;
this._cachedFlashInterval = 40; // resync state no more often than every X ms
this._waitUntilReadyQueue = [];
this.onready = function() {
this._flashaudio.flashElement.setBufferSize(this.bufferSize);
this._flashaudio.flashElement.setBufferThreshold(this.bufferThreshold);
while (this._waitUntilReadyQueue.length) {
var callback = this._waitUntilReadyQueue.shift();
callback.apply(this);
}
};
this.onlog = function(msg) {
console.log('AudioFeeder FlashBackend: ' + msg);
}
this.bufferThreshold = this.bufferSize * 2;
var events = {
'ready': 'sync',
'log': 'sync',
'starved': 'sync',
'bufferlow': 'async'
};
this._callbackName = 'AudioFeederFlashBackendCallback' + this._flashaudio.id;
var self = this;
window[this._callbackName] = (function(eventName) {
var method = events[eventName],
callback = this['on' + eventName];
if (method && callback) {
if (method === 'async') {
nextTick(callback.bind(this));
} else {
callback.apply(this, Array.prototype.slice.call(arguments, 1));
this._flushFlashBuffer();
}
}
}).bind(this);
};
/**
* Actual sample rate supported for output, in Hz
* Fixed to 44.1 kHz for Flash backend.
* @type {number}
* @readonly
*/
FlashBackend.prototype.rate = 44100;
/**
* Actual count of channels supported for output
* Fixed to stereo for Flash backend.
* @type {number}
* @readonly
*/
FlashBackend.prototype.channels = 2;
/**
* Buffer size hint.
* @type {number}
* @readonly
*/
FlashBackend.prototype.bufferSize = 4096;
/**
* Internal bufferThreshold property backing.
* @type {number}
* @access private
*/
FlashBackend.prototype._bufferThreshold = 8192;
/**
* Remaining sample count at which a 'bufferlow' event will be triggered.
*
* Will be pinged when falling below bufferThreshold or bufferSize,
* whichever is larger.
*
* @type {number}
*/
Object.defineProperty(FlashBackend.prototype, 'bufferThreshold', {
get: function getBufferThreshold() {
return this._bufferThreshold;
},
set: function setBufferThreshold(val) {
this._bufferThreshold = val | 0;
this.waitUntilReady((function() {
this._flashaudio.flashElement.setBufferThreshold(this._bufferThreshold);
}).bind(this));
}
});
/**
* Internal volume property backing.
* @type {number}
* @access private
*/
FlashBackend.prototype._volume = 1;
/**
* Volume multiplier, defaults to 1.0.
* @name volume
* @type {number}
*/
Object.defineProperty(FlashBackend.prototype, 'volume', {
get: function getVolume() {
return this._volume;
},
set: function setVolume(val) {
this._volume = +val;
this.waitUntilReady(this._flashVolumeUpdate.bind(this));
}
});
/**
* Internal muted property backing.
* @type {number}
* @access private
*/
FlashBackend.prototype._muted = false;
/**
* Is the backend currently set to mute output?
* When muted, this overrides the volume property.
*
* @type {boolean}
*/
Object.defineProperty(FlashBackend.prototype, 'muted', {
get: function getMuted() {
return this._muted;
},
set: function setMuted(val) {
this._muted = !!val;
this.waitUntilReady(this._flashVolumeUpdate.bind(this));
}
});
/**
* Are we paused/idle?
* @type {boolean}
* @access private
*/
FlashBackend.prototype._paused = true;
/**
* Pass the currently configured muted+volume state down to the Flash plugin
* @access private
*/
FlashBackend.prototype._flashVolumeUpdate = function() {
if (this._flashaudio && this._flashaudio.flashElement && this._flashaudio.flashElement.setVolume) {
this._flashaudio.flashElement.setVolume(this.muted ? 0 : this.volume);
}
}
/**
* Scaling and reordering of output for the Flash fallback.
* Input data must be pre-resampled to the correct sample rate.
*
* @param {SampleBuffer} samples - input data as separate channels of 32-bit float
* @returns {Int16Array} - interleaved stereo 16-bit signed integer output
* @access private
*
* @todo handle input with higher channel counts better
* @todo try sending floats to flash without losing precision?
*/
FlashBackend.prototype._resampleFlash = function(samples) {
var sampleincr = 1;
var samplecount = samples[0].length;
var newSamples = new Int16Array(samplecount * 2);
var chanLeft = samples[0];
var chanRight = this.channels > 1 ? samples[1] : chanLeft;
var volume = this.muted ? 0 : this.volume;
var multiplier = volume * 16384; // smaller than 32768 to allow some headroom from those floats
for(var s = 0; s < samplecount; s++) {
var idx = (s * sampleincr) | 0;
var idx_out = s * 2;
// Use a smaller
newSamples[idx_out] = chanLeft[idx] * multiplier;
newSamples[idx_out + 1] = chanRight[idx] * multiplier;
}
return newSamples;
};
var hexDigits = ['0', '1', '2', '3', '4', '5', '6', '7',
'8', '9', 'a', 'b', 'c', 'd', 'e', 'f'];
var hexBytes = [];
for (var i = 0; i < 256; i++) {
hexBytes[i] = hexDigits[(i & 0x0f)] +
hexDigits[(i & 0xf0) >> 4];
}
function hexString(buffer) {
var samples = new Uint8Array(buffer);
var digits = "",
len = samples.length;
for (var i = 0; i < len; i++) {
// Note that in IE 11 string concatenation is twice as fast as
// the traditional make-an-array-and-join here.
digits += hexBytes[samples[i]];
}
return digits;
}
/**
* Send any pending data off to the Flash plugin.
*
* @access private
*/
FlashBackend.prototype._flushFlashBuffer = function() {
var chunk = this._flashBuffer,
flashElement = this._flashaudio.flashElement;
if (this._cachedFlashState) {
this._cachedFlashState.samplesQueued += chunk.length / 8;
}
this._flashBuffer = '';
this._flushTimeout = null;
if (chunk.length > 0) {
this.waitUntilReady(function() {
flashElement.write(chunk);
});
}
};
/**
* Append a buffer of audio data to the output queue for playback.
*
* Audio data must be at the expected sample rate; resampling is done
* upstream in {@link AudioFeeder}.
*
* @param {SampleBuffer} sampleData - audio data at target sample rate
*/
FlashBackend.prototype.appendBuffer = function(sampleData) {
var resamples = this._resampleFlash(sampleData);
if (resamples.length > 0) {
var str = hexString(resamples.buffer);
this._flashBuffer += str;
if (!this._flushTimeout) {
// consolidate multiple consecutive tiny buffers in one pass;
// pushing data to Flash is relatively expensive on slow machines
this._flushTimeout = setTimeout(this._flushFlashBuffer.bind(this), this._flushInterval);
}
}
};
/**
* Get info about current playback state.
*
* @return {PlaybackState} - info about current playback state
*/
FlashBackend.prototype.getPlaybackState = function() {
if (this._flashaudio && this._flashaudio.flashElement && this._flashaudio.flashElement.write) {
var now = Date.now(),
delta = this._paused ? 0 : (now - this._cachedFlashTime),
state;
if (this._cachedFlashState && delta < this._cachedFlashInterval) {
var cachedFlashState = this._cachedFlashState;
state = {
playbackPosition: cachedFlashState.playbackPosition + delta / 1000,
samplesQueued: cachedFlashState.samplesQueued -
Math.max(0, Math.round(delta * this.rate / 1000)),
dropped: cachedFlashState.dropped,
delayed: cachedFlashState.delayed
};
} else {
state = this._flashaudio.flashElement.getPlaybackState();
this._cachedFlashState = state;
this._cachedFlashTime = now;
}
state.samplesQueued += this._flashBuffer.length / 8;
return state;
} else {
//console.log('getPlaybackState USED TOO EARLY');
return {
playbackPosition: 0,
samplesQueued: 0,
dropped: 0,
delayed: 0
};
}
};
/**
* Wait until the backend is ready to start, then call the callback.
*
* @param {function} callback - called on completion
* @todo handle fail case better?
*/
FlashBackend.prototype.waitUntilReady = function(callback) {
if (this._flashaudio && this._flashaudio.flashElement.write) {
callback.apply(this);
} else {
this._waitUntilReadyQueue.push(callback);
}
};
/**
* Start playback.
*
* Audio should have already been queued at this point,
* or starvation may occur immediately.
*/
FlashBackend.prototype.start = function() {
this._flashaudio.flashElement.start();
this._paused = false;
this._cachedFlashState = null;
};
/**
* Stop playback, but don't release resources or clear the buffers.
* We'll probably come back soon.
*/
FlashBackend.prototype.stop = function() {
this._flashaudio.flashElement.stop();
this._paused = true;
this._cachedFlashState = null;
};
/**
* Flush any queued data out of the system.
*/
FlashBackend.prototype.flush = function() {
this._flashaudio.flashElement.flush();
this._cachedFlashState = null;
};
/**
* Close out the playback system and release resources.
*/
FlashBackend.prototype.close = function() {
this.stop();
var wrapper = this._flashaudio.flashWrapper;
wrapper.parentNode.removeChild(wrapper);
this._flashaudio = null;
delete window[this._callbackName];
};
/**
* Synchronous callback for when we run out of input data
*
* @type function|null
*/
FlashBackend.prototype.onstarved = null;
/**
* Asynchronous callback for when the buffer runs low and
* should be refilled soon.
*
* @type function|null
*/
FlashBackend.prototype.onbufferlow = null;
/**
* Check if the browser appears to support Flash.
*
* Note this is somewhat optimistic, in that Flash may be supported
* but the dynamicaudio.swf file might not load, or it might load
* but there might be no audio devices, etc.
*
* Currently only checks for the ActiveX Flash plugin for Internet Explorer,
* as other target browsers support Web Audio API.
*
* @returns {boolean} - true if this browser appears to support Flash
*/
FlashBackend.isSupported = function() {
if (navigator.userAgent.indexOf('Trident') !== -1) {
// We only do the ActiveX test because we only need Flash in
// Internet Explorer 10/11. Other browsers use Web Audio directly
// (Edge, Safari) or native playback, so there's no need to test
// other ways of loading Flash.
try {
var obj = new ActiveXObject('ShockwaveFlash.ShockwaveFlash');
return true;
} catch(e) {
return false;
}
}
return false;
};
/** Flash fallback **/
/*
The Flash fallback is based on https://github.com/an146/dynamicaudio.js
This is the contents of the LICENSE file:
Copyright (c) 2010, Ben Firshman
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are met:
* Redistributions of source code must retain the above copyright notice, this
list of conditions and the following disclaimer.
* Redistributions in binary form must reproduce the above copyright notice,
this list of conditions and the following disclaimer in the documentation
and/or other materials provided with the distribution.
* The names of its contributors may not be used to endorse or promote products
derived from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR
ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
(INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON
ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
/**
* Wrapper class for instantiating Flash plugin.
*
* @constructor
* @param {Object} args - pass 'swf' to override default dynamicaudio.swf URL
* @access private
*/
function DynamicAudio(args) {
if (this instanceof arguments.callee) {
if (typeof this.init === "function") {
this.init.apply(this, (args && args.callee) ? args : arguments);
}
} else {
return new arguments.callee(arguments);
}
}
DynamicAudio.nextId = 1;
DynamicAudio.prototype = {
nextId: null,
swf: dynamicaudio_swf,
flashWrapper: null,
flashElement: null,
init: function(opts) {
var self = this;
self.id = DynamicAudio.nextId++;
if (opts && typeof opts.swf !== 'undefined') {
self.swf = opts.swf;
}
self.flashWrapper = document.createElement('div');
self.flashWrapper.id = 'dynamicaudio-flashwrapper-'+self.id;
// Credit to SoundManager2 for this:
var s = self.flashWrapper.style;
s.position = 'fixed';
s.width = '11px'; // must be at least 6px for flash to run fast
s.height = '11px';
s.bottom = s.left = '0px';
s.overflow = 'hidden';
self.flashElement = document.createElement('div');
self.flashElement.id = 'dynamicaudio-flashelement-'+self.id;
self.flashWrapper.appendChild(self.flashElement);
document.body.appendChild(self.flashWrapper);
var id = self.flashElement.id;
var params = '<param name="FlashVars" value="objectId=' + self.id + '">';
self.flashWrapper.innerHTML = "<object id='"+id+"' width='10' height='10' type='application/x-shockwave-flash' data='"+self.swf+"' style='visibility: visible;'><param name='allowscriptaccess' value='always'>" + params + "</object>";
self.flashElement = document.getElementById(id);
},
};
module.exports = FlashBackend;
})();
/***/ },
/* 5 */
/***/ function(module, exports, __webpack_require__) {
module.exports = __webpack_require__.p + "dynamicaudio.swf?version=0f98a83acdc72cf30968c16a018e1cf1";
/***/ }
/******/ ])
});
;
/***/ }),
/* 31 */
/***/ (function(module, exports, __webpack_require__) {
module.exports = __webpack_require__.p + "dynamicaudio.swf?version=0f98a83acdc72cf30968c16a018e1cf1";
/***/ }),
/* 32 */
/***/ (function(module, exports) {
/**
* Give as your 'process' function something that will trigger an async
* operation, then call the left() or right() methods to run another
* iteration, bisecting to the given direction.
*
* Caller is responsible for determining when done.
*
* @params options object {
* start: number,
* end: number,
* process: function(start, position, end)
* }
*/
function Bisector(options) {
var start = options.start,
end = options.end,
position = 0,
self = this,
n = 0;
function iterate() {
n++;
position = Math.floor((start + end) / 2);
return options.process(start, end, position);
}
self.start = function() {
iterate();
return self;
};
self.left = function() {
end = position;
return iterate();
};
self.right = function() {
start = position;
return iterate();
};
}
module.exports = Bisector;
/***/ }),
/* 33 */
/***/ (function(module, exports, __webpack_require__) {
/**
* Proxy object for web worker interface for codec classes.
*
* Used by the high-level player interface.
*
* @author Brion Vibber <brion@pobox.com>
* @copyright 2015
* @license MIT-style
*/
var OGVLoader = __webpack_require__(3);
var OGVWrapperCodec = (function(options) {
options = options || {};
var self = this,
suffix = '?version=' + encodeURIComponent(("1.4.2-20170425024925-504d7197")),
base = (typeof options.base === 'string') ? (options.base + '/') : '',
type = (typeof options.type === 'string') ? options.type : 'video/ogg',
processing = false,
demuxer = null,
videoDecoder = null,
audioDecoder = null,
flushIter = 0;
// Wrapper for callbacks to drop them after a flush
function flushSafe(func) {
var savedFlushIter = flushIter;
return function(arg) {
if (flushIter <= savedFlushIter) {
func(arg);
}
};
}
var loadedMetadata = false;
Object.defineProperty(self, 'loadedMetadata', {
get: function() {
return loadedMetadata;
}
});
Object.defineProperty(self, 'processing', {
get: function() {
return processing/*
|| (videoDecoder && videoDecoder.processing)
|| (audioDecoder && audioDecoder.processing)*/;
}
});
Object.defineProperty(self, 'duration', {
get: function() {
if (self.loadedMetadata) {
return demuxer.duration;
} else {
return NaN;
}
}
});
Object.defineProperty(self, 'hasAudio', {
get: function() {
return self.loadedMetadata && !!audioDecoder;
}
});
Object.defineProperty(self, 'audioReady', {
get: function() {
return self.hasAudio && demuxer.audioReady;
}
});
Object.defineProperty(self, 'audioTimestamp', {
get: function() {
return demuxer.audioTimestamp;
}
});
Object.defineProperty(self, 'audioFormat', {
get: function() {
if (self.hasAudio) {
return audioDecoder.audioFormat;
} else {
return null;
}
}
});
Object.defineProperty(self, 'audioBuffer', {
get: function() {
if (self.hasAudio) {
return audioDecoder.audioBuffer;
} else {
return null;
}
}
});
Object.defineProperty(self, 'hasVideo', {
get: function() {
return self.loadedMetadata && !!videoDecoder;
}
});
Object.defineProperty(self, 'frameReady', {
get: function() {
return self.hasVideo && demuxer.frameReady;
}
});
Object.defineProperty(self, 'frameTimestamp', {
get: function() {
return demuxer.frameTimestamp;
}
});
Object.defineProperty(self, 'keyframeTimestamp', {
get: function() {
return demuxer.keyframeTimestamp;
}
});
Object.defineProperty(self, 'nextKeyframeTimestamp', {
get: function() {
return demuxer.nextKeyframeTimestamp;
}
});
Object.defineProperty(self, 'videoFormat', {
get: function() {
if (self.hasVideo) {
return videoDecoder.videoFormat;
} else {
return null;
}
}
});
Object.defineProperty(self, 'frameBuffer', {
get: function() {
if (self.hasVideo) {
return videoDecoder.frameBuffer;
} else {
return null;
}
}
});
Object.defineProperty(self, 'seekable', {
get: function() {
return demuxer.seekable;
}
});
// - public methods
self.init = function(callback) {
var demuxerClassName;
if (options.type === 'video/webm') {
demuxerClassName = options.wasm ? 'OGVDemuxerWebMW' : 'OGVDemuxerWebM';
} else {
demuxerClassName = options.wasm ? 'OGVDemuxerOggW' : 'OGVDemuxerOgg';
}
processing = true;
OGVLoader.loadClass(demuxerClassName, function(demuxerClass) {
demuxer = new demuxerClass();
demuxer.onseek = function(offset) {
if (self.onseek) {
self.onseek(offset);
}
};
demuxer.init(function() {
processing = false;
callback();
});
});
};
self.close = function() {
if (demuxer) {
demuxer.close();
demuxer = null;
}
if (videoDecoder) {
videoDecoder.close();
videoDecoder = null;
}
if (audioDecoder) {
audioDecoder.close();
audioDecoder = null;
}
};
self.receiveInput = function(data, callback) {
demuxer.receiveInput(data, callback);
};
var audioClassMap = {
vorbis: options.wasm ? 'OGVDecoderAudioVorbisW' : 'OGVDecoderAudioVorbis',
opus: options.wasm ? 'OGVDecoderAudioOpusW' : 'OGVDecoderAudioOpus'
};
function loadAudioCodec(callback) {
if (demuxer.audioCodec) {
var className = audioClassMap[demuxer.audioCodec];
processing = true;
OGVLoader.loadClass(className, function(audioCodecClass) {
var audioOptions = {};
if (demuxer.audioFormat) {
audioOptions.audioFormat = demuxer.audioFormat;
}
audioDecoder = new audioCodecClass(audioOptions);
audioDecoder.init(function() {
loadedAudioMetadata = audioDecoder.loadedMetadata;
processing = false;
callback();
});
}, {
worker: options.worker
});
} else {
callback();
}
}
var videoClassMap = {
theora: options.wasm ? 'OGVDecoderVideoTheoraW' : 'OGVDecoderVideoTheora',
vp8: options.wasm ? 'OGVDecoderVideoVP8W' : (options.threading ? 'OGVDecoderVideoVP8MT' : 'OGVDecoderVideoVP8'),
vp9: options.wasm ? 'OGVDecoderVideoVP9W' : (options.threading ? 'OGVDecoderVideoVP9MT' : 'OGVDecoderVideoVP9')
};
function loadVideoCodec(callback) {
if (demuxer.videoCodec) {
var className = videoClassMap[demuxer.videoCodec];
processing = true;
OGVLoader.loadClass(className, function(videoCodecClass) {
var videoOptions = {};
if (demuxer.videoFormat) {
videoOptions.videoFormat = demuxer.videoFormat;
}
if (options.memoryLimit) {
videoOptions.memoryLimit = options.memoryLimit;
}
videoDecoder = new videoCodecClass(videoOptions);
videoDecoder.init(function() {
loadedVideoMetadata = videoDecoder.loadedMetadata;
processing = false;
callback();
});
}, {
worker: options.worker && !options.threading
});
} else {
callback();
}
}
var loadedDemuxerMetadata = false,
loadedAudioMetadata = false,
loadedVideoMetadata = false,
loadedAllMetadata = false;
self.process = function(callback) {
if (processing) {
throw new Error('reentrancy fail on OGVWrapperCodec.process');
}
processing = true;
var videoPacketCount = demuxer.videoPackets.length,
audioPacketCount = demuxer.audioPackets.length,
start = (window.performance ? performance.now() : Date.now());
function finish(result) {
processing = false;
var delta = (window.performance ? performance.now() : Date.now()) - start;
if (delta > 8) {
console.log('slow demux in ' + delta + ' ms; ' +
(demuxer.videoPackets.length - videoPacketCount) + ' +video packets, ' +
(demuxer.audioPackets.length - audioPacketCount) + ' +audio packets');
}
//console.log('demux returned ' + (result ? 'true' : 'false') + '; ' + demuxer.videoPackets.length + '; ' + demuxer.audioPackets.length);
callback(result);
}
function doProcessData() {
demuxer.process(finish);
}
if (demuxer.loadedMetadata && !loadedDemuxerMetadata) {
// Demuxer just reached its metadata. Load the relevant codecs!
loadAudioCodec(function() {
loadVideoCodec(function() {
loadedDemuxerMetadata = true;
loadedAudioMetadata = !audioDecoder;
loadedVideoMetadata = !videoDecoder;
loadedAllMetadata = loadedAudioMetadata && loadedVideoMetadata;
finish(true);
});
});
} else if (loadedDemuxerMetadata && !loadedAudioMetadata) {
if (audioDecoder.loadedMetadata) {
loadedAudioMetadata = true;
loadedAllMetadata = loadedAudioMetadata && loadedVideoMetadata;
finish(true);
} else if (demuxer.audioReady) {
demuxer.dequeueAudioPacket(function(packet) {
audioDecoder.processHeader(packet, function(ret) {
finish(true);
});
});
} else {
doProcessData();
}
} else if (loadedAudioMetadata && !loadedVideoMetadata) {
if (videoDecoder.loadedMetadata) {
loadedVideoMetadata = true;
loadedAllMetadata = loadedAudioMetadata && loadedVideoMetadata;
finish(true);
} else if (demuxer.frameReady) {
processing = true;
demuxer.dequeueVideoPacket(function(packet) {
videoDecoder.processHeader(packet, function() {
finish(true);
});
});
} else {
doProcessData();
}
} else if (loadedVideoMetadata && !self.loadedMetadata && loadedAllMetadata) {
// Ok we've found all the metadata there is. Enjoy.
loadedMetadata = true;
finish(true);
} else if (self.loadedMetadata && (!self.hasAudio || demuxer.audioReady) && (!self.hasVideo || demuxer.frameReady)) {
// Already queued up some packets. Go read them!
finish(true);
} else {
// We need to process more of the data we've already received,
// or ask for more if we ran out!
doProcessData();
}
};
self.decodeFrame = function(callback) {
var cb = flushSafe(callback),
timestamp = self.frameTimestamp,
keyframeTimestamp = self.keyframeTimestamp;
demuxer.dequeueVideoPacket(function(packet) {
videoDecoder.processFrame(packet, function(ok) {
// hack
if (videoDecoder.frameBuffer) {
videoDecoder.frameBuffer.timestamp = timestamp;
videoDecoder.frameBuffer.keyframeTimestamp = keyframeTimestamp;
}
cb(ok);
});
});
};
self.decodeAudio = function(callback) {
var cb = flushSafe(callback);
demuxer.dequeueAudioPacket(function(packet) {
audioDecoder.processAudio(packet, cb);
});
}
self.discardFrame = function(callback) {
demuxer.dequeueVideoPacket(function(packet) {
callback();
});
};
self.discardAudio = function(callback) {
demuxer.dequeueAudioPacket(function(packet) {
callback();
});
};
self.flush = function(callback) {
flushIter++;
demuxer.flush(callback);
};
self.getKeypointOffset = function(timeSeconds, callback) {
demuxer.getKeypointOffset(timeSeconds, callback);
};
self.seekToKeypoint = function(timeSeconds, callback) {
demuxer.seekToKeypoint(timeSeconds, flushSafe(callback));
}
self.onseek = null;
Object.defineProperty(self, "demuxerCpuTime", {
get: function() {
if (demuxer) {
return demuxer.cpuTime;
} else {
return 0;
}
}
});
Object.defineProperty(self, "audioCpuTime", {
get: function() {
if (audioDecoder) {
return audioDecoder.cpuTime;
} else {
return 0;
}
}
});
Object.defineProperty(self, "videoCpuTime", {
get: function() {
if (videoDecoder) {
return videoDecoder.cpuTime;
} else {
return 0;
}
}
});
return self;
});
module.exports = OGVWrapperCodec;
/***/ })
/******/ ])
});
;