mirror of
https://github.com/invoiceninja/invoiceninja.git
synced 2024-11-08 20:22:42 +01:00
6664 lines
251 KiB
JavaScript
6664 lines
251 KiB
JavaScript
|
/** @preserve jsPDF 0.9.0rc2 ( 2013-08-07T15:00 commit ID c9c47d1de98fabb0681ad9fba049ef644f8f22ba )
|
||
|
Copyright (c) 2010-2012 James Hall, james@snapshotmedia.co.uk, https://github.com/MrRio/jsPDF
|
||
|
Copyright (c) 2012 Willow Systems Corporation, willow-systems.com
|
||
|
MIT license.
|
||
|
*/
|
||
|
|
||
|
/*
|
||
|
* Permission is hereby granted, free of charge, to any person obtaining
|
||
|
* a copy of this software and associated documentation files (the
|
||
|
* "Software"), to deal in the Software without restriction, including
|
||
|
* without limitation the rights to use, copy, modify, merge, publish,
|
||
|
* distribute, sublicense, and/or sell copies of the Software, and to
|
||
|
* permit persons to whom the Software is furnished to do so, subject to
|
||
|
* the following conditions:
|
||
|
*
|
||
|
* The above copyright notice and this permission notice shall be
|
||
|
* included in all copies or substantial portions of the Software.
|
||
|
*
|
||
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
||
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
||
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
||
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
||
|
* LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
||
|
* OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
||
|
* WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
||
|
* ====================================================================
|
||
|
*/
|
||
|
|
||
|
|
||
|
/**
|
||
|
Creates new jsPDF document object instance
|
||
|
@class
|
||
|
@param orientation One of "portrait" or "landscape" (or shortcuts "p" (Default), "l")
|
||
|
@param unit Measurement unit to be used when coordinates are specified. One of "pt" (points), "mm" (Default), "cm", "in"
|
||
|
@param format One of 'a0', 'a1', 'a2', 'a3', 'a4' (Default) etc to 'a10', 'b0' to 'b10', 'c0' to 'c10', 'letter', 'government-letter', 'legal', 'junior-legal', 'ledger' or 'tabloid'
|
||
|
@returns {jsPDF}
|
||
|
@name jsPDF
|
||
|
*/
|
||
|
var jsPDF = (function () {
|
||
|
'use strict';
|
||
|
/*jslint browser:true, plusplus: true, bitwise: true, nomen: true */
|
||
|
/*global document: false, btoa, atob, zpipe, Uint8Array, ArrayBuffer, Blob, saveAs, adler32cs, Deflater */
|
||
|
|
||
|
// this will run on <=IE9, possibly some niche browsers
|
||
|
// new webkit-based, FireFox, IE10 already have native version of this.
|
||
|
if (typeof btoa === 'undefined') {
|
||
|
window.btoa = function (data) {
|
||
|
// DO NOT ADD UTF8 ENCODING CODE HERE!!!!
|
||
|
|
||
|
// UTF8 encoding encodes bytes over char code 128
|
||
|
// and, essentially, turns an 8-bit binary streams
|
||
|
// (that base64 can deal with) into 7-bit binary streams.
|
||
|
// (by default server does not know that and does not recode the data back to 8bit)
|
||
|
// You destroy your data.
|
||
|
|
||
|
// binary streams like jpeg image data etc, while stored in JavaScript strings,
|
||
|
// (which are 16bit arrays) are in 8bit format already.
|
||
|
// You do NOT need to char-encode that before base64 encoding.
|
||
|
|
||
|
// if you, by act of fate
|
||
|
// have string which has individual characters with code
|
||
|
// above 255 (pure unicode chars), encode that BEFORE you base64 here.
|
||
|
// you can use absolutely any approch there, as long as in the end,
|
||
|
// base64 gets an 8bit (char codes 0 - 255) stream.
|
||
|
// when you get it on the server after un-base64, you must
|
||
|
// UNencode it too, to get back to 16, 32bit or whatever original bin stream.
|
||
|
|
||
|
// Note, Yes, JavaScript strings are, in most cases UCS-2 -
|
||
|
// 16-bit character arrays. This does not mean, however,
|
||
|
// that you always have to UTF8 it before base64.
|
||
|
// it means that if you have actual characters anywhere in
|
||
|
// that string that have char code above 255, you need to
|
||
|
// recode *entire* string from 16-bit (or 32bit) to 8-bit array.
|
||
|
// You can do binary split to UTF16 (BE or LE)
|
||
|
// you can do utf8, you can split the thing by hand and prepend BOM to it,
|
||
|
// but whatever you do, make sure you mirror the opposite on
|
||
|
// the server. If server does not expect to post-process un-base64
|
||
|
// 8-bit binary stream, think very very hard about messing around with encoding.
|
||
|
|
||
|
// so, long story short:
|
||
|
// DO NOT ADD UTF8 ENCODING CODE HERE!!!!
|
||
|
|
||
|
/* @preserve
|
||
|
====================================================================
|
||
|
base64 encoder
|
||
|
MIT, GPL
|
||
|
|
||
|
version: 1109.2015
|
||
|
discuss at: http://phpjs.org/functions/base64_encode
|
||
|
+ original by: Tyler Akins (http://rumkin.com)
|
||
|
+ improved by: Bayron Guevara
|
||
|
+ improved by: Thunder.m
|
||
|
+ improved by: Kevin van Zonneveld (http://kevin.vanzonneveld.net)
|
||
|
+ bugfixed by: Pellentesque Malesuada
|
||
|
+ improved by: Kevin van Zonneveld (http://kevin.vanzonneveld.net)
|
||
|
+ improved by: Rafal Kukawski (http://kukawski.pl)
|
||
|
+ Daniel Dotsenko, Willow Systems Corp, willow-systems.com
|
||
|
====================================================================
|
||
|
*/
|
||
|
|
||
|
var b64 = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/=",
|
||
|
b64a = b64.split(''),
|
||
|
o1,
|
||
|
o2,
|
||
|
o3,
|
||
|
h1,
|
||
|
h2,
|
||
|
h3,
|
||
|
h4,
|
||
|
bits,
|
||
|
i = 0,
|
||
|
ac = 0,
|
||
|
enc = "",
|
||
|
tmp_arr = [],
|
||
|
r;
|
||
|
|
||
|
do { // pack three octets into four hexets
|
||
|
o1 = data.charCodeAt(i++);
|
||
|
o2 = data.charCodeAt(i++);
|
||
|
o3 = data.charCodeAt(i++);
|
||
|
|
||
|
bits = o1 << 16 | o2 << 8 | o3;
|
||
|
|
||
|
h1 = bits >> 18 & 0x3f;
|
||
|
h2 = bits >> 12 & 0x3f;
|
||
|
h3 = bits >> 6 & 0x3f;
|
||
|
h4 = bits & 0x3f;
|
||
|
|
||
|
// use hexets to index into b64, and append result to encoded string
|
||
|
tmp_arr[ac++] = b64a[h1] + b64a[h2] + b64a[h3] + b64a[h4];
|
||
|
} while (i < data.length);
|
||
|
|
||
|
enc = tmp_arr.join('');
|
||
|
r = data.length % 3;
|
||
|
return (r ? enc.slice(0, r - 3) : enc) + '==='.slice(r || 3);
|
||
|
// end of base64 encoder MIT, GPL
|
||
|
};
|
||
|
}
|
||
|
|
||
|
if (typeof atob === 'undefined') {
|
||
|
window.atob = function (data) {
|
||
|
// http://kevin.vanzonneveld.net
|
||
|
// + original by: Tyler Akins (http://rumkin.com)
|
||
|
// + improved by: Thunder.m
|
||
|
// + input by: Aman Gupta
|
||
|
// + improved by: Kevin van Zonneveld (http://kevin.vanzonneveld.net)
|
||
|
// + bugfixed by: Onno Marsman
|
||
|
// + bugfixed by: Pellentesque Malesuada
|
||
|
// + improved by: Kevin van Zonneveld (http://kevin.vanzonneveld.net)
|
||
|
// + input by: Brett Zamir (http://brett-zamir.me)
|
||
|
// + bugfixed by: Kevin van Zonneveld (http://kevin.vanzonneveld.net)
|
||
|
// * example 1: base64_decode('S2V2aW4gdmFuIFpvbm5ldmVsZA==');
|
||
|
// * returns 1: 'Kevin van Zonneveld'
|
||
|
// mozilla has this native
|
||
|
// - but breaks in 2.0.0.12!
|
||
|
//if (typeof this.window['atob'] == 'function') {
|
||
|
// return atob(data);
|
||
|
//}
|
||
|
var b64 = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/=",
|
||
|
o1,
|
||
|
o2,
|
||
|
o3,
|
||
|
h1,
|
||
|
h2,
|
||
|
h3,
|
||
|
h4,
|
||
|
bits,
|
||
|
i = 0,
|
||
|
ac = 0,
|
||
|
dec = "",
|
||
|
tmp_arr = [];
|
||
|
|
||
|
if (!data) {
|
||
|
return data;
|
||
|
}
|
||
|
|
||
|
data += '';
|
||
|
|
||
|
do { // unpack four hexets into three octets using index points in b64
|
||
|
h1 = b64.indexOf(data.charAt(i++));
|
||
|
h2 = b64.indexOf(data.charAt(i++));
|
||
|
h3 = b64.indexOf(data.charAt(i++));
|
||
|
h4 = b64.indexOf(data.charAt(i++));
|
||
|
|
||
|
bits = h1 << 18 | h2 << 12 | h3 << 6 | h4;
|
||
|
|
||
|
o1 = bits >> 16 & 0xff;
|
||
|
o2 = bits >> 8 & 0xff;
|
||
|
o3 = bits & 0xff;
|
||
|
|
||
|
if (h3 === 64) {
|
||
|
tmp_arr[ac++] = String.fromCharCode(o1);
|
||
|
} else if (h4 === 64) {
|
||
|
tmp_arr[ac++] = String.fromCharCode(o1, o2);
|
||
|
} else {
|
||
|
tmp_arr[ac++] = String.fromCharCode(o1, o2, o3);
|
||
|
}
|
||
|
} while (i < data.length);
|
||
|
dec = tmp_arr.join('');
|
||
|
return dec;
|
||
|
};
|
||
|
}
|
||
|
|
||
|
var getObjectLength = typeof Object.keys === 'function' ?
|
||
|
function (object) {
|
||
|
return Object.keys(object).length;
|
||
|
} :
|
||
|
function (object) {
|
||
|
var i = 0, e;
|
||
|
for (e in object) {
|
||
|
if (object.hasOwnProperty(e)) {
|
||
|
i++;
|
||
|
}
|
||
|
}
|
||
|
return i;
|
||
|
},
|
||
|
|
||
|
/**
|
||
|
PubSub implementation
|
||
|
|
||
|
@class
|
||
|
@name PubSub
|
||
|
*/
|
||
|
PubSub = function (context) {
|
||
|
/** @preserve
|
||
|
-----------------------------------------------------------------------------------------------
|
||
|
JavaScript PubSub library
|
||
|
2012 (c) ddotsenko@willowsystems.com
|
||
|
based on Peter Higgins (dante@dojotoolkit.org)
|
||
|
Loosely based on Dojo publish/subscribe API, limited in scope. Rewritten blindly.
|
||
|
Original is (c) Dojo Foundation 2004-2010. Released under either AFL or new BSD, see:
|
||
|
http://dojofoundation.org/license for more information.
|
||
|
-----------------------------------------------------------------------------------------------
|
||
|
*/
|
||
|
/**
|
||
|
@private
|
||
|
@fieldOf PubSub
|
||
|
*/
|
||
|
this.topics = {};
|
||
|
/**
|
||
|
Stores what will be `this` within the callback functions.
|
||
|
|
||
|
@private
|
||
|
@fieldOf PubSub#
|
||
|
*/
|
||
|
this.context = context;
|
||
|
/**
|
||
|
Allows caller to emit an event and pass arguments to event listeners.
|
||
|
@public
|
||
|
@function
|
||
|
@param topic {String} Name of the channel on which to voice this event
|
||
|
@param args Any number of arguments you want to pass to the listeners of this event.
|
||
|
@methodOf PubSub#
|
||
|
@name publish
|
||
|
*/
|
||
|
this.publish = function (topic, args) {
|
||
|
if (this.topics[topic]) {
|
||
|
var currentTopic = this.topics[topic],
|
||
|
toremove = [],
|
||
|
fn,
|
||
|
i,
|
||
|
l,
|
||
|
pair,
|
||
|
emptyFunc = function () {};
|
||
|
args = Array.prototype.slice.call(arguments, 1);
|
||
|
for (i = 0, l = currentTopic.length; i < l; i++) {
|
||
|
pair = currentTopic[i]; // this is a [function, once_flag] array
|
||
|
fn = pair[0];
|
||
|
if (pair[1]) { /* 'run once' flag set */
|
||
|
pair[0] = emptyFunc;
|
||
|
toremove.push(i);
|
||
|
}
|
||
|
fn.apply(this.context, args);
|
||
|
}
|
||
|
for (i = 0, l = toremove.length; i < l; i++) {
|
||
|
currentTopic.splice(toremove[i], 1);
|
||
|
}
|
||
|
}
|
||
|
};
|
||
|
/**
|
||
|
Allows listener code to subscribe to channel and be called when data is available
|
||
|
@public
|
||
|
@function
|
||
|
@param topic {String} Name of the channel on which to voice this event
|
||
|
@param callback {Function} Executable (function pointer) that will be ran when event is voiced on this channel.
|
||
|
@param once {Boolean} (optional. False by default) Flag indicating if the function is to be triggered only once.
|
||
|
@returns {Object} A token object that cen be used for unsubscribing.
|
||
|
@methodOf PubSub#
|
||
|
@name subscribe
|
||
|
*/
|
||
|
this.subscribe = function (topic, callback, once) {
|
||
|
if (!this.topics[topic]) {
|
||
|
this.topics[topic] = [[callback, once]];
|
||
|
} else {
|
||
|
this.topics[topic].push([callback, once]);
|
||
|
}
|
||
|
return {
|
||
|
"topic": topic,
|
||
|
"callback": callback
|
||
|
};
|
||
|
};
|
||
|
/**
|
||
|
Allows listener code to unsubscribe from a channel
|
||
|
@public
|
||
|
@function
|
||
|
@param token {Object} A token object that was returned by `subscribe` method
|
||
|
@methodOf PubSub#
|
||
|
@name unsubscribe
|
||
|
*/
|
||
|
this.unsubscribe = function (token) {
|
||
|
if (this.topics[token.topic]) {
|
||
|
var currentTopic = this.topics[token.topic], i, l;
|
||
|
|
||
|
for (i = 0, l = currentTopic.length; i < l; i++) {
|
||
|
if (currentTopic[i][0] === token.callback) {
|
||
|
currentTopic.splice(i, 1);
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
};
|
||
|
};
|
||
|
|
||
|
|
||
|
/**
|
||
|
@constructor
|
||
|
@private
|
||
|
*/
|
||
|
function jsPDF(orientation, unit, format, compressPdf) { /** String orientation, String unit, String format, Boolean compressed */
|
||
|
|
||
|
// Default parameter values
|
||
|
if (typeof orientation === 'undefined') {
|
||
|
orientation = 'p';
|
||
|
} else {
|
||
|
orientation = orientation.toString().toLowerCase();
|
||
|
}
|
||
|
if (typeof unit === 'undefined') { unit = 'mm'; }
|
||
|
if (typeof format === 'undefined') { format = 'a4'; }
|
||
|
if (typeof compressPdf === 'undefined' && typeof zpipe === 'undefined') { compressPdf = false; }
|
||
|
|
||
|
var format_as_string = format.toString().toLowerCase(),
|
||
|
version = '0.9.0rc2',
|
||
|
content = [],
|
||
|
content_length = 0,
|
||
|
compress = compressPdf,
|
||
|
pdfVersion = '1.3', // PDF Version
|
||
|
pageFormats = { // Size in pt of various paper formats
|
||
|
'a0': [2383.94, 3370.39],
|
||
|
'a1': [1683.78, 2383.94],
|
||
|
'a2': [1190.55, 1683.78],
|
||
|
'a3': [841.89, 1190.55],
|
||
|
'a4': [595.28, 841.89],
|
||
|
'a5': [419.53, 595.28],
|
||
|
'a6': [297.64, 419.53],
|
||
|
'a7': [209.76, 297.64],
|
||
|
'a8': [147.4 , 209.76],
|
||
|
'a9': [104.88, 147.4],
|
||
|
'a10': [73.7, 104.88],
|
||
|
'b0': [2834.65, 4008.19],
|
||
|
'b1': [2004.09, 2834.65],
|
||
|
'b2': [1417.32, 2004.09],
|
||
|
'b3': [1000.63, 1417.32],
|
||
|
'b4': [708.66, 1000.63],
|
||
|
'b5': [498.9, 708.66],
|
||
|
'b6': [354.33, 498.9],
|
||
|
'b7': [249.45, 354.33],
|
||
|
'b8': [175.75, 249.45],
|
||
|
'b9': [124.72, 175.75],
|
||
|
'b10': [87.87, 124.72],
|
||
|
'c0': [2599.37, 3676.54],
|
||
|
'c1': [1836.85, 2599.37],
|
||
|
'c2': [1298.27, 1836.85],
|
||
|
'c3': [918.43, 1298.27],
|
||
|
'c4': [649.13, 918.43],
|
||
|
'c5': [459.21, 649.13],
|
||
|
'c6': [323.15, 459.21],
|
||
|
'c7': [229.61, 323.15],
|
||
|
'c8': [161.57, 229.61],
|
||
|
'c9': [113.39, 161.57],
|
||
|
'c10': [79.37, 113.39],
|
||
|
'letter': [612, 792],
|
||
|
'government-letter': [576, 756],
|
||
|
'legal': [612, 1008],
|
||
|
'junior-legal': [576, 360],
|
||
|
'ledger': [1224, 792],
|
||
|
'tabloid': [792, 1224]
|
||
|
},
|
||
|
textColor = '0 g',
|
||
|
drawColor = '0 G',
|
||
|
page = 0,
|
||
|
pages = [],
|
||
|
objectNumber = 2, // 'n' Current object number
|
||
|
outToPages = false, // switches where out() prints. outToPages true = push to pages obj. outToPages false = doc builder content
|
||
|
offsets = [], // List of offsets. Activated and reset by buildDocument(). Pupulated by various calls buildDocument makes.
|
||
|
fonts = {}, // collection of font objects, where key is fontKey - a dynamically created label for a given font.
|
||
|
fontmap = {}, // mapping structure fontName > fontStyle > font key - performance layer. See addFont()
|
||
|
activeFontSize = 16,
|
||
|
activeFontKey, // will be string representing the KEY of the font as combination of fontName + fontStyle
|
||
|
lineWidth = 0.200025, // 2mm
|
||
|
lineHeightProportion = 1.15,
|
||
|
pageHeight,
|
||
|
pageWidth,
|
||
|
k, // Scale factor
|
||
|
documentProperties = {'title': '', 'subject': '', 'author': '', 'keywords': '', 'creator': ''},
|
||
|
lineCapID = 0,
|
||
|
lineJoinID = 0,
|
||
|
API = {},
|
||
|
events = new PubSub(API),
|
||
|
tmp,
|
||
|
plugin,
|
||
|
/////////////////////
|
||
|
// Private functions
|
||
|
/////////////////////
|
||
|
// simplified (speedier) replacement for sprintf's %.2f conversion
|
||
|
f2 = function (number) {
|
||
|
return number.toFixed(2);
|
||
|
},
|
||
|
// simplified (speedier) replacement for sprintf's %.3f conversion
|
||
|
f3 = function (number) {
|
||
|
return number.toFixed(3);
|
||
|
},
|
||
|
// simplified (speedier) replacement for sprintf's %02d
|
||
|
padd2 = function (number) {
|
||
|
var n = (number).toFixed(0);
|
||
|
if (number < 10) {
|
||
|
return '0' + n;
|
||
|
} else {
|
||
|
return n;
|
||
|
}
|
||
|
},
|
||
|
// simplified (speedier) replacement for sprintf's %02d
|
||
|
padd10 = function (number) {
|
||
|
var n = (number).toFixed(0);
|
||
|
if (n.length < 10) {
|
||
|
return new Array( 11 - n.length ).join('0') + n;
|
||
|
} else {
|
||
|
return n;
|
||
|
}
|
||
|
},
|
||
|
out = function (string) {
|
||
|
if (outToPages) { /* set by beginPage */
|
||
|
pages[page].push(string);
|
||
|
} else {
|
||
|
content.push(string);
|
||
|
content_length += string.length + 1; // +1 is for '\n' that will be used to join contents of content
|
||
|
}
|
||
|
},
|
||
|
newObject = function () {
|
||
|
// Begin a new object
|
||
|
objectNumber++;
|
||
|
offsets[objectNumber] = content_length;
|
||
|
out(objectNumber + ' 0 obj');
|
||
|
return objectNumber;
|
||
|
},
|
||
|
putStream = function (str) {
|
||
|
out('stream');
|
||
|
out(str);
|
||
|
out('endstream');
|
||
|
},
|
||
|
wPt,
|
||
|
hPt,
|
||
|
kids,
|
||
|
i,
|
||
|
putPages = function () {
|
||
|
wPt = pageWidth * k;
|
||
|
hPt = pageHeight * k;
|
||
|
|
||
|
// outToPages = false as set in endDocument(). out() writes to content.
|
||
|
|
||
|
var n, p, arr, uint, i, deflater, adler32;
|
||
|
for (n = 1; n <= page; n++) {
|
||
|
newObject();
|
||
|
out('<</Type /Page');
|
||
|
out('/Parent 1 0 R');
|
||
|
out('/Resources 2 0 R');
|
||
|
out('/Contents ' + (objectNumber + 1) + ' 0 R>>');
|
||
|
out('endobj');
|
||
|
|
||
|
// Page content
|
||
|
p = pages[n].join('\n');
|
||
|
newObject();
|
||
|
if (compress) {
|
||
|
arr = [];
|
||
|
for (i = 0; i < p.length; ++i) {
|
||
|
arr[i] = p.charCodeAt(i);
|
||
|
}
|
||
|
adler32 = adler32cs.from(p);
|
||
|
deflater = new Deflater(6);
|
||
|
deflater.append(new Uint8Array(arr));
|
||
|
p = deflater.flush();
|
||
|
arr = [new Uint8Array([120, 156]), new Uint8Array(p),
|
||
|
new Uint8Array([adler32 & 0xFF, (adler32 >> 8) & 0xFF, (adler32 >> 16) & 0xFF, (adler32 >> 24) & 0xFF])];
|
||
|
p = '';
|
||
|
for (i in arr) {
|
||
|
if (arr.hasOwnProperty(i)) {
|
||
|
p += String.fromCharCode.apply(null, arr[i]);
|
||
|
}
|
||
|
}
|
||
|
out('<</Length ' + p.length + ' /Filter [/FlateDecode]>>');
|
||
|
} else {
|
||
|
out('<</Length ' + p.length + '>>');
|
||
|
}
|
||
|
putStream(p);
|
||
|
out('endobj');
|
||
|
}
|
||
|
offsets[1] = content_length;
|
||
|
out('1 0 obj');
|
||
|
out('<</Type /Pages');
|
||
|
kids = '/Kids [';
|
||
|
for (i = 0; i < page; i++) {
|
||
|
kids += (3 + 2 * i) + ' 0 R ';
|
||
|
}
|
||
|
out(kids + ']');
|
||
|
out('/Count ' + page);
|
||
|
out('/MediaBox [0 0 ' + f2(wPt) + ' ' + f2(hPt) + ']');
|
||
|
out('>>');
|
||
|
out('endobj');
|
||
|
},
|
||
|
putFont = function (font) {
|
||
|
font.objectNumber = newObject();
|
||
|
out('<</BaseFont/' + font.PostScriptName + '/Type/Font');
|
||
|
if (typeof font.encoding === 'string') {
|
||
|
out('/Encoding/' + font.encoding);
|
||
|
}
|
||
|
out('/Subtype/Type1>>');
|
||
|
out('endobj');
|
||
|
},
|
||
|
putFonts = function () {
|
||
|
var fontKey;
|
||
|
for (fontKey in fonts) {
|
||
|
if (fonts.hasOwnProperty(fontKey)) {
|
||
|
putFont(fonts[fontKey]);
|
||
|
}
|
||
|
}
|
||
|
},
|
||
|
putXobjectDict = function () {
|
||
|
// Loop through images, or other data objects
|
||
|
events.publish('putXobjectDict');
|
||
|
},
|
||
|
putResourceDictionary = function () {
|
||
|
out('/ProcSet [/PDF /Text /ImageB /ImageC /ImageI]');
|
||
|
out('/Font <<');
|
||
|
// Do this for each font, the '1' bit is the index of the font
|
||
|
var fontKey;
|
||
|
for (fontKey in fonts) {
|
||
|
if (fonts.hasOwnProperty(fontKey)) {
|
||
|
out('/' + fontKey + ' ' + fonts[fontKey].objectNumber + ' 0 R');
|
||
|
}
|
||
|
}
|
||
|
out('>>');
|
||
|
out('/XObject <<');
|
||
|
putXobjectDict();
|
||
|
out('>>');
|
||
|
},
|
||
|
putResources = function () {
|
||
|
putFonts();
|
||
|
events.publish('putResources');
|
||
|
// Resource dictionary
|
||
|
offsets[2] = content_length;
|
||
|
out('2 0 obj');
|
||
|
out('<<');
|
||
|
putResourceDictionary();
|
||
|
out('>>');
|
||
|
out('endobj');
|
||
|
events.publish('postPutResources');
|
||
|
},
|
||
|
addToFontDictionary = function (fontKey, fontName, fontStyle) {
|
||
|
// this is mapping structure for quick font key lookup.
|
||
|
// returns the KEY of the font (ex: "F1") for a given pair of font name and type (ex: "Arial". "Italic")
|
||
|
var undef;
|
||
|
if (fontmap[fontName] === undef) {
|
||
|
fontmap[fontName] = {}; // fontStyle is a var interpreted and converted to appropriate string. don't wrap in quotes.
|
||
|
}
|
||
|
fontmap[fontName][fontStyle] = fontKey;
|
||
|
},
|
||
|
/**
|
||
|
FontObject describes a particular font as member of an instnace of jsPDF
|
||
|
|
||
|
It's a collection of properties like 'id' (to be used in PDF stream),
|
||
|
'fontName' (font's family name), 'fontStyle' (font's style variant label)
|
||
|
|
||
|
@class
|
||
|
@public
|
||
|
@property id {String} PDF-document-instance-specific label assinged to the font.
|
||
|
@property PostScriptName {String} PDF specification full name for the font
|
||
|
@property encoding {Object} Encoding_name-to-Font_metrics_object mapping.
|
||
|
@name FontObject
|
||
|
*/
|
||
|
FontObject = {},
|
||
|
addFont = function (PostScriptName, fontName, fontStyle, encoding) {
|
||
|
var fontKey = 'F' + (getObjectLength(fonts) + 1).toString(10),
|
||
|
// This is FontObject
|
||
|
font = fonts[fontKey] = {
|
||
|
'id': fontKey,
|
||
|
// , 'objectNumber': will be set by putFont()
|
||
|
'PostScriptName': PostScriptName,
|
||
|
'fontName': fontName,
|
||
|
'fontStyle': fontStyle,
|
||
|
'encoding': encoding,
|
||
|
'metadata': {}
|
||
|
};
|
||
|
|
||
|
addToFontDictionary(fontKey, fontName, fontStyle);
|
||
|
|
||
|
events.publish('addFont', font);
|
||
|
|
||
|
return fontKey;
|
||
|
},
|
||
|
addFonts = function () {
|
||
|
|
||
|
var HELVETICA = "helvetica",
|
||
|
TIMES = "times",
|
||
|
COURIER = "courier",
|
||
|
NORMAL = "normal",
|
||
|
BOLD = "bold",
|
||
|
ITALIC = "italic",
|
||
|
BOLD_ITALIC = "bolditalic",
|
||
|
encoding = 'StandardEncoding',
|
||
|
standardFonts = [
|
||
|
['Helvetica', HELVETICA, NORMAL],
|
||
|
['Helvetica-Bold', HELVETICA, BOLD],
|
||
|
['Helvetica-Oblique', HELVETICA, ITALIC],
|
||
|
['Helvetica-BoldOblique', HELVETICA, BOLD_ITALIC],
|
||
|
['Courier', COURIER, NORMAL],
|
||
|
['Courier-Bold', COURIER, BOLD],
|
||
|
['Courier-Oblique', COURIER, ITALIC],
|
||
|
['Courier-BoldOblique', COURIER, BOLD_ITALIC],
|
||
|
['Times-Roman', TIMES, NORMAL],
|
||
|
['Times-Bold', TIMES, BOLD],
|
||
|
['Times-Italic', TIMES, ITALIC],
|
||
|
['Times-BoldItalic', TIMES, BOLD_ITALIC]
|
||
|
],
|
||
|
i,
|
||
|
l,
|
||
|
fontKey,
|
||
|
parts;
|
||
|
for (i = 0, l = standardFonts.length; i < l; i++) {
|
||
|
fontKey = addFont(
|
||
|
standardFonts[i][0],
|
||
|
standardFonts[i][1],
|
||
|
standardFonts[i][2],
|
||
|
encoding
|
||
|
);
|
||
|
|
||
|
// adding aliases for standard fonts, this time matching the capitalization
|
||
|
parts = standardFonts[i][0].split('-');
|
||
|
addToFontDictionary(fontKey, parts[0], parts[1] || '');
|
||
|
}
|
||
|
|
||
|
events.publish('addFonts', {'fonts': fonts, 'dictionary': fontmap});
|
||
|
},
|
||
|
/**
|
||
|
|
||
|
@public
|
||
|
@function
|
||
|
@param text {String}
|
||
|
@param flags {Object} Encoding flags.
|
||
|
@returns {String} Encoded string
|
||
|
*/
|
||
|
to8bitStream = function (text, flags) {
|
||
|
/* PDF 1.3 spec:
|
||
|
"For text strings encoded in Unicode, the first two bytes must be 254 followed by
|
||
|
255, representing the Unicode byte order marker, U+FEFF. (This sequence conflicts
|
||
|
with the PDFDocEncoding character sequence thorn ydieresis, which is unlikely
|
||
|
to be a meaningful beginning of a word or phrase.) The remainder of the
|
||
|
string consists of Unicode character codes, according to the UTF-16 encoding
|
||
|
specified in the Unicode standard, version 2.0. Commonly used Unicode values
|
||
|
are represented as 2 bytes per character, with the high-order byte appearing first
|
||
|
in the string."
|
||
|
|
||
|
In other words, if there are chars in a string with char code above 255, we
|
||
|
recode the string to UCS2 BE - string doubles in length and BOM is prepended.
|
||
|
|
||
|
HOWEVER!
|
||
|
Actual *content* (body) text (as opposed to strings used in document properties etc)
|
||
|
does NOT expect BOM. There, it is treated as a literal GID (Glyph ID)
|
||
|
|
||
|
Because of Adobe's focus on "you subset your fonts!" you are not supposed to have
|
||
|
a font that maps directly Unicode (UCS2 / UTF16BE) code to font GID, but you could
|
||
|
fudge it with "Identity-H" encoding and custom CIDtoGID map that mimics Unicode
|
||
|
code page. There, however, all characters in the stream are treated as GIDs,
|
||
|
including BOM, which is the reason we need to skip BOM in content text (i.e. that
|
||
|
that is tied to a font).
|
||
|
|
||
|
To signal this "special" PDFEscape / to8bitStream handling mode,
|
||
|
API.text() function sets (unless you overwrite it with manual values
|
||
|
given to API.text(.., flags) )
|
||
|
flags.autoencode = true
|
||
|
flags.noBOM = true
|
||
|
|
||
|
*/
|
||
|
|
||
|
/*
|
||
|
`flags` properties relied upon:
|
||
|
.sourceEncoding = string with encoding label.
|
||
|
"Unicode" by default. = encoding of the incoming text.
|
||
|
pass some non-existing encoding name
|
||
|
(ex: 'Do not touch my strings! I know what I am doing.')
|
||
|
to make encoding code skip the encoding step.
|
||
|
.outputEncoding = Either valid PDF encoding name
|
||
|
(must be supported by jsPDF font metrics, otherwise no encoding)
|
||
|
or a JS object, where key = sourceCharCode, value = outputCharCode
|
||
|
missing keys will be treated as: sourceCharCode === outputCharCode
|
||
|
.noBOM
|
||
|
See comment higher above for explanation for why this is important
|
||
|
.autoencode
|
||
|
See comment higher above for explanation for why this is important
|
||
|
*/
|
||
|
|
||
|
var i, l, undef, sourceEncoding, encodingBlock, outputEncoding, newtext, isUnicode, ch, bch;
|
||
|
|
||
|
if (flags === undef) {
|
||
|
flags = {};
|
||
|
}
|
||
|
|
||
|
sourceEncoding = flags.sourceEncoding ? sourceEncoding : 'Unicode';
|
||
|
|
||
|
outputEncoding = flags.outputEncoding;
|
||
|
|
||
|
// This 'encoding' section relies on font metrics format
|
||
|
// attached to font objects by, among others,
|
||
|
// "Willow Systems' standard_font_metrics plugin"
|
||
|
// see jspdf.plugin.standard_font_metrics.js for format
|
||
|
// of the font.metadata.encoding Object.
|
||
|
// It should be something like
|
||
|
// .encoding = {'codePages':['WinANSI....'], 'WinANSI...':{code:code, ...}}
|
||
|
// .widths = {0:width, code:width, ..., 'fof':divisor}
|
||
|
// .kerning = {code:{previous_char_code:shift, ..., 'fof':-divisor},...}
|
||
|
if ((flags.autoencode || outputEncoding) &&
|
||
|
fonts[activeFontKey].metadata &&
|
||
|
fonts[activeFontKey].metadata[sourceEncoding] &&
|
||
|
fonts[activeFontKey].metadata[sourceEncoding].encoding
|
||
|
) {
|
||
|
encodingBlock = fonts[activeFontKey].metadata[sourceEncoding].encoding;
|
||
|
|
||
|
// each font has default encoding. Some have it clearly defined.
|
||
|
if (!outputEncoding && fonts[activeFontKey].encoding) {
|
||
|
outputEncoding = fonts[activeFontKey].encoding;
|
||
|
}
|
||
|
|
||
|
// Hmmm, the above did not work? Let's try again, in different place.
|
||
|
if (!outputEncoding && encodingBlock.codePages) {
|
||
|
outputEncoding = encodingBlock.codePages[0]; // let's say, first one is the default
|
||
|
}
|
||
|
|
||
|
if (typeof outputEncoding === 'string') {
|
||
|
outputEncoding = encodingBlock[outputEncoding];
|
||
|
}
|
||
|
// we want output encoding to be a JS Object, where
|
||
|
// key = sourceEncoding's character code and
|
||
|
// value = outputEncoding's character code.
|
||
|
if (outputEncoding) {
|
||
|
isUnicode = false;
|
||
|
newtext = [];
|
||
|
for (i = 0, l = text.length; i < l; i++) {
|
||
|
ch = outputEncoding[text.charCodeAt(i)];
|
||
|
if (ch) {
|
||
|
newtext.push(
|
||
|
String.fromCharCode(ch)
|
||
|
);
|
||
|
} else {
|
||
|
newtext.push(
|
||
|
text[i]
|
||
|
);
|
||
|
}
|
||
|
|
||
|
// since we are looping over chars anyway, might as well
|
||
|
// check for residual unicodeness
|
||
|
if (newtext[i].charCodeAt(0) >> 8) { /* more than 255 */
|
||
|
isUnicode = true;
|
||
|
}
|
||
|
}
|
||
|
text = newtext.join('');
|
||
|
}
|
||
|
}
|
||
|
|
||
|
i = text.length;
|
||
|
// isUnicode may be set to false above. Hence the triple-equal to undefined
|
||
|
while (isUnicode === undef && i !== 0) {
|
||
|
if (text.charCodeAt(i - 1) >> 8) { /* more than 255 */
|
||
|
isUnicode = true;
|
||
|
}
|
||
|
i--;
|
||
|
}
|
||
|
if (!isUnicode) {
|
||
|
return text;
|
||
|
} else {
|
||
|
newtext = flags.noBOM ? [] : [254, 255];
|
||
|
for (i = 0, l = text.length; i < l; i++) {
|
||
|
ch = text.charCodeAt(i);
|
||
|
bch = ch >> 8; // divide by 256
|
||
|
if (bch >> 8) { /* something left after dividing by 256 second time */
|
||
|
throw new Error("Character at position " + i.toString(10) + " of string '" + text + "' exceeds 16bits. Cannot be encoded into UCS-2 BE");
|
||
|
}
|
||
|
newtext.push(bch);
|
||
|
newtext.push(ch - (bch << 8));
|
||
|
}
|
||
|
return String.fromCharCode.apply(undef, newtext);
|
||
|
}
|
||
|
},
|
||
|
// Replace '/', '(', and ')' with pdf-safe versions
|
||
|
pdfEscape = function (text, flags) {
|
||
|
// doing to8bitStream does NOT make this PDF display unicode text. For that
|
||
|
// we also need to reference a unicode font and embed it - royal pain in the rear.
|
||
|
|
||
|
// There is still a benefit to to8bitStream - PDF simply cannot handle 16bit chars,
|
||
|
// which JavaScript Strings are happy to provide. So, while we still cannot display
|
||
|
// 2-byte characters property, at least CONDITIONALLY converting (entire string containing)
|
||
|
// 16bit chars to (USC-2-BE) 2-bytes per char + BOM streams we ensure that entire PDF
|
||
|
// is still parseable.
|
||
|
// This will allow immediate support for unicode in document properties strings.
|
||
|
return to8bitStream(text, flags).replace(/\\/g, '\\\\').replace(/\(/g, '\\(').replace(/\)/g, '\\)');
|
||
|
},
|
||
|
putInfo = function () {
|
||
|
out('/Producer (jsPDF ' + version + ')');
|
||
|
if (documentProperties.title) {
|
||
|
out('/Title (' + pdfEscape(documentProperties.title) + ')');
|
||
|
}
|
||
|
if (documentProperties.subject) {
|
||
|
out('/Subject (' + pdfEscape(documentProperties.subject) + ')');
|
||
|
}
|
||
|
if (documentProperties.author) {
|
||
|
out('/Author (' + pdfEscape(documentProperties.author) + ')');
|
||
|
}
|
||
|
if (documentProperties.keywords) {
|
||
|
out('/Keywords (' + pdfEscape(documentProperties.keywords) + ')');
|
||
|
}
|
||
|
if (documentProperties.creator) {
|
||
|
out('/Creator (' + pdfEscape(documentProperties.creator) + ')');
|
||
|
}
|
||
|
var created = new Date();
|
||
|
out('/CreationDate (D:' +
|
||
|
[
|
||
|
created.getFullYear(),
|
||
|
padd2(created.getMonth() + 1),
|
||
|
padd2(created.getDate()),
|
||
|
padd2(created.getHours()),
|
||
|
padd2(created.getMinutes()),
|
||
|
padd2(created.getSeconds())
|
||
|
].join('') +
|
||
|
')'
|
||
|
);
|
||
|
},
|
||
|
putCatalog = function () {
|
||
|
out('/Type /Catalog');
|
||
|
out('/Pages 1 0 R');
|
||
|
// @TODO: Add zoom and layout modes
|
||
|
out('/OpenAction [3 0 R /FitH null]');
|
||
|
out('/PageLayout /OneColumn');
|
||
|
events.publish('putCatalog');
|
||
|
},
|
||
|
putTrailer = function () {
|
||
|
out('/Size ' + (objectNumber + 1));
|
||
|
out('/Root ' + objectNumber + ' 0 R');
|
||
|
out('/Info ' + (objectNumber - 1) + ' 0 R');
|
||
|
},
|
||
|
beginPage = function () {
|
||
|
page++;
|
||
|
// Do dimension stuff
|
||
|
outToPages = true;
|
||
|
pages[page] = [];
|
||
|
},
|
||
|
_addPage = function () {
|
||
|
beginPage();
|
||
|
// Set line width
|
||
|
out(f2(lineWidth * k) + ' w');
|
||
|
// Set draw color
|
||
|
out(drawColor);
|
||
|
// resurrecting non-default line caps, joins
|
||
|
if (lineCapID !== 0) {
|
||
|
out(lineCapID.toString(10) + ' J');
|
||
|
}
|
||
|
if (lineJoinID !== 0) {
|
||
|
out(lineJoinID.toString(10) + ' j');
|
||
|
}
|
||
|
events.publish('addPage', {'pageNumber': page});
|
||
|
},
|
||
|
/**
|
||
|
Returns a document-specific font key - a label assigned to a
|
||
|
font name + font type combination at the time the font was added
|
||
|
to the font inventory.
|
||
|
|
||
|
Font key is used as label for the desired font for a block of text
|
||
|
to be added to the PDF document stream.
|
||
|
@private
|
||
|
@function
|
||
|
@param fontName {String} can be undefined on "falthy" to indicate "use current"
|
||
|
@param fontStyle {String} can be undefined on "falthy" to indicate "use current"
|
||
|
@returns {String} Font key.
|
||
|
*/
|
||
|
getFont = function (fontName, fontStyle) {
|
||
|
var key, undef;
|
||
|
|
||
|
if (fontName === undef) {
|
||
|
fontName = fonts[activeFontKey].fontName;
|
||
|
}
|
||
|
if (fontStyle === undef) {
|
||
|
fontStyle = fonts[activeFontKey].fontStyle;
|
||
|
}
|
||
|
|
||
|
try {
|
||
|
key = fontmap[fontName][fontStyle]; // returns a string like 'F3' - the KEY corresponding tot he font + type combination.
|
||
|
} catch (e) {
|
||
|
key = undef;
|
||
|
}
|
||
|
if (!key) {
|
||
|
throw new Error("Unable to look up font label for font '" + fontName + "', '" + fontStyle + "'. Refer to getFontList() for available fonts.");
|
||
|
}
|
||
|
|
||
|
return key;
|
||
|
},
|
||
|
buildDocument = function () {
|
||
|
|
||
|
outToPages = false; // switches out() to content
|
||
|
content = [];
|
||
|
offsets = [];
|
||
|
|
||
|
// putHeader()
|
||
|
out('%PDF-' + pdfVersion);
|
||
|
|
||
|
putPages();
|
||
|
|
||
|
putResources();
|
||
|
|
||
|
// Info
|
||
|
newObject();
|
||
|
out('<<');
|
||
|
putInfo();
|
||
|
out('>>');
|
||
|
out('endobj');
|
||
|
|
||
|
// Catalog
|
||
|
newObject();
|
||
|
out('<<');
|
||
|
putCatalog();
|
||
|
out('>>');
|
||
|
out('endobj');
|
||
|
|
||
|
// Cross-ref
|
||
|
var o = content_length, i;
|
||
|
out('xref');
|
||
|
out('0 ' + (objectNumber + 1));
|
||
|
out('0000000000 65535 f ');
|
||
|
for (i = 1; i <= objectNumber; i++) {
|
||
|
out(padd10(offsets[i]) + ' 00000 n ');
|
||
|
}
|
||
|
// Trailer
|
||
|
out('trailer');
|
||
|
out('<<');
|
||
|
putTrailer();
|
||
|
out('>>');
|
||
|
out('startxref');
|
||
|
out(o);
|
||
|
out('%%EOF');
|
||
|
|
||
|
outToPages = true;
|
||
|
|
||
|
return content.join('\n');
|
||
|
},
|
||
|
getStyle = function (style) {
|
||
|
// see Path-Painting Operators of PDF spec
|
||
|
var op = 'S'; // stroke
|
||
|
if (style === 'F') {
|
||
|
op = 'f'; // fill
|
||
|
} else if (style === 'FD' || style === 'DF') {
|
||
|
op = 'B'; // both
|
||
|
}
|
||
|
return op;
|
||
|
},
|
||
|
|
||
|
/**
|
||
|
Generates the PDF document.
|
||
|
Possible values:
|
||
|
datauristring (alias dataurlstring) - Data-Url-formatted data returned as string.
|
||
|
datauri (alias datauri) - Data-Url-formatted data pushed into current window's location (effectively reloading the window with contents of the PDF).
|
||
|
|
||
|
If `type` argument is undefined, output is raw body of resulting PDF returned as a string.
|
||
|
|
||
|
@param {String} type A string identifying one of the possible output types.
|
||
|
@param {Object} options An object providing some additional signalling to PDF generator.
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name output
|
||
|
*/
|
||
|
output = function (type, options) {
|
||
|
var undef, data, length, array, i, blob;
|
||
|
switch (type) {
|
||
|
case undef:
|
||
|
return buildDocument();
|
||
|
case 'save':
|
||
|
if (navigator.getUserMedia) {
|
||
|
if (window.URL === undefined) {
|
||
|
return API.output('dataurlnewwindow');
|
||
|
} else if (window.URL.createObjectURL === undefined) {
|
||
|
return API.output('dataurlnewwindow');
|
||
|
}
|
||
|
}
|
||
|
data = buildDocument();
|
||
|
|
||
|
// Need to add the file to BlobBuilder as a Uint8Array
|
||
|
length = data.length;
|
||
|
array = new Uint8Array(new ArrayBuffer(length));
|
||
|
|
||
|
for (i = 0; i < length; i++) {
|
||
|
array[i] = data.charCodeAt(i);
|
||
|
}
|
||
|
|
||
|
blob = new Blob([array], {type: "application/pdf"});
|
||
|
|
||
|
saveAs(blob, options);
|
||
|
break;
|
||
|
case 'datauristring':
|
||
|
case 'dataurlstring':
|
||
|
return 'data:application/pdf;base64,' + btoa(buildDocument());
|
||
|
case 'datauri':
|
||
|
case 'dataurl':
|
||
|
document.location.href = 'data:application/pdf;base64,' + btoa(buildDocument());
|
||
|
break;
|
||
|
case 'dataurlnewwindow':
|
||
|
window.open('data:application/pdf;base64,' + btoa(buildDocument()));
|
||
|
break;
|
||
|
default:
|
||
|
throw new Error('Output type "' + type + '" is not supported.');
|
||
|
}
|
||
|
// @TODO: Add different output options
|
||
|
};
|
||
|
|
||
|
if (unit === 'pt') {
|
||
|
k = 1;
|
||
|
} else if (unit === 'mm') {
|
||
|
k = 72 / 25.4;
|
||
|
} else if (unit === 'cm') {
|
||
|
k = 72 / 2.54;
|
||
|
} else if (unit === 'in') {
|
||
|
k = 72;
|
||
|
} else {
|
||
|
throw ('Invalid unit: ' + unit);
|
||
|
}
|
||
|
|
||
|
// Dimensions are stored as user units and converted to points on output
|
||
|
if (pageFormats.hasOwnProperty(format_as_string)) {
|
||
|
pageHeight = pageFormats[format_as_string][1] / k;
|
||
|
pageWidth = pageFormats[format_as_string][0] / k;
|
||
|
} else {
|
||
|
try {
|
||
|
pageHeight = format[1];
|
||
|
pageWidth = format[0];
|
||
|
} catch (err) {
|
||
|
throw ('Invalid format: ' + format);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if (orientation === 'p' || orientation === 'portrait') {
|
||
|
orientation = 'p';
|
||
|
if (pageWidth > pageHeight) {
|
||
|
tmp = pageWidth;
|
||
|
pageWidth = pageHeight;
|
||
|
pageHeight = tmp;
|
||
|
}
|
||
|
} else if (orientation === 'l' || orientation === 'landscape') {
|
||
|
orientation = 'l';
|
||
|
if (pageHeight > pageWidth) {
|
||
|
tmp = pageWidth;
|
||
|
pageWidth = pageHeight;
|
||
|
pageHeight = tmp;
|
||
|
}
|
||
|
} else {
|
||
|
throw ('Invalid orientation: ' + orientation);
|
||
|
}
|
||
|
|
||
|
|
||
|
|
||
|
//---------------------------------------
|
||
|
// Public API
|
||
|
|
||
|
/*
|
||
|
Object exposing internal API to plugins
|
||
|
@public
|
||
|
*/
|
||
|
API.internal = {
|
||
|
'pdfEscape': pdfEscape,
|
||
|
'getStyle': getStyle,
|
||
|
/**
|
||
|
Returns {FontObject} describing a particular font.
|
||
|
@public
|
||
|
@function
|
||
|
@param fontName {String} (Optional) Font's family name
|
||
|
@param fontStyle {String} (Optional) Font's style variation name (Example:"Italic")
|
||
|
@returns {FontObject}
|
||
|
*/
|
||
|
'getFont': function () { return fonts[getFont.apply(API, arguments)]; },
|
||
|
'getFontSize': function () { return activeFontSize; },
|
||
|
'getLineHeight': function () { return activeFontSize * lineHeightProportion; },
|
||
|
'btoa': btoa,
|
||
|
'write': function (string1, string2, string3, etc) {
|
||
|
out(
|
||
|
arguments.length === 1 ? string1 : Array.prototype.join.call(arguments, ' ')
|
||
|
);
|
||
|
},
|
||
|
'getCoordinateString': function (value) {
|
||
|
return f2(value * k);
|
||
|
},
|
||
|
'getVerticalCoordinateString': function (value) {
|
||
|
return f2((pageHeight - value) * k);
|
||
|
},
|
||
|
'collections': {},
|
||
|
'newObject': newObject,
|
||
|
'putStream': putStream,
|
||
|
'events': events,
|
||
|
// ratio that you use in multiplication of a given "size" number to arrive to 'point'
|
||
|
// units of measurement.
|
||
|
// scaleFactor is set at initialization of the document and calculated against the stated
|
||
|
// default measurement units for the document.
|
||
|
// If default is "mm", k is the number that will turn number in 'mm' into 'points' number.
|
||
|
// through multiplication.
|
||
|
'scaleFactor': k,
|
||
|
'pageSize': {'width': pageWidth, 'height': pageHeight},
|
||
|
'output': function (type, options) {
|
||
|
return output(type, options);
|
||
|
},
|
||
|
'getNumberOfPages': function () {return pages.length - 1; },
|
||
|
'pages': pages
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Adds (and transfers the focus to) new page to the PDF document.
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
|
||
|
@methodOf jsPDF#
|
||
|
@name addPage
|
||
|
*/
|
||
|
API.addPage = function () {
|
||
|
_addPage();
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Adds text to page. Supports adding multiline text when 'text' argument is an Array of Strings.
|
||
|
@function
|
||
|
@param {String|Array} text String or array of strings to be added to the page. Each line is shifted one line down per font, spacing settings declared before this call.
|
||
|
@param {Number} x Coordinate (in units declared at inception of PDF document) against left edge of the page
|
||
|
@param {Number} y Coordinate (in units declared at inception of PDF document) against upper edge of the page
|
||
|
@param {Object} flags Collection of settings signalling how the text must be encoded. Defaults are sane. If you think you want to pass some flags, you likely can read the source.
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name text
|
||
|
*/
|
||
|
API.text = function (text, x, y, flags) {
|
||
|
/**
|
||
|
* Inserts something like this into PDF
|
||
|
BT
|
||
|
/F1 16 Tf % Font name + size
|
||
|
16 TL % How many units down for next line in multiline text
|
||
|
0 g % color
|
||
|
28.35 813.54 Td % position
|
||
|
(line one) Tj
|
||
|
T* (line two) Tj
|
||
|
T* (line three) Tj
|
||
|
ET
|
||
|
*/
|
||
|
|
||
|
var undef, _first, _second, _third, newtext, str, i;
|
||
|
// Pre-August-2012 the order of arguments was function(x, y, text, flags)
|
||
|
// in effort to make all calls have similar signature like
|
||
|
// function(data, coordinates... , miscellaneous)
|
||
|
// this method had its args flipped.
|
||
|
// code below allows backward compatibility with old arg order.
|
||
|
if (typeof text === 'number') {
|
||
|
_first = y;
|
||
|
_second = text;
|
||
|
_third = x;
|
||
|
|
||
|
text = _first;
|
||
|
x = _second;
|
||
|
y = _third;
|
||
|
}
|
||
|
|
||
|
// If there are any newlines in text, we assume
|
||
|
// the user wanted to print multiple lines, so break the
|
||
|
// text up into an array. If the text is already an array,
|
||
|
// we assume the user knows what they are doing.
|
||
|
if (typeof text === 'string' && text.match(/[\n\r]/)) {
|
||
|
text = text.split(/\r\n|\r|\n/g);
|
||
|
}
|
||
|
|
||
|
if (typeof flags === 'undefined') {
|
||
|
flags = {'noBOM': true, 'autoencode': true};
|
||
|
} else {
|
||
|
|
||
|
if (flags.noBOM === undef) {
|
||
|
flags.noBOM = true;
|
||
|
}
|
||
|
|
||
|
if (flags.autoencode === undef) {
|
||
|
flags.autoencode = true;
|
||
|
}
|
||
|
|
||
|
}
|
||
|
|
||
|
if (typeof text === 'string') {
|
||
|
str = pdfEscape(text, flags);
|
||
|
} else if (text instanceof Array) { /* Array */
|
||
|
// we don't want to destroy original text array, so cloning it
|
||
|
newtext = text.concat();
|
||
|
// we do array.join('text that must not be PDFescaped")
|
||
|
// thus, pdfEscape each component separately
|
||
|
for (i = newtext.length - 1; i !== -1; i--) {
|
||
|
newtext[i] = pdfEscape(newtext[i], flags);
|
||
|
}
|
||
|
str = newtext.join(") Tj\nT* (");
|
||
|
} else {
|
||
|
throw new Error('Type of text must be string or Array. "' + text + '" is not recognized.');
|
||
|
}
|
||
|
// Using "'" ("go next line and render text" mark) would save space but would complicate our rendering code, templates
|
||
|
|
||
|
// BT .. ET does NOT have default settings for Tf. You must state that explicitely every time for BT .. ET
|
||
|
// if you want text transformation matrix (+ multiline) to work reliably (which reads sizes of things from font declarations)
|
||
|
// Thus, there is NO useful, *reliable* concept of "default" font for a page.
|
||
|
// The fact that "default" (reuse font used before) font worked before in basic cases is an accident
|
||
|
// - readers dealing smartly with brokenness of jsPDF's markup.
|
||
|
out(
|
||
|
'BT\n/' +
|
||
|
activeFontKey + ' ' + activeFontSize + ' Tf\n' + // font face, style, size
|
||
|
(activeFontSize * lineHeightProportion) + ' TL\n' + // line spacing
|
||
|
textColor +
|
||
|
'\n' + f2(x * k) + ' ' + f2((pageHeight - y) * k) + ' Td\n(' +
|
||
|
str +
|
||
|
') Tj\nET'
|
||
|
);
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
API.line = function (x1, y1, x2, y2) {
|
||
|
out(
|
||
|
f2(x1 * k) + ' ' + f2((pageHeight - y1) * k) + ' m ' +
|
||
|
f2(x2 * k) + ' ' + f2((pageHeight - y2) * k) + ' l S'
|
||
|
);
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Adds series of curves (straight lines or cubic bezier curves) to canvas, starting at `x`, `y` coordinates.
|
||
|
All data points in `lines` are relative to last line origin.
|
||
|
`x`, `y` become x1,y1 for first line / curve in the set.
|
||
|
For lines you only need to specify [x2, y2] - (ending point) vector against x1, y1 starting point.
|
||
|
For bezier curves you need to specify [x2,y2,x3,y3,x4,y4] - vectors to control points 1, 2, ending point. All vectors are against the start of the curve - x1,y1.
|
||
|
|
||
|
@example .lines([[2,2],[-2,2],[1,1,2,2,3,3],[2,1]], 212,110, 10) // line, line, bezier curve, line
|
||
|
@param {Array} lines Array of *vector* shifts as pairs (lines) or sextets (cubic bezier curves).
|
||
|
@param {Number} x Coordinate (in units declared at inception of PDF document) against left edge of the page
|
||
|
@param {Number} y Coordinate (in units declared at inception of PDF document) against upper edge of the page
|
||
|
@param {Number} scale (Defaults to [1.0,1.0]) x,y Scaling factor for all vectors. Elements can be any floating number Sub-one makes drawing smaller. Over-one grows the drawing. Negative flips the direction.
|
||
|
@param {String} style One of 'S' (the default), 'F', 'FD' or 'DF'. 'S' draws just the curve. 'F' fills the region defined by the curves. 'DF' or 'FD' draws the curves and fills the region.
|
||
|
@param {Boolean} closed If true, the path is closed with a straight line from the end of the last curve to the starting point.
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name lines
|
||
|
*/
|
||
|
API.lines = function (lines, x, y, scale, style, closed) {
|
||
|
var undef, _first, _second, _third, scalex, scaley, i, l, leg, x2, y2, x3, y3, x4, y4;
|
||
|
|
||
|
// Pre-August-2012 the order of arguments was function(x, y, lines, scale, style)
|
||
|
// in effort to make all calls have similar signature like
|
||
|
// function(content, coordinateX, coordinateY , miscellaneous)
|
||
|
// this method had its args flipped.
|
||
|
// code below allows backward compatibility with old arg order.
|
||
|
if (typeof lines === 'number') {
|
||
|
_first = y;
|
||
|
_second = lines;
|
||
|
_third = x;
|
||
|
|
||
|
lines = _first;
|
||
|
x = _second;
|
||
|
y = _third;
|
||
|
}
|
||
|
|
||
|
style = getStyle(style);
|
||
|
scale = scale === undef ? [1, 1] : scale;
|
||
|
|
||
|
// starting point
|
||
|
out(f3(x * k) + ' ' + f3((pageHeight - y) * k) + ' m ');
|
||
|
|
||
|
scalex = scale[0];
|
||
|
scaley = scale[1];
|
||
|
l = lines.length;
|
||
|
//, x2, y2 // bezier only. In page default measurement "units", *after* scaling
|
||
|
//, x3, y3 // bezier only. In page default measurement "units", *after* scaling
|
||
|
// ending point for all, lines and bezier. . In page default measurement "units", *after* scaling
|
||
|
x4 = x; // last / ending point = starting point for first item.
|
||
|
y4 = y; // last / ending point = starting point for first item.
|
||
|
|
||
|
for (i = 0; i < l; i++) {
|
||
|
leg = lines[i];
|
||
|
if (leg.length === 2) {
|
||
|
// simple line
|
||
|
x4 = leg[0] * scalex + x4; // here last x4 was prior ending point
|
||
|
y4 = leg[1] * scaley + y4; // here last y4 was prior ending point
|
||
|
out(f3(x4 * k) + ' ' + f3((pageHeight - y4) * k) + ' l');
|
||
|
} else {
|
||
|
// bezier curve
|
||
|
x2 = leg[0] * scalex + x4; // here last x4 is prior ending point
|
||
|
y2 = leg[1] * scaley + y4; // here last y4 is prior ending point
|
||
|
x3 = leg[2] * scalex + x4; // here last x4 is prior ending point
|
||
|
y3 = leg[3] * scaley + y4; // here last y4 is prior ending point
|
||
|
x4 = leg[4] * scalex + x4; // here last x4 was prior ending point
|
||
|
y4 = leg[5] * scaley + y4; // here last y4 was prior ending point
|
||
|
out(
|
||
|
f3(x2 * k) + ' ' +
|
||
|
f3((pageHeight - y2) * k) + ' ' +
|
||
|
f3(x3 * k) + ' ' +
|
||
|
f3((pageHeight - y3) * k) + ' ' +
|
||
|
f3(x4 * k) + ' ' +
|
||
|
f3((pageHeight - y4) * k) + ' c'
|
||
|
);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if (closed == true) {
|
||
|
out(' h');
|
||
|
}
|
||
|
|
||
|
// stroking / filling / both the path
|
||
|
out(style);
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Adds a rectangle to PDF
|
||
|
|
||
|
@param {Number} x Coordinate (in units declared at inception of PDF document) against left edge of the page
|
||
|
@param {Number} y Coordinate (in units declared at inception of PDF document) against upper edge of the page
|
||
|
@param {Number} w Width (in units declared at inception of PDF document)
|
||
|
@param {Number} h Height (in units declared at inception of PDF document)
|
||
|
@param {String} style (Defaults to active fill/stroke style) A string signalling if stroke, fill or both are to be applied.
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name rect
|
||
|
*/
|
||
|
API.rect = function (x, y, w, h, style) {
|
||
|
var op = getStyle(style);
|
||
|
out([
|
||
|
f2(x * k),
|
||
|
f2((pageHeight - y) * k),
|
||
|
f2(w * k),
|
||
|
f2(-h * k),
|
||
|
're',
|
||
|
op
|
||
|
].join(' '));
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Adds a triangle to PDF
|
||
|
|
||
|
@param {Number} x1 Coordinate (in units declared at inception of PDF document) against left edge of the page
|
||
|
@param {Number} y1 Coordinate (in units declared at inception of PDF document) against upper edge of the page
|
||
|
@param {Number} x2 Coordinate (in units declared at inception of PDF document) against left edge of the page
|
||
|
@param {Number} y2 Coordinate (in units declared at inception of PDF document) against upper edge of the page
|
||
|
@param {Number} x3 Coordinate (in units declared at inception of PDF document) against left edge of the page
|
||
|
@param {Number} y3 Coordinate (in units declared at inception of PDF document) against upper edge of the page
|
||
|
@param {String} style (Defaults to active fill/stroke style) A string signalling if stroke, fill or both are to be applied.
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name triangle
|
||
|
*/
|
||
|
API.triangle = function (x1, y1, x2, y2, x3, y3, style) {
|
||
|
this.lines(
|
||
|
[
|
||
|
[ x2 - x1, y2 - y1 ], // vector to point 2
|
||
|
[ x3 - x2, y3 - y2 ], // vector to point 3
|
||
|
[ x1 - x3, y1 - y3 ] // closing vector back to point 1
|
||
|
],
|
||
|
x1,
|
||
|
y1, // start of path
|
||
|
[1, 1],
|
||
|
style,
|
||
|
true
|
||
|
);
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Adds a rectangle with rounded corners to PDF
|
||
|
|
||
|
@param {Number} x Coordinate (in units declared at inception of PDF document) against left edge of the page
|
||
|
@param {Number} y Coordinate (in units declared at inception of PDF document) against upper edge of the page
|
||
|
@param {Number} w Width (in units declared at inception of PDF document)
|
||
|
@param {Number} h Height (in units declared at inception of PDF document)
|
||
|
@param {Number} rx Radius along x axis (in units declared at inception of PDF document)
|
||
|
@param {Number} rx Radius along y axis (in units declared at inception of PDF document)
|
||
|
@param {String} style (Defaults to active fill/stroke style) A string signalling if stroke, fill or both are to be applied.
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name roundedRect
|
||
|
*/
|
||
|
API.roundedRect = function (x, y, w, h, rx, ry, style) {
|
||
|
var MyArc = 4 / 3 * (Math.SQRT2 - 1);
|
||
|
this.lines(
|
||
|
[
|
||
|
[ (w - 2 * rx), 0 ],
|
||
|
[ (rx * MyArc), 0, rx, ry - (ry * MyArc), rx, ry ],
|
||
|
[ 0, (h - 2 * ry) ],
|
||
|
[ 0, (ry * MyArc), -(rx * MyArc), ry, -rx, ry],
|
||
|
[ (-w + 2 * rx), 0],
|
||
|
[ -(rx * MyArc), 0, -rx, -(ry * MyArc), -rx, -ry],
|
||
|
[ 0, (-h + 2 * ry)],
|
||
|
[ 0, -(ry * MyArc), (rx * MyArc), -ry, rx, -ry]
|
||
|
],
|
||
|
x + rx,
|
||
|
y, // start of path
|
||
|
[1, 1],
|
||
|
style
|
||
|
);
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Adds an ellipse to PDF
|
||
|
|
||
|
@param {Number} x Coordinate (in units declared at inception of PDF document) against left edge of the page
|
||
|
@param {Number} y Coordinate (in units declared at inception of PDF document) against upper edge of the page
|
||
|
@param {Number} rx Radius along x axis (in units declared at inception of PDF document)
|
||
|
@param {Number} rx Radius along y axis (in units declared at inception of PDF document)
|
||
|
@param {String} style (Defaults to active fill/stroke style) A string signalling if stroke, fill or both are to be applied.
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name ellipse
|
||
|
*/
|
||
|
API.ellipse = function (x, y, rx, ry, style) {
|
||
|
var op = getStyle(style),
|
||
|
lx = 4 / 3 * (Math.SQRT2 - 1) * rx,
|
||
|
ly = 4 / 3 * (Math.SQRT2 - 1) * ry;
|
||
|
|
||
|
out([
|
||
|
f2((x + rx) * k),
|
||
|
f2((pageHeight - y) * k),
|
||
|
'm',
|
||
|
f2((x + rx) * k),
|
||
|
f2((pageHeight - (y - ly)) * k),
|
||
|
f2((x + lx) * k),
|
||
|
f2((pageHeight - (y - ry)) * k),
|
||
|
f2(x * k),
|
||
|
f2((pageHeight - (y - ry)) * k),
|
||
|
'c'
|
||
|
].join(' '));
|
||
|
out([
|
||
|
f2((x - lx) * k),
|
||
|
f2((pageHeight - (y - ry)) * k),
|
||
|
f2((x - rx) * k),
|
||
|
f2((pageHeight - (y - ly)) * k),
|
||
|
f2((x - rx) * k),
|
||
|
f2((pageHeight - y) * k),
|
||
|
'c'
|
||
|
].join(' '));
|
||
|
out([
|
||
|
f2((x - rx) * k),
|
||
|
f2((pageHeight - (y + ly)) * k),
|
||
|
f2((x - lx) * k),
|
||
|
f2((pageHeight - (y + ry)) * k),
|
||
|
f2(x * k),
|
||
|
f2((pageHeight - (y + ry)) * k),
|
||
|
'c'
|
||
|
].join(' '));
|
||
|
out([
|
||
|
f2((x + lx) * k),
|
||
|
f2((pageHeight - (y + ry)) * k),
|
||
|
f2((x + rx) * k),
|
||
|
f2((pageHeight - (y + ly)) * k),
|
||
|
f2((x + rx) * k),
|
||
|
f2((pageHeight - y) * k),
|
||
|
'c',
|
||
|
op
|
||
|
].join(' '));
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Adds an circle to PDF
|
||
|
|
||
|
@param {Number} x Coordinate (in units declared at inception of PDF document) against left edge of the page
|
||
|
@param {Number} y Coordinate (in units declared at inception of PDF document) against upper edge of the page
|
||
|
@param {Number} r Radius (in units declared at inception of PDF document)
|
||
|
@param {String} style (Defaults to active fill/stroke style) A string signalling if stroke, fill or both are to be applied.
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name circle
|
||
|
*/
|
||
|
API.circle = function (x, y, r, style) {
|
||
|
return this.ellipse(x, y, r, r, style);
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Adds a properties to the PDF document
|
||
|
|
||
|
@param {Object} A property_name-to-property_value object structure.
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name setProperties
|
||
|
*/
|
||
|
API.setProperties = function (properties) {
|
||
|
// copying only those properties we can render.
|
||
|
var property;
|
||
|
for (property in documentProperties) {
|
||
|
if (documentProperties.hasOwnProperty(property) && properties[property]) {
|
||
|
documentProperties[property] = properties[property];
|
||
|
}
|
||
|
}
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Sets font size for upcoming text elements.
|
||
|
|
||
|
@param {Number} size Font size in points.
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name setFontSize
|
||
|
*/
|
||
|
API.setFontSize = function (size) {
|
||
|
activeFontSize = size;
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Sets text font face, variant for upcoming text elements.
|
||
|
See output of jsPDF.getFontList() for possible font names, styles.
|
||
|
|
||
|
@param {String} fontName Font name or family. Example: "times"
|
||
|
@param {String} fontStyle Font style or variant. Example: "italic"
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name setFont
|
||
|
*/
|
||
|
API.setFont = function (fontName, fontStyle) {
|
||
|
activeFontKey = getFont(fontName, fontStyle);
|
||
|
// if font is not found, the above line blows up and we never go further
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Switches font style or variant for upcoming text elements,
|
||
|
while keeping the font face or family same.
|
||
|
See output of jsPDF.getFontList() for possible font names, styles.
|
||
|
|
||
|
@param {String} style Font style or variant. Example: "italic"
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name setFontStyle
|
||
|
*/
|
||
|
API.setFontStyle = API.setFontType = function (style) {
|
||
|
var undef;
|
||
|
activeFontKey = getFont(undef, style);
|
||
|
// if font is not found, the above line blows up and we never go further
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Returns an object - a tree of fontName to fontStyle relationships available to
|
||
|
active PDF document.
|
||
|
|
||
|
@public
|
||
|
@function
|
||
|
@returns {Object} Like {'times':['normal', 'italic', ... ], 'arial':['normal', 'bold', ... ], ... }
|
||
|
@methodOf jsPDF#
|
||
|
@name getFontList
|
||
|
*/
|
||
|
API.getFontList = function () {
|
||
|
// TODO: iterate over fonts array or return copy of fontmap instead in case more are ever added.
|
||
|
var list = {},
|
||
|
fontName,
|
||
|
fontStyle,
|
||
|
tmp;
|
||
|
|
||
|
for (fontName in fontmap) {
|
||
|
if (fontmap.hasOwnProperty(fontName)) {
|
||
|
list[fontName] = tmp = [];
|
||
|
for (fontStyle in fontmap[fontName]) {
|
||
|
if (fontmap[fontName].hasOwnProperty(fontStyle)) {
|
||
|
tmp.push(fontStyle);
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
return list;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Sets line width for upcoming lines.
|
||
|
|
||
|
@param {Number} width Line width (in units declared at inception of PDF document)
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name setLineWidth
|
||
|
*/
|
||
|
API.setLineWidth = function (width) {
|
||
|
out((width * k).toFixed(2) + ' w');
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Sets the stroke color for upcoming elements.
|
||
|
|
||
|
Depending on the number of arguments given, Gray, RGB, or CMYK
|
||
|
color space is implied.
|
||
|
|
||
|
When only ch1 is given, "Gray" color space is implied and it
|
||
|
must be a value in the range from 0.00 (solid black) to to 1.00 (white)
|
||
|
if values are communicated as String types, or in range from 0 (black)
|
||
|
to 255 (white) if communicated as Number type.
|
||
|
The RGB-like 0-255 range is provided for backward compatibility.
|
||
|
|
||
|
When only ch1,ch2,ch3 are given, "RGB" color space is implied and each
|
||
|
value must be in the range from 0.00 (minimum intensity) to to 1.00
|
||
|
(max intensity) if values are communicated as String types, or
|
||
|
from 0 (min intensity) to to 255 (max intensity) if values are communicated
|
||
|
as Number types.
|
||
|
The RGB-like 0-255 range is provided for backward compatibility.
|
||
|
|
||
|
When ch1,ch2,ch3,ch4 are given, "CMYK" color space is implied and each
|
||
|
value must be a in the range from 0.00 (0% concentration) to to
|
||
|
1.00 (100% concentration)
|
||
|
|
||
|
Because JavaScript treats fixed point numbers badly (rounds to
|
||
|
floating point nearest to binary representation) it is highly advised to
|
||
|
communicate the fractional numbers as String types, not JavaScript Number type.
|
||
|
|
||
|
@param {Number|String} ch1 Color channel value
|
||
|
@param {Number|String} ch2 Color channel value
|
||
|
@param {Number|String} ch3 Color channel value
|
||
|
@param {Number|String} ch4 Color channel value
|
||
|
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name setDrawColor
|
||
|
*/
|
||
|
API.setDrawColor = function (ch1, ch2, ch3, ch4) {
|
||
|
var color;
|
||
|
if (ch2 === undefined || (ch4 === undefined && ch1 === ch2 === ch3)) {
|
||
|
// Gray color space.
|
||
|
if (typeof ch1 === 'string') {
|
||
|
color = ch1 + ' G';
|
||
|
} else {
|
||
|
color = f2(ch1 / 255) + ' G';
|
||
|
}
|
||
|
} else if (ch4 === undefined) {
|
||
|
// RGB
|
||
|
if (typeof ch1 === 'string') {
|
||
|
color = [ch1, ch2, ch3, 'RG'].join(' ');
|
||
|
} else {
|
||
|
color = [f2(ch1 / 255), f2(ch2 / 255), f2(ch3 / 255), 'RG'].join(' ');
|
||
|
}
|
||
|
} else {
|
||
|
// CMYK
|
||
|
if (typeof ch1 === 'string') {
|
||
|
color = [ch1, ch2, ch3, ch4, 'K'].join(' ');
|
||
|
} else {
|
||
|
color = [f2(ch1), f2(ch2), f2(ch3), f2(ch4), 'K'].join(' ');
|
||
|
}
|
||
|
}
|
||
|
|
||
|
out(color);
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Sets the fill color for upcoming elements.
|
||
|
|
||
|
Depending on the number of arguments given, Gray, RGB, or CMYK
|
||
|
color space is implied.
|
||
|
|
||
|
When only ch1 is given, "Gray" color space is implied and it
|
||
|
must be a value in the range from 0.00 (solid black) to to 1.00 (white)
|
||
|
if values are communicated as String types, or in range from 0 (black)
|
||
|
to 255 (white) if communicated as Number type.
|
||
|
The RGB-like 0-255 range is provided for backward compatibility.
|
||
|
|
||
|
When only ch1,ch2,ch3 are given, "RGB" color space is implied and each
|
||
|
value must be in the range from 0.00 (minimum intensity) to to 1.00
|
||
|
(max intensity) if values are communicated as String types, or
|
||
|
from 0 (min intensity) to to 255 (max intensity) if values are communicated
|
||
|
as Number types.
|
||
|
The RGB-like 0-255 range is provided for backward compatibility.
|
||
|
|
||
|
When ch1,ch2,ch3,ch4 are given, "CMYK" color space is implied and each
|
||
|
value must be a in the range from 0.00 (0% concentration) to to
|
||
|
1.00 (100% concentration)
|
||
|
|
||
|
Because JavaScript treats fixed point numbers badly (rounds to
|
||
|
floating point nearest to binary representation) it is highly advised to
|
||
|
communicate the fractional numbers as String types, not JavaScript Number type.
|
||
|
|
||
|
@param {Number|String} ch1 Color channel value
|
||
|
@param {Number|String} ch2 Color channel value
|
||
|
@param {Number|String} ch3 Color channel value
|
||
|
@param {Number|String} ch4 Color channel value
|
||
|
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name setFillColor
|
||
|
*/
|
||
|
API.setFillColor = function (ch1, ch2, ch3, ch4) {
|
||
|
var color;
|
||
|
|
||
|
if (ch2 === undefined || (ch4 === undefined && ch1 === ch2 === ch3)) {
|
||
|
// Gray color space.
|
||
|
if (typeof ch1 === 'string') {
|
||
|
color = ch1 + ' g';
|
||
|
} else {
|
||
|
color = f2(ch1 / 255) + ' g';
|
||
|
}
|
||
|
} else if (ch4 === undefined) {
|
||
|
// RGB
|
||
|
if (typeof ch1 === 'string') {
|
||
|
color = [ch1, ch2, ch3, 'rg'].join(' ');
|
||
|
} else {
|
||
|
color = [f2(ch1 / 255), f2(ch2 / 255), f2(ch3 / 255), 'rg'].join(' ');
|
||
|
}
|
||
|
} else {
|
||
|
// CMYK
|
||
|
if (typeof ch1 === 'string') {
|
||
|
color = [ch1, ch2, ch3, ch4, 'k'].join(' ');
|
||
|
} else {
|
||
|
color = [f2(ch1), f2(ch2), f2(ch3), f2(ch4), 'k'].join(' ');
|
||
|
}
|
||
|
}
|
||
|
|
||
|
out(color);
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Sets the text color for upcoming elements.
|
||
|
If only one, first argument is given,
|
||
|
treats the value as gray-scale color value.
|
||
|
|
||
|
@param {Number} r Red channel color value in range 0-255
|
||
|
@param {Number} g Green channel color value in range 0-255
|
||
|
@param {Number} b Blue channel color value in range 0-255
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name setTextColor
|
||
|
*/
|
||
|
API.setTextColor = function (r, g, b) {
|
||
|
if ((r === 0 && g === 0 && b === 0) || (typeof g === 'undefined')) {
|
||
|
textColor = f3(r / 255) + ' g';
|
||
|
} else {
|
||
|
textColor = [f3(r / 255), f3(g / 255), f3(b / 255), 'rg'].join(' ');
|
||
|
}
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Is an Object providing a mapping from human-readable to
|
||
|
integer flag values designating the varieties of line cap
|
||
|
and join styles.
|
||
|
|
||
|
@returns {Object}
|
||
|
@fieldOf jsPDF#
|
||
|
@name CapJoinStyles
|
||
|
*/
|
||
|
API.CapJoinStyles = {
|
||
|
0: 0,
|
||
|
'butt': 0,
|
||
|
'but': 0,
|
||
|
'miter': 0,
|
||
|
1: 1,
|
||
|
'round': 1,
|
||
|
'rounded': 1,
|
||
|
'circle': 1,
|
||
|
2: 2,
|
||
|
'projecting': 2,
|
||
|
'project': 2,
|
||
|
'square': 2,
|
||
|
'bevel': 2
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Sets the line cap styles
|
||
|
See {jsPDF.CapJoinStyles} for variants
|
||
|
|
||
|
@param {String|Number} style A string or number identifying the type of line cap
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name setLineCap
|
||
|
*/
|
||
|
API.setLineCap = function (style) {
|
||
|
var id = this.CapJoinStyles[style];
|
||
|
if (id === undefined) {
|
||
|
throw new Error("Line cap style of '" + style + "' is not recognized. See or extend .CapJoinStyles property for valid styles");
|
||
|
}
|
||
|
lineCapID = id;
|
||
|
out(id.toString(10) + ' J');
|
||
|
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
Sets the line join styles
|
||
|
See {jsPDF.CapJoinStyles} for variants
|
||
|
|
||
|
@param {String|Number} style A string or number identifying the type of line join
|
||
|
@function
|
||
|
@returns {jsPDF}
|
||
|
@methodOf jsPDF#
|
||
|
@name setLineJoin
|
||
|
*/
|
||
|
API.setLineJoin = function (style) {
|
||
|
var id = this.CapJoinStyles[style];
|
||
|
if (id === undefined) {
|
||
|
throw new Error("Line join style of '" + style + "' is not recognized. See or extend .CapJoinStyles property for valid styles");
|
||
|
}
|
||
|
lineJoinID = id;
|
||
|
out(id.toString(10) + ' j');
|
||
|
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
// Output is both an internal (for plugins) and external function
|
||
|
API.output = output;
|
||
|
|
||
|
/**
|
||
|
* Saves as PDF document. An alias of jsPDF.output('save', 'filename.pdf')
|
||
|
* @param {String} filename The filename including extension.
|
||
|
*
|
||
|
* @function
|
||
|
* @returns {jsPDF}
|
||
|
* @methodOf jsPDF#
|
||
|
* @name save
|
||
|
*/
|
||
|
API.save = function (filename) {
|
||
|
API.output('save', filename);
|
||
|
};
|
||
|
|
||
|
// applying plugins (more methods) ON TOP of built-in API.
|
||
|
// this is intentional as we allow plugins to override
|
||
|
// built-ins
|
||
|
for (plugin in jsPDF.API) {
|
||
|
if (jsPDF.API.hasOwnProperty(plugin)) {
|
||
|
if (plugin === 'events' && jsPDF.API.events.length) {
|
||
|
(function (events, newEvents) {
|
||
|
|
||
|
// jsPDF.API.events is a JS Array of Arrays
|
||
|
// where each Array is a pair of event name, handler
|
||
|
// Events were added by plugins to the jsPDF instantiator.
|
||
|
// These are always added to the new instance and some ran
|
||
|
// during instantiation.
|
||
|
|
||
|
var eventname, handler_and_args, i;
|
||
|
|
||
|
for (i = newEvents.length - 1; i !== -1; i--) {
|
||
|
// subscribe takes 3 args: 'topic', function, runonce_flag
|
||
|
// if undefined, runonce is false.
|
||
|
// users can attach callback directly,
|
||
|
// or they can attach an array with [callback, runonce_flag]
|
||
|
// that's what the "apply" magic is for below.
|
||
|
eventname = newEvents[i][0];
|
||
|
handler_and_args = newEvents[i][1];
|
||
|
events.subscribe.apply(
|
||
|
events,
|
||
|
[eventname].concat(
|
||
|
typeof handler_and_args === 'function' ?
|
||
|
[ handler_and_args ] :
|
||
|
handler_and_args
|
||
|
)
|
||
|
);
|
||
|
}
|
||
|
}(events, jsPDF.API.events));
|
||
|
} else {
|
||
|
API[plugin] = jsPDF.API[plugin];
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
/////////////////////////////////////////
|
||
|
// continuing initilisation of jsPDF Document object
|
||
|
/////////////////////////////////////////
|
||
|
|
||
|
|
||
|
// Add the first page automatically
|
||
|
addFonts();
|
||
|
activeFontKey = 'F1';
|
||
|
_addPage();
|
||
|
|
||
|
events.publish('initialized');
|
||
|
|
||
|
return API;
|
||
|
}
|
||
|
|
||
|
/**
|
||
|
jsPDF.API is a STATIC property of jsPDF class.
|
||
|
jsPDF.API is an object you can add methods and properties to.
|
||
|
The methods / properties you add will show up in new jsPDF objects.
|
||
|
|
||
|
One property is prepopulated. It is the 'events' Object. Plugin authors can add topics, callbacks to this object. These will be reassigned to all new instances of jsPDF.
|
||
|
Examples:
|
||
|
jsPDF.API.events['initialized'] = function(){ 'this' is API object }
|
||
|
jsPDF.API.events['addFont'] = function(added_font_object){ 'this' is API object }
|
||
|
|
||
|
@static
|
||
|
@public
|
||
|
@memberOf jsPDF
|
||
|
@name API
|
||
|
|
||
|
@example
|
||
|
jsPDF.API.mymethod = function(){
|
||
|
// 'this' will be ref to internal API object. see jsPDF source
|
||
|
// , so you can refer to built-in methods like so:
|
||
|
// this.line(....)
|
||
|
// this.text(....)
|
||
|
}
|
||
|
var pdfdoc = new jsPDF()
|
||
|
pdfdoc.mymethod() // <- !!!!!!
|
||
|
*/
|
||
|
jsPDF.API = {'events': []};
|
||
|
|
||
|
return jsPDF;
|
||
|
}());
|
||
|
/** @preserve
|
||
|
jsPDF addImage plugin (JPEG only at this time)
|
||
|
Copyright (c) 2012 https://github.com/siefkenj/
|
||
|
*/
|
||
|
|
||
|
/**
|
||
|
* Permission is hereby granted, free of charge, to any person obtaining
|
||
|
* a copy of this software and associated documentation files (the
|
||
|
* "Software"), to deal in the Software without restriction, including
|
||
|
* without limitation the rights to use, copy, modify, merge, publish,
|
||
|
* distribute, sublicense, and/or sell copies of the Software, and to
|
||
|
* permit persons to whom the Software is furnished to do so, subject to
|
||
|
* the following conditions:
|
||
|
*
|
||
|
* The above copyright notice and this permission notice shall be
|
||
|
* included in all copies or substantial portions of the Software.
|
||
|
*
|
||
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
||
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
||
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
||
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
||
|
* LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
||
|
* OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
||
|
* WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
||
|
* ====================================================================
|
||
|
*/
|
||
|
|
||
|
;(function(jsPDFAPI) {
|
||
|
'use strict'
|
||
|
|
||
|
var namespace = 'addImage_'
|
||
|
|
||
|
// takes a string imgData containing the raw bytes of
|
||
|
// a jpeg image and returns [width, height]
|
||
|
// Algorithm from: http://www.64lines.com/jpeg-width-height
|
||
|
var getJpegSize = function(imgData) {
|
||
|
'use strict'
|
||
|
var width, height;
|
||
|
// Verify we have a valid jpeg header 0xff,0xd8,0xff,0xe0,?,?,'J','F','I','F',0x00
|
||
|
if (!imgData.charCodeAt(0) === 0xff ||
|
||
|
!imgData.charCodeAt(1) === 0xd8 ||
|
||
|
!imgData.charCodeAt(2) === 0xff ||
|
||
|
!imgData.charCodeAt(3) === 0xe0 ||
|
||
|
!imgData.charCodeAt(6) === 'J'.charCodeAt(0) ||
|
||
|
!imgData.charCodeAt(7) === 'F'.charCodeAt(0) ||
|
||
|
!imgData.charCodeAt(8) === 'I'.charCodeAt(0) ||
|
||
|
!imgData.charCodeAt(9) === 'F'.charCodeAt(0) ||
|
||
|
!imgData.charCodeAt(10) === 0x00) {
|
||
|
throw new Error('getJpegSize requires a binary jpeg file')
|
||
|
}
|
||
|
var blockLength = imgData.charCodeAt(4)*256 + imgData.charCodeAt(5);
|
||
|
var i = 4, len = imgData.length;
|
||
|
while ( i < len ) {
|
||
|
i += blockLength;
|
||
|
if (imgData.charCodeAt(i) !== 0xff) {
|
||
|
throw new Error('getJpegSize could not find the size of the image');
|
||
|
}
|
||
|
if (imgData.charCodeAt(i+1) === 0xc0 || //(SOF) Huffman - Baseline DCT
|
||
|
imgData.charCodeAt(i+1) === 0xc1 || //(SOF) Huffman - Extended sequential DCT
|
||
|
imgData.charCodeAt(i+1) === 0xc2 || // Progressive DCT (SOF2)
|
||
|
imgData.charCodeAt(i+1) === 0xc3 || // Spatial (sequential) lossless (SOF3)
|
||
|
imgData.charCodeAt(i+1) === 0xc4 || // Differential sequential DCT (SOF5)
|
||
|
imgData.charCodeAt(i+1) === 0xc5 || // Differential progressive DCT (SOF6)
|
||
|
imgData.charCodeAt(i+1) === 0xc6 || // Differential spatial (SOF7)
|
||
|
imgData.charCodeAt(i+1) === 0xc7) {
|
||
|
height = imgData.charCodeAt(i+5)*256 + imgData.charCodeAt(i+6);
|
||
|
width = imgData.charCodeAt(i+7)*256 + imgData.charCodeAt(i+8);
|
||
|
return [width, height];
|
||
|
} else {
|
||
|
i += 2;
|
||
|
blockLength = imgData.charCodeAt(i)*256 + imgData.charCodeAt(i+1)
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
// Image functionality ported from pdf.js
|
||
|
, putImage = function(img) {
|
||
|
var objectNumber = this.internal.newObject()
|
||
|
, out = this.internal.write
|
||
|
, putStream = this.internal.putStream
|
||
|
|
||
|
img['n'] = objectNumber
|
||
|
|
||
|
out('<</Type /XObject')
|
||
|
out('/Subtype /Image')
|
||
|
out('/Width ' + img['w'])
|
||
|
out('/Height ' + img['h'])
|
||
|
if (img['cs'] === 'Indexed') {
|
||
|
out('/ColorSpace [/Indexed /DeviceRGB '
|
||
|
+ (img['pal'].length / 3 - 1) + ' ' + (objectNumber + 1)
|
||
|
+ ' 0 R]');
|
||
|
} else {
|
||
|
out('/ColorSpace /' + img['cs']);
|
||
|
if (img['cs'] === 'DeviceCMYK') {
|
||
|
out('/Decode [1 0 1 0 1 0 1 0]');
|
||
|
}
|
||
|
}
|
||
|
out('/BitsPerComponent ' + img['bpc']);
|
||
|
if ('f' in img) {
|
||
|
out('/Filter /' + img['f']);
|
||
|
}
|
||
|
if ('dp' in img) {
|
||
|
out('/DecodeParms <<' + img['dp'] + '>>');
|
||
|
}
|
||
|
if ('trns' in img && img['trns'].constructor == Array) {
|
||
|
var trns = '';
|
||
|
for ( var i = 0; i < img['trns'].length; i++) {
|
||
|
trns += (img[trns][i] + ' ' + img['trns'][i] + ' ');
|
||
|
out('/Mask [' + trns + ']');
|
||
|
}
|
||
|
}
|
||
|
if ('smask' in img) {
|
||
|
out('/SMask ' + (objectNumber + 1) + ' 0 R');
|
||
|
}
|
||
|
out('/Length ' + img['data'].length + '>>');
|
||
|
|
||
|
putStream(img['data']);
|
||
|
|
||
|
out('endobj');
|
||
|
}
|
||
|
, putResourcesCallback = function() {
|
||
|
var images = this.internal.collections[namespace + 'images']
|
||
|
for ( var i in images ) {
|
||
|
putImage.call(this, images[i])
|
||
|
}
|
||
|
}
|
||
|
, putXObjectsDictCallback = function(){
|
||
|
var images = this.internal.collections[namespace + 'images']
|
||
|
, out = this.internal.write
|
||
|
, image
|
||
|
for (var i in images) {
|
||
|
image = images[i]
|
||
|
out(
|
||
|
'/I' + image['i']
|
||
|
, image['n']
|
||
|
, '0'
|
||
|
, 'R'
|
||
|
)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
jsPDFAPI.addImage = function(imageData, format, x, y, w, h) {
|
||
|
'use strict'
|
||
|
if (typeof imageData === 'object' && imageData.nodeType === 1) {
|
||
|
var canvas = document.createElement('canvas');
|
||
|
canvas.width = imageData.clientWidth;
|
||
|
canvas.height = imageData.clientHeight;
|
||
|
|
||
|
var ctx = canvas.getContext('2d');
|
||
|
if (!ctx) {
|
||
|
throw ('addImage requires canvas to be supported by browser.');
|
||
|
}
|
||
|
ctx.drawImage(imageData, 0, 0, canvas.width, canvas.height);
|
||
|
imageData = canvas.toDataURL('image/jpeg');
|
||
|
format = "JPEG";
|
||
|
}
|
||
|
if (format.toUpperCase() !== 'JPEG') {
|
||
|
throw new Error('addImage currently only supports format \'JPEG\', not \''+format+'\'');
|
||
|
}
|
||
|
|
||
|
var imageIndex
|
||
|
, images = this.internal.collections[namespace + 'images']
|
||
|
, coord = this.internal.getCoordinateString
|
||
|
, vcoord = this.internal.getVerticalCoordinateString;
|
||
|
|
||
|
// Detect if the imageData is raw binary or Data URL
|
||
|
if (imageData.substring(0, 23) === 'data:image/jpeg;base64,') {
|
||
|
imageData = atob(imageData.replace('data:image/jpeg;base64,', ''));
|
||
|
}
|
||
|
|
||
|
if (images){
|
||
|
// this is NOT the first time this method is ran on this instance of jsPDF object.
|
||
|
imageIndex = Object.keys ?
|
||
|
Object.keys(images).length :
|
||
|
(function(o){
|
||
|
var i = 0
|
||
|
for (var e in o){if(o.hasOwnProperty(e)){ i++ }}
|
||
|
return i
|
||
|
})(images)
|
||
|
} else {
|
||
|
// this is the first time this method is ran on this instance of jsPDF object.
|
||
|
imageIndex = 0
|
||
|
this.internal.collections[namespace + 'images'] = images = {}
|
||
|
this.internal.events.subscribe('putResources', putResourcesCallback)
|
||
|
this.internal.events.subscribe('putXobjectDict', putXObjectsDictCallback)
|
||
|
}
|
||
|
|
||
|
var dims = getJpegSize(imageData);
|
||
|
var info = {
|
||
|
w : dims[0],
|
||
|
h : dims[1],
|
||
|
cs : 'DeviceRGB',
|
||
|
bpc : 8,
|
||
|
f : 'DCTDecode',
|
||
|
i : imageIndex,
|
||
|
data : imageData
|
||
|
// n: objectNumber will be added by putImage code
|
||
|
|
||
|
};
|
||
|
images[imageIndex] = info
|
||
|
if (!w && !h) {
|
||
|
w = -96;
|
||
|
h = -96;
|
||
|
}
|
||
|
if (w < 0) {
|
||
|
w = (-1) * info['w'] * 72 / w / this.internal.scaleFactor;
|
||
|
}
|
||
|
if (h < 0) {
|
||
|
h = (-1) * info['h'] * 72 / h / this.internal.scaleFactor;
|
||
|
}
|
||
|
if (w === 0) {
|
||
|
w = h * info['w'] / info['h'];
|
||
|
}
|
||
|
if (h === 0) {
|
||
|
h = w * info['h'] / info['w'];
|
||
|
}
|
||
|
|
||
|
this.internal.write(
|
||
|
'q'
|
||
|
, coord(w)
|
||
|
, '0 0'
|
||
|
, coord(h) // TODO: check if this should be shifted by vcoord
|
||
|
, coord(x)
|
||
|
, vcoord(y + h)
|
||
|
, 'cm /I'+info['i']
|
||
|
, 'Do Q'
|
||
|
)
|
||
|
|
||
|
return this
|
||
|
}
|
||
|
})(jsPDF.API)
|
||
|
/** @preserve
|
||
|
jsPDF fromHTML plugin. BETA stage. API subject to change. Needs browser, jQuery
|
||
|
Copyright (c) 2012 2012 Willow Systems Corporation, willow-systems.com
|
||
|
*/
|
||
|
/*
|
||
|
* Permission is hereby granted, free of charge, to any person obtaining
|
||
|
* a copy of this software and associated documentation files (the
|
||
|
* "Software"), to deal in the Software without restriction, including
|
||
|
* without limitation the rights to use, copy, modify, merge, publish,
|
||
|
* distribute, sublicense, and/or sell copies of the Software, and to
|
||
|
* permit persons to whom the Software is furnished to do so, subject to
|
||
|
* the following conditions:
|
||
|
*
|
||
|
* The above copyright notice and this permission notice shall be
|
||
|
* included in all copies or substantial portions of the Software.
|
||
|
*
|
||
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
||
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
||
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
||
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
||
|
* LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
||
|
* OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
||
|
* WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
||
|
* ====================================================================
|
||
|
*/
|
||
|
|
||
|
;(function(jsPDFAPI) {
|
||
|
'use strict'
|
||
|
|
||
|
|
||
|
if(!String.prototype.trim) {
|
||
|
String.prototype.trim = function () {
|
||
|
return this.replace(/^\s+|\s+$/g,'');
|
||
|
};
|
||
|
}
|
||
|
if(!String.prototype.trimLeft) {
|
||
|
String.prototype.trimLeft = function () {
|
||
|
return this.replace(/^\s+/g,'');
|
||
|
};
|
||
|
}
|
||
|
if(!String.prototype.trimRight) {
|
||
|
String.prototype.trimRight = function () {
|
||
|
return this.replace(/\s+$/g,'');
|
||
|
};
|
||
|
}
|
||
|
|
||
|
function PurgeWhiteSpace(array){
|
||
|
var i = 0, l = array.length, fragment
|
||
|
, lTrimmed = false
|
||
|
, rTrimmed = false
|
||
|
|
||
|
while (!lTrimmed && i !== l) {
|
||
|
fragment = array[i] = array[i].trimLeft()
|
||
|
if (fragment) {
|
||
|
// there is something left there.
|
||
|
lTrimmed = true
|
||
|
}
|
||
|
;i++;
|
||
|
}
|
||
|
|
||
|
i = l - 1
|
||
|
while (l && !rTrimmed && i !== -1) {
|
||
|
fragment = array[i] = array[i].trimRight()
|
||
|
if (fragment) {
|
||
|
// there is something left there.
|
||
|
rTrimmed = true
|
||
|
}
|
||
|
;i--;
|
||
|
}
|
||
|
|
||
|
var r = /\s+$/g
|
||
|
, trailingSpace = true // it's safe to assume we always trim start of display:block element's text.
|
||
|
|
||
|
for (i = 0; i !== l; i++) {
|
||
|
fragment = array[i].replace(/\s+/g, ' ')
|
||
|
// if (l > 1) {
|
||
|
// console.log(i, trailingSpace, fragment)
|
||
|
// }
|
||
|
if (trailingSpace) {
|
||
|
fragment = fragment.trimLeft()
|
||
|
}
|
||
|
if (fragment) {
|
||
|
// meaning, it was not reduced to ""
|
||
|
// if it was, we don't want to clear trailingSpace flag.
|
||
|
trailingSpace = r.test(fragment)
|
||
|
}
|
||
|
array[i] = fragment
|
||
|
}
|
||
|
|
||
|
return array
|
||
|
}
|
||
|
|
||
|
function Renderer(pdf, x, y, settings) {
|
||
|
this.pdf = pdf
|
||
|
this.x = x
|
||
|
this.y = y
|
||
|
this.settings = settings
|
||
|
|
||
|
this.init()
|
||
|
|
||
|
return this
|
||
|
}
|
||
|
|
||
|
Renderer.prototype.init = function(){
|
||
|
|
||
|
this.paragraph = {
|
||
|
'text': []
|
||
|
, 'style': []
|
||
|
}
|
||
|
|
||
|
this.pdf.internal.write(
|
||
|
'q'
|
||
|
)
|
||
|
}
|
||
|
|
||
|
Renderer.prototype.dispose = function(){
|
||
|
this.pdf.internal.write(
|
||
|
'Q' // closes the 'q' in init()
|
||
|
)
|
||
|
return {
|
||
|
'x':this.x, 'y':this.y // bottom left of last line. = upper left of what comes after us.
|
||
|
// TODO: we cannot traverse pages yet, but need to figure out how to communicate that when we do.
|
||
|
// TODO: add more stats: number of lines, paragraphs etc.
|
||
|
}
|
||
|
}
|
||
|
|
||
|
Renderer.prototype.splitFragmentsIntoLines = function(fragments, styles){
|
||
|
var defaultFontSize = 12 // points
|
||
|
, k = this.pdf.internal.scaleFactor // when multiplied by this, converts jsPDF instance units into 'points'
|
||
|
|
||
|
// var widths = options.widths ? options.widths : this.internal.getFont().metadata.Unicode.widths
|
||
|
// , kerning = options.kerning ? options.kerning : this.internal.getFont().metadata.Unicode.kerning
|
||
|
, fontMetricsCache = {}
|
||
|
, ff, fs
|
||
|
, fontMetrics
|
||
|
|
||
|
, fragment // string, element of `fragments`
|
||
|
, style // object with properties with names similar to CSS. Holds pertinent style info for given fragment
|
||
|
, fragmentSpecificMetrics // fontMetrics + some indent and sizing properties populated. We reuse it, hence the bother.
|
||
|
, fragmentLength // fragment's length in jsPDF units
|
||
|
, fragmentChopped // will be array - fragment split into "lines"
|
||
|
|
||
|
, line = [] // array of pairs of arrays [t,s], where t is text string, and s is style object for that t.
|
||
|
, lines = [line] // array of arrays of pairs of arrays
|
||
|
, currentLineLength = 0 // in jsPDF instance units (inches, cm etc)
|
||
|
, maxLineLength = this.settings.width // need to decide if this is the best way to know width of content.
|
||
|
|
||
|
// this loop sorts text fragments (and associated style)
|
||
|
// into lines. Some fragments will be chopped into smaller
|
||
|
// fragments to be spread over multiple lines.
|
||
|
while (fragments.length) {
|
||
|
|
||
|
fragment = fragments.shift()
|
||
|
style = styles.shift()
|
||
|
|
||
|
// if not empty string
|
||
|
if (fragment) {
|
||
|
|
||
|
ff = style['font-family']
|
||
|
fs = style['font-style']
|
||
|
|
||
|
fontMetrics = fontMetricsCache[ff+fs]
|
||
|
if (!fontMetrics) {
|
||
|
fontMetrics = this.pdf.internal.getFont(ff, fs).metadata.Unicode
|
||
|
fontMetricsCache[ff+fs] = fontMetrics
|
||
|
}
|
||
|
|
||
|
fragmentSpecificMetrics = {
|
||
|
'widths': fontMetrics.widths
|
||
|
, 'kerning': fontMetrics.kerning
|
||
|
|
||
|
// fontSize comes to us from CSS scraper as "proportion of normal" value
|
||
|
// , hence the multiplication
|
||
|
, 'fontSize': style['font-size'] * defaultFontSize
|
||
|
|
||
|
// // these should not matter as we provide the metrics manually
|
||
|
// // if we would not, we would need these:
|
||
|
// , 'fontName': style.fontName
|
||
|
// , 'fontStyle': style.fontStyle
|
||
|
|
||
|
// this is setting for "indent first line of paragraph", but we abuse it
|
||
|
// for continuing inline spans of text. Indent value = space in jsPDF instance units
|
||
|
// (whatever user passed to 'new jsPDF(orientation, units, size)
|
||
|
// already consumed on this line. May be zero, of course, for "start of line"
|
||
|
// it's used only on chopper, ignored in all "sizing" code
|
||
|
, 'textIndent': currentLineLength
|
||
|
}
|
||
|
|
||
|
// in user units (inch, cm etc.)
|
||
|
fragmentLength = this.pdf.getStringUnitWidth(
|
||
|
fragment
|
||
|
, fragmentSpecificMetrics
|
||
|
) * fragmentSpecificMetrics.fontSize / k
|
||
|
|
||
|
if (currentLineLength + fragmentLength > maxLineLength) {
|
||
|
// whatever is already on the line + this new fragment
|
||
|
// will be longer than max len for a line.
|
||
|
// Hence, chopping fragment into lines:
|
||
|
fragmentChopped = this.pdf.splitTextToSize(
|
||
|
fragment
|
||
|
, maxLineLength
|
||
|
, fragmentSpecificMetrics
|
||
|
)
|
||
|
|
||
|
line.push([fragmentChopped.shift(), style])
|
||
|
while (fragmentChopped.length){
|
||
|
line = [[fragmentChopped.shift(), style]]
|
||
|
lines.push(line)
|
||
|
}
|
||
|
|
||
|
currentLineLength = this.pdf.getStringUnitWidth(
|
||
|
// new line's first (and only) fragment's length is our new line length
|
||
|
line[0][0]
|
||
|
, fragmentSpecificMetrics
|
||
|
) * fragmentSpecificMetrics.fontSize / k
|
||
|
} else {
|
||
|
// nice, we can fit this fragment on current line. Less work for us...
|
||
|
line.push([fragment, style])
|
||
|
currentLineLength += fragmentLength
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
return lines
|
||
|
}
|
||
|
|
||
|
Renderer.prototype.RenderTextFragment = function(text, style) {
|
||
|
|
||
|
var defaultFontSize = 12
|
||
|
// , header = "/F1 16 Tf\n16 TL\n0 g"
|
||
|
, font = this.pdf.internal.getFont(style['font-family'], style['font-style'])
|
||
|
|
||
|
this.pdf.internal.write(
|
||
|
'/' + font.id // font key
|
||
|
, (defaultFontSize * style['font-size']).toFixed(2) // font size comes as float = proportion to normal.
|
||
|
, 'Tf' // font def marker
|
||
|
, '('+this.pdf.internal.pdfEscape(text)+') Tj'
|
||
|
)
|
||
|
}
|
||
|
|
||
|
Renderer.prototype.renderParagraph = function(){
|
||
|
|
||
|
var fragments = PurgeWhiteSpace( this.paragraph.text )
|
||
|
, styles = this.paragraph.style
|
||
|
, blockstyle = this.paragraph.blockstyle
|
||
|
, priorblockstype = this.paragraph.blockstyle || {}
|
||
|
this.paragraph = {'text':[], 'style':[], 'blockstyle':{}, 'priorblockstyle':blockstyle}
|
||
|
|
||
|
if (!fragments.join('').trim()) {
|
||
|
/* if it's empty string */
|
||
|
return
|
||
|
} // else there is something to draw
|
||
|
|
||
|
var lines = this.splitFragmentsIntoLines(fragments, styles)
|
||
|
, line // will be array of array pairs [[t,s],[t,s],[t,s]...] where t = text, s = style object
|
||
|
|
||
|
, maxLineHeight
|
||
|
, defaultFontSize = 12
|
||
|
, fontToUnitRatio = defaultFontSize / this.pdf.internal.scaleFactor
|
||
|
|
||
|
// these will be in pdf instance units
|
||
|
, paragraphspacing_before = (
|
||
|
// we only use margin-top potion that is larger than margin-bottom of previous elem
|
||
|
// because CSS margins don't stack, they overlap.
|
||
|
Math.max( ( blockstyle['margin-top'] || 0 ) - ( priorblockstype['margin-bottom'] || 0 ), 0 ) +
|
||
|
( blockstyle['padding-top'] || 0 )
|
||
|
) * fontToUnitRatio
|
||
|
, paragraphspacing_after = (
|
||
|
( blockstyle['margin-bottom'] || 0 ) + ( blockstyle['padding-bottom'] || 0 )
|
||
|
) * fontToUnitRatio
|
||
|
|
||
|
, out = this.pdf.internal.write
|
||
|
|
||
|
, i, l
|
||
|
|
||
|
this.y += paragraphspacing_before
|
||
|
|
||
|
out(
|
||
|
'q' // canning the scope
|
||
|
, 'BT' // Begin Text
|
||
|
// and this moves the text start to desired position.
|
||
|
, this.pdf.internal.getCoordinateString(this.x)
|
||
|
, this.pdf.internal.getVerticalCoordinateString(this.y)
|
||
|
, 'Td'
|
||
|
)
|
||
|
|
||
|
// looping through lines
|
||
|
while (lines.length) {
|
||
|
line = lines.shift()
|
||
|
|
||
|
maxLineHeight = 0
|
||
|
|
||
|
for (i = 0, l = line.length; i !== l; i++) {
|
||
|
if (line[i][0].trim()) {
|
||
|
maxLineHeight = Math.max(maxLineHeight, line[i][1]['line-height'], line[i][1]['font-size'])
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// current coordinates are "top left" corner of text box. Text must start from "lower left"
|
||
|
// so, lowering the current coord one line height.
|
||
|
out(
|
||
|
0
|
||
|
, (-1 * defaultFontSize * maxLineHeight).toFixed(2) // shifting down a line in native `points' means reducing y coordinate
|
||
|
, 'Td'
|
||
|
// , (defaultFontSize * maxLineHeight).toFixed(2) // line height comes as float = proportion to normal.
|
||
|
// , 'TL' // line height marker. Not sure we need it with "Td", but...
|
||
|
)
|
||
|
|
||
|
for (i = 0, l = line.length; i !== l; i++) {
|
||
|
if (line[i][0]) {
|
||
|
this.RenderTextFragment(line[i][0], line[i][1])
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// y is in user units (cm, inch etc)
|
||
|
// maxLineHeight is ratio of defaultFontSize
|
||
|
// defaultFontSize is in points always.
|
||
|
// this.internal.scaleFactor is ratio of user unit to points.
|
||
|
// Dividing by it converts points to user units.
|
||
|
// vertical offset will be in user units.
|
||
|
// this.y is in user units.
|
||
|
this.y += maxLineHeight * fontToUnitRatio
|
||
|
}
|
||
|
|
||
|
out(
|
||
|
'ET' // End Text
|
||
|
, 'Q' // restore scope
|
||
|
)
|
||
|
|
||
|
this.y += paragraphspacing_after
|
||
|
}
|
||
|
|
||
|
Renderer.prototype.setBlockBoundary = function(){
|
||
|
this.renderParagraph()
|
||
|
}
|
||
|
|
||
|
Renderer.prototype.setBlockStyle = function(css){
|
||
|
this.paragraph.blockstyle = css
|
||
|
}
|
||
|
|
||
|
Renderer.prototype.addText = function(text, css){
|
||
|
this.paragraph.text.push(text)
|
||
|
this.paragraph.style.push(css)
|
||
|
}
|
||
|
|
||
|
|
||
|
//=====================
|
||
|
// these are DrillForContent and friends
|
||
|
|
||
|
var FontNameDB = {
|
||
|
'helvetica':'helvetica'
|
||
|
, 'sans-serif':'helvetica'
|
||
|
, 'serif':'times'
|
||
|
, 'times':'times'
|
||
|
, 'times new roman':'times'
|
||
|
, 'monospace':'courier'
|
||
|
, 'courier':'courier'
|
||
|
}
|
||
|
, FontWeightMap = {"100":'normal',"200":'normal',"300":'normal',"400":'normal',"500":'bold',"600":'bold',"700":'bold',"800":'bold',"900":'bold',"normal":'normal',"bold":'bold',"bolder":'bold',"lighter":'normal'}
|
||
|
, FontStyleMap = {'normal':'normal','italic':'italic','oblique':'italic'}
|
||
|
, UnitedNumberMap = {'normal':1}
|
||
|
|
||
|
function ResolveFont(css_font_family_string){
|
||
|
var name
|
||
|
, parts = css_font_family_string.split(',') // yes, we don't care about , inside quotes
|
||
|
, part = parts.shift()
|
||
|
|
||
|
while (!name && part){
|
||
|
name = FontNameDB[ part.trim().toLowerCase() ]
|
||
|
part = parts.shift()
|
||
|
}
|
||
|
return name
|
||
|
}
|
||
|
|
||
|
// return ratio to "normal" font size. in other words, it's fraction of 16 pixels.
|
||
|
function ResolveUnitedNumber(css_line_height_string){
|
||
|
var undef
|
||
|
, normal = 16.00
|
||
|
, value = UnitedNumberMap[css_line_height_string]
|
||
|
if (value) {
|
||
|
return value
|
||
|
}
|
||
|
|
||
|
// not in cache, ok. need to parse it.
|
||
|
|
||
|
// Could it be a named value?
|
||
|
// we will use Windows 94dpi sizing with CSS2 suggested 1.2 step ratio
|
||
|
// where "normal" or "medium" is 16px
|
||
|
// see: http://style.cleverchimp.com/font_size_intervals/altintervals.html
|
||
|
value = ({
|
||
|
'xx-small':9
|
||
|
, 'x-small':11
|
||
|
, 'small':13
|
||
|
, 'medium':16
|
||
|
, 'large':19
|
||
|
, 'x-large':23
|
||
|
, 'xx-large':28
|
||
|
, 'auto':0
|
||
|
})[css_line_height_string]
|
||
|
if (value !== undef) {
|
||
|
// caching, returning
|
||
|
return UnitedNumberMap[css_line_height_string] = value / normal
|
||
|
}
|
||
|
|
||
|
// not in cache, ok. need to parse it.
|
||
|
// is it int?
|
||
|
if (value = parseFloat(css_line_height_string)) {
|
||
|
// caching, returning
|
||
|
return UnitedNumberMap[css_line_height_string] = value / normal
|
||
|
}
|
||
|
|
||
|
// must be a "united" value ('123em', '134px' etc.)
|
||
|
// most browsers convert it to px so, only handling the px
|
||
|
value = css_line_height_string.match( /([\d\.]+)(px)/ )
|
||
|
if (value.length === 3) {
|
||
|
// caching, returning
|
||
|
return UnitedNumberMap[css_line_height_string] = parseFloat( value[1] ) / normal
|
||
|
}
|
||
|
|
||
|
return UnitedNumberMap[css_line_height_string] = 1
|
||
|
}
|
||
|
|
||
|
function GetCSS(element){
|
||
|
var $e = $(element)
|
||
|
, css = {}
|
||
|
, tmp
|
||
|
|
||
|
css['font-family'] = ResolveFont( $e.css('font-family') ) || 'times'
|
||
|
css['font-style'] = FontStyleMap [ $e.css('font-style') ] || 'normal'
|
||
|
tmp = FontWeightMap[ $e.css('font-weight') ] || 'normal'
|
||
|
if (tmp === 'bold') {
|
||
|
if (css['font-style'] === 'normal') {
|
||
|
css['font-style'] = tmp
|
||
|
} else {
|
||
|
css['font-style'] = tmp + css['font-style'] // jsPDF's default fonts have it as "bolditalic"
|
||
|
}
|
||
|
}
|
||
|
|
||
|
css['font-size'] = ResolveUnitedNumber( $e.css('font-size') ) || 1 // ratio to "normal" size
|
||
|
css['line-height'] = ResolveUnitedNumber( $e.css('line-height') ) || 1 // ratio to "normal" size
|
||
|
|
||
|
css['display'] = $e.css('display') === 'inline' ? 'inline' : 'block'
|
||
|
|
||
|
if (css['display'] === 'block'){
|
||
|
css['margin-top'] = ResolveUnitedNumber( $e.css('margin-top') ) || 0
|
||
|
css['margin-bottom'] = ResolveUnitedNumber( $e.css('margin-bottom') ) || 0
|
||
|
css['padding-top'] = ResolveUnitedNumber( $e.css('padding-top') ) || 0
|
||
|
css['padding-bottom'] = ResolveUnitedNumber( $e.css('padding-bottom') ) || 0
|
||
|
}
|
||
|
|
||
|
return css
|
||
|
}
|
||
|
|
||
|
function elementHandledElsewhere(element, renderer, elementHandlers){
|
||
|
var isHandledElsewhere = false
|
||
|
|
||
|
var i, l, t
|
||
|
, handlers = elementHandlers['#'+element.id]
|
||
|
if (handlers) {
|
||
|
if (typeof handlers === 'function') {
|
||
|
isHandledElsewhere = handlers(element, renderer)
|
||
|
} else /* array */ {
|
||
|
i = 0
|
||
|
l = handlers.length
|
||
|
while (!isHandledElsewhere && i !== l){
|
||
|
isHandledElsewhere = handlers[i](element, renderer)
|
||
|
;i++;
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
handlers = elementHandlers[element.nodeName]
|
||
|
if (!isHandledElsewhere && handlers) {
|
||
|
if (typeof handlers === 'function') {
|
||
|
isHandledElsewhere = handlers(element, renderer)
|
||
|
} else /* array */ {
|
||
|
i = 0
|
||
|
l = handlers.length
|
||
|
while (!isHandledElsewhere && i !== l){
|
||
|
isHandledElsewhere = handlers[i](element, renderer)
|
||
|
;i++;
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
return isHandledElsewhere
|
||
|
}
|
||
|
|
||
|
function DrillForContent(element, renderer, elementHandlers){
|
||
|
var cns = element.childNodes
|
||
|
, cn
|
||
|
, fragmentCSS = GetCSS(element)
|
||
|
, isBlock = fragmentCSS.display === 'block'
|
||
|
|
||
|
if (isBlock) {
|
||
|
renderer.setBlockBoundary()
|
||
|
renderer.setBlockStyle(fragmentCSS)
|
||
|
}
|
||
|
|
||
|
for (var i = 0, l = cns.length; i < l ; i++){
|
||
|
cn = cns[i]
|
||
|
|
||
|
if (typeof cn === 'object') {
|
||
|
// Don't render the insides of script tags, they contain text nodes which then render
|
||
|
if (cn.nodeType === 1 && cn.nodeName != 'SCRIPT') {
|
||
|
if (!elementHandledElsewhere(cn, renderer, elementHandlers)) {
|
||
|
DrillForContent(cn, renderer, elementHandlers)
|
||
|
}
|
||
|
} else if (cn.nodeType === 3){
|
||
|
renderer.addText( cn.nodeValue, fragmentCSS )
|
||
|
}
|
||
|
} else if (typeof cn === 'string') {
|
||
|
renderer.addText( cn, fragmentCSS )
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if (isBlock) {
|
||
|
renderer.setBlockBoundary()
|
||
|
}
|
||
|
}
|
||
|
|
||
|
function process(pdf, element, x, y, settings) {
|
||
|
|
||
|
// we operate on DOM elems. So HTML-formatted strings need to pushed into one
|
||
|
if (typeof element === 'string') {
|
||
|
element = (function(element) {
|
||
|
var framename = "jsPDFhtmlText" + Date.now().toString() + (Math.random() * 1000).toFixed(0)
|
||
|
, visuallyhidden = 'position: absolute !important;' +
|
||
|
'clip: rect(1px 1px 1px 1px); /* IE6, IE7 */' +
|
||
|
'clip: rect(1px, 1px, 1px, 1px);' +
|
||
|
'padding:0 !important;' +
|
||
|
'border:0 !important;' +
|
||
|
'height: 1px !important;' +
|
||
|
'width: 1px !important; ' +
|
||
|
'top:auto;' +
|
||
|
'left:-100px;' +
|
||
|
'overflow: hidden;'
|
||
|
// TODO: clean up hidden div
|
||
|
, $hiddendiv = $(
|
||
|
'<div style="'+visuallyhidden+'">' +
|
||
|
'<iframe style="height:1px;width:1px" name="'+framename+'" />' +
|
||
|
'</div>'
|
||
|
).appendTo(document.body)
|
||
|
, $frame = window.frames[framename]
|
||
|
return $($frame.document.body).html(element)[0]
|
||
|
})( element )
|
||
|
}
|
||
|
|
||
|
var r = new Renderer( pdf, x, y, settings )
|
||
|
, a = DrillForContent( element, r, settings.elementHandlers )
|
||
|
|
||
|
return r.dispose()
|
||
|
|
||
|
}
|
||
|
|
||
|
|
||
|
/**
|
||
|
Converts HTML-formatted text into formatted PDF text.
|
||
|
|
||
|
Notes:
|
||
|
2012-07-18
|
||
|
Plugin relies on having browser, DOM around. The HTML is pushed into dom and traversed.
|
||
|
Plugin relies on jQuery for CSS extraction.
|
||
|
Targeting HTML output from Markdown templating, which is a very simple
|
||
|
markup - div, span, em, strong, p. No br-based paragraph separation supported explicitly (but still may work.)
|
||
|
Images, tables are NOT supported.
|
||
|
|
||
|
@public
|
||
|
@function
|
||
|
@param HTML {String or DOM Element} HTML-formatted text, or pointer to DOM element that is to be rendered into PDF.
|
||
|
@param x {Number} starting X coordinate in jsPDF instance's declared units.
|
||
|
@param y {Number} starting Y coordinate in jsPDF instance's declared units.
|
||
|
@param settings {Object} Additional / optional variables controlling parsing, rendering.
|
||
|
@returns {Object} jsPDF instance
|
||
|
*/
|
||
|
jsPDFAPI.fromHTML = function(HTML, x, y, settings) {
|
||
|
'use strict'
|
||
|
// `this` is _jsPDF object returned when jsPDF is inited (new jsPDF())
|
||
|
// `this.internal` is a collection of useful, specific-to-raw-PDF-stream functions.
|
||
|
// for example, `this.internal.write` function allowing you to write directly to PDF stream.
|
||
|
// `this.line`, `this.text` etc are available directly.
|
||
|
// so if your plugin just wraps complex series of this.line or this.text or other public API calls,
|
||
|
// you don't need to look into `this.internal`
|
||
|
// See _jsPDF object in jspdf.js for complete list of what's available to you.
|
||
|
|
||
|
// it is good practice to return ref to jsPDF instance to make
|
||
|
// the calls chainable.
|
||
|
// return this
|
||
|
|
||
|
// but in this case it is more usefull to return some stats about what we rendered.
|
||
|
return process(this, HTML, x, y, settings)
|
||
|
}
|
||
|
|
||
|
})(jsPDF.API)
|
||
|
/** @preserve
|
||
|
jsPDF Silly SVG plugin
|
||
|
Copyright (c) 2012 Willow Systems Corporation, willow-systems.com
|
||
|
*/
|
||
|
/**
|
||
|
* Permission is hereby granted, free of charge, to any person obtaining
|
||
|
* a copy of this software and associated documentation files (the
|
||
|
* "Software"), to deal in the Software without restriction, including
|
||
|
* without limitation the rights to use, copy, modify, merge, publish,
|
||
|
* distribute, sublicense, and/or sell copies of the Software, and to
|
||
|
* permit persons to whom the Software is furnished to do so, subject to
|
||
|
* the following conditions:
|
||
|
*
|
||
|
* The above copyright notice and this permission notice shall be
|
||
|
* included in all copies or substantial portions of the Software.
|
||
|
*
|
||
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
||
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
||
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
||
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
||
|
* LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
||
|
* OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
||
|
* WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
||
|
* ====================================================================
|
||
|
*/
|
||
|
|
||
|
;(function(jsPDFAPI) {
|
||
|
'use strict'
|
||
|
|
||
|
/**
|
||
|
Parses SVG XML and converts only some of the SVG elements into
|
||
|
PDF elements.
|
||
|
|
||
|
Supports:
|
||
|
paths
|
||
|
|
||
|
@public
|
||
|
@function
|
||
|
@param
|
||
|
@returns {Type}
|
||
|
*/
|
||
|
jsPDFAPI.addSVG = function(svgtext, x, y, w, h) {
|
||
|
// 'this' is _jsPDF object returned when jsPDF is inited (new jsPDF())
|
||
|
|
||
|
var undef
|
||
|
|
||
|
if (x === undef || x === undef) {
|
||
|
throw new Error("addSVG needs values for 'x' and 'y'")
|
||
|
}
|
||
|
|
||
|
function InjectCSS(cssbody, document) {
|
||
|
var styletag = document.createElement('style')
|
||
|
styletag.type = 'text/css'
|
||
|
if (styletag.styleSheet) {
|
||
|
// ie
|
||
|
styletag.styleSheet.cssText = cssbody
|
||
|
} else {
|
||
|
// others
|
||
|
styletag.appendChild(document.createTextNode(cssbody))
|
||
|
}
|
||
|
document.getElementsByTagName("head")[0].appendChild(styletag)
|
||
|
}
|
||
|
|
||
|
function createWorkerNode(document){
|
||
|
|
||
|
var frameID = 'childframe' // Date.now().toString() + '_' + (Math.random() * 100).toString()
|
||
|
, frame = document.createElement('iframe')
|
||
|
|
||
|
InjectCSS(
|
||
|
'.jsPDF_sillysvg_iframe {display:none;position:absolute;}'
|
||
|
, document
|
||
|
)
|
||
|
|
||
|
frame.name = frameID
|
||
|
frame.setAttribute("width", 0)
|
||
|
frame.setAttribute("height", 0)
|
||
|
frame.setAttribute("frameborder", "0")
|
||
|
frame.setAttribute("scrolling", "no")
|
||
|
frame.setAttribute("seamless", "seamless")
|
||
|
frame.setAttribute("class", "jsPDF_sillysvg_iframe")
|
||
|
|
||
|
document.body.appendChild(frame)
|
||
|
|
||
|
return frame
|
||
|
}
|
||
|
|
||
|
function attachSVGToWorkerNode(svgtext, frame){
|
||
|
var framedoc = ( frame.contentWindow || frame.contentDocument ).document
|
||
|
framedoc.write(svgtext)
|
||
|
framedoc.close()
|
||
|
return framedoc.getElementsByTagName('svg')[0]
|
||
|
}
|
||
|
|
||
|
function convertPathToPDFLinesArgs(path){
|
||
|
'use strict'
|
||
|
// we will use 'lines' method call. it needs:
|
||
|
// - starting coordinate pair
|
||
|
// - array of arrays of vector shifts (2-len for line, 6 len for bezier)
|
||
|
// - scale array [horizontal, vertical] ratios
|
||
|
// - style (stroke, fill, both)
|
||
|
|
||
|
var x = parseFloat(path[1])
|
||
|
, y = parseFloat(path[2])
|
||
|
, vectors = []
|
||
|
, position = 3
|
||
|
, len = path.length
|
||
|
|
||
|
while (position < len){
|
||
|
if (path[position] === 'c'){
|
||
|
vectors.push([
|
||
|
parseFloat(path[position + 1])
|
||
|
, parseFloat(path[position + 2])
|
||
|
, parseFloat(path[position + 3])
|
||
|
, parseFloat(path[position + 4])
|
||
|
, parseFloat(path[position + 5])
|
||
|
, parseFloat(path[position + 6])
|
||
|
])
|
||
|
position += 7
|
||
|
} else if (path[position] === 'l') {
|
||
|
vectors.push([
|
||
|
parseFloat(path[position + 1])
|
||
|
, parseFloat(path[position + 2])
|
||
|
])
|
||
|
position += 3
|
||
|
} else {
|
||
|
position += 1
|
||
|
}
|
||
|
}
|
||
|
return [x,y,vectors]
|
||
|
}
|
||
|
|
||
|
var workernode = createWorkerNode(document)
|
||
|
, svgnode = attachSVGToWorkerNode(svgtext, workernode)
|
||
|
, scale = [1,1]
|
||
|
, svgw = parseFloat(svgnode.getAttribute('width'))
|
||
|
, svgh = parseFloat(svgnode.getAttribute('height'))
|
||
|
|
||
|
if (svgw && svgh) {
|
||
|
// setting both w and h makes image stretch to size.
|
||
|
// this may distort the image, but fits your demanded size
|
||
|
if (w && h) {
|
||
|
scale = [w / svgw, h / svgh]
|
||
|
}
|
||
|
// if only one is set, that value is set as max and SVG
|
||
|
// is scaled proportionately.
|
||
|
else if (w) {
|
||
|
scale = [w / svgw, w / svgw]
|
||
|
} else if (h) {
|
||
|
scale = [h / svgh, h / svgh]
|
||
|
}
|
||
|
}
|
||
|
|
||
|
var i, l, tmp
|
||
|
, linesargs
|
||
|
, items = svgnode.childNodes
|
||
|
for (i = 0, l = items.length; i < l; i++) {
|
||
|
tmp = items[i]
|
||
|
if (tmp.tagName && tmp.tagName.toUpperCase() === 'PATH') {
|
||
|
linesargs = convertPathToPDFLinesArgs( tmp.getAttribute("d").split(' ') )
|
||
|
// path start x coordinate
|
||
|
linesargs[0] = linesargs[0] * scale[0] + x // where x is upper left X of image
|
||
|
// path start y coordinate
|
||
|
linesargs[1] = linesargs[1] * scale[1] + y // where y is upper left Y of image
|
||
|
// the rest of lines are vectors. these will adjust with scale value auto.
|
||
|
this.lines.call(
|
||
|
this
|
||
|
, linesargs[2] // lines
|
||
|
, linesargs[0] // starting x
|
||
|
, linesargs[1] // starting y
|
||
|
, scale
|
||
|
)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// clean up
|
||
|
// workernode.parentNode.removeChild(workernode)
|
||
|
|
||
|
return this
|
||
|
}
|
||
|
|
||
|
})(jsPDF.API)
|
||
|
/** @preserve
|
||
|
jsPDF split_text_to_size plugin
|
||
|
Copyright (c) 2012 Willow Systems Corporation, willow-systems.com
|
||
|
MIT license.
|
||
|
*/
|
||
|
/**
|
||
|
* Permission is hereby granted, free of charge, to any person obtaining
|
||
|
* a copy of this software and associated documentation files (the
|
||
|
* "Software"), to deal in the Software without restriction, including
|
||
|
* without limitation the rights to use, copy, modify, merge, publish,
|
||
|
* distribute, sublicense, and/or sell copies of the Software, and to
|
||
|
* permit persons to whom the Software is furnished to do so, subject to
|
||
|
* the following conditions:
|
||
|
*
|
||
|
* The above copyright notice and this permission notice shall be
|
||
|
* included in all copies or substantial portions of the Software.
|
||
|
*
|
||
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
||
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
||
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
||
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
||
|
* LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
||
|
* OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
||
|
* WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
||
|
* ====================================================================
|
||
|
*/
|
||
|
|
||
|
;(function(API) {
|
||
|
'use strict'
|
||
|
|
||
|
/**
|
||
|
Returns an array of length matching length of the 'word' string, with each
|
||
|
cell ocupied by the width of the char in that position.
|
||
|
|
||
|
@function
|
||
|
@param word {String}
|
||
|
@param widths {Object}
|
||
|
@param kerning {Object}
|
||
|
@returns {Array}
|
||
|
*/
|
||
|
var getCharWidthsArray = API.getCharWidthsArray = function(text, options){
|
||
|
|
||
|
if (!options) {
|
||
|
options = {}
|
||
|
}
|
||
|
|
||
|
var widths = options.widths ? options.widths : this.internal.getFont().metadata.Unicode.widths
|
||
|
, widthsFractionOf = widths.fof ? widths.fof : 1
|
||
|
, kerning = options.kerning ? options.kerning : this.internal.getFont().metadata.Unicode.kerning
|
||
|
, kerningFractionOf = kerning.fof ? kerning.fof : 1
|
||
|
|
||
|
// console.log("widths, kergnings", widths, kerning)
|
||
|
|
||
|
var i, l
|
||
|
, char_code
|
||
|
, char_width
|
||
|
, prior_char_code = 0 // for kerning
|
||
|
, default_char_width = widths[0] || widthsFractionOf
|
||
|
, output = []
|
||
|
|
||
|
for (i = 0, l = text.length; i < l; i++) {
|
||
|
char_code = text.charCodeAt(i)
|
||
|
output.push(
|
||
|
( widths[char_code] || default_char_width ) / widthsFractionOf +
|
||
|
( kerning[char_code] && kerning[char_code][prior_char_code] || 0 ) / kerningFractionOf
|
||
|
)
|
||
|
prior_char_code = char_code
|
||
|
}
|
||
|
|
||
|
return output
|
||
|
}
|
||
|
var getArraySum = function(array){
|
||
|
var i = array.length
|
||
|
, output = 0
|
||
|
while(i){
|
||
|
;i--;
|
||
|
output += array[i]
|
||
|
}
|
||
|
return output
|
||
|
}
|
||
|
/**
|
||
|
Returns a widths of string in a given font, if the font size is set as 1 point.
|
||
|
|
||
|
In other words, this is "proportional" value. For 1 unit of font size, the length
|
||
|
of the string will be that much.
|
||
|
|
||
|
Multiply by font size to get actual width in *points*
|
||
|
Then divide by 72 to get inches or divide by (72/25.6) to get 'mm' etc.
|
||
|
|
||
|
@public
|
||
|
@function
|
||
|
@param
|
||
|
@returns {Type}
|
||
|
*/
|
||
|
var getStringUnitWidth = API.getStringUnitWidth = function(text, options) {
|
||
|
return getArraySum(getCharWidthsArray.call(this, text, options))
|
||
|
}
|
||
|
|
||
|
/**
|
||
|
returns array of lines
|
||
|
*/
|
||
|
var splitLongWord = function(word, widths_array, firstLineMaxLen, maxLen){
|
||
|
var answer = []
|
||
|
|
||
|
// 1st, chop off the piece that can fit on the hanging line.
|
||
|
var i = 0
|
||
|
, l = word.length
|
||
|
, workingLen = 0
|
||
|
while (i !== l && workingLen + widths_array[i] < firstLineMaxLen){
|
||
|
workingLen += widths_array[i]
|
||
|
;i++;
|
||
|
}
|
||
|
// this is first line.
|
||
|
answer.push(word.slice(0, i))
|
||
|
|
||
|
// 2nd. Split the rest into maxLen pieces.
|
||
|
var startOfLine = i
|
||
|
workingLen = 0
|
||
|
while (i !== l){
|
||
|
if (workingLen + widths_array[i] > maxLen) {
|
||
|
answer.push(word.slice(startOfLine, i))
|
||
|
workingLen = 0
|
||
|
startOfLine = i
|
||
|
}
|
||
|
workingLen += widths_array[i]
|
||
|
;i++;
|
||
|
}
|
||
|
if (startOfLine !== i) {
|
||
|
answer.push(word.slice(startOfLine, i))
|
||
|
}
|
||
|
|
||
|
return answer
|
||
|
}
|
||
|
|
||
|
// Note, all sizing inputs for this function must be in "font measurement units"
|
||
|
// By default, for PDF, it's "point".
|
||
|
var splitParagraphIntoLines = function(text, maxlen, options){
|
||
|
// at this time works only on Western scripts, ones with space char
|
||
|
// separating the words. Feel free to expand.
|
||
|
|
||
|
if (!options) {
|
||
|
options = {}
|
||
|
}
|
||
|
|
||
|
var spaceCharWidth = getCharWidthsArray(' ', options)[0]
|
||
|
|
||
|
var words = text.split(' ')
|
||
|
|
||
|
var line = []
|
||
|
, lines = [line]
|
||
|
, line_length = options.textIndent || 0
|
||
|
, separator_length = 0
|
||
|
, current_word_length = 0
|
||
|
, word
|
||
|
, widths_array
|
||
|
|
||
|
var i, l, tmp
|
||
|
for (i = 0, l = words.length; i < l; i++) {
|
||
|
word = words[i]
|
||
|
widths_array = getCharWidthsArray(word, options)
|
||
|
current_word_length = getArraySum(widths_array)
|
||
|
|
||
|
if (line_length + separator_length + current_word_length > maxlen) {
|
||
|
if (current_word_length > maxlen) {
|
||
|
// this happens when you have space-less long URLs for example.
|
||
|
// we just chop these to size. We do NOT insert hiphens
|
||
|
tmp = splitLongWord(word, widths_array, maxlen - (line_length + separator_length), maxlen)
|
||
|
// first line we add to existing line object
|
||
|
line.push(tmp.shift()) // it's ok to have extra space indicator there
|
||
|
// last line we make into new line object
|
||
|
line = [tmp.pop()]
|
||
|
// lines in the middle we apped to lines object as whole lines
|
||
|
while(tmp.length){
|
||
|
lines.push([tmp.shift()]) // single fragment occupies whole line
|
||
|
}
|
||
|
current_word_length = getArraySum( widths_array.slice(word.length - line[0].length) )
|
||
|
} else {
|
||
|
// just put it on a new line
|
||
|
line = [word]
|
||
|
}
|
||
|
|
||
|
// now we attach new line to lines
|
||
|
lines.push(line)
|
||
|
|
||
|
line_length = current_word_length
|
||
|
separator_length = spaceCharWidth
|
||
|
|
||
|
} else {
|
||
|
line.push(word)
|
||
|
|
||
|
line_length += separator_length + current_word_length
|
||
|
separator_length = spaceCharWidth
|
||
|
}
|
||
|
}
|
||
|
|
||
|
var output = []
|
||
|
for (i = 0, l = lines.length; i < l; i++) {
|
||
|
output.push( lines[i].join(' ') )
|
||
|
}
|
||
|
return output
|
||
|
|
||
|
}
|
||
|
|
||
|
/**
|
||
|
Splits a given string into an array of strings. Uses 'size' value
|
||
|
(in measurement units declared as default for the jsPDF instance)
|
||
|
and the font's "widths" and "Kerning" tables, where availabe, to
|
||
|
determine display length of a given string for a given font.
|
||
|
|
||
|
We use character's 100% of unit size (height) as width when Width
|
||
|
table or other default width is not available.
|
||
|
|
||
|
@public
|
||
|
@function
|
||
|
@param text {String} Unencoded, regular JavaScript (Unicode, UTF-16 / UCS-2) string.
|
||
|
@param size {Number} Nominal number, measured in units default to this instance of jsPDF.
|
||
|
@param options {Object} Optional flags needed for chopper to do the right thing.
|
||
|
@returns {Array} with strings chopped to size.
|
||
|
*/
|
||
|
API.splitTextToSize = function(text, maxlen, options) {
|
||
|
'use strict'
|
||
|
|
||
|
if (!options) {
|
||
|
options = {}
|
||
|
}
|
||
|
|
||
|
var fsize = options.fontSize || this.internal.getFontSize()
|
||
|
, newOptions = (function(options){
|
||
|
var widths = {0:1}
|
||
|
, kerning = {}
|
||
|
|
||
|
if (!options.widths || !options.kerning) {
|
||
|
var f = this.internal.getFont(options.fontName, options.fontStyle)
|
||
|
, encoding = 'Unicode'
|
||
|
// NOT UTF8, NOT UTF16BE/LE, NOT UCS2BE/LE
|
||
|
// Actual JavaScript-native String's 16bit char codes used.
|
||
|
// no multi-byte logic here
|
||
|
|
||
|
if (f.metadata[encoding]) {
|
||
|
return {
|
||
|
widths: f.metadata[encoding].widths || widths
|
||
|
, kerning: f.metadata[encoding].kerning || kerning
|
||
|
}
|
||
|
}
|
||
|
} else {
|
||
|
return {
|
||
|
widths: options.widths
|
||
|
, kerning: options.kerning
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// then use default values
|
||
|
return {
|
||
|
widths: widths
|
||
|
, kerning: kerning
|
||
|
}
|
||
|
}).call(this, options)
|
||
|
|
||
|
// first we split on end-of-line chars
|
||
|
var paragraphs
|
||
|
if (text.match(/[\n\r]/)) {
|
||
|
paragraphs = text.split(/\r\n|\r|\n/g)
|
||
|
} else {
|
||
|
paragraphs = [text]
|
||
|
}
|
||
|
|
||
|
// now we convert size (max length of line) into "font size units"
|
||
|
// at present time, the "font size unit" is always 'point'
|
||
|
// 'proportional' means, "in proportion to font size"
|
||
|
var fontUnit_maxLen = 1.0 * this.internal.scaleFactor * maxlen / fsize
|
||
|
// at this time, fsize is always in "points" regardless of the default measurement unit of the doc.
|
||
|
// this may change in the future?
|
||
|
// until then, proportional_maxlen is likely to be in 'points'
|
||
|
|
||
|
// If first line is to be indented (shorter or longer) than maxLen
|
||
|
// we indicate that by using CSS-style "text-indent" option.
|
||
|
// here it's in font units too (which is likely 'points')
|
||
|
// it can be negative (which makes the first line longer than maxLen)
|
||
|
newOptions.textIndent = options.textIndent ?
|
||
|
options.textIndent * 1.0 * this.internal.scaleFactor / fsize :
|
||
|
0
|
||
|
|
||
|
var i, l
|
||
|
, output = []
|
||
|
for (i = 0, l = paragraphs.length; i < l; i++) {
|
||
|
output = output.concat(
|
||
|
splitParagraphIntoLines(
|
||
|
paragraphs[i]
|
||
|
, fontUnit_maxLen
|
||
|
, newOptions
|
||
|
)
|
||
|
)
|
||
|
}
|
||
|
|
||
|
return output
|
||
|
}
|
||
|
|
||
|
})(jsPDF.API);
|
||
|
/** @preserve
|
||
|
jsPDF standard_fonts_metrics plugin
|
||
|
Copyright (c) 2012 Willow Systems Corporation, willow-systems.com
|
||
|
MIT license.
|
||
|
*/
|
||
|
/**
|
||
|
* Permission is hereby granted, free of charge, to any person obtaining
|
||
|
* a copy of this software and associated documentation files (the
|
||
|
* "Software"), to deal in the Software without restriction, including
|
||
|
* without limitation the rights to use, copy, modify, merge, publish,
|
||
|
* distribute, sublicense, and/or sell copies of the Software, and to
|
||
|
* permit persons to whom the Software is furnished to do so, subject to
|
||
|
* the following conditions:
|
||
|
*
|
||
|
* The above copyright notice and this permission notice shall be
|
||
|
* included in all copies or substantial portions of the Software.
|
||
|
*
|
||
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
||
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
||
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
||
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
||
|
* LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
||
|
* OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
||
|
* WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
||
|
* ====================================================================
|
||
|
*/
|
||
|
|
||
|
;(function(API) {
|
||
|
'use strict'
|
||
|
|
||
|
/*
|
||
|
# reference (Python) versions of 'compress' and 'uncompress'
|
||
|
# only 'uncompress' function is featured lower as JavaScript
|
||
|
# if you want to unit test "roundtrip", just transcribe the reference
|
||
|
# 'compress' function from Python into JavaScript
|
||
|
|
||
|
def compress(data):
|
||
|
|
||
|
keys = '0123456789abcdef'
|
||
|
values = 'klmnopqrstuvwxyz'
|
||
|
mapping = dict(zip(keys, values))
|
||
|
vals = []
|
||
|
for key in data.keys():
|
||
|
value = data[key]
|
||
|
try:
|
||
|
keystring = hex(key)[2:]
|
||
|
keystring = keystring[:-1] + mapping[keystring[-1:]]
|
||
|
except:
|
||
|
keystring = key.join(["'","'"])
|
||
|
#print('Keystring is %s' % keystring)
|
||
|
|
||
|
try:
|
||
|
if value < 0:
|
||
|
valuestring = hex(value)[3:]
|
||
|
numberprefix = '-'
|
||
|
else:
|
||
|
valuestring = hex(value)[2:]
|
||
|
numberprefix = ''
|
||
|
valuestring = numberprefix + valuestring[:-1] + mapping[valuestring[-1:]]
|
||
|
except:
|
||
|
if type(value) == dict:
|
||
|
valuestring = compress(value)
|
||
|
else:
|
||
|
raise Exception("Don't know what to do with value type %s" % type(value))
|
||
|
|
||
|
vals.append(keystring+valuestring)
|
||
|
|
||
|
return '{' + ''.join(vals) + '}'
|
||
|
|
||
|
def uncompress(data):
|
||
|
|
||
|
decoded = '0123456789abcdef'
|
||
|
encoded = 'klmnopqrstuvwxyz'
|
||
|
mapping = dict(zip(encoded, decoded))
|
||
|
|
||
|
sign = +1
|
||
|
stringmode = False
|
||
|
stringparts = []
|
||
|
|
||
|
output = {}
|
||
|
|
||
|
activeobject = output
|
||
|
parentchain = []
|
||
|
|
||
|
keyparts = ''
|
||
|
valueparts = ''
|
||
|
|
||
|
key = None
|
||
|
|
||
|
ending = set(encoded)
|
||
|
|
||
|
i = 1
|
||
|
l = len(data) - 1 # stripping starting, ending {}
|
||
|
while i != l: # stripping {}
|
||
|
# -, {, }, ' are special.
|
||
|
|
||
|
ch = data[i]
|
||
|
i += 1
|
||
|
|
||
|
if ch == "'":
|
||
|
if stringmode:
|
||
|
# end of string mode
|
||
|
stringmode = False
|
||
|
key = ''.join(stringparts)
|
||
|
else:
|
||
|
# start of string mode
|
||
|
stringmode = True
|
||
|
stringparts = []
|
||
|
elif stringmode == True:
|
||
|
#print("Adding %s to stringpart" % ch)
|
||
|
stringparts.append(ch)
|
||
|
|
||
|
elif ch == '{':
|
||
|
# start of object
|
||
|
parentchain.append( [activeobject, key] )
|
||
|
activeobject = {}
|
||
|
key = None
|
||
|
#DEBUG = True
|
||
|
elif ch == '}':
|
||
|
# end of object
|
||
|
parent, key = parentchain.pop()
|
||
|
parent[key] = activeobject
|
||
|
key = None
|
||
|
activeobject = parent
|
||
|
#DEBUG = False
|
||
|
|
||
|
elif ch == '-':
|
||
|
sign = -1
|
||
|
else:
|
||
|
# must be number
|
||
|
if key == None:
|
||
|
#debug("In Key. It is '%s', ch is '%s'" % (keyparts, ch))
|
||
|
if ch in ending:
|
||
|
#debug("End of key")
|
||
|
keyparts += mapping[ch]
|
||
|
key = int(keyparts, 16) * sign
|
||
|
sign = +1
|
||
|
keyparts = ''
|
||
|
else:
|
||
|
keyparts += ch
|
||
|
else:
|
||
|
#debug("In value. It is '%s', ch is '%s'" % (valueparts, ch))
|
||
|
if ch in ending:
|
||
|
#debug("End of value")
|
||
|
valueparts += mapping[ch]
|
||
|
activeobject[key] = int(valueparts, 16) * sign
|
||
|
sign = +1
|
||
|
key = None
|
||
|
valueparts = ''
|
||
|
else:
|
||
|
valueparts += ch
|
||
|
|
||
|
#debug(activeobject)
|
||
|
|
||
|
return output
|
||
|
|
||
|
*/
|
||
|
|
||
|
/**
|
||
|
Uncompresses data compressed into custom, base16-like format.
|
||
|
@public
|
||
|
@function
|
||
|
@param
|
||
|
@returns {Type}
|
||
|
*/
|
||
|
var uncompress = function(data){
|
||
|
|
||
|
var decoded = '0123456789abcdef'
|
||
|
, encoded = 'klmnopqrstuvwxyz'
|
||
|
, mapping = {}
|
||
|
|
||
|
for (var i = 0; i < encoded.length; i++){
|
||
|
mapping[encoded[i]] = decoded[i]
|
||
|
}
|
||
|
|
||
|
var undef
|
||
|
, output = {}
|
||
|
, sign = 1
|
||
|
, stringparts // undef. will be [] in string mode
|
||
|
|
||
|
, activeobject = output
|
||
|
, parentchain = []
|
||
|
, parent_key_pair
|
||
|
, keyparts = ''
|
||
|
, valueparts = ''
|
||
|
, key // undef. will be Truthy when Key is resolved.
|
||
|
, datalen = data.length - 1 // stripping ending }
|
||
|
, ch
|
||
|
|
||
|
i = 1 // stripping starting {
|
||
|
|
||
|
while (i != datalen){
|
||
|
// - { } ' are special.
|
||
|
|
||
|
ch = data[i]
|
||
|
i += 1
|
||
|
|
||
|
if (ch == "'"){
|
||
|
if (stringparts){
|
||
|
// end of string mode
|
||
|
key = stringparts.join('')
|
||
|
stringparts = undef
|
||
|
} else {
|
||
|
// start of string mode
|
||
|
stringparts = []
|
||
|
}
|
||
|
} else if (stringparts){
|
||
|
stringparts.push(ch)
|
||
|
} else if (ch == '{'){
|
||
|
// start of object
|
||
|
parentchain.push( [activeobject, key] )
|
||
|
activeobject = {}
|
||
|
key = undef
|
||
|
} else if (ch == '}'){
|
||
|
// end of object
|
||
|
parent_key_pair = parentchain.pop()
|
||
|
parent_key_pair[0][parent_key_pair[1]] = activeobject
|
||
|
key = undef
|
||
|
activeobject = parent_key_pair[0]
|
||
|
} else if (ch == '-'){
|
||
|
sign = -1
|
||
|
} else {
|
||
|
// must be number
|
||
|
if (key === undef) {
|
||
|
if (mapping.hasOwnProperty(ch)){
|
||
|
keyparts += mapping[ch]
|
||
|
key = parseInt(keyparts, 16) * sign
|
||
|
sign = +1
|
||
|
keyparts = ''
|
||
|
} else {
|
||
|
keyparts += ch
|
||
|
}
|
||
|
} else {
|
||
|
if (mapping.hasOwnProperty(ch)){
|
||
|
valueparts += mapping[ch]
|
||
|
activeobject[key] = parseInt(valueparts, 16) * sign
|
||
|
sign = +1
|
||
|
key = undef
|
||
|
valueparts = ''
|
||
|
} else {
|
||
|
valueparts += ch
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
} // end while
|
||
|
|
||
|
return output
|
||
|
}
|
||
|
|
||
|
// encoding = 'Unicode'
|
||
|
// NOT UTF8, NOT UTF16BE/LE, NOT UCS2BE/LE. NO clever BOM behavior
|
||
|
// Actual 16bit char codes used.
|
||
|
// no multi-byte logic here
|
||
|
|
||
|
// Unicode characters to WinAnsiEncoding:
|
||
|
// {402: 131, 8211: 150, 8212: 151, 8216: 145, 8217: 146, 8218: 130, 8220: 147, 8221: 148, 8222: 132, 8224: 134, 8225: 135, 8226: 149, 8230: 133, 8364: 128, 8240:137, 8249: 139, 8250: 155, 710: 136, 8482: 153, 338: 140, 339: 156, 732: 152, 352: 138, 353: 154, 376: 159, 381: 142, 382: 158}
|
||
|
// as you can see, all Unicode chars are outside of 0-255 range. No char code conflicts.
|
||
|
// this means that you can give Win cp1252 encoded strings to jsPDF for rendering directly
|
||
|
// as well as give strings with some (supported by these fonts) Unicode characters and
|
||
|
// these will be mapped to win cp1252
|
||
|
// for example, you can send char code (cp1252) 0x80 or (unicode) 0x20AC, getting "Euro" glyph displayed in both cases.
|
||
|
|
||
|
var encodingBlock = {
|
||
|
'codePages': ['WinAnsiEncoding']
|
||
|
, 'WinAnsiEncoding': uncompress("{19m8n201n9q201o9r201s9l201t9m201u8m201w9n201x9o201y8o202k8q202l8r202m9p202q8p20aw8k203k8t203t8v203u9v2cq8s212m9t15m8w15n9w2dw9s16k8u16l9u17s9z17x8y17y9y}")
|
||
|
}
|
||
|
, encodings = {'Unicode':{
|
||
|
'Courier': encodingBlock
|
||
|
, 'Courier-Bold': encodingBlock
|
||
|
, 'Courier-BoldOblique': encodingBlock
|
||
|
, 'Courier-Oblique': encodingBlock
|
||
|
, 'Helvetica': encodingBlock
|
||
|
, 'Helvetica-Bold': encodingBlock
|
||
|
, 'Helvetica-BoldOblique': encodingBlock
|
||
|
, 'Helvetica-Oblique': encodingBlock
|
||
|
, 'Times-Roman': encodingBlock
|
||
|
, 'Times-Bold': encodingBlock
|
||
|
, 'Times-BoldItalic': encodingBlock
|
||
|
, 'Times-Italic': encodingBlock
|
||
|
// , 'Symbol'
|
||
|
// , 'ZapfDingbats'
|
||
|
}}
|
||
|
/**
|
||
|
Resources:
|
||
|
Font metrics data is reprocessed derivative of contents of
|
||
|
"Font Metrics for PDF Core 14 Fonts" package, which exhibits the following copyright and license:
|
||
|
|
||
|
Copyright (c) 1989, 1990, 1991, 1992, 1993, 1997 Adobe Systems Incorporated. All Rights Reserved.
|
||
|
|
||
|
This file and the 14 PostScript(R) AFM files it accompanies may be used,
|
||
|
copied, and distributed for any purpose and without charge, with or without
|
||
|
modification, provided that all copyright notices are retained; that the AFM
|
||
|
files are not distributed without this file; that all modifications to this
|
||
|
file or any of the AFM files are prominently noted in the modified file(s);
|
||
|
and that this paragraph is not modified. Adobe Systems has no responsibility
|
||
|
or obligation to support the use of the AFM files.
|
||
|
|
||
|
*/
|
||
|
, fontMetrics = {'Unicode':{
|
||
|
// all sizing numbers are n/fontMetricsFractionOf = one font size unit
|
||
|
// this means that if fontMetricsFractionOf = 1000, and letter A's width is 476, it's
|
||
|
// width is 476/1000 or 47.6% of its height (regardless of font size)
|
||
|
// At this time this value applies to "widths" and "kerning" numbers.
|
||
|
|
||
|
// char code 0 represents "default" (average) width - use it for chars missing in this table.
|
||
|
// key 'fof' represents the "fontMetricsFractionOf" value
|
||
|
|
||
|
'Courier-Oblique': uncompress("{'widths'{k3w'fof'6o}'kerning'{'fof'-6o}}")
|
||
|
, 'Times-BoldItalic': uncompress("{'widths'{k3o2q4ycx2r201n3m201o6o201s2l201t2l201u2l201w3m201x3m201y3m2k1t2l2r202m2n2n3m2o3m2p5n202q6o2r1w2s2l2t2l2u3m2v3t2w1t2x2l2y1t2z1w3k3m3l3m3m3m3n3m3o3m3p3m3q3m3r3m3s3m203t2l203u2l3v2l3w3t3x3t3y3t3z3m4k5n4l4m4m4m4n4m4o4s4p4m4q4m4r4s4s4y4t2r4u3m4v4m4w3x4x5t4y4s4z4s5k3x5l4s5m4m5n3r5o3x5p4s5q4m5r5t5s4m5t3x5u3x5v2l5w1w5x2l5y3t5z3m6k2l6l3m6m3m6n2w6o3m6p2w6q2l6r3m6s3r6t1w6u1w6v3m6w1w6x4y6y3r6z3m7k3m7l3m7m2r7n2r7o1w7p3r7q2w7r4m7s3m7t2w7u2r7v2n7w1q7x2n7y3t202l3mcl4mal2ram3man3mao3map3mar3mas2lat4uau1uav3maw3way4uaz2lbk2sbl3t'fof'6obo2lbp3tbq3mbr1tbs2lbu1ybv3mbz3mck4m202k3mcm4mcn4mco4mcp4mcq5ycr4mcs4mct4mcu4mcv4mcw2r2m3rcy2rcz2rdl4sdm4sdn4sdo4sdp4sdq4sds4sdt4sdu4sdv4sdw4sdz3mek3mel3mem3men3meo3mep3meq4ser2wes2wet2weu2wev2wew1wex1wey1wez1wfl3rfm3mfn3mfo3mfp3mfq3mfr3tfs3mft3rfu3rfv3rfw3rfz2w203k6o212m6o2dw2l2cq2l3t3m3u2l17s3x19m3m}'kerning'{cl{4qu5kt5qt5rs17ss5ts}201s{201ss}201t{cks4lscmscnscoscpscls2wu2yu201ts}201x{2wu2yu}2k{201ts}2w{4qx5kx5ou5qx5rs17su5tu}2x{17su5tu5ou}2y{4qx5kx5ou5qx5rs17ss5ts}'fof'-6ofn{17sw5tw5ou5qw5rs}7t{cksclscmscnscoscps4ls}3u{17su5tu5os5qs}3v{17su5tu5os5qs}7p{17su5tu}ck{4qu5kt5qt5rs17ss5ts}4l{4qu5kt5qt5rs17ss5ts}cm{4qu5kt5qt5rs17ss5ts}cn{4qu5kt5qt5rs17ss5ts}co{4qu5kt5qt5rs17ss5ts}cp{4qu5kt5qt5rs17ss5ts}6l{4qu5ou5qw5rt17su5tu}5q{ckuclucmucnucoucpu4lu}5r{ckuclucmucnucoucpu4lu}7q{cksclscmscnscoscps4ls}6p{4qu5ou5qw5rt17sw5tw}ek{4qu5ou5qw5rt17su5tu}el{4qu5ou5qw5rt17su5tu}em{4qu5ou5qw5rt17su5tu}en{4qu5ou5qw5rt17su5tu}eo{4qu5ou5qw5rt17su5tu}ep{4qu5ou5qw5rt17su5tu}es{17ss5ts5qs4qu}et{4qu5ou5qw5rt17sw5tw}eu{4qu5ou5qw5rt17ss5ts}ev{17ss5ts5qs4qu}6z{17sw5tw5ou5qw5rs}fm{17sw5tw5ou5qw5rs}7n{201ts}fo{17sw5tw5ou5qw5rs}fp{17sw5tw5ou5qw5rs}fq{17sw5tw5ou5qw5rs}7r{cksclscmscnscoscps4ls}fs{17sw5tw5ou5qw5rs}ft{17su5tu}fu{17su5tu}fv{17su5tu}fw{17su5tu}fz{cksclscmscnscoscps4ls}}}")
|
||
|
, 'Helvetica-Bold': uncompress("{'widths'{k3s2q4scx1w201n3r201o6o201s1w201t1w201u1w201w3m201x3m201y3m2k1w2l2l202m2n2n3r2o3r2p5t202q6o2r1s2s2l2t2l2u2r2v3u2w1w2x2l2y1w2z1w3k3r3l3r3m3r3n3r3o3r3p3r3q3r3r3r3s3r203t2l203u2l3v2l3w3u3x3u3y3u3z3x4k6l4l4s4m4s4n4s4o4s4p4m4q3x4r4y4s4s4t1w4u3r4v4s4w3x4x5n4y4s4z4y5k4m5l4y5m4s5n4m5o3x5p4s5q4m5r5y5s4m5t4m5u3x5v2l5w1w5x2l5y3u5z3r6k2l6l3r6m3x6n3r6o3x6p3r6q2l6r3x6s3x6t1w6u1w6v3r6w1w6x5t6y3x6z3x7k3x7l3x7m2r7n3r7o2l7p3x7q3r7r4y7s3r7t3r7u3m7v2r7w1w7x2r7y3u202l3rcl4sal2lam3ran3rao3rap3rar3ras2lat4tau2pav3raw3uay4taz2lbk2sbl3u'fof'6obo2lbp3xbq3rbr1wbs2lbu2obv3rbz3xck4s202k3rcm4scn4sco4scp4scq6ocr4scs4mct4mcu4mcv4mcw1w2m2zcy1wcz1wdl4sdm4ydn4ydo4ydp4ydq4yds4ydt4sdu4sdv4sdw4sdz3xek3rel3rem3ren3reo3rep3req5ter3res3ret3reu3rev3rew1wex1wey1wez1wfl3xfm3xfn3xfo3xfp3xfq3xfr3ufs3xft3xfu3xfv3xfw3xfz3r203k6o212m6o2dw2l2cq2l3t3r3u2l17s4m19m3r}'kerning'{cl{4qs5ku5ot5qs17sv5tv}201t{2ww4wy2yw}201w{2ks}201x{2ww4wy2yw}2k{201ts201xs}2w{7qs4qu5kw5os5qw5rs17su5tu7tsfzs}2x{5ow5qs}2y{7qs4qu5kw5os5qw5rs17su5tu7tsfzs}'fof'-6o7p{17su5tu5ot}ck{4qs5ku5ot5qs17sv5tv}4l{4qs5ku5ot5qs17sv5tv}cm{4qs5ku5ot5qs17sv5tv}cn{4qs5ku5ot5qs17sv5tv}co{4qs5ku5ot5qs17sv5tv}cp{4qs5ku5ot5qs17sv5tv}6l{17st5tt5os}17s{2kwclvcmvcnvcovcpv4lv4wwckv}5o{2kucltcmtcntcotcpt4lt4wtckt}5q{2ksclscmscnscoscps4ls4wvcks}5r{2ks4ws}5t{2kwclvcmvcnvcovcpv4lv4wwckv}eo{17st5tt5os}fu{17su5tu5ot}6p{17ss5ts}ek{17st5tt5os}el{17st5tt5os}em{17st5tt5os}en{17st5tt5os}6o{201ts}ep{17st5tt5os}es{17ss5ts}et{17ss5ts}eu{17ss5ts}ev{17ss5ts}6z{17su5tu5os5qt}fm{17su5tu5os5qt}fn{17su5tu5os5qt}fo{17su5tu5os5qt}fp{17su5tu5os5qt}fq{17su5tu5os5qt}fs{17su5tu5os5qt}ft{17su5tu5ot}7m{5os}fv{17su5tu5ot}fw{17su5tu5ot}}}")
|
||
|
, 'Courier': uncompress("{'widths'{k3w'fof'6o}'kerning'{'fof'-6o}}")
|
||
|
, 'Courier-BoldOblique': uncompress("{'widths'{k3w'fof'6o}'kerning'{'fof'-6o}}")
|
||
|
, 'Times-Bold': uncompress("{'widths'{k3q2q5ncx2r201n3m201o6o201s2l201t2l201u2l201w3m201x3m201y3m2k1t2l2l202m2n2n3m2o3m2p6o202q6o2r1w2s2l2t2l2u3m2v3t2w1t2x2l2y1t2z1w3k3m3l3m3m3m3n3m3o3m3p3m3q3m3r3m3s3m203t2l203u2l3v2l3w3t3x3t3y3t3z3m4k5x4l4s4m4m4n4s4o4s4p4m4q3x4r4y4s4y4t2r4u3m4v4y4w4m4x5y4y4s4z4y5k3x5l4y5m4s5n3r5o4m5p4s5q4s5r6o5s4s5t4s5u4m5v2l5w1w5x2l5y3u5z3m6k2l6l3m6m3r6n2w6o3r6p2w6q2l6r3m6s3r6t1w6u2l6v3r6w1w6x5n6y3r6z3m7k3r7l3r7m2w7n2r7o2l7p3r7q3m7r4s7s3m7t3m7u2w7v2r7w1q7x2r7y3o202l3mcl4sal2lam3man3mao3map3mar3mas2lat4uau1yav3maw3tay4uaz2lbk2sbl3t'fof'6obo2lbp3rbr1tbs2lbu2lbv3mbz3mck4s202k3mcm4scn4sco4scp4scq6ocr4scs4mct4mcu4mcv4mcw2r2m3rcy2rcz2rdl4sdm4ydn4ydo4ydp4ydq4yds4ydt4sdu4sdv4sdw4sdz3rek3mel3mem3men3meo3mep3meq4ser2wes2wet2weu2wev2wew1wex1wey1wez1wfl3rfm3mfn3mfo3mfp3mfq3mfr3tfs3mft3rfu3rfv3rfw3rfz3m203k6o212m6o2dw2l2cq2l3t3m3u2l17s4s19m3m}'kerning'{cl{4qt5ks5ot5qy5rw17sv5tv}201t{cks4lscmscnscoscpscls4wv}2k{201ts}2w{4qu5ku7mu5os5qx5ru17su5tu}2x{17su5tu5ou5qs}2y{4qv5kv7mu5ot5qz5ru17su5tu}'fof'-6o7t{cksclscmscnscoscps4ls}3u{17su5tu5os5qu}3v{17su5tu5os5qu}fu{17su5tu5ou5qu}7p{17su5tu5ou5qu}ck{4qt5ks5ot5qy5rw17sv5tv}4l{4qt5ks5ot5qy5rw17sv5tv}cm{4qt5ks5ot5qy5rw17sv5tv}cn{4qt5ks5ot5qy5rw17sv5tv}co{4qt5ks5ot5qy5rw17sv5tv}cp{4qt5ks5ot5qy5rw17sv5tv}6l{17st5tt5ou5qu}17s{ckuclucmucnucoucpu4lu4wu}5o{ckuclucmucnucoucpu4lu4wu}5q{ckzclzcmzcnzcozcpz4lz4wu}5r{ckxclxcmxcnxcoxcpx4lx4wu}5t{ckuclucmucnucoucpu4lu4wu}7q{ckuclucmucnucoucpu4lu}6p{17sw5tw5ou5qu}ek{17st5tt5qu}el{17st5tt5ou5qu}em{17st5tt5qu}en{17st5tt5qu}eo{17st5tt5qu}ep{17st5tt5ou5qu}es{17ss5ts5qu}et{17sw5tw5ou5qu}eu{17sw5tw5ou5qu}ev{17ss5ts5qu}6z{17sw5tw5ou5qu5rs}fm{17sw5tw5ou5qu5rs}fn{17sw5tw5ou5qu5rs}fo{17sw5tw5ou5qu5rs}fp{17sw5tw5ou5qu5rs}fq{17sw5tw5ou5qu5rs}7r{cktcltcmtcntcotcpt4lt5os}fs{17sw5tw5ou5qu5rs}ft{17su5tu5ou5qu}7m{5os}fv{17su5tu5ou5qu}fw{17su5tu5ou5qu}fz{cksclscmscnscoscps4ls}}}")
|
||
|
//, 'Symbol': uncompress("{'widths'{k3uaw4r19m3m2k1t2l2l202m2y2n3m2p5n202q6o3k3m2s2l2t2l2v3r2w1t3m3m2y1t2z1wbk2sbl3r'fof'6o3n3m3o3m3p3m3q3m3r3m3s3m3t3m3u1w3v1w3w3r3x3r3y3r3z2wbp3t3l3m5v2l5x2l5z3m2q4yfr3r7v3k7w1o7x3k}'kerning'{'fof'-6o}}")
|
||
|
, 'Helvetica': uncompress("{'widths'{k3p2q4mcx1w201n3r201o6o201s1q201t1q201u1q201w2l201x2l201y2l2k1w2l1w202m2n2n3r2o3r2p5t202q6o2r1n2s2l2t2l2u2r2v3u2w1w2x2l2y1w2z1w3k3r3l3r3m3r3n3r3o3r3p3r3q3r3r3r3s3r203t2l203u2l3v1w3w3u3x3u3y3u3z3r4k6p4l4m4m4m4n4s4o4s4p4m4q3x4r4y4s4s4t1w4u3m4v4m4w3r4x5n4y4s4z4y5k4m5l4y5m4s5n4m5o3x5p4s5q4m5r5y5s4m5t4m5u3x5v1w5w1w5x1w5y2z5z3r6k2l6l3r6m3r6n3m6o3r6p3r6q1w6r3r6s3r6t1q6u1q6v3m6w1q6x5n6y3r6z3r7k3r7l3r7m2l7n3m7o1w7p3r7q3m7r4s7s3m7t3m7u3m7v2l7w1u7x2l7y3u202l3rcl4mal2lam3ran3rao3rap3rar3ras2lat4tau2pav3raw3uay4taz2lbk2sbl3u'fof'6obo2lbp3rbr1wbs2lbu2obv3rbz3xck4m202k3rcm4mcn4mco4mcp4mcq6ocr4scs4mct4mcu4mcv4mcw1w2m2ncy1wcz1wdl4sdm4ydn4ydo4ydp4ydq4yds4ydt4sdu4sdv4sdw4sdz3xek3rel3rem3ren3reo3rep3req5ter3mes3ret3reu3rev3rew1wex1wey1wez1wfl3rfm3rfn3rfo3rfp3rfq3rfr3ufs3xft3rfu3rfv3rfw3rfz3m203k6o212m6o2dw2l2cq2l3t3r3u1w17s4m19m3r}'kerning'{5q{4wv}cl{4qs5kw5ow5qs17sv5tv}201t{2wu4w1k2yu}201x{2wu4wy2yu}17s{2ktclucmucnu4otcpu4lu4wycoucku}2w{7qs4qz5k1m17sy5ow5qx5rsfsu5ty7tufzu}2x{17sy5ty5oy5qs}2y{7qs4qz5k1m17sy5ow5qx5rsfsu5ty7tufzu}'fof'-6o7p{17sv5tv5ow}ck{4qs5kw5ow5qs17sv5tv}4l{4qs5kw5ow5qs17sv5tv}cm{4qs5kw5ow5qs17sv5tv}cn{4qs5kw5ow5qs17sv5tv}co{4qs5kw5ow5qs17sv5tv}cp{4qs5kw5ow5qs17sv5tv}6l{17sy5ty5ow}do{17st5tt}4z{17st5tt}7s{fst}dm{17st5tt}dn{17st5tt}5o{ckwclwcmwcnwcowcpw4lw4wv}dp{17st5tt}dq{17st5tt}7t{5ow}ds{17st5tt}5t{2ktclucmucnu4otcpu4lu4wycoucku}fu{17sv5tv5ow}6p{17sy5ty5ow5qs}ek{17sy5ty5ow}el{17sy5ty5ow}em{17sy5ty5ow}en{5ty}eo{17sy5ty5ow}ep{17sy5ty5ow}es{17sy5ty5qs}et{17sy5ty5ow5qs}eu{17sy5ty5ow5qs}ev{17sy5ty5ow5qs}6z{17sy5ty5ow5qs}fm{17sy5ty5ow5qs}fn{17sy5ty5ow5qs}fo{17sy5ty5ow5qs}fp{17sy5ty5qs}fq{17sy5ty5ow5qs}7r{5ow}fs{17sy5ty5ow5qs}ft{17sv5tv5ow}7m{5ow}fv{17sv5tv5ow}fw{17sv5tv5ow}}}")
|
||
|
, 'Helvetica-BoldOblique': uncompress("{'widths'{k3s2q4scx1w201n3r201o6o201s1w201t1w201u1w201w3m201x3m201y3m2k1w2l2l202m2n2n3r2o3r2p5t202q6o2r1s2s2l2t2l2u2r2v3u2w1w2x2l2y1w2z1w3k3r3l3r3m3r3n3r3o3r3p3r3q3r3r3r3s3r203t2l203u2l3v2l3w3u3x3u3y3u3z3x4k6l4l4s4m4s4n4s4o4s4p4m4q3x4r4y4s4s4t1w4u3r4v4s4w3x4x5n4y4s4z4y5k4m5l4y5m4s5n4m5o3x5p4s5q4m5r5y5s4m5t4m5u3x5v2l5w1w5x2l5y3u5z3r6k2l6l3r6m3x6n3r6o3x6p3r6q2l6r3x6s3x6t1w6u1w6v3r6w1w6x5t6y3x6z3x7k3x7l3x7m2r7n3r7o2l7p3x7q3r7r4y7s3r7t3r7u3m7v2r7w1w7x2r7y3u202l3rcl4sal2lam3ran3rao3rap3rar3ras2lat4tau2pav3raw3uay4taz2lbk2sbl3u'fof'6obo2lbp3xbq3rbr1wbs2lbu2obv3rbz3xck4s202k3rcm4scn4sco4scp4scq6ocr4scs4mct4mcu4mcv4mcw1w2m2zcy1wcz1wdl4sdm4ydn4ydo4ydp4ydq4yds4ydt4sdu4sdv4sdw4sdz3xek3rel3rem3ren3reo3rep3req5ter3res3ret3reu3rev3rew1wex1wey1wez1wfl3xfm3xfn3xfo3xfp3xfq3xfr3ufs3xft3xfu3xfv3xfw3xfz3r203k6o212m6o2dw2l2cq2l3t3r3u2l17s4m19m3r}'kerning'{cl{4qs5ku5ot5qs17sv5tv}201t{2ww4wy2yw}201w{2ks}201x{2ww4wy2yw}2k{201ts201xs}2w{7qs4qu5kw5os5qw5rs17su5tu7tsfzs}2x{5ow5qs}2y{7qs4qu5kw5os5qw5rs17su5tu7tsfzs}'fof'-6o7p{17su5tu5ot}ck{4qs5ku5ot5qs17sv5tv}4l{4qs5ku5ot5qs17sv5tv}cm{4qs5ku5ot5qs17sv5tv}cn{4qs5ku5ot5qs17sv5tv}co{4qs5ku5ot5qs17sv5tv}cp{4qs5ku5ot5qs17sv5tv}6l{17st5tt5os}17s{2kwclvcmvcnvcovcpv4lv4wwckv}5o{2kucltcmtcntcotcpt4lt4wtckt}5q{2ksclscmscnscoscps4ls4wvcks}5r{2ks4ws}5t{2kwclvcmvcnvcovcpv4lv4wwckv}eo{17st5tt5os}fu{17su5tu5ot}6p{17ss5ts}ek{17st5tt5os}el{17st5tt5os}em{17st5tt5os}en{17st5tt5os}6o{201ts}ep{17st5tt5os}es{17ss5ts}et{17ss5ts}eu{17ss5ts}ev{17ss5ts}6z{17su5tu5os5qt}fm{17su5tu5os5qt}fn{17su5tu5os5qt}fo{17su5tu5os5qt}fp{17su5tu5os5qt}fq{17su5tu5os5qt}fs{17su5tu5os5qt}ft{17su5tu5ot}7m{5os}fv{17su5tu5ot}fw{17su5tu5ot}}}")
|
||
|
//, 'ZapfDingbats': uncompress("{'widths'{k4u2k1w'fof'6o}'kerning'{'fof'-6o}}")
|
||
|
, 'Courier-Bold': uncompress("{'widths'{k3w'fof'6o}'kerning'{'fof'-6o}}")
|
||
|
, 'Times-Italic': uncompress("{'widths'{k3n2q4ycx2l201n3m201o5t201s2l201t2l201u2l201w3r201x3r201y3r2k1t2l2l202m2n2n3m2o3m2p5n202q5t2r1p2s2l2t2l2u3m2v4n2w1t2x2l2y1t2z1w3k3m3l3m3m3m3n3m3o3m3p3m3q3m3r3m3s3m203t2l203u2l3v2l3w4n3x4n3y4n3z3m4k5w4l3x4m3x4n4m4o4s4p3x4q3x4r4s4s4s4t2l4u2w4v4m4w3r4x5n4y4m4z4s5k3x5l4s5m3x5n3m5o3r5p4s5q3x5r5n5s3x5t3r5u3r5v2r5w1w5x2r5y2u5z3m6k2l6l3m6m3m6n2w6o3m6p2w6q1w6r3m6s3m6t1w6u1w6v2w6w1w6x4s6y3m6z3m7k3m7l3m7m2r7n2r7o1w7p3m7q2w7r4m7s2w7t2w7u2r7v2s7w1v7x2s7y3q202l3mcl3xal2ram3man3mao3map3mar3mas2lat4wau1vav3maw4nay4waz2lbk2sbl4n'fof'6obo2lbp3mbq3obr1tbs2lbu1zbv3mbz3mck3x202k3mcm3xcn3xco3xcp3xcq5tcr4mcs3xct3xcu3xcv3xcw2l2m2ucy2lcz2ldl4mdm4sdn4sdo4sdp4sdq4sds4sdt4sdu4sdv4sdw4sdz3mek3mel3mem3men3meo3mep3meq4mer2wes2wet2weu2wev2wew1wex1wey1wez1wfl3mfm3mfn3mfo3mfp3mfq3mfr4nfs3mft3mfu3mfv3mfw3mfz2w203k6o212m6m2dw2l2cq2l3t3m3u2l17s3r19m3m}'kerning'{cl{5kt4qw}201s{201sw}201t{201tw2wy2yy6q-t}201x{2wy2yy}2k{201tw}2w{7qs4qy7rs5ky7mw5os5qx5ru17su5tu}2x{17ss5ts5os}2y{7qs4qy7rs5ky7mw5os5qx5ru17su5tu}'fof'-6o6t{17ss5ts5qs}7t{5os}3v{5qs}7p{17su5tu5qs}ck{5kt4qw}4l{5kt4qw}cm{5kt4qw}cn{5kt4qw}co{5kt4qw}cp{5kt4qw}6l{4qs5ks5ou5qw5ru17su5tu}17s{2ks}5q{ckvclvcmvcnvcovcpv4lv}5r{ckuclucmucnucoucpu4lu}5t{2ks}6p{4qs5ks5ou5qw5ru17su5tu}ek{4qs5ks5ou5qw5ru17su5tu}el{4qs5ks5ou5qw5ru17su5tu}em{4qs5ks5ou5qw5ru17su5tu}en{4qs5ks5ou5qw5ru17su5tu}eo{4qs5ks5ou5qw5ru17su5tu}ep{4qs5ks5ou5qw5ru17su5tu}es{5ks5qs4qs}et{4qs5ks5ou5qw5ru17su5tu}eu{4qs5ks5qw5ru17su5tu}ev{5ks5qs4qs}ex{17ss5ts5qs}6z{4qv5ks5ou5qw5ru17su5tu}fm{4qv5ks5ou5qw5ru17su5tu}fn{4qv5ks5ou5qw5ru17su5tu}fo{4qv5ks5ou5qw5ru17su5tu}fp{4qv5ks5ou5qw5ru17su5tu}fq{4qv5ks5ou5qw5ru17su5tu}7r{5os}fs{4qv5ks5ou5qw5ru17su5tu}ft{17su5tu5qs}fu{17su5tu5qs}fv{17su5tu5qs}fw{17su5tu5qs}}}")
|
||
|
, 'Times-Roman': uncompress("{'widths'{k3n2q4ycx2l201n3m201o6o201s2l201t2l201u2l201w2w201x2w201y2w2k1t2l2l202m2n2n3m2o3m2p5n202q6o2r1m2s2l2t2l2u3m2v3s2w1t2x2l2y1t2z1w3k3m3l3m3m3m3n3m3o3m3p3m3q3m3r3m3s3m203t2l203u2l3v1w3w3s3x3s3y3s3z2w4k5w4l4s4m4m4n4m4o4s4p3x4q3r4r4s4s4s4t2l4u2r4v4s4w3x4x5t4y4s4z4s5k3r5l4s5m4m5n3r5o3x5p4s5q4s5r5y5s4s5t4s5u3x5v2l5w1w5x2l5y2z5z3m6k2l6l2w6m3m6n2w6o3m6p2w6q2l6r3m6s3m6t1w6u1w6v3m6w1w6x4y6y3m6z3m7k3m7l3m7m2l7n2r7o1w7p3m7q3m7r4s7s3m7t3m7u2w7v3k7w1o7x3k7y3q202l3mcl4sal2lam3man3mao3map3mar3mas2lat4wau1vav3maw3say4waz2lbk2sbl3s'fof'6obo2lbp3mbq2xbr1tbs2lbu1zbv3mbz2wck4s202k3mcm4scn4sco4scp4scq5tcr4mcs3xct3xcu3xcv3xcw2l2m2tcy2lcz2ldl4sdm4sdn4sdo4sdp4sdq4sds4sdt4sdu4sdv4sdw4sdz3mek2wel2wem2wen2weo2wep2weq4mer2wes2wet2weu2wev2wew1wex1wey1wez1wfl3mfm3mfn3mfo3mfp3mfq3mfr3sfs3mft3mfu3mfv3mfw3mfz3m203k6o212m6m2dw2l2cq2l3t3m3u1w17s4s19m3m}'kerning'{cl{4qs5ku17sw5ou5qy5rw201ss5tw201ws}201s{201ss}201t{ckw4lwcmwcnwcowcpwclw4wu201ts}2k{201ts}2w{4qs5kw5os5qx5ru17sx5tx}2x{17sw5tw5ou5qu}2y{4qs5kw5os5qx5ru17sx5tx}'fof'-6o7t{ckuclucmucnucoucpu4lu5os5rs}3u{17su5tu5qs}3v{17su5tu5qs}7p{17sw5tw5qs}ck{4qs5ku17sw5ou5qy5rw201ss5tw201ws}4l{4qs5ku17sw5ou5qy5rw201ss5tw201ws}cm{4qs5ku17sw5ou5qy5rw201ss5tw201ws}cn{4qs5ku17sw5ou5qy5rw201ss5tw201ws}co{4qs5ku17sw5ou5qy5rw201ss5tw201ws}cp{4qs5ku17sw5ou5qy5rw201ss5tw201ws}6l{17su5tu5os5qw5rs}17s{2ktclvcmvcnvcovcpv4lv4wuckv}5o{ckwclwcmwcnwcowcpw4lw4wu}5q{ckyclycmycnycoycpy4ly4wu5ms}5r{cktcltcmtcntcotcpt4lt4ws}5t{2ktclvcmvcnvcovcpv4lv4wuckv}7q{cksclscmscnscoscps4ls}6p{17su5tu5qw5rs}ek{5qs5rs}el{17su5tu5os5qw5rs}em{17su5tu5os5qs5rs}en{17su5qs5rs}eo{5qs5rs}ep{17su5tu5os5qw5rs}es{5qs}et{17su5tu5qw5rs}eu{17su5tu5qs5rs}ev{5qs}6z{17sv5tv5os5qx5rs}fm{5os5qt5rs}fn{17sv5tv5os5qx5rs}fo{17sv5tv5os5qx5rs}fp{5os5qt5rs}fq{5os5qt5rs}7r{ckuclucmucnucoucpu4lu5os}fs{17sv5tv5os5qx5rs}ft{17ss5ts5qs}fu{17sw5tw5qs}fv{17sw5tw5qs}fw{17ss5ts5qs}fz{ckuclucmucnucoucpu4lu5os5rs}}}")
|
||
|
, 'Helvetica-Oblique': uncompress("{'widths'{k3p2q4mcx1w201n3r201o6o201s1q201t1q201u1q201w2l201x2l201y2l2k1w2l1w202m2n2n3r2o3r2p5t202q6o2r1n2s2l2t2l2u2r2v3u2w1w2x2l2y1w2z1w3k3r3l3r3m3r3n3r3o3r3p3r3q3r3r3r3s3r203t2l203u2l3v1w3w3u3x3u3y3u3z3r4k6p4l4m4m4m4n4s4o4s4p4m4q3x4r4y4s4s4t1w4u3m4v4m4w3r4x5n4y4s4z4y5k4m5l4y5m4s5n4m5o3x5p4s5q4m5r5y5s4m5t4m5u3x5v1w5w1w5x1w5y2z5z3r6k2l6l3r6m3r6n3m6o3r6p3r6q1w6r3r6s3r6t1q6u1q6v3m6w1q6x5n6y3r6z3r7k3r7l3r7m2l7n3m7o1w7p3r7q3m7r4s7s3m7t3m7u3m7v2l7w1u7x2l7y3u202l3rcl4mal2lam3ran3rao3rap3rar3ras2lat4tau2pav3raw3uay4taz2lbk2sbl3u'fof'6obo2lbp3rbr1wbs2lbu2obv3rbz3xck4m202k3rcm4mcn4mco4mcp4mcq6ocr4scs4mct4mcu4mcv4mcw1w2m2ncy1wcz1wdl4sdm4ydn4ydo4ydp4ydq4yds4ydt4sdu4sdv4sdw4sdz3xek3rel3rem3ren3reo3rep3req5ter3mes3ret3reu3rev3rew1wex1wey1wez1wfl3rfm3rfn3rfo3rfp3rfq3rfr3ufs3xft3rfu3rfv3rfw3rfz3m203k6o212m6o2dw2l2cq2l3t3r3u1w17s4m19m3r}'kerning'{5q{4wv}cl{4qs5kw5ow5qs17sv5tv}201t{2wu4w1k2yu}201x{2wu4wy2yu}17s{2ktclucmucnu4otcpu4lu4wycoucku}2w{7qs4qz5k1m17sy5ow5qx5rsfsu5ty7tufzu}2x{17sy5ty5oy5qs}2y{7qs4qz5k1m17sy5ow5qx5rsfsu5ty7tufzu}'fof'-6o7p{17sv5tv5ow}ck{4qs5kw5ow5qs17sv5tv}4l{4qs5kw5ow5qs17sv5tv}cm{4qs5kw5ow5qs17sv5tv}cn{4qs5kw5ow5qs17sv5tv}co{4qs5kw5ow5qs17sv5tv}cp{4qs5kw5ow5qs17sv5tv}6l{17sy5ty5ow}do{17st5tt}4z{17st5tt}7s{fst}dm{17st5tt}dn{17st5tt}5o{ckwclwcmwcnwcowcpw4lw4wv}dp{17st5tt}dq{17st5tt}7t{5ow}ds{17st5tt}5t{2ktclucmucnu4otcpu4lu4wycoucku}fu{17sv5tv5ow}6p{17sy5ty5ow5qs}ek{17sy5ty5ow}el{17sy5ty5ow}em{17sy5ty5ow}en{5ty}eo{17sy5ty5ow}ep{17sy5ty5ow}es{17sy5ty5qs}et{17sy5ty5ow5qs}eu{17sy5ty5ow5qs}ev{17sy5ty5ow5qs}6z{17sy5ty5ow5qs}fm{17sy5ty5ow5qs}fn{17sy5ty5ow5qs}fo{17sy5ty5ow5qs}fp{17sy5ty5qs}fq{17sy5ty5ow5qs}7r{5ow}fs{17sy5ty5ow5qs}ft{17sv5tv5ow}7m{5ow}fv{17sv5tv5ow}fw{17sv5tv5ow}}}")
|
||
|
}};
|
||
|
|
||
|
/*
|
||
|
This event handler is fired when a new jsPDF object is initialized
|
||
|
This event handler appends metrics data to standard fonts within
|
||
|
that jsPDF instance. The metrics are mapped over Unicode character
|
||
|
codes, NOT CIDs or other codes matching the StandardEncoding table of the
|
||
|
standard PDF fonts.
|
||
|
Future:
|
||
|
Also included is the encoding maping table, converting Unicode (UCS-2, UTF-16)
|
||
|
char codes to StandardEncoding character codes. The encoding table is to be used
|
||
|
somewhere around "pdfEscape" call.
|
||
|
*/
|
||
|
|
||
|
API.events.push([
|
||
|
'addFonts'
|
||
|
,function(fontManagementObjects) {
|
||
|
// fontManagementObjects is {
|
||
|
// 'fonts':font_ID-keyed hash of font objects
|
||
|
// , 'dictionary': lookup object, linking ["FontFamily"]['Style'] to font ID
|
||
|
//}
|
||
|
var font
|
||
|
, fontID
|
||
|
, metrics
|
||
|
, unicode_section
|
||
|
, encoding = 'Unicode'
|
||
|
, encodingBlock
|
||
|
|
||
|
for (fontID in fontManagementObjects.fonts){
|
||
|
if (fontManagementObjects.fonts.hasOwnProperty(fontID)) {
|
||
|
font = fontManagementObjects.fonts[fontID]
|
||
|
|
||
|
// // we only ship 'Unicode' mappings and metrics. No need for loop.
|
||
|
// // still, leaving this for the future.
|
||
|
|
||
|
// for (encoding in fontMetrics){
|
||
|
// if (fontMetrics.hasOwnProperty(encoding)) {
|
||
|
|
||
|
metrics = fontMetrics[encoding][font.PostScriptName]
|
||
|
if (metrics) {
|
||
|
if (font.metadata[encoding]) {
|
||
|
unicode_section = font.metadata[encoding]
|
||
|
} else {
|
||
|
unicode_section = font.metadata[encoding] = {}
|
||
|
}
|
||
|
|
||
|
unicode_section.widths = metrics.widths
|
||
|
unicode_section.kerning = metrics.kerning
|
||
|
}
|
||
|
// }
|
||
|
// }
|
||
|
// for (encoding in encodings){
|
||
|
// if (encodings.hasOwnProperty(encoding)) {
|
||
|
encodingBlock = encodings[encoding][font.PostScriptName]
|
||
|
if (encodingBlock) {
|
||
|
if (font.metadata[encoding]) {
|
||
|
unicode_section = font.metadata[encoding]
|
||
|
} else {
|
||
|
unicode_section = font.metadata[encoding] = {}
|
||
|
}
|
||
|
|
||
|
unicode_section.encoding = encodingBlock
|
||
|
if (encodingBlock.codePages && encodingBlock.codePages.length) {
|
||
|
font.encoding = encodingBlock.codePages[0]
|
||
|
}
|
||
|
}
|
||
|
// }
|
||
|
// }
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
]) // end of adding event handler
|
||
|
|
||
|
})(jsPDF.API);
|
||
|
/** ====================================================================
|
||
|
* jsPDF Cell plugin
|
||
|
* Copyright (c) 2013 Youssef Beddad, youssef.beddad@gmail.com
|
||
|
* 2013 Eduardo Menezes de Morais, eduardo.morais@usp.br
|
||
|
*
|
||
|
* Permission is hereby granted, free of charge, to any person obtaining
|
||
|
* a copy of this software and associated documentation files (the
|
||
|
* "Software"), to deal in the Software without restriction, including
|
||
|
* without limitation the rights to use, copy, modify, merge, publish,
|
||
|
* distribute, sublicense, and/or sell copies of the Software, and to
|
||
|
* permit persons to whom the Software is furnished to do so, subject to
|
||
|
* the following conditions:
|
||
|
*
|
||
|
* The above copyright notice and this permission notice shall be
|
||
|
* included in all copies or substantial portions of the Software.
|
||
|
*
|
||
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
||
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
||
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
||
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
||
|
* LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
||
|
* OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
||
|
* WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
||
|
* ====================================================================
|
||
|
*/
|
||
|
|
||
|
(function (jsPDFAPI) {
|
||
|
'use strict';
|
||
|
/*jslint browser:true */
|
||
|
/*global document: false, jsPDF */
|
||
|
|
||
|
var fontName,
|
||
|
fontSize,
|
||
|
fontStyle,
|
||
|
padding = 3,
|
||
|
margin = 13,
|
||
|
headerFunction,
|
||
|
lastCellPos = { x: undefined, y: undefined, w: undefined, h: undefined, ln: undefined },
|
||
|
pages = 1,
|
||
|
setLastCellPosition = function (x, y, w, h, ln) {
|
||
|
lastCellPos = { 'x': x, 'y': y, 'w': w, 'h': h, 'ln': ln };
|
||
|
},
|
||
|
getLastCellPosition = function () {
|
||
|
return lastCellPos;
|
||
|
};
|
||
|
|
||
|
jsPDFAPI.setHeaderFunction = function (func) {
|
||
|
headerFunction = func;
|
||
|
};
|
||
|
|
||
|
jsPDFAPI.getTextDimensions = function (txt) {
|
||
|
fontName = this.internal.getFont().fontName;
|
||
|
fontSize = this.internal.getFontSize();
|
||
|
fontStyle = this.internal.getFont().fontStyle;
|
||
|
|
||
|
// 1 pixel = 0.264583 mm and 1 mm = 72/25.4 point
|
||
|
var px2pt = 0.264583 * 72 / 25.4,
|
||
|
dimensions,
|
||
|
text;
|
||
|
|
||
|
text = document.createElement('font');
|
||
|
text.id = "jsPDFCell";
|
||
|
text.style.fontStyle = fontStyle;
|
||
|
text.style.fontName = fontName;
|
||
|
text.style.fontSize = fontSize + 'pt';
|
||
|
text.innerText = txt;
|
||
|
|
||
|
document.body.appendChild(text);
|
||
|
|
||
|
dimensions = { w: (text.offsetWidth + 1) * px2pt, h: (text.offsetHeight + 1) * px2pt};
|
||
|
|
||
|
document.body.removeChild(text);
|
||
|
|
||
|
return dimensions;
|
||
|
};
|
||
|
|
||
|
jsPDFAPI.cellAddPage = function () {
|
||
|
this.addPage();
|
||
|
setLastCellPosition(undefined, undefined, undefined, undefined, undefined);
|
||
|
pages += 1;
|
||
|
};
|
||
|
|
||
|
jsPDFAPI.cellInitialize = function () {
|
||
|
lastCellPos = { x: undefined, y: undefined, w: undefined, h: undefined, ln: undefined };
|
||
|
pages = 1;
|
||
|
};
|
||
|
|
||
|
jsPDFAPI.cell = function (x, y, w, h, txt, ln, align) {
|
||
|
var curCell = getLastCellPosition();
|
||
|
|
||
|
// If this is not the first cell, we must change its position
|
||
|
if (curCell.ln !== undefined) {
|
||
|
|
||
|
if (curCell.ln === ln) {
|
||
|
//Same line
|
||
|
x = curCell.x + curCell.w;
|
||
|
y = curCell.y;
|
||
|
} else {
|
||
|
//New line
|
||
|
if ((curCell.y + curCell.h + h + margin) >= this.internal.pageSize.height) {
|
||
|
this.cellAddPage();
|
||
|
|
||
|
if (this.printHeaders && this.tableHeaderRow) {
|
||
|
this.printHeaderRow(ln);
|
||
|
}
|
||
|
}
|
||
|
//We ignore the passed y: the lines may have diferent heights
|
||
|
y = (getLastCellPosition().y + getLastCellPosition().h);
|
||
|
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if (txt[0] !== '') {
|
||
|
if (this.printingHeaderRow) {
|
||
|
this.rect(x, y, w, h, 'FD');
|
||
|
} else {
|
||
|
this.rect(x, y, w, h);
|
||
|
}
|
||
|
if (align === 'right') {
|
||
|
if (txt instanceof Array) {
|
||
|
for(var i = 0; i<txt.length; i++) {
|
||
|
var currentLine = txt[i];
|
||
|
var textSize = this.getStringUnitWidth(currentLine) * this.internal.getFontSize();
|
||
|
this.text(currentLine, x + w - textSize - padding, y + this.internal.getLineHeight()*(i+1));
|
||
|
}
|
||
|
}
|
||
|
} else {
|
||
|
this.text(txt, x + padding, y + this.internal.getLineHeight());
|
||
|
}
|
||
|
}
|
||
|
setLastCellPosition(x, y, w, h, ln);
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
* Return an array containing all of the owned keys of an Object
|
||
|
* @type {Function}
|
||
|
* @return {String[]} of Object keys
|
||
|
*/
|
||
|
jsPDFAPI.getKeys = (typeof Object.keys === 'function')
|
||
|
? function (object) {
|
||
|
if (!object) {
|
||
|
return [];
|
||
|
}
|
||
|
return Object.keys(object);
|
||
|
}
|
||
|
: function (object) {
|
||
|
var keys = [],
|
||
|
property;
|
||
|
|
||
|
for (property in object) {
|
||
|
if (object.hasOwnProperty(property)) {
|
||
|
keys.push(property);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
return keys;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
* Return the maximum value from an array
|
||
|
* @param array
|
||
|
* @param comparisonFn
|
||
|
* @returns {*}
|
||
|
*/
|
||
|
jsPDFAPI.arrayMax = function (array, comparisonFn) {
|
||
|
var max = array[0],
|
||
|
i,
|
||
|
ln,
|
||
|
item;
|
||
|
|
||
|
for (i = 0, ln = array.length; i < ln; i += 1) {
|
||
|
item = array[i];
|
||
|
|
||
|
if (comparisonFn) {
|
||
|
if (comparisonFn(max, item) === -1) {
|
||
|
max = item;
|
||
|
}
|
||
|
} else {
|
||
|
if (item > max) {
|
||
|
max = item;
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
return max;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
* Create a table from a set of data.
|
||
|
* @param {Object[]} data As array of objects containing key-value pairs
|
||
|
* @param {String[]} [headers] Omit or null to auto-generate headers at a performance cost
|
||
|
* @param {Object} [config.printHeaders] True to print column headers at the top of every page
|
||
|
* @param {Object} [config.autoSize] True to dynamically set the column widths to match the widest cell value
|
||
|
* @param {Object} [config.autoStretch] True to force the table to fit the width of the page
|
||
|
*/
|
||
|
jsPDFAPI.table = function (data, headers, config) {
|
||
|
|
||
|
var headerNames = [],
|
||
|
headerPrompts = [],
|
||
|
header,
|
||
|
autoSize,
|
||
|
printHeaders,
|
||
|
autoStretch,
|
||
|
i,
|
||
|
ln,
|
||
|
columnMatrix = {},
|
||
|
columnWidths = {},
|
||
|
columnData,
|
||
|
column,
|
||
|
columnMinWidths = [],
|
||
|
j,
|
||
|
tableHeaderConfigs = [],
|
||
|
model,
|
||
|
jln,
|
||
|
func;
|
||
|
|
||
|
/**
|
||
|
* @property {Number} lnMod
|
||
|
* Keep track of the current line number modifier used when creating cells
|
||
|
*/
|
||
|
this.lnMod = 0;
|
||
|
|
||
|
if (config) {
|
||
|
autoSize = config.autoSize || false;
|
||
|
printHeaders = this.printHeaders = config.printHeaders || true;
|
||
|
autoStretch = config.autoStretch || true;
|
||
|
}
|
||
|
|
||
|
if (!data) {
|
||
|
throw 'No data for PDF table';
|
||
|
}
|
||
|
|
||
|
// Set headers
|
||
|
if (headers === undefined || (headers === null)) {
|
||
|
|
||
|
// No headers defined so we derive from data
|
||
|
headerNames = this.getKeys(data[0]);
|
||
|
|
||
|
} else if (headers[0] && (typeof headers[0] !== 'string')) {
|
||
|
|
||
|
// Split header configs into names and prompts
|
||
|
for (i = 0, ln = headers.length; i < ln; i += 1) {
|
||
|
header = headers[i];
|
||
|
headerNames.push(header.name);
|
||
|
headerPrompts.push(header.prompt);
|
||
|
columnWidths[header.name] = header.width;
|
||
|
}
|
||
|
|
||
|
} else {
|
||
|
headerNames = headers;
|
||
|
}
|
||
|
|
||
|
if (config.autoSize) {
|
||
|
|
||
|
// Create Columns Matrix
|
||
|
|
||
|
func = function (rec) {
|
||
|
return rec[header];
|
||
|
};
|
||
|
|
||
|
for (i = 0, ln = headerNames.length; i < ln; i += 1) {
|
||
|
header = headerNames[i];
|
||
|
|
||
|
columnMatrix[header] = data.map(
|
||
|
func
|
||
|
);
|
||
|
|
||
|
// get header width
|
||
|
columnMinWidths.push(this.getTextDimensions(headerPrompts[i] || header).w);
|
||
|
|
||
|
column = columnMatrix[header];
|
||
|
|
||
|
// get cell widths
|
||
|
for (j = 0, ln = column.length; j < ln; j += 1) {
|
||
|
columnData = column[j];
|
||
|
|
||
|
columnMinWidths.push(this.getTextDimensions(columnData).w);
|
||
|
}
|
||
|
|
||
|
// get final column width
|
||
|
columnWidths[header] = jsPDFAPI.arrayMax(columnMinWidths);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// -- Construct the table
|
||
|
|
||
|
if (config.printHeaders) {
|
||
|
var lineHeight = this.calculateLineHeight(headerNames, columnWidths, headerPrompts.length?headerPrompts:headerNames);
|
||
|
|
||
|
// Construct the header row
|
||
|
for (i = 0, ln = headerNames.length; i < ln; i += 1) {
|
||
|
header = headerNames[i];
|
||
|
tableHeaderConfigs.push([margin, margin, columnWidths[header], lineHeight, String(headerPrompts.length ? headerPrompts[i] : header)]);
|
||
|
}
|
||
|
|
||
|
// Store the table header config
|
||
|
this.setTableHeaderRow(tableHeaderConfigs);
|
||
|
|
||
|
// Print the header for the start of the table
|
||
|
this.printHeaderRow(1);
|
||
|
}
|
||
|
|
||
|
// Construct the data rows
|
||
|
for (i = 0, ln = data.length; i < ln; i += 1) {
|
||
|
var lineHeight;
|
||
|
model = data[i];
|
||
|
lineHeight = this.calculateLineHeight(headerNames, columnWidths, model);
|
||
|
|
||
|
for (j = 0, jln = headerNames.length; j < jln; j += 1) {
|
||
|
header = headerNames[j];
|
||
|
this.cell(margin, margin, columnWidths[header], lineHeight, model[header], i + 2, headers[j].align);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
* Calculate the height for containing the highest column
|
||
|
* @param {String[]} headerNames is the header, used as keys to the data
|
||
|
* @param {Integer[]} columnWidths is size of each column
|
||
|
* @param {Object[]} model is the line of data we want to calculate the height of
|
||
|
*/
|
||
|
jsPDFAPI.calculateLineHeight = function (headerNames, columnWidths, model) {
|
||
|
var header, lineHeight = 0;
|
||
|
for (var j = 0; j < headerNames.length; j++) {
|
||
|
header = headerNames[j];
|
||
|
model[header] = this.splitTextToSize(String(model[header]), columnWidths[header] - padding);
|
||
|
var h = this.internal.getLineHeight() * model[header].length + padding;
|
||
|
if (h > lineHeight)
|
||
|
lineHeight = h;
|
||
|
}
|
||
|
return lineHeight;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
* Store the config for outputting a table header
|
||
|
* @param {Object[]} config
|
||
|
* An array of cell configs that would define a header row: Each config matches the config used by jsPDFAPI.cell
|
||
|
* except the ln parameter is excluded
|
||
|
*/
|
||
|
jsPDFAPI.setTableHeaderRow = function (config) {
|
||
|
this.tableHeaderRow = config;
|
||
|
};
|
||
|
|
||
|
/**
|
||
|
* Output the store header row
|
||
|
* @param lineNumber The line number to output the header at
|
||
|
*/
|
||
|
jsPDFAPI.printHeaderRow = function (lineNumber) {
|
||
|
if (!this.tableHeaderRow) {
|
||
|
throw 'Property tableHeaderRow does not exist.';
|
||
|
}
|
||
|
|
||
|
var tableHeaderCell,
|
||
|
tmpArray,
|
||
|
i,
|
||
|
ln;
|
||
|
|
||
|
this.printingHeaderRow = true;
|
||
|
if (headerFunction !== undefined) {
|
||
|
var position = headerFunction(this, pages);
|
||
|
setLastCellPosition(position[0], position[1], position[2], position[3], -1);
|
||
|
}
|
||
|
|
||
|
this.setFontStyle('bold');
|
||
|
for (i = 0, ln = this.tableHeaderRow.length; i < ln; i += 1) {
|
||
|
this.setFillColor(200,200,200);
|
||
|
|
||
|
tableHeaderCell = this.tableHeaderRow[i];
|
||
|
tmpArray = [].concat(tableHeaderCell);
|
||
|
|
||
|
this.cell.apply(this, tmpArray.concat(lineNumber));
|
||
|
}
|
||
|
this.setFontStyle('normal');
|
||
|
this.printingHeaderRow = false;
|
||
|
};
|
||
|
|
||
|
})(jsPDF.API);
|
||
|
/** ====================================================================
|
||
|
* jsPDF total_pages plugin
|
||
|
* Copyright (c) 2013 Eduardo Menezes de Morais, eduardo.morais@usp.br
|
||
|
*
|
||
|
* Permission is hereby granted, free of charge, to any person obtaining
|
||
|
* a copy of this software and associated documentation files (the
|
||
|
* "Software"), to deal in the Software without restriction, including
|
||
|
* without limitation the rights to use, copy, modify, merge, publish,
|
||
|
* distribute, sublicense, and/or sell copies of the Software, and to
|
||
|
* permit persons to whom the Software is furnished to do so, subject to
|
||
|
* the following conditions:
|
||
|
*
|
||
|
* The above copyright notice and this permission notice shall be
|
||
|
* included in all copies or substantial portions of the Software.
|
||
|
*
|
||
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
||
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
||
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
||
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
||
|
* LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
||
|
* OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
||
|
* WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
||
|
* ====================================================================
|
||
|
*/
|
||
|
|
||
|
(function(jsPDFAPI) {
|
||
|
'use strict';
|
||
|
|
||
|
jsPDFAPI.putTotalPages = function(pageExpression) {
|
||
|
'use strict';
|
||
|
var replaceExpression = new RegExp(pageExpression, 'g');
|
||
|
for (var n = 1; n <= this.internal.getNumberOfPages(); n++) {
|
||
|
for (var i = 0; i < this.internal.pages[n].length; i++)
|
||
|
this.internal.pages[n][i] = this.internal.pages[n][i].replace(replaceExpression, this.internal.getNumberOfPages());
|
||
|
}
|
||
|
return this;
|
||
|
};
|
||
|
|
||
|
})(jsPDF.API);
|
||
|
/* BlobBuilder.js
|
||
|
* A BlobBuilder implementation.
|
||
|
* 2012-04-21
|
||
|
*
|
||
|
* By Eli Grey, http://eligrey.com
|
||
|
* License: X11/MIT
|
||
|
* See LICENSE.md
|
||
|
*/
|
||
|
|
||
|
/*global self, unescape */
|
||
|
/*jslint bitwise: true, regexp: true, confusion: true, es5: true, vars: true, white: true,
|
||
|
plusplus: true */
|
||
|
|
||
|
/*! @source http://purl.eligrey.com/github/BlobBuilder.js/blob/master/BlobBuilder.js */
|
||
|
|
||
|
var BlobBuilder = BlobBuilder || self.WebKitBlobBuilder || self.MozBlobBuilder || self.MSBlobBuilder || (function(view) {
|
||
|
"use strict";
|
||
|
var
|
||
|
get_class = function(object) {
|
||
|
return Object.prototype.toString.call(object).match(/^\[object\s(.*)\]$/)[1];
|
||
|
}
|
||
|
, FakeBlobBuilder = function(){
|
||
|
this.data = [];
|
||
|
}
|
||
|
, FakeBlob = function(data, type, encoding) {
|
||
|
this.data = data;
|
||
|
this.size = data.length;
|
||
|
this.type = type;
|
||
|
this.encoding = encoding;
|
||
|
}
|
||
|
, FBB_proto = FakeBlobBuilder.prototype
|
||
|
, FB_proto = FakeBlob.prototype
|
||
|
, FileReaderSync = view.FileReaderSync
|
||
|
, FileException = function(type) {
|
||
|
this.code = this[this.name = type];
|
||
|
}
|
||
|
, file_ex_codes = (
|
||
|
"NOT_FOUND_ERR SECURITY_ERR ABORT_ERR NOT_READABLE_ERR ENCODING_ERR "
|
||
|
+ "NO_MODIFICATION_ALLOWED_ERR INVALID_STATE_ERR SYNTAX_ERR"
|
||
|
).split(" ")
|
||
|
, file_ex_code = file_ex_codes.length
|
||
|
, real_URL = view.URL || view.webkitURL || view
|
||
|
, real_create_object_URL = real_URL.createObjectURL
|
||
|
, real_revoke_object_URL = real_URL.revokeObjectURL
|
||
|
, URL = real_URL
|
||
|
, btoa = view.btoa
|
||
|
, atob = view.atob
|
||
|
, can_apply_typed_arrays = false
|
||
|
, can_apply_typed_arrays_test = function(pass) {
|
||
|
can_apply_typed_arrays = !pass;
|
||
|
}
|
||
|
|
||
|
, ArrayBuffer = view.ArrayBuffer
|
||
|
, Uint8Array = view.Uint8Array
|
||
|
;
|
||
|
FakeBlobBuilder.fake = FB_proto.fake = true;
|
||
|
while (file_ex_code--) {
|
||
|
FileException.prototype[file_ex_codes[file_ex_code]] = file_ex_code + 1;
|
||
|
}
|
||
|
try {
|
||
|
if (Uint8Array) {
|
||
|
can_apply_typed_arrays_test.apply(0, new Uint8Array(1));
|
||
|
}
|
||
|
} catch (ex) {}
|
||
|
if (!real_URL.createObjectURL) {
|
||
|
URL = view.URL = {};
|
||
|
}
|
||
|
URL.createObjectURL = function(blob) {
|
||
|
var
|
||
|
type = blob.type
|
||
|
, data_URI_header
|
||
|
;
|
||
|
if (type === null) {
|
||
|
type = "application/octet-stream";
|
||
|
}
|
||
|
if (blob instanceof FakeBlob) {
|
||
|
data_URI_header = "data:" + type;
|
||
|
if (blob.encoding === "base64") {
|
||
|
return data_URI_header + ";base64," + blob.data;
|
||
|
} else if (blob.encoding === "URI") {
|
||
|
return data_URI_header + "," + decodeURIComponent(blob.data);
|
||
|
} if (btoa) {
|
||
|
return data_URI_header + ";base64," + btoa(blob.data);
|
||
|
} else {
|
||
|
return data_URI_header + "," + encodeURIComponent(blob.data);
|
||
|
}
|
||
|
} else if (real_create_object_URL) {
|
||
|
return real_create_object_URL.call(real_URL, blob);
|
||
|
}
|
||
|
};
|
||
|
URL.revokeObjectURL = function(object_URL) {
|
||
|
if (object_URL.substring(0, 5) !== "data:" && real_revoke_object_URL) {
|
||
|
real_revoke_object_URL.call(real_URL, object_URL);
|
||
|
}
|
||
|
};
|
||
|
FBB_proto.append = function(data/*, endings*/) {
|
||
|
var bb = this.data;
|
||
|
// decode data to a binary string
|
||
|
if (Uint8Array && data instanceof ArrayBuffer) {
|
||
|
if (can_apply_typed_arrays) {
|
||
|
bb.push(String.fromCharCode.apply(String, new Uint8Array(data)));
|
||
|
} else {
|
||
|
var
|
||
|
str = ""
|
||
|
, buf = new Uint8Array(data)
|
||
|
, i = 0
|
||
|
, buf_len = buf.length
|
||
|
;
|
||
|
for (; i < buf_len; i++) {
|
||
|
str += String.fromCharCode(buf[i]);
|
||
|
}
|
||
|
}
|
||
|
} else if (get_class(data) === "Blob" || get_class(data) === "File") {
|
||
|
if (FileReaderSync) {
|
||
|
var fr = new FileReaderSync;
|
||
|
bb.push(fr.readAsBinaryString(data));
|
||
|
} else {
|
||
|
// async FileReader won't work as BlobBuilder is sync
|
||
|
throw new FileException("NOT_READABLE_ERR");
|
||
|
}
|
||
|
} else if (data instanceof FakeBlob) {
|
||
|
if (data.encoding === "base64" && atob) {
|
||
|
bb.push(atob(data.data));
|
||
|
} else if (data.encoding === "URI") {
|
||
|
bb.push(decodeURIComponent(data.data));
|
||
|
} else if (data.encoding === "raw") {
|
||
|
bb.push(data.data);
|
||
|
}
|
||
|
} else {
|
||
|
if (typeof data !== "string") {
|
||
|
data += ""; // convert unsupported types to strings
|
||
|
}
|
||
|
// decode UTF-16 to binary string
|
||
|
bb.push(unescape(encodeURIComponent(data)));
|
||
|
}
|
||
|
};
|
||
|
FBB_proto.getBlob = function(type) {
|
||
|
if (!arguments.length) {
|
||
|
type = null;
|
||
|
}
|
||
|
return new FakeBlob(this.data.join(""), type, "raw");
|
||
|
};
|
||
|
FBB_proto.toString = function() {
|
||
|
return "[object BlobBuilder]";
|
||
|
};
|
||
|
FB_proto.slice = function(start, end, type) {
|
||
|
var args = arguments.length;
|
||
|
if (args < 3) {
|
||
|
type = null;
|
||
|
}
|
||
|
return new FakeBlob(
|
||
|
this.data.slice(start, args > 1 ? end : this.data.length)
|
||
|
, type
|
||
|
, this.encoding
|
||
|
);
|
||
|
};
|
||
|
FB_proto.toString = function() {
|
||
|
return "[object Blob]";
|
||
|
};
|
||
|
return FakeBlobBuilder;
|
||
|
}(self));
|
||
|
/* FileSaver.js
|
||
|
* A saveAs() FileSaver implementation.
|
||
|
* 2013-01-23
|
||
|
*
|
||
|
* By Eli Grey, http://eligrey.com
|
||
|
* License: X11/MIT
|
||
|
* See LICENSE.md
|
||
|
*/
|
||
|
|
||
|
/*global self */
|
||
|
/*jslint bitwise: true, regexp: true, confusion: true, es5: true, vars: true, white: true,
|
||
|
plusplus: true */
|
||
|
|
||
|
/*! @source http://purl.eligrey.com/github/FileSaver.js/blob/master/FileSaver.js */
|
||
|
|
||
|
var saveAs = saveAs
|
||
|
|| (navigator.msSaveBlob && navigator.msSaveBlob.bind(navigator))
|
||
|
|| (function(view) {
|
||
|
"use strict";
|
||
|
var
|
||
|
doc = view.document
|
||
|
// only get URL when necessary in case BlobBuilder.js hasn't overridden it yet
|
||
|
, get_URL = function() {
|
||
|
return view.URL || view.webkitURL || view;
|
||
|
}
|
||
|
, URL = view.URL || view.webkitURL || view
|
||
|
, save_link = doc.createElementNS("http://www.w3.org/1999/xhtml", "a")
|
||
|
, can_use_save_link = "download" in save_link
|
||
|
, click = function(node) {
|
||
|
var event = doc.createEvent("MouseEvents");
|
||
|
event.initMouseEvent(
|
||
|
"click", true, false, view, 0, 0, 0, 0, 0
|
||
|
, false, false, false, false, 0, null
|
||
|
);
|
||
|
return node.dispatchEvent(event); // false if event was cancelled
|
||
|
}
|
||
|
, webkit_req_fs = view.webkitRequestFileSystem
|
||
|
, req_fs = view.requestFileSystem || webkit_req_fs || view.mozRequestFileSystem
|
||
|
, throw_outside = function (ex) {
|
||
|
(view.setImmediate || view.setTimeout)(function() {
|
||
|
throw ex;
|
||
|
}, 0);
|
||
|
}
|
||
|
, force_saveable_type = "application/octet-stream"
|
||
|
, fs_min_size = 0
|
||
|
, deletion_queue = []
|
||
|
, process_deletion_queue = function() {
|
||
|
var i = deletion_queue.length;
|
||
|
while (i--) {
|
||
|
var file = deletion_queue[i];
|
||
|
if (typeof file === "string") { // file is an object URL
|
||
|
URL.revokeObjectURL(file);
|
||
|
} else { // file is a File
|
||
|
file.remove();
|
||
|
}
|
||
|
}
|
||
|
deletion_queue.length = 0; // clear queue
|
||
|
}
|
||
|
, dispatch = function(filesaver, event_types, event) {
|
||
|
event_types = [].concat(event_types);
|
||
|
var i = event_types.length;
|
||
|
while (i--) {
|
||
|
var listener = filesaver["on" + event_types[i]];
|
||
|
if (typeof listener === "function") {
|
||
|
try {
|
||
|
listener.call(filesaver, event || filesaver);
|
||
|
} catch (ex) {
|
||
|
throw_outside(ex);
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
, FileSaver = function(blob, name) {
|
||
|
// First try a.download, then web filesystem, then object URLs
|
||
|
var
|
||
|
filesaver = this
|
||
|
, type = blob.type
|
||
|
, blob_changed = false
|
||
|
, object_url
|
||
|
, target_view
|
||
|
, get_object_url = function() {
|
||
|
var object_url = get_URL().createObjectURL(blob);
|
||
|
deletion_queue.push(object_url);
|
||
|
return object_url;
|
||
|
}
|
||
|
, dispatch_all = function() {
|
||
|
dispatch(filesaver, "writestart progress write writeend".split(" "));
|
||
|
}
|
||
|
// on any filesys errors revert to saving with object URLs
|
||
|
, fs_error = function() {
|
||
|
// don't create more object URLs than needed
|
||
|
if (blob_changed || !object_url) {
|
||
|
object_url = get_object_url(blob);
|
||
|
}
|
||
|
if (target_view) {
|
||
|
target_view.location.href = object_url;
|
||
|
}
|
||
|
filesaver.readyState = filesaver.DONE;
|
||
|
dispatch_all();
|
||
|
}
|
||
|
, abortable = function(func) {
|
||
|
return function() {
|
||
|
if (filesaver.readyState !== filesaver.DONE) {
|
||
|
return func.apply(this, arguments);
|
||
|
}
|
||
|
};
|
||
|
}
|
||
|
, create_if_not_found = {create: true, exclusive: false}
|
||
|
, slice
|
||
|
;
|
||
|
filesaver.readyState = filesaver.INIT;
|
||
|
if (!name) {
|
||
|
name = "download";
|
||
|
}
|
||
|
if (can_use_save_link) {
|
||
|
object_url = get_object_url(blob);
|
||
|
save_link.href = object_url;
|
||
|
save_link.download = name;
|
||
|
if (click(save_link)) {
|
||
|
filesaver.readyState = filesaver.DONE;
|
||
|
dispatch_all();
|
||
|
return;
|
||
|
}
|
||
|
}
|
||
|
// Object and web filesystem URLs have a problem saving in Google Chrome when
|
||
|
// viewed in a tab, so I force save with application/octet-stream
|
||
|
// http://code.google.com/p/chromium/issues/detail?id=91158
|
||
|
if (view.chrome && type && type !== force_saveable_type) {
|
||
|
slice = blob.slice || blob.webkitSlice;
|
||
|
blob = slice.call(blob, 0, blob.size, force_saveable_type);
|
||
|
blob_changed = true;
|
||
|
}
|
||
|
// Since I can't be sure that the guessed media type will trigger a download
|
||
|
// in WebKit, I append .download to the filename.
|
||
|
// https://bugs.webkit.org/show_bug.cgi?id=65440
|
||
|
if (webkit_req_fs && name !== "download") {
|
||
|
name += ".download";
|
||
|
}
|
||
|
if (type === force_saveable_type || webkit_req_fs) {
|
||
|
target_view = view;
|
||
|
} else {
|
||
|
target_view = view.open();
|
||
|
}
|
||
|
if (!req_fs) {
|
||
|
fs_error();
|
||
|
return;
|
||
|
}
|
||
|
fs_min_size += blob.size;
|
||
|
req_fs(view.TEMPORARY, fs_min_size, abortable(function(fs) {
|
||
|
fs.root.getDirectory("saved", create_if_not_found, abortable(function(dir) {
|
||
|
var save = function() {
|
||
|
dir.getFile(name, create_if_not_found, abortable(function(file) {
|
||
|
file.createWriter(abortable(function(writer) {
|
||
|
writer.onwriteend = function(event) {
|
||
|
target_view.location.href = file.toURL();
|
||
|
deletion_queue.push(file);
|
||
|
filesaver.readyState = filesaver.DONE;
|
||
|
dispatch(filesaver, "writeend", event);
|
||
|
};
|
||
|
writer.onerror = function() {
|
||
|
var error = writer.error;
|
||
|
if (error.code !== error.ABORT_ERR) {
|
||
|
fs_error();
|
||
|
}
|
||
|
};
|
||
|
"writestart progress write abort".split(" ").forEach(function(event) {
|
||
|
writer["on" + event] = filesaver["on" + event];
|
||
|
});
|
||
|
writer.write(blob);
|
||
|
filesaver.abort = function() {
|
||
|
writer.abort();
|
||
|
filesaver.readyState = filesaver.DONE;
|
||
|
};
|
||
|
filesaver.readyState = filesaver.WRITING;
|
||
|
}), fs_error);
|
||
|
}), fs_error);
|
||
|
};
|
||
|
dir.getFile(name, {create: false}, abortable(function(file) {
|
||
|
// delete file if it already exists
|
||
|
file.remove();
|
||
|
save();
|
||
|
}), abortable(function(ex) {
|
||
|
if (ex.code === ex.NOT_FOUND_ERR) {
|
||
|
save();
|
||
|
} else {
|
||
|
fs_error();
|
||
|
}
|
||
|
}));
|
||
|
}), fs_error);
|
||
|
}), fs_error);
|
||
|
}
|
||
|
, FS_proto = FileSaver.prototype
|
||
|
, saveAs = function(blob, name) {
|
||
|
return new FileSaver(blob, name);
|
||
|
}
|
||
|
;
|
||
|
FS_proto.abort = function() {
|
||
|
var filesaver = this;
|
||
|
filesaver.readyState = filesaver.DONE;
|
||
|
dispatch(filesaver, "abort");
|
||
|
};
|
||
|
FS_proto.readyState = FS_proto.INIT = 0;
|
||
|
FS_proto.WRITING = 1;
|
||
|
FS_proto.DONE = 2;
|
||
|
|
||
|
FS_proto.error =
|
||
|
FS_proto.onwritestart =
|
||
|
FS_proto.onprogress =
|
||
|
FS_proto.onwrite =
|
||
|
FS_proto.onabort =
|
||
|
FS_proto.onerror =
|
||
|
FS_proto.onwriteend =
|
||
|
null;
|
||
|
|
||
|
view.addEventListener("unload", process_deletion_queue, false);
|
||
|
return saveAs;
|
||
|
}(self));
|
||
|
/*
|
||
|
Copyright (c) 2013 Gildas Lormeau. All rights reserved.
|
||
|
|
||
|
Redistribution and use in source and binary forms, with or without
|
||
|
modification, are permitted provided that the following conditions are met:
|
||
|
|
||
|
1. Redistributions of source code must retain the above copyright notice,
|
||
|
this list of conditions and the following disclaimer.
|
||
|
|
||
|
2. Redistributions in binary form must reproduce the above copyright
|
||
|
notice, this list of conditions and the following disclaimer in
|
||
|
the documentation and/or other materials provided with the distribution.
|
||
|
|
||
|
3. The names of the authors may not be used to endorse or promote products
|
||
|
derived from this software without specific prior written permission.
|
||
|
|
||
|
THIS SOFTWARE IS PROVIDED ``AS IS'' AND ANY EXPRESSED OR IMPLIED WARRANTIES,
|
||
|
INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND
|
||
|
FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL JCRAFT,
|
||
|
INC. OR ANY CONTRIBUTORS TO THIS SOFTWARE BE LIABLE FOR ANY DIRECT, INDIRECT,
|
||
|
INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA,
|
||
|
OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
|
||
|
LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
|
||
|
NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE,
|
||
|
EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||
|
*/
|
||
|
|
||
|
/*
|
||
|
* This program is based on JZlib 1.0.2 ymnk, JCraft,Inc.
|
||
|
* JZlib is based on zlib-1.1.3, so all credit should go authors
|
||
|
* Jean-loup Gailly(jloup@gzip.org) and Mark Adler(madler@alumni.caltech.edu)
|
||
|
* and contributors of zlib.
|
||
|
*/
|
||
|
|
||
|
// Global
|
||
|
|
||
|
var MAX_BITS = 15;
|
||
|
var D_CODES = 30;
|
||
|
var BL_CODES = 19;
|
||
|
|
||
|
var LENGTH_CODES = 29;
|
||
|
var LITERALS = 256;
|
||
|
var L_CODES = (LITERALS + 1 + LENGTH_CODES);
|
||
|
var HEAP_SIZE = (2 * L_CODES + 1);
|
||
|
|
||
|
var END_BLOCK = 256;
|
||
|
|
||
|
// Bit length codes must not exceed MAX_BL_BITS bits
|
||
|
var MAX_BL_BITS = 7;
|
||
|
|
||
|
// repeat previous bit length 3-6 times (2 bits of repeat count)
|
||
|
var REP_3_6 = 16;
|
||
|
|
||
|
// repeat a zero length 3-10 times (3 bits of repeat count)
|
||
|
var REPZ_3_10 = 17;
|
||
|
|
||
|
// repeat a zero length 11-138 times (7 bits of repeat count)
|
||
|
var REPZ_11_138 = 18;
|
||
|
|
||
|
// The lengths of the bit length codes are sent in order of decreasing
|
||
|
// probability, to avoid transmitting the lengths for unused bit
|
||
|
// length codes.
|
||
|
|
||
|
var Buf_size = 8 * 2;
|
||
|
|
||
|
// JZlib version : "1.0.2"
|
||
|
var Z_DEFAULT_COMPRESSION = -1;
|
||
|
|
||
|
// compression strategy
|
||
|
var Z_FILTERED = 1;
|
||
|
var Z_HUFFMAN_ONLY = 2;
|
||
|
var Z_DEFAULT_STRATEGY = 0;
|
||
|
|
||
|
var Z_NO_FLUSH = 0;
|
||
|
var Z_PARTIAL_FLUSH = 1;
|
||
|
var Z_FULL_FLUSH = 3;
|
||
|
var Z_FINISH = 4;
|
||
|
|
||
|
var Z_OK = 0;
|
||
|
var Z_STREAM_END = 1;
|
||
|
var Z_NEED_DICT = 2;
|
||
|
var Z_STREAM_ERROR = -2;
|
||
|
var Z_DATA_ERROR = -3;
|
||
|
var Z_BUF_ERROR = -5;
|
||
|
|
||
|
// Tree
|
||
|
|
||
|
// see definition of array dist_code below
|
||
|
var _dist_code = [ 0, 1, 2, 3, 4, 4, 5, 5, 6, 6, 6, 6, 7, 7, 7, 7, 8, 8, 8, 8, 8, 8, 8, 8, 9, 9, 9, 9, 9, 9, 9, 9, 10, 10, 10, 10, 10, 10, 10, 10, 10, 10,
|
||
|
10, 10, 10, 10, 10, 10, 11, 11, 11, 11, 11, 11, 11, 11, 11, 11, 11, 11, 11, 11, 11, 11, 12, 12, 12, 12, 12, 12, 12, 12, 12, 12, 12, 12, 12, 12, 12,
|
||
|
12, 12, 12, 12, 12, 12, 12, 12, 12, 12, 12, 12, 12, 12, 12, 12, 12, 13, 13, 13, 13, 13, 13, 13, 13, 13, 13, 13, 13, 13, 13, 13, 13, 13, 13, 13, 13,
|
||
|
13, 13, 13, 13, 13, 13, 13, 13, 13, 13, 13, 13, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14,
|
||
|
14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14, 14,
|
||
|
14, 14, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15,
|
||
|
15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 15, 0, 0, 16, 17, 18, 18, 19, 19,
|
||
|
20, 20, 20, 20, 21, 21, 21, 21, 22, 22, 22, 22, 22, 22, 22, 22, 23, 23, 23, 23, 23, 23, 23, 23, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24,
|
||
|
24, 24, 24, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26,
|
||
|
26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27,
|
||
|
27, 27, 27, 27, 27, 27, 27, 27, 27, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28,
|
||
|
28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 28, 29,
|
||
|
29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29,
|
||
|
29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29, 29 ];
|
||
|
|
||
|
function Tree() {
|
||
|
'use strict';
|
||
|
var that = this;
|
||
|
|
||
|
// dyn_tree; // the dynamic tree
|
||
|
// max_code; // largest code with non zero frequency
|
||
|
// stat_desc; // the corresponding static tree
|
||
|
|
||
|
// Compute the optimal bit lengths for a tree and update the total bit
|
||
|
// length
|
||
|
// for the current block.
|
||
|
// IN assertion: the fields freq and dad are set, heap[heap_max] and
|
||
|
// above are the tree nodes sorted by increasing frequency.
|
||
|
// OUT assertions: the field len is set to the optimal bit length, the
|
||
|
// array bl_count contains the frequencies for each bit length.
|
||
|
// The length opt_len is updated; static_len is also updated if stree is
|
||
|
// not null.
|
||
|
function gen_bitlen(s) {
|
||
|
var tree = that.dyn_tree;
|
||
|
var stree = that.stat_desc.static_tree;
|
||
|
var extra = that.stat_desc.extra_bits;
|
||
|
var base = that.stat_desc.extra_base;
|
||
|
var max_length = that.stat_desc.max_length;
|
||
|
var h; // heap index
|
||
|
var n, m; // iterate over the tree elements
|
||
|
var bits; // bit length
|
||
|
var xbits; // extra bits
|
||
|
var f; // frequency
|
||
|
var overflow = 0; // number of elements with bit length too large
|
||
|
|
||
|
for (bits = 0; bits <= MAX_BITS; bits++)
|
||
|
s.bl_count[bits] = 0;
|
||
|
|
||
|
// In a first pass, compute the optimal bit lengths (which may
|
||
|
// overflow in the case of the bit length tree).
|
||
|
tree[s.heap[s.heap_max] * 2 + 1] = 0; // root of the heap
|
||
|
|
||
|
for (h = s.heap_max + 1; h < HEAP_SIZE; h++) {
|
||
|
n = s.heap[h];
|
||
|
bits = tree[tree[n * 2 + 1] * 2 + 1] + 1;
|
||
|
if (bits > max_length) {
|
||
|
bits = max_length;
|
||
|
overflow++;
|
||
|
}
|
||
|
tree[n * 2 + 1] = bits;
|
||
|
// We overwrite tree[n*2+1] which is no longer needed
|
||
|
|
||
|
if (n > that.max_code)
|
||
|
continue; // not a leaf node
|
||
|
|
||
|
s.bl_count[bits]++;
|
||
|
xbits = 0;
|
||
|
if (n >= base)
|
||
|
xbits = extra[n - base];
|
||
|
f = tree[n * 2];
|
||
|
s.opt_len += f * (bits + xbits);
|
||
|
if (stree)
|
||
|
s.static_len += f * (stree[n * 2 + 1] + xbits);
|
||
|
}
|
||
|
if (overflow === 0)
|
||
|
return;
|
||
|
|
||
|
// This happens for example on obj2 and pic of the Calgary corpus
|
||
|
// Find the first bit length which could increase:
|
||
|
do {
|
||
|
bits = max_length - 1;
|
||
|
while (s.bl_count[bits] === 0)
|
||
|
bits--;
|
||
|
s.bl_count[bits]--; // move one leaf down the tree
|
||
|
s.bl_count[bits + 1] += 2; // move one overflow item as its brother
|
||
|
s.bl_count[max_length]--;
|
||
|
// The brother of the overflow item also moves one step up,
|
||
|
// but this does not affect bl_count[max_length]
|
||
|
overflow -= 2;
|
||
|
} while (overflow > 0);
|
||
|
|
||
|
for (bits = max_length; bits !== 0; bits--) {
|
||
|
n = s.bl_count[bits];
|
||
|
while (n !== 0) {
|
||
|
m = s.heap[--h];
|
||
|
if (m > that.max_code)
|
||
|
continue;
|
||
|
if (tree[m * 2 + 1] != bits) {
|
||
|
s.opt_len += (bits - tree[m * 2 + 1]) * tree[m * 2];
|
||
|
tree[m * 2 + 1] = bits;
|
||
|
}
|
||
|
n--;
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Reverse the first len bits of a code, using straightforward code (a
|
||
|
// faster
|
||
|
// method would use a table)
|
||
|
// IN assertion: 1 <= len <= 15
|
||
|
function bi_reverse(code, // the value to invert
|
||
|
len // its bit length
|
||
|
) {
|
||
|
var res = 0;
|
||
|
do {
|
||
|
res |= code & 1;
|
||
|
code >>>= 1;
|
||
|
res <<= 1;
|
||
|
} while (--len > 0);
|
||
|
return res >>> 1;
|
||
|
}
|
||
|
|
||
|
// Generate the codes for a given tree and bit counts (which need not be
|
||
|
// optimal).
|
||
|
// IN assertion: the array bl_count contains the bit length statistics for
|
||
|
// the given tree and the field len is set for all tree elements.
|
||
|
// OUT assertion: the field code is set for all tree elements of non
|
||
|
// zero code length.
|
||
|
function gen_codes(tree, // the tree to decorate
|
||
|
max_code, // largest code with non zero frequency
|
||
|
bl_count // number of codes at each bit length
|
||
|
) {
|
||
|
var next_code = []; // next code value for each
|
||
|
// bit length
|
||
|
var code = 0; // running code value
|
||
|
var bits; // bit index
|
||
|
var n; // code index
|
||
|
var len;
|
||
|
|
||
|
// The distribution counts are first used to generate the code values
|
||
|
// without bit reversal.
|
||
|
for (bits = 1; bits <= MAX_BITS; bits++) {
|
||
|
next_code[bits] = code = ((code + bl_count[bits - 1]) << 1);
|
||
|
}
|
||
|
|
||
|
// Check that the bit counts in bl_count are consistent. The last code
|
||
|
// must be all ones.
|
||
|
// Assert (code + bl_count[MAX_BITS]-1 == (1<<MAX_BITS)-1,
|
||
|
// "inconsistent bit counts");
|
||
|
// Tracev((stderr,"\ngen_codes: max_code %d ", max_code));
|
||
|
|
||
|
for (n = 0; n <= max_code; n++) {
|
||
|
len = tree[n * 2 + 1];
|
||
|
if (len === 0)
|
||
|
continue;
|
||
|
// Now reverse the bits
|
||
|
tree[n * 2] = bi_reverse(next_code[len]++, len);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Construct one Huffman tree and assigns the code bit strings and lengths.
|
||
|
// Update the total bit length for the current block.
|
||
|
// IN assertion: the field freq is set for all tree elements.
|
||
|
// OUT assertions: the fields len and code are set to the optimal bit length
|
||
|
// and corresponding code. The length opt_len is updated; static_len is
|
||
|
// also updated if stree is not null. The field max_code is set.
|
||
|
that.build_tree = function(s) {
|
||
|
var tree = that.dyn_tree;
|
||
|
var stree = that.stat_desc.static_tree;
|
||
|
var elems = that.stat_desc.elems;
|
||
|
var n, m; // iterate over heap elements
|
||
|
var max_code = -1; // largest code with non zero frequency
|
||
|
var node; // new node being created
|
||
|
|
||
|
// Construct the initial heap, with least frequent element in
|
||
|
// heap[1]. The sons of heap[n] are heap[2*n] and heap[2*n+1].
|
||
|
// heap[0] is not used.
|
||
|
s.heap_len = 0;
|
||
|
s.heap_max = HEAP_SIZE;
|
||
|
|
||
|
for (n = 0; n < elems; n++) {
|
||
|
if (tree[n * 2] !== 0) {
|
||
|
s.heap[++s.heap_len] = max_code = n;
|
||
|
s.depth[n] = 0;
|
||
|
} else {
|
||
|
tree[n * 2 + 1] = 0;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// The pkzip format requires that at least one distance code exists,
|
||
|
// and that at least one bit should be sent even if there is only one
|
||
|
// possible code. So to avoid special checks later on we force at least
|
||
|
// two codes of non zero frequency.
|
||
|
while (s.heap_len < 2) {
|
||
|
node = s.heap[++s.heap_len] = max_code < 2 ? ++max_code : 0;
|
||
|
tree[node * 2] = 1;
|
||
|
s.depth[node] = 0;
|
||
|
s.opt_len--;
|
||
|
if (stree)
|
||
|
s.static_len -= stree[node * 2 + 1];
|
||
|
// node is 0 or 1 so it does not have extra bits
|
||
|
}
|
||
|
that.max_code = max_code;
|
||
|
|
||
|
// The elements heap[heap_len/2+1 .. heap_len] are leaves of the tree,
|
||
|
// establish sub-heaps of increasing lengths:
|
||
|
|
||
|
for (n = Math.floor(s.heap_len / 2); n >= 1; n--)
|
||
|
s.pqdownheap(tree, n);
|
||
|
|
||
|
// Construct the Huffman tree by repeatedly combining the least two
|
||
|
// frequent nodes.
|
||
|
|
||
|
node = elems; // next internal node of the tree
|
||
|
do {
|
||
|
// n = node of least frequency
|
||
|
n = s.heap[1];
|
||
|
s.heap[1] = s.heap[s.heap_len--];
|
||
|
s.pqdownheap(tree, 1);
|
||
|
m = s.heap[1]; // m = node of next least frequency
|
||
|
|
||
|
s.heap[--s.heap_max] = n; // keep the nodes sorted by frequency
|
||
|
s.heap[--s.heap_max] = m;
|
||
|
|
||
|
// Create a new node father of n and m
|
||
|
tree[node * 2] = (tree[n * 2] + tree[m * 2]);
|
||
|
s.depth[node] = Math.max(s.depth[n], s.depth[m]) + 1;
|
||
|
tree[n * 2 + 1] = tree[m * 2 + 1] = node;
|
||
|
|
||
|
// and insert the new node in the heap
|
||
|
s.heap[1] = node++;
|
||
|
s.pqdownheap(tree, 1);
|
||
|
} while (s.heap_len >= 2);
|
||
|
|
||
|
s.heap[--s.heap_max] = s.heap[1];
|
||
|
|
||
|
// At this point, the fields freq and dad are set. We can now
|
||
|
// generate the bit lengths.
|
||
|
|
||
|
gen_bitlen(s);
|
||
|
|
||
|
// The field len is now set, we can generate the bit codes
|
||
|
gen_codes(tree, that.max_code, s.bl_count);
|
||
|
};
|
||
|
|
||
|
}
|
||
|
|
||
|
Tree._length_code = [ 0, 1, 2, 3, 4, 5, 6, 7, 8, 8, 9, 9, 10, 10, 11, 11, 12, 12, 12, 12, 13, 13, 13, 13, 14, 14, 14, 14, 15, 15, 15, 15, 16, 16, 16, 16,
|
||
|
16, 16, 16, 16, 17, 17, 17, 17, 17, 17, 17, 17, 18, 18, 18, 18, 18, 18, 18, 18, 19, 19, 19, 19, 19, 19, 19, 19, 20, 20, 20, 20, 20, 20, 20, 20, 20,
|
||
|
20, 20, 20, 20, 20, 20, 20, 21, 21, 21, 21, 21, 21, 21, 21, 21, 21, 21, 21, 21, 21, 21, 21, 22, 22, 22, 22, 22, 22, 22, 22, 22, 22, 22, 22, 22, 22,
|
||
|
22, 22, 23, 23, 23, 23, 23, 23, 23, 23, 23, 23, 23, 23, 23, 23, 23, 23, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24,
|
||
|
24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 24, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25, 25,
|
||
|
25, 25, 25, 25, 25, 25, 25, 25, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26, 26,
|
||
|
26, 26, 26, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 28 ];
|
||
|
|
||
|
Tree.base_length = [ 0, 1, 2, 3, 4, 5, 6, 7, 8, 10, 12, 14, 16, 20, 24, 28, 32, 40, 48, 56, 64, 80, 96, 112, 128, 160, 192, 224, 0 ];
|
||
|
|
||
|
Tree.base_dist = [ 0, 1, 2, 3, 4, 6, 8, 12, 16, 24, 32, 48, 64, 96, 128, 192, 256, 384, 512, 768, 1024, 1536, 2048, 3072, 4096, 6144, 8192, 12288, 16384,
|
||
|
24576 ];
|
||
|
|
||
|
// Mapping from a distance to a distance code. dist is the distance - 1 and
|
||
|
// must not have side effects. _dist_code[256] and _dist_code[257] are never
|
||
|
// used.
|
||
|
Tree.d_code = function(dist) {
|
||
|
return ((dist) < 256 ? _dist_code[dist] : _dist_code[256 + ((dist) >>> 7)]);
|
||
|
};
|
||
|
|
||
|
// extra bits for each length code
|
||
|
Tree.extra_lbits = [ 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 2, 2, 2, 2, 3, 3, 3, 3, 4, 4, 4, 4, 5, 5, 5, 5, 0 ];
|
||
|
|
||
|
// extra bits for each distance code
|
||
|
Tree.extra_dbits = [ 0, 0, 0, 0, 1, 1, 2, 2, 3, 3, 4, 4, 5, 5, 6, 6, 7, 7, 8, 8, 9, 9, 10, 10, 11, 11, 12, 12, 13, 13 ];
|
||
|
|
||
|
// extra bits for each bit length code
|
||
|
Tree.extra_blbits = [ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2, 3, 7 ];
|
||
|
|
||
|
Tree.bl_order = [ 16, 17, 18, 0, 8, 7, 9, 6, 10, 5, 11, 4, 12, 3, 13, 2, 14, 1, 15 ];
|
||
|
|
||
|
// StaticTree
|
||
|
|
||
|
function StaticTree(static_tree, extra_bits, extra_base, elems, max_length) {
|
||
|
var that = this;
|
||
|
that.static_tree = static_tree;
|
||
|
that.extra_bits = extra_bits;
|
||
|
that.extra_base = extra_base;
|
||
|
that.elems = elems;
|
||
|
that.max_length = max_length;
|
||
|
}
|
||
|
|
||
|
StaticTree.static_ltree = [ 12, 8, 140, 8, 76, 8, 204, 8, 44, 8, 172, 8, 108, 8, 236, 8, 28, 8, 156, 8, 92, 8, 220, 8, 60, 8, 188, 8, 124, 8, 252, 8, 2, 8,
|
||
|
130, 8, 66, 8, 194, 8, 34, 8, 162, 8, 98, 8, 226, 8, 18, 8, 146, 8, 82, 8, 210, 8, 50, 8, 178, 8, 114, 8, 242, 8, 10, 8, 138, 8, 74, 8, 202, 8, 42,
|
||
|
8, 170, 8, 106, 8, 234, 8, 26, 8, 154, 8, 90, 8, 218, 8, 58, 8, 186, 8, 122, 8, 250, 8, 6, 8, 134, 8, 70, 8, 198, 8, 38, 8, 166, 8, 102, 8, 230, 8,
|
||
|
22, 8, 150, 8, 86, 8, 214, 8, 54, 8, 182, 8, 118, 8, 246, 8, 14, 8, 142, 8, 78, 8, 206, 8, 46, 8, 174, 8, 110, 8, 238, 8, 30, 8, 158, 8, 94, 8,
|
||
|
222, 8, 62, 8, 190, 8, 126, 8, 254, 8, 1, 8, 129, 8, 65, 8, 193, 8, 33, 8, 161, 8, 97, 8, 225, 8, 17, 8, 145, 8, 81, 8, 209, 8, 49, 8, 177, 8, 113,
|
||
|
8, 241, 8, 9, 8, 137, 8, 73, 8, 201, 8, 41, 8, 169, 8, 105, 8, 233, 8, 25, 8, 153, 8, 89, 8, 217, 8, 57, 8, 185, 8, 121, 8, 249, 8, 5, 8, 133, 8,
|
||
|
69, 8, 197, 8, 37, 8, 165, 8, 101, 8, 229, 8, 21, 8, 149, 8, 85, 8, 213, 8, 53, 8, 181, 8, 117, 8, 245, 8, 13, 8, 141, 8, 77, 8, 205, 8, 45, 8,
|
||
|
173, 8, 109, 8, 237, 8, 29, 8, 157, 8, 93, 8, 221, 8, 61, 8, 189, 8, 125, 8, 253, 8, 19, 9, 275, 9, 147, 9, 403, 9, 83, 9, 339, 9, 211, 9, 467, 9,
|
||
|
51, 9, 307, 9, 179, 9, 435, 9, 115, 9, 371, 9, 243, 9, 499, 9, 11, 9, 267, 9, 139, 9, 395, 9, 75, 9, 331, 9, 203, 9, 459, 9, 43, 9, 299, 9, 171, 9,
|
||
|
427, 9, 107, 9, 363, 9, 235, 9, 491, 9, 27, 9, 283, 9, 155, 9, 411, 9, 91, 9, 347, 9, 219, 9, 475, 9, 59, 9, 315, 9, 187, 9, 443, 9, 123, 9, 379,
|
||
|
9, 251, 9, 507, 9, 7, 9, 263, 9, 135, 9, 391, 9, 71, 9, 327, 9, 199, 9, 455, 9, 39, 9, 295, 9, 167, 9, 423, 9, 103, 9, 359, 9, 231, 9, 487, 9, 23,
|
||
|
9, 279, 9, 151, 9, 407, 9, 87, 9, 343, 9, 215, 9, 471, 9, 55, 9, 311, 9, 183, 9, 439, 9, 119, 9, 375, 9, 247, 9, 503, 9, 15, 9, 271, 9, 143, 9,
|
||
|
399, 9, 79, 9, 335, 9, 207, 9, 463, 9, 47, 9, 303, 9, 175, 9, 431, 9, 111, 9, 367, 9, 239, 9, 495, 9, 31, 9, 287, 9, 159, 9, 415, 9, 95, 9, 351, 9,
|
||
|
223, 9, 479, 9, 63, 9, 319, 9, 191, 9, 447, 9, 127, 9, 383, 9, 255, 9, 511, 9, 0, 7, 64, 7, 32, 7, 96, 7, 16, 7, 80, 7, 48, 7, 112, 7, 8, 7, 72, 7,
|
||
|
40, 7, 104, 7, 24, 7, 88, 7, 56, 7, 120, 7, 4, 7, 68, 7, 36, 7, 100, 7, 20, 7, 84, 7, 52, 7, 116, 7, 3, 8, 131, 8, 67, 8, 195, 8, 35, 8, 163, 8,
|
||
|
99, 8, 227, 8 ];
|
||
|
|
||
|
StaticTree.static_dtree = [ 0, 5, 16, 5, 8, 5, 24, 5, 4, 5, 20, 5, 12, 5, 28, 5, 2, 5, 18, 5, 10, 5, 26, 5, 6, 5, 22, 5, 14, 5, 30, 5, 1, 5, 17, 5, 9, 5,
|
||
|
25, 5, 5, 5, 21, 5, 13, 5, 29, 5, 3, 5, 19, 5, 11, 5, 27, 5, 7, 5, 23, 5 ];
|
||
|
|
||
|
StaticTree.static_l_desc = new StaticTree(StaticTree.static_ltree, Tree.extra_lbits, LITERALS + 1, L_CODES, MAX_BITS);
|
||
|
|
||
|
StaticTree.static_d_desc = new StaticTree(StaticTree.static_dtree, Tree.extra_dbits, 0, D_CODES, MAX_BITS);
|
||
|
|
||
|
StaticTree.static_bl_desc = new StaticTree(null, Tree.extra_blbits, 0, BL_CODES, MAX_BL_BITS);
|
||
|
|
||
|
// Deflate
|
||
|
|
||
|
var MAX_MEM_LEVEL = 9;
|
||
|
var DEF_MEM_LEVEL = 8;
|
||
|
|
||
|
function Config(good_length, max_lazy, nice_length, max_chain, func) {
|
||
|
var that = this;
|
||
|
that.good_length = good_length;
|
||
|
that.max_lazy = max_lazy;
|
||
|
that.nice_length = nice_length;
|
||
|
that.max_chain = max_chain;
|
||
|
that.func = func;
|
||
|
}
|
||
|
|
||
|
var STORED = 0;
|
||
|
var FAST = 1;
|
||
|
var SLOW = 2;
|
||
|
var config_table = [ new Config(0, 0, 0, 0, STORED), new Config(4, 4, 8, 4, FAST), new Config(4, 5, 16, 8, FAST), new Config(4, 6, 32, 32, FAST),
|
||
|
new Config(4, 4, 16, 16, SLOW), new Config(8, 16, 32, 32, SLOW), new Config(8, 16, 128, 128, SLOW), new Config(8, 32, 128, 256, SLOW),
|
||
|
new Config(32, 128, 258, 1024, SLOW), new Config(32, 258, 258, 4096, SLOW) ];
|
||
|
|
||
|
var z_errmsg = [ "need dictionary", // Z_NEED_DICT
|
||
|
// 2
|
||
|
"stream end", // Z_STREAM_END 1
|
||
|
"", // Z_OK 0
|
||
|
"", // Z_ERRNO (-1)
|
||
|
"stream error", // Z_STREAM_ERROR (-2)
|
||
|
"data error", // Z_DATA_ERROR (-3)
|
||
|
"", // Z_MEM_ERROR (-4)
|
||
|
"buffer error", // Z_BUF_ERROR (-5)
|
||
|
"",// Z_VERSION_ERROR (-6)
|
||
|
"" ];
|
||
|
|
||
|
// block not completed, need more input or more output
|
||
|
var NeedMore = 0;
|
||
|
|
||
|
// block flush performed
|
||
|
var BlockDone = 1;
|
||
|
|
||
|
// finish started, need only more output at next deflate
|
||
|
var FinishStarted = 2;
|
||
|
|
||
|
// finish done, accept no more input or output
|
||
|
var FinishDone = 3;
|
||
|
|
||
|
// preset dictionary flag in zlib header
|
||
|
var PRESET_DICT = 0x20;
|
||
|
|
||
|
var INIT_STATE = 42;
|
||
|
var BUSY_STATE = 113;
|
||
|
var FINISH_STATE = 666;
|
||
|
|
||
|
// The deflate compression method
|
||
|
var Z_DEFLATED = 8;
|
||
|
|
||
|
var STORED_BLOCK = 0;
|
||
|
var STATIC_TREES = 1;
|
||
|
var DYN_TREES = 2;
|
||
|
|
||
|
var MIN_MATCH = 3;
|
||
|
var MAX_MATCH = 258;
|
||
|
var MIN_LOOKAHEAD = (MAX_MATCH + MIN_MATCH + 1);
|
||
|
|
||
|
function smaller(tree, n, m, depth) {
|
||
|
var tn2 = tree[n * 2];
|
||
|
var tm2 = tree[m * 2];
|
||
|
return (tn2 < tm2 || (tn2 == tm2 && depth[n] <= depth[m]));
|
||
|
}
|
||
|
|
||
|
function Deflate() {
|
||
|
|
||
|
var that = this;
|
||
|
var strm; // pointer back to this zlib stream
|
||
|
var status; // as the name implies
|
||
|
// pending_buf; // output still pending
|
||
|
var pending_buf_size; // size of pending_buf
|
||
|
// pending_out; // next pending byte to output to the stream
|
||
|
// pending; // nb of bytes in the pending buffer
|
||
|
var method; // STORED (for zip only) or DEFLATED
|
||
|
var last_flush; // value of flush param for previous deflate call
|
||
|
|
||
|
var w_size; // LZ77 window size (32K by default)
|
||
|
var w_bits; // log2(w_size) (8..16)
|
||
|
var w_mask; // w_size - 1
|
||
|
|
||
|
var window;
|
||
|
// Sliding window. Input bytes are read into the second half of the window,
|
||
|
// and move to the first half later to keep a dictionary of at least wSize
|
||
|
// bytes. With this organization, matches are limited to a distance of
|
||
|
// wSize-MAX_MATCH bytes, but this ensures that IO is always
|
||
|
// performed with a length multiple of the block size. Also, it limits
|
||
|
// the window size to 64K, which is quite useful on MSDOS.
|
||
|
// To do: use the user input buffer as sliding window.
|
||
|
|
||
|
var window_size;
|
||
|
// Actual size of window: 2*wSize, except when the user input buffer
|
||
|
// is directly used as sliding window.
|
||
|
|
||
|
var prev;
|
||
|
// Link to older string with same hash index. To limit the size of this
|
||
|
// array to 64K, this link is maintained only for the last 32K strings.
|
||
|
// An index in this array is thus a window index modulo 32K.
|
||
|
|
||
|
var head; // Heads of the hash chains or NIL.
|
||
|
|
||
|
var ins_h; // hash index of string to be inserted
|
||
|
var hash_size; // number of elements in hash table
|
||
|
var hash_bits; // log2(hash_size)
|
||
|
var hash_mask; // hash_size-1
|
||
|
|
||
|
// Number of bits by which ins_h must be shifted at each input
|
||
|
// step. It must be such that after MIN_MATCH steps, the oldest
|
||
|
// byte no longer takes part in the hash key, that is:
|
||
|
// hash_shift * MIN_MATCH >= hash_bits
|
||
|
var hash_shift;
|
||
|
|
||
|
// Window position at the beginning of the current output block. Gets
|
||
|
// negative when the window is moved backwards.
|
||
|
|
||
|
var block_start;
|
||
|
|
||
|
var match_length; // length of best match
|
||
|
var prev_match; // previous match
|
||
|
var match_available; // set if previous match exists
|
||
|
var strstart; // start of string to insert
|
||
|
var match_start; // start of matching string
|
||
|
var lookahead; // number of valid bytes ahead in window
|
||
|
|
||
|
// Length of the best match at previous step. Matches not greater than this
|
||
|
// are discarded. This is used in the lazy match evaluation.
|
||
|
var prev_length;
|
||
|
|
||
|
// To speed up deflation, hash chains are never searched beyond this
|
||
|
// length. A higher limit improves compression ratio but degrades the speed.
|
||
|
var max_chain_length;
|
||
|
|
||
|
// Attempt to find a better match only when the current match is strictly
|
||
|
// smaller than this value. This mechanism is used only for compression
|
||
|
// levels >= 4.
|
||
|
var max_lazy_match;
|
||
|
|
||
|
// Insert new strings in the hash table only if the match length is not
|
||
|
// greater than this length. This saves time but degrades compression.
|
||
|
// max_insert_length is used only for compression levels <= 3.
|
||
|
|
||
|
var level; // compression level (1..9)
|
||
|
var strategy; // favor or force Huffman coding
|
||
|
|
||
|
// Use a faster search when the previous match is longer than this
|
||
|
var good_match;
|
||
|
|
||
|
// Stop searching when current match exceeds this
|
||
|
var nice_match;
|
||
|
|
||
|
var dyn_ltree; // literal and length tree
|
||
|
var dyn_dtree; // distance tree
|
||
|
var bl_tree; // Huffman tree for bit lengths
|
||
|
|
||
|
var l_desc = new Tree(); // desc for literal tree
|
||
|
var d_desc = new Tree(); // desc for distance tree
|
||
|
var bl_desc = new Tree(); // desc for bit length tree
|
||
|
|
||
|
// that.heap_len; // number of elements in the heap
|
||
|
// that.heap_max; // element of largest frequency
|
||
|
// The sons of heap[n] are heap[2*n] and heap[2*n+1]. heap[0] is not used.
|
||
|
// The same heap array is used to build all trees.
|
||
|
|
||
|
// Depth of each subtree used as tie breaker for trees of equal frequency
|
||
|
that.depth = [];
|
||
|
|
||
|
var l_buf; // index for literals or lengths */
|
||
|
|
||
|
// Size of match buffer for literals/lengths. There are 4 reasons for
|
||
|
// limiting lit_bufsize to 64K:
|
||
|
// - frequencies can be kept in 16 bit counters
|
||
|
// - if compression is not successful for the first block, all input
|
||
|
// data is still in the window so we can still emit a stored block even
|
||
|
// when input comes from standard input. (This can also be done for
|
||
|
// all blocks if lit_bufsize is not greater than 32K.)
|
||
|
// - if compression is not successful for a file smaller than 64K, we can
|
||
|
// even emit a stored file instead of a stored block (saving 5 bytes).
|
||
|
// This is applicable only for zip (not gzip or zlib).
|
||
|
// - creating new Huffman trees less frequently may not provide fast
|
||
|
// adaptation to changes in the input data statistics. (Take for
|
||
|
// example a binary file with poorly compressible code followed by
|
||
|
// a highly compressible string table.) Smaller buffer sizes give
|
||
|
// fast adaptation but have of course the overhead of transmitting
|
||
|
// trees more frequently.
|
||
|
// - I can't count above 4
|
||
|
var lit_bufsize;
|
||
|
|
||
|
var last_lit; // running index in l_buf
|
||
|
|
||
|
// Buffer for distances. To simplify the code, d_buf and l_buf have
|
||
|
// the same number of elements. To use different lengths, an extra flag
|
||
|
// array would be necessary.
|
||
|
|
||
|
var d_buf; // index of pendig_buf
|
||
|
|
||
|
// that.opt_len; // bit length of current block with optimal trees
|
||
|
// that.static_len; // bit length of current block with static trees
|
||
|
var matches; // number of string matches in current block
|
||
|
var last_eob_len; // bit length of EOB code for last block
|
||
|
|
||
|
// Output buffer. bits are inserted starting at the bottom (least
|
||
|
// significant bits).
|
||
|
var bi_buf;
|
||
|
|
||
|
// Number of valid bits in bi_buf. All bits above the last valid bit
|
||
|
// are always zero.
|
||
|
var bi_valid;
|
||
|
|
||
|
// number of codes at each bit length for an optimal tree
|
||
|
that.bl_count = [];
|
||
|
|
||
|
// heap used to build the Huffman trees
|
||
|
that.heap = [];
|
||
|
|
||
|
dyn_ltree = [];
|
||
|
dyn_dtree = [];
|
||
|
bl_tree = [];
|
||
|
|
||
|
function lm_init() {
|
||
|
var i;
|
||
|
window_size = 2 * w_size;
|
||
|
|
||
|
head[hash_size - 1] = 0;
|
||
|
for (i = 0; i < hash_size - 1; i++) {
|
||
|
head[i] = 0;
|
||
|
}
|
||
|
|
||
|
// Set the default configuration parameters:
|
||
|
max_lazy_match = config_table[level].max_lazy;
|
||
|
good_match = config_table[level].good_length;
|
||
|
nice_match = config_table[level].nice_length;
|
||
|
max_chain_length = config_table[level].max_chain;
|
||
|
|
||
|
strstart = 0;
|
||
|
block_start = 0;
|
||
|
lookahead = 0;
|
||
|
match_length = prev_length = MIN_MATCH - 1;
|
||
|
match_available = 0;
|
||
|
ins_h = 0;
|
||
|
}
|
||
|
|
||
|
function init_block() {
|
||
|
var i;
|
||
|
// Initialize the trees.
|
||
|
for (i = 0; i < L_CODES; i++)
|
||
|
dyn_ltree[i * 2] = 0;
|
||
|
for (i = 0; i < D_CODES; i++)
|
||
|
dyn_dtree[i * 2] = 0;
|
||
|
for (i = 0; i < BL_CODES; i++)
|
||
|
bl_tree[i * 2] = 0;
|
||
|
|
||
|
dyn_ltree[END_BLOCK * 2] = 1;
|
||
|
that.opt_len = that.static_len = 0;
|
||
|
last_lit = matches = 0;
|
||
|
}
|
||
|
|
||
|
// Initialize the tree data structures for a new zlib stream.
|
||
|
function tr_init() {
|
||
|
|
||
|
l_desc.dyn_tree = dyn_ltree;
|
||
|
l_desc.stat_desc = StaticTree.static_l_desc;
|
||
|
|
||
|
d_desc.dyn_tree = dyn_dtree;
|
||
|
d_desc.stat_desc = StaticTree.static_d_desc;
|
||
|
|
||
|
bl_desc.dyn_tree = bl_tree;
|
||
|
bl_desc.stat_desc = StaticTree.static_bl_desc;
|
||
|
|
||
|
bi_buf = 0;
|
||
|
bi_valid = 0;
|
||
|
last_eob_len = 8; // enough lookahead for inflate
|
||
|
|
||
|
// Initialize the first block of the first file:
|
||
|
init_block();
|
||
|
}
|
||
|
|
||
|
// Restore the heap property by moving down the tree starting at node k,
|
||
|
// exchanging a node with the smallest of its two sons if necessary,
|
||
|
// stopping
|
||
|
// when the heap property is re-established (each father smaller than its
|
||
|
// two sons).
|
||
|
that.pqdownheap = function(tree, // the tree to restore
|
||
|
k // node to move down
|
||
|
) {
|
||
|
var heap = that.heap;
|
||
|
var v = heap[k];
|
||
|
var j = k << 1; // left son of k
|
||
|
while (j <= that.heap_len) {
|
||
|
// Set j to the smallest of the two sons:
|
||
|
if (j < that.heap_len && smaller(tree, heap[j + 1], heap[j], that.depth)) {
|
||
|
j++;
|
||
|
}
|
||
|
// Exit if v is smaller than both sons
|
||
|
if (smaller(tree, v, heap[j], that.depth))
|
||
|
break;
|
||
|
|
||
|
// Exchange v with the smallest son
|
||
|
heap[k] = heap[j];
|
||
|
k = j;
|
||
|
// And continue down the tree, setting j to the left son of k
|
||
|
j <<= 1;
|
||
|
}
|
||
|
heap[k] = v;
|
||
|
};
|
||
|
|
||
|
// Scan a literal or distance tree to determine the frequencies of the codes
|
||
|
// in the bit length tree.
|
||
|
function scan_tree(tree,// the tree to be scanned
|
||
|
max_code // and its largest code of non zero frequency
|
||
|
) {
|
||
|
var n; // iterates over all tree elements
|
||
|
var prevlen = -1; // last emitted length
|
||
|
var curlen; // length of current code
|
||
|
var nextlen = tree[0 * 2 + 1]; // length of next code
|
||
|
var count = 0; // repeat count of the current code
|
||
|
var max_count = 7; // max repeat count
|
||
|
var min_count = 4; // min repeat count
|
||
|
|
||
|
if (nextlen === 0) {
|
||
|
max_count = 138;
|
||
|
min_count = 3;
|
||
|
}
|
||
|
tree[(max_code + 1) * 2 + 1] = 0xffff; // guard
|
||
|
|
||
|
for (n = 0; n <= max_code; n++) {
|
||
|
curlen = nextlen;
|
||
|
nextlen = tree[(n + 1) * 2 + 1];
|
||
|
if (++count < max_count && curlen == nextlen) {
|
||
|
continue;
|
||
|
} else if (count < min_count) {
|
||
|
bl_tree[curlen * 2] += count;
|
||
|
} else if (curlen !== 0) {
|
||
|
if (curlen != prevlen)
|
||
|
bl_tree[curlen * 2]++;
|
||
|
bl_tree[REP_3_6 * 2]++;
|
||
|
} else if (count <= 10) {
|
||
|
bl_tree[REPZ_3_10 * 2]++;
|
||
|
} else {
|
||
|
bl_tree[REPZ_11_138 * 2]++;
|
||
|
}
|
||
|
count = 0;
|
||
|
prevlen = curlen;
|
||
|
if (nextlen === 0) {
|
||
|
max_count = 138;
|
||
|
min_count = 3;
|
||
|
} else if (curlen == nextlen) {
|
||
|
max_count = 6;
|
||
|
min_count = 3;
|
||
|
} else {
|
||
|
max_count = 7;
|
||
|
min_count = 4;
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Construct the Huffman tree for the bit lengths and return the index in
|
||
|
// bl_order of the last bit length code to send.
|
||
|
function build_bl_tree() {
|
||
|
var max_blindex; // index of last bit length code of non zero freq
|
||
|
|
||
|
// Determine the bit length frequencies for literal and distance trees
|
||
|
scan_tree(dyn_ltree, l_desc.max_code);
|
||
|
scan_tree(dyn_dtree, d_desc.max_code);
|
||
|
|
||
|
// Build the bit length tree:
|
||
|
bl_desc.build_tree(that);
|
||
|
// opt_len now includes the length of the tree representations, except
|
||
|
// the lengths of the bit lengths codes and the 5+5+4 bits for the
|
||
|
// counts.
|
||
|
|
||
|
// Determine the number of bit length codes to send. The pkzip format
|
||
|
// requires that at least 4 bit length codes be sent. (appnote.txt says
|
||
|
// 3 but the actual value used is 4.)
|
||
|
for (max_blindex = BL_CODES - 1; max_blindex >= 3; max_blindex--) {
|
||
|
if (bl_tree[Tree.bl_order[max_blindex] * 2 + 1] !== 0)
|
||
|
break;
|
||
|
}
|
||
|
// Update opt_len to include the bit length tree and counts
|
||
|
that.opt_len += 3 * (max_blindex + 1) + 5 + 5 + 4;
|
||
|
|
||
|
return max_blindex;
|
||
|
}
|
||
|
|
||
|
// Output a byte on the stream.
|
||
|
// IN assertion: there is enough room in pending_buf.
|
||
|
function put_byte(p) {
|
||
|
that.pending_buf[that.pending++] = p;
|
||
|
}
|
||
|
|
||
|
function put_short(w) {
|
||
|
put_byte(w & 0xff);
|
||
|
put_byte((w >>> 8) & 0xff);
|
||
|
}
|
||
|
|
||
|
function putShortMSB(b) {
|
||
|
put_byte((b >> 8) & 0xff);
|
||
|
put_byte((b & 0xff) & 0xff);
|
||
|
}
|
||
|
|
||
|
function send_bits(value, length) {
|
||
|
var val, len = length;
|
||
|
if (bi_valid > Buf_size - len) {
|
||
|
val = value;
|
||
|
// bi_buf |= (val << bi_valid);
|
||
|
bi_buf |= ((val << bi_valid) & 0xffff);
|
||
|
put_short(bi_buf);
|
||
|
bi_buf = val >>> (Buf_size - bi_valid);
|
||
|
bi_valid += len - Buf_size;
|
||
|
} else {
|
||
|
// bi_buf |= (value) << bi_valid;
|
||
|
bi_buf |= (((value) << bi_valid) & 0xffff);
|
||
|
bi_valid += len;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
function send_code(c, tree) {
|
||
|
var c2 = c * 2;
|
||
|
send_bits(tree[c2] & 0xffff, tree[c2 + 1] & 0xffff);
|
||
|
}
|
||
|
|
||
|
// Send a literal or distance tree in compressed form, using the codes in
|
||
|
// bl_tree.
|
||
|
function send_tree(tree,// the tree to be sent
|
||
|
max_code // and its largest code of non zero frequency
|
||
|
) {
|
||
|
var n; // iterates over all tree elements
|
||
|
var prevlen = -1; // last emitted length
|
||
|
var curlen; // length of current code
|
||
|
var nextlen = tree[0 * 2 + 1]; // length of next code
|
||
|
var count = 0; // repeat count of the current code
|
||
|
var max_count = 7; // max repeat count
|
||
|
var min_count = 4; // min repeat count
|
||
|
|
||
|
if (nextlen === 0) {
|
||
|
max_count = 138;
|
||
|
min_count = 3;
|
||
|
}
|
||
|
|
||
|
for (n = 0; n <= max_code; n++) {
|
||
|
curlen = nextlen;
|
||
|
nextlen = tree[(n + 1) * 2 + 1];
|
||
|
if (++count < max_count && curlen == nextlen) {
|
||
|
continue;
|
||
|
} else if (count < min_count) {
|
||
|
do {
|
||
|
send_code(curlen, bl_tree);
|
||
|
} while (--count !== 0);
|
||
|
} else if (curlen !== 0) {
|
||
|
if (curlen != prevlen) {
|
||
|
send_code(curlen, bl_tree);
|
||
|
count--;
|
||
|
}
|
||
|
send_code(REP_3_6, bl_tree);
|
||
|
send_bits(count - 3, 2);
|
||
|
} else if (count <= 10) {
|
||
|
send_code(REPZ_3_10, bl_tree);
|
||
|
send_bits(count - 3, 3);
|
||
|
} else {
|
||
|
send_code(REPZ_11_138, bl_tree);
|
||
|
send_bits(count - 11, 7);
|
||
|
}
|
||
|
count = 0;
|
||
|
prevlen = curlen;
|
||
|
if (nextlen === 0) {
|
||
|
max_count = 138;
|
||
|
min_count = 3;
|
||
|
} else if (curlen == nextlen) {
|
||
|
max_count = 6;
|
||
|
min_count = 3;
|
||
|
} else {
|
||
|
max_count = 7;
|
||
|
min_count = 4;
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Send the header for a block using dynamic Huffman trees: the counts, the
|
||
|
// lengths of the bit length codes, the literal tree and the distance tree.
|
||
|
// IN assertion: lcodes >= 257, dcodes >= 1, blcodes >= 4.
|
||
|
function send_all_trees(lcodes, dcodes, blcodes) {
|
||
|
var rank; // index in bl_order
|
||
|
|
||
|
send_bits(lcodes - 257, 5); // not +255 as stated in appnote.txt
|
||
|
send_bits(dcodes - 1, 5);
|
||
|
send_bits(blcodes - 4, 4); // not -3 as stated in appnote.txt
|
||
|
for (rank = 0; rank < blcodes; rank++) {
|
||
|
send_bits(bl_tree[Tree.bl_order[rank] * 2 + 1], 3);
|
||
|
}
|
||
|
send_tree(dyn_ltree, lcodes - 1); // literal tree
|
||
|
send_tree(dyn_dtree, dcodes - 1); // distance tree
|
||
|
}
|
||
|
|
||
|
// Flush the bit buffer, keeping at most 7 bits in it.
|
||
|
function bi_flush() {
|
||
|
if (bi_valid == 16) {
|
||
|
put_short(bi_buf);
|
||
|
bi_buf = 0;
|
||
|
bi_valid = 0;
|
||
|
} else if (bi_valid >= 8) {
|
||
|
put_byte(bi_buf & 0xff);
|
||
|
bi_buf >>>= 8;
|
||
|
bi_valid -= 8;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Send one empty static block to give enough lookahead for inflate.
|
||
|
// This takes 10 bits, of which 7 may remain in the bit buffer.
|
||
|
// The current inflate code requires 9 bits of lookahead. If the
|
||
|
// last two codes for the previous block (real code plus EOB) were coded
|
||
|
// on 5 bits or less, inflate may have only 5+3 bits of lookahead to decode
|
||
|
// the last real code. In this case we send two empty static blocks instead
|
||
|
// of one. (There are no problems if the previous block is stored or fixed.)
|
||
|
// To simplify the code, we assume the worst case of last real code encoded
|
||
|
// on one bit only.
|
||
|
function _tr_align() {
|
||
|
send_bits(STATIC_TREES << 1, 3);
|
||
|
send_code(END_BLOCK, StaticTree.static_ltree);
|
||
|
|
||
|
bi_flush();
|
||
|
|
||
|
// Of the 10 bits for the empty block, we have already sent
|
||
|
// (10 - bi_valid) bits. The lookahead for the last real code (before
|
||
|
// the EOB of the previous block) was thus at least one plus the length
|
||
|
// of the EOB plus what we have just sent of the empty static block.
|
||
|
if (1 + last_eob_len + 10 - bi_valid < 9) {
|
||
|
send_bits(STATIC_TREES << 1, 3);
|
||
|
send_code(END_BLOCK, StaticTree.static_ltree);
|
||
|
bi_flush();
|
||
|
}
|
||
|
last_eob_len = 7;
|
||
|
}
|
||
|
|
||
|
// Save the match info and tally the frequency counts. Return true if
|
||
|
// the current block must be flushed.
|
||
|
function _tr_tally(dist, // distance of matched string
|
||
|
lc // match length-MIN_MATCH or unmatched char (if dist==0)
|
||
|
) {
|
||
|
var out_length, in_length, dcode;
|
||
|
that.pending_buf[d_buf + last_lit * 2] = (dist >>> 8) & 0xff;
|
||
|
that.pending_buf[d_buf + last_lit * 2 + 1] = dist & 0xff;
|
||
|
|
||
|
that.pending_buf[l_buf + last_lit] = lc & 0xff;
|
||
|
last_lit++;
|
||
|
|
||
|
if (dist === 0) {
|
||
|
// lc is the unmatched char
|
||
|
dyn_ltree[lc * 2]++;
|
||
|
} else {
|
||
|
matches++;
|
||
|
// Here, lc is the match length - MIN_MATCH
|
||
|
dist--; // dist = match distance - 1
|
||
|
dyn_ltree[(Tree._length_code[lc] + LITERALS + 1) * 2]++;
|
||
|
dyn_dtree[Tree.d_code(dist) * 2]++;
|
||
|
}
|
||
|
|
||
|
if ((last_lit & 0x1fff) === 0 && level > 2) {
|
||
|
// Compute an upper bound for the compressed length
|
||
|
out_length = last_lit * 8;
|
||
|
in_length = strstart - block_start;
|
||
|
for (dcode = 0; dcode < D_CODES; dcode++) {
|
||
|
out_length += dyn_dtree[dcode * 2] * (5 + Tree.extra_dbits[dcode]);
|
||
|
}
|
||
|
out_length >>>= 3;
|
||
|
if ((matches < Math.floor(last_lit / 2)) && out_length < Math.floor(in_length / 2))
|
||
|
return true;
|
||
|
}
|
||
|
|
||
|
return (last_lit == lit_bufsize - 1);
|
||
|
// We avoid equality with lit_bufsize because of wraparound at 64K
|
||
|
// on 16 bit machines and because stored blocks are restricted to
|
||
|
// 64K-1 bytes.
|
||
|
}
|
||
|
|
||
|
// Send the block data compressed using the given Huffman trees
|
||
|
function compress_block(ltree, dtree) {
|
||
|
var dist; // distance of matched string
|
||
|
var lc; // match length or unmatched char (if dist === 0)
|
||
|
var lx = 0; // running index in l_buf
|
||
|
var code; // the code to send
|
||
|
var extra; // number of extra bits to send
|
||
|
|
||
|
if (last_lit !== 0) {
|
||
|
do {
|
||
|
dist = ((that.pending_buf[d_buf + lx * 2] << 8) & 0xff00) | (that.pending_buf[d_buf + lx * 2 + 1] & 0xff);
|
||
|
lc = (that.pending_buf[l_buf + lx]) & 0xff;
|
||
|
lx++;
|
||
|
|
||
|
if (dist === 0) {
|
||
|
send_code(lc, ltree); // send a literal byte
|
||
|
} else {
|
||
|
// Here, lc is the match length - MIN_MATCH
|
||
|
code = Tree._length_code[lc];
|
||
|
|
||
|
send_code(code + LITERALS + 1, ltree); // send the length
|
||
|
// code
|
||
|
extra = Tree.extra_lbits[code];
|
||
|
if (extra !== 0) {
|
||
|
lc -= Tree.base_length[code];
|
||
|
send_bits(lc, extra); // send the extra length bits
|
||
|
}
|
||
|
dist--; // dist is now the match distance - 1
|
||
|
code = Tree.d_code(dist);
|
||
|
|
||
|
send_code(code, dtree); // send the distance code
|
||
|
extra = Tree.extra_dbits[code];
|
||
|
if (extra !== 0) {
|
||
|
dist -= Tree.base_dist[code];
|
||
|
send_bits(dist, extra); // send the extra distance bits
|
||
|
}
|
||
|
} // literal or match pair ?
|
||
|
|
||
|
// Check that the overlay between pending_buf and d_buf+l_buf is
|
||
|
// ok:
|
||
|
} while (lx < last_lit);
|
||
|
}
|
||
|
|
||
|
send_code(END_BLOCK, ltree);
|
||
|
last_eob_len = ltree[END_BLOCK * 2 + 1];
|
||
|
}
|
||
|
|
||
|
// Flush the bit buffer and align the output on a byte boundary
|
||
|
function bi_windup() {
|
||
|
if (bi_valid > 8) {
|
||
|
put_short(bi_buf);
|
||
|
} else if (bi_valid > 0) {
|
||
|
put_byte(bi_buf & 0xff);
|
||
|
}
|
||
|
bi_buf = 0;
|
||
|
bi_valid = 0;
|
||
|
}
|
||
|
|
||
|
// Copy a stored block, storing first the length and its
|
||
|
// one's complement if requested.
|
||
|
function copy_block(buf, // the input data
|
||
|
len, // its length
|
||
|
header // true if block header must be written
|
||
|
) {
|
||
|
bi_windup(); // align on byte boundary
|
||
|
last_eob_len = 8; // enough lookahead for inflate
|
||
|
|
||
|
if (header) {
|
||
|
put_short(len);
|
||
|
put_short(~len);
|
||
|
}
|
||
|
|
||
|
that.pending_buf.set(window.subarray(buf, buf + len), that.pending);
|
||
|
that.pending += len;
|
||
|
}
|
||
|
|
||
|
// Send a stored block
|
||
|
function _tr_stored_block(buf, // input block
|
||
|
stored_len, // length of input block
|
||
|
eof // true if this is the last block for a file
|
||
|
) {
|
||
|
send_bits((STORED_BLOCK << 1) + (eof ? 1 : 0), 3); // send block type
|
||
|
copy_block(buf, stored_len, true); // with header
|
||
|
}
|
||
|
|
||
|
// Determine the best encoding for the current block: dynamic trees, static
|
||
|
// trees or store, and output the encoded block to the zip file.
|
||
|
function _tr_flush_block(buf, // input block, or NULL if too old
|
||
|
stored_len, // length of input block
|
||
|
eof // true if this is the last block for a file
|
||
|
) {
|
||
|
var opt_lenb, static_lenb;// opt_len and static_len in bytes
|
||
|
var max_blindex = 0; // index of last bit length code of non zero freq
|
||
|
|
||
|
// Build the Huffman trees unless a stored block is forced
|
||
|
if (level > 0) {
|
||
|
// Construct the literal and distance trees
|
||
|
l_desc.build_tree(that);
|
||
|
|
||
|
d_desc.build_tree(that);
|
||
|
|
||
|
// At this point, opt_len and static_len are the total bit lengths
|
||
|
// of
|
||
|
// the compressed block data, excluding the tree representations.
|
||
|
|
||
|
// Build the bit length tree for the above two trees, and get the
|
||
|
// index
|
||
|
// in bl_order of the last bit length code to send.
|
||
|
max_blindex = build_bl_tree();
|
||
|
|
||
|
// Determine the best encoding. Compute first the block length in
|
||
|
// bytes
|
||
|
opt_lenb = (that.opt_len + 3 + 7) >>> 3;
|
||
|
static_lenb = (that.static_len + 3 + 7) >>> 3;
|
||
|
|
||
|
if (static_lenb <= opt_lenb)
|
||
|
opt_lenb = static_lenb;
|
||
|
} else {
|
||
|
opt_lenb = static_lenb = stored_len + 5; // force a stored block
|
||
|
}
|
||
|
|
||
|
if ((stored_len + 4 <= opt_lenb) && buf != -1) {
|
||
|
// 4: two words for the lengths
|
||
|
// The test buf != NULL is only necessary if LIT_BUFSIZE > WSIZE.
|
||
|
// Otherwise we can't have processed more than WSIZE input bytes
|
||
|
// since
|
||
|
// the last block flush, because compression would have been
|
||
|
// successful. If LIT_BUFSIZE <= WSIZE, it is never too late to
|
||
|
// transform a block into a stored block.
|
||
|
_tr_stored_block(buf, stored_len, eof);
|
||
|
} else if (static_lenb == opt_lenb) {
|
||
|
send_bits((STATIC_TREES << 1) + (eof ? 1 : 0), 3);
|
||
|
compress_block(StaticTree.static_ltree, StaticTree.static_dtree);
|
||
|
} else {
|
||
|
send_bits((DYN_TREES << 1) + (eof ? 1 : 0), 3);
|
||
|
send_all_trees(l_desc.max_code + 1, d_desc.max_code + 1, max_blindex + 1);
|
||
|
compress_block(dyn_ltree, dyn_dtree);
|
||
|
}
|
||
|
|
||
|
// The above check is made mod 2^32, for files larger than 512 MB
|
||
|
// and uLong implemented on 32 bits.
|
||
|
|
||
|
init_block();
|
||
|
|
||
|
if (eof) {
|
||
|
bi_windup();
|
||
|
}
|
||
|
}
|
||
|
|
||
|
function flush_block_only(eof) {
|
||
|
_tr_flush_block(block_start >= 0 ? block_start : -1, strstart - block_start, eof);
|
||
|
block_start = strstart;
|
||
|
strm.flush_pending();
|
||
|
}
|
||
|
|
||
|
// Fill the window when the lookahead becomes insufficient.
|
||
|
// Updates strstart and lookahead.
|
||
|
//
|
||
|
// IN assertion: lookahead < MIN_LOOKAHEAD
|
||
|
// OUT assertions: strstart <= window_size-MIN_LOOKAHEAD
|
||
|
// At least one byte has been read, or avail_in === 0; reads are
|
||
|
// performed for at least two bytes (required for the zip translate_eol
|
||
|
// option -- not supported here).
|
||
|
function fill_window() {
|
||
|
var n, m;
|
||
|
var p;
|
||
|
var more; // Amount of free space at the end of the window.
|
||
|
|
||
|
do {
|
||
|
more = (window_size - lookahead - strstart);
|
||
|
|
||
|
// Deal with !@#$% 64K limit:
|
||
|
if (more === 0 && strstart === 0 && lookahead === 0) {
|
||
|
more = w_size;
|
||
|
} else if (more == -1) {
|
||
|
// Very unlikely, but possible on 16 bit machine if strstart ==
|
||
|
// 0
|
||
|
// and lookahead == 1 (input done one byte at time)
|
||
|
more--;
|
||
|
|
||
|
// If the window is almost full and there is insufficient
|
||
|
// lookahead,
|
||
|
// move the upper half to the lower one to make room in the
|
||
|
// upper half.
|
||
|
} else if (strstart >= w_size + w_size - MIN_LOOKAHEAD) {
|
||
|
window.set(window.subarray(w_size, w_size + w_size), 0);
|
||
|
|
||
|
match_start -= w_size;
|
||
|
strstart -= w_size; // we now have strstart >= MAX_DIST
|
||
|
block_start -= w_size;
|
||
|
|
||
|
// Slide the hash table (could be avoided with 32 bit values
|
||
|
// at the expense of memory usage). We slide even when level ==
|
||
|
// 0
|
||
|
// to keep the hash table consistent if we switch back to level
|
||
|
// > 0
|
||
|
// later. (Using level 0 permanently is not an optimal usage of
|
||
|
// zlib, so we don't care about this pathological case.)
|
||
|
|
||
|
n = hash_size;
|
||
|
p = n;
|
||
|
do {
|
||
|
m = (head[--p] & 0xffff);
|
||
|
head[p] = (m >= w_size ? m - w_size : 0);
|
||
|
} while (--n !== 0);
|
||
|
|
||
|
n = w_size;
|
||
|
p = n;
|
||
|
do {
|
||
|
m = (prev[--p] & 0xffff);
|
||
|
prev[p] = (m >= w_size ? m - w_size : 0);
|
||
|
// If n is not on any hash chain, prev[n] is garbage but
|
||
|
// its value will never be used.
|
||
|
} while (--n !== 0);
|
||
|
more += w_size;
|
||
|
}
|
||
|
|
||
|
if (strm.avail_in === 0)
|
||
|
return;
|
||
|
|
||
|
// If there was no sliding:
|
||
|
// strstart <= WSIZE+MAX_DIST-1 && lookahead <= MIN_LOOKAHEAD - 1 &&
|
||
|
// more == window_size - lookahead - strstart
|
||
|
// => more >= window_size - (MIN_LOOKAHEAD-1 + WSIZE + MAX_DIST-1)
|
||
|
// => more >= window_size - 2*WSIZE + 2
|
||
|
// In the BIG_MEM or MMAP case (not yet supported),
|
||
|
// window_size == input_size + MIN_LOOKAHEAD &&
|
||
|
// strstart + s->lookahead <= input_size => more >= MIN_LOOKAHEAD.
|
||
|
// Otherwise, window_size == 2*WSIZE so more >= 2.
|
||
|
// If there was sliding, more >= WSIZE. So in all cases, more >= 2.
|
||
|
|
||
|
n = strm.read_buf(window, strstart + lookahead, more);
|
||
|
lookahead += n;
|
||
|
|
||
|
// Initialize the hash value now that we have some input:
|
||
|
if (lookahead >= MIN_MATCH) {
|
||
|
ins_h = window[strstart] & 0xff;
|
||
|
ins_h = (((ins_h) << hash_shift) ^ (window[strstart + 1] & 0xff)) & hash_mask;
|
||
|
}
|
||
|
// If the whole input has less than MIN_MATCH bytes, ins_h is
|
||
|
// garbage,
|
||
|
// but this is not important since only literal bytes will be
|
||
|
// emitted.
|
||
|
} while (lookahead < MIN_LOOKAHEAD && strm.avail_in !== 0);
|
||
|
}
|
||
|
|
||
|
// Copy without compression as much as possible from the input stream,
|
||
|
// return
|
||
|
// the current block state.
|
||
|
// This function does not insert new strings in the dictionary since
|
||
|
// uncompressible data is probably not useful. This function is used
|
||
|
// only for the level=0 compression option.
|
||
|
// NOTE: this function should be optimized to avoid extra copying from
|
||
|
// window to pending_buf.
|
||
|
function deflate_stored(flush) {
|
||
|
// Stored blocks are limited to 0xffff bytes, pending_buf is limited
|
||
|
// to pending_buf_size, and each stored block has a 5 byte header:
|
||
|
|
||
|
var max_block_size = 0xffff;
|
||
|
var max_start;
|
||
|
|
||
|
if (max_block_size > pending_buf_size - 5) {
|
||
|
max_block_size = pending_buf_size - 5;
|
||
|
}
|
||
|
|
||
|
// Copy as much as possible from input to output:
|
||
|
while (true) {
|
||
|
// Fill the window as much as possible:
|
||
|
if (lookahead <= 1) {
|
||
|
fill_window();
|
||
|
if (lookahead === 0 && flush == Z_NO_FLUSH)
|
||
|
return NeedMore;
|
||
|
if (lookahead === 0)
|
||
|
break; // flush the current block
|
||
|
}
|
||
|
|
||
|
strstart += lookahead;
|
||
|
lookahead = 0;
|
||
|
|
||
|
// Emit a stored block if pending_buf will be full:
|
||
|
max_start = block_start + max_block_size;
|
||
|
if (strstart === 0 || strstart >= max_start) {
|
||
|
// strstart === 0 is possible when wraparound on 16-bit machine
|
||
|
lookahead = (strstart - max_start);
|
||
|
strstart = max_start;
|
||
|
|
||
|
flush_block_only(false);
|
||
|
if (strm.avail_out === 0)
|
||
|
return NeedMore;
|
||
|
|
||
|
}
|
||
|
|
||
|
// Flush if we may have to slide, otherwise block_start may become
|
||
|
// negative and the data will be gone:
|
||
|
if (strstart - block_start >= w_size - MIN_LOOKAHEAD) {
|
||
|
flush_block_only(false);
|
||
|
if (strm.avail_out === 0)
|
||
|
return NeedMore;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
flush_block_only(flush == Z_FINISH);
|
||
|
if (strm.avail_out === 0)
|
||
|
return (flush == Z_FINISH) ? FinishStarted : NeedMore;
|
||
|
|
||
|
return flush == Z_FINISH ? FinishDone : BlockDone;
|
||
|
}
|
||
|
|
||
|
function longest_match(cur_match) {
|
||
|
var chain_length = max_chain_length; // max hash chain length
|
||
|
var scan = strstart; // current string
|
||
|
var match; // matched string
|
||
|
var len; // length of current match
|
||
|
var best_len = prev_length; // best match length so far
|
||
|
var limit = strstart > (w_size - MIN_LOOKAHEAD) ? strstart - (w_size - MIN_LOOKAHEAD) : 0;
|
||
|
var _nice_match = nice_match;
|
||
|
|
||
|
// Stop when cur_match becomes <= limit. To simplify the code,
|
||
|
// we prevent matches with the string of window index 0.
|
||
|
|
||
|
var wmask = w_mask;
|
||
|
|
||
|
var strend = strstart + MAX_MATCH;
|
||
|
var scan_end1 = window[scan + best_len - 1];
|
||
|
var scan_end = window[scan + best_len];
|
||
|
|
||
|
// The code is optimized for HASH_BITS >= 8 and MAX_MATCH-2 multiple of
|
||
|
// 16.
|
||
|
// It is easy to get rid of this optimization if necessary.
|
||
|
|
||
|
// Do not waste too much time if we already have a good match:
|
||
|
if (prev_length >= good_match) {
|
||
|
chain_length >>= 2;
|
||
|
}
|
||
|
|
||
|
// Do not look for matches beyond the end of the input. This is
|
||
|
// necessary
|
||
|
// to make deflate deterministic.
|
||
|
if (_nice_match > lookahead)
|
||
|
_nice_match = lookahead;
|
||
|
|
||
|
do {
|
||
|
match = cur_match;
|
||
|
|
||
|
// Skip to next match if the match length cannot increase
|
||
|
// or if the match length is less than 2:
|
||
|
if (window[match + best_len] != scan_end || window[match + best_len - 1] != scan_end1 || window[match] != window[scan]
|
||
|
|| window[++match] != window[scan + 1])
|
||
|
continue;
|
||
|
|
||
|
// The check at best_len-1 can be removed because it will be made
|
||
|
// again later. (This heuristic is not always a win.)
|
||
|
// It is not necessary to compare scan[2] and match[2] since they
|
||
|
// are always equal when the other bytes match, given that
|
||
|
// the hash keys are equal and that HASH_BITS >= 8.
|
||
|
scan += 2;
|
||
|
match++;
|
||
|
|
||
|
// We check for insufficient lookahead only every 8th comparison;
|
||
|
// the 256th check will be made at strstart+258.
|
||
|
do {
|
||
|
} while (window[++scan] == window[++match] && window[++scan] == window[++match] && window[++scan] == window[++match]
|
||
|
&& window[++scan] == window[++match] && window[++scan] == window[++match] && window[++scan] == window[++match]
|
||
|
&& window[++scan] == window[++match] && window[++scan] == window[++match] && scan < strend);
|
||
|
|
||
|
len = MAX_MATCH - (strend - scan);
|
||
|
scan = strend - MAX_MATCH;
|
||
|
|
||
|
if (len > best_len) {
|
||
|
match_start = cur_match;
|
||
|
best_len = len;
|
||
|
if (len >= _nice_match)
|
||
|
break;
|
||
|
scan_end1 = window[scan + best_len - 1];
|
||
|
scan_end = window[scan + best_len];
|
||
|
}
|
||
|
|
||
|
} while ((cur_match = (prev[cur_match & wmask] & 0xffff)) > limit && --chain_length !== 0);
|
||
|
|
||
|
if (best_len <= lookahead)
|
||
|
return best_len;
|
||
|
return lookahead;
|
||
|
}
|
||
|
|
||
|
// Compress as much as possible from the input stream, return the current
|
||
|
// block state.
|
||
|
// This function does not perform lazy evaluation of matches and inserts
|
||
|
// new strings in the dictionary only for unmatched strings or for short
|
||
|
// matches. It is used only for the fast compression options.
|
||
|
function deflate_fast(flush) {
|
||
|
// short hash_head = 0; // head of the hash chain
|
||
|
var hash_head = 0; // head of the hash chain
|
||
|
var bflush; // set if current block must be flushed
|
||
|
|
||
|
while (true) {
|
||
|
// Make sure that we always have enough lookahead, except
|
||
|
// at the end of the input file. We need MAX_MATCH bytes
|
||
|
// for the next match, plus MIN_MATCH bytes to insert the
|
||
|
// string following the next match.
|
||
|
if (lookahead < MIN_LOOKAHEAD) {
|
||
|
fill_window();
|
||
|
if (lookahead < MIN_LOOKAHEAD && flush == Z_NO_FLUSH) {
|
||
|
return NeedMore;
|
||
|
}
|
||
|
if (lookahead === 0)
|
||
|
break; // flush the current block
|
||
|
}
|
||
|
|
||
|
// Insert the string window[strstart .. strstart+2] in the
|
||
|
// dictionary, and set hash_head to the head of the hash chain:
|
||
|
if (lookahead >= MIN_MATCH) {
|
||
|
ins_h = (((ins_h) << hash_shift) ^ (window[(strstart) + (MIN_MATCH - 1)] & 0xff)) & hash_mask;
|
||
|
|
||
|
// prev[strstart&w_mask]=hash_head=head[ins_h];
|
||
|
hash_head = (head[ins_h] & 0xffff);
|
||
|
prev[strstart & w_mask] = head[ins_h];
|
||
|
head[ins_h] = strstart;
|
||
|
}
|
||
|
|
||
|
// Find the longest match, discarding those <= prev_length.
|
||
|
// At this point we have always match_length < MIN_MATCH
|
||
|
|
||
|
if (hash_head !== 0 && ((strstart - hash_head) & 0xffff) <= w_size - MIN_LOOKAHEAD) {
|
||
|
// To simplify the code, we prevent matches with the string
|
||
|
// of window index 0 (in particular we have to avoid a match
|
||
|
// of the string with itself at the start of the input file).
|
||
|
if (strategy != Z_HUFFMAN_ONLY) {
|
||
|
match_length = longest_match(hash_head);
|
||
|
}
|
||
|
// longest_match() sets match_start
|
||
|
}
|
||
|
if (match_length >= MIN_MATCH) {
|
||
|
// check_match(strstart, match_start, match_length);
|
||
|
|
||
|
bflush = _tr_tally(strstart - match_start, match_length - MIN_MATCH);
|
||
|
|
||
|
lookahead -= match_length;
|
||
|
|
||
|
// Insert new strings in the hash table only if the match length
|
||
|
// is not too large. This saves time but degrades compression.
|
||
|
if (match_length <= max_lazy_match && lookahead >= MIN_MATCH) {
|
||
|
match_length--; // string at strstart already in hash table
|
||
|
do {
|
||
|
strstart++;
|
||
|
|
||
|
ins_h = ((ins_h << hash_shift) ^ (window[(strstart) + (MIN_MATCH - 1)] & 0xff)) & hash_mask;
|
||
|
// prev[strstart&w_mask]=hash_head=head[ins_h];
|
||
|
hash_head = (head[ins_h] & 0xffff);
|
||
|
prev[strstart & w_mask] = head[ins_h];
|
||
|
head[ins_h] = strstart;
|
||
|
|
||
|
// strstart never exceeds WSIZE-MAX_MATCH, so there are
|
||
|
// always MIN_MATCH bytes ahead.
|
||
|
} while (--match_length !== 0);
|
||
|
strstart++;
|
||
|
} else {
|
||
|
strstart += match_length;
|
||
|
match_length = 0;
|
||
|
ins_h = window[strstart] & 0xff;
|
||
|
|
||
|
ins_h = (((ins_h) << hash_shift) ^ (window[strstart + 1] & 0xff)) & hash_mask;
|
||
|
// If lookahead < MIN_MATCH, ins_h is garbage, but it does
|
||
|
// not
|
||
|
// matter since it will be recomputed at next deflate call.
|
||
|
}
|
||
|
} else {
|
||
|
// No match, output a literal byte
|
||
|
|
||
|
bflush = _tr_tally(0, window[strstart] & 0xff);
|
||
|
lookahead--;
|
||
|
strstart++;
|
||
|
}
|
||
|
if (bflush) {
|
||
|
|
||
|
flush_block_only(false);
|
||
|
if (strm.avail_out === 0)
|
||
|
return NeedMore;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
flush_block_only(flush == Z_FINISH);
|
||
|
if (strm.avail_out === 0) {
|
||
|
if (flush == Z_FINISH)
|
||
|
return FinishStarted;
|
||
|
else
|
||
|
return NeedMore;
|
||
|
}
|
||
|
return flush == Z_FINISH ? FinishDone : BlockDone;
|
||
|
}
|
||
|
|
||
|
// Same as above, but achieves better compression. We use a lazy
|
||
|
// evaluation for matches: a match is finally adopted only if there is
|
||
|
// no better match at the next window position.
|
||
|
function deflate_slow(flush) {
|
||
|
// short hash_head = 0; // head of hash chain
|
||
|
var hash_head = 0; // head of hash chain
|
||
|
var bflush; // set if current block must be flushed
|
||
|
var max_insert;
|
||
|
|
||
|
// Process the input block.
|
||
|
while (true) {
|
||
|
// Make sure that we always have enough lookahead, except
|
||
|
// at the end of the input file. We need MAX_MATCH bytes
|
||
|
// for the next match, plus MIN_MATCH bytes to insert the
|
||
|
// string following the next match.
|
||
|
|
||
|
if (lookahead < MIN_LOOKAHEAD) {
|
||
|
fill_window();
|
||
|
if (lookahead < MIN_LOOKAHEAD && flush == Z_NO_FLUSH) {
|
||
|
return NeedMore;
|
||
|
}
|
||
|
if (lookahead === 0)
|
||
|
break; // flush the current block
|
||
|
}
|
||
|
|
||
|
// Insert the string window[strstart .. strstart+2] in the
|
||
|
// dictionary, and set hash_head to the head of the hash chain:
|
||
|
|
||
|
if (lookahead >= MIN_MATCH) {
|
||
|
ins_h = (((ins_h) << hash_shift) ^ (window[(strstart) + (MIN_MATCH - 1)] & 0xff)) & hash_mask;
|
||
|
// prev[strstart&w_mask]=hash_head=head[ins_h];
|
||
|
hash_head = (head[ins_h] & 0xffff);
|
||
|
prev[strstart & w_mask] = head[ins_h];
|
||
|
head[ins_h] = strstart;
|
||
|
}
|
||
|
|
||
|
// Find the longest match, discarding those <= prev_length.
|
||
|
prev_length = match_length;
|
||
|
prev_match = match_start;
|
||
|
match_length = MIN_MATCH - 1;
|
||
|
|
||
|
if (hash_head !== 0 && prev_length < max_lazy_match && ((strstart - hash_head) & 0xffff) <= w_size - MIN_LOOKAHEAD) {
|
||
|
// To simplify the code, we prevent matches with the string
|
||
|
// of window index 0 (in particular we have to avoid a match
|
||
|
// of the string with itself at the start of the input file).
|
||
|
|
||
|
if (strategy != Z_HUFFMAN_ONLY) {
|
||
|
match_length = longest_match(hash_head);
|
||
|
}
|
||
|
// longest_match() sets match_start
|
||
|
|
||
|
if (match_length <= 5 && (strategy == Z_FILTERED || (match_length == MIN_MATCH && strstart - match_start > 4096))) {
|
||
|
|
||
|
// If prev_match is also MIN_MATCH, match_start is garbage
|
||
|
// but we will ignore the current match anyway.
|
||
|
match_length = MIN_MATCH - 1;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// If there was a match at the previous step and the current
|
||
|
// match is not better, output the previous match:
|
||
|
if (prev_length >= MIN_MATCH && match_length <= prev_length) {
|
||
|
max_insert = strstart + lookahead - MIN_MATCH;
|
||
|
// Do not insert strings in hash table beyond this.
|
||
|
|
||
|
// check_match(strstart-1, prev_match, prev_length);
|
||
|
|
||
|
bflush = _tr_tally(strstart - 1 - prev_match, prev_length - MIN_MATCH);
|
||
|
|
||
|
// Insert in hash table all strings up to the end of the match.
|
||
|
// strstart-1 and strstart are already inserted. If there is not
|
||
|
// enough lookahead, the last two strings are not inserted in
|
||
|
// the hash table.
|
||
|
lookahead -= prev_length - 1;
|
||
|
prev_length -= 2;
|
||
|
do {
|
||
|
if (++strstart <= max_insert) {
|
||
|
ins_h = (((ins_h) << hash_shift) ^ (window[(strstart) + (MIN_MATCH - 1)] & 0xff)) & hash_mask;
|
||
|
// prev[strstart&w_mask]=hash_head=head[ins_h];
|
||
|
hash_head = (head[ins_h] & 0xffff);
|
||
|
prev[strstart & w_mask] = head[ins_h];
|
||
|
head[ins_h] = strstart;
|
||
|
}
|
||
|
} while (--prev_length !== 0);
|
||
|
match_available = 0;
|
||
|
match_length = MIN_MATCH - 1;
|
||
|
strstart++;
|
||
|
|
||
|
if (bflush) {
|
||
|
flush_block_only(false);
|
||
|
if (strm.avail_out === 0)
|
||
|
return NeedMore;
|
||
|
}
|
||
|
} else if (match_available !== 0) {
|
||
|
|
||
|
// If there was no match at the previous position, output a
|
||
|
// single literal. If there was a match but the current match
|
||
|
// is longer, truncate the previous match to a single literal.
|
||
|
|
||
|
bflush = _tr_tally(0, window[strstart - 1] & 0xff);
|
||
|
|
||
|
if (bflush) {
|
||
|
flush_block_only(false);
|
||
|
}
|
||
|
strstart++;
|
||
|
lookahead--;
|
||
|
if (strm.avail_out === 0)
|
||
|
return NeedMore;
|
||
|
} else {
|
||
|
// There is no previous match to compare with, wait for
|
||
|
// the next step to decide.
|
||
|
|
||
|
match_available = 1;
|
||
|
strstart++;
|
||
|
lookahead--;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if (match_available !== 0) {
|
||
|
bflush = _tr_tally(0, window[strstart - 1] & 0xff);
|
||
|
match_available = 0;
|
||
|
}
|
||
|
flush_block_only(flush == Z_FINISH);
|
||
|
|
||
|
if (strm.avail_out === 0) {
|
||
|
if (flush == Z_FINISH)
|
||
|
return FinishStarted;
|
||
|
else
|
||
|
return NeedMore;
|
||
|
}
|
||
|
|
||
|
return flush == Z_FINISH ? FinishDone : BlockDone;
|
||
|
}
|
||
|
|
||
|
function deflateReset(strm) {
|
||
|
strm.total_in = strm.total_out = 0;
|
||
|
strm.msg = null; //
|
||
|
|
||
|
that.pending = 0;
|
||
|
that.pending_out = 0;
|
||
|
|
||
|
status = BUSY_STATE;
|
||
|
|
||
|
last_flush = Z_NO_FLUSH;
|
||
|
|
||
|
tr_init();
|
||
|
lm_init();
|
||
|
return Z_OK;
|
||
|
}
|
||
|
|
||
|
that.deflateInit = function(strm, _level, bits, _method, memLevel, _strategy) {
|
||
|
if (!_method)
|
||
|
_method = Z_DEFLATED;
|
||
|
if (!memLevel)
|
||
|
memLevel = DEF_MEM_LEVEL;
|
||
|
if (!_strategy)
|
||
|
_strategy = Z_DEFAULT_STRATEGY;
|
||
|
|
||
|
// byte[] my_version=ZLIB_VERSION;
|
||
|
|
||
|
//
|
||
|
// if (!version || version[0] != my_version[0]
|
||
|
// || stream_size != sizeof(z_stream)) {
|
||
|
// return Z_VERSION_ERROR;
|
||
|
// }
|
||
|
|
||
|
strm.msg = null;
|
||
|
|
||
|
if (_level == Z_DEFAULT_COMPRESSION)
|
||
|
_level = 6;
|
||
|
|
||
|
if (memLevel < 1 || memLevel > MAX_MEM_LEVEL || _method != Z_DEFLATED || bits < 9 || bits > 15 || _level < 0 || _level > 9 || _strategy < 0
|
||
|
|| _strategy > Z_HUFFMAN_ONLY) {
|
||
|
return Z_STREAM_ERROR;
|
||
|
}
|
||
|
|
||
|
strm.dstate = that;
|
||
|
|
||
|
w_bits = bits;
|
||
|
w_size = 1 << w_bits;
|
||
|
w_mask = w_size - 1;
|
||
|
|
||
|
hash_bits = memLevel + 7;
|
||
|
hash_size = 1 << hash_bits;
|
||
|
hash_mask = hash_size - 1;
|
||
|
hash_shift = Math.floor((hash_bits + MIN_MATCH - 1) / MIN_MATCH);
|
||
|
|
||
|
window = new Uint8Array(w_size * 2);
|
||
|
prev = [];
|
||
|
head = [];
|
||
|
|
||
|
lit_bufsize = 1 << (memLevel + 6); // 16K elements by default
|
||
|
|
||
|
// We overlay pending_buf and d_buf+l_buf. This works since the average
|
||
|
// output size for (length,distance) codes is <= 24 bits.
|
||
|
that.pending_buf = new Uint8Array(lit_bufsize * 4);
|
||
|
pending_buf_size = lit_bufsize * 4;
|
||
|
|
||
|
d_buf = Math.floor(lit_bufsize / 2);
|
||
|
l_buf = (1 + 2) * lit_bufsize;
|
||
|
|
||
|
level = _level;
|
||
|
|
||
|
strategy = _strategy;
|
||
|
method = _method & 0xff;
|
||
|
|
||
|
return deflateReset(strm);
|
||
|
};
|
||
|
|
||
|
that.deflateEnd = function() {
|
||
|
if (status != INIT_STATE && status != BUSY_STATE && status != FINISH_STATE) {
|
||
|
return Z_STREAM_ERROR;
|
||
|
}
|
||
|
// Deallocate in reverse order of allocations:
|
||
|
that.pending_buf = null;
|
||
|
head = null;
|
||
|
prev = null;
|
||
|
window = null;
|
||
|
// free
|
||
|
that.dstate = null;
|
||
|
return status == BUSY_STATE ? Z_DATA_ERROR : Z_OK;
|
||
|
};
|
||
|
|
||
|
that.deflateParams = function(strm, _level, _strategy) {
|
||
|
var err = Z_OK;
|
||
|
|
||
|
if (_level == Z_DEFAULT_COMPRESSION) {
|
||
|
_level = 6;
|
||
|
}
|
||
|
if (_level < 0 || _level > 9 || _strategy < 0 || _strategy > Z_HUFFMAN_ONLY) {
|
||
|
return Z_STREAM_ERROR;
|
||
|
}
|
||
|
|
||
|
if (config_table[level].func != config_table[_level].func && strm.total_in !== 0) {
|
||
|
// Flush the last buffer:
|
||
|
err = strm.deflate(Z_PARTIAL_FLUSH);
|
||
|
}
|
||
|
|
||
|
if (level != _level) {
|
||
|
level = _level;
|
||
|
max_lazy_match = config_table[level].max_lazy;
|
||
|
good_match = config_table[level].good_length;
|
||
|
nice_match = config_table[level].nice_length;
|
||
|
max_chain_length = config_table[level].max_chain;
|
||
|
}
|
||
|
strategy = _strategy;
|
||
|
return err;
|
||
|
};
|
||
|
|
||
|
that.deflateSetDictionary = function(strm, dictionary, dictLength) {
|
||
|
var length = dictLength;
|
||
|
var n, index = 0;
|
||
|
|
||
|
if (!dictionary || status != INIT_STATE)
|
||
|
return Z_STREAM_ERROR;
|
||
|
|
||
|
if (length < MIN_MATCH)
|
||
|
return Z_OK;
|
||
|
if (length > w_size - MIN_LOOKAHEAD) {
|
||
|
length = w_size - MIN_LOOKAHEAD;
|
||
|
index = dictLength - length; // use the tail of the dictionary
|
||
|
}
|
||
|
window.set(dictionary.subarray(index, index + length), 0);
|
||
|
|
||
|
strstart = length;
|
||
|
block_start = length;
|
||
|
|
||
|
// Insert all strings in the hash table (except for the last two bytes).
|
||
|
// s->lookahead stays null, so s->ins_h will be recomputed at the next
|
||
|
// call of fill_window.
|
||
|
|
||
|
ins_h = window[0] & 0xff;
|
||
|
ins_h = (((ins_h) << hash_shift) ^ (window[1] & 0xff)) & hash_mask;
|
||
|
|
||
|
for (n = 0; n <= length - MIN_MATCH; n++) {
|
||
|
ins_h = (((ins_h) << hash_shift) ^ (window[(n) + (MIN_MATCH - 1)] & 0xff)) & hash_mask;
|
||
|
prev[n & w_mask] = head[ins_h];
|
||
|
head[ins_h] = n;
|
||
|
}
|
||
|
return Z_OK;
|
||
|
};
|
||
|
|
||
|
that.deflate = function(_strm, flush) {
|
||
|
var i, header, level_flags, old_flush, bstate;
|
||
|
|
||
|
if (flush > Z_FINISH || flush < 0) {
|
||
|
return Z_STREAM_ERROR;
|
||
|
}
|
||
|
|
||
|
if (!_strm.next_out || (!_strm.next_in && _strm.avail_in !== 0) || (status == FINISH_STATE && flush != Z_FINISH)) {
|
||
|
_strm.msg = z_errmsg[Z_NEED_DICT - (Z_STREAM_ERROR)];
|
||
|
return Z_STREAM_ERROR;
|
||
|
}
|
||
|
if (_strm.avail_out === 0) {
|
||
|
_strm.msg = z_errmsg[Z_NEED_DICT - (Z_BUF_ERROR)];
|
||
|
return Z_BUF_ERROR;
|
||
|
}
|
||
|
|
||
|
strm = _strm; // just in case
|
||
|
old_flush = last_flush;
|
||
|
last_flush = flush;
|
||
|
|
||
|
// Write the zlib header
|
||
|
if (status == INIT_STATE) {
|
||
|
header = (Z_DEFLATED + ((w_bits - 8) << 4)) << 8;
|
||
|
level_flags = ((level - 1) & 0xff) >> 1;
|
||
|
|
||
|
if (level_flags > 3)
|
||
|
level_flags = 3;
|
||
|
header |= (level_flags << 6);
|
||
|
if (strstart !== 0)
|
||
|
header |= PRESET_DICT;
|
||
|
header += 31 - (header % 31);
|
||
|
|
||
|
status = BUSY_STATE;
|
||
|
putShortMSB(header);
|
||
|
}
|
||
|
|
||
|
// Flush as much pending output as possible
|
||
|
if (that.pending !== 0) {
|
||
|
strm.flush_pending();
|
||
|
if (strm.avail_out === 0) {
|
||
|
// console.log(" avail_out==0");
|
||
|
// Since avail_out is 0, deflate will be called again with
|
||
|
// more output space, but possibly with both pending and
|
||
|
// avail_in equal to zero. There won't be anything to do,
|
||
|
// but this is not an error situation so make sure we
|
||
|
// return OK instead of BUF_ERROR at next call of deflate:
|
||
|
last_flush = -1;
|
||
|
return Z_OK;
|
||
|
}
|
||
|
|
||
|
// Make sure there is something to do and avoid duplicate
|
||
|
// consecutive
|
||
|
// flushes. For repeated and useless calls with Z_FINISH, we keep
|
||
|
// returning Z_STREAM_END instead of Z_BUFF_ERROR.
|
||
|
} else if (strm.avail_in === 0 && flush <= old_flush && flush != Z_FINISH) {
|
||
|
strm.msg = z_errmsg[Z_NEED_DICT - (Z_BUF_ERROR)];
|
||
|
return Z_BUF_ERROR;
|
||
|
}
|
||
|
|
||
|
// User must not provide more input after the first FINISH:
|
||
|
if (status == FINISH_STATE && strm.avail_in !== 0) {
|
||
|
_strm.msg = z_errmsg[Z_NEED_DICT - (Z_BUF_ERROR)];
|
||
|
return Z_BUF_ERROR;
|
||
|
}
|
||
|
|
||
|
// Start a new block or continue the current one.
|
||
|
if (strm.avail_in !== 0 || lookahead !== 0 || (flush != Z_NO_FLUSH && status != FINISH_STATE)) {
|
||
|
bstate = -1;
|
||
|
switch (config_table[level].func) {
|
||
|
case STORED:
|
||
|
bstate = deflate_stored(flush);
|
||
|
break;
|
||
|
case FAST:
|
||
|
bstate = deflate_fast(flush);
|
||
|
break;
|
||
|
case SLOW:
|
||
|
bstate = deflate_slow(flush);
|
||
|
break;
|
||
|
default:
|
||
|
}
|
||
|
|
||
|
if (bstate == FinishStarted || bstate == FinishDone) {
|
||
|
status = FINISH_STATE;
|
||
|
}
|
||
|
if (bstate == NeedMore || bstate == FinishStarted) {
|
||
|
if (strm.avail_out === 0) {
|
||
|
last_flush = -1; // avoid BUF_ERROR next call, see above
|
||
|
}
|
||
|
return Z_OK;
|
||
|
// If flush != Z_NO_FLUSH && avail_out === 0, the next call
|
||
|
// of deflate should use the same flush parameter to make sure
|
||
|
// that the flush is complete. So we don't have to output an
|
||
|
// empty block here, this will be done at next call. This also
|
||
|
// ensures that for a very small output buffer, we emit at most
|
||
|
// one empty block.
|
||
|
}
|
||
|
|
||
|
if (bstate == BlockDone) {
|
||
|
if (flush == Z_PARTIAL_FLUSH) {
|
||
|
_tr_align();
|
||
|
} else { // FULL_FLUSH or SYNC_FLUSH
|
||
|
_tr_stored_block(0, 0, false);
|
||
|
// For a full flush, this empty block will be recognized
|
||
|
// as a special marker by inflate_sync().
|
||
|
if (flush == Z_FULL_FLUSH) {
|
||
|
// state.head[s.hash_size-1]=0;
|
||
|
for (i = 0; i < hash_size/*-1*/; i++)
|
||
|
// forget history
|
||
|
head[i] = 0;
|
||
|
}
|
||
|
}
|
||
|
strm.flush_pending();
|
||
|
if (strm.avail_out === 0) {
|
||
|
last_flush = -1; // avoid BUF_ERROR at next call, see above
|
||
|
return Z_OK;
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if (flush != Z_FINISH)
|
||
|
return Z_OK;
|
||
|
return Z_STREAM_END;
|
||
|
};
|
||
|
}
|
||
|
|
||
|
// ZStream
|
||
|
|
||
|
function ZStream() {
|
||
|
var that = this;
|
||
|
that.next_in_index = 0;
|
||
|
that.next_out_index = 0;
|
||
|
// that.next_in; // next input byte
|
||
|
that.avail_in = 0; // number of bytes available at next_in
|
||
|
that.total_in = 0; // total nb of input bytes read so far
|
||
|
// that.next_out; // next output byte should be put there
|
||
|
that.avail_out = 0; // remaining free space at next_out
|
||
|
that.total_out = 0; // total nb of bytes output so far
|
||
|
// that.msg;
|
||
|
// that.dstate;
|
||
|
}
|
||
|
|
||
|
ZStream.prototype = {
|
||
|
deflateInit : function(level, bits) {
|
||
|
var that = this;
|
||
|
that.dstate = new Deflate();
|
||
|
if (!bits)
|
||
|
bits = MAX_BITS;
|
||
|
return that.dstate.deflateInit(that, level, bits);
|
||
|
},
|
||
|
|
||
|
deflate : function(flush) {
|
||
|
var that = this;
|
||
|
if (!that.dstate) {
|
||
|
return Z_STREAM_ERROR;
|
||
|
}
|
||
|
return that.dstate.deflate(that, flush);
|
||
|
},
|
||
|
|
||
|
deflateEnd : function() {
|
||
|
var that = this;
|
||
|
if (!that.dstate)
|
||
|
return Z_STREAM_ERROR;
|
||
|
var ret = that.dstate.deflateEnd();
|
||
|
that.dstate = null;
|
||
|
return ret;
|
||
|
},
|
||
|
|
||
|
deflateParams : function(level, strategy) {
|
||
|
var that = this;
|
||
|
if (!that.dstate)
|
||
|
return Z_STREAM_ERROR;
|
||
|
return that.dstate.deflateParams(that, level, strategy);
|
||
|
},
|
||
|
|
||
|
deflateSetDictionary : function(dictionary, dictLength) {
|
||
|
var that = this;
|
||
|
if (!that.dstate)
|
||
|
return Z_STREAM_ERROR;
|
||
|
return that.dstate.deflateSetDictionary(that, dictionary, dictLength);
|
||
|
},
|
||
|
|
||
|
// Read a new buffer from the current input stream, update the
|
||
|
// total number of bytes read. All deflate() input goes through
|
||
|
// this function so some applications may wish to modify it to avoid
|
||
|
// allocating a large strm->next_in buffer and copying from it.
|
||
|
// (See also flush_pending()).
|
||
|
read_buf : function(buf, start, size) {
|
||
|
var that = this;
|
||
|
var len = that.avail_in;
|
||
|
if (len > size)
|
||
|
len = size;
|
||
|
if (len === 0)
|
||
|
return 0;
|
||
|
that.avail_in -= len;
|
||
|
buf.set(that.next_in.subarray(that.next_in_index, that.next_in_index + len), start);
|
||
|
that.next_in_index += len;
|
||
|
that.total_in += len;
|
||
|
return len;
|
||
|
},
|
||
|
|
||
|
// Flush as much pending output as possible. All deflate() output goes
|
||
|
// through this function so some applications may wish to modify it
|
||
|
// to avoid allocating a large strm->next_out buffer and copying into it.
|
||
|
// (See also read_buf()).
|
||
|
flush_pending : function() {
|
||
|
var that = this;
|
||
|
var len = that.dstate.pending;
|
||
|
|
||
|
if (len > that.avail_out)
|
||
|
len = that.avail_out;
|
||
|
if (len === 0)
|
||
|
return;
|
||
|
|
||
|
// if (that.dstate.pending_buf.length <= that.dstate.pending_out || that.next_out.length <= that.next_out_index
|
||
|
// || that.dstate.pending_buf.length < (that.dstate.pending_out + len) || that.next_out.length < (that.next_out_index +
|
||
|
// len)) {
|
||
|
// console.log(that.dstate.pending_buf.length + ", " + that.dstate.pending_out + ", " + that.next_out.length + ", " +
|
||
|
// that.next_out_index + ", " + len);
|
||
|
// console.log("avail_out=" + that.avail_out);
|
||
|
// }
|
||
|
|
||
|
that.next_out.set(that.dstate.pending_buf.subarray(that.dstate.pending_out, that.dstate.pending_out + len), that.next_out_index);
|
||
|
|
||
|
that.next_out_index += len;
|
||
|
that.dstate.pending_out += len;
|
||
|
that.total_out += len;
|
||
|
that.avail_out -= len;
|
||
|
that.dstate.pending -= len;
|
||
|
if (that.dstate.pending === 0) {
|
||
|
that.dstate.pending_out = 0;
|
||
|
}
|
||
|
}
|
||
|
};
|
||
|
|
||
|
// Deflater
|
||
|
|
||
|
function Deflater(level) {
|
||
|
var that = this;
|
||
|
var z = new ZStream();
|
||
|
var bufsize = 512;
|
||
|
var flush = Z_NO_FLUSH;
|
||
|
var buf = new Uint8Array(bufsize);
|
||
|
|
||
|
if (typeof level == "undefined")
|
||
|
level = Z_DEFAULT_COMPRESSION;
|
||
|
z.deflateInit(level);
|
||
|
z.next_out = buf;
|
||
|
|
||
|
that.append = function(data, onprogress) {
|
||
|
var err, buffers = [], lastIndex = 0, bufferIndex = 0, bufferSize = 0, array;
|
||
|
if (!data.length)
|
||
|
return;
|
||
|
z.next_in_index = 0;
|
||
|
z.next_in = data;
|
||
|
z.avail_in = data.length;
|
||
|
do {
|
||
|
z.next_out_index = 0;
|
||
|
z.avail_out = bufsize;
|
||
|
err = z.deflate(flush);
|
||
|
if (err != Z_OK)
|
||
|
throw "deflating: " + z.msg;
|
||
|
if (z.next_out_index)
|
||
|
if (z.next_out_index == bufsize)
|
||
|
buffers.push(new Uint8Array(buf));
|
||
|
else
|
||
|
buffers.push(new Uint8Array(buf.subarray(0, z.next_out_index)));
|
||
|
bufferSize += z.next_out_index;
|
||
|
if (onprogress && z.next_in_index > 0 && z.next_in_index != lastIndex) {
|
||
|
onprogress(z.next_in_index);
|
||
|
lastIndex = z.next_in_index;
|
||
|
}
|
||
|
} while (z.avail_in > 0 || z.avail_out === 0);
|
||
|
array = new Uint8Array(bufferSize);
|
||
|
buffers.forEach(function(chunk) {
|
||
|
array.set(chunk, bufferIndex);
|
||
|
bufferIndex += chunk.length;
|
||
|
});
|
||
|
return array;
|
||
|
};
|
||
|
that.flush = function() {
|
||
|
var err, buffers = [], bufferIndex = 0, bufferSize = 0, array;
|
||
|
do {
|
||
|
z.next_out_index = 0;
|
||
|
z.avail_out = bufsize;
|
||
|
err = z.deflate(Z_FINISH);
|
||
|
if (err != Z_STREAM_END && err != Z_OK)
|
||
|
throw "deflating: " + z.msg;
|
||
|
if (bufsize - z.avail_out > 0)
|
||
|
buffers.push(new Uint8Array(buf.subarray(0, z.next_out_index)));
|
||
|
bufferSize += z.next_out_index;
|
||
|
} while (z.avail_in > 0 || z.avail_out === 0);
|
||
|
z.deflateEnd();
|
||
|
array = new Uint8Array(bufferSize);
|
||
|
buffers.forEach(function(chunk) {
|
||
|
array.set(chunk, bufferIndex);
|
||
|
bufferIndex += chunk.length;
|
||
|
});
|
||
|
return array;
|
||
|
};
|
||
|
}/*
|
||
|
* Copyright (c) 2012 chick307 <chick307@gmail.com>
|
||
|
*
|
||
|
* Licensed under the MIT License.
|
||
|
* http://opensource.org/licenses/mit-license
|
||
|
*/
|
||
|
|
||
|
void function(global, callback) {
|
||
|
if (typeof module === 'object') {
|
||
|
module.exports = callback();
|
||
|
} else if (typeof define === 'function') {
|
||
|
define(callback);
|
||
|
} else {
|
||
|
global.adler32cs = callback();
|
||
|
}
|
||
|
}(this, function() {
|
||
|
var _hasArrayBuffer = typeof ArrayBuffer === 'function' &&
|
||
|
typeof Uint8Array === 'function';
|
||
|
|
||
|
var _Buffer = null, _isBuffer = (function() {
|
||
|
if (!_hasArrayBuffer)
|
||
|
return function _isBuffer() { return false };
|
||
|
|
||
|
try {
|
||
|
var buffer = require('buffer');
|
||
|
if (typeof buffer.Buffer === 'function')
|
||
|
_Buffer = buffer.Buffer;
|
||
|
} catch (error) {}
|
||
|
|
||
|
return function _isBuffer(value) {
|
||
|
return value instanceof ArrayBuffer ||
|
||
|
_Buffer !== null && value instanceof _Buffer;
|
||
|
};
|
||
|
}());
|
||
|
|
||
|
var _utf8ToBinary = (function() {
|
||
|
if (_Buffer !== null) {
|
||
|
return function _utf8ToBinary(utf8String) {
|
||
|
return new _Buffer(utf8String, 'utf8').toString('binary');
|
||
|
};
|
||
|
} else {
|
||
|
return function _utf8ToBinary(utf8String) {
|
||
|
return unescape(encodeURIComponent(utf8String));
|
||
|
};
|
||
|
}
|
||
|
}());
|
||
|
|
||
|
var MOD = 65521;
|
||
|
|
||
|
var _update = function _update(checksum, binaryString) {
|
||
|
var a = checksum & 0xFFFF, b = checksum >>> 16;
|
||
|
for (var i = 0, length = binaryString.length; i < length; i++) {
|
||
|
a = (a + (binaryString.charCodeAt(i) & 0xFF)) % MOD;
|
||
|
b = (b + a) % MOD;
|
||
|
}
|
||
|
return (b << 16 | a) >>> 0;
|
||
|
};
|
||
|
|
||
|
var _updateUint8Array = function _updateUint8Array(checksum, uint8Array) {
|
||
|
var a = checksum & 0xFFFF, b = checksum >>> 16;
|
||
|
for (var i = 0, length = uint8Array.length, x; i < length; i++) {
|
||
|
a = (a + uint8Array[i]) % MOD;
|
||
|
b = (b + a) % MOD;
|
||
|
}
|
||
|
return (b << 16 | a) >>> 0
|
||
|
};
|
||
|
|
||
|
var exports = {};
|
||
|
|
||
|
var Adler32 = exports.Adler32 = (function() {
|
||
|
var ctor = function Adler32(checksum) {
|
||
|
if (!(this instanceof ctor)) {
|
||
|
throw new TypeError(
|
||
|
'Constructor cannot called be as a function.');
|
||
|
}
|
||
|
if (!isFinite(checksum = checksum == null ? 1 : +checksum)) {
|
||
|
throw new Error(
|
||
|
'First arguments needs to be a finite number.');
|
||
|
}
|
||
|
this.checksum = checksum >>> 0;
|
||
|
};
|
||
|
|
||
|
var proto = ctor.prototype = {};
|
||
|
proto.constructor = ctor;
|
||
|
|
||
|
ctor.from = function(from) {
|
||
|
from.prototype = proto;
|
||
|
return from;
|
||
|
}(function from(binaryString) {
|
||
|
if (!(this instanceof ctor)) {
|
||
|
throw new TypeError(
|
||
|
'Constructor cannot called be as a function.');
|
||
|
}
|
||
|
if (binaryString == null)
|
||
|
throw new Error('First argument needs to be a string.');
|
||
|
this.checksum = _update(1, binaryString.toString());
|
||
|
});
|
||
|
|
||
|
ctor.fromUtf8 = function(fromUtf8) {
|
||
|
fromUtf8.prototype = proto;
|
||
|
return fromUtf8;
|
||
|
}(function fromUtf8(utf8String) {
|
||
|
if (!(this instanceof ctor)) {
|
||
|
throw new TypeError(
|
||
|
'Constructor cannot called be as a function.');
|
||
|
}
|
||
|
if (utf8String == null)
|
||
|
throw new Error('First argument needs to be a string.');
|
||
|
var binaryString = _utf8ToBinary(utf8String.toString());
|
||
|
this.checksum = _update(1, binaryString);
|
||
|
});
|
||
|
|
||
|
if (_hasArrayBuffer) {
|
||
|
ctor.fromBuffer = function(fromBuffer) {
|
||
|
fromBuffer.prototype = proto;
|
||
|
return fromBuffer;
|
||
|
}(function fromBuffer(buffer) {
|
||
|
if (!(this instanceof ctor)) {
|
||
|
throw new TypeError(
|
||
|
'Constructor cannot called be as a function.');
|
||
|
}
|
||
|
if (!_isBuffer(buffer))
|
||
|
throw new Error('First argument needs to be ArrayBuffer.');
|
||
|
var array = new Uint8Array(buffer);
|
||
|
return this.checksum = _updateUint8Array(1, array);
|
||
|
});
|
||
|
}
|
||
|
|
||
|
proto.update = function update(binaryString) {
|
||
|
if (binaryString == null)
|
||
|
throw new Error('First argument needs to be a string.');
|
||
|
binaryString = binaryString.toString();
|
||
|
return this.checksum = _update(this.checksum, binaryString);
|
||
|
};
|
||
|
|
||
|
proto.updateUtf8 = function updateUtf8(utf8String) {
|
||
|
if (utf8String == null)
|
||
|
throw new Error('First argument needs to be a string.');
|
||
|
var binaryString = _utf8ToBinary(utf8String.toString());
|
||
|
return this.checksum = _update(this.checksum, binaryString);
|
||
|
};
|
||
|
|
||
|
if (_hasArrayBuffer) {
|
||
|
proto.updateBuffer = function updateBuffer(buffer) {
|
||
|
if (!_isBuffer(buffer))
|
||
|
throw new Error('First argument needs to be ArrayBuffer.');
|
||
|
var array = new Uint8Array(buffer);
|
||
|
return this.checksum = _updateUint8Array(this.checksum, array);
|
||
|
};
|
||
|
}
|
||
|
|
||
|
proto.clone = function clone() {
|
||
|
return new Adler32(this.checksum);
|
||
|
};
|
||
|
|
||
|
return ctor;
|
||
|
}());
|
||
|
|
||
|
exports.from = function from(binaryString) {
|
||
|
if (binaryString == null)
|
||
|
throw new Error('First argument needs to be a string.');
|
||
|
return _update(1, binaryString.toString());
|
||
|
};
|
||
|
|
||
|
exports.fromUtf8 = function fromUtf8(utf8String) {
|
||
|
if (utf8String == null)
|
||
|
throw new Error('First argument needs to be a string.');
|
||
|
var binaryString = _utf8ToBinary(utf8String.toString());
|
||
|
return _update(1, binaryString);
|
||
|
};
|
||
|
|
||
|
if (_hasArrayBuffer) {
|
||
|
exports.fromBuffer = function fromBuffer(buffer) {
|
||
|
if (!_isBuffer(buffer))
|
||
|
throw new Error('First argument need to be ArrayBuffer.');
|
||
|
var array = new Uint8Array(buffer);
|
||
|
return _updateUint8Array(1, array);
|
||
|
};
|
||
|
}
|
||
|
|
||
|
return exports;
|
||
|
});
|