626 lines
25 KiB
JavaScript
626 lines
25 KiB
JavaScript
var fs = require("fs");
|
|
var zlib = require("zlib");
|
|
var fd_slicer = require("fd-slicer");
|
|
var util = require("util");
|
|
var EventEmitter = require("events").EventEmitter;
|
|
var Transform = require("stream").Transform;
|
|
var PassThrough = require("stream").PassThrough;
|
|
var Writable = require("stream").Writable;
|
|
|
|
exports.open = open;
|
|
exports.fromFd = fromFd;
|
|
exports.fromBuffer = fromBuffer;
|
|
exports.fromRandomAccessReader = fromRandomAccessReader;
|
|
exports.dosDateTimeToDate = dosDateTimeToDate;
|
|
exports.ZipFile = ZipFile;
|
|
exports.Entry = Entry;
|
|
exports.RandomAccessReader = RandomAccessReader;
|
|
|
|
function open(path, options, callback) {
|
|
if (typeof options === "function") {
|
|
callback = options;
|
|
options = null;
|
|
}
|
|
if (options == null) options = {};
|
|
if (options.autoClose == null) options.autoClose = true;
|
|
if (options.lazyEntries == null) options.lazyEntries = false;
|
|
if (callback == null) callback = defaultCallback;
|
|
fs.open(path, "r", function(err, fd) {
|
|
if (err) return callback(err);
|
|
fromFd(fd, options, function(err, zipfile) {
|
|
if (err) fs.close(fd, defaultCallback);
|
|
callback(err, zipfile);
|
|
});
|
|
});
|
|
}
|
|
|
|
function fromFd(fd, options, callback) {
|
|
if (typeof options === "function") {
|
|
callback = options;
|
|
options = null;
|
|
}
|
|
if (options == null) options = {};
|
|
if (options.autoClose == null) options.autoClose = false;
|
|
if (options.lazyEntries == null) options.lazyEntries = false;
|
|
if (callback == null) callback = defaultCallback;
|
|
fs.fstat(fd, function(err, stats) {
|
|
if (err) return callback(err);
|
|
var reader = fd_slicer.createFromFd(fd, {autoClose: true});
|
|
fromRandomAccessReader(reader, stats.size, options, callback);
|
|
});
|
|
}
|
|
|
|
function fromBuffer(buffer, options, callback) {
|
|
if (typeof options === "function") {
|
|
callback = options;
|
|
options = null;
|
|
}
|
|
if (options == null) options = {};
|
|
options.autoClose = false;
|
|
if (options.lazyEntries == null) options.lazyEntries = false;
|
|
// i got your open file right here.
|
|
var reader = fd_slicer.createFromBuffer(buffer);
|
|
fromRandomAccessReader(reader, buffer.length, options, callback);
|
|
}
|
|
|
|
function fromRandomAccessReader(reader, totalSize, options, callback) {
|
|
if (typeof options === "function") {
|
|
callback = options;
|
|
options = null;
|
|
}
|
|
if (options == null) options = {};
|
|
if (options.autoClose == null) options.autoClose = true;
|
|
if (options.lazyEntries == null) options.lazyEntries = false;
|
|
if (callback == null) callback = defaultCallback;
|
|
if (typeof totalSize !== "number") throw new Error("expected totalSize parameter to be a number");
|
|
if (totalSize > Number.MAX_SAFE_INTEGER) {
|
|
throw new Error("zip file too large. only file sizes up to 2^52 are supported due to JavaScript's Number type being an IEEE 754 double.");
|
|
}
|
|
|
|
// the matching unref() call is in zipfile.close()
|
|
reader.ref();
|
|
|
|
// eocdr means End of Central Directory Record.
|
|
// search backwards for the eocdr signature.
|
|
// the last field of the eocdr is a variable-length comment.
|
|
// the comment size is encoded in a 2-byte field in the eocdr, which we can't find without trudging backwards through the comment to find it.
|
|
// as a consequence of this design decision, it's possible to have ambiguous zip file metadata if a coherent eocdr was in the comment.
|
|
// we search backwards for a eocdr signature, and hope that whoever made the zip file was smart enough to forbid the eocdr signature in the comment.
|
|
var eocdrWithoutCommentSize = 22;
|
|
var maxCommentSize = 0x10000; // 2-byte size
|
|
var bufferSize = Math.min(eocdrWithoutCommentSize + maxCommentSize, totalSize);
|
|
var buffer = new Buffer(bufferSize);
|
|
var bufferReadStart = totalSize - buffer.length;
|
|
readAndAssertNoEof(reader, buffer, 0, bufferSize, bufferReadStart, function(err) {
|
|
if (err) return callback(err);
|
|
for (var i = bufferSize - eocdrWithoutCommentSize; i >= 0; i -= 1) {
|
|
if (buffer.readUInt32LE(i) !== 0x06054b50) continue;
|
|
// found eocdr
|
|
var eocdrBuffer = buffer.slice(i);
|
|
|
|
// 0 - End of central directory signature = 0x06054b50
|
|
// 4 - Number of this disk
|
|
var diskNumber = eocdrBuffer.readUInt16LE(4);
|
|
if (diskNumber !== 0) return callback(new Error("multi-disk zip files are not supported: found disk number: " + diskNumber));
|
|
// 6 - Disk where central directory starts
|
|
// 8 - Number of central directory records on this disk
|
|
// 10 - Total number of central directory records
|
|
var entryCount = eocdrBuffer.readUInt16LE(10);
|
|
// 12 - Size of central directory (bytes)
|
|
// 16 - Offset of start of central directory, relative to start of archive
|
|
var centralDirectoryOffset = eocdrBuffer.readUInt32LE(16);
|
|
// 20 - Comment length
|
|
var commentLength = eocdrBuffer.readUInt16LE(20);
|
|
var expectedCommentLength = eocdrBuffer.length - eocdrWithoutCommentSize;
|
|
if (commentLength !== expectedCommentLength) {
|
|
return callback(new Error("invalid comment length. expected: " + expectedCommentLength + ". found: " + commentLength));
|
|
}
|
|
// 22 - Comment
|
|
// the encoding is always cp437.
|
|
var comment = bufferToString(eocdrBuffer, 22, eocdrBuffer.length, false);
|
|
|
|
if (!(entryCount === 0xffff || centralDirectoryOffset === 0xffffffff)) {
|
|
return callback(null, new ZipFile(reader, centralDirectoryOffset, totalSize, entryCount, comment, options.autoClose, options.lazyEntries));
|
|
}
|
|
|
|
// ZIP64 format
|
|
|
|
// ZIP64 Zip64 end of central directory locator
|
|
var zip64EocdlBuffer = new Buffer(20);
|
|
var zip64EocdlOffset = bufferReadStart + i - zip64EocdlBuffer.length;
|
|
readAndAssertNoEof(reader, zip64EocdlBuffer, 0, zip64EocdlBuffer.length, zip64EocdlOffset, function(err) {
|
|
if (err) return callback(err);
|
|
|
|
// 0 - zip64 end of central dir locator signature = 0x07064b50
|
|
if (zip64EocdlBuffer.readUInt32LE(0) !== 0x07064b50) {
|
|
return callback(new Error("invalid ZIP64 End of Central Directory Locator signature"));
|
|
}
|
|
// 4 - number of the disk with the start of the zip64 end of central directory
|
|
// 8 - relative offset of the zip64 end of central directory record
|
|
var zip64EocdrOffset = readUInt64LE(zip64EocdlBuffer, 8);
|
|
// 16 - total number of disks
|
|
|
|
// ZIP64 end of central directory record
|
|
var zip64EocdrBuffer = new Buffer(56);
|
|
readAndAssertNoEof(reader, zip64EocdrBuffer, 0, zip64EocdrBuffer.length, zip64EocdrOffset, function(err) {
|
|
if (err) return callback(err);
|
|
|
|
// 0 - zip64 end of central dir signature 4 bytes (0x06064b50)
|
|
if (zip64EocdrBuffer.readUInt32LE(0) !== 0x06064b50) return callback(new Error("invalid ZIP64 end of central directory record signature"));
|
|
// 4 - size of zip64 end of central directory record 8 bytes
|
|
// 12 - version made by 2 bytes
|
|
// 14 - version needed to extract 2 bytes
|
|
// 16 - number of this disk 4 bytes
|
|
// 20 - number of the disk with the start of the central directory 4 bytes
|
|
// 24 - total number of entries in the central directory on this disk 8 bytes
|
|
// 32 - total number of entries in the central directory 8 bytes
|
|
entryCount = readUInt64LE(zip64EocdrBuffer, 32);
|
|
// 40 - size of the central directory 8 bytes
|
|
// 48 - offset of start of central directory with respect to the starting disk number 8 bytes
|
|
centralDirectoryOffset = readUInt64LE(zip64EocdrBuffer, 48);
|
|
// 56 - zip64 extensible data sector (variable size)
|
|
return callback(null, new ZipFile(reader, centralDirectoryOffset, totalSize, entryCount, comment, options.autoClose, options.lazyEntries));
|
|
});
|
|
});
|
|
return;
|
|
}
|
|
callback(new Error("end of central directory record signature not found"));
|
|
});
|
|
}
|
|
|
|
util.inherits(ZipFile, EventEmitter);
|
|
function ZipFile(reader, centralDirectoryOffset, fileSize, entryCount, comment, autoClose, lazyEntries) {
|
|
var self = this;
|
|
EventEmitter.call(self);
|
|
self.reader = reader;
|
|
// forward close events
|
|
self.reader.on("error", function(err) {
|
|
// error closing the fd
|
|
emitError(self, err);
|
|
});
|
|
self.reader.once("close", function() {
|
|
self.emit("close");
|
|
});
|
|
self.readEntryCursor = centralDirectoryOffset;
|
|
self.fileSize = fileSize;
|
|
self.entryCount = entryCount;
|
|
self.comment = comment;
|
|
self.entriesRead = 0;
|
|
self.autoClose = !!autoClose;
|
|
self.lazyEntries = !!lazyEntries;
|
|
self.isOpen = true;
|
|
self.emittedError = false;
|
|
|
|
if (!self.lazyEntries) self.readEntry();
|
|
}
|
|
ZipFile.prototype.close = function() {
|
|
if (!this.isOpen) return;
|
|
this.isOpen = false;
|
|
this.reader.unref();
|
|
};
|
|
|
|
function emitErrorAndAutoClose(self, err) {
|
|
if (self.autoClose) self.close();
|
|
emitError(self, err);
|
|
}
|
|
function emitError(self, err) {
|
|
if (self.emittedError) return;
|
|
self.emittedError = true;
|
|
self.emit("error", err);
|
|
}
|
|
|
|
ZipFile.prototype.readEntry = function() {
|
|
var self = this;
|
|
if (self.entryCount === self.entriesRead) {
|
|
// done with metadata
|
|
setImmediate(function() {
|
|
if (self.autoClose) self.close();
|
|
if (self.emittedError) return;
|
|
self.emit("end");
|
|
});
|
|
return;
|
|
}
|
|
if (self.emittedError) return;
|
|
var buffer = new Buffer(46);
|
|
readAndAssertNoEof(self.reader, buffer, 0, buffer.length, self.readEntryCursor, function(err) {
|
|
if (err) return emitErrorAndAutoClose(self, err);
|
|
if (self.emittedError) return;
|
|
var entry = new Entry();
|
|
// 0 - Central directory file header signature
|
|
var signature = buffer.readUInt32LE(0);
|
|
if (signature !== 0x02014b50) return emitErrorAndAutoClose(self, new Error("invalid central directory file header signature: 0x" + signature.toString(16)));
|
|
// 4 - Version made by
|
|
entry.versionMadeBy = buffer.readUInt16LE(4);
|
|
// 6 - Version needed to extract (minimum)
|
|
entry.versionNeededToExtract = buffer.readUInt16LE(6);
|
|
// 8 - General purpose bit flag
|
|
entry.generalPurposeBitFlag = buffer.readUInt16LE(8);
|
|
// 10 - Compression method
|
|
entry.compressionMethod = buffer.readUInt16LE(10);
|
|
// 12 - File last modification time
|
|
entry.lastModFileTime = buffer.readUInt16LE(12);
|
|
// 14 - File last modification date
|
|
entry.lastModFileDate = buffer.readUInt16LE(14);
|
|
// 16 - CRC-32
|
|
entry.crc32 = buffer.readUInt32LE(16);
|
|
// 20 - Compressed size
|
|
entry.compressedSize = buffer.readUInt32LE(20);
|
|
// 24 - Uncompressed size
|
|
entry.uncompressedSize = buffer.readUInt32LE(24);
|
|
// 28 - File name length (n)
|
|
entry.fileNameLength = buffer.readUInt16LE(28);
|
|
// 30 - Extra field length (m)
|
|
entry.extraFieldLength = buffer.readUInt16LE(30);
|
|
// 32 - File comment length (k)
|
|
entry.fileCommentLength = buffer.readUInt16LE(32);
|
|
// 34 - Disk number where file starts
|
|
// 36 - Internal file attributes
|
|
entry.internalFileAttributes = buffer.readUInt16LE(36);
|
|
// 38 - External file attributes
|
|
entry.externalFileAttributes = buffer.readUInt32LE(38);
|
|
// 42 - Relative offset of local file header
|
|
entry.relativeOffsetOfLocalHeader = buffer.readUInt32LE(42);
|
|
|
|
self.readEntryCursor += 46;
|
|
|
|
buffer = new Buffer(entry.fileNameLength + entry.extraFieldLength + entry.fileCommentLength);
|
|
readAndAssertNoEof(self.reader, buffer, 0, buffer.length, self.readEntryCursor, function(err) {
|
|
if (err) return emitErrorAndAutoClose(self, err);
|
|
if (self.emittedError) return;
|
|
// 46 - File name
|
|
var isUtf8 = entry.generalPurposeBitFlag & 0x800
|
|
try {
|
|
entry.fileName = bufferToString(buffer, 0, entry.fileNameLength, isUtf8);
|
|
} catch (e) {
|
|
return emitErrorAndAutoClose(self, e);
|
|
}
|
|
|
|
// 46+n - Extra field
|
|
var fileCommentStart = entry.fileNameLength + entry.extraFieldLength;
|
|
var extraFieldBuffer = buffer.slice(entry.fileNameLength, fileCommentStart);
|
|
entry.extraFields = [];
|
|
var i = 0;
|
|
while (i < extraFieldBuffer.length) {
|
|
var headerId = extraFieldBuffer.readUInt16LE(i + 0);
|
|
var dataSize = extraFieldBuffer.readUInt16LE(i + 2);
|
|
var dataStart = i + 4;
|
|
var dataEnd = dataStart + dataSize;
|
|
var dataBuffer = new Buffer(dataSize);
|
|
extraFieldBuffer.copy(dataBuffer, 0, dataStart, dataEnd);
|
|
entry.extraFields.push({
|
|
id: headerId,
|
|
data: dataBuffer,
|
|
});
|
|
i = dataEnd;
|
|
}
|
|
|
|
// 46+n+m - File comment
|
|
try {
|
|
entry.fileComment = bufferToString(buffer, fileCommentStart, fileCommentStart + entry.fileCommentLength, isUtf8);
|
|
} catch (e) {
|
|
return emitErrorAndAutoClose(self, e);
|
|
}
|
|
|
|
self.readEntryCursor += buffer.length;
|
|
self.entriesRead += 1;
|
|
|
|
if (entry.uncompressedSize === 0xffffffff ||
|
|
entry.compressedSize === 0xffffffff ||
|
|
entry.relativeOffsetOfLocalHeader === 0xffffffff) {
|
|
// ZIP64 format
|
|
// find the Zip64 Extended Information Extra Field
|
|
var zip64EiefBuffer = null;
|
|
for (var i = 0; i < entry.extraFields.length; i++) {
|
|
var extraField = entry.extraFields[i];
|
|
if (extraField.id === 0x0001) {
|
|
zip64EiefBuffer = extraField.data;
|
|
break;
|
|
}
|
|
}
|
|
if (zip64EiefBuffer == null) return emitErrorAndAutoClose(self, new Error("expected Zip64 Extended Information Extra Field"));
|
|
var index = 0;
|
|
// 0 - Original Size 8 bytes
|
|
if (entry.uncompressedSize === 0xffffffff) {
|
|
if (index + 8 > zip64EiefBuffer.length) return emitErrorAndAutoClose(self, new Error("Zip64 Extended Information Extra Field does not include Original Size"));
|
|
entry.uncompressedSize = readUInt64LE(zip64EiefBuffer, index);
|
|
index += 8;
|
|
}
|
|
// 8 - Compressed Size 8 bytes
|
|
if (entry.compressedSize === 0xffffffff) {
|
|
if (index + 8 > zip64EiefBuffer.length) return emitErrorAndAutoClose(self, new Error("Zip64 Extended Information Extra Field does not include Compressed Size"));
|
|
entry.compressedSize = readUInt64LE(zip64EiefBuffer, index);
|
|
index += 8;
|
|
}
|
|
// 16 - Relative Header Offset 8 bytes
|
|
if (entry.relativeOffsetOfLocalHeader === 0xffffffff) {
|
|
if (index + 8 > zip64EiefBuffer.length) return emitErrorAndAutoClose(self, new Error("Zip64 Extended Information Extra Field does not include Relative Header Offset"));
|
|
entry.relativeOffsetOfLocalHeader = readUInt64LE(zip64EiefBuffer, index);
|
|
index += 8;
|
|
}
|
|
// 24 - Disk Start Number 4 bytes
|
|
}
|
|
|
|
// validate file size
|
|
if (entry.compressionMethod === 0) {
|
|
if (entry.compressedSize !== entry.uncompressedSize) {
|
|
var msg = "compressed/uncompressed size mismatch for stored file: " + entry.compressedSize + " != " + entry.uncompressedSize;
|
|
return emitErrorAndAutoClose(self, new Error(msg));
|
|
}
|
|
}
|
|
|
|
// validate file name
|
|
if (entry.fileName.indexOf("\\") !== -1) return emitErrorAndAutoClose(self, new Error("invalid characters in fileName: " + entry.fileName));
|
|
if (/^[a-zA-Z]:/.test(entry.fileName) || /^\//.test(entry.fileName)) return emitErrorAndAutoClose(self, new Error("absolute path: " + entry.fileName));
|
|
if (entry.fileName.split("/").indexOf("..") !== -1) return emitErrorAndAutoClose(self, new Error("invalid relative path: " + entry.fileName));
|
|
self.emit("entry", entry);
|
|
|
|
if (!self.lazyEntries) self.readEntry();
|
|
});
|
|
});
|
|
};
|
|
|
|
ZipFile.prototype.openReadStream = function(entry, callback) {
|
|
var self = this;
|
|
if (!self.isOpen) return callback(new Error("closed"));
|
|
// make sure we don't lose the fd before we open the actual read stream
|
|
self.reader.ref();
|
|
var buffer = new Buffer(30);
|
|
readAndAssertNoEof(self.reader, buffer, 0, buffer.length, entry.relativeOffsetOfLocalHeader, function(err) {
|
|
try {
|
|
if (err) return callback(err);
|
|
// 0 - Local file header signature = 0x04034b50
|
|
var signature = buffer.readUInt32LE(0);
|
|
if (signature !== 0x04034b50) return callback(new Error("invalid local file header signature: 0x" + signature.toString(16)));
|
|
// all this should be redundant
|
|
// 4 - Version needed to extract (minimum)
|
|
// 6 - General purpose bit flag
|
|
// 8 - Compression method
|
|
// 10 - File last modification time
|
|
// 12 - File last modification date
|
|
// 14 - CRC-32
|
|
// 18 - Compressed size
|
|
// 22 - Uncompressed size
|
|
// 26 - File name length (n)
|
|
var fileNameLength = buffer.readUInt16LE(26);
|
|
// 28 - Extra field length (m)
|
|
var extraFieldLength = buffer.readUInt16LE(28);
|
|
// 30 - File name
|
|
// 30+n - Extra field
|
|
var localFileHeaderEnd = entry.relativeOffsetOfLocalHeader + buffer.length + fileNameLength + extraFieldLength;
|
|
var compressed;
|
|
if (entry.compressionMethod === 0) {
|
|
// 0 - The file is stored (no compression)
|
|
compressed = false;
|
|
} else if (entry.compressionMethod === 8) {
|
|
// 8 - The file is Deflated
|
|
compressed = true;
|
|
} else {
|
|
return callback(new Error("unsupported compression method: " + entry.compressionMethod));
|
|
}
|
|
var fileDataStart = localFileHeaderEnd;
|
|
var fileDataEnd = fileDataStart + entry.compressedSize;
|
|
if (entry.compressedSize !== 0) {
|
|
// bounds check now, because the read streams will probably not complain loud enough.
|
|
// since we're dealing with an unsigned offset plus an unsigned size,
|
|
// we only have 1 thing to check for.
|
|
if (fileDataEnd > self.fileSize) {
|
|
return callback(new Error("file data overflows file bounds: " +
|
|
fileDataStart + " + " + entry.compressedSize + " > " + self.fileSize));
|
|
}
|
|
}
|
|
var readStream = self.reader.createReadStream({start: fileDataStart, end: fileDataEnd});
|
|
var endpointStream = readStream;
|
|
if (compressed) {
|
|
var destroyed = false;
|
|
var inflateFilter = zlib.createInflateRaw();
|
|
readStream.on("error", function(err) {
|
|
// setImmediate here because errors can be emitted during the first call to pipe()
|
|
setImmediate(function() {
|
|
if (!destroyed) inflateFilter.emit("error", err);
|
|
});
|
|
});
|
|
|
|
var checkerStream = new AssertByteCountStream(entry.uncompressedSize);
|
|
inflateFilter.on("error", function(err) {
|
|
// forward zlib errors to the client-visible stream
|
|
setImmediate(function() {
|
|
if (!destroyed) checkerStream.emit("error", err);
|
|
});
|
|
});
|
|
checkerStream.destroy = function() {
|
|
destroyed = true;
|
|
inflateFilter.unpipe(checkerStream);
|
|
readStream.unpipe(inflateFilter);
|
|
// TODO: the inflateFilter now causes a memory leak. see Issue #27.
|
|
readStream.destroy();
|
|
};
|
|
endpointStream = readStream.pipe(inflateFilter).pipe(checkerStream);
|
|
}
|
|
callback(null, endpointStream);
|
|
} finally {
|
|
self.reader.unref();
|
|
}
|
|
});
|
|
};
|
|
|
|
function Entry() {
|
|
}
|
|
Entry.prototype.getLastModDate = function() {
|
|
return dosDateTimeToDate(this.lastModFileDate, this.lastModFileTime);
|
|
};
|
|
|
|
function dosDateTimeToDate(date, time) {
|
|
var day = date & 0x1f; // 1-31
|
|
var month = (date >> 5 & 0xf) - 1; // 1-12, 0-11
|
|
var year = (date >> 9 & 0x7f) + 1980; // 0-128, 1980-2108
|
|
|
|
var millisecond = 0;
|
|
var second = (time & 0x1f) * 2; // 0-29, 0-58 (even numbers)
|
|
var minute = time >> 5 & 0x3f; // 0-59
|
|
var hour = time >> 11 & 0x1f; // 0-23
|
|
|
|
return new Date(year, month, day, hour, minute, second, millisecond);
|
|
}
|
|
|
|
function readAndAssertNoEof(reader, buffer, offset, length, position, callback) {
|
|
if (length === 0) {
|
|
// fs.read will throw an out-of-bounds error if you try to read 0 bytes from a 0 byte file
|
|
return setImmediate(function() { callback(null, new Buffer(0)); });
|
|
}
|
|
reader.read(buffer, offset, length, position, function(err, bytesRead) {
|
|
if (err) return callback(err);
|
|
if (bytesRead < length) return callback(new Error("unexpected EOF"));
|
|
callback();
|
|
});
|
|
}
|
|
|
|
util.inherits(AssertByteCountStream, Transform);
|
|
function AssertByteCountStream(byteCount) {
|
|
Transform.call(this);
|
|
this.actualByteCount = 0;
|
|
this.expectedByteCount = byteCount;
|
|
}
|
|
AssertByteCountStream.prototype._transform = function(chunk, encoding, cb) {
|
|
this.actualByteCount += chunk.length;
|
|
if (this.actualByteCount > this.expectedByteCount) {
|
|
var msg = "too many bytes in the stream. expected " + this.expectedByteCount + ". got at least " + this.actualByteCount;
|
|
return cb(new Error(msg));
|
|
}
|
|
cb(null, chunk);
|
|
};
|
|
AssertByteCountStream.prototype._flush = function(cb) {
|
|
if (this.actualByteCount < this.expectedByteCount) {
|
|
var msg = "not enough bytes in the stream. expected " + this.expectedByteCount + ". got only " + this.actualByteCount;
|
|
return cb(new Error(msg));
|
|
}
|
|
cb();
|
|
};
|
|
|
|
util.inherits(RandomAccessReader, EventEmitter);
|
|
function RandomAccessReader() {
|
|
EventEmitter.call(this);
|
|
this.refCount = 0;
|
|
}
|
|
RandomAccessReader.prototype.ref = function() {
|
|
this.refCount += 1;
|
|
};
|
|
RandomAccessReader.prototype.unref = function() {
|
|
var self = this;
|
|
self.refCount -= 1;
|
|
|
|
if (self.refCount > 0) return;
|
|
if (self.refCount < 0) throw new Error("invalid unref");
|
|
|
|
self.close(onCloseDone);
|
|
|
|
function onCloseDone(err) {
|
|
if (err) return self.emit('error', err);
|
|
self.emit('close');
|
|
}
|
|
};
|
|
RandomAccessReader.prototype.createReadStream = function(options) {
|
|
var start = options.start;
|
|
var end = options.end;
|
|
if (start === end) {
|
|
var emptyStream = new PassThrough();
|
|
setImmediate(function() {
|
|
emptyStream.end();
|
|
});
|
|
return emptyStream;
|
|
}
|
|
var stream = this._readStreamForRange(start, end);
|
|
|
|
var destroyed = false;
|
|
var refUnrefFilter = new RefUnrefFilter(this);
|
|
stream.on("error", function(err) {
|
|
setImmediate(function() {
|
|
if (!destroyed) refUnrefFilter.emit("error", err);
|
|
});
|
|
});
|
|
refUnrefFilter.destroy = function() {
|
|
stream.unpipe(refUnrefFilter);
|
|
refUnrefFilter.unref();
|
|
stream.destroy();
|
|
};
|
|
|
|
var byteCounter = new AssertByteCountStream(end - start);
|
|
refUnrefFilter.on("error", function(err) {
|
|
setImmediate(function() {
|
|
if (!destroyed) byteCounter.emit("error", err);
|
|
});
|
|
});
|
|
byteCounter.destroy = function() {
|
|
destroyed = true;
|
|
refUnrefFilter.unpipe(byteCounter);
|
|
refUnrefFilter.destroy();
|
|
};
|
|
|
|
return stream.pipe(refUnrefFilter).pipe(byteCounter);
|
|
};
|
|
RandomAccessReader.prototype._readStreamForRange = function(start, end) {
|
|
throw new Error("not implemented");
|
|
};
|
|
RandomAccessReader.prototype.read = function(buffer, offset, length, position, callback) {
|
|
var readStream = this.createReadStream({start: position, end: position + length});
|
|
var writeStream = new Writable();
|
|
var written = 0;
|
|
writeStream._write = function(chunk, encoding, cb) {
|
|
chunk.copy(buffer, offset + written, 0, chunk.length);
|
|
written += chunk.length;
|
|
cb();
|
|
};
|
|
writeStream.on("finish", callback);
|
|
readStream.on("error", function(error) {
|
|
callback(error);
|
|
});
|
|
readStream.pipe(writeStream);
|
|
};
|
|
RandomAccessReader.prototype.close = function(callback) {
|
|
setImmediate(callback);
|
|
};
|
|
|
|
util.inherits(RefUnrefFilter, PassThrough);
|
|
function RefUnrefFilter(context) {
|
|
PassThrough.call(this);
|
|
this.context = context;
|
|
this.context.ref();
|
|
this.unreffedYet = false;
|
|
}
|
|
RefUnrefFilter.prototype._flush = function(cb) {
|
|
this.unref();
|
|
cb();
|
|
};
|
|
RefUnrefFilter.prototype.unref = function(cb) {
|
|
if (this.unreffedYet) return;
|
|
this.unreffedYet = true;
|
|
this.context.unref();
|
|
};
|
|
|
|
var cp437 = '\u0000☺☻♥♦♣♠•◘○◙♂♀♪♫☼►◄↕‼¶§▬↨↑↓→←∟↔▲▼ !"#$%&\'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[\\]^_`abcdefghijklmnopqrstuvwxyz{|}~⌂ÇüéâäàåçêëèïîìÄÅÉæÆôöòûùÿÖÜ¢£¥₧ƒáíóúñѪº¿⌐¬½¼¡«»░▒▓│┤╡╢╖╕╣║╗╝╜╛┐└┴┬├─┼╞╟╚╔╩╦╠═╬╧╨╤╥╙╘╒╓╫╪┘┌█▄▌▐▀αßΓπΣσµτΦΘΩδ∞φε∩≡±≥≤⌠⌡÷≈°∙·√ⁿ²■ ';
|
|
function bufferToString(buffer, start, end, isUtf8) {
|
|
if (isUtf8) {
|
|
return buffer.toString("utf8", start, end);
|
|
} else {
|
|
var result = "";
|
|
for (var i = start; i < end; i++) {
|
|
result += cp437[buffer[i]];
|
|
}
|
|
return result;
|
|
}
|
|
}
|
|
|
|
function readUInt64LE(buffer, offset) {
|
|
// there is no native function for this, because we can't actually store 64-bit integers precisely.
|
|
// after 53 bits, JavaScript's Number type (IEEE 754 double) can't store individual integers anymore.
|
|
// but since 53 bits is a whole lot more than 32 bits, we do our best anyway.
|
|
var lower32 = buffer.readUInt32LE(offset);
|
|
var upper32 = buffer.readUInt32LE(offset + 4);
|
|
// we can't use bitshifting here, because JavaScript bitshifting only works on 32-bit integers.
|
|
return upper32 * 0x100000000 + lower32;
|
|
// as long as we're bounds checking the result of this function against the total file size,
|
|
// we'll catch any overflow errors, because we already made sure the total file size was within reason.
|
|
}
|
|
|
|
function defaultCallback(err) {
|
|
if (err) throw err;
|
|
}
|