unblocker
Version:
Web proxy for evading internet censorship & general-purpose library for rewriting remote websites.
91 lines (72 loc) • 3.54 kB
JavaScript
var PassThrough = require('stream').PassThrough;
var zlib = require('zlib');
var contentTypes = require('./content-types.js');
var debug = require('debug')('unblocker:decompress');
module.exports = function(config) {
function acceptableCompression(data) {
// deflate is tricky so we're only going to ask for gzip if the client allows it
if (data.headers['accept-encoding'] && data.headers['accept-encoding'].indexOf('gzip' != -1)) {
data.headers['accept-encoding'] = 'gzip';
} else {
delete data.headers['accept-encoding'];
}
}
function shouldProcess(data) {
// no body to process on 204 No Content or 304 Not Modified
if ([204, 304].includes(data.remoteResponse.statusCode)) {
return false;
}
var headers = data.headers;
// skip requests that declair a 0-length body
if (parseInt(headers['content-length'], 10) === 0) {
return false;
}
// decompress if it's gzipped or deflate'd
return (
headers['content-encoding'] == 'gzip' ||
headers['content-encoding'] == 'deflate'
);
}
function decompressResponse(data) {
if (contentTypes.shouldProcess(config, data) && shouldProcess(data)) {
debug('decompressing %s encoding and deleting content-encoding header', data.headers['content-encoding']);
// https://github.com/nfriedly/node-unblocker/pull/105
// zlib streams can throw if given a source with no content
// we try to avoid processing those, but it could happen.
// because of that, we're using a pass through stream as a placeholder.
// we won't create the zlib stream until we know we have data to process.
var sourceStream = data.stream;
var placeHolder = new PassThrough();
data.stream = placeHolder;
var handleData = function handleData() {
var firstChunk = sourceStream.read();
// 'readable' is also called when a stream ends with no data
// this is how you tell when that's the case:
if (firstChunk === null) {
placeHolder.end();
return;
}
var decompressStream;
if (data.headers['content-encoding'] == 'deflate') {
// https://github.com/nfriedly/node-unblocker/issues/12
// inflateRaw seems to work here wheras inflate and unzip do not.
// todo: validate this against other sites - if some require raw and others require non-raw, then maybe just rewrite the accept-encoding header to gzip only
decompressStream = zlib.createInflateRaw();
} else {
decompressStream = zlib.createUnzip();
}
decompressStream.write(firstChunk);
// continue piping through the placeholder since the next stream in the chain is listening to it, not our new decompressStream
sourceStream.pipe(decompressStream).pipe(placeHolder);
};
// if we do get data, create a decompression stream and pipe through it
sourceStream.once('readable', handleData);
delete data.headers['content-encoding'];
}
}
return {
handleRequest: acceptableCompression,
handleResponse: decompressResponse
};
};
;