crawler
Version:
Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously
20 lines (15 loc) • 464 B
JavaScript
const Crawler = require('../lib/crawler');
const expect = require('chai').expect;
const given = require('mocha-testdata');
describe('Unsupported function test', function() {
given.async('onDrain','preRequest','cache')
.it('should throw exception when using unsupported options', function(done,key) {
const opt = {
maxConnections: 10
};
opt[key] = {};
expect(function(){Crawler(opt);}).to.throw(Error);
done();
});
});
;