node-spider
Version:
Generic web crawler powered by Node.js
40 lines (31 loc) • 1.26 kB
Markdown
# Changelog
## 1.4.1
- Fixed bug when adding referrer
## 1.4.0
- Implemented `addReferrer` option
- Implemented `keepAlive` option
- Implemented `xhr` option
## 1.3.0
- Fixed document.$ (cheerio instance) was being recreated on every call
- Implemented git hooks via ghooks module
- Added jshint script, along with it's rc config and ignore files
- Added a pre-commit hook to lint the code
- Added tests powered by mocha+chai
- Added a pre-push hook to test the code
## 1.2.2
- request() errors hanging the flow fixed, thanks @arve0
- require('./lib/Document') fixed, it's now lower case, thanks @arve0
## 1.2.1
- `logs` cannot be `true` anymore, just specify a stream, or don't
- All options are passed to `request`, so `headers` and `encoding` work implicitely now
- Updated the docs
## 1.2.0
- Added `catchErrors` option, if true handlers will by try-catched and errors sent to `error` callback
- `logs` option can be any stream and the logs are written to it (stdout, stderr, file, etc)
## 1.1.0
- Added `delay` option to wait after each request
- Added `encoding` option to set the requests encoding
## 1.0.0
- Updated dependencies versions
- Duplicated URLs are ignored unless `allowDuplicates` is `true`
- Reversed `error()` callback's arguments