rate-limiter-flexible
Version:
Flexible API rate limiter backed by Redis for distributed node.js applications
232 lines (197 loc) • 6.06 kB
JavaScript
const RateLimiterStoreAbstract = require('./RateLimiterStoreAbstract');
const RateLimiterRes = require('./RateLimiterRes');
class RateLimiterMongo extends RateLimiterStoreAbstract {
/**
*
* @param {Object} opts
* Defaults {
* indexKeyPrefix: {attr1: 1, attr2: 1}
* ... see other in RateLimiterStoreAbstract
*
* mongo: MongoClient
* }
*/
constructor(opts) {
super(opts);
this.dbName = opts.dbName;
this.tableName = opts.tableName;
this.indexKeyPrefix = opts.indexKeyPrefix;
if (opts.mongo) {
this.client = opts.mongo;
} else {
this.client = opts.storeClient;
}
if (typeof this.client.then === 'function') {
// If Promise
this.client
.then((conn) => {
this.client = conn;
this._initCollection();
});
} else {
this._initCollection();
}
}
get dbName() {
return this._dbName;
}
set dbName(value) {
this._dbName = typeof value === 'undefined' ? RateLimiterMongo.getDbName() : value;
}
static getDbName() {
return 'node-rate-limiter-flexible';
}
get tableName() {
return this._tableName;
}
set tableName(value) {
this._tableName = typeof value === 'undefined' ? this.keyPrefix : value;
}
get client() {
return this._client;
}
set client(value) {
if (typeof value === 'undefined') {
throw new Error('mongo is not set');
}
this._client = value;
}
get indexKeyPrefix() {
return this._indexKeyPrefix;
}
set indexKeyPrefix(obj) {
this._indexKeyPrefix = obj || {};
}
_initCollection() {
const db = typeof this.client.db === 'function'
? this.client.db(this.dbName)
: this.client.db;
const collection = db.collection(this.tableName);
collection.createIndex({ expire: -1 }, { expireAfterSeconds: 0 });
collection.createIndex(Object.assign({}, this.indexKeyPrefix, { key: 1 }), { unique: true });
this._collection = collection;
}
_getRateLimiterRes(rlKey, changedPoints, result) {
const res = new RateLimiterRes();
let doc;
if (typeof result.value === 'undefined') {
if (result._id) {
doc = result;
} else {
[doc] = result.ops; // ops set on replaceOne
}
} else {
doc = result.value;
}
res.isFirstInDuration = doc.points === changedPoints;
res.consumedPoints = doc.points;
res.remainingPoints = Math.max(this.points - res.consumedPoints, 0);
res.msBeforeNext = Math.max(new Date(doc.expire).getTime() - Date.now(), 0);
return res;
}
_upsert(key, points, msDuration, forceExpire = false, options = {}) {
if (!this._collection) {
return Promise.reject(Error('Mongo connection is not established'));
}
const docAttrs = options.attrs || {};
let where;
let upsertData;
if (forceExpire) {
where = { key };
where = Object.assign(where, docAttrs);
upsertData = {
$set: {
key,
points,
expire: new Date(Date.now() + msDuration),
},
};
upsertData.$set = Object.assign(upsertData.$set, docAttrs);
} else {
where = {
expire: { $gt: new Date() },
key,
};
where = Object.assign(where, docAttrs);
upsertData = {
$setOnInsert: {
key,
expire: new Date(Date.now() + msDuration),
},
$inc: { points },
};
upsertData.$setOnInsert = Object.assign(upsertData.$setOnInsert, docAttrs);
}
/*
* 1. Find actual limit and increment points
* 2. If limit expired, but Mongo doesn't clean doc by TTL yet, try to replace limit doc completely
* 3. If 2 or more Mongo threads try to insert the new limit doc, only the first succeed
* 4. Try to upsert from step 1. Actual limit is created now, points are incremented without problems
*/
return new Promise((resolve, reject) => {
this._collection.findOneAndUpdate(
where,
upsertData,
{
upsert: true,
returnOriginal: false,
} // eslint-disable-line comma-dangle
).then((res) => {
resolve(res);
}).catch((errUpsert) => {
if (errUpsert && errUpsert.code === 11000) { // E11000 duplicate key error collection
const replaceWhere = Object.assign({
expire: { $lte: new Date() }, // try to replace OLD limit doc
key,
}, docAttrs);
const replaceTo = Object.assign({
key,
points,
expire: new Date(Date.now() + msDuration),
}, docAttrs);
this._collection.replaceOne(
replaceWhere,
replaceTo,
{
upsert: true,
returnOriginal: false,
} // eslint-disable-line comma-dangle
).then((res) => {
resolve(res);
}).catch((errReplace) => {
if (errReplace && errReplace.code === 11000) { // E11000 duplicate key error collection
this._upsert(key, points, msDuration, forceExpire)
.then(res => resolve(res))
.catch(err => reject(err));
} else {
reject(errReplace);
}
});
} else {
reject(errUpsert);
}
});
});
}
_get(rlKey, options = {}) {
if (!this._collection) {
return Promise.reject(Error('Mongo connection is not established'));
}
const docAttrs = options.attrs || {};
const where = Object.assign({
key: rlKey,
expire: { $gt: new Date() },
}, docAttrs);
return this._collection.findOne(where);
}
_delete(rlKey, options = {}) {
if (!this._collection) {
return Promise.reject(Error('Mongo connection is not established'));
}
const docAttrs = options.attrs || {};
const where = Object.assign({ key: rlKey }, docAttrs);
return this._collection.deleteOne(where)
.then(res => res.result.n > 0);
}
}
module.exports = RateLimiterMongo;