
Security News
Attackers Are Hunting High-Impact Node.js Maintainers in a Coordinated Social Engineering Campaign
Multiple high-impact npm maintainers confirm they have been targeted in the same social engineering campaign that compromised Axios.
rate-limiter-memory
Advanced tools
Node.js rate limiter by key and protection from DDoS and Brute-Force attacks in process Memory, Redis, MongoDb, Memcached, MySQL, PostgreSQL, Cluster or PM
fork from rate-limiter-flexible and keep only RateLimiterMemory.
rate-limiter-flexible counts and limits the number of actions by key and protects from DDoS and brute force attacks at any scale.
It works with Redis, Prisma, DynamoDB, process Memory, Cluster or PM2, Memcached, MongoDB, MySQL, and PostgreSQL.
Memory limiter also works in the browser.
Atomic increments. All operations in memory or distributed environment use atomic increments against race conditions.
Fast. Average request takes 0.7ms in Cluster and 2.5ms in Distributed application. See benchmarks.
Flexible. Combine limiters, block key for some duration, delay actions, manage failover with insurance options, configure smart key blocking in memory and many others.
Ready for growth. It provides a unified API for all limiters. Whenever your application grows, it is ready. Prepare your limiters in minutes.
Friendly. No matter which node package you prefer: redis or ioredis, sequelize/typeorm or knex, memcached, native driver or mongoose. It works with all of them.
In-memory blocks. Avoid extra requests to store with inMemoryBlockOnConsumed.
Allow traffic bursts with BurstyRateLimiter.
Deno compatible See this example
It uses a fixed window, as it is much faster than a rolling window. See comparative benchmarks with other libraries here
npm i --save rate-limiter-flexible
yarn add rate-limiter-flexible
// CommonJS
const { RateLimiterMemory } = require("rate-limiter-flexible");
// or
// ECMAScript
import { RateLimiterMemory } from "rate-limiter-flexible";
// or
import RateLimiterMemory from "rate-limiter-flexible/lib/RateLimiterMemory.js";
Points can be consumed by IP address, user ID, authorisation token, API route or any other string.
const opts = {
points: 6, // 6 points
duration: 1, // Per second
};
const rateLimiter = new RateLimiterMemory(opts);
rateLimiter.consume(remoteAddress, 2) // consume 2 points
.then((rateLimiterRes) => {
// 2 points consumed
})
.catch((rateLimiterRes) => {
// Not enough points to consume
});
The Promise's resolve and reject callbacks both return an instance of the RateLimiterRes class if there is no error.
Object attributes:
RateLimiterRes = {
msBeforeNext: 250, // Number of milliseconds before next action can be done
remainingPoints: 0, // Number of remaining points in current duration
consumedPoints: 5, // Number of consumed points in current duration
isFirstInDuration: false, // action is first in current duration
}
You may want to set HTTP headers for the response:
const headers = {
"Retry-After": rateLimiterRes.msBeforeNext / 1000,
"X-RateLimit-Limit": opts.points,
"X-RateLimit-Remaining": rateLimiterRes.remainingPoints,
"X-RateLimit-Reset": new Date(Date.now() + rateLimiterRes.msBeforeNext)
}
get, set, block, delete, penalty and reward methodsFull documentation is on Wiki
Some copy/paste examples on Wiki:
See releases for detailed changelog.
points
Default: 4
Maximum number of points that can be consumed over duration
duration
Default: 1
Number of seconds before consumed points are reset.
Points are never reset if duration is set to 0.
storeClient
Required for store limiters
Must be redis, ioredis, memcached, mongodb, pg, mysql2, mysql or any other related pool or connection.
knex, if you use it.Smooth out traffic peaks:
Specific:
Read detailed description on Wiki.
RateLimiterRes or null.secDuration seconds.Average latency during test of pure NodeJS endpoint in cluster of 4 workers with everything set up on one server.
1000 concurrent clients with maximum 2000 requests per sec during 30 seconds.
1. Memory 0.34 ms
2. Cluster 0.69 ms
3. Redis 2.45 ms
4. Memcached 3.89 ms
5. Mongo 4.75 ms
500 concurrent clients with maximum 1000 req per sec during 30 seconds
6. PostgreSQL 7.48 ms (with connection pool max 100)
7. MySQL 14.59 ms (with connection pool 100)
Note, you can speed up limiters with inMemoryBlockOnConsumed option.
Appreciated, feel free!
Make sure you've launched npm run eslint before creating PR, all errors have to be fixed.
You can try to run npm run eslint-fix to fix some issues.
Any new limiter with storage must be extended from RateLimiterStoreAbstract.
It has to implement 4 methods:
_getRateLimiterRes parses raw data from store to RateLimiterRes object.
_upsert may be atomic or non-atomic upsert (increment). It inserts or updates the value by key and returns raw data.
If it doesn't make an atomic upsert (increment), the class should be suffixed with NonAtomic, e.g. RateLimiterRedisNonAtomic.
It must support forceExpire mode to overwrite key expiration time.
_get returns raw data by key or null if there is no key.
_delete deletes all key-related data and returns true on deleted, false if key is not found.
All other methods depends on the store. See RateLimiterRedis or RateLimiterPostgres for examples.
Note: all changes should be covered by tests.
FAQs
Node.js rate limiter by key and protection from DDoS and Brute-Force attacks in process Memory, Redis, MongoDb, Memcached, MySQL, PostgreSQL, Cluster or PM
The npm package rate-limiter-memory receives a total of 1 weekly downloads. As such, rate-limiter-memory popularity was classified as not popular.
We found that rate-limiter-memory demonstrated a not healthy version release cadence and project activity because the last version was released a year ago. It has 1 open source maintainer collaborating on the project.
Did you know?

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Security News
Multiple high-impact npm maintainers confirm they have been targeted in the same social engineering campaign that compromised Axios.

Security News
Axios compromise traced to social engineering, showing how attacks on maintainers can bypass controls and expose the broader software supply chain.

Security News
Node.js has paused its bug bounty program after funding ended, removing payouts for vulnerability reports but keeping its security process unchanged.