Huge News!Announcing our $40M Series B led by Abstract Ventures.Learn More
Socket
Sign inDemoInstall
Socket

keuss

Package Overview
Dependencies
Maintainers
1
Versions
76
Alerts
File Explorer

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

keuss

Job Queues an pipelines on selectable backends (for now: mongodb and redis) for node.js

  • 1.4.3
  • Source
  • npm
  • Socket score

Version published
Weekly downloads
19
increased by46.15%
Maintainers
1
Weekly downloads
 
Created
Source

keuss

Job Queues an pipelines on selectable backends (for now: mongodb and redis) for node.js

Contents

About

Keuss is an attempt or experiment to provide a serverless, persistent and kigh-available queue middleware supporting delays/schedule, using mongodb and redis to provide most of the backend needs. As of now, it has evolved into a rather capable and complete queue middleware

The underlying idea is that the key to provide persistency, HA and load balance is to have a storage subsystem that provides that, and use it to store your queues. Instead of reinventing the wheel by building such as storage I simply tried to adapt what's already out there

Modelling a queue with mongodb, for example, proved easy enough. It resulted simple, cheap and mongodb provides great persistency and HA, and decent support for load balancing. Although using Redis provided similar results, in both cases the load balancing part was somewhat incomplete: the whole thing lacked a bus to signal all clients about, for example, when an insertion in a particular queue takes place. Without this layer a certain amount of polling is needed, so it's obviously a Nice Thing To Have

Keuss ended up being a somewhat serverless queue system, where the server or common parts are bare storage systems such as redis or mongodb. There is no need for any extra keuss server in between clients and storage (although an actual keuss-server does exist, serving a different purpose on top of plain keuss). Thus, all keuss actually lays at the client side

Concepts

Queue

a Queue is more of an interface, a definition of what it can do. Keuss queues are capable of:

  • insert one element
  • schedule an element: insert one element with a not-before datetime
  • get an element, and block for some specified time if no element is available
  • reserve an element, and block for some specified time if no element is available
  • commit (remove) or rollback (return back) a previously reserved element
  • get element count
  • get element count whose not-before datetime is in the future (scheduled elements)
  • get usage stats: elements inserted, elements extracted

element here translates to any js object. Internally, it's usually managed as json

Pipeline

A pipeline is an enhanced queue that provides an extra operation: pass an element to another queue atomically. In an scenario where processors are linked with queues, it is usually a good feature to allow the 'commit element in incoming queue, insert element in the next queue' to be atomic. This removes chances for race conditions, or message losses

The pipeline concept is, indeed, an extension of the reserve-commit model; it is so far implemented only atop mongodb, and it is anyway considered as a 'low-level' feature, best used by means of specialized classes to encapsulate the aforementioned processors

Storage

Storage or Backend provides almost-complete queue primitives, fully functional and already usable as is. Keuss comes with 5 backends, with various levels of features and performance:

  • mongo, a mongodb-based backend that provides te full set of queue features, still with decent performance
  • redis-oq, backed using an ordered queue on top of redis (made in turn with a sorted set, a hash and some lua). Provides all queue features including reserve-commit-rollback. Noticeable faster than mongodb
  • redis-list, backed using a redis list. Does not offer reserve-commit-rollback nor the ability to schedule, but is much faster than redis-oq
  • pl-mongo, a version of the mongo backend that provides pipelining capabilities (the queues it produces are also pipelines).
  • ps-mongo, a verion of mongo backend where elements are not physically deleted from the collection when extracted; instead, they are just marked as processed and later deleted automatically using a mongodb TTL index

As mentioned before, persistence and HA depends exclusively on the underliying system: mongodb provides production-grade HA and persistence while using potentially gigantic queues, and with redis one can balance performance and simplicity over reliability and durability, by using standalone redis, redis sentinel or redis cluster. Keuss uses ioredis as redis driver, which supports all 3 cases

The following table shows the capabilities of each backend:

backenddelay/schedulereserve/commitpipelininghistoryspeed
redis-list----+++++
redis-oqxx--+++
mongoxx--++
pl-mongoxxx-+
ps-mongoxx-x++
Signaller

Signaller provides a bus interconnecting all keuss clients, so events can be shared. Keuss provides 2 signallers:

  • local : provides in-proccess messaging, useful only for simple cases or testing
  • redis-pubsub: uses the pubsub subsystem provided by redis
  • mongo-capped: uses pubsub on top of a mongodb capped collection, using mubsub

So far, the only events published by keuss is element inserted in queue X, which allows other clients waiting for elements to be available to wake up and retry. A client will not fire an event if another one of the same type (same client, same queue) was already fired less than 50ms ago

Stats

Stats provides counters and metrics on queues, shared among keuss clients. So far, only 'elements inserted' and 'elements got' are maintained. Two options are provided:

  • mem: very simple in-process, memory based
  • redis: backed by redis hashes. Modifications are buffered in memory and flushed every 100ms
  • mongo: backed by mongodb usnig one object per queue inside a singel collection. Modifications are buffered in memory and flushed every 100ms

How all fits together

  • Queues, or rather clients to individual queues, are created using a backend as factory
  • Backends need to be intialized before being used. Exact inittialization details depend on each backend
  • When creating a queue, a signaller and a stats are assigned to it. The actual class/type to be used can be specified at the queue's creation moment, or at the backend initialization moment. By default local and mem, respectively, are used
  • Queues are created on-demand, and are never destroyed as far as keuss is concerned. They do exist as long as the underlying backend kepts them in existence: for example, redis queues dissapear as such when they become empty
  • Pipelines are, strictly speaking, just enhanced queues; as such they behave and can be used as a queue

More info on pipelines here

Install

npm install keuss

Usage

Factory API

Backends, which work as queue factories, have the following operations

Initialization
var QM = require ('keuss/backends/<backend>');

MQ (opts, function (err, factory) {
  // factory contains the actual factory, initialized
})

where 'opts' is an object containing default values for queue creation (such as pollInterval, signaller or stats), plus the following backend-dependent values:

  • backends mongo, pl-mongo and ps-mongo
    • url: mongodb url to use, defaults to mongodb://localhost:27017/keuss
  • backends redis-list and redis-oq
    • redis: data to create a redis connection to the Redis acting as backend, see below
  • backend ps-mongo
    • ttl: time to keep consumed elements in the collection after being removed. Defauls to 3600 secs
Queue creation
// factory has been initialized
var q = factory.queue (<name>, <options>);

Where:

  • name: string to be used as queue name. Queues with the same name are in fact the same queue if they're backed in the same factory type using the same initialization data (mongodb url or redis conn-data)
  • options: the options passed at backend initialization are used as default values:
    • pollInterval: rearm or poll period in millisecs for get operations, defaults to 15000 (see Working with no signallers below)
    • signaller: signaller to use for the queue
      • provider: signaller factory
      • opts: options for the signaller factory (see below)
    • stats: stats store to use for this queue
      • provider: stats factory
      • opts: options for the stats factory (see below)

Signaller

Signaller factory is passed to queues either in queue creation or in backend init, inside opts.signaller. Note that the result fo the new operation is indeed the factory; the result of the require is therefore a metafactory

var signal_redis_pubsub = require ('../signal/redis-pubsub');

var local_redis_opts = {
  Redis: {
    port: 6379,
    host: 'localhost',
    db: 6
  }
};

var f_opts = {
  signaller: {
    provider: signal_redis_pubsub,
    opts: local_redis_opts
  }
  .
  .
  .
}

MQ (f_opts, function (err, factory) {
  // queues created by factory here will use a redis pubsub signaller, hosted at redis at localhost, db 6

The signaller has no public api per se; it is considered just a piece of infrastructure to glue queues together

Stats

Stats factory is passed to queues either in queue creation or in backend init, inside opts.signaller. Note that the result fo the new operation is indeed the factory; the result of the require is therefore a metafactory

var local_redis_pubsub = require ('../signal/redis-pubsub');

var local_redis_opts = {
  Redis: {
    port: 6379,
    host: 'localhost',
    db: 6
  }
};

var f_opts = {
  stats: {
    provider: signal_redis_pubsub,
    opts: local_redis_opts
  }
  .
  .
  .
}

MQ (f_opts, function (err, factory) {
  // queues created by factory here will use a redis-backed stats, hosted at redis at localhost, db 6

Stats objects, as of now, store the numer of elements inserted and the number of elements extracted; they are created behind the scenes and tied to queue instances, and the stats-related interface is in fact part fo the queues' interface

Queue API

Get Stats
q.stats (function (err, res) {
  ...
})
  • res contains usage stats (elements insterted, elements extracted)
Queue name
var qnane = q.name () 
Queue type
var qtype = q.type ()

returns a string with the type of the queue (the type of backend who was used to create it)

Queue occupation
q.size (function (err, res){
  ...
})

res contains the number of elements in the queue that are already elligible (that is, excluding scheduled elements with a schedule time in the future)

Total Queue occupation
q.totalSize (function (err, res){
  ...
})

res contains the number of elements in the queue (that is, including scheduled elements with a schedule time in the future)

Time of schedule of next message
q.next_t (function (err, res){
  ...
})

Returns a Date, or null if queue is empty. Queues with no support for schedule/delay always return null

Add element to queue
q.push (payload, opts, function (err, res) {
  ...
})

Adds payload to the queue, calls passed callback upon completion. Callback's res will contain the id assigned to the inserted element, if the backup provides one

Possible opts:

  • mature: unix timestamp where the element would be elligible for extraction. It is guaranteed that the element won't be extracted before this time
  • delay: delay in seconds to calculate the mature timestamp, if mature is not provided. For example, a delay=120 guarantees the element won't be extracted until 120 secs have elapsed at least
  • tries: value to initialize the retry counter, defaults to 0 (still no retries).

note: mature and delay have no effect if the backend does not support delay/schedule

Get element from queue
var tr = q.pop (cid, opts, function (err, res) {
  ...
})

Obtains an element from the queue. Callback is called with the element obtained if any, or if an error happened. If defined, the operation will wait for opts.timeout seconds for an element to appear in the queue before bailing out (with both err and res being null). However, it immediately returns an id that can be used to cancel the operation at anytime

cid is an string that identifies the consumer entity, which is considered to be a simple label

Possible opts:

  • timeout: milliseconds to wat for an elligible element to appear in the queue to be returned. If not defined it will wait forever
  • reserve: if true the element is only reserved, not completely returned. This means either ok or ko operations are needed upon the obtained element once processed, otherwise the element will be rolled back (and made available again) at some point in the future (this is only available on backends capable of reserve/commit)
Cancel a pending Pop
var tr = q.pop (cid, opts, function (err, res) {...});
.
.
.
q.cancel (tr);

Cancels a pending pop operation, identified by the value returned by pop()

Commit a reserved element
q.ok (id, function (err, res) {
  ...
})

commits a reserved element by its id (the id would be at res._id on the res param of pop() operation). This effectively erases the element from the queue

Rolls back a reserved element
q.ko (id, next_t, function (err, res) {
  ...
})

rolls back a reserved element by its id (the id would be at res._id on the res param of pop() operation). This effectively makes the element available again at the queue, marking to be mature at next_t (next_t being a millsec-unixtime). If no next_t is specified or a null is passed, 'now' is assumed.

Redis connections

Keuss relies on ioredis for connecting to redis. Anytime a redis connection is needed, keuss will create it from the opts object passed:

  • if opts is a function, it is executed. It is expected to return a redis connection
  • if it's an object and contains a 'Redis' field, this field is used to create a new ioredis Redis object, as in return new Redis (opts.Redis)
  • if it's an object and contains a 'Cluster' field, this field is used to create a new ioredis Redis.Cluster object, as in return new Redis.Cluster (opts.Cluster)
  • else, a ioredis Redis object is created with opts as param, as in return new Redis (opts)

Examples:

  • default options
    var MQ = require ('keuss/backends/redis-list');
    var factory_opts = {};
         
    MQ (factory_opts, function (err, factory) {
      ...
    });
    
  • specific redis params for ioredis Redis client
    var MQ = require ('keuss/backends/redis-list');
    var factory_opts = {
      redis: {
        Redis: {
          port: 12293,
          host: 'some-redis-instance.somewhere.com',
          family: 4,
          password: 'xxxx',
          db: 0
        }
      }
    };
         
    MQ (factory_opts, function (err, factory) {
      ...
    });
    
  • use a factory function
    var MQ = require ('keuss/backends/redis-list');
    var Redis = require ('ioredis');
    var factory_opts = {
      redis: function () {
        return new Redis ({
          port: 12293,
          host: 'some-redis-instance.somewhere.com',
          family: 4,
          password: 'xxxx',
          db: 0
        })
      }
    };
         
    MQ (factory_opts, function (err, factory) {
      ...
    });
    

Reserve & (commit | rollback)

Working with no signallers

Even when using signallers, get operations on queue never block or wait forever; waiting get operations rearm themselves every 15000 millisec (or whatever specified in the pollInterval). This feature provides the ability to work with more than one process without signallers, geting a maximum latency of pollInterval millisecs, but also provides a safe backup in the event of signalling lost for whatever reason

Examples

A set of funcioning examples can be found inside the examples directory

Keywords

FAQs

Package last updated on 06 Aug 2018

Did you know?

Socket

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

  • Package Alerts
  • Integrations
  • Docs
  • Pricing
  • FAQ
  • Roadmap
  • Changelog

Packages

npm

Stay in touch

Get open source security insights delivered straight into your inbox.


  • Terms
  • Privacy
  • Security

Made with ⚡️ by Socket Inc