Research
Security News
Quasar RAT Disguised as an npm Package for Detecting Vulnerabilities in Ethereum Smart Contracts
Socket researchers uncover a malicious npm package posing as a tool for detecting vulnerabilities in Etherium smart contracts.
A cargo
with queue
-like parallel processing and capacity optimization.
Creates a quargo
object with the specified capacity
, concurrency
and delay
. Tasks added to the quargo
will be processed altogether (up to the capacity
limit) in parallel batches (up to the concurrency
limit). If all workers are in progress, the task is queued until one becomes available. If the quargo
hasn’t reached capacity
, the task is queued for delay
milliseconds. Once a worker has completed some tasks, each callback of those tasks is called.
Quargo passes an array of tasks to one of a group of workers, repeating when the worker is finished.
capacity
tasks at once.delay
milliseconds after being pushed to the quargo
capacity
is reached or delay
has passed, depending on workers availability.worker(tasks, callback)
- An asynchronous function for processing an array of queued tasks, which must call its callback(err)
argument when finished, with an optional err
argument.capacity
- An integer for determining how many tasks should be processed per round.concurrency
- An optional integer for determining how many worker functions should be run in parallel; if omitted, the default is 1
.delay
- An optional integer for determining how long should the quargo
wait to reach capacity
; if omitted, the default is 0
.The quargo
object returned has the following properties and methods:
length()
- a function returning the number of items waiting to be processed.running()
- a function returning true if there are items being processed, false if not.idle()
- a function returning false if there are items waiting or being processed, or true if not.concurrency
- an integer for determining how many worker
functions should be run in parallel. This property can be changed after a quargo
is created to alter the concurrency on-the-fly.push(task, [callback])
- adds task
to the quargo
. Calls callback
once the worker
has finished processing the task. Instead of a single task, a tasks
array can be submitted. The respective callback is used for every task in the list.empty
- A callback that is called when the last item from the quargo
is given to a worker
.drain
- A callback that is called when the last item from the quargo
has returned from the worker
.quargo(worker, capacity)
quargo(worker, capacity, concurrency)
quargo(worker, capacity, concurrency, delay)
quargo(worker, options)
quargo(options, worker)
Possible options
are
capacity
concurrency
delay
empty
callbackdrain
callbackObject | Tasks per worker (capacity) | Workers per object (concurrency) |
---|---|---|
queue | 1 | x |
cargo | y | 1 |
quargo | y | x |
quargo(worker, capacity)
Equivalent to async.cargo(worker, capacity)
quargo(worker, 1, concurrency)
Equivalent to async.queue(worker, concurrency)
quargo(worker, capacity, concurrency)
Roughly equivalent to using a queue and a cargo together
var queue = async.queue(worker, concurrency);
var cargo = async.cargo(function(tasks, cargoCb) {
queue.push(tasks);
cargoCb(); // call immediately
}, capacity);
cargo.push(task, taskCb);
In the async
version, taskCb
will never be called (it would mean passing cargoCb
to queue.push(tasks, cargoCb)
, which therefore waits for the worker to complete before pushing other tasks to the queue, making the queue useless).
quargo(worker, capacity, concurrency, delay)
Instead of processing tasks on next tick as async.cargo
does, quargo
waits for delay
milliseconds before processing tasks.
If capacity
is reached before delay
, delay
is ignored and tasks are processed immediately. This is the capacity optimization of quargo
.
You may send a huge amount of data to StatHat. Do you make an HTTP request every-time, when you could send them by batch instead? async.cargo
won’t help though, I doubt all stats arrive during the same tick.
You send data to BigQuery but the rate at which you would like to send it exceeds their limit? You need a quargo
in your life to send by batches instead.
AWS SQS is super cheap. Using a quargo
to delete messages by batch makes it cheaper.
FAQs
A cargo with queue-like parallel processing and capacity optimization.
We found that quargo demonstrated a not healthy version release cadence and project activity because the last version was released a year ago. It has 1 open source maintainer collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Research
Security News
Socket researchers uncover a malicious npm package posing as a tool for detecting vulnerabilities in Etherium smart contracts.
Security News
Research
A supply chain attack on Rspack's npm packages injected cryptomining malware, potentially impacting thousands of developers.
Research
Security News
Socket researchers discovered a malware campaign on npm delivering the Skuld infostealer via typosquatted packages, exposing sensitive data.