
Security News
OpenClaw Skill Marketplace Emerges as Active Malware Vector
Security researchers report widespread abuse of OpenClaw skills to deliver info-stealing malware, exposing a new supply chain risk as agent ecosystems scale.
A lightweight node spider. Supports:
import { Crawler, userAgent } from 'ngrab'
import cheerio from 'cheerio'
// For example, crawling the hottest projects on Github
let crawler = new Crawler({
// required && unique
name: 'myCrawler',
// enable bloom filter
bloom: true,
// set random intervals(ms) between requests
interval: () => (Math.random() * 16 + 4) * 1000, // [4s, 20s]
// initial Link
startUrls: ['https://github.com/trending'],
})
// download(name, cb)
crawler.download('trending', async ({ req, res, followLinks, resolveLink }) => {
if (!res) return
// parsing HTML strings
let $ = cheerio.load(res.body.toString())
// extract data
let repoList: Array<{ name: string; href: string }> = [],
$rows = $('.Box-row')
if ($rows.length) {
$rows.each(function (index) {
let $item = $(this)
repoList.push({
name: $('.lh-condensed a .text-normal', $item)
.text()
.replace(/\s+/g, ' ')
.trim(),
href: $('.lh-condensed a', $item).attr('href') as string,
})
})
// print
console.log(repoList) // or store in your Database
// follow links
// repoList.forEach((v) => followLinks(resolveLink(v.href)))
}
})
// start crawling
crawler.run()
The request hook will execute before each request:
// request(name, cb)
crawler.request('headers', async (context) => {
// set custom headers
Object.assign(context.req.headers, {
'Cache-Control': 'no-cache',
'User-Agent': userAgent(), // set random UserAgent
Accept: '*/*',
'Accept-Encoding': 'gzip, deflate, compress',
Connection: 'keep-alive',
})
})
Instead of parsing everything in 'crawler.download()', you can split the parsing code into different routes:
crawler.route({
url: 'https://github.com/trending', // for trending page (compatible with minimatch)
async download(({req, res})){
// parsing ...
}
})
crawler.route({
url: 'https://github.com/*/*', // for repository page
async download(({req, res})){
// parsing ...
}
})
crawler.route({
url: 'https://github.com/*/*/issues', // for issues page
async download(({req, res})){
// parsing ...
}
})
You can provider a proxy server getter when initializing the crawler:
let crawler = new Crawler({
name: 'myCrawler',
startUrls: ['https://github.com/trending'],
async proxy() {
let url = await getProxyUrlFromSomeWhere()
// The return value will be used as a proxy when sending a request
return url
},
})
FAQs
A lightweight node spider
We found that ngrab demonstrated a not healthy version release cadence and project activity because the last version was released a year ago. It has 1 open source maintainer collaborating on the project.
Did you know?

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Security News
Security researchers report widespread abuse of OpenClaw skills to deliver info-stealing malware, exposing a new supply chain risk as agent ecosystems scale.

Security News
Claude Opus 4.6 has uncovered more than 500 open source vulnerabilities, raising new considerations for disclosure, triage, and patching at scale.

Research
/Security News
Malicious dYdX client packages were published to npm and PyPI after a maintainer compromise, enabling wallet credential theft and remote code execution.