Huge News!Announcing our $40M Series B led by Abstract Ventures.Learn More
Socket
Sign inDemoInstall
Socket

crawler-ninja-expired

Package Overview
Dependencies
Maintainers
1
Versions
15
Alerts
File Explorer

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

crawler-ninja-expired

Expired domains finder for crawler.ninja

  • 0.1.1
  • Source
  • npm
  • Socket score

Version published
Weekly downloads
1
decreased by-75%
Maintainers
1
Weekly downloads
 
Created
Source

Expired Domains Finder

This Crawler.Ninja plugin aims to find expired domains by crawling web sites.

Help & Forks welcomed ! or please wait ... work in progress !

How to install

$ npm install crawler-ninja crawler-ninja-expired simple-proxies --save

Crash course

var proxyLoader = require("simple-proxies/lib/proxyfileloader");
var crawler     = require("crawler-ninja");
var ep          = require("crawler-ninja-expired");

var proxyFile = "proxies.txt";

// Load proxies
var config = proxyLoader.config()
                        .setProxyFile(proxyFile)
                        .setCheckProxies(true)
                        .setRemoveInvalidProxies(true);

proxyLoader.loadProxyFile(config, function(error, proxyList) {
    if (error) {
      console.log(error);

    }
    else {
       crawl(proxyList);
    }

});


function crawl(proxyList){
    var c = new crawler.Crawler({
        externalLinks : true,
        images : false,
        scripts : false,
        links : false, //link tags used for css, canonical, ...
        followRedirect : true,
        proxyList : proxyList
    });

    var expired = new ep.Plugin(c);

    c.on("end", function() {

        var end = new Date();
        console.log("Well done Sir !, done in : " + (end - start));
        // the attributes expireds is a map with a key that match to the expired domains
        console.log(ed.expireds.keys())



    });

    var start = new Date();
    c.queue({url : "http://www.site.com"});
}





Using proxies is not mandatory but it is recommanded.Remove the attributes proxyList in the crawler constructor if you don't want to use proxies.

You can find all the crawler options on this page.

Rough todolist

  • Check if the domain is available and/or status (pending, ... )
  • Get Pagerank, TrustFlow and Citation Flow
  • Get the number of linked domains, anchor text infos, ...
  • Use Riak as default persistence layer

ChangeLog

0.1.0

  • Basic implementation that logs domain that match to dns error

Keywords

FAQs

Package last updated on 14 May 2015

Did you know?

Socket

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

  • Package Alerts
  • Integrations
  • Docs
  • Pricing
  • FAQ
  • Roadmap
  • Changelog

Packages

npm

Stay in touch

Get open source security insights delivered straight into your inbox.


  • Terms
  • Privacy
  • Security

Made with ⚡️ by Socket Inc