libxget-js
Non-interractive, chunk-based, web content retriever
Installing
Via NPM:
npm install libxget
npm install -g libxget
This installs a CLI command accessible with the xget
command.
$ xget -v
v0.10.0
Usage
CLI
The xget
command, utilizes the library to retrieve web content by its chunks according to specification
xget https://google.com/doodle.png
xget https://google.com/doodle.png image.png
xget https://myserver.io/runtime.log --no-bar | less
xget https://service.com/movie.mp4 | vlc -
Use --help
to see full usage documentation.
Programmatically
import xget from "libxget";
xget("https://github.com/microsoft/TypeScript/archive/master.zip", {
chunks: 10,
retries: 10,
}).pipe(fs.createWriteStream("master.zip"));
Get the master branch of the Typescript repository.
With 10 simultaneous downloads. Retrying each one to a max of 10.
How it works
|progress| |=========|
/- xresilient[axios] -> || part || -> || cache || -\
/- xresilient[axios] -> || part || -> || cache || -\
/- xresilient[axios] -> || part || -> || cache || -\
URL -> xresilient[axios] -> || part || -> || cache || -> chunkmerger [ -> hasher ] -> output
\- xresilient[axios] -> || part || -> || cache || -/
\- xresilient[axios] -> || part || -> || cache || -/
\- xresilient[axios] -> || part || -> || cache || -/
|progress| |=========|
xget, using the axios library first infers from an abrupt GET response whether or not the server supports byte-ranges through Accept-Ranges or Content-Range.
In the event that it does, it opens N connections feeding in non-overlapping segments of the resource. In order to retry broken connections, xget wraps the request generator in xresilient streams to ensure proper retries and probable completion of each chunk stream. The streams are piped and tracked through the progress bar and into a caching stream and then all chunks are merged sequentially, in-place and in-order and piped into an optional hasher and finally the output.
The purpose of the caching stream is to ensure that other chunks can begin while the merger is still writing previous chunks. Liberating the download speed from the write speed, recieved chunks are buffered in memory to a maximum cache limit.
The cacher also comes after the progress bar to ensure it properly measures the download speed and not the disk write speed, in the case where disk speed is slower than the network, although unlikely.
The purpose of the hasher is to check the integrity of the merged chunks while writing instead of that being a separate process. Very useful for large-file downloads, so you only process the file once, while downloading.
API
xget(url[, options])
chunks
: <number> Maximum number of non-overlapping chunk connections. Default: 5
retries
: <number> Number of retries for each chunk. Default: 5
timeout
: <number> Network response timeout (ms). Default: 20000
start
: <number> Position to start feeding the stream from. Default: 0
auto
: <boolean> Whether or not to start the request automatically or wait for a request.start()
call (useful when chaining events you want to fire in order). Default: true
size
: <number> Number of bytes to stream off the response.hash
: <string> Hash algorithm to use to create a crypto.Hash instance computing the stream hash.cache
: <number> Whether or not to use an in-memory cache to enable read-aheads of pending chunks.cacheSize
: <boolean> Custom maximum cache size (bytes).use
: <object> Key-value pairs of middlewares with which to pipe the response object through. keys are strings, values are Transformer generating functions (Alternatively, use the xget.use() method).with
: <object> Key-value pairs of middlewares with which to pipe the dataslice object through. keys are strings, values are functions whose return values are accessible within the store. (Alternatively, use the xget.with() method).headHandler
: <HeadHandler> An interceptor for the initial headers, useful for programmatically defining a range offset;
xget.store: Map
A map whose keys and values are tags and return types of content processed within the withStack of the xget object.
xget(URL)
.with("variable", () => 5)
.once("set", (store) => {
console.log(store.get("variable"));
})
.pipe(FILE);
A readonly property that tells whether or not the xget instance has ended.
A readonly property that tells whether or not the xget instance has been loaded.
xget.bytesRead: Number
A readonly property that tells how many bytes has been processed by the underlying streams.
The core multi-chunk request instance.
new xget.XGETStream(url[, options])
Event: 'end'
The 'end'
event is emitted after the data from the URL has been fully flushed.
Event: 'set'
store
: <xget.store> The shared internal data store.
The 'set'
event is emitted after all the middlewares defined in the with
option of the XGETOptions or with the xget.with() method.
This event is fired after the 'loaded'
event.
Event: 'error'
err
: <Error> The error instance.
The 'error'
event is emitted once a chunk has met it's maximum number of retries.
At which point, it would abruptly destroy other chunk connections.
Event: 'retry'
retrySlice
:
meta
: <boolean> Whether or not the error causing the retry was caused while getting the URL metadata. i.e before any streams are employed.index
: <number> The index count of the chunk.retryCount
: <number> The number of retry iterations so far.maxRetries
: <number> The maximum number of retries possible.bytesRead
: <number> The number of bytes previously read (if any).totalBytes
: <number> The total number of bytes that are to be read by the stream.lastErr
: <Error> The error emitted by the previous stream.store
: <xget.store> The shared internal data store.
The 'retry'
event is emitted by every chunk once it has been re-initialized underneath.
Based on the spec of the xresilient module, chunks are reinitialized once an error event is met.
Event: 'loaded'
loadData
: <LoadData> The pre-computed config for the loaded data slice.
This is emitted right after the initial headers data is gotten, preprocessed, parsed and used to tailor the configuration for the chunk setup.
This loadData
contains information like the actual size of the remote file and whether or not the server supports multiple connections, chunking, file resumption, etc.
This event is fired after calling the headHandler and prior to the 'set'
event.
xget.start()
Starts the request process if options.auto
was set to false.
Returns true
if the request was started, false
if it had already been started.
xget.getHash([encoding])
encoding
: <string> The character encoding to use. Default: 'hex'
- Returns: <Buffer> | <string>
Calculates the digest of all data that has been processed by the library and its middleware transformers.
This, creates a deep copy of the internal state of the current crypto.Hash object of which it calculates the digest.
This ensures you can get a hash of an instancce of the data even while still streaming from the URL response.
xget.getHashAlgorithm()
Returns the hash algorithm if any is in use.
xget.setHeadHandler()
fn
: <HeadHandler> Handler to be set.- Returns: <boolean> Whether or not the handler was successfully set.
Sets an interceptor for the initial headers, useful for programmatically defining a range offset. Returns false
if the request has already been loaded, true
if successfully set.
xget.setCacheCapacity()
Set maximum capacity for internal cache.
xget.use(tag, handler)
Add a named handler to the use middleware stack whose return value would be used to transform the response stream in a series of pipes.
The handler
method is called after the stream is requested from and we start pumping the underlying request
instances for a data response stream.
The core expects the handler
to return a stream.Duplex instance. (A readable, writable stream) to transform or passthrough the raw data streams along the way.
xget(URL)
.use("compressor", () => zlib.createGzip())
.pipe(createWriteStreamSomehow());
xget.with(tag, handler)
Add a named handler
to the with middleware stack whose return value would be stored within the store after execution.
xget(URL)
.with("bar", ({ size }) => progressBar(size))
.use("bar", (_, store) => store.get("bar").genStream())
.once("set", (store) => store.get("bar").print("Downloading..."))
.pipe(createWriteStreamSomehow());
xget.getErrContext(err)
err
: <Error>- Returns: <Object>
raw
: <Error>tag
: <string> The tag of the middleware function as defined.source
: <'xget:with'
> | <'xget:use'
> The type of middleware from which the error was emitted.
Extract data from an error if it was either thrown from within a UseMiddlewareFn or a WithMiddlewareFn function.
xget(URL)
.use('errorThrower', () => {
throw new Error('Custom error being thrown');
})
.once('error', err => {
const ({tag, source}) = xget.getErrContext(err);
if (source)
console.log(`Error thrown from within the [${tag}] method of the [${source}] middlware`);
})
.pipe(createWriteStreamSomehow());
props
: <object>
chunks
: <number> Number of chunks the resource can simultaneously provide.headers
: <IncomingHttpHeaders> GET headers from the URL.start
: <number> Relayed .start
field from XGETOptions.totalSize
: <number> Actual size of the resource without an offset.acceptsRanges
: <boolean> Whether or not the URL resource accepts byte ranges.
- Returns: <number | void> An offset to begin streaming from. Analogous to the
.start
field in XGETOptions. If void, defaults to .start
or 0
;
An interceptor for the initial GET data, useful for programmatically defining a range offset.
url
: <string> The URL specified.size
: <number> Finite number returned if server responds appropriately, else Infinity
.start
: <number> Sticks to specification if server allows chunking via content-ranges
else, resets to 0
.chunkable
: <number> Whether or not the URL feed can be chunked, supporting simultaneous connections.totalSize
: <number> Actual size of the resource without an offset.chunkStack
: <ChunkLoadInstance[]> The chunkstack array.headers
: <IncomingHttpHeaders> The headers object.
ChunkLoadInstance: Object
min
: <number> The minimum extent for the chunk segment range.max
: <number> The maximum extent for the chunk segment range.size
: <number> The total size of the chunk segment.stream
: <ResilientStream> A resilient stream that wraps around a request instance.
WithMiddlewareFn: Function
This handler
is called immediately after metadata from URL is loaded that describes the response.
That is, pre-streaming data from the GET response like size (content-length), content-type, filename (content-disposition), whether or not it's chunkable (accept-ranges, content-range) and a couple of other criterias.
This information is passed into a handler whose return value is filed within the store referenced by the tag
.
UseMiddlewareFn: Function
CLI Info
- To avoid the terminal being cluttered while using pipes, direct other chained binaries'
stdout
and stderr
to /dev/null
xget https://myserver.com/movie.mp4 | vlc - > /dev/null 2>&1
Development
Building
Feel free to clone, use in adherance to the license. Pull requests are very much welcome.
git clone https://github.com/miraclx/libxget-js.git
cd libxget-js
npm install
License
Apache 2.0 © Miraculous Owonubi (@miraclx) <omiraculous@gmail.com>