Huge News!Announcing our $40M Series B led by Abstract Ventures.Learn More
Socket
Sign inDemoInstall
Socket

knox-mpu-alt

Package Overview
Dependencies
Maintainers
3
Versions
8
Alerts
File Explorer

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

knox-mpu-alt

Provide multi part upload functionality to Amazon S3 using the knox library. Forked from knox-mpu.

  • 0.1.11
  • Source
  • npm
  • Socket score

Version published
Weekly downloads
453
decreased by-8.85%
Maintainers
3
Weekly downloads
 
Created
Source

knox-mpu-alt

This is a fork of https://github.com/nathanoehlman/knox-mpu. If/when the relevant PRs associated w/ this module get merged, we can fold it back into the main repo (we just couldn't wait any longer for it).

~mike and scott

A Node.js client designed to make large file uploads to Amazon S3 via the MultiPartUpload API simple and easy. It's built on top of the excellent Knox library from the guys over at LearnBoost.

Features

  • Simple and easy to use
  • Pipe either a file, or a stream directly to S3 (No need to know the content length first!)
  • Automatically separates a file/stream into appropriate sized segments for upload
  • Asynchronous uploading of segments
  • Handy events to track your upload progress

Planned

  • Better error handling (reuploading failed parts, etc)

Installing

Installation is done via NPM, by running npm install knox-mpu-alt

Examples

Uploading a stream

To upload a stream, simply pass the stream when constructing the MultiPartUpload. The upload will then listen to the stream, and create parts from incoming data stream. When a part reaches the minimum part size, it will attempt to upload it to S3.


// Create a Knox client first
var client = knox.createClient({ ... }),
    upload = null;


upload = new MultiPartUpload(
            {
                client: client,
                objectName: 'destination.txt', // Amazon S3 object name
                stream: stream
            },
            // Callback handler
            function(err, body) {
                // If successful, will return body, containing Location, Bucket, Key, ETag and size of the object
                /*
                  {
                      Location: 'http://Example-Bucket.s3.amazonaws.com/destination.txt',
                      Bucket: 'Example-Bucket',
                      Key: 'destination.txt',
                      ETag: '"3858f62230ac3c915f300c664312c11f-9"',
                      size: 7242880
                  }
                */
            }
        );
Uploading a file

To upload a file, pass the path to the file in the constructor. Knox-mpu will split the file into parts and upload them.


// Create a Knox client first
var client = knox.createClient({ ... }),
    upload = null;


upload = new MultiPartUpload(
            {
                client: client,
                objectName: 'destination.txt', // Amazon S3 object name
                file: ... // path to the file
            },
            // Callback handler
            function(err, body) {
                // If successful, will return body, containing Location, Bucket, Key, ETag and size of the object
                /*
                  {
                      Location: 'http://Example-Bucket.s3.amazonaws.com/destination.txt',
                      Bucket: 'Example-Bucket',
                      Key: 'destination.txt',
                      ETag: '"3858f62230ac3c915f300c664312c11f-9"',
                      size: 7242880
                  }
                */
            }
        );

Options

The following options can be passed to the MultiPartUpload constructor -

  • client Required The knox client to use for this upload request
  • objectName Required The destination object name/path on S3 for this upload
  • stream The stream to upload (required if file is not being supplied)
  • file The path to the file (required if stream is not being supplied)
  • headers Any additional headers to include on the requests
  • partSize The minimum size of the parts to upload (default to 5MB).
  • batchSize The maximum number of concurrent parts that can be uploading at any one time (default is 4)
  • maxUploadSize The maximum size of the file to upload (default inifinity). Useful if there is a stream with unknown length.
  • noDisk If true, parts will be kept in-memory instead of written to temp files (default to false).
  • maxRetries Number of times to retry failed part upload (default is 0 for no retry).

Events

The MultiPartUpload will emit a number of events -

  • initiated Emitted when the multi part upload has been initiated, and received an upload ID. Passes the upload id through as the first argument to the event
  • uploading Emitted each time a part starts uploading. The part id is passed as the first argument.
  • uploaded Emitted each time a part finishes uploading. Passes through an object containing the part id and Amazon ETag for the uploaded part.
  • error Emitted each time a part upload fails. Passes an object containing the part id and error message
  • completed Emitted when the upload has completed successfully. Contains the object information from Amazon S3 (location, bucket, key and ETag)

Keywords

FAQs

Package last updated on 27 Feb 2015

Did you know?

Socket

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

  • Package Alerts
  • Integrations
  • Docs
  • Pricing
  • FAQ
  • Roadmap
  • Changelog

Packages

npm

Stay in touch

Get open source security insights delivered straight into your inbox.


  • Terms
  • Privacy
  • Security

Made with ⚡️ by Socket Inc