Socket
Socket
Sign inDemoInstall

gcs-resumable-upload

Package Overview
Dependencies
63
Maintainers
6
Versions
68
Alerts
File Explorer

Advanced tools

Install Socket

Detect and block malicious and high-risk dependencies

Install

    gcs-resumable-upload

Upload a file to Google Cloud Storage with built-in resumable behavior


Version published
Weekly downloads
263K
decreased by-11.71%
Maintainers
6
Install size
4.02 MB
Created
Weekly downloads
 

Changelog

Source

v1.1.0

03-26-2019 07:13 PDT

New Features

  • feat: support ConfigStore configPath option (#194)

Internal / Testing Changes

  • chore: publish to npm using wombat (#197)
  • build: use per-repo npm publish token (#195)
  • build: Add docuploader credentials to node publish jobs (#192)
  • build: use node10 to run samples-test, system-test etc (#190)
  • build: update release configuration

Readme

Source

gcs-resumable-upload

Upload a file to Google Cloud Storage with built-in resumable behavior

$ npm install gcs-resumable-upload
const {upload} = require('gcs-resumable-upload');
const fs = require('fs');

fs.createReadStream('titanic.mov')
  .pipe(upload({ bucket: 'legally-owned-movies', file: 'titanic.mov' }))
  .on('finish', () => {
    // Uploaded!
  });

Or from the command line:

$ npm install -g gcs-resumable-upload
$ cat titanic.mov | gcs-upload legally-owned-movies titanic.mov

If somewhere during the operation, you lose your connection to the internet or your tough-guy brother slammed your laptop shut when he saw what you were uploading, the next time you try to upload to that file, it will resume automatically from where you left off.

How it works

This module stores a file using ConfigStore that is written to when you first start an upload. It is aliased by the file name you are uploading to and holds the first 16kb chunk of data* as well as the unique resumable upload URI. (Resumable uploads are complicated)

If your upload was interrupted, next time you run the code, we ask the API how much data it has already, then simply dump all of the data coming through the pipe that it already has.

After the upload completes, the entry in the config file is removed. Done!

* The first 16kb chunk is stored to validate if you are sending the same data when you resume the upload. If not, a new resumable upload is started with the new data.

Authentication

Oh, right. This module uses google-auth-library and accepts all of the configuration that module does to strike up a connection as config.authConfig. See authConfig.

API

{upload} = require('gcs-resumable-upload')


upload(config)

config
  • Type: object

Configuration object.

config.authClient

If you want to re-use an auth client from google-auth-library, pass an instance here.

config.authConfig
  • Type: object
  • Optional

See authConfig.

config.bucket
  • Type: string
  • Required

The name of the destination bucket.

config.configPath
  • Type: string
  • Optional

Where the gcs-resumable-upload configuration file should be stored on your system. This maps to the configstore option by the same name.

config.file
  • Type: string
  • Required

The name of the destination file.

config.generation
  • Type: number
  • Optional

This will cause the upload to fail if the current generation of the remote object does not match the one provided here.

config.key
  • Type: string|buffer
  • Optional

A customer-supplied encryption key.

config.kmsKeyName
  • Type: string
  • Optional

Resource name of the Cloud KMS key, of the form projects/my-project/locations/global/keyRings/my-kr/cryptoKeys/my-key, that will be used to encrypt the object. Overrides the object metadata's kms_key_name value, if any.

config.metadata
  • Type: object
  • Optional

Any metadata you wish to set on the object.

config.metadata.contentLength

Set the length of the file being uploaded.

config.metadata.contentType

Set the content type of the incoming data.

config.offset
  • Type: number
  • Optional

The starting byte of the upload stream, for resuming an interrupted upload.

config.origin
  • Type: string
  • Optional

Set an Origin header when creating the resumable upload URI.

config.predefinedAcl
  • Type: string
  • Optional

Apply a predefined set of access controls to the created file.

Acceptable values are:

  • authenticatedRead - Object owner gets OWNER access, and allAuthenticatedUsers get READER access.
  • bucketOwnerFullControl - Object owner gets OWNER access, and project team owners get OWNER access.
  • bucketOwnerRead - Object owner gets OWNER access, and project team owners get READER access.
  • private - Object owner gets OWNER access.
  • projectPrivate - Object owner gets OWNER access, and project team members get access according to their roles.
  • publicRead - Object owner gets OWNER access, and allUsers get READER access.
config.private
  • Type: boolean
  • Optional

Make the uploaded file private. (Alias for config.predefinedAcl = 'private')

config.public
  • Type: boolean
  • Optional

Make the uploaded file public. (Alias for config.predefinedAcl = 'publicRead')

config.uri
  • Type: string
  • Optional

If you already have a resumable URI from a previously-created resumable upload, just pass it in here and we'll use that.

config.userProject
  • Type: string
  • Optional

If the bucket being accessed has requesterPays functionality enabled, this can be set to control which project is billed for the access of this file.

--

Events
.on('error', function (err) {})
err
  • Type: Error

Invoked if the authorization failed, the request failed, or the file wasn't successfully uploaded.

.on('response', function (response) {})
resp
  • Type: Object

The response object from Gaxios.

metadata
  • Type: Object

The file's new metadata.

.on('finish', function () {})

The file was uploaded successfully.


{createURI} = require('gcs-resumable-upload);

createURI(config, callback)
callback(err, resumableURI)
callback.err
  • Type: Error

Invoked if the authorization failed or the request to start a resumable session failed.

callback.resumableURI
  • Type: String

The resumable upload session URI.

Keywords

FAQs

Last updated on 26 Mar 2019

Did you know?

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

  • Package Alerts
  • Integrations
  • Docs
  • Pricing
  • FAQ
  • Roadmap

Stay in touch

Get open source security insights delivered straight into your inbox.


  • Terms
  • Privacy
  • Security

Made with ⚡️ by Socket Inc