Comparing version 6.2.2 to 7.0.0-alpha.0
@@ -0,0 +0,0 @@ import { ITokenizer, IFileInfo, IReadChunkOptions } from './types'; |
@@ -1,9 +0,6 @@ | ||
"use strict"; | ||
Object.defineProperty(exports, "__esModule", { value: true }); | ||
exports.AbstractTokenizer = void 0; | ||
const peek_readable_1 = require("peek-readable"); | ||
import { EndOfStreamError } from 'peek-readable'; | ||
/** | ||
* Core tokenizer | ||
*/ | ||
class AbstractTokenizer { | ||
export class AbstractTokenizer { | ||
constructor(fileInfo) { | ||
@@ -27,3 +24,3 @@ /** | ||
if (len < token.len) | ||
throw new peek_readable_1.EndOfStreamError(); | ||
throw new EndOfStreamError(); | ||
return token.get(uint8Array, 0); | ||
@@ -41,3 +38,3 @@ } | ||
if (len < token.len) | ||
throw new peek_readable_1.EndOfStreamError(); | ||
throw new EndOfStreamError(); | ||
return token.get(uint8Array, 0); | ||
@@ -53,3 +50,3 @@ } | ||
if (len < token.len) | ||
throw new peek_readable_1.EndOfStreamError(); | ||
throw new EndOfStreamError(); | ||
return token.get(this.numBuffer, 0); | ||
@@ -65,3 +62,3 @@ } | ||
if (len < token.len) | ||
throw new peek_readable_1.EndOfStreamError(); | ||
throw new EndOfStreamError(); | ||
return token.get(this.numBuffer, 0); | ||
@@ -93,2 +90,1 @@ } | ||
} | ||
exports.AbstractTokenizer = AbstractTokenizer; |
@@ -0,0 +0,0 @@ import { IFileInfo, IReadChunkOptions } from './types'; |
@@ -1,7 +0,4 @@ | ||
"use strict"; | ||
Object.defineProperty(exports, "__esModule", { value: true }); | ||
exports.BufferTokenizer = void 0; | ||
const peek_readable_1 = require("peek-readable"); | ||
const AbstractTokenizer_1 = require("./AbstractTokenizer"); | ||
class BufferTokenizer extends AbstractTokenizer_1.AbstractTokenizer { | ||
import { EndOfStreamError } from 'peek-readable'; | ||
import { AbstractTokenizer } from './AbstractTokenizer'; | ||
export class BufferTokenizer extends AbstractTokenizer { | ||
/** | ||
@@ -44,3 +41,3 @@ * Construct BufferTokenizer | ||
if ((!options.mayBeLess) && bytes2read < options.length) { | ||
throw new peek_readable_1.EndOfStreamError(); | ||
throw new EndOfStreamError(); | ||
} | ||
@@ -56,2 +53,1 @@ else { | ||
} | ||
exports.BufferTokenizer = BufferTokenizer; |
@@ -0,0 +0,0 @@ /// <reference types="node" /> |
@@ -1,8 +0,4 @@ | ||
"use strict"; | ||
Object.defineProperty(exports, "__esModule", { value: true }); | ||
exports.fromBuffer = exports.fromStream = exports.EndOfStreamError = void 0; | ||
const ReadStreamTokenizer_1 = require("./ReadStreamTokenizer"); | ||
const BufferTokenizer_1 = require("./BufferTokenizer"); | ||
var peek_readable_1 = require("peek-readable"); | ||
Object.defineProperty(exports, "EndOfStreamError", { enumerable: true, get: function () { return peek_readable_1.EndOfStreamError; } }); | ||
import { ReadStreamTokenizer } from './ReadStreamTokenizer'; | ||
import { BufferTokenizer } from './BufferTokenizer'; | ||
export { EndOfStreamError } from 'peek-readable'; | ||
/** | ||
@@ -15,7 +11,6 @@ * Construct ReadStreamTokenizer from given Stream. | ||
*/ | ||
function fromStream(stream, fileInfo) { | ||
export function fromStream(stream, fileInfo) { | ||
fileInfo = fileInfo ? fileInfo : {}; | ||
return new ReadStreamTokenizer_1.ReadStreamTokenizer(stream, fileInfo); | ||
return new ReadStreamTokenizer(stream, fileInfo); | ||
} | ||
exports.fromStream = fromStream; | ||
/** | ||
@@ -27,5 +22,4 @@ * Construct ReadStreamTokenizer from given Buffer. | ||
*/ | ||
function fromBuffer(uint8Array, fileInfo) { | ||
return new BufferTokenizer_1.BufferTokenizer(uint8Array, fileInfo); | ||
export function fromBuffer(uint8Array, fileInfo) { | ||
return new BufferTokenizer(uint8Array, fileInfo); | ||
} | ||
exports.fromBuffer = fromBuffer; |
@@ -0,0 +0,0 @@ /// <reference types="node" /> |
@@ -1,8 +0,5 @@ | ||
"use strict"; | ||
Object.defineProperty(exports, "__esModule", { value: true }); | ||
exports.fromFile = exports.FileTokenizer = void 0; | ||
const AbstractTokenizer_1 = require("./AbstractTokenizer"); | ||
const peek_readable_1 = require("peek-readable"); | ||
const fs = require("./FsPromise"); | ||
class FileTokenizer extends AbstractTokenizer_1.AbstractTokenizer { | ||
import { AbstractTokenizer } from './AbstractTokenizer'; | ||
import { EndOfStreamError } from 'peek-readable'; | ||
import * as fs from './FsPromise'; | ||
export class FileTokenizer extends AbstractTokenizer { | ||
constructor(fd, fileInfo) { | ||
@@ -44,3 +41,3 @@ super(fileInfo); | ||
if (res.bytesRead < length && (!options || !options.mayBeLess)) { | ||
throw new peek_readable_1.EndOfStreamError(); | ||
throw new EndOfStreamError(); | ||
} | ||
@@ -59,3 +56,3 @@ return res.bytesRead; | ||
if ((!options.mayBeLess) && res.bytesRead < options.length) { | ||
throw new peek_readable_1.EndOfStreamError(); | ||
throw new EndOfStreamError(); | ||
} | ||
@@ -68,4 +65,3 @@ return res.bytesRead; | ||
} | ||
exports.FileTokenizer = FileTokenizer; | ||
async function fromFile(sourceFilePath) { | ||
export async function fromFile(sourceFilePath) { | ||
const stat = await fs.stat(sourceFilePath); | ||
@@ -78,2 +74,1 @@ if (!stat.isFile) { | ||
} | ||
exports.fromFile = fromFile; |
@@ -0,0 +0,0 @@ /** |
@@ -1,11 +0,8 @@ | ||
"use strict"; | ||
/** | ||
* Module convert fs functions to promise based functions | ||
*/ | ||
Object.defineProperty(exports, "__esModule", { value: true }); | ||
exports.readFile = exports.writeFileSync = exports.writeFile = exports.read = exports.open = exports.close = exports.stat = exports.createReadStream = exports.pathExists = void 0; | ||
const fs = require("fs"); | ||
exports.pathExists = fs.existsSync; | ||
exports.createReadStream = fs.createReadStream; | ||
async function stat(path) { | ||
import * as fs from 'fs'; | ||
export const pathExists = fs.existsSync; | ||
export const createReadStream = fs.createReadStream; | ||
export async function stat(path) { | ||
return new Promise((resolve, reject) => { | ||
@@ -20,4 +17,3 @@ fs.stat(path, (err, stats) => { | ||
} | ||
exports.stat = stat; | ||
async function close(fd) { | ||
export async function close(fd) { | ||
return new Promise((resolve, reject) => { | ||
@@ -32,4 +28,3 @@ fs.close(fd, err => { | ||
} | ||
exports.close = close; | ||
async function open(path, mode) { | ||
export async function open(path, mode) { | ||
return new Promise((resolve, reject) => { | ||
@@ -44,4 +39,3 @@ fs.open(path, mode, (err, fd) => { | ||
} | ||
exports.open = open; | ||
async function read(fd, buffer, offset, length, position) { | ||
export async function read(fd, buffer, offset, length, position) { | ||
return new Promise((resolve, reject) => { | ||
@@ -56,4 +50,3 @@ fs.read(fd, buffer, offset, length, position, (err, bytesRead, _buffer) => { | ||
} | ||
exports.read = read; | ||
async function writeFile(path, data) { | ||
export async function writeFile(path, data) { | ||
return new Promise((resolve, reject) => { | ||
@@ -68,8 +61,6 @@ fs.writeFile(path, data, err => { | ||
} | ||
exports.writeFile = writeFile; | ||
function writeFileSync(path, data) { | ||
export function writeFileSync(path, data) { | ||
fs.writeFileSync(path, data); | ||
} | ||
exports.writeFileSync = writeFileSync; | ||
async function readFile(path) { | ||
export async function readFile(path) { | ||
return new Promise((resolve, reject) => { | ||
@@ -84,2 +75,1 @@ fs.readFile(path, (err, buffer) => { | ||
} | ||
exports.readFile = readFile; |
@@ -0,0 +0,0 @@ /// <reference types="node" /> |
@@ -1,11 +0,5 @@ | ||
"use strict"; | ||
Object.defineProperty(exports, "__esModule", { value: true }); | ||
exports.fromStream = exports.fromBuffer = exports.EndOfStreamError = exports.fromFile = void 0; | ||
const fs = require("./FsPromise"); | ||
const core = require("./core"); | ||
var FileTokenizer_1 = require("./FileTokenizer"); | ||
Object.defineProperty(exports, "fromFile", { enumerable: true, get: function () { return FileTokenizer_1.fromFile; } }); | ||
var core_1 = require("./core"); | ||
Object.defineProperty(exports, "EndOfStreamError", { enumerable: true, get: function () { return core_1.EndOfStreamError; } }); | ||
Object.defineProperty(exports, "fromBuffer", { enumerable: true, get: function () { return core_1.fromBuffer; } }); | ||
import * as fs from './FsPromise'; | ||
import * as core from './core'; | ||
export { fromFile } from './FileTokenizer'; | ||
export { EndOfStreamError, fromBuffer } from './core'; | ||
/** | ||
@@ -18,3 +12,3 @@ * Construct ReadStreamTokenizer from given Stream. | ||
*/ | ||
async function fromStream(stream, fileInfo) { | ||
export async function fromStream(stream, fileInfo) { | ||
fileInfo = fileInfo ? fileInfo : {}; | ||
@@ -28,2 +22,1 @@ if (stream.path) { | ||
} | ||
exports.fromStream = fromStream; |
@@ -0,0 +0,0 @@ /// <reference types="node" /> |
@@ -1,11 +0,8 @@ | ||
"use strict"; | ||
Object.defineProperty(exports, "__esModule", { value: true }); | ||
exports.ReadStreamTokenizer = void 0; | ||
const AbstractTokenizer_1 = require("./AbstractTokenizer"); | ||
const peek_readable_1 = require("peek-readable"); | ||
import { AbstractTokenizer } from './AbstractTokenizer'; | ||
import { EndOfStreamError, StreamReader } from 'peek-readable'; | ||
const maxBufferSize = 256000; | ||
class ReadStreamTokenizer extends AbstractTokenizer_1.AbstractTokenizer { | ||
export class ReadStreamTokenizer extends AbstractTokenizer { | ||
constructor(stream, fileInfo) { | ||
super(fileInfo); | ||
this.streamReader = new peek_readable_1.StreamReader(stream); | ||
this.streamReader = new StreamReader(stream); | ||
} | ||
@@ -57,3 +54,3 @@ /** | ||
if ((!options || !options.mayBeLess) && bytesRead < length) { | ||
throw new peek_readable_1.EndOfStreamError(); | ||
throw new EndOfStreamError(); | ||
} | ||
@@ -88,3 +85,3 @@ return bytesRead; | ||
catch (err) { | ||
if (options && options.mayBeLess && err instanceof peek_readable_1.EndOfStreamError) { | ||
if (options && options.mayBeLess && err instanceof EndOfStreamError) { | ||
return 0; | ||
@@ -95,3 +92,3 @@ } | ||
if ((!options.mayBeLess) && bytesRead < options.length) { | ||
throw new peek_readable_1.EndOfStreamError(); | ||
throw new EndOfStreamError(); | ||
} | ||
@@ -117,2 +114,1 @@ } | ||
} | ||
exports.ReadStreamTokenizer = ReadStreamTokenizer; |
@@ -0,0 +0,0 @@ /// <reference types="node" /> |
@@ -1,2 +0,1 @@ | ||
"use strict"; | ||
Object.defineProperty(exports, "__esModule", { value: true }); | ||
export {}; |
{ | ||
"name": "strtok3", | ||
"version": "6.2.2", | ||
"version": "7.0.0-alpha.0", | ||
"description": "A promise based streaming tokenizer", | ||
@@ -29,3 +29,3 @@ "author": { | ||
"engines": { | ||
"node": ">=10" | ||
"node": ">=12.20" | ||
}, | ||
@@ -38,2 +38,6 @@ "repository": { | ||
"main": "lib/index.js", | ||
"exports": { | ||
".": "./lib/index.js", | ||
"./core": "./lib/core.js" | ||
}, | ||
"types": "lib/index.d.ts", | ||
@@ -73,3 +77,3 @@ "files": [ | ||
"dependencies": { | ||
"peek-readable": "^4.0.0" | ||
"peek-readable": "^5.0.0-alpha.0" | ||
}, | ||
@@ -76,0 +80,0 @@ "keywords": [ |
@@ -0,0 +0,0 @@ ![Node.js CI](https://github.com/Borewit/strtok3/workflows/Node.js%20CI/badge.svg) |
Sorry, the diff of this file is not supported yet
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package
Filesystem access
Supply chain riskAccesses the file system, and could potentially read sensitive data.
Found 1 instance in 1 package
No v1
QualityPackage is not semver >=1. This means it is not stable and does not support ^ ranges.
Found 1 instance in 1 package
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package
Filesystem access
Supply chain riskAccesses the file system, and could potentially read sensitive data.
Found 1 instance in 1 package
48636
727
2
+ Addedpeek-readable@5.2.0(transitive)
- Removedpeek-readable@4.1.0(transitive)
Updatedpeek-readable@^5.0.0-alpha.0