Comparing version 7.0.0-alpha.1 to 7.0.0-alpha.2
@@ -82,5 +82,10 @@ import { EndOfStreamError } from 'peek-readable'; | ||
normalizeOptions(uint8Array, options) { | ||
options = Object.assign({ offset: 0, length: uint8Array.length - ((options && options.offset) ? options.offset : 0), position: this.position }, options); | ||
options = { | ||
offset: 0, | ||
length: uint8Array.length - ((options && options.offset) ? options.offset : 0), | ||
position: this.position, | ||
...options | ||
}; | ||
return options; | ||
} | ||
} |
import { IFileInfo, IReadChunkOptions } from './types'; | ||
import { AbstractTokenizer } from './AbstractTokenizer'; | ||
import { AbstractTokenizer } from './AbstractTokenizer.js'; | ||
export declare class BufferTokenizer extends AbstractTokenizer { | ||
@@ -4,0 +4,0 @@ private uint8Array; |
import { EndOfStreamError } from 'peek-readable'; | ||
import { AbstractTokenizer } from './AbstractTokenizer'; | ||
import { AbstractTokenizer } from './AbstractTokenizer.js'; | ||
export class BufferTokenizer extends AbstractTokenizer { | ||
@@ -4,0 +4,0 @@ /** |
/// <reference types="node" /> | ||
import { ReadStreamTokenizer } from './ReadStreamTokenizer'; | ||
import * as Stream from 'stream'; | ||
import { BufferTokenizer } from './BufferTokenizer'; | ||
import { IFileInfo } from './types'; | ||
import { ReadStreamTokenizer } from './ReadStreamTokenizer.js'; | ||
import * as Stream from 'node:stream'; | ||
import { BufferTokenizer } from './BufferTokenizer.js'; | ||
import { IFileInfo } from './types.js'; | ||
export { EndOfStreamError } from 'peek-readable'; | ||
@@ -7,0 +7,0 @@ export { ITokenizer, IFileInfo } from './types'; |
@@ -1,3 +0,3 @@ | ||
import { ReadStreamTokenizer } from './ReadStreamTokenizer'; | ||
import { BufferTokenizer } from './BufferTokenizer'; | ||
import { ReadStreamTokenizer } from './ReadStreamTokenizer.js'; | ||
import { BufferTokenizer } from './BufferTokenizer.js'; | ||
export { EndOfStreamError } from 'peek-readable'; | ||
@@ -4,0 +4,0 @@ /** |
/// <reference types="node" /> | ||
import { AbstractTokenizer } from './AbstractTokenizer'; | ||
import { AbstractTokenizer } from './AbstractTokenizer.js'; | ||
import { IFileInfo, IReadChunkOptions } from './types'; | ||
@@ -4,0 +4,0 @@ export declare class FileTokenizer extends AbstractTokenizer { |
@@ -1,4 +0,4 @@ | ||
import { AbstractTokenizer } from './AbstractTokenizer'; | ||
import { AbstractTokenizer } from './AbstractTokenizer.js'; | ||
import { EndOfStreamError } from 'peek-readable'; | ||
import * as fs from './FsPromise'; | ||
import * as fs from './FsPromise.js'; | ||
export class FileTokenizer extends AbstractTokenizer { | ||
@@ -5,0 +5,0 @@ constructor(fd, fileInfo) { |
@@ -5,3 +5,3 @@ /** | ||
/// <reference types="node" /> | ||
import * as fs from 'fs'; | ||
import * as fs from 'node:fs'; | ||
export interface IReadResult { | ||
@@ -8,0 +8,0 @@ bytesRead: number; |
/** | ||
* Module convert fs functions to promise based functions | ||
*/ | ||
import * as fs from 'fs'; | ||
import * as fs from 'node:fs'; | ||
export const pathExists = fs.existsSync; | ||
@@ -6,0 +6,0 @@ export const createReadStream = fs.createReadStream; |
/// <reference types="node" /> | ||
import * as Stream from 'stream'; | ||
import { ReadStreamTokenizer } from './ReadStreamTokenizer'; | ||
import * as core from './core'; | ||
export { fromFile } from './FileTokenizer'; | ||
export { ITokenizer, EndOfStreamError, fromBuffer, IFileInfo } from './core'; | ||
import { ReadStreamTokenizer } from './ReadStreamTokenizer.js'; | ||
import * as core from './core.js'; | ||
export { fromFile } from './FileTokenizer.js'; | ||
export { ITokenizer, EndOfStreamError, fromBuffer, IFileInfo } from './core.js'; | ||
export { IToken, IGetToken } from '@tokenizer/token'; | ||
@@ -8,0 +8,0 @@ /** |
@@ -1,5 +0,5 @@ | ||
import * as fs from './FsPromise'; | ||
import * as core from './core'; | ||
export { fromFile } from './FileTokenizer'; | ||
export { EndOfStreamError, fromBuffer } from './core'; | ||
import * as fs from './FsPromise.js'; | ||
import * as core from './core.js'; | ||
export { fromFile } from './FileTokenizer.js'; | ||
export { EndOfStreamError, fromBuffer } from './core.js'; | ||
/** | ||
@@ -6,0 +6,0 @@ * Construct ReadStreamTokenizer from given Stream. |
/// <reference types="node" /> | ||
import { AbstractTokenizer } from './AbstractTokenizer'; | ||
import * as Stream from 'stream'; | ||
import { AbstractTokenizer } from './AbstractTokenizer.js'; | ||
import * as Stream from 'node:stream'; | ||
import { IFileInfo, IReadChunkOptions } from './types'; | ||
@@ -5,0 +5,0 @@ export declare class ReadStreamTokenizer extends AbstractTokenizer { |
@@ -1,2 +0,2 @@ | ||
import { AbstractTokenizer } from './AbstractTokenizer'; | ||
import { AbstractTokenizer } from './AbstractTokenizer.js'; | ||
import { EndOfStreamError, StreamReader } from 'peek-readable'; | ||
@@ -3,0 +3,0 @@ const maxBufferSize = 256000; |
{ | ||
"name": "strtok3", | ||
"version": "7.0.0-alpha.1", | ||
"version": "7.0.0-alpha.2", | ||
"description": "A promise based streaming tokenizer", | ||
@@ -20,5 +20,6 @@ "author": { | ||
"eslint": "eslint lib test --ext .ts --ignore-pattern *.d.ts", | ||
"fix": "eslint lib test --ext .ts --ignore-pattern *.d.ts --fix", | ||
"lint-md": "remark -u preset-lint-recommended .", | ||
"lint": "npm run lint-md && npm run eslint", | ||
"test": "mocha --require ts-node/register --require source-map-support/register --full-trace test/test.ts", | ||
"test": "mocha", | ||
"test-coverage": "nyc npm run test", | ||
@@ -66,7 +67,6 @@ "send-coveralls": "nyc report --reporter=text-lcov | coveralls", | ||
"eslint-plugin-prefer-arrow": "^1.2.3", | ||
"mocha": "^9.0.2", | ||
"mocha": "^9.0.3", | ||
"nyc": "^15.1.0", | ||
"remark-cli": "^9.0.0", | ||
"remark-preset-lint-recommended": "^5.0.0", | ||
"source-map-support": "^0.5.16", | ||
"token-types": "^4.1.0", | ||
@@ -78,3 +78,3 @@ "ts-node": "^10.1.0", | ||
"dependencies": { | ||
"peek-readable": "^5.0.0-alpha.1" | ||
"peek-readable": "^5.0.0-alpha.2" | ||
}, | ||
@@ -81,0 +81,0 @@ "keywords": [ |
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package
Filesystem access
Supply chain riskAccesses the file system, and could potentially read sensitive data.
Found 1 instance in 1 package
48722
22
732
0
Updatedpeek-readable@^5.0.0-alpha.2