Comparing version 1.2.0 to 1.2.1
/// <reference types="node" /> | ||
import { IGetToken, IToken } from 'token-types'; | ||
import { IGetToken, IToken } from "token-types"; | ||
import { ITokenizer } from "./"; | ||
@@ -4,0 +4,0 @@ export declare abstract class AbstractTokenizer implements ITokenizer { |
"use strict"; | ||
Object.defineProperty(exports, "__esModule", { value: true }); | ||
var _1 = require("./"); | ||
var AbstractTokenizer = /** @class */ (function () { | ||
@@ -11,3 +12,5 @@ function AbstractTokenizer() { | ||
var buffer = new Buffer(token.len); | ||
return this.readBuffer(buffer, 0, token.len, position).then(function (res) { | ||
return this.readBuffer(buffer, 0, token.len, position).then(function (len) { | ||
if (len < token.len) | ||
throw _1.EndOfFile; | ||
return token.get(buffer, 0); | ||
@@ -14,0 +17,0 @@ }); |
@@ -38,2 +38,3 @@ "use strict"; | ||
} | ||
this.peekOffset = this.position; // clear peek pointer | ||
if (!length) { | ||
@@ -71,6 +72,3 @@ length = buffer.length; | ||
return fs.read(this.fd, buffer, offset, length, this.peekOffset).then(function (res) { | ||
if (res.bytesRead < length) | ||
throw _1.EndOfFile; | ||
_this.peekOffset += res.bytesRead; | ||
// debug("Read:" + buffer.slice(offset, length).toString("hex")); | ||
return res.bytesRead; | ||
@@ -77,0 +75,0 @@ }); |
@@ -42,3 +42,3 @@ "use strict"; | ||
.catch(function (err) { | ||
if (err === then_read_stream_1.StreamReader.EndOfStream) | ||
if (err === then_read_stream_1.EndOfStream) | ||
throw _1.EndOfFile; | ||
@@ -61,5 +61,5 @@ else | ||
} | ||
return this.streamReader.peek(buffer, offset, length) // ToDo: looks like wrong return type is defined in fs.read | ||
return this.streamReader.peek(buffer, offset, length) | ||
.catch(function (err) { | ||
if (err === then_read_stream_1.StreamReader.EndOfStream) | ||
if (err === then_read_stream_1.EndOfStream) | ||
throw _1.EndOfFile; | ||
@@ -66,0 +66,0 @@ else |
{ | ||
"name": "strtok3", | ||
"version": "1.2.0", | ||
"version": "1.2.1", | ||
"description": "A promise based streaming tokenizer", | ||
@@ -37,3 +37,3 @@ "author": { | ||
"@types/mocha": "^2.2.42", | ||
"@types/node": "^8.0.25", | ||
"@types/node": "^8.0.26", | ||
"chai": "^4.1.1", | ||
@@ -51,3 +51,3 @@ "coveralls": "^2.13.1", | ||
"fs-extra": "^4.0.1", | ||
"then-read-stream": "^0.9.6", | ||
"then-read-stream": "^1.0.1", | ||
"token-types": "^0.9.1" | ||
@@ -54,0 +54,0 @@ }, |
Sorry, the diff of this file is too big to display
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package
3478
134718
+ Addedthen-read-stream@1.5.1(transitive)
- Removedthen-read-stream@0.9.6(transitive)
Updatedthen-read-stream@^1.0.1