Comparing version 1.1.3 to 1.2.0
@@ -6,3 +6,3 @@ /// <reference types="node" /> | ||
fileSize?: number; | ||
offset: number; | ||
position: number; | ||
private numBuffer; | ||
@@ -9,0 +9,0 @@ /** |
"use strict"; | ||
Object.defineProperty(exports, "__esModule", { value: true }); | ||
var AbstractTokenizer = (function () { | ||
var AbstractTokenizer = /** @class */ (function () { | ||
function AbstractTokenizer() { | ||
this.offset = 0; | ||
this.position = 0; | ||
this.numBuffer = new Buffer(4); | ||
@@ -7,0 +7,0 @@ } |
@@ -16,3 +16,3 @@ "use strict"; | ||
var _1 = require("./"); | ||
var FileTokenizer = (function (_super) { | ||
var FileTokenizer = /** @class */ (function (_super) { | ||
__extends(FileTokenizer, _super); | ||
@@ -37,3 +37,3 @@ function FileTokenizer(fd, fileSize) { | ||
if (position) { | ||
this.offset = position; | ||
this.position = position; | ||
} | ||
@@ -43,6 +43,6 @@ if (!length) { | ||
} | ||
return fs.read(this.fd, buffer, offset, length, this.offset).then(function (res) { | ||
return fs.read(this.fd, buffer, offset, length, this.position).then(function (res) { | ||
if (res.bytesRead < length) | ||
throw _1.EndOfFile; | ||
_this.offset += res.bytesRead; | ||
_this.position += res.bytesRead; | ||
// debug("Read:" + buffer.slice(offset, length).toString("hex")); | ||
@@ -66,4 +66,4 @@ return res.bytesRead; | ||
} | ||
else if (this.peekOffset < this.offset) { | ||
this.peekOffset = this.offset; | ||
else if (this.peekOffset < this.position) { | ||
this.peekOffset = this.position; | ||
} | ||
@@ -85,3 +85,3 @@ if (!length) { | ||
FileTokenizer.prototype.ignore = function (length) { | ||
this.offset += length; | ||
this.position += length; | ||
return Promise.resolve(null); | ||
@@ -88,0 +88,0 @@ }; |
@@ -19,3 +19,3 @@ /// <reference types="node" /> | ||
*/ | ||
offset: number; | ||
position: number; | ||
/** | ||
@@ -22,0 +22,0 @@ * Peek (read ahead) buffer from tokenizer |
@@ -11,3 +11,3 @@ "use strict"; | ||
exports.EndOfFile = new Error("End-Of-File"); | ||
var IgnoreType = (function () { | ||
var IgnoreType = /** @class */ (function () { | ||
/** | ||
@@ -14,0 +14,0 @@ * @param len Number of bytes to ignore (skip) |
@@ -16,3 +16,3 @@ "use strict"; | ||
var then_read_stream_1 = require("then-read-stream"); | ||
var ReadStreamTokenizer = (function (_super) { | ||
var ReadStreamTokenizer = /** @class */ (function (_super) { | ||
__extends(ReadStreamTokenizer, _super); | ||
@@ -39,3 +39,3 @@ function ReadStreamTokenizer(stream, fileSize) { | ||
.then(function (bytesRead) { | ||
_this.offset += bytesRead; | ||
_this.position += bytesRead; | ||
return bytesRead; | ||
@@ -42,0 +42,0 @@ }) |
{ | ||
"name": "strtok3", | ||
"version": "1.1.3", | ||
"version": "1.2.0", | ||
"description": "A promise based streaming tokenizer", | ||
@@ -34,7 +34,6 @@ "author": { | ||
"devDependencies": { | ||
"@types/chai": "^4.0.3", | ||
"@types/es6-promise": "0.0.32", | ||
"@types/chai": "^4.0.4", | ||
"@types/fs-extra": "^4.0.0", | ||
"@types/mocha": "^2.2.41", | ||
"@types/node": "^8.0.24", | ||
"@types/mocha": "^2.2.42", | ||
"@types/node": "^8.0.25", | ||
"chai": "^4.1.1", | ||
@@ -45,9 +44,10 @@ "coveralls": "^2.13.1", | ||
"ts-node": "^3.3.0", | ||
"tslint": "^5.6.0", | ||
"tslint": "^5.7.0", | ||
"typescript": "^2.5.1" | ||
}, | ||
"dependencies": { | ||
"@types/es6-promise": "0.0.33", | ||
"es6-promise": "^4.1.1", | ||
"fs-extra": "^4.0.1", | ||
"then-read-stream": "^0.9.5", | ||
"then-read-stream": "^0.9.6", | ||
"token-types": "^0.9.1" | ||
@@ -54,0 +54,0 @@ }, |
Sorry, the diff of this file is too big to display
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package
11
134802
5
3477
+ Added@types/es6-promise@0.0.33
+ Added@types/es6-promise@0.0.33(transitive)
Updatedthen-read-stream@^0.9.6