feat: Created a mini nodeJS server with NewMan for testing without PostMan GUI.

This will mimic a run in a CD/CI environment or docker container.
This commit is contained in:
Simon Priet
2021-09-08 14:01:19 +02:00
parent 5fbd7c88fa
commit e69a613a37
5610 changed files with 740417 additions and 3 deletions

21
node_modules/csv-parse/LICENSE generated vendored Normal file
View File

@@ -0,0 +1,21 @@
The MIT License (MIT)
Copyright (c) 2010 Adaltas
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
SOFTWARE.

27
node_modules/csv-parse/README.md generated vendored Normal file
View File

@@ -0,0 +1,27 @@
# CSV Parser for Node.js
[![Build Status](https://api.travis-ci.org/adaltas/node-csv-parse.svg)](https://travis-ci.org/#!/adaltas/node-csv-parse) [![NPM](https://img.shields.io/npm/dm/csv-parse)](https://www.npmjs.com/package/csv-parse) [![NPM](https://img.shields.io/npm/v/csv-parse)](https://www.npmjs.com/package/csv-parse)
Part of the [CSV module](https://csv.js.org/), this project is a parser converting CSV text input into arrays or objects. It implements the Node.js [`stream.Transform` API](http://nodejs.org/api/stream.html#stream_class_stream_transform). It also provides a simple callback-based API for convenience. It is both extremely easy to use and powerful. It was first released in 2010 and is used against big data sets by a large community.
## Documentation
* [Project homepage](http://csv.js.org/parse/)
* [API](http://csv.js.org/parse/api/)
* [Options](http://csv.js.org/parse/options/)
* [Info properties](http://csv.js.org/parse/info/)
* [Common errors](http://csv.js.org/parse/errors/)
* [Examples](http://csv.js.org/project/examples/)
## Features
* Follow the Node.js streaming API
* Simplicity with the optional callback API
* Support delimiters, quotes, escape characters and comments
* Line breaks discovery
* Support big datasets
* Complete test coverage and samples for inspiration
* No external dependencies
* Work nicely with the [csv-generate](https://csv.js.org/generate/), [stream-transform](https://csv.js.org/transform/) and [csv-stringify](https://csv.js.org/stringify/) packages
* MIT License

65
node_modules/csv-parse/lib/ResizeableBuffer.js generated vendored Normal file
View File

@@ -0,0 +1,65 @@
class ResizeableBuffer{
constructor(size=100){
this.size = size
this.length = 0
this.buf = Buffer.alloc(size)
}
prepend(val){
if(Buffer.isBuffer(val)){
const length = this.length + val.length
if(length >= this.size){
this.resize()
if(length >= this.size){
throw Error('INVALID_BUFFER_STATE')
}
}
const buf = this.buf
this.buf = Buffer.alloc(this.size)
val.copy(this.buf, 0)
buf.copy(this.buf, val.length)
this.length += val.length
}else{
const length = this.length++
if(length === this.size){
this.resize()
}
const buf = this.clone()
this.buf[0] = val
buf.copy(this.buf,1, 0, length)
}
}
append(val){
const length = this.length++
if(length === this.size){
this.resize()
}
this.buf[length] = val
}
clone(){
return Buffer.from(this.buf.slice(0, this.length))
}
resize(){
const length = this.length
this.size = this.size * 2
const buf = Buffer.alloc(this.size)
this.buf.copy(buf,0, 0, length)
this.buf = buf
}
toString(encoding){
if(encoding){
return this.buf.slice(0, this.length).toString(encoding)
}else{
return Uint8Array.prototype.slice.call(this.buf.slice(0, this.length))
}
}
toJSON(){
return this.toString('utf8')
}
reset(){
this.length = 0
}
}
module.exports = ResizeableBuffer

8131
node_modules/csv-parse/lib/browser/index.js generated vendored Normal file

File diff suppressed because it is too large Load Diff

8168
node_modules/csv-parse/lib/browser/sync.js generated vendored Normal file

File diff suppressed because it is too large Load Diff

102
node_modules/csv-parse/lib/es5/ResizeableBuffer.js generated vendored Normal file
View File

@@ -0,0 +1,102 @@
"use strict";
function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError("Cannot call a class as a function"); } }
function _defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } }
function _createClass(Constructor, protoProps, staticProps) { if (protoProps) _defineProperties(Constructor.prototype, protoProps); if (staticProps) _defineProperties(Constructor, staticProps); return Constructor; }
var ResizeableBuffer = /*#__PURE__*/function () {
function ResizeableBuffer() {
var size = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : 100;
_classCallCheck(this, ResizeableBuffer);
this.size = size;
this.length = 0;
this.buf = Buffer.alloc(size);
}
_createClass(ResizeableBuffer, [{
key: "prepend",
value: function prepend(val) {
if (Buffer.isBuffer(val)) {
var length = this.length + val.length;
if (length >= this.size) {
this.resize();
if (length >= this.size) {
throw Error('INVALID_BUFFER_STATE');
}
}
var buf = this.buf;
this.buf = Buffer.alloc(this.size);
val.copy(this.buf, 0);
buf.copy(this.buf, val.length);
this.length += val.length;
} else {
var _length = this.length++;
if (_length === this.size) {
this.resize();
}
var _buf = this.clone();
this.buf[0] = val;
_buf.copy(this.buf, 1, 0, _length);
}
}
}, {
key: "append",
value: function append(val) {
var length = this.length++;
if (length === this.size) {
this.resize();
}
this.buf[length] = val;
}
}, {
key: "clone",
value: function clone() {
return Buffer.from(this.buf.slice(0, this.length));
}
}, {
key: "resize",
value: function resize() {
var length = this.length;
this.size = this.size * 2;
var buf = Buffer.alloc(this.size);
this.buf.copy(buf, 0, 0, length);
this.buf = buf;
}
}, {
key: "toString",
value: function toString(encoding) {
if (encoding) {
return this.buf.slice(0, this.length).toString(encoding);
} else {
return Uint8Array.prototype.slice.call(this.buf.slice(0, this.length));
}
}
}, {
key: "toJSON",
value: function toJSON() {
return this.toString('utf8');
}
}, {
key: "reset",
value: function reset() {
this.length = 0;
}
}]);
return ResizeableBuffer;
}();
module.exports = ResizeableBuffer;

268
node_modules/csv-parse/lib/es5/index.d.ts generated vendored Normal file
View File

@@ -0,0 +1,268 @@
// Original definitions in https://github.com/DefinitelyTyped/DefinitelyTyped by: David Muller <https://github.com/davidm77>
/// <reference types="node" />
import * as stream from "stream";
export = parse;
declare function parse(input: Buffer | string, options?: parse.Options, callback?: parse.Callback): parse.Parser;
declare function parse(input: Buffer | string, callback?: parse.Callback): parse.Parser;
declare function parse(options?: parse.Options, callback?: parse.Callback): parse.Parser;
declare function parse(callback?: parse.Callback): parse.Parser;
declare namespace parse {
type Callback = (err: Error | undefined, records: any | undefined, info: Info) => void;
interface Parser extends stream.Transform {}
class Parser {
constructor(options: Options);
__push(line: any): any;
__write(chars: any, end: any, callback: any): any;
readonly options: Options
readonly info: Info;
}
interface CastingContext {
readonly column: number | string;
readonly empty_lines: number;
readonly error: CsvError;
readonly header: boolean;
readonly index: number;
readonly quoting: boolean;
readonly lines: number;
readonly records: number;
readonly invalid_field_length: number;
}
type CastingFunction = (value: string, context: CastingContext) => any;
type CastingDateFunction = (value: string, context: CastingContext) => Date;
type ColumnOption = string | undefined | null | false | { name: string };
interface Options {
/**
* If true, the parser will attempt to convert read data types to native types.
* @deprecated Use {@link cast}
*/
auto_parse?: boolean | CastingFunction;
autoParse?: boolean | CastingFunction;
/**
* If true, the parser will attempt to convert read data types to dates. It requires the "auto_parse" option.
* @deprecated Use {@link cast_date}
*/
auto_parse_date?: boolean | CastingDateFunction;
autoParseDate?: boolean | CastingDateFunction;
/**
* If true, detect and exclude the byte order mark (BOM) from the CSV input if present.
*/
bom?: boolean;
/**
* If true, the parser will attempt to convert input string to native types.
* If a function, receive the value as first argument, a context as second argument and return a new value. More information about the context properties is available below.
*/
cast?: boolean | CastingFunction;
/**
* If true, the parser will attempt to convert input string to dates.
* If a function, receive the value as argument and return a new value. It requires the "auto_parse" option. Be careful, it relies on Date.parse.
*/
cast_date?: boolean | CastingDateFunction;
castDate?: boolean | CastingDateFunction;
/**
* List of fields as an array,
* a user defined callback accepting the first line and returning the column names or true if autodiscovered in the first CSV line,
* default to null,
* affect the result data set in the sense that records will be objects instead of arrays.
*/
columns?: ColumnOption[] | boolean | ((record: any) => ColumnOption[]);
/**
* Convert values into an array of values when columns are activated and
* when multiple columns of the same name are found.
*/
columns_duplicates_to_array?: boolean;
columnsDuplicatesToArray?: boolean;
/**
* Treat all the characters after this one as a comment, default to '' (disabled).
*/
comment?: string;
/**
* Set the field delimiter. One character only, defaults to comma.
*/
delimiter?: string | string[] | Buffer;
/**
* Set the source and destination encoding, a value of `null` returns buffer instead of strings.
*/
encoding?: string | null;
/**
* Set the escape character, one character only, defaults to double quotes.
*/
escape?: string | null | false | Buffer;
/**
* Start handling records from the requested number of records.
*/
from?: number;
/**
* Start handling records from the requested line number.
*/
from_line?: number;
fromLine?: number;
/**
* Don't interpret delimiters as such in the last field according to the number of fields calculated from the number of columns, the option require the presence of the `column` option when `true`.
*/
ignore_last_delimiters?: boolean | number;
/**
* Generate two properties `info` and `record` where `info` is a snapshot of the info object at the time the record was created and `record` is the parsed array or object.
*/
info?: boolean;
/**
* If true, ignore whitespace immediately following the delimiter (i.e. left-trim all fields), defaults to false.
* Does not remove whitespace in a quoted field.
*/
ltrim?: boolean;
/**
* Maximum numer of characters to be contained in the field and line buffers before an exception is raised,
* used to guard against a wrong delimiter or record_delimiter,
* default to 128000 characters.
*/
max_record_size?: number;
maxRecordSize?: number;
/**
* Name of header-record title to name objects by.
*/
objname?: string;
/**
* Alter and filter records by executing a user defined function.
*/
on_record?: (record: any, context: CastingContext) => any;
onRecord?: (record: any, context: CastingContext) => any;
/**
* Optional character surrounding a field, one character only, defaults to double quotes.
*/
quote?: string | boolean | Buffer | null;
/**
* Generate two properties raw and row where raw is the original CSV row content and row is the parsed array or object.
*/
raw?: boolean;
/**
* Preserve quotes inside unquoted field.
*/
relax?: boolean;
/**
* Discard inconsistent columns count, default to false.
*/
relax_column_count?: boolean;
relaxColumnCount?: boolean;
/**
* Discard inconsistent columns count when the record contains less fields than expected, default to false.
*/
relax_column_count_less?: boolean;
relaxColumnCountLess?: boolean;
/**
* Discard inconsistent columns count when the record contains more fields than expected, default to false.
*/
relax_column_count_more?: boolean;
relaxColumnCountMore?: boolean;
/**
* One or multiple characters used to delimit record rows; defaults to auto discovery if not provided.
* Supported auto discovery method are Linux ("\n"), Apple ("\r") and Windows ("\r\n") row delimiters.
*/
record_delimiter?: string | string[] | Buffer | Buffer[];
recordDelimiter?: string | string[] | Buffer | Buffer[];
/**
* If true, ignore whitespace immediately preceding the delimiter (i.e. right-trim all fields), defaults to false.
* Does not remove whitespace in a quoted field.
*/
rtrim?: boolean;
/**
* Dont generate empty values for empty lines.
* Defaults to false
*/
skip_empty_lines?: boolean;
skipEmptyLines?: boolean;
/**
* Skip a line with error found inside and directly go process the next line.
*/
skip_lines_with_error?: boolean;
skipLinesWithError?: boolean;
/**
* Don't generate records for lines containing empty column values (column matching /\s*\/), defaults to false.
*/
skip_lines_with_empty_values?: boolean;
skipLinesWithEmptyValues?: boolean;
/**
* Stop handling records after the requested number of records.
*/
to?: number;
/**
* Stop handling records after the requested line number.
*/
to_line?: number;
toLine?: number;
/**
* If true, ignore whitespace immediately around the delimiter, defaults to false.
* Does not remove whitespace in a quoted field.
*/
trim?: boolean;
}
interface Info {
/**
* Count the number of lines being fully commented.
*/
readonly comment_lines: number;
/**
* Count the number of processed empty lines.
*/
readonly empty_lines: number;
/**
* The number of lines encountered in the source dataset, start at 1 for the first line.
*/
readonly lines: number;
/**
* Count the number of processed records.
*/
readonly records: number;
/**
* Count of the number of processed bytes.
*/
readonly bytes: number;
/**
* Number of non uniform records when `relax_column_count` is true.
*/
readonly invalid_field_length: number;
}
class CsvError extends Error {
readonly code: CsvErrorCode;
[key: string]: any;
constructor(code: CsvErrorCode, message: string | string[], options?: Options, ...contexts: any[]);
}
type CsvErrorCode =
'CSV_INVALID_OPTION_BOM'
| 'CSV_INVALID_OPTION_CAST'
| 'CSV_INVALID_OPTION_CAST_DATE'
| 'CSV_INVALID_OPTION_COLUMNS'
| 'CSV_INVALID_OPTION_COLUMNS_DUPLICATES_TO_ARRAY'
| 'CSV_INVALID_OPTION_COMMENT'
| 'CSV_INVALID_OPTION_DELIMITER'
| 'CSV_INVALID_OPTION_ON_RECORD'
| 'CSV_INVALID_CLOSING_QUOTE'
| 'INVALID_OPENING_QUOTE'
| 'CSV_INVALID_COLUMN_MAPPING'
| 'CSV_INVALID_ARGUMENT'
| 'CSV_INVALID_COLUMN_DEFINITION'
| 'CSV_MAX_RECORD_SIZE'
| 'CSV_NON_TRIMABLE_CHAR_AFTER_CLOSING_QUOTE'
| 'CSV_QUOTE_NOT_CLOSED'
| 'CSV_INCONSISTENT_RECORD_LENGTH'
| 'CSV_RECORD_DONT_MATCH_COLUMNS_LENGTH'
| 'CSV_OPTION_COLUMNS_MISSING_NAME'
}

1530
node_modules/csv-parse/lib/es5/index.js generated vendored Normal file

File diff suppressed because it is too large Load Diff

6
node_modules/csv-parse/lib/es5/sync.d.ts generated vendored Normal file
View File

@@ -0,0 +1,6 @@
import * as csvParse from './index';
export = parse;
declare function parse(input: Buffer | string, options?: csvParse.Options): any;
declare namespace parse {}

33
node_modules/csv-parse/lib/es5/sync.js generated vendored Normal file
View File

@@ -0,0 +1,33 @@
"use strict";
var parse = require('.');
module.exports = function (data) {
var options = arguments.length > 1 && arguments[1] !== undefined ? arguments[1] : {};
if (typeof data === 'string') {
data = Buffer.from(data);
}
var records = options && options.objname ? {} : [];
var parser = new parse.Parser(options);
parser.push = function (record) {
if (record === null) {
return;
}
if (options.objname === undefined) records.push(record);else {
records[record[0]] = record[1];
}
};
var err1 = parser.__parse(data, false);
if (err1 !== undefined) throw err1;
var err2 = parser.__parse(undefined, true);
if (err2 !== undefined) throw err2;
return records;
};

268
node_modules/csv-parse/lib/index.d.ts generated vendored Normal file
View File

@@ -0,0 +1,268 @@
// Original definitions in https://github.com/DefinitelyTyped/DefinitelyTyped by: David Muller <https://github.com/davidm77>
/// <reference types="node" />
import * as stream from "stream";
export = parse;
declare function parse(input: Buffer | string, options?: parse.Options, callback?: parse.Callback): parse.Parser;
declare function parse(input: Buffer | string, callback?: parse.Callback): parse.Parser;
declare function parse(options?: parse.Options, callback?: parse.Callback): parse.Parser;
declare function parse(callback?: parse.Callback): parse.Parser;
declare namespace parse {
type Callback = (err: Error | undefined, records: any | undefined, info: Info) => void;
interface Parser extends stream.Transform {}
class Parser {
constructor(options: Options);
__push(line: any): any;
__write(chars: any, end: any, callback: any): any;
readonly options: Options
readonly info: Info;
}
interface CastingContext {
readonly column: number | string;
readonly empty_lines: number;
readonly error: CsvError;
readonly header: boolean;
readonly index: number;
readonly quoting: boolean;
readonly lines: number;
readonly records: number;
readonly invalid_field_length: number;
}
type CastingFunction = (value: string, context: CastingContext) => any;
type CastingDateFunction = (value: string, context: CastingContext) => Date;
type ColumnOption = string | undefined | null | false | { name: string };
interface Options {
/**
* If true, the parser will attempt to convert read data types to native types.
* @deprecated Use {@link cast}
*/
auto_parse?: boolean | CastingFunction;
autoParse?: boolean | CastingFunction;
/**
* If true, the parser will attempt to convert read data types to dates. It requires the "auto_parse" option.
* @deprecated Use {@link cast_date}
*/
auto_parse_date?: boolean | CastingDateFunction;
autoParseDate?: boolean | CastingDateFunction;
/**
* If true, detect and exclude the byte order mark (BOM) from the CSV input if present.
*/
bom?: boolean;
/**
* If true, the parser will attempt to convert input string to native types.
* If a function, receive the value as first argument, a context as second argument and return a new value. More information about the context properties is available below.
*/
cast?: boolean | CastingFunction;
/**
* If true, the parser will attempt to convert input string to dates.
* If a function, receive the value as argument and return a new value. It requires the "auto_parse" option. Be careful, it relies on Date.parse.
*/
cast_date?: boolean | CastingDateFunction;
castDate?: boolean | CastingDateFunction;
/**
* List of fields as an array,
* a user defined callback accepting the first line and returning the column names or true if autodiscovered in the first CSV line,
* default to null,
* affect the result data set in the sense that records will be objects instead of arrays.
*/
columns?: ColumnOption[] | boolean | ((record: any) => ColumnOption[]);
/**
* Convert values into an array of values when columns are activated and
* when multiple columns of the same name are found.
*/
columns_duplicates_to_array?: boolean;
columnsDuplicatesToArray?: boolean;
/**
* Treat all the characters after this one as a comment, default to '' (disabled).
*/
comment?: string;
/**
* Set the field delimiter. One character only, defaults to comma.
*/
delimiter?: string | string[] | Buffer;
/**
* Set the source and destination encoding, a value of `null` returns buffer instead of strings.
*/
encoding?: string | null;
/**
* Set the escape character, one character only, defaults to double quotes.
*/
escape?: string | null | false | Buffer;
/**
* Start handling records from the requested number of records.
*/
from?: number;
/**
* Start handling records from the requested line number.
*/
from_line?: number;
fromLine?: number;
/**
* Don't interpret delimiters as such in the last field according to the number of fields calculated from the number of columns, the option require the presence of the `column` option when `true`.
*/
ignore_last_delimiters?: boolean | number;
/**
* Generate two properties `info` and `record` where `info` is a snapshot of the info object at the time the record was created and `record` is the parsed array or object.
*/
info?: boolean;
/**
* If true, ignore whitespace immediately following the delimiter (i.e. left-trim all fields), defaults to false.
* Does not remove whitespace in a quoted field.
*/
ltrim?: boolean;
/**
* Maximum numer of characters to be contained in the field and line buffers before an exception is raised,
* used to guard against a wrong delimiter or record_delimiter,
* default to 128000 characters.
*/
max_record_size?: number;
maxRecordSize?: number;
/**
* Name of header-record title to name objects by.
*/
objname?: string;
/**
* Alter and filter records by executing a user defined function.
*/
on_record?: (record: any, context: CastingContext) => any;
onRecord?: (record: any, context: CastingContext) => any;
/**
* Optional character surrounding a field, one character only, defaults to double quotes.
*/
quote?: string | boolean | Buffer | null;
/**
* Generate two properties raw and row where raw is the original CSV row content and row is the parsed array or object.
*/
raw?: boolean;
/**
* Preserve quotes inside unquoted field.
*/
relax?: boolean;
/**
* Discard inconsistent columns count, default to false.
*/
relax_column_count?: boolean;
relaxColumnCount?: boolean;
/**
* Discard inconsistent columns count when the record contains less fields than expected, default to false.
*/
relax_column_count_less?: boolean;
relaxColumnCountLess?: boolean;
/**
* Discard inconsistent columns count when the record contains more fields than expected, default to false.
*/
relax_column_count_more?: boolean;
relaxColumnCountMore?: boolean;
/**
* One or multiple characters used to delimit record rows; defaults to auto discovery if not provided.
* Supported auto discovery method are Linux ("\n"), Apple ("\r") and Windows ("\r\n") row delimiters.
*/
record_delimiter?: string | string[] | Buffer | Buffer[];
recordDelimiter?: string | string[] | Buffer | Buffer[];
/**
* If true, ignore whitespace immediately preceding the delimiter (i.e. right-trim all fields), defaults to false.
* Does not remove whitespace in a quoted field.
*/
rtrim?: boolean;
/**
* Dont generate empty values for empty lines.
* Defaults to false
*/
skip_empty_lines?: boolean;
skipEmptyLines?: boolean;
/**
* Skip a line with error found inside and directly go process the next line.
*/
skip_lines_with_error?: boolean;
skipLinesWithError?: boolean;
/**
* Don't generate records for lines containing empty column values (column matching /\s*\/), defaults to false.
*/
skip_lines_with_empty_values?: boolean;
skipLinesWithEmptyValues?: boolean;
/**
* Stop handling records after the requested number of records.
*/
to?: number;
/**
* Stop handling records after the requested line number.
*/
to_line?: number;
toLine?: number;
/**
* If true, ignore whitespace immediately around the delimiter, defaults to false.
* Does not remove whitespace in a quoted field.
*/
trim?: boolean;
}
interface Info {
/**
* Count the number of lines being fully commented.
*/
readonly comment_lines: number;
/**
* Count the number of processed empty lines.
*/
readonly empty_lines: number;
/**
* The number of lines encountered in the source dataset, start at 1 for the first line.
*/
readonly lines: number;
/**
* Count the number of processed records.
*/
readonly records: number;
/**
* Count of the number of processed bytes.
*/
readonly bytes: number;
/**
* Number of non uniform records when `relax_column_count` is true.
*/
readonly invalid_field_length: number;
}
class CsvError extends Error {
readonly code: CsvErrorCode;
[key: string]: any;
constructor(code: CsvErrorCode, message: string | string[], options?: Options, ...contexts: any[]);
}
type CsvErrorCode =
'CSV_INVALID_OPTION_BOM'
| 'CSV_INVALID_OPTION_CAST'
| 'CSV_INVALID_OPTION_CAST_DATE'
| 'CSV_INVALID_OPTION_COLUMNS'
| 'CSV_INVALID_OPTION_COLUMNS_DUPLICATES_TO_ARRAY'
| 'CSV_INVALID_OPTION_COMMENT'
| 'CSV_INVALID_OPTION_DELIMITER'
| 'CSV_INVALID_OPTION_ON_RECORD'
| 'CSV_INVALID_CLOSING_QUOTE'
| 'INVALID_OPENING_QUOTE'
| 'CSV_INVALID_COLUMN_MAPPING'
| 'CSV_INVALID_ARGUMENT'
| 'CSV_INVALID_COLUMN_DEFINITION'
| 'CSV_MAX_RECORD_SIZE'
| 'CSV_NON_TRIMABLE_CHAR_AFTER_CLOSING_QUOTE'
| 'CSV_QUOTE_NOT_CLOSED'
| 'CSV_INCONSISTENT_RECORD_LENGTH'
| 'CSV_RECORD_DONT_MATCH_COLUMNS_LENGTH'
| 'CSV_OPTION_COLUMNS_MISSING_NAME'
}

1268
node_modules/csv-parse/lib/index.js generated vendored Normal file

File diff suppressed because it is too large Load Diff

6
node_modules/csv-parse/lib/sync.d.ts generated vendored Normal file
View File

@@ -0,0 +1,6 @@
import * as csvParse from './index';
export = parse;
declare function parse(input: Buffer | string, options?: csvParse.Options): any;
declare namespace parse {}

25
node_modules/csv-parse/lib/sync.js generated vendored Normal file
View File

@@ -0,0 +1,25 @@
const parse = require('.')
module.exports = function(data, options={}){
if(typeof data === 'string'){
data = Buffer.from(data)
}
const records = options && options.objname ? {} : []
const parser = new parse.Parser(options)
parser.push = function(record){
if(record === null){
return
}
if(options.objname === undefined)
records.push(record)
else{
records[record[0]] = record[1]
}
}
const err1 = parser.__parse(data, false)
if(err1 !== undefined) throw err1
const err2 = parser.__parse(undefined, true)
if(err2 !== undefined) throw err2
return records
}

99
node_modules/csv-parse/package.json generated vendored Normal file
View File

@@ -0,0 +1,99 @@
{
"version": "4.16.3",
"name": "csv-parse",
"description": "CSV parsing implementing the Node.js `stream.Transform` API",
"keywords": [
"csv",
"parse",
"parser",
"convert",
"tsv",
"stream"
],
"coffeelintConfig": {
"indentation": {
"level": "error",
"value": 2
},
"line_endings": {
"level": "error",
"value": "unix"
},
"max_line_length": {
"level": "ignore"
}
},
"author": "David Worms <david@adaltas.com> (https://www.adaltas.com)",
"contributors": [
"David Worms <david@adaltas.com> (https://www.adaltas.com)",
"Will White (https://github.com/willwhite)",
"Justin Latimer (https://github.com/justinlatimer)",
"jonseymour (https://github.com/jonseymour)",
"pascalopitz (https://github.com/pascalopitz)",
"Josh Pschorr (https://github.com/jpschorr)",
"Elad Ben-Israel (https://github.com/eladb)",
"Philippe Plantier (https://github.com/phipla)",
"Tim Oxley (https://github.com/timoxley)",
"Damon Oehlman (https://github.com/DamonOehlman)",
"Alexandru Topliceanu (https://github.com/topliceanu)",
"Visup (https://github.com/visup)",
"Edmund von der Burg (https://github.com/evdb)",
"Douglas Christopher Wilson (https://github.com/dougwilson)",
"Joe Eaves (https://github.com/Joeasaurus)",
"Mark Stosberg (https://github.com/markstos)"
],
"license": "MIT",
"repository": {
"type": "git",
"url": "https://github.com/wdavidw/node-csv-parse"
},
"homepage": "https://csv.js.org/parse/",
"devDependencies": {
"@babel/cli": "^7.14.8",
"@babel/core": "^7.15.0",
"@babel/preset-env": "^7.15.0",
"@types/mocha": "^9.0.0",
"@types/node": "^16.7.8",
"babelify": "^10.0.0",
"browserify": "^17.0.0",
"coffeescript": "^2.5.1",
"csv-generate": "^3.4.3",
"csv-spectrum": "^1.0.0",
"each": "^1.2.2",
"eslint": "^7.32.0",
"express": "^4.17.1",
"mocha": "^9.1.1",
"pad": "^3.2.0",
"should": "^13.2.3",
"stream-transform": "^2.1.3",
"ts-node": "^10.2.1",
"typescript": "^4.4.2"
},
"files": [
"/lib"
],
"main": "./lib",
"mocha": {
"throw-deprecation": true,
"require": [
"should",
"coffeescript/register",
"ts-node/register"
],
"inline-diffs": true,
"timeout": 40000,
"reporter": "spec",
"recursive": true
},
"scripts": {
"build:babel": "cd lib && babel *.js -d es5 && cd ..",
"build:browserify": "browserify lib/index.js --transform babelify --standalone parse > lib/browser/index.js && browserify lib/sync.js --transform babelify --standalone parse > lib/browser/sync.js",
"build": "npm run build:babel && npm run build:browserify",
"preversion": "cp lib/*.ts lib/es5 && git add lib/es5/*.ts",
"pretest": "npm run build",
"lint": "eslint lib/*.js",
"test": "npm run lint && TS_NODE_COMPILER_OPTIONS='{\"strictNullChecks\":true}' mocha test/**/*.{coffee,ts}"
},
"types": "./lib/index.d.ts",
"gitHead": "0fd5209b6862655c384cda052abf38019b959e70"
}