/
dataset.ts
678 lines (584 loc) · 24.7 KB
/
dataset.ts
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
import { MAX_PAYLOAD_SIZE_BYTES } from '@apify/consts';
import type { DatasetClient, DatasetInfo, Dictionary, StorageClient } from '@crawlee/types';
import { stringify } from 'csv-stringify/sync';
import ow from 'ow';
import { checkStorageAccess } from './access_checking';
import { KeyValueStore } from './key_value_store';
import type { StorageManagerOptions } from './storage_manager';
import { StorageManager } from './storage_manager';
import { purgeDefaultStorages } from './utils';
import { Configuration } from '../configuration';
import { log, type Log } from '../log';
import type { Awaitable } from '../typedefs';
/** @internal */
export const DATASET_ITERATORS_DEFAULT_LIMIT = 10000;
const SAFETY_BUFFER_PERCENT = 0.01 / 100; // 0.01%
/**
* Accepts a JSON serializable object as an input, validates its serializability,
* and validates its serialized size against limitBytes. Optionally accepts its index
* in an array to provide better error messages. Returns serialized object.
* @ignore
*/
export function checkAndSerialize<T>(item: T, limitBytes: number, index?: number): string {
const s = typeof index === 'number' ? ` at index ${index} ` : ' ';
const isItemObject = item && typeof item === 'object' && !Array.isArray(item);
if (!isItemObject) {
throw new Error(`Data item${s}is not an object. You can push only objects into a dataset.`);
}
let payload;
try {
payload = JSON.stringify(item);
} catch (e) {
const err = e as Error;
throw new Error(`Data item${s}is not serializable to JSON.\nCause: ${err.message}`);
}
const bytes = Buffer.byteLength(payload);
if (bytes > limitBytes) {
throw new Error(`Data item${s}is too large (size: ${bytes} bytes, limit: ${limitBytes} bytes)`);
}
return payload;
}
/**
* Takes an array of JSONs (payloads) as input and produces an array of JSON strings
* where each string is a JSON array of payloads with a maximum size of limitBytes per one
* JSON array. Fits as many payloads as possible into a single JSON array and then moves
* on to the next, preserving item order.
*
* The function assumes that none of the items is larger than limitBytes and does not validate.
* @ignore
*/
export function chunkBySize(items: string[], limitBytes: number): string[] {
if (!items.length) return [];
if (items.length === 1) return items;
// Split payloads into buckets of valid size.
let lastChunkBytes = 2; // Add 2 bytes for [] wrapper.
const chunks: (string | string[])[] = [];
for (const payload of items) {
const bytes = Buffer.byteLength(payload);
if (bytes <= limitBytes && (bytes + 2) > limitBytes) {
// Handle cases where wrapping with [] would fail, but solo object is fine.
chunks.push(payload);
lastChunkBytes = bytes;
} else if (lastChunkBytes + bytes <= limitBytes) {
// ensure array
if (!Array.isArray(chunks[chunks.length - 1])) {
chunks.push([]);
}
(chunks[chunks.length - 1] as string[]).push(payload);
lastChunkBytes += bytes + 1; // Add 1 byte for ',' separator.
} else {
chunks.push([payload]);
lastChunkBytes = bytes + 2; // Add 2 bytes for [] wrapper.
}
}
// Stringify array chunks.
return chunks.map((chunk) => (typeof chunk === 'string' ? chunk : `[${chunk.join(',')}]`));
}
export interface DatasetDataOptions {
/**
* Number of array elements that should be skipped at the start.
* @default 0
*/
offset?: number;
/**
* Maximum number of array elements to return.
* @default 250000
*/
limit?: number;
/**
* If `true` then the objects are sorted by `createdAt` in descending order.
* Otherwise they are sorted in ascending order.
* @default false
*/
desc?: boolean;
/**
* An array of field names that will be included in the result. If omitted, all fields are included in the results.
*/
fields?: string[];
/**
* Specifies a name of the field in the result objects that will be used to unwind the resulting objects.
* By default, the results are returned as they are.
*/
unwind?: string;
/**
* If `true` then the function returns only non-empty items and skips hidden fields (i.e. fields starting with `#` character).
* Note that the `clean` parameter is a shortcut for `skipHidden: true` and `skipEmpty: true` options.
* @default false
*/
clean?: boolean;
/**
* If `true` then the function doesn't return hidden fields (fields starting with "#" character).
* @default false
*/
skipHidden?: boolean;
/**
* If `true` then the function doesn't return empty items.
* Note that in this case the returned number of items might be lower than limit parameter and pagination must be done using the `limit` value.
* @default false
*/
skipEmpty?: boolean;
}
export interface DatasetExportOptions extends Omit<DatasetDataOptions, 'offset' | 'limit'> {}
export interface DatasetIteratorOptions extends Omit<DatasetDataOptions, 'offset' | 'limit' | 'clean' | 'skipHidden' | 'skipEmpty'> {
/** @internal */
offset?: number;
/**
* @default 10000
* @internal
*/
limit?: number;
/** @internal */
clean?: boolean;
/** @internal */
skipHidden?: boolean;
/** @internal */
skipEmpty?: boolean;
/** @internal */
format?: string;
}
export interface DatasetExportToOptions extends DatasetExportOptions {
fromDataset?: string;
toKVS?: string;
}
/**
* The `Dataset` class represents a store for structured data where each object stored has the same attributes,
* such as online store products or real estate offers. You can imagine it as a table,
* where each object is a row and its attributes are columns.
* Dataset is an append-only storage - you can only add new records to it but you cannot modify or remove existing records.
* Typically it is used to store crawling results.
*
* Do not instantiate this class directly, use the
* {@apilink Dataset.open} function instead.
*
* `Dataset` stores its data either on local disk or in the Apify cloud,
* depending on whether the `APIFY_LOCAL_STORAGE_DIR` or `APIFY_TOKEN` environment variables are set.
*
* If the `APIFY_LOCAL_STORAGE_DIR` environment variable is set, the data is stored in
* the local directory in the following files:
* ```
* {APIFY_LOCAL_STORAGE_DIR}/datasets/{DATASET_ID}/{INDEX}.json
* ```
* Note that `{DATASET_ID}` is the name or ID of the dataset. The default dataset has ID: `default`,
* unless you override it by setting the `APIFY_DEFAULT_DATASET_ID` environment variable.
* Each dataset item is stored as a separate JSON file, where `{INDEX}` is a zero-based index of the item in the dataset.
*
* If the `APIFY_TOKEN` environment variable is set but `APIFY_LOCAL_STORAGE_DIR` not, the data is stored in the
* [Apify Dataset](https://docs.apify.com/storage/dataset)
* cloud storage. Note that you can force usage of the cloud storage also by passing the `forceCloud`
* option to {@apilink Dataset.open} function,
* even if the `APIFY_LOCAL_STORAGE_DIR` variable is set.
*
* **Example usage:**
*
* ```javascript
* // Write a single row to the default dataset
* await Dataset.pushData({ col1: 123, col2: 'val2' });
*
* // Open a named dataset
* const dataset = await Dataset.open('some-name');
*
* // Write a single row
* await dataset.pushData({ foo: 'bar' });
*
* // Write multiple rows
* await dataset.pushData([
* { foo: 'bar2', col2: 'val2' },
* { col3: 123 },
* ]);
*
* // Export the entirety of the dataset to one file in the key-value store
* await dataset.exportToCSV('MY-DATA');
* ```
* @category Result Stores
*/
export class Dataset<Data extends Dictionary = Dictionary> {
id: string;
name?: string;
client: DatasetClient<Data>;
log: Log = log.child({ prefix: 'Dataset' });
/**
* @internal
*/
constructor(options: DatasetOptions, readonly config = Configuration.getGlobalConfig()) {
this.id = options.id;
this.name = options.name;
this.client = options.client.dataset(this.id) as DatasetClient<Data>;
}
/**
* Stores an object or an array of objects to the dataset.
* The function returns a promise that resolves when the operation finishes.
* It has no result, but throws on invalid args or other errors.
*
* **IMPORTANT**: Make sure to use the `await` keyword when calling `pushData()`,
* otherwise the crawler process might finish before the data is stored!
*
* The size of the data is limited by the receiving API and therefore `pushData()` will only
* allow objects whose JSON representation is smaller than 9MB. When an array is passed,
* none of the included objects
* may be larger than 9MB, but the array itself may be of any size.
*
* The function internally
* chunks the array into separate items and pushes them sequentially.
* The chunking process is stable (keeps order of data), but it does not provide a transaction
* safety mechanism. Therefore, in the event of an uploading error (after several automatic retries),
* the function's Promise will reject and the dataset will be left in a state where some of
* the items have already been saved to the dataset while other items from the source array were not.
* To overcome this limitation, the developer may, for example, read the last item saved in the dataset
* and re-attempt the save of the data from this item onwards to prevent duplicates.
* @param data Object or array of objects containing data to be stored in the default dataset.
* The objects must be serializable to JSON and the JSON representation of each object must be smaller than 9MB.
*/
async pushData(data: Data | Data[]): Promise<void> {
checkStorageAccess();
ow(data, 'data', ow.object);
const dispatch = async (payload: string) => this.client.pushItems(payload);
const limit = MAX_PAYLOAD_SIZE_BYTES - Math.ceil(MAX_PAYLOAD_SIZE_BYTES * SAFETY_BUFFER_PERCENT);
// Handle singular Objects
if (!Array.isArray(data)) {
const payload = checkAndSerialize(data, limit);
return dispatch(payload);
}
// Handle Arrays
const payloads = data.map((item, index) => checkAndSerialize(item, limit, index));
const chunks = chunkBySize(payloads, limit);
// Invoke client in series to preserve order of data
for (const chunk of chunks) {
await dispatch(chunk);
}
}
/**
* Returns {@apilink DatasetContent} object holding the items in the dataset based on the provided parameters.
*/
async getData(options: DatasetDataOptions = {}): Promise<DatasetContent<Data>> {
checkStorageAccess();
try {
return await this.client.listItems(options);
} catch (e) {
const error = e as Error;
if (error.message.includes('Cannot create a string longer than')) {
throw new Error('dataset.getData(): The response is too large for parsing. You can fix this by lowering the "limit" option.');
}
throw e;
}
}
/**
* Returns all the data from the dataset. This will iterate through the whole dataset
* via the `listItems()` client method, which gives you only paginated results.
*/
async export(options: DatasetExportOptions = {}): Promise<Data[]> {
checkStorageAccess();
const items: Data[] = [];
const fetchNextChunk = async (offset = 0): Promise<void> => {
const limit = 1000;
const value = await this.client.listItems({ offset, limit, ...options });
if (value.count === 0) {
return;
}
items.push(...value.items);
if (value.total > offset + value.count) {
return fetchNextChunk(offset + value.count);
}
};
await fetchNextChunk();
return items;
}
/**
* Save the entirety of the dataset's contents into one file within a key-value store.
*
* @param key The name of the value to save the data in.
* @param [options] An optional options object where you can provide the dataset and target KVS name.
* @param [contentType] Only JSON and CSV are supported currently, defaults to JSON.
*/
async exportTo(key: string, options?: DatasetExportToOptions, contentType?: string): Promise<Data[]> {
const kvStore = await KeyValueStore.open(options?.toKVS ?? null, { config: this.config });
const items = await this.export(options);
if (contentType === 'text/csv') {
const value = stringify([
Object.keys(items[0]),
...items.map((item) => Object.values(item)),
]);
await kvStore.setValue(key, value, { contentType });
return items;
}
if (contentType === 'application/json') {
await kvStore.setValue(key, items);
return items;
}
throw new Error(`Unsupported content type: ${contentType}`);
return items;
}
/**
* Save entire default dataset's contents into one JSON file within a key-value store.
*
* @param key The name of the value to save the data in.
* @param [options] An optional options object where you can provide the target KVS name.
*/
async exportToJSON(key: string, options?: Omit<DatasetExportToOptions, 'fromDataset'>) {
await this.exportTo(key, options, 'application/json');
}
/**
* Save entire default dataset's contents into one CSV file within a key-value store.
*
* @param key The name of the value to save the data in.
* @param [options] An optional options object where you can provide the target KVS name.
*/
async exportToCSV(key: string, options?: Omit<DatasetExportToOptions, 'fromDataset'>) {
await this.exportTo(key, options, 'text/csv');
}
/**
* Save entire default dataset's contents into one JSON file within a key-value store.
*
* @param key The name of the value to save the data in.
* @param [options] An optional options object where you can provide the dataset and target KVS name.
*/
static async exportToJSON(key: string, options?: DatasetExportToOptions) {
checkStorageAccess();
const dataset = await this.open(options?.fromDataset);
await dataset.exportToJSON(key, options);
}
/**
* Save entire default dataset's contents into one CSV file within a key-value store.
*
* @param key The name of the value to save the data in.
* @param [options] An optional options object where you can provide the dataset and target KVS name.
*/
static async exportToCSV(key: string, options?: DatasetExportToOptions) {
checkStorageAccess();
const dataset = await this.open(options?.fromDataset);
await dataset.exportToCSV(key, options);
}
/**
* Returns an object containing general information about the dataset.
*
* The function returns the same object as the Apify API Client's
* [getDataset](https://docs.apify.com/api/apify-client-js/latest#ApifyClient-datasets-getDataset)
* function, which in turn calls the
* [Get dataset](https://apify.com/docs/api/v2#/reference/datasets/dataset/get-dataset)
* API endpoint.
*
* **Example:**
* ```
* {
* id: "WkzbQMuFYuamGv3YF",
* name: "my-dataset",
* userId: "wRsJZtadYvn4mBZmm",
* createdAt: new Date("2015-12-12T07:34:14.202Z"),
* modifiedAt: new Date("2015-12-13T08:36:13.202Z"),
* accessedAt: new Date("2015-12-14T08:36:13.202Z"),
* itemCount: 14,
* }
* ```
*/
async getInfo(): Promise<DatasetInfo | undefined> {
checkStorageAccess();
return this.client.get();
}
/**
* Iterates over dataset items, yielding each in turn to an `iteratee` function.
* Each invocation of `iteratee` is called with two arguments: `(item, index)`.
*
* If the `iteratee` function returns a Promise then it is awaited before the next call.
* If it throws an error, the iteration is aborted and the `forEach` function throws the error.
*
* **Example usage**
* ```javascript
* const dataset = await Dataset.open('my-results');
* await dataset.forEach(async (item, index) => {
* console.log(`Item at ${index}: ${JSON.stringify(item)}`);
* });
* ```
*
* @param iteratee A function that is called for every item in the dataset.
* @param [options] All `forEach()` parameters.
* @param [index] Specifies the initial index number passed to the `iteratee` function.
* @default 0
*/
async forEach(iteratee: DatasetConsumer<Data>, options: DatasetIteratorOptions = {}, index = 0): Promise<void> {
checkStorageAccess();
if (!options.offset) options.offset = 0;
if (options.format && options.format !== 'json') throw new Error('Dataset.forEach/map/reduce() support only a "json" format.');
if (!options.limit) options.limit = DATASET_ITERATORS_DEFAULT_LIMIT;
const { items, total, limit, offset } = await this.getData(options);
for (const item of items) {
await iteratee(item, index++);
}
const newOffset = offset + limit;
if (newOffset >= total) return;
const newOpts = { ...options, offset: newOffset };
return this.forEach(iteratee, newOpts, index);
}
/**
* Produces a new array of values by mapping each value in list through a transformation function `iteratee()`.
* Each invocation of `iteratee()` is called with two arguments: `(element, index)`.
*
* If `iteratee` returns a `Promise` then it's awaited before a next call.
*
* @param iteratee
* @param [options] All `map()` parameters.
*/
async map<R>(iteratee: DatasetMapper<Data, R>, options: DatasetIteratorOptions = {}): Promise<R[]> {
checkStorageAccess();
const result: R[] = [];
await this.forEach(async (item, index) => {
const res = await iteratee(item, index);
result.push(res);
}, options);
return result;
}
/**
* Reduces a list of values down to a single value.
*
* Memo is the initial state of the reduction, and each successive step of it should be returned by `iteratee()`.
* The `iteratee()` is passed three arguments: the `memo`, then the `value` and `index` of the iteration.
*
* If no `memo` is passed to the initial invocation of reduce, the `iteratee()` is not invoked on the first element of the list.
* The first element is instead passed as the memo in the invocation of the `iteratee()` on the next element in the list.
*
* If `iteratee()` returns a `Promise` then it's awaited before a next call.
*
* @param iteratee
* @param memo Initial state of the reduction.
* @param [options] All `reduce()` parameters.
*/
async reduce<T>(iteratee: DatasetReducer<T, Data>, memo: T, options: DatasetIteratorOptions = {}): Promise<T> {
checkStorageAccess();
let currentMemo: T = memo;
const wrappedFunc: DatasetConsumer<Data> = async (item, index) => {
return Promise
.resolve()
.then(() => {
return !index && currentMemo === undefined
? item
: iteratee(currentMemo, item, index);
})
.then((newMemo) => {
currentMemo = newMemo as T;
});
};
await this.forEach(wrappedFunc, options);
return currentMemo;
}
/**
* Removes the dataset either from the Apify cloud storage or from the local directory,
* depending on the mode of operation.
*/
async drop(): Promise<void> {
checkStorageAccess();
await this.client.delete();
const manager = StorageManager.getManager(Dataset, this.config);
manager.closeStorage(this);
}
/**
* Opens a dataset and returns a promise resolving to an instance of the {@apilink Dataset} class.
*
* Datasets are used to store structured data where each object stored has the same attributes,
* such as online store products or real estate offers.
* The actual data is stored either on the local filesystem or in the cloud.
*
* For more details and code examples, see the {@apilink Dataset} class.
*
* @param [datasetIdOrName]
* ID or name of the dataset to be opened. If `null` or `undefined`,
* the function returns the default dataset associated with the crawler run.
* @param [options] Storage manager options.
*/
static async open<Data extends Dictionary = Dictionary>(datasetIdOrName?: string | null, options: StorageManagerOptions = {}): Promise<Dataset<Data>> {
checkStorageAccess();
ow(datasetIdOrName, ow.optional.string);
ow(options, ow.object.exactShape({
config: ow.optional.object.instanceOf(Configuration),
storageClient: ow.optional.object,
}));
options.config ??= Configuration.getGlobalConfig();
options.storageClient ??= options.config.getStorageClient();
await purgeDefaultStorages({ onlyPurgeOnce: true, client: options.storageClient, config: options.config });
const manager = StorageManager.getManager<Dataset<Data>>(this, options.config);
return manager.openStorage(datasetIdOrName, options.storageClient);
}
/**
* Stores an object or an array of objects to the default {@apilink Dataset} of the current crawler run.
*
* This is just a convenient shortcut for {@apilink Dataset.pushData}.
* For example, calling the following code:
* ```javascript
* await Dataset.pushData({ myValue: 123 });
* ```
*
* is equivalent to:
* ```javascript
* const dataset = await Dataset.open();
* await dataset.pushData({ myValue: 123 });
* ```
*
* For more information, see {@apilink Dataset.open} and {@apilink Dataset.pushData}
*
* **IMPORTANT**: Make sure to use the `await` keyword when calling `pushData()`,
* otherwise the crawler process might finish before the data are stored!
*
* @param item Object or array of objects containing data to be stored in the default dataset.
* The objects must be serializable to JSON and the JSON representation of each object must be smaller than 9MB.
* @ignore
*/
static async pushData<Data extends Dictionary = Dictionary>(item: Data | Data[]): Promise<void> {
const dataset = await this.open();
return dataset.pushData(item);
}
/**
* Returns {@apilink DatasetContent} object holding the items in the dataset based on the provided parameters.
*/
static async getData<Data extends Dictionary = Dictionary>(options: DatasetDataOptions = {}): Promise<DatasetContent<Data>> {
const dataset = await this.open();
return dataset.getData(options);
}
}
/**
* User-function used in the `Dataset.forEach()` API.
*/
export interface DatasetConsumer<Data> {
/**
* @param item Current {@apilink Dataset} entry being processed.
* @param index Position of current {@apilink Dataset} entry.
*/
(item: Data, index: number): Awaitable<void>;
}
/**
* User-function used in the `Dataset.map()` API.
*/
export interface DatasetMapper<Data, R> {
/**
* User-function used in the `Dataset.map()` API.
* @param item Current {@apilink Dataset} entry being processed.
* @param index Position of current {@apilink Dataset} entry.
*/
(item: Data, index: number): Awaitable<R>;
}
/**
* User-function used in the `Dataset.reduce()` API.
*/
export interface DatasetReducer<T, Data> {
/**
* @param memo Previous state of the reduction.
* @param item Current {@apilink Dataset} entry being processed.
* @param index Position of current {@apilink Dataset} entry.
*/
(memo: T, item: Data, index: number): Awaitable<T>;
}
export interface DatasetOptions {
id: string;
name?: string;
client: StorageClient;
}
export interface DatasetContent<Data> {
/** Total count of entries in the dataset. */
total: number;
/** Count of dataset entries returned in this set. */
count: number;
/** Position of the first returned entry in the dataset. */
offset: number;
/** Maximum number of dataset entries requested. */
limit: number;
/** Dataset entries based on chosen format parameter. */
items: Data[];
/** Should the results be in descending order. */
desc?: boolean;
}