-
Notifications
You must be signed in to change notification settings - Fork 120
/
apiscript.js
1371 lines (1249 loc) · 59.8 KB
/
apiscript.js
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
// Flags
// whether to check for duplicate translation or not during create
var checkduplicate = true;
// whether json is required in translation or not
var jsonrequired = true
// whether to generate latin or not using translate.py script
var generateLatin = false
const fs = require('fs');
const path = require('path');
// Requires for md5 hash generations for fonts to check duplicates
const crypto = require('crypto');
// Requires for python or other system binaries to launch
const {
spawnSync
} = require('child_process');
const {
firefox
} = require('playwright');
// https://www.npmjs.com/package/extract-zip
// Required to extract the downloaded fonts .zip file from fontsquirrel
const extract = require('extract-zip')
// Required to get font details
const opentype = require('opentype.js');
// Folder that holds all the translations that needs to be added
var startDir = path.join(__dirname, "start")
// Folder that holds all the quran editions
var editionsFolder = "editions"
var editionsDir = path.join(__dirname, editionsFolder)
// Stores the files for download and as backup
var databaseDir = path.join(__dirname, 'database')
// Stores translations in line by line format of 6236 lines
var linebylineDir = path.join(databaseDir, 'linebyline')
// Directory containing all the fonts
var fontsDir = path.join(__dirname, 'fonts')
var startUrl = "https://cdn.jsdelivr.net/gh/fawazahmed0/quran-api@"
var apiVersion = '1'
// API url
var url = startUrl + apiVersion + "/"
// Stores the translation files snippets and it's json,retreieves them from linebylineDir
var jsondb = {}
// spaces to be used for prettify/json.stringify
var prettyindent = 4
// Creating line to [chapter,verseNo] mappings
// Array containing number of verses in each chapters
var chaplength = [7, 286, 200, 176, 120, 165, 206, 75, 129, 109, 123, 111, 43, 52, 99, 128, 111, 110, 98, 135, 112, 78, 118, 64, 77, 227, 93, 88, 69, 60, 34, 30, 73, 54, 45, 83, 182, 88, 75, 85, 54, 53, 89, 59, 37, 35, 38, 29, 18, 45, 60, 49, 62, 55, 78, 96, 29, 22, 24, 13, 14, 11, 11, 18, 12, 12, 30, 52, 52, 44, 28, 28, 20, 56, 40, 31, 50, 40, 46, 42, 29, 19, 36, 25, 22, 17, 19, 26, 30, 20, 15, 21, 11, 8, 8, 19, 5, 8, 8, 11, 11, 8, 3, 9, 5, 4, 7, 3, 6, 3, 5, 4, 5, 6]
// contains chapter verse mappings for each line
var mappings = []
// Number of verses in quran
const VERSE_LENGTH = 6236
for (i = 1; i <= 114; i++) {
for (j = 1; j <= chaplength[i - 1]; j++) {
mappings.push([i, j])
}
}
// capitalizes all the first letters in a sentense
var capitalize = words => words.split(' ').map(w => w[0].toUpperCase() + w.substring(1)).join(' ')
// stores iso codes
var isocodes;
//stores maqra,juz etc start and end
var qinfo;
// stores the google language codes
var gLangCodes;
// https://stackoverflow.com/a/5767589
// access node command line args
var argarr = process.argv.slice(2);
// Page to add translation text and get the direction of text and also for font generation
var page
// browser variable, to allow easily closing the browser from anywhere in the script
var browser
// We will make few directories, incase if they doesn't exists, this will help to run this script even if we
// partially cloned this repo
fs.mkdirSync(startDir, {
recursive: true
});
fs.mkdirSync(editionsDir, {
recursive: true
});
fs.mkdirSync(fontsDir, {
recursive: true
});
fs.mkdirSync(path.join(databaseDir, "originals"), {
recursive: true
});
fs.mkdirSync(path.join(databaseDir, "chapterverse"), {
recursive: true
});
// function that will run on running this script
async function start() {
logmsg("\nBEGIN:\n" + process.argv.join(' '), true)
// Print the help and how to use the script file and arguments, same as given in contribute
if (argarr[0] == undefined)
helpPrint()
else if ("" + argarr[0].toLowerCase().trim() == "create")
await create()
else if (argarr[0].toLowerCase().trim() == "update")
await create(true)
else if (argarr[0].toLowerCase().trim() == "search")
search(argarr.slice(1))
else if (argarr[0].toLowerCase().trim() == "delete") {
// storing the data in jsondb, so editionsListingsGen function can work and create editions.json
await jsonDB()
deleteEditions(argarr.slice(1))
} else if (argarr[0].toLowerCase().trim() == "fontsgen")
fontsGen()
else
helpPrint()
}
// calling start()
start()
// Prints the information on how to use this tool, mirror whatever is written in contribution.md
function helpPrint() {
var filename = path.basename(__filename);
console.log("\nUsage: node ", filename, " [arguments]")
console.log("\n\narguments:")
console.log("\ncreate\ncreates the database in REST architecture, paste your files in start directory and then run this command\nExample: node ", filename, " create")
console.log("\nupdate\nupdates the database, copy the edition that needs to be edited from database/chapterverse directory and paste that edition in start directory and then perform any editing you want in the file and then run this command\nExample: node ", filename, " update")
console.log("\ndelete\ndeletes the edition from database\nExample: node ", filename, " delete editionNameToDelete")
console.log("\nsearch\nsearches the provided line in database\nExample: node ", filename, ' search "verseToSearch"')
console.log("\nfontsgen\ngenerates the fonts, paste your fonts in start direcotry and then run this command\nExample: node ", filename, ' fontsgen')
}
// function that will generate the editions, it will take the files from startDir
async function create(update) {
// saving database snippet, filename and it's json data in jsondb variable
await jsonDB()
// saving isocodes in json
isocodes = fs.readFileSync(path.join(__dirname, 'isocodes', 'iso-codes.min.json')).toString();
isocodes = JSON.parse(isocodes)
// saving quran data such as how many rukus, etc, this will be used to generate the rukus endpoint
qinfo = fs.readFileSync(path.join(__dirname, 'info.min.json')).toString();
qinfo = JSON.parse(qinfo)
// saving google translate language codes in json
gLangCodes = fs.readFileSync(path.join(__dirname, 'isocodes', 'google-codes.min.json')).toString();
gLangCodes = JSON.parse(gLangCodes)
// Launching browser as we will need it for checking direction of the language
await launchBrowser()
// Saving flags used
logmsg("\nFlags Used\ncheckduplicate:" + checkduplicate + "\njsonrequired:" + jsonrequired + "\ngenerateLatin:" + generateLatin, true)
// Starting to read files in startDir
for (var filename of fs.readdirSync(startDir)) {
// we don't want to read .gitkeep, it is used as a placeholder for start direcotory to exist in git
if (filename == '.gitkeep')
continue;
logmsg("\nStarting to create files for " + filename)
// Reading the file and retrieving as array, filteredarr, and jsondata inside it
// filterarr doesn't contain jsondata and empty lines in it
var [orgarr, cleanarr, jsondata] = readDBTxt(path.join(startDir, filename))
if (!jsondata) {
logmsg("\nNo JSON found in file " + filename + " or please enter the json in correct format", true)
jsondata = {}
if (jsonrequired) {
var tempjson = '{"author":"Name of Author","language":"Name of language","source":"","comments":""}'
logmsg("\nAdd json at end of file in the following format:\n\n" + JSON.stringify(JSON.parse(tempjson), null, prettyindent))
continue
}
}
if (!Array.isArray(cleanarr)) {
logmsg("\nproblem in " + filename + " format, so ignoring this file")
continue
}
// Now we have to check and make sure same copy doesn't exists in the repo, here we will use the linebylineDir to check
var duplicatefilename = checkduplicateTrans(cleanarr)
// We don't want to check for duplicates during update operation
if (duplicatefilename && !update) {
logmsg("\nThis file " + filename + " seems to be a duplicate copy of edition " + duplicatefilename.replace(/(\.[^\.]*$)/i, ""))
if (checkduplicate)
continue
else
logmsg("\ncheckduplicate is set to false, so a duplicate copy of this translation will be created in the database")
}
// Cleaning and lowercasing the json
jsondata = cleanifyObject(jsondata)
// This will store the language name and isocode
var temp = []
// Cleaning or defining the language
if (!jsondata['language']) {
// detect langauge if it's not already defined
temp = isoLangMap(detectLang(cleanarr))
} else {
// mapping the langauge name to iso language name
temp = isoLangMap([jsondata['language']])
// if the above fails, then we will have to detect the language
if (!Array.isArray(temp))
temp = isoLangMap(detectLang(cleanarr))
}
// If the language is still not detected, we will skip this translation
if (!Array.isArray(temp)) {
logmsg("\nPlease specify the proper iso name of the language in json, Skipping this translation " + filename)
continue;
}
// Assigning isoname of the language and it's isocode
jsondata['language'] = temp[0]
jsondata['iso'] = temp[1]
// if this is update operation
if (update) {
if (!fs.existsSync(path.join(linebylineDir, filename))) {
logmsg("\nEdition with name " + filename.replace(/(\.[^\.]*$)/i, "") + " does not exist in the database")
continue
} else if (jsondata['name'] && filename.replace(/(\.[^\.]*$)/i, "") != jsondata['name']) {
logmsg("\nYou are trying to change edition name to " + jsondata['name'] + " this should be done only in very rare cases only")
if (fs.existsSync(path.join(linebylineDir, jsondata['name'] + '.txt'))) {
logmsg("\nA Edition with same name as " + jsondata['name'] + " exists in the database, you will have to give a new edition name")
continue
}
}
// Path to edition-la and edition-lad
var latinDPath = path.join(linebylineDir, filename.replace(/(\.[^\.]*$)/i, "-lad$1"))
var latinPath = path.join(linebylineDir, filename.replace(/(\.[^\.]*$)/i, "-la$1"))
// stores the index as key and text as value , it will stores lines having edited changes
var uniqueobj = {}
var fulllatinarr = [];
if (fs.existsSync(latinDPath))
[ , fulllatinarr] = readDBTxt(latinDPath)
else if (fs.existsSync(latinPath))
[ , fulllatinarr] = readDBTxt(latinPath)
// if the edition-la or edition-lad existed
if (fulllatinarr) {
// stores the old edition data, this will be used to compare the lines which are having changes
// so that only the changed line will be used for latin generation, as it's expensive process
var [ , oldEditionArr] = readDBTxt(path.join(linebylineDir, filename))
// storing unique/edited lines with their index in uniqueobj
for (var i = 0; i < oldEditionArr.length; i++) {
if (oldEditionArr[i] != cleanarr[i])
uniqueobj[i] = cleanarr[i]
}
}
// delete the old editions
var oldEditionName = filename.replace(/(\.[^\.]*$)/i, "")
deleteEditions([oldEditionName])
}
// generating edition
logmsg("\nGenerating Edition")
// if edition name exists in the file during update process, we will use that as edition name
if (update && jsondata['name'])
var genJSON = await generateEdition(cleanarr, jsondata, jsondata['name'])
else
var genJSON = await generateEdition(cleanarr, jsondata)
if (update) {
// if this is update operation, then we will give more preference to json data from file, instead of generated json data
for (var [key, val] of Object.entries(genJSON)) {
// we want the autogenerated link, linkmin and direction values
if (jsondata[key] && !key.includes('link') && key != 'direction')
genJSON[key] = jsondata[key]
}
}
// if the language is latin diacritical, then we will generate a non diacritical version attaching -la to it
if (isLatin(cleanarr) && isDiacritic(cleanarr)) {
logmsg("\nGenerating non diacritical Edition")
var nonDiacriticarr = cleanarr.join('\n').normalize('NFD').replace(/[\u0300-\u036f]/g, '').split('\n')
// generating edition
var latinGenJSON = await generateEdition(nonDiacriticarr, jsondata, genJSON['name'] + '-la')
}
// we will try to generate latin script if the language is not latin, it will make it easier for users to read the translation
// who studied in english, but have a different mother tongue and are not well versed in reading the mother tongue script, for example me
else if (!isLatin(cleanarr) && generateLatin) {
logmsg("\nPlease wait trying to generate latin script for this translation, it will take 5-10mins to complete")
var genLatinarr = []
// if this is create operation or if the latin script for the edition doesn't exist, we will try building one
if (!update || fulllatinarr.length == 0) {
genLatinarr = await genLatin(cleanarr, genJSON['name'])
logmsg("\ninside new gen", true)
} else if (Object.keys(uniqueobj).length == 0) {
// if there are no edited lines in the updated translation, maybe only json data was updated in the file
// So we will use the old latin translation
genLatinarr = fulllatinarr
logmsg("\ninside no gen", true)
} else {
// generating latin only for edited lines
var latinreturn = await genLatin(Object.values(uniqueobj), genJSON['name'])
var i = 0
logmsg("\ninside few gen", true)
// The return latin script should be an array and the no of lines we passed, should be returned back
if (Array.isArray(latinreturn) && Object.keys(uniqueobj).length == latinreturn.length) {
for (var key of Object.keys(uniqueobj))
fulllatinarr[key] = latinreturn[i++]
}
genLatinarr = fulllatinarr
}
// if the latin script was generated, then we will use that to generate editions with -la and -lad appended to it
if (Array.isArray(genLatinarr) && isLatin(genLatinarr) && genLatinarr.length == VERSE_LENGTH) {
// assuming the generated latin is non diacritical
var nonDiacriticarr = genLatinarr
logmsg("\nCreating Latin Script for the language")
// generating diacritical variant and non diacritical variant if the generated latin script is diacritical
if (isDiacritic(genLatinarr)) {
logmsg("\nGenerating diacritical and non diacritical Editions for this language")
await generateEdition(genLatinarr, jsondata, genJSON['name'] + '-lad')
// generating non diacritical variant array
nonDiacriticarr = genLatinarr.join('\n').normalize('NFD').replace(/[\u0300-\u036f]/g, '').split('\n')
}
// Generate non diacritical variant
var latinGenJSON = await generateEdition(nonDiacriticarr, jsondata, genJSON['name'] + '-la')
}
}
// move the file for which update/create have been completed from startDir to originals dir
fs.renameSync(path.join(startDir, filename), path.join(databaseDir, "originals", filename))
}
// Generate the editions.json
editionsListingsGen()
// close the browser when everything is done
await browser.close();
}
// This function is a wrapper to generate json and generate the files in the database
async function generateEdition(arr, jsondata, editionName) {
var genJSON = {}
// use the editionName to generateJSON if it's defined
if (editionName)
genJSON = await generateJSON(arr, jsondata, editionName)
else
genJSON = await generateJSON(arr, jsondata)
// generate files in database
generateFiles(arr, genJSON)
// save the json data and snippet inside the jsondb variable
await jsonDB(genJSON['name'] + '.txt')
logmsg("\n Generated edition " + genJSON['name'])
return genJSON
}
// Generate the files and folder for the edition in REST architecture
function generateFiles(arr, json) {
// We will generate the files and folders only if we are in github actions where CI env is set to trues and not on dev environment
if(process.env.CI){
for (var key of Object.keys(qinfo)) {
// we don't want to generate for verses,chapters, as it's generation logic is different
if (key != "verses" && key != "chapters" && key != "sajdas") {
fs.mkdirSync(path.join(editionsDir, json['name'], key), {
recursive: true
});
for (var j = 0; j < qinfo[key]['references'].length; j++) {
var fromchap = qinfo[key]['references'][j].start.chapter
var fromverse = qinfo[key]['references'][j].start.verse
if (qinfo[key]['references'][j + 1]) {
var tochap = qinfo[key]['references'][j + 1].start.chapter
var toverse = qinfo[key]['references'][j + 1].start.verse
} else {
var tochap = ""
var toverse = ""
}
// writing the chapter and verse in the specified folder
chapVerseWriter(fromchap, fromverse, tochap, toverse, arr, path.join(editionsDir, json['name'], key, j + 1 + ''), key)
}
}
}
// save whole quran json
chapVerseWriter(1, 1, "", "", arr, path.join(editionsDir, json['name']), 'quran')
var k = 0
// creating chapter and single verse json
for (var i = 1; i <= 114; i++) {
fs.mkdirSync(path.join(editionsDir, json['name'], i + ''), {
recursive: true
});
// writing chapter json
chapVerseWriter(i, 1, i, chaplength[i - 1], arr, path.join(editionsDir, json['name'], i + ''), 'chapter', true)
for (var j = 1; j <= chaplength[i - 1]; j++) {
// creating single verse json
var vjson = {}
vjson['chapter'] = i
vjson['verse'] = j
vjson['text'] = arr[k++]
fs.writeFileSync(path.join(editionsDir, json['name'], i + '', j + ".min.json"), JSON.stringify(vjson))
fs.writeFileSync(path.join(editionsDir, json['name'], i + '', j + ".json"), JSON.stringify(vjson, null, prettyindent))
}
}
// attaching chap|verseno|versetext to array
var chapversearray = arr.map((value, index) => mappings[index][0] + '|' + mappings[index][1] + '|' + value)
// saving in chapterverse folder as back
fs.writeFileSync(path.join(databaseDir, 'chapterverse', json['name'] + ".txt"), chapversearray.join('\n') + '\n' + JSON.stringify(json, null, prettyindent))
}
// saving in linebylineDir as back
fs.writeFileSync(path.join(linebylineDir, json['name'] + ".txt"), arr.join('\n') + '\n' + JSON.stringify(json, null, prettyindent))
}
// validates the translation and returns a clean translation without the numbers etc
function validateCleanTrans(arr, filename) {
var filterarr = arr.filter(elem => !/^\s*$/.test(elem))
// In proper format if the number of lines are even 6236 after filtering empty lines
if (filterarr.length == VERSE_LENGTH)
return cleanTrans(filterarr)
// assuming it's having few empty verses in it, we will print which lines and chapter/verse is having problem
else if(arr.length == VERSE_LENGTH)
PrintEmptyVerse(arr, filename)
// asuuming there is number pattern of verses such as 1|1|Praise be to God
else {
var j = 0;
var stop = 0
// specifies the limit, i.e next number of lines to search for next verse
var limit = 5
// Saving the original arr as we will be modifying it
var orgarr = [...arr]
// Filtering the array from empty lines, uncomment this if the input translation has so many empty lines in it
arr = arr.filter(elem => !/^\s*$/.test(elem))
// stores the last line string which had valid number pattern like 1|1|Praise be to God
// setting this to first line, incase the translation is without number patters and in wrong format, it will print error in line 0
var laststr = orgarr[0]
// stores the numbers from line
// var numsarr;
// This will store arr which has valid lines, it will help in backtrack incase the limit is reached and we wanted to check the lastcorrect index for the search value
var temparr = []
// Stores the last index with valid number pattern
var lastindex = 0;
// Stores the regex verse pattern
var versePattern;
// stores the regex split values in an array
var splitval = []
for (i = 0; i < arr.length; i++) {
// Make versePattern for only if the mappings is defined and we are not trying to access which doesn't exists
if (mappings[j])
versePattern = new RegExp('[^0-9]' + mappings[j][1] + '[^0-9]');
// Checking number patter of verses, j is the line number and mappings[j][1] accesses the verse number in that specific line number
// we are checking that this line has same verse number as in mappings
if (mappings[j] && versePattern.test(' ' + arr[i] + ' ')) {
j++;
// storing the line with valid number pattern
laststr = arr[i];
// resetting the stop as we found a valid verse line
stop = 0
// Saving the index with valid number pattern to help in backtracking
lastindex = i
// Saving the valid arr to help in backtracking
temparr = [...arr]
} else {
// merging the newline content assuming it's part of verse, as we did not find next verse pattern
arr[i - 1] = arr[i - 1] + " " + arr[i]
// console.log("merged " + arr[i - 1])
// logmsg("\nmerged the below line \n" + arr[i - 1], true)
// deleting the current line as it is already merged with previous line
arr.splice(i, 1)
// Going back to the previous line containing the merged content
i--;
// if the limit is reached and we still did not find the next verse or if the next loop will not happen
// and we have not yet got the right pattern verses, we will backtrack using temparray
// and check for search value at valid index, to make sure that next verse was not merged with the valid laststr
if (stop++ == limit || (i + 1 == arr.length && j != mappings.length)) {
// Making sure temparr is not empty
if (temparr.length != 0)
// Checking whether the last valid index had the verse pattern in it
splitval = temparr[lastindex].split(versePattern)
// If the verse pattern exists, then we will cut and push that line into new line, otherwise we will stop, as it might be missing some verses in it
if (splitval[1]) {
temparr.splice(lastindex, 1, splitval[0], temparr[lastindex].replace(splitval[0], ""))
logmsg("Two verse on same line " + temparr.slice(lastindex, lastindex + 3), true)
// Saving valid array, in next loop the number pattern will be detected and the process will go as usual
arr = [...temparr]
} else {
break;
}
}
}
}
// if the above loop went till end i.e 6236 lines without finding any invalid verse pattern line, it means the file is in proper format
if (j == VERSE_LENGTH){
// Just making sure it doesn't have any empty lines
var cleantrans = cleanTrans(arr)
var filterarr = cleantrans.filter(elem => !/^\s*$/.test(elem))
// If the translation has empty lines, we will let them know which lines have problem
if(filterarr.length!=VERSE_LENGTH)
PrintEmptyVerse(cleantrans, filename)
else
return filterarr
}
else
logmsg("\nerror while checking the " + filename + " it might be missing chapter " + mappings[j][0] + " and verse " + mappings[j][1] + " check at roughly lineno " + (orgarr.indexOf(laststr) + 1) + " after or before the line '" + laststr + "' ,error will be somewhere near this line")
}
}
// Cleaning translation from numbers, special symbols etc
function cleanTrans(arr) {
for (i = 0; i < arr.length; i++) {
// https://en.wikipedia.org/wiki/List_of_Unicode_characters#Basic_Latin
// This will remove all special symbols and numbers from starting and ending of verse
arr[i] = arr[i].replace(/^[\u0020-\u0040|\u005b-\u0060|\u007b-\u007e|\s|\n|\r|\p{N}]{1,20}/u, " ").replace(/^\s*\w{1}\s*(\.|\)|\}|>|\])+[\u0020-\u0040|\u005b-\u0060|\u007b-\u007e|\s|\n|\r|\p{N}]{0,7}/ui, " ").replace(/[\u0020-\u0040|\u005b-\u0060|\u007b-\u007e|\s|\n|\r|\p{N}]{1,15}$/u, " ").replace(/[\r\n]/g, " ").replace(/\s\s+/g, " ").trim()
// Checking partially open/close bracket exists or not at begninning of verse
var bracket1 = arr[i].match(/^[^\[|\(|\<|\{]+(\]|\)|\>|\})/)
// Checking partially open/close bracket exists or not at end of verse
var bracket2 = arr[i].match(/(\[|\(|\<|\{)[^\]|\)|\>|\}]+$/)
// closing partially open/close bracket in the verse
// closing partially open/close bracket at the beginning of verse
if (bracket1)
arr[i] = getOppoBracket(bracket1[0].slice(-1)) + arr[i]
// closing partially open/close bracket at the end of verse
if (bracket2)
arr[i] = arr[i] + getOppoBracket(bracket2[0].slice(0, 1))
}
return arr
}
// clean the string from special symbols,numbers,multiple spaces etc , this is used for string comparision
function cleanify(str) {
return str.replace(/[\u0020-\u0040|\u005b-\u0060|\u007b-\u007e|\s|\n|\p{N}]+/ugi, " ").replace(/^\s*\w{1}\s+/i, " ").replace(/\s\s+/g, " ").trim().toLowerCase()
}
// Prints the empty line in the translation, so user can fix it
function PrintEmptyVerse(arr, filename){
arr.forEach((elem,index) => {
if(/^\s*$/.test(elem))
logmsg("\nerror while checking the " + filename + " it might be missing chapter " + mappings[index][0] + " and verse " + mappings[index][1] + ", you will find an empty verse, check near line \n"+arr[index-1])
});
}
// returns opposite bracket
function getOppoBracket(str) {
switch (str) {
case '(':
return ')'
case ')':
return '('
case '<':
return '>'
case '>':
return '<'
case '[':
return ']'
case ']':
return '['
case '{':
return '}'
case '}':
return '{'
default:
return ''
}
}
// function to delete list of editions from the database
// This will also remove the auto generated -la and -lad of edition
function deleteEditions(arr) {
var deleted = false
for (var val of arr) {
for (var editionname of [val, val + '-la', val + '-lad']) {
// array containing paths to delete
var pathsarr = []
pathsarr.push(path.join(editionsDir, editionname))
pathsarr.push(path.join(editionsDir, editionname + '.json'))
pathsarr.push(path.join(editionsDir, editionname + '.min.json'))
pathsarr.push(path.join(linebylineDir, editionname + '.txt'))
pathsarr.push(path.join(databaseDir, 'chapterverse', editionname + '.txt'))
for (var pathToDelete of pathsarr) {
if (fs.existsSync(pathToDelete)) {
deleted = true
if (fs.statSync(pathToDelete).isDirectory())
fs.rmdirSync(pathToDelete, {
recursive: true
})
else
fs.unlinkSync(pathToDelete)
}
}
// Deleting also from temporary jsondb variable
delete jsondb[editionname + '.txt']
logmsg("\n deletion completed for " + editionname)
}
}
// Generate the editions.json if any of the file was deleted
if (deleted)
editionsListingsGen()
}
// reads the jsondb variable to generate editions.json
function editionsListingsGen() {
var newjsondb = {}
// we will always keep the editions.json in sorted order, so it's easier to find
var sortedkeys = Object.keys(jsondb).sort()
for (var name of sortedkeys) {
// removing .txt from filename and replace dash with underscore as many programming languages doesn't support - (dash) in json object key
var newname = name.replace(/\..*/gi, "").replace(/-/gi, "_")
newjsondb[newname] = jsondb[name]['jsondata']
}
fs.writeFileSync(editionsDir + ".json", JSON.stringify(newjsondb, null, prettyindent))
fs.writeFileSync(editionsDir + ".min.json", JSON.stringify(newjsondb))
logmsg("\neditions.json and editions.min.json generated")
}
// Generates font and it's listings fonts.json
async function fontsGen() {
// we will excute all of this only if startDir has files in it, as all the operations are pretty expensive
// remember there is one .gitkeep file already, so files in startDir should be more than 1
if (fs.readdirSync(startDir).length > 1) {
// Get all FontNames without extension and in lower case
var fontNoExtension = fs.readdirSync(fontsDir).map(elem => elem.replace(/\.[^\.]*$/, "").toLowerCase())
// Removing duplicate values
fontNoExtension = [...new Set(fontNoExtension)];
// Rename all files with suffix -org in startdir, as we use -org in our code to denote original files
// And rename duplicate filenames in startsDir comparing with fontNoExtension array
// And rename to standard format
for (var val of fs.readdirSync(startDir)) {
// extension of file
var extension = val.match(/\.[^\.]*$/gi) || [""]
// removing all the underscores,extension etc and making it to standard format
var name = val.replace(extension[0], "").replace(/[^A-Z0-9]/gi, " ").replace(/([A-Z]+)/g, " $1").trim().replace(/\s\s*/g, "-").toLowerCase()
// Rename the filename if it ends with -org
while (/-org$/i.test(name))
name = name.replace(/-org$/i, "")
// Checking for duplicateNames and renaming them in startDir
if (fontNoExtension.includes(name) || fontNoExtension.includes(name + '-org')) {
for (var i = 1;; i++) {
// Get the number at the ending of the name, so we can increment it if it exists
var lastNum = name.match(/\d+$/) || [0]
lastNum = parseInt(lastNum[0])
// Increament the above number if it exists to get a new name
var newFileName = name.replace(/\d+$/, "") + (lastNum + 1)
if (!fontNoExtension.includes(newFileName) && !fontNoExtension.includes(newFileName + '-org'))
break;
}
// assigning the newFileName of the file to the name
name = newFileName
}
// rename the file with the standard generated name
fs.renameSync(path.join(startDir, val), path.join(startDir, name + extension[0]))
}
// Directory that will store the files temporarily
var tempDir = path.join(__dirname, 'temp')
var fontsDirCrypto = dirHash(fontsDir)
var startDirCrypto = dirHash(startDir)
// We have to remove the .gitkeep from the json object
var gitkeepKey = Object.keys(startDirCrypto).find(key => startDirCrypto[key] == '.gitkeep')
delete startDirCrypto[gitkeepKey]
var startDirHashArr = Object.keys(startDirCrypto)
// array containing duplicate hashs
var duplicateHash = Object.keys(fontsDirCrypto).filter(elem => startDirHashArr.includes(elem));
for (var val of duplicateHash) {
logmsg("\nfont " + startDirCrypto[val] + " is a duplicate of " + fontsDirCrypto[val] + " hence we are ignoring it")
// Removing the hash and filename from the json, as we will not generate for duplicate fonts
delete startDirCrypto[val]
}
// Execute the below only if the start directory json has values in it
if (Object.keys(startDirCrypto).length > 0) {
logmsg("\n\nGenerating fonts Please wait, it will take around 10-15mins\n\n" + "we will generate fonts for " + Object.values(startDirCrypto).join(', '))
await launchBrowser('https://www.fontsquirrel.com/tools/webfont-generator', tempDir)
}
// array containing full path of font files in startDir
var fullPathArr = Object.values(startDirCrypto).map(elem => path.join(startDir, elem))
// We will generate files one by one
for (var fontwithPath of fullPathArr) {
// Delete temp if it exists, to clean previous partial data due to script error
fs.rmdirSync(tempDir, {
recursive: true
})
// Making the temporary directory
fs.mkdirSync(tempDir, {
recursive: true
});
logmsg("\nStarting Generation for " + path.basename(fontwithPath))
try {
var downloadedZip = await downloadFonts([fontwithPath])
// extract zip to tempDir
if (downloadedZip) {
await extract(downloadedZip, {
dir: tempDir
})
logmsg("\nGeneration Complete for " + path.basename(fontwithPath))
}
} catch (error) {
logmsg("\nThere was error for " + path.basename(fontwithPath) + "\n" + error, true)
logmsg("\nSeems like the fonts generation did not go well for " + path.basename(fontwithPath) + ", anyways we will still add the font \nassuming the fontsquirrel doesn't support generation for these fonts")
// close and relaunch browser if the error is big, and not recoverable
await browser.close()
await launchBrowser('https://www.fontsquirrel.com/tools/webfont-generator', tempDir)
}
// move all the generated files ending with valid font extensions to the fonts directory
for (var val of fs.readdirSync(tempDir)) {
if (/\.svg$/i.test(val) || /\.eot$/i.test(val) || /\.ttf$/i.test(val) || /\.woff$/i.test(val) || /\.woff2$/i.test(val))
fs.renameSync(path.join(tempDir, val), path.join(fontsDir, val))
}
// move all the valid fonts from startDir to fontsDir with -org suffix
// extension of file
var extension = path.basename(fontwithPath).match(/\.[^\.]*$/gi) || [""]
var newName = path.basename(fontwithPath).replace(extension[0], "") + '-org' + extension[0]
fs.renameSync(fontwithPath, path.join(fontsDir, newName))
// Delete the tempDir
fs.rmdirSync(tempDir, {
recursive: true
})
}
// closing the browser incase, if it was launched
if (fullPathArr.length > 0)
await browser.close()
} // End of if
// Now have to generate fonts.json listings
fontsListingsGen()
}
// Generates the fonts.json and fonts.min.json by reading from the fontsDir
function fontsListingsGen() {
// I might need this code in future, incase the fontsquirrel renames the uploaded fonts to non standard forms
// var fontsarr = fs.readdirSync(fontsDir)
// for (var fontname of fontsarr) {
// Getting the extension of fontname
// var extension = fontname.match(/\.[^\.]*$/gi) || [""]
// Replacing the special symbols,spaces etc with - and lowering the case
// var name = fontname.replace(extension[0], "").replace(/[^A-Z0-9]/gi, " ").replace(/([A-Z]+)/g, " $1").trim().replace(/\s\s*/g, "-").toLowerCase() + extension[0].toLowerCase().trim()
// renaming the fonts to proper names and removing special symbols etc
// fs.renameSync(path.join(fontsDir, fontname), path.join(fontsDir, name))
// }
// getting sorted array of fonts, all the fonts have already been renamed to standard form in fontsgen()
var fontsarr = fs.readdirSync(fontsDir).sort()
var fontjson = {}
// Return the metadata of fonts in fontsDir of only those filenames passes the given regex
var fontsMetaJSON = fontsMeta(fontsDir, /\-org\.?[a-z]*$/i)
// generating fontjson
for (var fontname of fontsarr) {
// Getting the extension of fontname
var extension = fontname.match(/\.[^\.]*$/gi) || [""]
var fontWithNoExtension = fontname.replace(extension[0], "")
// Removing the extension from fontname and also changing the dash to underscore as many programming languages doesn't support - (dash) in json
// Also we are removing the -org from it, as original files will have -org in it
var keyname = fontWithNoExtension.replace(/-org$/i, "").replace(/-/gi, "_")
// we don't want to define again if it's already defined, otherwise it will replace the previous values
if (!fontjson[keyname])
fontjson[keyname] = {}
var innerkey = extension[0].trim().substring(1)
// if this fontname endsWith -org , then it means this was the original file used to generate the other fonts
// we will also assign the metadata stored in fontsMetaJSON
if (fontWithNoExtension.endsWith('-org')) {
// set key to original for fonts that ends with -org
innerkey = 'original'
fontjson[keyname]['name'] = keyname.replace(/_/gi, "-")
fontjson[keyname]['font'] = fontsMetaJSON[fontname]['fontname']
fontjson[keyname]['designer'] = fontsMetaJSON[fontname]['designer']
fontjson[keyname]['source'] = fontsMetaJSON[fontname]['source']
fontjson[keyname]['version'] = fontsMetaJSON[fontname]['version']
}
fontjson[keyname][innerkey] = url + 'fonts/' + fontname
}
fs.writeFileSync(path.join(__dirname, "fonts.json"), JSON.stringify(fontjson, null, prettyindent))
fs.writeFileSync(path.join(__dirname, "fonts.min.json"), JSON.stringify(fontjson))
logmsg("\nfonts.json and fonts.min.json generated")
}
// Takes input arg as path to fonts directory and regex of filenames to consider in fonts directory and returns object containing meta data about fonts
// If no nameregex is defined it returns the meta data of all fonts in directory
function fontsMeta(PathToFontsDir, nameregex) {
// stores fonts meta data such as fontname, authorname, version etc
var metaobj = {}
var fontsarr = fs.readdirSync(PathToFontsDir)
// If nameregex is defined, then we will only consider those files which pases the regex for filename
if (nameregex)
fontsarr = fontsarr.filter(e => nameregex.test(e))
for (var filename of fontsarr) {
// Initializing the metaobj with filename as key
metaobj[filename] = {
"fontname": "",
"designer": "",
"source": "",
"version": ""
}
// stores font object returned by opentypejs
var font = {}
// Loading fonts using opentypejs
try {
font = opentype.loadSync(path.join(PathToFontsDir, filename))
} catch (e) {}
// Assigning the values
if (font && font.tables && font.tables.name) {
if (font.tables.name.fontFamily) {
metaobj[filename]['fontname'] = font.tables.name.fontFamily.en || ""
if (font.tables.name.fontSubfamily && font.tables.name.fontSubfamily.en)
metaobj[filename]['fontname'] = metaobj[filename]['fontname'] + ' ' + font.tables.name.fontSubfamily.en
}
if (font.tables.name.version && font.tables.name.version.en) {
try {
metaobj[filename]['version'] = parseFloat(font.tables.name.version.en.match(/\d+\.?\d*/)[0])
} catch (e) {}
}
if (font.tables.name.designer)
metaobj[filename]['designer'] = font.tables.name.designer.en || ""
if (font.tables.name.manufacturer && font.tables.name.manufacturer.en &&
font.tables.name.manufacturer.en.trim() != metaobj[filename]['designer'].trim())
metaobj[filename]['designer'] = metaobj[filename]['designer'] + ' , ' + font.tables.name.manufacturer.en
if (font.tables.name.designerURL)
metaobj[filename]['source'] = font.tables.name.designerURL.en || ""
if (font.tables.name.manufacturerURL && font.tables.name.manufacturerURL.en &&
font.tables.name.manufacturerURL.en.trim() != metaobj[filename]['source'].trim())
metaobj[filename]['source'] = metaobj[filename]['source'] + ' , ' + font.tables.name.manufacturerURL.en
//replace any unnecessary commas at front or back of string
metaobj[filename]['designer'] = metaobj[filename]['designer'].trim().replace(/^,|,$/g, '')
metaobj[filename]['source'] = metaobj[filename]['source'].trim().replace(/^,|,$/g, '')
}
}
return metaobj
}
// Take arg as array of paths for which fonts needs to be generated, and then it downloads the generated contenct in the default
// download directory of the browser, which is tempDir in our case, and it returns the filename with path of the download file which is zip
// For some reason passing multiple paths, causes problem to fontsquirrel, so we are passing array of single path each time to solve this issue
async function downloadFonts(pathArr) {
if (pathArr.length == 0)
return
// Reloading the page, to make sure we get whole new page, and all the old details are removed
await page.reload({
timeout: 60000
})
// This function generates fonts using fontsquirrel webfont generator
//https://github.com/microsoft/playwright/issues/2351
await page.check('input[value="expert"]');
await page.check('input[value="ttf"]');
await page.check('input[value="svg"]');
await page.check('input[value="eotz"]');
await page.check('input[value="woff"]');
await page.check('input[value="woff2"]');
// This will keep the fonts size reduced
await page.check('input[value="keep"]');
// This will avoid subsetting the font to latin only languages
await page.check('input[id="no_subsetting"]');
await page.check('input[name="agreement"]');
await page.fill('input[name="filename_suffix"]', '');
// We will add single random file first
await page.setInputFiles('input[type="file"]', pathArr);
// If there are many fonts to be uploaded and the size of fonts is large, then it will take more
// Time and this uploadwaitTime time has to be increased then
// Giving 600 seconds for each file upload
var uploadwaitTime = 600000
try {
// dismiss all the dialogs that will popup due to font being already webfont
// https://playwright.dev/#version=v1.3.0&path=docs%2Fapi.md&q=class-dialog
page.on('dialog', async dialog => {
try{await dialog.dismiss()}catch(error){
logmsg("\nThere was error in dialog dismiss for " + pathArr + "\n" + error, true)
logmsg("\nError in dialog dismiss, seems like either this is already a webfont or a blacklisted font which is not supported")
}
});
// https://playwright.dev/#version=v1.3.0&path=docs%2Fnetwork.md&q=handle-file-downloads
const [download] = await Promise.all([
page.waitForEvent('download', {
timeout: uploadwaitTime
}), // wait for download to start
page.click('#ffgen_downloadbtn', {
timeout: uploadwaitTime
})
]);
var downloadedFilePath = await download.path();
return downloadedFilePath
} catch (error) {
logmsg("\nThere was error for " + pathArr + "\n" + error, true)
return
}
}
// https://stackoverflow.com/a/11869589/2437224
// https://stackoverflow.com/a/37227430/2437224
// https://stackoverflow.com/questions/46441667/reading-binary-data-in-node-js
// Takes directory as an input and return an object containing list of md5 hash as keys and it's filenames as values
// Used for checking duplicate files
function dirHash(pathToDir) {
// JSON that will store md5 hash as keys and it's filename as values
var cryptoJSON = {}
for (var filename of fs.readdirSync(pathToDir)) {
var buf = fs.readFileSync(path.join(pathToDir, filename))
var hash = crypto.createHash('md5').update(buf, "binary").digest('hex');
cryptoJSON[hash] = filename
}
return cryptoJSON
}
// Stores the translation files snippets and it's json,retreieves them from linebylineDir
async function jsonDB(singlefile) {
for (var filename of fs.readdirSync(linebylineDir)) {
// if single file is defined, we will break the loop at end, we will only read that particular files data into jsondb object
if (singlefile)
filename = singlefile
var filepath = path.join(linebylineDir, filename)
var fileSize = fs.statSync(filepath).size
// read the first 2% bytes of file to be stored as snippet in jsondb object
var data = await streamRead(filepath, 0, parseInt(fileSize*0.02))
jsondb[filename] = {}
// taking verse from line 11 to 20 and storing it for searching and duplicate detection
jsondb[filename]['snippet'] = data.split(/\r?\n/).slice(10, 20).join('\n')
// reading last 6k bytes of file to fetch json
data = await streamRead(filepath, fileSize - 6000)
// parse the json
jsondb[filename]['jsondata'] = getJSONInArray(data.split(/\r?\n/))[0]
// break the loop, as we only wanted to add one file
if (singlefile)
break;
}
}
// reads the file using streams, start is the starting byte and end is the bytes to read
async function streamRead(pathtofile, start, end) {
var readstream;
if (start && !end)
readstream = fs.createReadStream(pathtofile, {
start: start
});
else if (!start && end)
readstream = fs.createReadStream(pathtofile, {
end: end
});
else if (!start && !end)
readstream = fs.createReadStream(pathtofile);
else
readstream = fs.createReadStream(pathtofile, {
start: start,
end: end
});
var data = ''
for await (var chunk of readstream) {
data = data + chunk.toString()
}
return data
}
// gets the JSON from end of array, returns [jsondata, i], where i is the position from end where jsondata was parsed successfully
function getJSONInArray(arr) {
var i = 0
while (!isValidJSON(arr.slice(--i).join('\n')) && i > -100);
if (i != -100)
return [JSON.parse(arr.slice(i).join('\n')), i]
}
// Checks for duplicate files in the database
function checkduplicateTrans(arr) {
for (var filename of fs.readdirSync(linebylineDir)) {
if (cleanify(arr.join('\n')).includes(cleanify(jsondb[filename]['snippet'])))
return filename
}
}
// Generates the json with standard naming conventions
async function generateJSON(arr, newjson, editionName) {
var isocode = newjson['iso']
// Deleting iso key, as it might create a bug in the future, as this key was added later to solve an issue in actions enviroment
delete newjson['iso']
// capitalize first letters
newjson['language'] = capitalize(newjson['language'])
// If values are undefined we will assign it as empty string
newjson['author'] = newjson['author'] || "unknown"
// Removing special symbols and diacritics from authors name
newjson['author'] = newjson['author'].normalize('NFD').replace(/[\u0300-\u036f]/g, '').replace(/[^A-Za-z\s\.\,]+/gi, " ").replace(/\s\s+/gi, " ").toLowerCase().trim()
newjson['author'] = capitalize(newjson['author'])
// If values are undefined we will assign it as empty string
newjson['source'] = newjson['source'] || ""
newjson['comments'] = newjson['comments'] || ""