-
Notifications
You must be signed in to change notification settings - Fork 3
/
package.json
51 lines (51 loc) · 1.22 KB
/
package.json
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
{
"name": "robot-directives",
"description": "Parse robot directives within HTML meta and/or HTTP headers.",
"version": "0.5.0-alpha",
"license": "MIT",
"author": "Steven Vachon <contact@svachon.com> (https://svachon.com)",
"repository": "github:stevenvachon/robot-directives",
"main": "lib",
"browser": "lib-es5",
"dependencies": {
"deep-freeze-node": "^1.1.3",
"isbot": "^2.2.1",
"useragent": "^2.3.0"
},
"devDependencies": {
"@babel/cli": "^7.4.3",
"@babel/core": "^7.4.3",
"@babel/preset-env": "^7.4.3",
"chai": "^4.2.0",
"coveralls": "^3.0.3",
"mocha": "^6.1.4",
"nyc": "^14.0.0"
},
"engines": {
"node": ">= 8"
},
"scripts": {
"ci": "npm run test && nyc report --reporter=text-lcov | coveralls",
"posttest": "nyc report --reporter=text-summary --reporter=html",
"prepublishOnly": "npm test && babel lib/ --out-dir=lib-es5/ --presets=@babel/env --source-maps",
"test": "nyc --silent mocha test.js --check-leaks --bail"
},
"files": [
"lib",
"lib-es5"
],
"keywords": [
"crawlers",
"header",
"html",
"http",
"meta",
"metadata",
"nofollow",
"noindex",
"robots",
"robots.txt",
"seo",
"spiders"
]
}