Skip to content
Permalink
Branch: master
Find file Copy path
Find file Copy path
Fetching contributors…
Cannot retrieve contributors at this time
executable file 104 lines (99 sloc) 2.15 KB
// Configuration file for MediaWiki Jobrunner
// This file is managed by Puppet
{
"groups": {
"basic": {
// Number of runner processes in this group
"runners": 19,
// Job types to include ("*" means "all")
"include": [
"*"
],
// Job types to exempt (useful when combined with "*")
"exclude": [
"AssembleUploadChunks",
"ParsoidCacheUpdateJobOnDependencyChange",
"ParsoidCacheUpdateJobOnEdit",
"PublishStashedFile",
"gwtoolsetGWTFileBackendCleanupJob",
"gwtoolsetUploadMediafileJob",
"gwtoolsetUploadMetadataJob",
"uploadFromUrl",
"webVideoTranscode"
],
"low-priority": [
"cirrusSearchLinksUpdate",
"htmlCacheUpdate",
"refreshLinks"
]
},
"gwt": {
"runners": 1,
"include": [
"gwtoolsetUploadMetadataJob",
"gwtoolsetUploadMediafileJob",
"gwtoolsetGWTFileBackendCleanupJob"
]
},
"parsoid": {
"runners": 20,
"include": [
"ParsoidCacheUpdateJobOnEdit"
],
"low-priority": [
"ParsoidCacheUpdateJobOnDependencyChange"
]
},
"transcode": {
"runners": 0,
"include": [
"webVideoTranscode"
]
},
"upload": {
"runners": 7,
"include": [
"AssembleUploadChunks",
"PublishStashedFile",
"uploadFromUrl"
]
}
},
"limits": {
// How many times to let jobs be recycled before abandoning
"attempts": {
"*": 3
},
// How long jobs can be claimed before being recycled
"claimTTL": {
"*": 3600,
"webVideoTranscode": 86400
},
// runJobs.php process time limits
"real": {
"*": 300,
"webVideoTranscode": 86400
},
// runJobs.php memory limits
"memory": {
"*": "300M"
}
},
"redis": {
// Ready queue trackers
"aggregators": [
"rdb1001.eqiad.wmnet",
"rdb1003.eqiad.wmnet"
],
// Main queue servers
"queues": [
"rdb1001.eqiad.wmnet",
"rdb1003.eqiad.wmnet"
],
"password": "XXXXX"
},
// Address to send statsd metrics to
"statsd": "statsd.eqiad.wmnet:8125",
// Command used to launch a runner for a given job queue
"dispatcher": "php /srv/mediawiki/multiversion/MWScript.php runJobs.php --wiki=%(db)x --type=%(type)x --maxtime=%(maxtime)x --memory-limit=%(maxmem)x --result=json"
}
You can’t perform that action at this time.