MediaWiki REL1_32
JobRunner.php
Go to the documentation of this file.
1<?php
26use Liuggio\StatsdClient\Factory\StatsdDataFactory;
27use Psr\Log\LoggerAwareInterface;
28use Psr\Log\LoggerInterface;
29use Wikimedia\ScopedCallback;
32
39class JobRunner implements LoggerAwareInterface {
41 protected $config;
43 protected $debug;
44
48 protected $logger;
49
50 const MAX_ALLOWED_LAG = 3; // abort if more than this much DB lag is present
51 const LAG_CHECK_PERIOD = 1.0; // check replica DB lag this many seconds
52 const ERROR_BACKOFF_TTL = 1; // seconds to back off a queue due to errors
53 const READONLY_BACKOFF_TTL = 30; // seconds to back off a queue due to read-only errors
54
58 public function setDebugHandler( $debug ) {
59 $this->debug = $debug;
60 }
61
66 public function setLogger( LoggerInterface $logger ) {
67 $this->logger = $logger;
68 }
69
73 public function __construct( LoggerInterface $logger = null ) {
74 if ( $logger === null ) {
75 $logger = LoggerFactory::getInstance( 'runJobs' );
76 }
77 $this->setLogger( $logger );
78 $this->config = MediaWikiServices::getInstance()->getMainConfig();
79 }
80
105 public function run( array $options ) {
106 $jobClasses = $this->config->get( 'JobClasses' );
107 $profilerLimits = $this->config->get( 'TrxProfilerLimits' );
108
109 $response = [ 'jobs' => [], 'reached' => 'none-ready' ];
110
111 $type = $options['type'] ?? false;
112 $maxJobs = $options['maxJobs'] ?? false;
113 $maxTime = $options['maxTime'] ?? false;
114 $noThrottle = isset( $options['throttle'] ) && !$options['throttle'];
115
116 // Bail if job type is invalid
117 if ( $type !== false && !isset( $jobClasses[$type] ) ) {
118 $response['reached'] = 'none-possible';
119 return $response;
120 }
121
122 // Bail out if DB is in read-only mode
123 if ( wfReadOnly() ) {
124 $response['reached'] = 'read-only';
125 return $response;
126 }
127
128 $lbFactory = MediaWikiServices::getInstance()->getDBLoadBalancerFactory();
129 if ( $lbFactory->hasTransactionRound() ) {
130 throw new LogicException( __METHOD__ . ' called with an active transaction round.' );
131 }
132 // Bail out if there is too much DB lag.
133 // This check should not block as we want to try other wiki queues.
134 list( , $maxLag ) = $lbFactory->getMainLB()->getMaxLag();
135 if ( $maxLag >= self::MAX_ALLOWED_LAG ) {
136 $response['reached'] = 'replica-lag-limit';
137 return $response;
138 }
139
140 // Catch huge single updates that lead to replica DB lag
141 $trxProfiler = Profiler::instance()->getTransactionProfiler();
142 $trxProfiler->setLogger( LoggerFactory::getInstance( 'DBPerformance' ) );
143 $trxProfiler->setExpectations( $profilerLimits['JobRunner'], __METHOD__ );
144
145 // Some jobs types should not run until a certain timestamp
146 $backoffs = []; // map of (type => UNIX expiry)
147 $backoffDeltas = []; // map of (type => seconds)
148 $wait = 'wait'; // block to read backoffs the first time
149
150 $group = JobQueueGroup::singleton();
151 $stats = MediaWikiServices::getInstance()->getStatsdDataFactory();
152 $jobsPopped = 0;
153 $timeMsTotal = 0;
154 $startTime = microtime( true ); // time since jobs started running
155 $lastCheckTime = 1; // timestamp of last replica DB check
156 do {
157 // Sync the persistent backoffs with concurrent runners
158 $backoffs = $this->syncBackoffDeltas( $backoffs, $backoffDeltas, $wait );
159 $blacklist = $noThrottle ? [] : array_keys( $backoffs );
160 $wait = 'nowait'; // less important now
161
162 if ( $type === false ) {
163 $job = $group->pop(
164 JobQueueGroup::TYPE_DEFAULT,
165 JobQueueGroup::USE_CACHE,
166 $blacklist
167 );
168 } elseif ( in_array( $type, $blacklist ) ) {
169 $job = false; // requested queue in backoff state
170 } else {
171 $job = $group->pop( $type ); // job from a single queue
172 }
173
174 if ( $job ) { // found a job
175 ++$jobsPopped;
176 $popTime = time();
177 $jType = $job->getType();
178
179 WebRequest::overrideRequestId( $job->getRequestId() );
180
181 // Back off of certain jobs for a while (for throttling and for errors)
182 $ttw = $this->getBackoffTimeToWait( $job );
183 if ( $ttw > 0 ) {
184 // Always add the delta for other runners in case the time running the
185 // job negated the backoff for each individually but not collectively.
186 $backoffDeltas[$jType] = isset( $backoffDeltas[$jType] )
187 ? $backoffDeltas[$jType] + $ttw
188 : $ttw;
189 $backoffs = $this->syncBackoffDeltas( $backoffs, $backoffDeltas, $wait );
190 }
191
192 $info = $this->executeJob( $job, $lbFactory, $stats, $popTime );
193 if ( $info['status'] !== false || !$job->allowRetries() ) {
194 $group->ack( $job ); // succeeded or job cannot be retried
195 }
196
197 // Back off of certain jobs for a while (for throttling and for errors)
198 if ( $info['status'] === false && mt_rand( 0, 49 ) == 0 ) {
199 $ttw = max( $ttw, $this->getErrorBackoffTTL( $info['error'] ) );
200 $backoffDeltas[$jType] = isset( $backoffDeltas[$jType] )
201 ? $backoffDeltas[$jType] + $ttw
202 : $ttw;
203 }
204
205 $response['jobs'][] = [
206 'type' => $jType,
207 'status' => ( $info['status'] === false ) ? 'failed' : 'ok',
208 'error' => $info['error'],
209 'time' => $info['timeMs']
210 ];
211 $timeMsTotal += $info['timeMs'];
212
213 // Break out if we hit the job count or wall time limits...
214 if ( $maxJobs && $jobsPopped >= $maxJobs ) {
215 $response['reached'] = 'job-limit';
216 break;
217 } elseif ( $maxTime && ( microtime( true ) - $startTime ) > $maxTime ) {
218 $response['reached'] = 'time-limit';
219 break;
220 }
221
222 // Don't let any of the main DB replica DBs get backed up.
223 // This only waits for so long before exiting and letting
224 // other wikis in the farm (on different masters) get a chance.
225 $timePassed = microtime( true ) - $lastCheckTime;
226 if ( $timePassed >= self::LAG_CHECK_PERIOD || $timePassed < 0 ) {
227 $success = $lbFactory->waitForReplication( [
228 'ifWritesSince' => $lastCheckTime,
229 'timeout' => self::MAX_ALLOWED_LAG,
230 ] );
231 if ( !$success ) {
232 $response['reached'] = 'replica-lag-limit';
233 break;
234 }
235 $lastCheckTime = microtime( true );
236 }
237
238 // Bail if near-OOM instead of in a job
239 if ( !$this->checkMemoryOK() ) {
240 $response['reached'] = 'memory-limit';
241 break;
242 }
243 }
244 } while ( $job ); // stop when there are no jobs
245
246 // Sync the persistent backoffs for the next runJobs.php pass
247 if ( $backoffDeltas ) {
248 $this->syncBackoffDeltas( $backoffs, $backoffDeltas, 'wait' );
249 }
250
251 $response['backoffs'] = $backoffs;
252 $response['elapsed'] = $timeMsTotal;
253
254 return $response;
255 }
256
261 private function getErrorBackoffTTL( $error ) {
262 return strpos( $error, 'DBReadOnlyError' ) !== false
265 }
266
274 private function executeJob( Job $job, LBFactory $lbFactory, $stats, $popTime ) {
275 $jType = $job->getType();
276 $msg = $job->toString() . " STARTING";
277 $this->logger->debug( $msg, [
278 'job_type' => $job->getType(),
279 ] );
280 $this->debugCallback( $msg );
281
282 // Run the job...
283 $rssStart = $this->getMaxRssKb();
284 $jobStartTime = microtime( true );
285 try {
286 $fnameTrxOwner = get_class( $job ) . '::run'; // give run() outer scope
287 if ( !$job->hasExecutionFlag( $job::JOB_NO_EXPLICIT_TRX_ROUND ) ) {
288 $lbFactory->beginMasterChanges( $fnameTrxOwner );
289 }
290 $status = $job->run();
291 $error = $job->getLastError();
292 $this->commitMasterChanges( $lbFactory, $job, $fnameTrxOwner );
293 // Run any deferred update tasks; doUpdates() manages transactions itself
294 DeferredUpdates::doUpdates();
295 } catch ( Exception $e ) {
296 MWExceptionHandler::rollbackMasterChangesAndLog( $e );
297 $status = false;
298 $error = get_class( $e ) . ': ' . $e->getMessage();
299 }
300 // Always attempt to call teardown() even if Job throws exception.
301 try {
302 $job->teardown( $status );
303 } catch ( Exception $e ) {
304 MWExceptionHandler::logException( $e );
305 }
306
307 // Commit all outstanding connections that are in a transaction
308 // to get a fresh repeatable read snapshot on every connection.
309 // Note that jobs are still responsible for handling replica DB lag.
310 $lbFactory->flushReplicaSnapshots( __METHOD__ );
311 // Clear out title cache data from prior snapshots
312 MediaWikiServices::getInstance()->getLinkCache()->clear();
313 $timeMs = intval( ( microtime( true ) - $jobStartTime ) * 1000 );
314 $rssEnd = $this->getMaxRssKb();
315
316 // Record how long jobs wait before getting popped
317 $readyTs = $job->getReadyTimestamp();
318 if ( $readyTs ) {
319 $pickupDelay = max( 0, $popTime - $readyTs );
320 $stats->timing( 'jobqueue.pickup_delay.all', 1000 * $pickupDelay );
321 $stats->timing( "jobqueue.pickup_delay.$jType", 1000 * $pickupDelay );
322 }
323 // Record root job age for jobs being run
324 $rootTimestamp = $job->getRootJobParams()['rootJobTimestamp'];
325 if ( $rootTimestamp ) {
326 $age = max( 0, $popTime - wfTimestamp( TS_UNIX, $rootTimestamp ) );
327 $stats->timing( "jobqueue.pickup_root_age.$jType", 1000 * $age );
328 }
329 // Track the execution time for jobs
330 $stats->timing( "jobqueue.run.$jType", $timeMs );
331 // Track RSS increases for jobs (in case of memory leaks)
332 if ( $rssStart && $rssEnd ) {
333 $stats->updateCount( "jobqueue.rss_delta.$jType", $rssEnd - $rssStart );
334 }
335
336 if ( $status === false ) {
337 $msg = $job->toString() . " t={job_duration} error={job_error}";
338 $this->logger->error( $msg, [
339 'job_type' => $job->getType(),
340 'job_duration' => $timeMs,
341 'job_error' => $error,
342 ] );
343
344 $msg = $job->toString() . " t=$timeMs error={$error}";
345 $this->debugCallback( $msg );
346 } else {
347 $msg = $job->toString() . " t={job_duration} good";
348 $this->logger->info( $msg, [
349 'job_type' => $job->getType(),
350 'job_duration' => $timeMs,
351 ] );
352
353 $msg = $job->toString() . " t=$timeMs good";
354 $this->debugCallback( $msg );
355 }
356
357 return [ 'status' => $status, 'error' => $error, 'timeMs' => $timeMs ];
358 }
359
363 private function getMaxRssKb() {
364 $info = wfGetRusage() ?: [];
365 // see https://linux.die.net/man/2/getrusage
366 return isset( $info['ru_maxrss'] ) ? (int)$info['ru_maxrss'] : null;
367 }
368
374 private function getBackoffTimeToWait( Job $job ) {
375 $throttling = $this->config->get( 'JobBackoffThrottling' );
376
377 if ( !isset( $throttling[$job->getType()] ) || $job instanceof DuplicateJob ) {
378 return 0; // not throttled
379 }
380
381 $itemsPerSecond = $throttling[$job->getType()];
382 if ( $itemsPerSecond <= 0 ) {
383 return 0; // not throttled
384 }
385
386 $seconds = 0;
387 if ( $job->workItemCount() > 0 ) {
388 $exactSeconds = $job->workItemCount() / $itemsPerSecond;
389 // use randomized rounding
390 $seconds = floor( $exactSeconds );
391 $remainder = $exactSeconds - $seconds;
392 $seconds += ( mt_rand() / mt_getrandmax() < $remainder ) ? 1 : 0;
393 }
394
395 return (int)$seconds;
396 }
397
406 private function loadBackoffs( array $backoffs, $mode = 'wait' ) {
407 $file = wfTempDir() . '/mw-runJobs-backoffs.json';
408 if ( is_file( $file ) ) {
409 $noblock = ( $mode === 'nowait' ) ? LOCK_NB : 0;
410 $handle = fopen( $file, 'rb' );
411 if ( !flock( $handle, LOCK_SH | $noblock ) ) {
412 fclose( $handle );
413 return $backoffs; // don't wait on lock
414 }
415 $content = stream_get_contents( $handle );
416 flock( $handle, LOCK_UN );
417 fclose( $handle );
418 $ctime = microtime( true );
419 $cBackoffs = json_decode( $content, true ) ?: [];
420 foreach ( $cBackoffs as $type => $timestamp ) {
421 if ( $timestamp < $ctime ) {
422 unset( $cBackoffs[$type] );
423 }
424 }
425 } else {
426 $cBackoffs = [];
427 }
428
429 return $cBackoffs;
430 }
431
443 private function syncBackoffDeltas( array $backoffs, array &$deltas, $mode = 'wait' ) {
444 if ( !$deltas ) {
445 return $this->loadBackoffs( $backoffs, $mode );
446 }
447
448 $noblock = ( $mode === 'nowait' ) ? LOCK_NB : 0;
449 $file = wfTempDir() . '/mw-runJobs-backoffs.json';
450 $handle = fopen( $file, 'wb+' );
451 if ( !flock( $handle, LOCK_EX | $noblock ) ) {
452 fclose( $handle );
453 return $backoffs; // don't wait on lock
454 }
455 $ctime = microtime( true );
456 $content = stream_get_contents( $handle );
457 $cBackoffs = json_decode( $content, true ) ?: [];
458 foreach ( $deltas as $type => $seconds ) {
459 $cBackoffs[$type] = isset( $cBackoffs[$type] ) && $cBackoffs[$type] >= $ctime
460 ? $cBackoffs[$type] + $seconds
461 : $ctime + $seconds;
462 }
463 foreach ( $cBackoffs as $type => $timestamp ) {
464 if ( $timestamp < $ctime ) {
465 unset( $cBackoffs[$type] );
466 }
467 }
468 ftruncate( $handle, 0 );
469 fwrite( $handle, json_encode( $cBackoffs ) );
470 flock( $handle, LOCK_UN );
471 fclose( $handle );
472
473 $deltas = [];
474
475 return $cBackoffs;
476 }
477
483 private function checkMemoryOK() {
484 static $maxBytes = null;
485 if ( $maxBytes === null ) {
486 $m = [];
487 if ( preg_match( '!^(\d+)(k|m|g|)$!i', ini_get( 'memory_limit' ), $m ) ) {
488 list( , $num, $unit ) = $m;
489 $conv = [ 'g' => 1073741824, 'm' => 1048576, 'k' => 1024, '' => 1 ];
490 $maxBytes = $num * $conv[strtolower( $unit )];
491 } else {
492 $maxBytes = 0;
493 }
494 }
495 $usedBytes = memory_get_usage();
496 if ( $maxBytes && $usedBytes >= 0.95 * $maxBytes ) {
497 $msg = "Detected excessive memory usage ({used_bytes}/{max_bytes}).";
498 $this->logger->error( $msg, [
499 'used_bytes' => $usedBytes,
500 'max_bytes' => $maxBytes,
501 ] );
502
503 $msg = "Detected excessive memory usage ($usedBytes/$maxBytes).";
504 $this->debugCallback( $msg );
505
506 return false;
507 }
508
509 return true;
510 }
511
516 private function debugCallback( $msg ) {
517 if ( $this->debug ) {
518 call_user_func_array( $this->debug, [ wfTimestamp( TS_DB ) . " $msg\n" ] );
519 }
520 }
521
533 private function commitMasterChanges( LBFactory $lbFactory, Job $job, $fnameTrxOwner ) {
534 $syncThreshold = $this->config->get( 'JobSerialCommitThreshold' );
535
536 $time = false;
537 $lb = $lbFactory->getMainLB();
538 if ( $syncThreshold !== false && $lb->getServerCount() > 1 ) {
539 // Generally, there is one master connection to the local DB
540 $dbwSerial = $lb->getAnyOpenConnection( $lb->getWriterIndex() );
541 // We need natively blocking fast locks
542 if ( $dbwSerial && $dbwSerial->namedLocksEnqueue() ) {
543 $time = $dbwSerial->pendingWriteQueryDuration( $dbwSerial::ESTIMATE_DB_APPLY );
544 if ( $time < $syncThreshold ) {
545 $dbwSerial = false;
546 }
547 } else {
548 $dbwSerial = false;
549 }
550 } else {
551 // There are no replica DBs or writes are all to foreign DB (we don't handle that)
552 $dbwSerial = false;
553 }
554
555 if ( !$dbwSerial ) {
556 $lbFactory->commitMasterChanges(
557 $fnameTrxOwner,
558 // Abort if any transaction was too big
559 [ 'maxWriteDuration' => $this->config->get( 'MaxJobDBWriteDuration' ) ]
560 );
561
562 return;
563 }
564
565 $ms = intval( 1000 * $time );
566
567 $msg = $job->toString() . " COMMIT ENQUEUED [{job_commit_write_ms}ms of writes]";
568 $this->logger->info( $msg, [
569 'job_type' => $job->getType(),
570 'job_commit_write_ms' => $ms,
571 ] );
572
573 $msg = $job->toString() . " COMMIT ENQUEUED [{$ms}ms of writes]";
574 $this->debugCallback( $msg );
575
576 // Wait for an exclusive lock to commit
577 if ( !$dbwSerial->lock( 'jobrunner-serial-commit', $fnameTrxOwner, 30 ) ) {
578 // This will trigger a rollback in the main loop
579 throw new DBError( $dbwSerial, "Timed out waiting on commit queue." );
580 }
581 $unlocker = new ScopedCallback( function () use ( $dbwSerial, $fnameTrxOwner ) {
582 $dbwSerial->unlock( 'jobrunner-serial-commit', $fnameTrxOwner );
583 } );
584
585 // Wait for the replica DBs to catch up
586 $pos = $lb->getMasterPos();
587 if ( $pos ) {
588 $lb->waitForAll( $pos );
589 }
590
591 // Actually commit the DB master changes
592 $lbFactory->commitMasterChanges(
593 $fnameTrxOwner,
594 // Abort if any transaction was too big
595 [ 'maxWriteDuration' => $this->config->get( 'MaxJobDBWriteDuration' ) ]
596 );
597 ScopedCallback::consume( $unlocker );
598 }
599}
wfTempDir()
Tries to get the system directory for temporary files.
wfGetRusage()
Get system resource usage of current request context.
wfReadOnly()
Check whether the wiki is in read-only mode.
wfTimestamp( $outputtype=TS_UNIX, $ts=0)
Get a timestamp string in one of various formats.
No-op job that does nothing.
Job queue runner utility methods.
Definition JobRunner.php:39
const LAG_CHECK_PERIOD
Definition JobRunner.php:51
setDebugHandler( $debug)
Definition JobRunner.php:58
const READONLY_BACKOFF_TTL
Definition JobRunner.php:53
run(array $options)
Run jobs of the specified number/type for the specified time.
const MAX_ALLOWED_LAG
Definition JobRunner.php:50
__construct(LoggerInterface $logger=null)
Definition JobRunner.php:73
callable null $debug
Debug output handler.
Definition JobRunner.php:43
setLogger(LoggerInterface $logger)
Definition JobRunner.php:66
syncBackoffDeltas(array $backoffs, array &$deltas, $mode='wait')
Merge the current backoff expiries from persistent storage.
getErrorBackoffTTL( $error)
executeJob(Job $job, LBFactory $lbFactory, $stats, $popTime)
debugCallback( $msg)
Log the job message.
commitMasterChanges(LBFactory $lbFactory, Job $job, $fnameTrxOwner)
Issue a commit on all masters who are currently in a transaction and have made changes to the databas...
getBackoffTimeToWait(Job $job)
checkMemoryOK()
Make sure that this script is not too close to the memory usage limit.
const ERROR_BACKOFF_TTL
Definition JobRunner.php:52
loadBackoffs(array $backoffs, $mode='wait')
Get the previous backoff expiries from persistent storage On I/O or lock acquisition failure this ret...
Config $config
Definition JobRunner.php:41
Class to both describe a background job and handle jobs.
Definition Job.php:30
PSR-3 logger instance factory.
MediaWikiServices is the service locator for the application scope of MediaWiki.
Database error base class.
Definition DBError.php:30
An interface for generating database load balancers.
Definition LBFactory.php:39
beginMasterChanges( $fname=__METHOD__)
Flush any master transaction snapshots and set DBO_TRX (if DBO_DEFAULT is set)
getMainLB( $domain=false)
commitMasterChanges( $fname=__METHOD__, array $options=[])
Commit changes and clear view snapshots on all master connections.
flushReplicaSnapshots( $fname=__METHOD__)
Commit all replica DB transactions so as to flush any REPEATABLE-READ or SSI snapshot.
deferred txt A few of the database updates required by various functions here can be deferred until after the result page is displayed to the user For updating the view updating the linked to tables after a etc PHP does not yet have any way to tell the server to actually return and disconnect while still running these but it might have such a feature in the future We handle these by creating a deferred update object and putting those objects on a global list
Definition deferred.txt:11
see documentation in includes Linker php for Linker::makeImageLink & $time
Definition hooks.txt:1841
null means default in associative array with keys and values unescaped Should be merged with default with a value of false meaning to suppress the attribute in associative array with keys and values unescaped & $options
Definition hooks.txt:2050
Status::newGood()` to allow deletion, and then `return false` from the hook function. Ensure you consume the 'ChangeTagAfterDelete' hook to carry out custom deletion actions. $tag:name of the tag $user:user initiating the action & $status:Status object. See above. 'ChangeTagsListActive':Allows you to nominate which of the tags your extension uses are in active use. & $tags:list of all active tags. Append to this array. 'ChangeTagsAfterUpdateTags':Called after tags have been updated with the ChangeTags::updateTags function. Params:$addedTags:tags effectively added in the update $removedTags:tags effectively removed in the update $prevTags:tags that were present prior to the update $rc_id:recentchanges table id $rev_id:revision table id $log_id:logging table id $params:tag params $rc:RecentChange being tagged when the tagging accompanies the action, or null $user:User who performed the tagging when the tagging is subsequent to the action, or null 'ChangeTagsAllowedAdd':Called when checking if a user can add tags to a change. & $allowedTags:List of all the tags the user is allowed to add. Any tags the user wants to add( $addTags) that are not in this array will cause it to fail. You may add or remove tags to this array as required. $addTags:List of tags user intends to add. $user:User who is adding the tags. 'ChangeUserGroups':Called before user groups are changed. $performer:The User who will perform the change $user:The User whose groups will be changed & $add:The groups that will be added & $remove:The groups that will be removed 'Collation::factory':Called if $wgCategoryCollation is an unknown collation. $collationName:Name of the collation in question & $collationObject:Null. Replace with a subclass of the Collation class that implements the collation given in $collationName. 'ConfirmEmailComplete':Called after a user 's email has been confirmed successfully. $user:user(object) whose email is being confirmed 'ContentAlterParserOutput':Modify parser output for a given content object. Called by Content::getParserOutput after parsing has finished. Can be used for changes that depend on the result of the parsing but have to be done before LinksUpdate is called(such as adding tracking categories based on the rendered HTML). $content:The Content to render $title:Title of the page, as context $parserOutput:ParserOutput to manipulate 'ContentGetParserOutput':Customize parser output for a given content object, called by AbstractContent::getParserOutput. May be used to override the normal model-specific rendering of page content. $content:The Content to render $title:Title of the page, as context $revId:The revision ID, as context $options:ParserOptions for rendering. To avoid confusing the parser cache, the output can only depend on parameters provided to this hook function, not on global state. $generateHtml:boolean, indicating whether full HTML should be generated. If false, generation of HTML may be skipped, but other information should still be present in the ParserOutput object. & $output:ParserOutput, to manipulate or replace 'ContentHandlerDefaultModelFor':Called when the default content model is determined for a given title. May be used to assign a different model for that title. $title:the Title in question & $model:the model name. Use with CONTENT_MODEL_XXX constants. 'ContentHandlerForModelID':Called when a ContentHandler is requested for a given content model name, but no entry for that model exists in $wgContentHandlers. Note:if your extension implements additional models via this hook, please use GetContentModels hook to make them known to core. $modeName:the requested content model name & $handler:set this to a ContentHandler object, if desired. 'ContentModelCanBeUsedOn':Called to determine whether that content model can be used on a given page. This is especially useful to prevent some content models to be used in some special location. $contentModel:ID of the content model in question $title:the Title in question. & $ok:Output parameter, whether it is OK to use $contentModel on $title. Handler functions that modify $ok should generally return false to prevent further hooks from further modifying $ok. 'ContribsPager::getQueryInfo':Before the contributions query is about to run & $pager:Pager object for contributions & $queryInfo:The query for the contribs Pager 'ContribsPager::reallyDoQuery':Called before really executing the query for My Contributions & $data:an array of results of all contribs queries $pager:The ContribsPager object hooked into $offset:Index offset, inclusive $limit:Exact query limit $descending:Query direction, false for ascending, true for descending 'ContributionsLineEnding':Called before a contributions HTML line is finished $page:SpecialPage object for contributions & $ret:the HTML line $row:the DB row for this line & $classes:the classes to add to the surrounding< li > & $attribs:associative array of other HTML attributes for the< li > element. Currently only data attributes reserved to MediaWiki are allowed(see Sanitizer::isReservedDataAttribute). 'ContributionsToolLinks':Change tool links above Special:Contributions $id:User identifier $title:User page title & $tools:Array of tool links $specialPage:SpecialPage instance for context and services. Can be either SpecialContributions or DeletedContributionsPage. Extensions should type hint against a generic SpecialPage though. 'ConvertContent':Called by AbstractContent::convert when a conversion to another content model is requested. Handler functions that modify $result should generally return false to disable further attempts at conversion. $content:The Content object to be converted. $toModel:The ID of the content model to convert to. $lossy: boolean indicating whether lossy conversion is allowed. & $result:Output parameter, in case the handler function wants to provide a converted Content object. Note that $result->getContentModel() must return $toModel. 'ContentSecurityPolicyDefaultSource':Modify the allowed CSP load sources. This affects all directives except for the script directive. If you want to add a script source, see ContentSecurityPolicyScriptSource hook. & $defaultSrc:Array of Content-Security-Policy allowed sources $policyConfig:Current configuration for the Content-Security-Policy header $mode:ContentSecurityPolicy::REPORT_ONLY_MODE or ContentSecurityPolicy::FULL_MODE depending on type of header 'ContentSecurityPolicyDirectives':Modify the content security policy directives. Use this only if ContentSecurityPolicyDefaultSource and ContentSecurityPolicyScriptSource do not meet your needs. & $directives:Array of CSP directives $policyConfig:Current configuration for the CSP header $mode:ContentSecurityPolicy::REPORT_ONLY_MODE or ContentSecurityPolicy::FULL_MODE depending on type of header 'ContentSecurityPolicyScriptSource':Modify the allowed CSP script sources. Note that you also have to use ContentSecurityPolicyDefaultSource if you want non-script sources to be loaded from whatever you add. & $scriptSrc:Array of CSP directives $policyConfig:Current configuration for the CSP header $mode:ContentSecurityPolicy::REPORT_ONLY_MODE or ContentSecurityPolicy::FULL_MODE depending on type of header 'CustomEditor':When invoking the page editor Return true to allow the normal editor to be used, or false if implementing a custom editor, e.g. for a special namespace, etc. $article:Article being edited $user:User performing the edit 'DatabaseOraclePostInit':Called after initialising an Oracle database $db:the DatabaseOracle object 'DeletedContribsPager::reallyDoQuery':Called before really executing the query for Special:DeletedContributions Similar to ContribsPager::reallyDoQuery & $data:an array of results of all contribs queries $pager:The DeletedContribsPager object hooked into $offset:Index offset, inclusive $limit:Exact query limit $descending:Query direction, false for ascending, true for descending 'DeletedContributionsLineEnding':Called before a DeletedContributions HTML line is finished. Similar to ContributionsLineEnding $page:SpecialPage object for DeletedContributions & $ret:the HTML line $row:the DB row for this line & $classes:the classes to add to the surrounding< li > & $attribs:associative array of other HTML attributes for the< li > element. Currently only data attributes reserved to MediaWiki are allowed(see Sanitizer::isReservedDataAttribute). 'DeleteUnknownPreferences':Called by the cleanupPreferences.php maintenance script to build a WHERE clause with which to delete preferences that are not known about. This hook is used by extensions that have dynamically-named preferences that should not be deleted in the usual cleanup process. For example, the Gadgets extension creates preferences prefixed with 'gadget-', and so anything with that prefix is excluded from the deletion. &where:An array that will be passed as the $cond parameter to IDatabase::select() to determine what will be deleted from the user_properties table. $db:The IDatabase object, useful for accessing $db->buildLike() etc. 'DifferenceEngineAfterLoadNewText':called in DifferenceEngine::loadNewText() after the new revision 's content has been loaded into the class member variable $differenceEngine->mNewContent but before returning true from this function. $differenceEngine:DifferenceEngine object 'DifferenceEngineLoadTextAfterNewContentIsLoaded':called in DifferenceEngine::loadText() after the new revision 's content has been loaded into the class member variable $differenceEngine->mNewContent but before checking if the variable 's value is null. This hook can be used to inject content into said class member variable. $differenceEngine:DifferenceEngine object 'DifferenceEngineMarkPatrolledLink':Allows extensions to change the "mark as patrolled" link which is shown both on the diff header as well as on the bottom of a page, usually wrapped in a span element which has class="patrollink". $differenceEngine:DifferenceEngine object & $markAsPatrolledLink:The "mark as patrolled" link HTML(string) $rcid:Recent change ID(rc_id) for this change(int) 'DifferenceEngineMarkPatrolledRCID':Allows extensions to possibly change the rcid parameter. For example the rcid might be set to zero due to the user being the same as the performer of the change but an extension might still want to show it under certain conditions. & $rcid:rc_id(int) of the change or 0 $differenceEngine:DifferenceEngine object $change:RecentChange object $user:User object representing the current user 'DifferenceEngineNewHeader':Allows extensions to change the $newHeader variable, which contains information about the new revision, such as the revision 's author, whether the revision was marked as a minor edit or not, etc. $differenceEngine:DifferenceEngine object & $newHeader:The string containing the various #mw-diff-otitle[1-5] divs, which include things like revision author info, revision comment, RevisionDelete link and more $formattedRevisionTools:Array containing revision tools, some of which may have been injected with the DiffRevisionTools hook $nextlink:String containing the link to the next revision(if any) $status
Definition hooks.txt:1071
this hook is for auditing only $response
Definition hooks.txt:813
processing should stop and the error should be shown to the user * false
Definition hooks.txt:187
returning false will NOT prevent logging $e
Definition hooks.txt:2226
Interface for configuration instances.
Definition Config.php:28
The wiki should then use memcached to cache various data To use multiple just add more items to the array To increase the weight of a make its entry a array("192.168.0.1:11211", 2))
$content
if(count( $args)< 1) $job