76 parent::__construct( $params );
77 $section = isset( $params[
'sectionsByWiki'][$this->
wiki] )
80 if ( !isset( $params[
'partitionsBySection'][
$section] ) ) {
81 throw new MWException(
"No configuration for section '$section'." );
83 $this->maxPartitionsTry = isset( $params[
'maxPartitionsTry'] )
84 ? $params[
'maxPartitionsTry']
87 $partitionMap = $params[
'partitionsBySection'][
$section];
88 arsort( $partitionMap, SORT_NUMERIC );
91 foreach ( [
'class',
'sectionsByWiki',
'maxPartitionsTry',
92 'partitionsBySection',
'configByPartition', ]
as $o
94 unset( $baseConfig[$o] );
97 unset( $baseConfig[
'aggregator'] );
99 foreach ( $partitionMap
as $partition => $w ) {
100 if ( !isset( $params[
'configByPartition'][$partition] ) ) {
101 throw new MWException(
"No configuration for partition '$partition'." );
104 $baseConfig + $params[
'configByPartition'][$partition] );
107 $this->partitionRing =
new HashRing( $partitionMap );
112 return [
'undefined',
'random',
'timestamp' ];
120 foreach ( $this->partitionQueues
as $queue ) {
121 if ( !$queue->supportsDelayedJobs() ) {
132 foreach ( $this->partitionQueues
as $queue ) {
134 $empty = $empty && $queue->doIsEmpty();
169 foreach ( $this->partitionQueues
as $queue ) {
171 $count += $queue->$method();
189 for ( $i = $this->maxPartitionsTry; $i > 0 && count( $jobsLeft ); --$i ) {
198 if ( count( $jobsLeft ) ) {
200 "Could not insert job(s), {$this->maxPartitionsTry} partitions tried." );
217 $uJobsByPartition = [];
219 foreach ( $jobs
as $key =>
$job ) {
220 if (
$job->ignoreDuplicates() ) {
223 unset( $jobs[$key] );
227 if (
$flags & self::QOS_ATOMIC ) {
228 $nuJobBatches = [ $jobs ];
233 $nuJobBatches = array_chunk( $jobs, 300 );
237 foreach ( $uJobsByPartition
as $partition => $jobBatch ) {
239 $queue = $this->partitionQueues[$partition];
242 $queue->doBatchPush( $jobBatch,
$flags | self::QOS_ATOMIC );
249 throw new JobQueueError(
"Could not insert job(s), no partitions available." );
251 $jobsLeft = array_merge( $jobsLeft, $jobBatch );
256 foreach ( $nuJobBatches
as $jobBatch ) {
258 $queue = $this->partitionQueues[$partition];
261 $queue->doBatchPush( $jobBatch,
$flags | self::QOS_ATOMIC );
268 throw new JobQueueError(
"Could not insert job(s), no partitions available." );
270 $jobsLeft = array_merge( $jobsLeft, $jobBatch );
278 $partitionsTry = $this->partitionRing->getLiveLocationWeights();
281 while ( count( $partitionsTry ) ) {
283 if ( $partition ===
false ) {
288 $queue = $this->partitionQueues[$partition];
297 $job->metadata[
'QueuePartition'] = $partition;
301 unset( $partitionsTry[$partition] );
310 if ( !isset( $job->metadata[
'QueuePartition'] ) ) {
311 throw new MWException(
"The given job has no defined partition name." );
314 $this->partitionQueues[$job->metadata[
'QueuePartition']]->ack( $job );
319 $partition = $this->partitionRing->getLiveLocation( $signature );
321 return $this->partitionQueues[$partition]->doIsRootJobOldDuplicate( $job );
323 if ( $this->partitionRing->ejectFromLiveRing( $partition, 5 ) ) {
324 $partition = $this->partitionRing->getLiveLocation( $signature );
325 return $this->partitionQueues[$partition]->doIsRootJobOldDuplicate( $job );
334 $partition = $this->partitionRing->getLiveLocation( $signature );
336 return $this->partitionQueues[$partition]->doDeduplicateRootJob( $job );
338 if ( $this->partitionRing->ejectFromLiveRing( $partition, 5 ) ) {
339 $partition = $this->partitionRing->getLiveLocation( $signature );
340 return $this->partitionQueues[$partition]->doDeduplicateRootJob( $job );
350 foreach ( $this->partitionQueues
as $queue ) {
365 foreach ( $this->partitionQueues
as $queue ) {
367 $queue->waitForBackups();
378 foreach ( $this->partitionQueues
as $queue ) {
379 $queue->doFlushCaches();
384 $iterator =
new AppendIterator();
387 foreach ( $this->partitionQueues
as $queue ) {
388 $iterator->append( $queue->getAllQueuedJobs() );
395 $iterator =
new AppendIterator();
398 foreach ( $this->partitionQueues
as $queue ) {
399 $iterator->append( $queue->getAllDelayedJobs() );
406 $iterator =
new AppendIterator();
409 foreach ( $this->partitionQueues
as $queue ) {
410 $iterator->append( $queue->getAllAcquiredJobs() );
417 $iterator =
new AppendIterator();
420 foreach ( $this->partitionQueues
as $queue ) {
421 $iterator->append( $queue->getAllAbandonedJobs() );
428 return "JobQueueFederated:wiki:{$this->wiki}" .
429 sha1(
serialize( array_keys( $this->partitionQueues ) ) );
437 foreach ( $this->partitionQueues
as $queue ) {
439 $nonEmpty = $queue->doGetSiblingQueuesWithJobs( $types );
440 if ( is_array( $nonEmpty ) ) {
445 if ( count(
$result ) == count( $types ) ) {
455 return array_values(
$result );
462 foreach ( $this->partitionQueues
as $queue ) {
464 $sizes = $queue->doGetSiblingQueueSizes( $types );
465 if ( is_array( $sizes ) ) {
466 foreach ( $sizes
as $type => $size ) {
483 wfDebugLog(
'JobQueueFederated', $e->getMessage() .
"\n" . $e->getTraceAsString() );
494 if ( $down >= count( $this->partitionQueues ) ) {
the array() calling protocol came about after MediaWiki 1.4rc1.
getCoalesceLocationInternal()
Class to handle enqueueing and running of background jobs for federated queues.
div flags Integer display flags(NO_ACTION_LINK, NO_EXTRA_USER_LINKS) 'LogException'returning false will NOT prevent logging $e
Class to both describe a background job and handle jobs.
it s the revision text itself In either if gzip is the revision text is gzipped $flags
getLiveLocationWeights()
Get the map of "live" locations to weight (ignores 0-weight items)
The index of the header message $result[1]=The index of the body text message $result[2 through n]=Parameters passed to body text message.Please note the header message cannot receive/use parameters. 'ImportHandleLogItemXMLTag':When parsing a XML tag in a log item.Return false to stop further processing of the tag $reader:XMLReader object $logInfo:Array of information 'ImportHandlePageXMLTag':When parsing a XML tag in a page.Return false to stop further processing of the tag $reader:XMLReader object &$pageInfo:Array of information 'ImportHandleRevisionXMLTag':When parsing a XML tag in a page revision.Return false to stop further processing of the tag $reader:XMLReader object $pageInfo:Array of page information $revisionInfo:Array of revision information 'ImportHandleToplevelXMLTag':When parsing a top level XML tag.Return false to stop further processing of the tag $reader:XMLReader object 'ImportHandleUploadXMLTag':When parsing a XML tag in a file upload.Return false to stop further processing of the tag $reader:XMLReader object $revisionInfo:Array of information 'ImportLogInterwikiLink':Hook to change the interwiki link used in log entries and edit summaries for transwiki imports.&$fullInterwikiPrefix:Interwiki prefix, may contain colons.&$pageTitle:String that contains page title. 'ImportSources':Called when reading from the $wgImportSources configuration variable.Can be used to lazy-load the import sources list.&$importSources:The value of $wgImportSources.Modify as necessary.See the comment in DefaultSettings.php for the detail of how to structure this array. 'InfoAction':When building information to display on the action=info page.$context:IContextSource object &$pageInfo:Array of information 'InitializeArticleMaybeRedirect':MediaWiki check to see if title is a redirect.&$title:Title object for the current page &$request:WebRequest &$ignoreRedirect:boolean to skip redirect check &$target:Title/string of redirect target &$article:Article object 'InternalParseBeforeLinks':during Parser's internalParse method before links but after nowiki/noinclude/includeonly/onlyinclude and other processings.&$parser:Parser object &$text:string containing partially parsed text &$stripState:Parser's internal StripState object 'InternalParseBeforeSanitize':during Parser's internalParse method just before the parser removes unwanted/dangerous HTML tags and after nowiki/noinclude/includeonly/onlyinclude and other processings.Ideal for syntax-extensions after template/parser function execution which respect nowiki and HTML-comments.&$parser:Parser object &$text:string containing partially parsed text &$stripState:Parser's internal StripState object 'InterwikiLoadPrefix':When resolving if a given prefix is an interwiki or not.Return true without providing an interwiki to continue interwiki search.$prefix:interwiki prefix we are looking for.&$iwData:output array describing the interwiki with keys iw_url, iw_local, iw_trans and optionally iw_api and iw_wikiid. 'InvalidateEmailComplete':Called after a user's email has been invalidated successfully.$user:user(object) whose email is being invalidated 'IRCLineURL':When constructing the URL to use in an IRC notification.Callee may modify $url and $query, URL will be constructed as $url.$query &$url:URL to index.php &$query:Query string $rc:RecentChange object that triggered url generation 'IsFileCacheable':Override the result of Article::isFileCacheable()(if true) &$article:article(object) being checked 'IsTrustedProxy':Override the result of IP::isTrustedProxy() &$ip:IP being check &$result:Change this value to override the result of IP::isTrustedProxy() 'IsUploadAllowedFromUrl':Override the result of UploadFromUrl::isAllowedUrl() $url:URL used to upload from &$allowed:Boolean indicating if uploading is allowed for given URL 'isValidEmailAddr':Override the result of Sanitizer::validateEmail(), for instance to return false if the domain name doesn't match your organization.$addr:The e-mail address entered by the user &$result:Set this and return false to override the internal checks 'isValidPassword':Override the result of User::isValidPassword() $password:The password entered by the user &$result:Set this and return false to override the internal checks $user:User the password is being validated for 'Language::getMessagesFileName':$code:The language code or the language we're looking for a messages file for &$file:The messages file path, you can override this to change the location. 'LanguageGetMagic':DEPRECATED!Use $magicWords in a file listed in $wgExtensionMessagesFiles instead.Use this to define synonyms of magic words depending of the language &$magicExtensions:associative array of magic words synonyms $lang:language code(string) 'LanguageGetNamespaces':Provide custom ordering for namespaces or remove namespaces.Do not use this hook to add namespaces.Use CanonicalNamespaces for that.&$namespaces:Array of namespaces indexed by their numbers 'LanguageGetSpecialPageAliases':DEPRECATED!Use $specialPageAliases in a file listed in $wgExtensionMessagesFiles instead.Use to define aliases of special pages names depending of the language &$specialPageAliases:associative array of magic words synonyms $lang:language code(string) 'LanguageGetTranslatedLanguageNames':Provide translated language names.&$names:array of language code=> language name $code:language of the preferred translations 'LanguageLinks':Manipulate a page's language links.This is called in various places to allow extensions to define the effective language links for a page.$title:The page's Title.&$links:Associative array mapping language codes to prefixed links of the form"language:title".&$linkFlags:Associative array mapping prefixed links to arrays of flags.Currently unused, but planned to provide support for marking individual language links in the UI, e.g.for featured articles. 'LanguageSelector':Hook to change the language selector available on a page.$out:The output page.$cssClassName:CSS class name of the language selector. 'LinkBegin':DEPRECATED!Use HtmlPageLinkRendererBegin instead.Used when generating internal and interwiki links in Linker::link(), before processing starts.Return false to skip default processing and return $ret.See documentation for Linker::link() for details on the expected meanings of parameters.$skin:the Skin object $target:the Title that the link is pointing to &$html:the contents that the< a > tag should have(raw HTML) $result
Prior to maintenance scripts were a hodgepodge of code that had no cohesion or formal method of action Beginning maintenance scripts have been cleaned up to use a unified class Directory structure How to run a script How to write your own DIRECTORY STRUCTURE The maintenance directory of a MediaWiki installation contains several all of which have unique purposes HOW TO RUN A SCRIPT Ridiculously just call php someScript php that s in the top level maintenance directory if not default wiki
Convenience class for weighted consistent hash rings.
wfDebugLog($logGroup, $text, $dest= 'all', array $context=[])
Send a line to a supplementary debug log file, if configured, or main debug log if not...
doIsRootJobOldDuplicate(Job $job)
ejectFromLiveRing($location, $ttl)
Remove a location from the "live" hash ring.
getLiveRing()
Get the "live" hash ring (which does not include ejected locations)
logException(Exception $e)
doGetSiblingQueuesWithJobs(array $types)
tryJobInsertions(array $jobs, HashRing &$partitionRing, $flags)
__construct(array $params)
This document is intended to provide useful advice for parties seeking to redistribute MediaWiki to end users It s targeted particularly at maintainers for Linux since it s been observed that distribution packages of MediaWiki often break We ve consistently had to recommend that users seeking support use official tarballs instead of their distribution s and this often solves whatever problem the user is having It would be nice if this could such as
usually copyright or history_copyright This message must be in HTML not wikitext if the section is included from a template $section
JobQueue[] $partitionQueues
(partition name => JobQueue) reverse sorted by weight
int $maxPartitionsTry
Maximum number of partitions to try.
injection txt This is an overview of how MediaWiki makes use of dependency injection The design described here grew from the discussion of RFC T384 The term dependency this means that anything an object needs to operate should be injected from the the object itself should only know narrow no concrete implementation of the logic it relies on The requirement to inject everything typically results in an architecture that based on two main types of and essentially stateless service objects that use other service objects to operate on the value objects As of the beginning MediaWiki is only starting to use the DI approach Much of the code still relies on global state or direct resulting in a highly cyclical dependency which acts as the top level factory for services in MediaWiki which can be used to gain access to default instances of various services MediaWikiServices however also allows new services to be defined and default services to be redefined Services are defined or redefined by providing a callback the instantiator that will return a new instance of the service When it will create an instance of MediaWikiServices and populate it with the services defined in the files listed by thereby bootstrapping the DI framework Per $wgServiceWiringFiles lists includes ServiceWiring php
static factory(array $params)
Get a job queue object of the specified type.
Class to handle enqueueing and running of background jobs.
doBatchPush(array $jobs, $flags)
doDeduplicateRootJob(IJobSpecification $job)
Job queue task description interface.
static pickRandom($weights)
Given an array of non-normalised probabilities, this function will select an element and return the a...
doGetSiblingQueueSizes(array $types)
getCrossPartitionSum($type, $method)
throwErrorIfAllPartitionsDown($down)
Throw an error if no partitions available.
getLiveLocation($item)
Get the location of an item on the "live" ring.