MediaWiki  1.27.2
KafkaHandler.php
Go to the documentation of this file.
1 <?php
21 namespace MediaWiki\Logger\Monolog;
22 
29 
48 class KafkaHandler extends AbstractProcessingHandler {
52  protected $produce;
53 
57  protected $options;
58 
62  protected $partitions = [];
63 
67  private static $defaultOptions = [
68  'alias' => [], // map from monolog channel to kafka topic
69  'swallowExceptions' => false, // swallow exceptions sending records
70  'logExceptions' => null, // A PSR3 logger to inform about errors
71  ];
72 
79  public function __construct(
80  Produce $produce, array $options, $level = Logger::DEBUG, $bubble = true
81  ) {
82  parent::__construct( $level, $bubble );
83  $this->produce = $produce;
84  $this->options = array_merge( self::$defaultOptions, $options );
85  }
86 
97  public static function factory(
98  $kafkaServers, array $options = [], $level = Logger::DEBUG, $bubble = true
99  ) {
100  $metadata = new MetaDataFromKafka( $kafkaServers );
101  $produce = new Produce( $metadata );
102 
103  if ( isset( $options['sendTimeout'] ) ) {
104  $timeOut = $options['sendTimeout'];
105  $produce->getClient()->setStreamOption( 'SendTimeoutSec', 0 );
106  $produce->getClient()->setStreamOption( 'SendTimeoutUSec',
107  intval( $timeOut * 1000000 )
108  );
109  }
110  if ( isset( $options['recvTimeout'] ) ) {
111  $timeOut = $options['recvTimeout'];
112  $produce->getClient()->setStreamOption( 'RecvTimeoutSec', 0 );
113  $produce->getClient()->setStreamOption( 'RecvTimeoutUSec',
114  intval( $timeOut * 1000000 )
115  );
116  }
117  if ( isset( $options['logExceptions'] ) && is_string( $options['logExceptions'] ) ) {
118  $options['logExceptions'] = LoggerFactory::getInstance( $options['logExceptions'] );
119  }
120 
121  return new self( $produce, $options, $level, $bubble );
122  }
123 
127  protected function write( array $record ) {
128  if ( $record['formatted'] !== null ) {
129  $this->addMessages( $record['channel'], [ $record['formatted'] ] );
130  $this->send();
131  }
132  }
133 
137  public function handleBatch( array $batch ) {
138  $channels = [];
139  foreach ( $batch as $record ) {
140  if ( $record['level'] < $this->level ) {
141  continue;
142  }
143  $channels[$record['channel']][] = $this->processRecord( $record );
144  }
145 
146  $formatter = $this->getFormatter();
147  foreach ( $channels as $channel => $records ) {
148  $messages = [];
149  foreach ( $records as $idx => $record ) {
150  $message = $formatter->format( $record );
151  if ( $message !== null ) {
152  $messages[] = $message;
153  }
154  }
155  if ( $messages ) {
156  $this->addMessages( $channel, $messages );
157  }
158  }
159 
160  $this->send();
161  }
162 
166  protected function send() {
167  try {
168  $this->produce->send();
169  } catch ( \Kafka\Exception $e ) {
170  $ignore = $this->warning(
171  'Error sending records to kafka: {exception}',
172  [ 'exception' => $e ] );
173  if ( !$ignore ) {
174  throw $e;
175  }
176  }
177  }
178 
184  protected function getRandomPartition( $topic ) {
185  if ( !array_key_exists( $topic, $this->partitions ) ) {
186  try {
187  $partitions = $this->produce->getAvailablePartitions( $topic );
188  } catch ( \Kafka\Exception $e ) {
189  $ignore = $this->warning(
190  'Error getting metadata for kafka topic {topic}: {exception}',
191  [ 'topic' => $topic, 'exception' => $e ] );
192  if ( $ignore ) {
193  return null;
194  }
195  throw $e;
196  }
197  if ( $partitions ) {
198  $key = array_rand( $partitions );
199  $this->partitions[$topic] = $partitions[$key];
200  } else {
201  $details = $this->produce->getClient()->getTopicDetail( $topic );
202  $ignore = $this->warning(
203  'No partitions available for kafka topic {topic}',
204  [ 'topic' => $topic, 'kafka' => $details ]
205  );
206  if ( !$ignore ) {
207  throw new \RuntimeException( "No partitions available for kafka topic $topic" );
208  }
209  $this->partitions[$topic] = null;
210  }
211  }
212  return $this->partitions[$topic];
213  }
214 
221  protected function addMessages( $channel, array $records ) {
222  if ( isset( $this->options['alias'][$channel] ) ) {
223  $topic = $this->options['alias'][$channel];
224  } else {
225  $topic = "monolog_$channel";
226  }
227  $partition = $this->getRandomPartition( $topic );
228  if ( $partition !== null ) {
229  $this->produce->setMessages( $topic, $partition, $records );
230  }
231  }
232 
238  protected function warning( $message, array $context = [] ) {
239  if ( $this->options['logExceptions'] instanceof LoggerInterface ) {
240  $this->options['logExceptions']->warning( $message, $context );
241  }
242  return $this->options['swallowExceptions'];
243  }
244 }
the array() calling protocol came about after MediaWiki 1.4rc1.
magic word the default is to use $key to get the and $key value or $key value text $key value html to format the value $key
Definition: hooks.txt:2321
static array __construct(Produce $produce, array $options, $level=Logger::DEBUG, $bubble=true)
Apache License January AND DISTRIBUTION Definitions License shall mean the terms and conditions for use
div flags Integer display flags(NO_ACTION_LINK, NO_EXTRA_USER_LINKS) 'LogException'returning false will NOT prevent logging $e
Definition: hooks.txt:1932
array $options
Optional handler configuration.
static getInstance($channel)
Get a named logger instance from the currently configured logger factory.
static array $defaultOptions
defaults for constructor options
array $partitions
Map from topic name to partition this request produces to.
We ve cleaned up the code here by removing clumps of infrequently used code and moving them off somewhere else It s much easier for someone working with this code to see what s _really_ going and make changes or fix bugs In we can take all the code that deals with the little used title reversing options(say) and put it in one place.Instead of having little title-reversing if-blocks spread all over the codebase in showAnArticle
$batch
Definition: linkcache.txt:23
send()
Send any records in the kafka client internal queue.
IContextSource $context
Definition: MediaWiki.php:32
static factory($kafkaServers, array $options=[], $level=Logger::DEBUG, $bubble=true)
Constructs the necessary support objects and returns a KafkaHandler instance.
This document is intended to provide useful advice for parties seeking to redistribute MediaWiki to end users It s targeted particularly at maintainers for Linux since it s been observed that distribution packages of MediaWiki often break We ve consistently had to recommend that users seeking support use official tarballs instead of their distribution s and this often solves whatever problem the user is having It would be nice if this could such as
Definition: distributors.txt:9
injection txt This is an overview of how MediaWiki makes use of dependency injection The design described here grew from the discussion of RFC T384 The term dependency this means that anything an object needs to operate should be injected from the the object itself should only know narrow no concrete implementation of the logic it relies on The requirement to inject everything typically results in an architecture that based on two main types of and essentially stateless service objects that use other service objects to operate on the value objects As of the beginning MediaWiki is only starting to use the DI approach Much of the code still relies on global state or direct resulting in a highly cyclical dependency which acts as the top level factory for services in MediaWiki which can be used to gain access to default instances of various services MediaWikiServices however also allows new services to be defined and default services to be redefined Services are defined or redefined by providing a callback the instantiator that will return a new instance of the service When it will create an instance of MediaWikiServices and populate it with the services defined in the files listed by thereby bootstrapping the DI framework Per $wgServiceWiringFiles lists includes ServiceWiring php
Definition: injection.txt:35
Log handler sends log events to a kafka server.
MediaWiki Logger LoggerFactory implements a PSR[0] compatible message logging system Named Psr Log LoggerInterface instances can be obtained from the MediaWiki Logger LoggerFactory::getInstance() static method.MediaWiki\Logger\LoggerFactory expects a class implementing the MediaWiki\Logger\Spi interface to act as a factory for new Psr\Log\LoggerInterface instances.The"Spi"in MediaWiki\Logger\Spi stands for"service provider interface".An SPI is an API intended to be implemented or extended by a third party.This software design pattern is intended to enable framework extension and replaceable components.It is specifically used in the MediaWiki\Logger\LoggerFactory service to allow alternate PSR-3 logging implementations to be easily integrated with MediaWiki.The service provider interface allows the backend logging library to be implemented in multiple ways.The $wgMWLoggerDefaultSpi global provides the classname of the default MediaWiki\Logger\Spi implementation to be loaded at runtime.This can either be the name of a class implementing the MediaWiki\Logger\Spi with a zero argument const ructor or a callable that will return an MediaWiki\Logger\Spi instance.Alternately the MediaWiki\Logger\LoggerFactory MediaWiki Logger LoggerFactory
Definition: logger.txt:5
$messages
Produce $produce
Sends requests to kafka.
warning($message, array $context=[])
addMessages($channel, array $records)
Adds records for a channel to the Kafka client internal queue.