Remove HWLDFWordAccumulator, deprecated in 1.28
[lhc/web/wiklou.git] / includes / MediaWiki.php
1 <?php
2 /**
3 * Helper class for the index.php entry point.
4 *
5 * This program is free software; you can redistribute it and/or modify
6 * it under the terms of the GNU General Public License as published by
7 * the Free Software Foundation; either version 2 of the License, or
8 * (at your option) any later version.
9 *
10 * This program is distributed in the hope that it will be useful,
11 * but WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 * GNU General Public License for more details.
14 *
15 * You should have received a copy of the GNU General Public License along
16 * with this program; if not, write to the Free Software Foundation, Inc.,
17 * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
18 * http://www.gnu.org/copyleft/gpl.html
19 *
20 * @file
21 */
22
23 use MediaWiki\Logger\LoggerFactory;
24 use Psr\Log\LoggerInterface;
25 use MediaWiki\MediaWikiServices;
26 use Wikimedia\Rdbms\ChronologyProtector;
27 use Wikimedia\Rdbms\LBFactory;
28 use Wikimedia\Rdbms\DBConnectionError;
29 use Liuggio\StatsdClient\Sender\SocketSender;
30
31 /**
32 * The MediaWiki class is the helper class for the index.php entry point.
33 */
34 class MediaWiki {
35 /**
36 * @var IContextSource
37 */
38 private $context;
39
40 /**
41 * @var Config
42 */
43 private $config;
44
45 /**
46 * @var string Cache what action this request is
47 */
48 private $action;
49
50 /**
51 * @param IContextSource|null $context
52 */
53 public function __construct( IContextSource $context = null ) {
54 if ( !$context ) {
55 $context = RequestContext::getMain();
56 }
57
58 $this->context = $context;
59 $this->config = $context->getConfig();
60 }
61
62 /**
63 * Parse the request to get the Title object
64 *
65 * @throws MalformedTitleException If a title has been provided by the user, but is invalid.
66 * @return Title Title object to be $wgTitle
67 */
68 private function parseTitle() {
69 $request = $this->context->getRequest();
70 $curid = $request->getInt( 'curid' );
71 $title = $request->getVal( 'title' );
72 $action = $request->getVal( 'action' );
73
74 if ( $request->getCheck( 'search' ) ) {
75 // Compatibility with old search URLs which didn't use Special:Search
76 // Just check for presence here, so blank requests still
77 // show the search page when using ugly URLs (T10054).
78 $ret = SpecialPage::getTitleFor( 'Search' );
79 } elseif ( $curid ) {
80 // URLs like this are generated by RC, because rc_title isn't always accurate
81 $ret = Title::newFromID( $curid );
82 } else {
83 $ret = Title::newFromURL( $title );
84 // Alias NS_MEDIA page URLs to NS_FILE...we only use NS_MEDIA
85 // in wikitext links to tell Parser to make a direct file link
86 if ( !is_null( $ret ) && $ret->getNamespace() == NS_MEDIA ) {
87 $ret = Title::makeTitle( NS_FILE, $ret->getDBkey() );
88 }
89 $contLang = MediaWikiServices::getInstance()->getContentLanguage();
90 // Check variant links so that interwiki links don't have to worry
91 // about the possible different language variants
92 if (
93 $contLang->hasVariants() && !is_null( $ret ) && $ret->getArticleID() == 0
94 ) {
95 $contLang->findVariantLink( $title, $ret );
96 }
97 }
98
99 // If title is not provided, always allow oldid and diff to set the title.
100 // If title is provided, allow oldid and diff to override the title, unless
101 // we are talking about a special page which might use these parameters for
102 // other purposes.
103 if ( $ret === null || !$ret->isSpecialPage() ) {
104 // We can have urls with just ?diff=,?oldid= or even just ?diff=
105 $oldid = $request->getInt( 'oldid' );
106 $oldid = $oldid ?: $request->getInt( 'diff' );
107 // Allow oldid to override a changed or missing title
108 if ( $oldid ) {
109 $rev = Revision::newFromId( $oldid );
110 $ret = $rev ? $rev->getTitle() : $ret;
111 }
112 }
113
114 // Use the main page as default title if nothing else has been provided
115 if ( $ret === null
116 && strval( $title ) === ''
117 && !$request->getCheck( 'curid' )
118 && $action !== 'delete'
119 ) {
120 $ret = Title::newMainPage();
121 }
122
123 if ( $ret === null || ( $ret->getDBkey() == '' && !$ret->isExternal() ) ) {
124 // If we get here, we definitely don't have a valid title; throw an exception.
125 // Try to get detailed invalid title exception first, fall back to MalformedTitleException.
126 Title::newFromTextThrow( $title );
127 throw new MalformedTitleException( 'badtitletext', $title );
128 }
129
130 return $ret;
131 }
132
133 /**
134 * Get the Title object that we'll be acting on, as specified in the WebRequest
135 * @return Title
136 */
137 public function getTitle() {
138 if ( !$this->context->hasTitle() ) {
139 try {
140 $this->context->setTitle( $this->parseTitle() );
141 } catch ( MalformedTitleException $ex ) {
142 $this->context->setTitle( SpecialPage::getTitleFor( 'Badtitle' ) );
143 }
144 }
145 return $this->context->getTitle();
146 }
147
148 /**
149 * Returns the name of the action that will be executed.
150 *
151 * @return string Action
152 */
153 public function getAction() {
154 if ( $this->action === null ) {
155 $this->action = Action::getActionName( $this->context );
156 }
157
158 return $this->action;
159 }
160
161 /**
162 * Performs the request.
163 * - bad titles
164 * - read restriction
165 * - local interwiki redirects
166 * - redirect loop
167 * - special pages
168 * - normal pages
169 *
170 * @throws MWException|PermissionsError|BadTitleError|HttpError
171 * @return void
172 */
173 private function performRequest() {
174 global $wgTitle;
175
176 $request = $this->context->getRequest();
177 $requestTitle = $title = $this->context->getTitle();
178 $output = $this->context->getOutput();
179 $user = $this->context->getUser();
180
181 if ( $request->getVal( 'printable' ) === 'yes' ) {
182 $output->setPrintable();
183 }
184
185 $unused = null; // To pass it by reference
186 Hooks::run( 'BeforeInitialize', [ &$title, &$unused, &$output, &$user, $request, $this ] );
187
188 // Invalid titles. T23776: The interwikis must redirect even if the page name is empty.
189 if ( is_null( $title ) || ( $title->getDBkey() == '' && !$title->isExternal() )
190 || $title->isSpecial( 'Badtitle' )
191 ) {
192 $this->context->setTitle( SpecialPage::getTitleFor( 'Badtitle' ) );
193 try {
194 $this->parseTitle();
195 } catch ( MalformedTitleException $ex ) {
196 throw new BadTitleError( $ex );
197 }
198 throw new BadTitleError();
199 }
200
201 // Check user's permissions to read this page.
202 // We have to check here to catch special pages etc.
203 // We will check again in Article::view().
204 $permErrors = $title->isSpecial( 'RunJobs' )
205 ? [] // relies on HMAC key signature alone
206 : $title->getUserPermissionsErrors( 'read', $user );
207 if ( count( $permErrors ) ) {
208 // T34276: allowing the skin to generate output with $wgTitle or
209 // $this->context->title set to the input title would allow anonymous users to
210 // determine whether a page exists, potentially leaking private data. In fact, the
211 // curid and oldid request parameters would allow page titles to be enumerated even
212 // when they are not guessable. So we reset the title to Special:Badtitle before the
213 // permissions error is displayed.
214
215 // The skin mostly uses $this->context->getTitle() these days, but some extensions
216 // still use $wgTitle.
217 $badTitle = SpecialPage::getTitleFor( 'Badtitle' );
218 $this->context->setTitle( $badTitle );
219 $wgTitle = $badTitle;
220
221 throw new PermissionsError( 'read', $permErrors );
222 }
223
224 // Interwiki redirects
225 if ( $title->isExternal() ) {
226 $rdfrom = $request->getVal( 'rdfrom' );
227 if ( $rdfrom ) {
228 $url = $title->getFullURL( [ 'rdfrom' => $rdfrom ] );
229 } else {
230 $query = $request->getValues();
231 unset( $query['title'] );
232 $url = $title->getFullURL( $query );
233 }
234 // Check for a redirect loop
235 if ( !preg_match( '/^' . preg_quote( $this->config->get( 'Server' ), '/' ) . '/', $url )
236 && $title->isLocal()
237 ) {
238 // 301 so google et al report the target as the actual url.
239 $output->redirect( $url, 301 );
240 } else {
241 $this->context->setTitle( SpecialPage::getTitleFor( 'Badtitle' ) );
242 try {
243 $this->parseTitle();
244 } catch ( MalformedTitleException $ex ) {
245 throw new BadTitleError( $ex );
246 }
247 throw new BadTitleError();
248 }
249 // Handle any other redirects.
250 // Redirect loops, titleless URL, $wgUsePathInfo URLs, and URLs with a variant
251 } elseif ( !$this->tryNormaliseRedirect( $title ) ) {
252 // Prevent information leak via Special:MyPage et al (T109724)
253 $spFactory = MediaWikiServices::getInstance()->getSpecialPageFactory();
254 if ( $title->isSpecialPage() ) {
255 $specialPage = $spFactory->getPage( $title->getDBkey() );
256 if ( $specialPage instanceof RedirectSpecialPage ) {
257 $specialPage->setContext( $this->context );
258 if ( $this->config->get( 'HideIdentifiableRedirects' )
259 && $specialPage->personallyIdentifiableTarget()
260 ) {
261 list( , $subpage ) = $spFactory->resolveAlias( $title->getDBkey() );
262 $target = $specialPage->getRedirect( $subpage );
263 // target can also be true. We let that case fall through to normal processing.
264 if ( $target instanceof Title ) {
265 $query = $specialPage->getRedirectQuery() ?: [];
266 $request = new DerivativeRequest( $this->context->getRequest(), $query );
267 $request->setRequestURL( $this->context->getRequest()->getRequestURL() );
268 $this->context->setRequest( $request );
269 // Do not varnish cache these. May vary even for anons
270 $this->context->getOutput()->lowerCdnMaxage( 0 );
271 $this->context->setTitle( $target );
272 $wgTitle = $target;
273 // Reset action type cache. (Special pages have only view)
274 $this->action = null;
275 $title = $target;
276 $output->addJsConfigVars( [
277 'wgInternalRedirectTargetUrl' => $target->getFullURL( $query ),
278 ] );
279 $output->addModules( 'mediawiki.action.view.redirect' );
280 }
281 }
282 }
283 }
284
285 // Special pages ($title may have changed since if statement above)
286 if ( $title->isSpecialPage() ) {
287 // Actions that need to be made when we have a special pages
288 $spFactory->executePath( $title, $this->context );
289 } else {
290 // ...otherwise treat it as an article view. The article
291 // may still be a wikipage redirect to another article or URL.
292 $article = $this->initializeArticle();
293 if ( is_object( $article ) ) {
294 $this->performAction( $article, $requestTitle );
295 } elseif ( is_string( $article ) ) {
296 $output->redirect( $article );
297 } else {
298 throw new MWException( "Shouldn't happen: MediaWiki::initializeArticle()"
299 . " returned neither an object nor a URL" );
300 }
301 }
302 }
303 }
304
305 /**
306 * Handle redirects for uncanonical title requests.
307 *
308 * Handles:
309 * - Redirect loops.
310 * - No title in URL.
311 * - $wgUsePathInfo URLs.
312 * - URLs with a variant.
313 * - Other non-standard URLs (as long as they have no extra query parameters).
314 *
315 * Behaviour:
316 * - Normalise title values:
317 * /wiki/Foo%20Bar -> /wiki/Foo_Bar
318 * - Normalise empty title:
319 * /wiki/ -> /wiki/Main
320 * /w/index.php?title= -> /wiki/Main
321 * - Don't redirect anything with query parameters other than 'title' or 'action=view'.
322 *
323 * @param Title $title
324 * @return bool True if a redirect was set.
325 * @throws HttpError
326 */
327 private function tryNormaliseRedirect( Title $title ) {
328 $request = $this->context->getRequest();
329 $output = $this->context->getOutput();
330
331 if ( $request->getVal( 'action', 'view' ) != 'view'
332 || $request->wasPosted()
333 || ( $request->getCheck( 'title' )
334 && $title->getPrefixedDBkey() == $request->getVal( 'title' ) )
335 || count( $request->getValueNames( [ 'action', 'title' ] ) )
336 || !Hooks::run( 'TestCanonicalRedirect', [ $request, $title, $output ] )
337 ) {
338 return false;
339 }
340
341 if ( $title->isSpecialPage() ) {
342 list( $name, $subpage ) = MediaWikiServices::getInstance()->getSpecialPageFactory()->
343 resolveAlias( $title->getDBkey() );
344 if ( $name ) {
345 $title = SpecialPage::getTitleFor( $name, $subpage );
346 }
347 }
348 // Redirect to canonical url, make it a 301 to allow caching
349 $targetUrl = wfExpandUrl( $title->getFullURL(), PROTO_CURRENT );
350 if ( $targetUrl == $request->getFullRequestURL() ) {
351 $message = "Redirect loop detected!\n\n" .
352 "This means the wiki got confused about what page was " .
353 "requested; this sometimes happens when moving a wiki " .
354 "to a new server or changing the server configuration.\n\n";
355
356 if ( $this->config->get( 'UsePathInfo' ) ) {
357 $message .= "The wiki is trying to interpret the page " .
358 "title from the URL path portion (PATH_INFO), which " .
359 "sometimes fails depending on the web server. Try " .
360 "setting \"\$wgUsePathInfo = false;\" in your " .
361 "LocalSettings.php, or check that \$wgArticlePath " .
362 "is correct.";
363 } else {
364 $message .= "Your web server was detected as possibly not " .
365 "supporting URL path components (PATH_INFO) correctly; " .
366 "check your LocalSettings.php for a customized " .
367 "\$wgArticlePath setting and/or toggle \$wgUsePathInfo " .
368 "to true.";
369 }
370 throw new HttpError( 500, $message );
371 }
372 $output->setCdnMaxage( 1200 );
373 $output->redirect( $targetUrl, '301' );
374 return true;
375 }
376
377 /**
378 * Initialize the main Article object for "standard" actions (view, etc)
379 * Create an Article object for the page, following redirects if needed.
380 *
381 * @return Article|string An Article, or a string to redirect to another URL
382 */
383 private function initializeArticle() {
384 $title = $this->context->getTitle();
385 if ( $this->context->canUseWikiPage() ) {
386 // Try to use request context wiki page, as there
387 // is already data from db saved in per process
388 // cache there from this->getAction() call.
389 $page = $this->context->getWikiPage();
390 } else {
391 // This case should not happen, but just in case.
392 // @TODO: remove this or use an exception
393 $page = WikiPage::factory( $title );
394 $this->context->setWikiPage( $page );
395 wfWarn( "RequestContext::canUseWikiPage() returned false" );
396 }
397
398 // Make GUI wrapper for the WikiPage
399 $article = Article::newFromWikiPage( $page, $this->context );
400
401 // Skip some unnecessary code if the content model doesn't support redirects
402 if ( !ContentHandler::getForTitle( $title )->supportsRedirects() ) {
403 return $article;
404 }
405
406 $request = $this->context->getRequest();
407
408 // Namespace might change when using redirects
409 // Check for redirects ...
410 $action = $request->getVal( 'action', 'view' );
411 $file = ( $page instanceof WikiFilePage ) ? $page->getFile() : null;
412 if ( ( $action == 'view' || $action == 'render' ) // ... for actions that show content
413 && !$request->getVal( 'oldid' ) // ... and are not old revisions
414 && !$request->getVal( 'diff' ) // ... and not when showing diff
415 && $request->getVal( 'redirect' ) != 'no' // ... unless explicitly told not to
416 // ... and the article is not a non-redirect image page with associated file
417 && !( is_object( $file ) && $file->exists() && !$file->getRedirected() )
418 ) {
419 // Give extensions a change to ignore/handle redirects as needed
420 $ignoreRedirect = $target = false;
421
422 Hooks::run( 'InitializeArticleMaybeRedirect',
423 [ &$title, &$request, &$ignoreRedirect, &$target, &$article ] );
424 $page = $article->getPage(); // reflect any hook changes
425
426 // Follow redirects only for... redirects.
427 // If $target is set, then a hook wanted to redirect.
428 if ( !$ignoreRedirect && ( $target || $page->isRedirect() ) ) {
429 // Is the target already set by an extension?
430 $target = $target ?: $page->followRedirect();
431 if ( is_string( $target ) && !$this->config->get( 'DisableHardRedirects' ) ) {
432 // we'll need to redirect
433 return $target;
434 }
435 if ( is_object( $target ) ) {
436 // Rewrite environment to redirected article
437 $rpage = WikiPage::factory( $target );
438 $rpage->loadPageData();
439 if ( $rpage->exists() || ( is_object( $file ) && !$file->isLocal() ) ) {
440 $rarticle = Article::newFromWikiPage( $rpage, $this->context );
441 $rarticle->setRedirectedFrom( $title );
442
443 $article = $rarticle;
444 $this->context->setTitle( $target );
445 $this->context->setWikiPage( $article->getPage() );
446 }
447 }
448 } else {
449 // Article may have been changed by hook
450 $this->context->setTitle( $article->getTitle() );
451 $this->context->setWikiPage( $article->getPage() );
452 }
453 }
454
455 return $article;
456 }
457
458 /**
459 * Perform one of the "standard" actions
460 *
461 * @param Page $page
462 * @param Title $requestTitle The original title, before any redirects were applied
463 */
464 private function performAction( Page $page, Title $requestTitle ) {
465 $request = $this->context->getRequest();
466 $output = $this->context->getOutput();
467 $title = $this->context->getTitle();
468 $user = $this->context->getUser();
469
470 if ( !Hooks::run( 'MediaWikiPerformAction',
471 [ $output, $page, $title, $user, $request, $this ] )
472 ) {
473 return;
474 }
475
476 $act = $this->getAction();
477 $action = Action::factory( $act, $page, $this->context );
478
479 if ( $action instanceof Action ) {
480 // Narrow DB query expectations for this HTTP request
481 $trxLimits = $this->config->get( 'TrxProfilerLimits' );
482 $trxProfiler = Profiler::instance()->getTransactionProfiler();
483 if ( $request->wasPosted() && !$action->doesWrites() ) {
484 $trxProfiler->setExpectations( $trxLimits['POST-nonwrite'], __METHOD__ );
485 $request->markAsSafeRequest();
486 }
487
488 # Let CDN cache things if we can purge them.
489 if ( $this->config->get( 'UseSquid' ) &&
490 in_array(
491 // Use PROTO_INTERNAL because that's what getCdnUrls() uses
492 wfExpandUrl( $request->getRequestURL(), PROTO_INTERNAL ),
493 $requestTitle->getCdnUrls()
494 )
495 ) {
496 $output->setCdnMaxage( $this->config->get( 'SquidMaxage' ) );
497 }
498
499 $action->show();
500 return;
501 }
502
503 // If we've not found out which action it is by now, it's unknown
504 $output->setStatusCode( 404 );
505 $output->showErrorPage( 'nosuchaction', 'nosuchactiontext' );
506 }
507
508 /**
509 * Run the current MediaWiki instance; index.php just calls this
510 */
511 public function run() {
512 try {
513 $this->setDBProfilingAgent();
514 try {
515 $this->main();
516 } catch ( ErrorPageError $e ) {
517 // T64091: while exceptions are convenient to bubble up GUI errors,
518 // they are not internal application faults. As with normal requests, this
519 // should commit, print the output, do deferred updates, jobs, and profiling.
520 $this->doPreOutputCommit();
521 $e->report(); // display the GUI error
522 }
523 } catch ( Exception $e ) {
524 $context = $this->context;
525 $action = $context->getRequest()->getVal( 'action', 'view' );
526 if (
527 $e instanceof DBConnectionError &&
528 $context->hasTitle() &&
529 $context->getTitle()->canExist() &&
530 in_array( $action, [ 'view', 'history' ], true ) &&
531 HTMLFileCache::useFileCache( $this->context, HTMLFileCache::MODE_OUTAGE )
532 ) {
533 // Try to use any (even stale) file during outages...
534 $cache = new HTMLFileCache( $context->getTitle(), $action );
535 if ( $cache->isCached() ) {
536 $cache->loadFromFileCache( $context, HTMLFileCache::MODE_OUTAGE );
537 print MWExceptionRenderer::getHTML( $e );
538 exit;
539 }
540 }
541
542 MWExceptionHandler::handleException( $e );
543 } catch ( Error $e ) {
544 // Type errors and such: at least handle it now and clean up the LBFactory state
545 MWExceptionHandler::handleException( $e );
546 }
547
548 $this->doPostOutputShutdown( 'normal' );
549 }
550
551 private function setDBProfilingAgent() {
552 $services = MediaWikiServices::getInstance();
553 // Add a comment for easy SHOW PROCESSLIST interpretation
554 $name = $this->context->getUser()->getName();
555 $services->getDBLoadBalancerFactory()->setAgentName(
556 mb_strlen( $name ) > 15 ? mb_substr( $name, 0, 15 ) . '...' : $name
557 );
558 }
559
560 /**
561 * @see MediaWiki::preOutputCommit()
562 * @param callable|null $postCommitWork [default: null]
563 * @since 1.26
564 */
565 public function doPreOutputCommit( callable $postCommitWork = null ) {
566 self::preOutputCommit( $this->context, $postCommitWork );
567 }
568
569 /**
570 * This function commits all DB and session changes as needed *before* the
571 * client can receive a response (in case DB commit fails) and thus also before
572 * the response can trigger a subsequent related request by the client
573 *
574 * If there is a significant amount of content to flush, it can be done in $postCommitWork
575 *
576 * @param IContextSource $context
577 * @param callable|null $postCommitWork [default: null]
578 * @since 1.27
579 */
580 public static function preOutputCommit(
581 IContextSource $context, callable $postCommitWork = null
582 ) {
583 // Either all DBs should commit or none
584 ignore_user_abort( true );
585
586 $config = $context->getConfig();
587 $request = $context->getRequest();
588 $output = $context->getOutput();
589 $lbFactory = MediaWikiServices::getInstance()->getDBLoadBalancerFactory();
590
591 // Commit all changes
592 $lbFactory->commitMasterChanges(
593 __METHOD__,
594 // Abort if any transaction was too big
595 [ 'maxWriteDuration' => $config->get( 'MaxUserDBWriteDuration' ) ]
596 );
597 wfDebug( __METHOD__ . ': primary transaction round committed' );
598
599 // Run updates that need to block the user or affect output (this is the last chance)
600 DeferredUpdates::doUpdates( 'enqueue', DeferredUpdates::PRESEND );
601 wfDebug( __METHOD__ . ': pre-send deferred updates completed' );
602 // T214471: persist the session to avoid race conditions on subsequent requests
603 $request->getSession()->save();
604
605 // Should the client return, their request should observe the new ChronologyProtector
606 // DB positions. This request might be on a foreign wiki domain, so synchronously update
607 // the DB positions in all datacenters to be safe. If this output is not a redirect,
608 // then OutputPage::output() will be relatively slow, meaning that running it in
609 // $postCommitWork should help mask the latency of those updates.
610 $flags = $lbFactory::SHUTDOWN_CHRONPROT_SYNC;
611 $strategy = 'cookie+sync';
612
613 $allowHeaders = !( $output->isDisabled() || headers_sent() );
614 if ( $output->getRedirect() && $lbFactory->hasOrMadeRecentMasterChanges( INF ) ) {
615 // OutputPage::output() will be fast, so $postCommitWork is useless for masking
616 // the latency of synchronously updating the DB positions in all datacenters.
617 // Try to make use of the time the client spends following redirects instead.
618 $domainDistance = self::getUrlDomainDistance( $output->getRedirect() );
619 if ( $domainDistance === 'local' && $allowHeaders ) {
620 $flags = $lbFactory::SHUTDOWN_CHRONPROT_ASYNC;
621 $strategy = 'cookie'; // use same-domain cookie and keep the URL uncluttered
622 } elseif ( $domainDistance === 'remote' ) {
623 $flags = $lbFactory::SHUTDOWN_CHRONPROT_ASYNC;
624 $strategy = 'cookie+url'; // cross-domain cookie might not work
625 }
626 }
627
628 // Record ChronologyProtector positions for DBs affected in this request at this point
629 $cpIndex = null;
630 $cpClientId = null;
631 $lbFactory->shutdown( $flags, $postCommitWork, $cpIndex, $cpClientId );
632 wfDebug( __METHOD__ . ': LBFactory shutdown completed' );
633
634 if ( $cpIndex > 0 ) {
635 if ( $allowHeaders ) {
636 $now = time();
637 $expires = $now + ChronologyProtector::POSITION_COOKIE_TTL;
638 $options = [ 'prefix' => '' ];
639 $value = LBFactory::makeCookieValueFromCPIndex( $cpIndex, $now, $cpClientId );
640 $request->response()->setCookie( 'cpPosIndex', $value, $expires, $options );
641 }
642
643 if ( $strategy === 'cookie+url' ) {
644 if ( $output->getRedirect() ) { // sanity
645 $safeUrl = $lbFactory->appendShutdownCPIndexAsQuery(
646 $output->getRedirect(),
647 $cpIndex
648 );
649 $output->redirect( $safeUrl );
650 } else {
651 $e = new LogicException( "No redirect; cannot append cpPosIndex parameter." );
652 MWExceptionHandler::logException( $e );
653 }
654 }
655 }
656
657 // Set a cookie to tell all CDN edge nodes to "stick" the user to the DC that handles this
658 // POST request (e.g. the "master" data center). Also have the user briefly bypass CDN so
659 // ChronologyProtector works for cacheable URLs.
660 if ( $request->wasPosted() && $lbFactory->hasOrMadeRecentMasterChanges() ) {
661 $expires = time() + $config->get( 'DataCenterUpdateStickTTL' );
662 $options = [ 'prefix' => '' ];
663 $request->response()->setCookie( 'UseDC', 'master', $expires, $options );
664 $request->response()->setCookie( 'UseCDNCache', 'false', $expires, $options );
665 }
666
667 // Avoid letting a few seconds of replica DB lag cause a month of stale data. This logic is
668 // also intimately related to the value of $wgCdnReboundPurgeDelay.
669 if ( $lbFactory->laggedReplicaUsed() ) {
670 $maxAge = $config->get( 'CdnMaxageLagged' );
671 $output->lowerCdnMaxage( $maxAge );
672 $request->response()->header( "X-Database-Lagged: true" );
673 wfDebugLog( 'replication', "Lagged DB used; CDN cache TTL limited to $maxAge seconds" );
674 }
675
676 // Avoid long-term cache pollution due to message cache rebuild timeouts (T133069)
677 if ( MessageCache::singleton()->isDisabled() ) {
678 $maxAge = $config->get( 'CdnMaxageSubstitute' );
679 $output->lowerCdnMaxage( $maxAge );
680 $request->response()->header( "X-Response-Substitute: true" );
681 }
682 }
683
684 /**
685 * @param string $url
686 * @return string Either "local", "remote" if in the farm, "external" otherwise
687 */
688 private static function getUrlDomainDistance( $url ) {
689 $clusterWiki = WikiMap::getWikiFromUrl( $url );
690 if ( WikiMap::isCurrentWikiId( $clusterWiki ) ) {
691 return 'local'; // the current wiki
692 }
693 if ( $clusterWiki !== false ) {
694 return 'remote'; // another wiki in this cluster/farm
695 }
696
697 return 'external';
698 }
699
700 /**
701 * This function does work that can be done *after* the
702 * user gets the HTTP response so they don't block on it
703 *
704 * This manages deferred updates, job insertion,
705 * final commit, and the logging of profiling data
706 *
707 * @param string $mode Use 'fast' to always skip job running
708 * @since 1.26
709 */
710 public function doPostOutputShutdown( $mode = 'normal' ) {
711 // Record backend request timing
712 $timing = $this->context->getTiming();
713 $timing->mark( 'requestShutdown' );
714
715 // Perform the last synchronous operations...
716 try {
717 // Show visible profiling data if enabled (which cannot be post-send)
718 Profiler::instance()->logDataPageOutputOnly();
719 } catch ( Exception $e ) {
720 // An error may already have been shown in run(), so just log it to be safe
721 MWExceptionHandler::rollbackMasterChangesAndLog( $e );
722 }
723
724 // Disable WebResponse setters for post-send processing (T191537).
725 WebResponse::disableForPostSend();
726
727 $blocksHttpClient = true;
728 // Defer everything else if possible...
729 $callback = function () use ( $mode, &$blocksHttpClient ) {
730 try {
731 $this->restInPeace( $mode, $blocksHttpClient );
732 } catch ( Exception $e ) {
733 // If this is post-send, then displaying errors can cause broken HTML
734 MWExceptionHandler::rollbackMasterChangesAndLog( $e );
735 }
736 };
737
738 if ( function_exists( 'register_postsend_function' ) ) {
739 // https://github.com/facebook/hhvm/issues/1230
740 register_postsend_function( $callback );
741 /** @noinspection PhpUnusedLocalVariableInspection */
742 $blocksHttpClient = false;
743 } else {
744 if ( function_exists( 'fastcgi_finish_request' ) ) {
745 fastcgi_finish_request();
746 /** @noinspection PhpUnusedLocalVariableInspection */
747 $blocksHttpClient = false;
748 } else {
749 // Either all DB and deferred updates should happen or none.
750 // The latter should not be cancelled due to client disconnect.
751 ignore_user_abort( true );
752 }
753
754 $callback();
755 }
756 }
757
758 private function main() {
759 global $wgTitle;
760
761 $output = $this->context->getOutput();
762 $request = $this->context->getRequest();
763
764 // Send Ajax requests to the Ajax dispatcher.
765 if ( $request->getVal( 'action' ) === 'ajax' ) {
766 // Set a dummy title, because $wgTitle == null might break things
767 $title = Title::makeTitle( NS_SPECIAL, 'Badtitle/performing an AJAX call in '
768 . __METHOD__
769 );
770 $this->context->setTitle( $title );
771 $wgTitle = $title;
772
773 $dispatcher = new AjaxDispatcher( $this->config );
774 $dispatcher->performAction( $this->context->getUser() );
775
776 return;
777 }
778
779 // Get title from request parameters,
780 // is set on the fly by parseTitle the first time.
781 $title = $this->getTitle();
782 $action = $this->getAction();
783 $wgTitle = $title;
784
785 // Set DB query expectations for this HTTP request
786 $trxLimits = $this->config->get( 'TrxProfilerLimits' );
787 $trxProfiler = Profiler::instance()->getTransactionProfiler();
788 $trxProfiler->setLogger( LoggerFactory::getInstance( 'DBPerformance' ) );
789 if ( $request->hasSafeMethod() ) {
790 $trxProfiler->setExpectations( $trxLimits['GET'], __METHOD__ );
791 } else {
792 $trxProfiler->setExpectations( $trxLimits['POST'], __METHOD__ );
793 }
794
795 // If the user has forceHTTPS set to true, or if the user
796 // is in a group requiring HTTPS, or if they have the HTTPS
797 // preference set, redirect them to HTTPS.
798 // Note: Do this after $wgTitle is setup, otherwise the hooks run from
799 // isLoggedIn() will do all sorts of weird stuff.
800 if (
801 $request->getProtocol() == 'http' &&
802 // switch to HTTPS only when supported by the server
803 preg_match( '#^https://#', wfExpandUrl( $request->getRequestURL(), PROTO_HTTPS ) ) &&
804 (
805 $request->getSession()->shouldForceHTTPS() ||
806 // Check the cookie manually, for paranoia
807 $request->getCookie( 'forceHTTPS', '' ) ||
808 // check for prefixed version that was used for a time in older MW versions
809 $request->getCookie( 'forceHTTPS' ) ||
810 // Avoid checking the user and groups unless it's enabled.
811 (
812 $this->context->getUser()->isLoggedIn()
813 && $this->context->getUser()->requiresHTTPS()
814 )
815 )
816 ) {
817 $oldUrl = $request->getFullRequestURL();
818 $redirUrl = preg_replace( '#^http://#', 'https://', $oldUrl );
819
820 // ATTENTION: This hook is likely to be removed soon due to overall design of the system.
821 if ( Hooks::run( 'BeforeHttpsRedirect', [ $this->context, &$redirUrl ] ) ) {
822 if ( $request->wasPosted() ) {
823 // This is weird and we'd hope it almost never happens. This
824 // means that a POST came in via HTTP and policy requires us
825 // redirecting to HTTPS. It's likely such a request is going
826 // to fail due to post data being lost, but let's try anyway
827 // and just log the instance.
828
829 // @todo FIXME: See if we could issue a 307 or 308 here, need
830 // to see how clients (automated & browser) behave when we do
831 wfDebugLog( 'RedirectedPosts', "Redirected from HTTP to HTTPS: $oldUrl" );
832 }
833 // Setup dummy Title, otherwise OutputPage::redirect will fail
834 $title = Title::newFromText( 'REDIR', NS_MAIN );
835 $this->context->setTitle( $title );
836 // Since we only do this redir to change proto, always send a vary header
837 $output->addVaryHeader( 'X-Forwarded-Proto' );
838 $output->redirect( $redirUrl );
839 $output->output();
840
841 return;
842 }
843 }
844
845 if ( $title->canExist() && HTMLFileCache::useFileCache( $this->context ) ) {
846 // Try low-level file cache hit
847 $cache = new HTMLFileCache( $title, $action );
848 if ( $cache->isCacheGood( /* Assume up to date */ ) ) {
849 // Check incoming headers to see if client has this cached
850 $timestamp = $cache->cacheTimestamp();
851 if ( !$output->checkLastModified( $timestamp ) ) {
852 $cache->loadFromFileCache( $this->context );
853 }
854 // Do any stats increment/watchlist stuff, assuming user is viewing the
855 // latest revision (which should always be the case for file cache)
856 $this->context->getWikiPage()->doViewUpdates( $this->context->getUser() );
857 // Tell OutputPage that output is taken care of
858 $output->disable();
859
860 return;
861 }
862 }
863
864 // Actually do the work of the request and build up any output
865 $this->performRequest();
866
867 // GUI-ify and stash the page output in MediaWiki::doPreOutputCommit() while
868 // ChronologyProtector synchronizes DB positions or replicas across all datacenters.
869 $buffer = null;
870 $outputWork = function () use ( $output, &$buffer ) {
871 if ( $buffer === null ) {
872 $buffer = $output->output( true );
873 }
874
875 return $buffer;
876 };
877
878 // Now commit any transactions, so that unreported errors after
879 // output() don't roll back the whole DB transaction and so that
880 // we avoid having both success and error text in the response
881 $this->doPreOutputCommit( $outputWork );
882
883 // Now send the actual output
884 print $outputWork();
885 }
886
887 /**
888 * Ends this task peacefully
889 * @param string $mode Use 'fast' to always skip job running
890 * @param bool $blocksHttpClient Whether this blocks an HTTP response to a client
891 */
892 public function restInPeace( $mode = 'fast', $blocksHttpClient = true ) {
893 $lbFactory = MediaWikiServices::getInstance()->getDBLoadBalancerFactory();
894 // Assure deferred updates are not in the main transaction
895 $lbFactory->commitMasterChanges( __METHOD__ );
896
897 // Loosen DB query expectations since the HTTP client is unblocked
898 $trxProfiler = Profiler::instance()->getTransactionProfiler();
899 $trxProfiler->redefineExpectations(
900 $this->context->getRequest()->hasSafeMethod()
901 ? $this->config->get( 'TrxProfilerLimits' )['PostSend-GET']
902 : $this->config->get( 'TrxProfilerLimits' )['PostSend-POST'],
903 __METHOD__
904 );
905
906 // Do any deferred jobs; preferring to run them now if a client will not wait on them
907 DeferredUpdates::doUpdates( $blocksHttpClient ? 'enqueue' : 'run' );
908
909 // Now that everything specific to this request is done,
910 // try to occasionally run jobs (if enabled) from the queues
911 if ( $mode === 'normal' ) {
912 $this->triggerJobs();
913 }
914
915 // Log profiling data, e.g. in the database or UDP
916 wfLogProfilingData();
917
918 // Commit and close up!
919 $lbFactory->commitMasterChanges( __METHOD__ );
920 $lbFactory->shutdown( LBFactory::SHUTDOWN_NO_CHRONPROT );
921
922 wfDebug( "Request ended normally\n" );
923 }
924
925 /**
926 * Send out any buffered statsd data according to sampling rules
927 *
928 * @param IBufferingStatsdDataFactory $stats
929 * @param Config $config
930 * @throws ConfigException
931 * @since 1.31
932 */
933 public static function emitBufferedStatsdData(
934 IBufferingStatsdDataFactory $stats, Config $config
935 ) {
936 if ( $config->get( 'StatsdServer' ) && $stats->hasData() ) {
937 try {
938 $statsdServer = explode( ':', $config->get( 'StatsdServer' ), 2 );
939 $statsdHost = $statsdServer[0];
940 $statsdPort = $statsdServer[1] ?? 8125;
941 $statsdSender = new SocketSender( $statsdHost, $statsdPort );
942 $statsdClient = new SamplingStatsdClient( $statsdSender, true, false );
943 $statsdClient->setSamplingRates( $config->get( 'StatsdSamplingRates' ) );
944 $statsdClient->send( $stats->getData() );
945
946 $stats->clearData(); // empty buffer for the next round
947 } catch ( Exception $ex ) {
948 MWExceptionHandler::logException( $ex );
949 }
950 }
951 }
952
953 /**
954 * Potentially open a socket and sent an HTTP request back to the server
955 * to run a specified number of jobs. This registers a callback to cleanup
956 * the socket once it's done.
957 */
958 public function triggerJobs() {
959 $jobRunRate = $this->config->get( 'JobRunRate' );
960 if ( $this->getTitle()->isSpecial( 'RunJobs' ) ) {
961 return; // recursion guard
962 } elseif ( $jobRunRate <= 0 || wfReadOnly() ) {
963 return;
964 }
965
966 if ( $jobRunRate < 1 ) {
967 $max = mt_getrandmax();
968 if ( mt_rand( 0, $max ) > $max * $jobRunRate ) {
969 return; // the higher the job run rate, the less likely we return here
970 }
971 $n = 1;
972 } else {
973 $n = intval( $jobRunRate );
974 }
975
976 $logger = LoggerFactory::getInstance( 'runJobs' );
977
978 try {
979 if ( $this->config->get( 'RunJobsAsync' ) ) {
980 // Send an HTTP request to the job RPC entry point if possible
981 $invokedWithSuccess = $this->triggerAsyncJobs( $n, $logger );
982 if ( !$invokedWithSuccess ) {
983 // Fall back to blocking on running the job(s)
984 $logger->warning( "Jobs switched to blocking; Special:RunJobs disabled" );
985 $this->triggerSyncJobs( $n, $logger );
986 }
987 } else {
988 $this->triggerSyncJobs( $n, $logger );
989 }
990 } catch ( JobQueueError $e ) {
991 // Do not make the site unavailable (T88312)
992 MWExceptionHandler::logException( $e );
993 }
994 }
995
996 /**
997 * @param int $n Number of jobs to try to run
998 * @param LoggerInterface $runJobsLogger
999 */
1000 private function triggerSyncJobs( $n, LoggerInterface $runJobsLogger ) {
1001 $trxProfiler = Profiler::instance()->getTransactionProfiler();
1002 $old = $trxProfiler->setSilenced( true );
1003 try {
1004 $runner = new JobRunner( $runJobsLogger );
1005 $runner->run( [ 'maxJobs' => $n ] );
1006 } finally {
1007 $trxProfiler->setSilenced( $old );
1008 }
1009 }
1010
1011 /**
1012 * @param int $n Number of jobs to try to run
1013 * @param LoggerInterface $runJobsLogger
1014 * @return bool Success
1015 */
1016 private function triggerAsyncJobs( $n, LoggerInterface $runJobsLogger ) {
1017 // Do not send request if there are probably no jobs
1018 $group = JobQueueGroup::singleton();
1019 if ( !$group->queuesHaveJobs( JobQueueGroup::TYPE_DEFAULT ) ) {
1020 return true;
1021 }
1022
1023 $query = [ 'title' => 'Special:RunJobs',
1024 'tasks' => 'jobs', 'maxjobs' => $n, 'sigexpiry' => time() + 5 ];
1025 $query['signature'] = SpecialRunJobs::getQuerySignature(
1026 $query, $this->config->get( 'SecretKey' ) );
1027
1028 $errno = $errstr = null;
1029 $info = wfParseUrl( $this->config->get( 'CanonicalServer' ) );
1030 $host = $info ? $info['host'] : null;
1031 $port = 80;
1032 if ( isset( $info['scheme'] ) && $info['scheme'] == 'https' ) {
1033 $host = "tls://" . $host;
1034 $port = 443;
1035 }
1036 if ( isset( $info['port'] ) ) {
1037 $port = $info['port'];
1038 }
1039
1040 Wikimedia\suppressWarnings();
1041 $sock = $host ? fsockopen(
1042 $host,
1043 $port,
1044 $errno,
1045 $errstr,
1046 // If it takes more than 100ms to connect to ourselves there is a problem...
1047 0.100
1048 ) : false;
1049 Wikimedia\restoreWarnings();
1050
1051 $invokedWithSuccess = true;
1052 if ( $sock ) {
1053 $special = MediaWikiServices::getInstance()->getSpecialPageFactory()->
1054 getPage( 'RunJobs' );
1055 $url = $special->getPageTitle()->getCanonicalURL( $query );
1056 $req = (
1057 "POST $url HTTP/1.1\r\n" .
1058 "Host: {$info['host']}\r\n" .
1059 "Connection: Close\r\n" .
1060 "Content-Length: 0\r\n\r\n"
1061 );
1062
1063 $runJobsLogger->info( "Running $n job(s) via '$url'" );
1064 // Send a cron API request to be performed in the background.
1065 // Give up if this takes too long to send (which should be rare).
1066 stream_set_timeout( $sock, 2 );
1067 $bytes = fwrite( $sock, $req );
1068 if ( $bytes !== strlen( $req ) ) {
1069 $invokedWithSuccess = false;
1070 $runJobsLogger->error( "Failed to start cron API (socket write error)" );
1071 } else {
1072 // Do not wait for the response (the script should handle client aborts).
1073 // Make sure that we don't close before that script reaches ignore_user_abort().
1074 $start = microtime( true );
1075 $status = fgets( $sock );
1076 $sec = microtime( true ) - $start;
1077 if ( !preg_match( '#^HTTP/\d\.\d 202 #', $status ) ) {
1078 $invokedWithSuccess = false;
1079 $runJobsLogger->error( "Failed to start cron API: received '$status' ($sec)" );
1080 }
1081 }
1082 fclose( $sock );
1083 } else {
1084 $invokedWithSuccess = false;
1085 $runJobsLogger->error( "Failed to start cron API (socket error $errno): $errstr" );
1086 }
1087
1088 return $invokedWithSuccess;
1089 }
1090 }