<?php
-
/**
- * Various HTTP related functions
+ * HTTP handling class
+ * @defgroup HTTP HTTP
+ * @file
+ * @ingroup HTTP
*/
+
class Http {
- static function get( $url, $timeout = 'default' ) {
- return request( "GET", $url, $timeout );
+ const SYNC_DOWNLOAD = 1; // syncronys upload (in a single request)
+ const ASYNC_DOWNLOAD = 2; // asynchronous upload we should spawn out another process and monitor progress if possible)
+
+ var $body = '';
+ public static function request( $url, $opts = array() ) {
+ $req = new HttpRequest( $url, $opts );
+ $status = $req->doRequest();
+ if( $status->isOK() ){
+ return $status->value;
+ } else {
+ return false;
+ }
}
- static function post( $url, $timeout = 'default' ) {
- return request( "POST", $url, $timeout );
+ /**
+ * Simple wrapper for Http::request( 'GET' )
+ */
+ public static function get( $url, $opts = array() ) {
+ $opt['method'] = 'GET';
+ return Http::request( $url, $opts );
}
/**
- * Get the contents of a file by HTTP
- *
- * if $timeout is 'default', $wgHTTPTimeout is used
+ * Simple wrapper for Http::request( 'POST' )
*/
- static function request( $method, $url, $timeout = 'default' ) {
- global $wgHTTPTimeout, $wgHTTPProxy, $wgVersion, $wgTitle;
+ public static function post( $url, $opts = array() ) {
+ $opts['method'] = 'POST';
+ return Http::request( $url, $opts );
+ }
- # Use curl if available
- if ( function_exists( 'curl_init' ) ) {
- $c = curl_init( $url );
- if ( wfIsLocalURL( $url ) ) {
- curl_setopt( $c, CURLOPT_PROXY, 'localhost:80' );
- } else if ($wgHTTPProxy) {
- curl_setopt($c, CURLOPT_PROXY, $wgHTTPProxy);
- }
+ public static function doDownload( $url, $target_file_path , $dl_mode = self::SYNC_DOWNLOAD , $redirectCount = 0 ){
+ global $wgPhpCliPath, $wgMaxUploadSize, $wgMaxRedirects;
+ // do a quick check to HEAD to insure the file size is not > $wgMaxUploadSize
+ $head = get_headers( $url, 1 );
- if ( $timeout == 'default' ) {
- $timeout = $wgHTTPTimeout;
+ // check for redirects:
+ if( isset( $head['Location'] ) && strrpos( $head[0], '302' ) !== false ){
+ if( $redirectCount < $wgMaxRedirects ){
+ if( UploadFromUrl::isValidURI( $head['Location'] ) ){
+ return self::doDownload( $head['Location'], $target_file_path , $dl_mode, $redirectCount++ );
+ } else {
+ return Status::newFatal( 'upload-proto-error' );
+ }
+ } else {
+ return Status::newFatal( 'upload-too-many-redirects' );
}
- curl_setopt( $c, CURLOPT_TIMEOUT, $timeout );
- curl_setopt( $c, CURLOPT_USERAGENT, "MediaWiki/$wgVersion" );
- if ( $method == 'POST' )
- curl_setopt( $c, CURLOPT_POST, true );
- else
- curl_setopt( $c, CURLOPT_CUSTOMREQUEST, $method );
-
- # Set the referer to $wgTitle, even in command-line mode
- # This is useful for interwiki transclusion, where the foreign
- # server wants to know what the referring page is.
- # $_SERVER['REQUEST_URI'] gives a less reliable indication of the
- # referring page.
- if ( is_object( $wgTitle ) ) {
- curl_setopt( $c, CURLOPT_REFERER, $wgTitle->getFullURL() );
+ }
+ // we did not get a 200 ok response:
+ if( strrpos( $head[0], '200 OK' ) === false ){
+ return Status::newFatal( 'upload-http-error', htmlspecialchars( $head[0] ) );
+ }
+
+ $content_length = ( isset( $head['Content-Length'] ) ) ? $head['Content-Length'] : null;
+ if( $content_length ){
+ if( $content_length > $wgMaxUploadSize ){
+ return Status::newFatal( 'requested file length ' . $content_length . ' is greater than $wgMaxUploadSize: ' . $wgMaxUploadSize );
}
+ }
- ob_start();
- curl_exec( $c );
- $text = ob_get_contents();
- ob_end_clean();
+ // check if we can find phpCliPath (for doing a background shell request to php to do the download:
+ if( $wgPhpCliPath && wfShellExecEnabled() && $dl_mode == self::ASYNC_DOWNLOAD ){
+ wfDebug( __METHOD__ . "\ASYNC_DOWNLOAD\n" );
+ // setup session and shell call:
+ return self::initBackgroundDownload( $url, $target_file_path, $content_length );
+ } else if( $dl_mode == self::SYNC_DOWNLOAD ){
+ wfDebug( __METHOD__ . "\nSYNC_DOWNLOAD\n" );
+ // SYNC_DOWNLOAD download as much as we can in the time we have to execute
+ $opts['method'] = 'GET';
+ $opts['target_file_path'] = $target_file_path;
+ $req = new HttpRequest( $url, $opts );
+ return $req->doRequest();
+ }
+ }
- # Don't return the text of error messages, return false on error
- if ( curl_getinfo( $c, CURLINFO_HTTP_CODE ) != 200 ) {
- $text = false;
- }
- curl_close( $c );
- } else {
- # Otherwise use file_get_contents, or its compatibility function from GlobalFunctions.php
- # This may take 3 minutes to time out, and doesn't have local fetch capabilities
+ /**
+ * a non blocking request (generally an exit point in the application)
+ * should write to a file location and give updates
+ *
+ */
+ private function initBackgroundDownload( $url, $target_file_path, $content_length = null ){
+ global $wgMaxUploadSize, $IP, $wgPhpCliPath;
+ $status = Status::newGood();
+
+ // generate a session id with all the details for the download (pid, target_file_path )
+ $upload_session_key = self::getUploadSessionKey();
+ $session_id = session_id();
+
+ // store the url and target path:
+ $_SESSION['wsDownload'][$upload_session_key]['url'] = $url;
+ $_SESSION['wsDownload'][$upload_session_key]['target_file_path'] = $target_file_path;
+
+ if( $content_length )
+ $_SESSION['wsDownload'][$upload_session_key]['content_length'] = $content_length;
- $opts = array('http' => array( 'method' => $method ) );
- $ctx = stream_context_create($opts);
+ // set initial loaded bytes:
+ $_SESSION['wsDownload'][$upload_session_key]['loaded'] = 0;
- $url_fopen = ini_set( 'allow_url_fopen', 1 );
- $text = file_get_contents( $url, false, $ctx );
- ini_set( 'allow_url_fopen', $url_fopen );
+ // run the background download request:
+ $cmd = $wgPhpCliPath . ' ' . $IP . "/maintenance/http_session_download.php --sid {$session_id} --usk {$upload_session_key}";
+ $pid = wfShellBackgroundExec( $cmd, $retval );
+ // the pid is not of much use since we won't be visiting this same apache any-time soon.
+ if( !$pid )
+ return Status::newFatal( 'could not run background shell exec' );
+
+ // update the status value with the $upload_session_key (for the user to check on the status of the upload)
+ $status->value = $upload_session_key;
+
+ // return good status
+ return $status;
+ }
+
+ function getUploadSessionKey(){
+ $key = mt_rand( 0, 0x7fffffff );
+ $_SESSION['wsUploadData'][$key] = array();
+ return $key;
+ }
+
+ /**
+ * used to run a session based download. Is initiated via the shell.
+ *
+ * @param $session_id String: the session id to grab download details from
+ * @param $upload_session_key String: the key of the given upload session
+ * (a given client could have started a few http uploads at once)
+ */
+ public static function doSessionIdDownload( $session_id, $upload_session_key ){
+ global $wgUser, $wgEnableWriteAPI, $wgAsyncHTTPTimeout;
+ wfDebug( __METHOD__ . "\n\ndoSessionIdDownload\n\n" );
+ // set session to the provided key:
+ session_id( $session_id );
+ // start the session
+ if( session_start() === false ){
+ wfDebug( __METHOD__ . ' could not start session' );
+ }
+ //get all the vars we need from session_id
+ if(!isset($_SESSION[ 'wsDownload' ][$upload_session_key])){
+ wfDebug( __METHOD__ .' Error:could not find upload session');
+ exit();
+ }
+ // setup the global user from the session key we just inherited
+ $wgUser = User::newFromSession();
+
+ // grab the session data to setup the request:
+ $sd =& $_SESSION['wsDownload'][$upload_session_key];
+ // close down the session so we can other http queries can get session updates:
+ session_write_close();
+
+ $req = new HttpRequest( $sd['url'], array(
+ 'target_file_path' => $sd['target_file_path'],
+ 'upload_session_key'=> $upload_session_key,
+ 'timeout' => $wgAsyncHTTPTimeout
+ ) );
+ // run the actual request .. (this can take some time)
+ wfDebug( __METHOD__ . "do Request: " . $sd['url'] . ' tf: ' . $sd['target_file_path'] );
+ $status = $req->doRequest();
+ //wfDebug("done with req status is: ". $status->isOK(). ' '.$status->getWikiText(). "\n");
+
+ // start up the session again:
+ if( session_start() === false ){
+ wfDebug( __METHOD__ . ' ERROR:: Could not start session');
}
- return $text;
+ // grab the updated session data pointer
+ $sd =& $_SESSION['wsDownload'][$upload_session_key];
+ // if error update status:
+ if( !$status->isOK() ){
+ $sd['apiUploadResult'] = ApiFormatJson::getJsonEncode(
+ array( 'error' => $status->getWikiText() )
+ );
+ }
+ // if status okay process upload using fauxReq to api:
+ if( $status->isOK() ){
+ // setup the FauxRequest
+ $fauxReqData = $sd['mParams'];
+ $fauxReqData['action'] = 'upload';
+ $fauxReqData['format'] = 'json';
+ $fauxReqData['internalhttpsession'] = $upload_session_key;
+
+ // evil but no other clean way about it:
+ $faxReq = new FauxRequest( $fauxReqData, true );
+ $processor = new ApiMain( $faxReq, $wgEnableWriteAPI );
+
+ //init the mUpload var for the $processor
+ $processor->execute();
+ $processor->getResult()->cleanUpUTF8();
+ $printer = $processor->createPrinterByName( 'json' );
+ $printer->initPrinter( false );
+ ob_start();
+ $printer->execute();
+ $apiUploadResult = ob_get_clean();
+
+ wfDebug( __METHOD__ . "\n\n got api result:: $apiUploadResult \n" );
+ // the status updates runner will grab the result form the session:
+ $sd['apiUploadResult'] = $apiUploadResult;
+ }
+ // close the session:
+ session_write_close();
}
/**
* Check if the URL can be served by localhost
+ * @param $url string Full url to check
+ * @return bool
*/
- static function isLocalURL( $url ) {
+ public static function isLocalURL( $url ) {
global $wgCommandLineMode, $wgConf;
if ( $wgCommandLineMode ) {
return false;
}
return false;
}
+
+ /**
+ * Return a standard user-agent we can use for external requests.
+ */
+ public static function userAgent() {
+ global $wgVersion;
+ return "MediaWiki/$wgVersion";
+ }
+}
+class HttpRequest{
+ var $target_file_path;
+ var $upload_session_key;
+
+ function __construct( $url, $opt ){
+ global $wgSyncHTTPTimeout;
+ $this->url = $url;
+ // set the timeout to default sync timeout (unless the timeout option is provided)
+ $this->timeout = ( isset( $opt['timeout'] ) ) ? $opt['timeout'] : $wgSyncHTTPTimeout;
+ $this->method = ( isset( $opt['method'] ) ) ? $opt['method'] : 'GET';
+ $this->target_file_path = ( isset( $opt['target_file_path'] ) ) ? $opt['target_file_path'] : false;
+ $this->upload_session_key = ( isset( $opt['upload_session_key'] ) ) ? $opt['upload_session_key'] : false;
+ }
+
+ /**
+ * Get the contents of a file by HTTP
+ * @param $url string Full URL to act on
+ * @param $Opt associative array Optional array of options:
+ * 'method' => 'GET', 'POST' etc.
+ * 'target_file_path' => if curl should output to a target file
+ * 'adapter' => 'curl', 'soket'
+ */
+ public function doRequest() {
+ # Use curl if available
+ if ( function_exists( 'curl_init' ) ) {
+ return $this->doCurlReq();
+ } else {
+ return $this->doPhpReq();
+ }
+ }
+
+ private function doCurlReq(){
+ global $wgHTTPProxy, $wgTitle;
+
+ $status = Status::newGood();
+ $c = curl_init( $this->url );
+
+ // proxy setup:
+ if ( Http::isLocalURL( $this->url ) ) {
+ curl_setopt( $c, CURLOPT_PROXY, 'localhost:80' );
+ } else if ( $wgHTTPProxy ) {
+ curl_setopt( $c, CURLOPT_PROXY, $wgHTTPProxy );
+ }
+
+ curl_setopt( $c, CURLOPT_TIMEOUT, $this->timeout );
+ curl_setopt( $c, CURLOPT_USERAGENT, Http::userAgent() );
+
+ if ( $this->method == 'POST' ) {
+ curl_setopt( $c, CURLOPT_POST, true );
+ curl_setopt( $c, CURLOPT_POSTFIELDS, '' );
+ } else {
+ curl_setopt( $c, CURLOPT_CUSTOMREQUEST, $this->method );
+ }
+
+ # Set the referer to $wgTitle, even in command-line mode
+ # This is useful for interwiki transclusion, where the foreign
+ # server wants to know what the referring page is.
+ # $_SERVER['REQUEST_URI'] gives a less reliable indication of the
+ # referring page.
+ if ( is_object( $wgTitle ) ) {
+ curl_setopt( $c, CURLOPT_REFERER, $wgTitle->getFullURL() );
+ }
+
+ // set the write back function (if we are writing to a file)
+ if( $this->target_file_path ){
+ $cwrite = new simpleFileWriter( $this->target_file_path, $this->upload_session_key );
+ if( !$cwrite->status->isOK() ){
+ wfDebug( __METHOD__ . "ERROR in setting up simpleFileWriter\n" );
+ $status = $cwrite->status;
+ }
+ curl_setopt( $c, CURLOPT_WRITEFUNCTION, array( $cwrite, 'callbackWriteBody' ) );
+ }
+
+ // start output grabber:
+ if( !$this->target_file_path )
+ ob_start();
+
+ //run the actual curl_exec:
+ try {
+ if ( false === curl_exec( $c ) ) {
+ $error_txt ='Error sending request: #' . curl_errno( $c ) .' '. curl_error( $c );
+ wfDebug( __METHOD__ . $error_txt . "\n" );
+ $status = Status::newFatal( $error_txt );
+ }
+ } catch ( Exception $e ) {
+ // do something with curl exec error?
+ }
+ // if direct request output the results to the stats value:
+ if( !$this->target_file_path && $status->isOK() ){
+ $status->value = ob_get_contents();
+ ob_end_clean();
+ }
+ // if we wrote to a target file close up or return error
+ if( $this->target_file_path ){
+ $cwrite->close();
+ if( !$cwrite->status->isOK() ){
+ return $cwrite->status;
+ }
+ }
+
+ # Don't return the text of error messages, return false on error
+ $retcode = curl_getinfo( $c, CURLINFO_HTTP_CODE );
+ if ( $retcode != 200 ) {
+ wfDebug( __METHOD__ . ": HTTP return code $retcode\n" );
+ $status = Status::newFatal( "HTTP return code $retcode\n" );
+ }
+ # Don't return truncated output
+ $errno = curl_errno( $c );
+ if ( $errno != CURLE_OK ) {
+ $errstr = curl_error( $c );
+ wfDebug( __METHOD__ . ": CURL error code $errno: $errstr\n" );
+ $status = Status::newFatal( " CURL error code $errno: $errstr\n" );
+ }
+ curl_close( $c );
+
+ // return the result obj
+ return $status;
+ }
+
+ public function doPhpReq(){
+ #$use file_get_contents...
+ # This doesn't have local fetch capabilities...
+
+ $headers = array( "User-Agent: " . self :: userAgent() );
+ if( strcasecmp( $method, 'post' ) == 0 ) {
+ // Required for HTTP 1.0 POSTs
+ $headers[] = "Content-Length: 0";
+ }
+ $opts = array(
+ 'http' => array(
+ 'method' => $method,
+ 'header' => implode( "\r\n", $headers ),
+ 'timeout' => $timeout ) );
+ $ctx = stream_context_create( $opts );
+
+ $status->value = file_get_contents( $url, false, $ctx );
+ if( !$status->value ){
+ $status->error( 'file_get_contents-failed' );
+ }
+ return $status;
+ }
+
}
-?>
+
+/**
+ * a simpleFileWriter with session id updates
+ */
+class simpleFileWriter {
+ var $target_file_path;
+ var $status = null;
+ var $session_id = null;
+ var $session_update_interval = 0; // how often to update the session while downloading
+
+ function simpleFileWriter( $target_file_path, $upload_session_key ){
+ $this->target_file_path = $target_file_path;
+ $this->upload_session_key = $upload_session_key;
+ $this->status = Status::newGood();
+ // open the file:
+ $this->fp = fopen( $this->target_file_path, 'w' );
+ if( $this->fp === false ){
+ $this->status = Status::newFatal( 'HTTP::could-not-open-file-for-writing' );
+ }
+ // true start time
+ $this->prevTime = time();
+ }
+
+ public function callbackWriteBody($ch, $data_packet){
+ global $wgMaxUploadSize;
+
+ // write out the content
+ if( fwrite( $this->fp, $data_packet ) === false ){
+ wfDebug( __METHOD__ ." ::could-not-write-to-file\n" );
+ $this->status = Status::newFatal( 'HTTP::could-not-write-to-file' );
+ return 0;
+ }
+
+ // check file size:
+ clearstatcache();
+ $this->current_fsize = filesize( $this->target_file_path );
+
+ if( $this->current_fsize > $wgMaxUploadSize ){
+ wfDebug( __METHOD__ . " ::http download too large\n" );
+ $this->status = Status::newFatal( 'HTTP::file-has-grown-beyond-upload-limit-killing: downloaded more than ' .
+ Language::formatSize( $wgMaxUploadSize ) . ' ' );
+ return 0;
+ }
+
+ // if more than session_update_interval second have passed update_session_progress
+ if( $this->upload_session_key && ( ( time() - $this->prevTime ) > $this->session_update_interval ) ) {
+ $this->prevTime = time();
+ $session_status = $this->update_session_progress();
+ if( !$session_status->isOK() ){
+ $this->status = $session_status;
+ wfDebug( __METHOD__ . ' update session failed or was canceled');
+ return 0;
+ }
+ }
+ return strlen( $data_packet );
+ }
+
+ public function update_session_progress(){
+ $status = Status::newGood();
+ // start the session
+ if( session_start() === false){
+ wfDebug( __METHOD__ . ' could not start session' );
+ exit( 0 );
+ }
+ $sd =& $_SESSION['wsDownload'][$this->upload_session_key];
+ // check if the user canceled the request:
+ if( $sd['user_cancel'] == true ){
+ // kill the download
+ return Status::newFatal( 'user-canceled-request' );
+ }
+ // update the progress bytes download so far:
+ $sd['loaded'] = $this->current_fsize;
+ wfDebug( __METHOD__ . ': set session loaded amount to: ' . $sd['loaded'] . "\n");
+ // close down the session so we can other http queries can get session updates:
+ session_write_close();
+ return $status;
+ }
+
+ public function close(){
+ // do a final session update:
+ $this->update_session_progress();
+ // close up the file handle:
+ if( false === fclose( $this->fp ) ){
+ $this->status = Status::newFatal( 'HTTP::could-not-close-file' );
+ }
+ }
+
+}
\ No newline at end of file