Ori.livneh has uploaded a new change for review.

  https://gerrit.wikimedia.org/r/240316

Change subject: Improved caching for LESS file compilation
......................................................................

Improved caching for LESS file compilation

Caching the output of a LESS compiler is tricky, because a LESS file may
include additional LESS files via @imports, in which case the cache needs
to vary as the contents of those files vary (and not just the contents of
the primary LESS file).

To solve this, we first introduce a utility class, FileContentsHasher. This
class is essentially a smart version of md5_file() -- given one or more file
names, it computes a hash digest of their contents. It tries to avoid
re-reading files by caching the hash digest in APC and re-using it as long as
the files' mtimes have not changed. This is the same approach I used in
I5ceb8537c.

Next, we use this class in ResourceLoaderFileModule in the following way:
whenever we compile a LESS file, we cache the result as an associative array
with the following keys:

* `files` : the list of files whose contents influenced the compiled CSS.
* `hash`  : a hash digest of the combined contents of those files.
* `css`   : the CSS output of the compiler itself.

Before using a cached value, we verify that it is still current by asking
FileContentHasher for a hash of the combined contents of all referenced files,
and we compare that against the value of the `hash` key of the cached entry.

Bug: T112035
Change-Id: I1ff61153ddb95ed17e543bd4af7dd13fa3352861
---
M autoload.php
A includes/FileContentsHasher.php
M includes/resourceloader/ResourceLoaderFileModule.php
3 files changed, 125 insertions(+), 3 deletions(-)


  git pull ssh://gerrit.wikimedia.org:29418/mediawiki/core 
refs/changes/16/240316/1

diff --git a/autoload.php b/autoload.php
index 4bed014..4ad921a 100644
--- a/autoload.php
+++ b/autoload.php
@@ -436,6 +436,7 @@
        'FileBackendStoreShardListIterator' => __DIR__ . 
'/includes/filebackend/FileBackendStore.php',
        'FileBasedSiteLookup' => __DIR__ . 
'/includes/site/FileBasedSiteLookup.php',
        'FileCacheBase' => __DIR__ . '/includes/cache/FileCacheBase.php',
+       'FileContentsHasher' => __DIR__ . '/includes/FileContentsHasher.php',
        'FileDeleteForm' => __DIR__ . '/includes/FileDeleteForm.php',
        'FileDependency' => __DIR__ . '/includes/cache/CacheDependency.php',
        'FileDuplicateSearchPage' => __DIR__ . 
'/includes/specials/SpecialFileDuplicateSearch.php',
diff --git a/includes/FileContentsHasher.php b/includes/FileContentsHasher.php
new file mode 100644
index 0000000..4d20059
--- /dev/null
+++ b/includes/FileContentsHasher.php
@@ -0,0 +1,90 @@
+<?php
+/**
+ * Generate hash digests of file contents to help with cache invalidation.
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License as published by
+ * the Free Software Foundation; either version 2 of the License, or
+ * (at your option) any later version.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License along
+ * with this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
+ * http://www.gnu.org/copyleft/gpl.html
+ *
+ * @file
+ */
+class FileContentsHasher {
+
+       /** @var BagOStuff */
+       protected $cache;
+
+       /**
+        * Constructor.
+        *
+        * @param BagOStuff $cache Cache object to use for storing computed 
hashes.
+        */
+       public function __construct( BagOStuff $cache = null ) {
+               $this->cache = $cache ?: ObjectCache::newAccelerator( 
CACHE_ANYTHING );
+       }
+
+       /**
+        * Get a hash of a file's contents, either by retrieving a previously-
+        * computed hash from the cache, or by computing a hash from the file.
+        *
+        * @param string $filePath Full path to the file.
+        * @param string $algo Name of selected hashing algorithm.
+        * @return string|bool Hash of file contents, or false if the file 
could not be read.
+        */
+       public function getFileContentsHash( $filePath, $algo = 'md4' ) {
+               $mtime = filemtime( $filePath );
+               if ( $mtime === false ) {
+                       return false;
+               }
+
+               $cacheKey = wfGlobalCacheKey( __CLASS__, $filePath, $algo );
+               $cachedHash = $this->cache->get( $cacheKey );
+
+               if ( isset( $cachedHash['mtime'] ) && $cachedHash['mtime'] >= 
$mtime ) {
+                       return $cachedHash['hash'];
+               }
+
+               $contents = file_get_contents( $filePath );
+               if ( $contents === false ) {
+                       return false;
+               }
+
+               $hash = hash( $algo, $contents );
+               $cache->set( $cacheKey, array(
+                       'mtime' => $mtime,
+                       'hash'  => $hash
+               ), 60 * 60 * 24 );
+
+               return $hash;
+       }
+
+       /**
+        * Get a hash of the combined contents of multiple files, either by
+        * retrieving a previously-computed hash from the cache, or by computing
+        * a hash from the files.
+        *
+        * @param array $filePaths Array of paths to files.
+        * @param string $algo Name of selected hashing algorithm.
+        * @return string|bool Hash of files' contents, or false if no file 
could not be read.
+        */
+       public function getFilesContentHash( array $filePaths, $algo = 'md4' ) {
+               sort( $filePaths );
+               $that = $this;
+               $hashes = array_map( function ( $filePath ) use ( $algo, $that 
) {
+                       return $that->getFileContentsHash( $filePath, $algo ) 
?: '';
+               }, $filePaths );
+
+               $hashes = implode( '', $hashes );
+               return $hashes ? hash( $algo, $hashes ) : false;
+       }
+}
diff --git a/includes/resourceloader/ResourceLoaderFileModule.php 
b/includes/resourceloader/ResourceLoaderFileModule.php
index 7fbc1cb..dbc4dad 100644
--- a/includes/resourceloader/ResourceLoaderFileModule.php
+++ b/includes/resourceloader/ResourceLoaderFileModule.php
@@ -966,12 +966,43 @@
         * @return string CSS source
         */
        protected function compileLessFile( $fileName, $compiler = null ) {
+               static $cache, $hasher;
+
+               if ( !$cache ) {
+                       $cache = ObjectCache::newAccelerator( CACHE_ANYTHING );
+                       $hasher = new FileContentsHasher( $cache );
+               }
+
+               // Construct a cache key from the LESS file name and a hash 
digest
+               // of the LESS variables used for compilation.
+               $varsHash = md5( serialize( ResourceLoader::getLessVars( 
$this->getConfig() ) ) );
+               $cacheKey = wfGlobalCacheKey( 'LESS', $fileName, $varsHash );
+               $cachedCompile = $cache->get( $cacheKey );
+
+               // If we got a cached value, we have to validate it by getting a
+               // checksum of all the files that were loaded by the parser and
+               // ensuring it matches the cached entry's.
+               if ( isset( $cachedCompile['hash'] ) ) {
+                       $contentHash = $hasher->getFilesContentHash( 
$cachedCompile['files'] );
+                       if ( $contentHash === $cachedCompile['hash'] ) {
+                               return $cachedCompile['css'];
+                       }
+               }
+
                if ( !$compiler ) {
                        $compiler = $this->getLessCompiler();
                }
-               $result = $compiler->parseFile( $fileName )->getCss();
-               $this->localFileRefs += array_keys( $compiler->AllParsedFiles() 
);
-               return $result;
+
+               $css = $compiler->parseFile( $fileName )->getCss();
+               $files = array_keys( $compiler->AllParsedFiles() );
+               $this->localFileRefs += $files;
+               $cache->set( $cacheKey, array(
+                       'css'   => $css,
+                       'files' => $files,
+                       'hash'  => $hasher->getFilesContentHash( $files ),
+               ), 60 * 60 * 24 );
+
+               return $css;
        }
 
        /**

-- 
To view, visit https://gerrit.wikimedia.org/r/240316
To unsubscribe, visit https://gerrit.wikimedia.org/r/settings

Gerrit-MessageType: newchange
Gerrit-Change-Id: I1ff61153ddb95ed17e543bd4af7dd13fa3352861
Gerrit-PatchSet: 1
Gerrit-Project: mediawiki/core
Gerrit-Branch: master
Gerrit-Owner: Ori.livneh <[email protected]>

_______________________________________________
MediaWiki-commits mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/mediawiki-commits

Reply via email to