A loader is a node module that exports a function. This function is called when a resource should be transformed by this loader. The given function will have access to the Loader API using the this
context provided to it.
Before we dig into the different types of loaders, their usage, and examples, let's take a look at the three ways you can develop and test a loader locally.
To test a single loader, you can simply use path
to resolve
a local file within a rule object:
webpack.config.js
const path = require('path');
module.exports = {
//...
module: {
rules: [
{
test: /\.js$/,
use: [
{
loader: path.resolve('path/to/loader.js'),
options: {/* ... */}
}
]
}
]
}
};
To test multiple, you can utilize the resolveLoader.modules
configuration to update where webpack will search for loaders. For example, if you had a local /loaders
directory in your project:
webpack.config.js
const path = require('path');
module.exports = {
//...
resolveLoader: {
modules: [
'node_modules',
path.resolve(__dirname, 'loaders')
]
}
};
By the way, if you've already created a separate repository and package for your loader, you could npm link
it to the project in which you'd like to test it out.
When a single loader is applied to the resource, the loader is called with only one parameter – a string containing the content of the resource file.
Synchronous loaders can simply return
a single value representing the transformed module. In more complex cases, the loader can return any number of values by using the this.callback(err, values...)
function. Errors are either passed to the this.callback
function or thrown in a sync loader.
The loader is expected to give back one or two values. The first value is a resulting JavaScript code as string or buffer. The second optional value is a SourceMap as JavaScript object.
When multiple loaders are chained, it is important to remember that they are executed in reverse order – either right to left or bottom to top depending on array format.
So, in the following example, the foo-loader
would be passed the raw resource and the bar-loader
would receive the output of the foo-loader
and return the final transformed module and a source map if necessary.
webpack.config.js
module.exports = {
//...
module: {
rules: [
{
test: /\.js/,
use: [
'bar-loader',
'foo-loader'
]
}
]
}
};
The following guidelines should be followed when writing a loader. They are ordered in terms of importance and some only apply in certain scenarios, read the detailed sections that follow for more information.
Loaders should do only a single task. This not only makes the job of maintaining each loader easier, but also allows them to be chained for usage in more scenarios.
Take advantage of the fact that loaders can be chained together. Instead of writing a single loader that tackles five tasks, write five simpler loaders that divide this effort. Isolating them not only keeps each individual loader simple, but may allow for them to be used for something you hadn't thought of originally.
Take the case of rendering a template file with data specified via loader options or query parameters. It could be written as a single loader that compiles the template from source, executes it and returns a module that exports a string containing the HTML code. However, in accordance with guidelines, a simple apply-loader
exists that can be chained with other open source loaders:
pug-loader
: Convert template to a module that exports a function.apply-loader
: Executes the function with loader options and returns raw HTML.html-loader
: Accepts HTML and outputs a valid JavaScript module.Keep the output modular. Loader generated modules should respect the same design principles as normal modules.
Make sure the loader does not retain state between module transformations. Each run should always be independent of other compiled modules as well as previous compilations of the same module.
Take advantage of the loader-utils
package. It provides a variety of useful tools but one of the most common is retrieving the options passed to the loader. Along with loader-utils
, the schema-utils
package should be used for consistent JSON Schema based validation of loader options. Here's a brief example that utilizes both:
loader.js
import { getOptions } from 'loader-utils';
import { validate } from 'schema-utils';
const schema = {
type: 'object',
properties: {
test: {
type: 'string'
}
}
};
export default function(source) {
const options = getOptions(this);
validate(schema, options, {
name: 'Example Loader',
baseDataPath: 'options'
});
// Apply some transformations to the source...
return `export default ${ JSON.stringify(source) }`;
}
If a loader uses external resources (i.e. by reading from filesystem), they must indicate it. This information is used to invalidate cacheable loaders and recompile in watch mode. Here's a brief example of how to accomplish this using the addDependency
method:
loader.js
import path from 'path';
export default function(source) {
var callback = this.async();
var headerPath = path.resolve('header.js');
this.addDependency(headerPath);
fs.readFile(headerPath, 'utf-8', function(err, header) {
if(err) return callback(err);
callback(null, header + '\n' + source);
});
}
Depending on the type of module, there may be a different schema used to specify dependencies. In CSS for example, the @import
and url(...)
statements are used. These dependencies should be resolved by the module system.
This can be done in one of two ways:
require
statements.this.resolve
function to resolve the path.The css-loader
is a good example of the first approach. It transforms dependencies to require
s, by replacing @import
statements with a require
to the other stylesheet and url(...)
with a require
to the referenced file.
In the case of the less-loader
, it cannot transform each @import
to a require
because all .less
files must be compiled in one pass for variables and mixin tracking. Therefore, the less-loader
extends the less compiler with custom path resolving logic. It then takes advantage of the second approach, this.resolve
, to resolve the dependency through webpack.
Avoid generating common code in every module the loader processes. Instead, create a runtime file in the loader and generate a require
to that shared module:
src/loader-runtime.js
const {someOtherModule} = require('./some-other-module');
module.exports = function runtime(params) {
const x = params.y * 2;
return someOtherModule(params, x);
};
src/loader.js
import runtime from './loader-runtime.js';
export default function loader(source) {
// Custom loader logic
return `${runtime({
source,
y: Math.random()
})}`;
}
Don't insert absolute paths into the module code as they break hashing when the root for the project is moved. There's a stringifyRequest
method in loader-utils
which can be used to convert an absolute path to a relative one.
If the loader you're working on is a simple wrapper around another package, then you should include the package as a peerDependency
. This approach allows the application's developer to specify the exact version in the package.json
if desired.
For instance, the sass-loader
specifies node-sass
as peer dependency like so:
{
"peerDependencies": {
"node-sass": "^4.0.0"
}
}
So you've written a loader, followed the guidelines above, and have it set up to run locally. What's next? Let's go through a simple unit testing example to ensure our loader is working the way we expect. We'll be using the Jest framework to do this. We'll also install babel-jest
and some presets that will allow us to use the import
/ export
and async
/ await
. Let's start by installing and saving these as a devDependencies
:
npm install --save-dev jest babel-jest @babel/core @babel/preset-env
babel.config.js
module.exports = {
presets: [
[
'@babel/preset-env',
{
targets: {
node: 'current',
},
},
],
],
};
Our loader will process .txt
files and simply replace any instance of [name]
with the name
option given to the loader. Then it will output a valid JavaScript module containing the text as its default export:
src/loader.js
import { getOptions } from 'loader-utils';
export default function loader(source) {
const options = getOptions(this);
source = source.replace(/\[name\]/g, options.name);
return `export default ${ JSON.stringify(source) }`;
}
We'll use this loader to process the following file:
test/example.txt
Hey [name]!
Pay close attention to this next step as we'll be using the Node.js API and memfs
to execute webpack. This lets us avoid emitting output
to disk and will give us access to the stats
data which we can use to grab our transformed module:
npm install --save-dev webpack memfs
test/compiler.js
import path from 'path';
import webpack from 'webpack';
import { createFsFromVolume, Volume } from 'memfs';
export default (fixture, options = {}) => {
const compiler = webpack({
context: __dirname,
entry: `./${fixture}`,
output: {
path: path.resolve(__dirname),
filename: 'bundle.js',
},
module: {
rules: [{
test: /\.txt$/,
use: {
loader: path.resolve(__dirname, '../src/loader.js'),
options,
}
}]
}
});
compiler.outputFileSystem = createFsFromVolume(new Volume());
compiler.outputFileSystem.join = path.join.bind(path);
return new Promise((resolve, reject) => {
compiler.run((err, stats) => {
if (err) reject(err);
if (stats.hasErrors()) reject(stats.toJson().errors);
resolve(stats);
});
});
};
And now, finally, we can write our test and add an npm script to run it:
test/loader.test.js
/**
* @jest-environment node
*/
import compiler from './compiler.js';
test('Inserts name and outputs JavaScript', async () => {
const stats = await compiler('example.txt', { name: 'Alice' });
const output = stats.toJson({source: true}).modules[0].source;
expect(output).toBe('export default "Hey Alice!\\n"');
});
package.json
{
"scripts": {
"test": "jest"
},
"jest": {
"testEnvironment": "node"
}
}
With everything in place, we can run it and see if our new loader passes the test:
PASS test/loader.test.js
✓ Inserts name and outputs JavaScript (229ms)
Test Suites: 1 passed, 1 total
Tests: 1 passed, 1 total
Snapshots: 0 total
Time: 1.853s, estimated 2s
Ran all test suites.
It worked! At this point you should be ready to start developing, testing, and deploying your own loaders. We hope that you'll share your creations with the rest of the community!