workerpool vs piscina vs bull
Offloading CPU-Intensive Tasks in Node.js Applications
workerpoolpiscinabullSimilar Packages:

Offloading CPU-Intensive Tasks in Node.js Applications

bull, piscina, and workerpool are all Node.js libraries designed to manage work distribution across multiple threads or processes, but they serve different architectural purposes. bull is a robust queue system built on Redis that enables job scheduling, retries, and distributed processing. piscina is a modern, high-performance worker thread pool implementation that simplifies running JavaScript tasks in parallel using Node.js Worker Threads. workerpool provides a flexible abstraction for executing functions in child processes or worker threads, supporting both local and remote (cluster-based) execution strategies.

Npm Package Weekly Downloads Trend

3 Years

Github Stars Ranking

Stat Detail

Package
Downloads
Stars
Size
Issues
Publish
License
workerpool6,089,4982,289591 kB334 months agoApache-2.0
piscina3,448,1355,095406 kB114 months agoMIT
bull652,45616,251309 kB147a year agoMIT

Offloading CPU-Intensive Work: bull vs piscina vs workerpool

Node.js runs JavaScript in a single thread, which means heavy computations can block your event loop and degrade responsiveness. To avoid this, developers use libraries like bull, piscina, and workerpool to move work off the main thread. But these tools solve different problems β€” one is a job queue, another is a worker thread pool, and the third is a generic parallel execution helper. Let’s break down how they work and when to use each.

🧠 Core Purpose: What Problem Does Each Solve?

bull: A Redis-Based Job Queue

bull isn’t about parallelism per se β€” it’s about reliable, asynchronous job processing. It stores jobs in Redis, allowing them to persist across restarts, be shared across services, and support advanced features like retries, delays, and priorities.

// bull: enqueue a job
const Queue = require('bull');
const emailQueue = new Queue('email');

// Add a job to the queue
await emailQueue.add({ to: 'user@example.com', subject: 'Welcome!' });

// Process jobs (can run in a separate process or server)
emailQueue.process(async (job) => {
  await sendEmail(job.data.to, job.data.subject);
});

βœ… Use bull when you care about durability, distribution, and workflow control β€” not just speed.

piscina: High-Performance Worker Thread Pool

piscina gives you a ready-to-use pool of Node.js Worker Threads with minimal overhead. It’s built for CPU-heavy tasks that need to run fast within a single application instance.

// piscina: run a task in a worker thread
const Piscina = require('piscina');
const piscina = new Piscina({ filename: path.resolve(__dirname, 'worker.js') });

// worker.js exports a function
// export async function processData(data) { /* ... */ }

const result = await piscina.run({ input: 'large dataset' }, { name: 'processData' });

βœ… Use piscina when you need low-latency, in-process parallelism for tasks like hashing, compression, or math-heavy operations.

workerpool: Flexible Task Execution in Workers or Processes

workerpool abstracts away whether you’re using child processes or worker threads, letting you call functions as if they were local β€” even if they run remotely.

// workerpool: execute a function in a worker
const workerpool = require('workerpool');
const pool = workerpool.pool();

// Define a function in a worker file (e.g., 'mathWorker.js')
// function square(n) { return n * n; }

const result = await pool.exec('square', [5], { worker: 'mathWorker.js' });
// result === 25

βœ… Use workerpool when you want a simple, unified API for offloading work without worrying about the underlying concurrency model.

βš™οΈ Concurrency Model: Threads, Processes, or Distributed?

PackageConcurrency ModelShared Memory?Cross-Process?
bullDistributed (via Redis)βŒβœ…
piscinaWorker Threads (in-process)βœ… (via transferable objects)❌
workerpoolWorker Threads or Child Processes⚠️ (depends on mode)βœ… (with cluster)
  • bull assumes workers may live on different machines. Communication happens through Redis, so no shared memory.
  • piscina uses Worker Threads, which share memory via ArrayBuffer transfers β€” very fast, but limited to one machine.
  • workerpool lets you choose: type: 'thread' for Worker Threads, or type: 'process' for full isolation (slower, but safer for unstable code).

πŸ’Ύ Persistence & Reliability

Only bull offers persistent job storage. If your app crashes:

  • With bull: Jobs stay in Redis and resume when the worker restarts.
  • With piscina/workerpool: In-flight tasks are lost. No retry, no history.

This makes bull essential for critical background work (e.g., payment processing), while the others suit ephemeral, best-effort tasks (e.g., real-time analytics).

πŸ“¦ API Style: Promises, Events, or Function Calls?

bull uses an event-driven, queue-based model:

queue.on('completed', (job) => console.log('Done:', job.id));
queue.on('failed', (job, err) => console.error('Failed:', err));

You add jobs and listen for outcomes β€” great for long-running or batched work.

piscina uses async/await with named exports:

// Main thread
const result = await piscina.run(data, { name: 'transform' });

// Worker file must export `transform`
export async function transform(data) { return data.map(x => x * 2); }

Clean and direct β€” feels like calling a local function.

workerpool uses dynamic function invocation:

// Call any exported function by name
const result = await pool.exec('myFunction', [arg1, arg2]);

More flexible than piscina (no need to pre-declare entry points), but slightly more runtime overhead.

πŸ› οΈ Error Handling

  • bull: Built-in retry mechanisms, failure tracking, and dead-letter queues.

    queue.process({ attempts: 3, backoff: 'exponential' }, async (job) => { /* ... */ });
    
  • piscina: Errors bubble up as rejected promises. You handle them like any async call.

    try {
      await piscina.run(...);
    } catch (err) {
      // Handle worker error
    }
    
  • workerpool: Same as piscina β€” errors reject the promise.

If you need automatic retries, only bull provides that out of the box.

🌐 Real-World Scenarios

Scenario 1: Sending Transactional Emails

  • Need: Guaranteed delivery, retries on failure, scale across servers.
  • βœ… Use bull β€” enqueue email jobs, let workers handle delivery with retry logic.

Scenario 2: Real-Time Video Frame Processing

  • Need: Low-latency, high-throughput computation on each frame.
  • βœ… Use piscina β€” spin up a thread pool, pass frames via SharedArrayBuffer for zero-copy.

Scenario 3: Running User-Defined Scripts Safely

  • Need: Isolate untrusted code, prevent main thread crashes.
  • βœ… Use workerpool with type: 'process' β€” each script runs in its own process.

πŸ“‰ When Not to Use Each

  • Don’t use bull for short-lived, non-critical tasks β€” Redis adds operational complexity you may not need.
  • Don’t use piscina if your task involves heavy I/O (like file reads) β€” threads won’t help much; consider clustering instead.
  • Don’t use workerpool if you need fine-grained control over thread lifecycle or maximum performance β€” piscina is leaner and faster.

πŸ“Š Summary Table

Featurebullpiscinaworkerpool
Primary Use CaseDistributed job queueIn-process CPU parallelismGeneric offloading
Persistenceβœ… (Redis)❌❌
Retry Logicβœ… Built-in❌ Manual❌ Manual
Concurrency ModelDistributed workersWorker ThreadsThreads or Processes
API StyleEvent-driven queuePromise + named exportsDynamic function calls
Cross-Machineβœ…βŒβœ… (with cluster mode)
Best ForEmail, reports, ETLMath, crypto, parsingSafe script execution

πŸ’‘ Final Guidance

  • If your task must not be lost and might run on multiple servers, go with bull.
  • If you’re maximizing CPU usage on one machine and need speed, pick piscina.
  • If you want a simple way to run functions off-thread without committing to threads or processes, use workerpool.

All three are mature, actively maintained, and solve real problems β€” but they’re not interchangeable. Choose based on whether you need reliability, raw speed, or flexibility.

How to Choose: workerpool vs piscina vs bull

  • workerpool:

    Choose workerpool when you need flexibility to run tasks in either child processes or worker threads and want a simple function-call abstraction without managing thread lifecycle manually. It works well for moderate parallelism needs in monolithic apps where you don’t require Redis-backed persistence or advanced queue semantics.

  • piscina:

    Choose piscina when you need maximum performance for CPU-bound tasks within a single Node.js instance and want a clean, promise-based API over Worker Threads. It’s well-suited for real-time data transformation, cryptographic operations, or parsing large JSON payloads without blocking the main thread.

  • bull:

    Choose bull when you need durable, persistent job queues with features like delayed jobs, priority levels, rate limiting, and retry logic β€” especially in distributed systems where multiple services must coordinate background work. It’s ideal for email sending, image processing pipelines, or any task that must survive process restarts and scale across machines.

README for workerpool

workerpool

NPM Version NPM Downloads NPM License

workerpool offers an easy way to create a pool of workers for both dynamically offloading computations as well as managing a pool of dedicated workers. workerpool basically implements a thread pool pattern. There is a pool of workers to execute tasks. New tasks are put in a queue. A worker executes one task at a time, and once finished, picks a new task from the queue. Workers can be accessed via a natural, promise based proxy, as if they are available straight in the main application.

workerpool runs on Node.js and in the browser.

Features

  • Easy to use
  • Runs in the browser and on node.js
  • Dynamically offload functions to a worker
  • Access workers via a proxy
  • Cancel running tasks
  • Set a timeout on tasks
  • Handles crashed workers
  • Small: 9 kB minified and gzipped
  • Supports transferable objects (only for web workers and worker_threads)

Why

JavaScript is based upon a single event loop which handles one event at a time. Jeremy Epstein explains this clearly:

In Node.js everything runs in parallel, except your code. What this means is that all I/O code that you write in Node.js is non-blocking, while (conversely) all non-I/O code that you write in Node.js is blocking.

This means that CPU heavy tasks will block other tasks from being executed. In case of a browser environment, the browser will not react to user events like a mouse click while executing a CPU intensive task (the browser "hangs"). In case of a node.js server, the server will not respond to any new request while executing a single, heavy request.

For front-end processes, this is not a desired situation. Therefore, CPU intensive tasks should be offloaded from the main event loop onto dedicated workers. In a browser environment, Web Workers can be used. In node.js, child processes and worker_threads are available. An application should be split in separate, decoupled parts, which can run independent of each other in a parallelized way. Effectively, this results in an architecture which achieves concurrency by means of isolated processes and message passing.

Install

Install via npm:

npm install workerpool

Load

To load workerpool in a node.js application (both main application as well as workers):

const workerpool = require('workerpool');

To load workerpool in the browser:

<script src="workerpool.js"></script>

To load workerpool in a web worker in the browser:

importScripts('workerpool.js');

Setting up the workerpool with React or webpack5 requires additional configuration steps, as outlined in the webpack5 section.

Use

Offload functions dynamically

In the following example there is a function add, which is offloaded dynamically to a worker to be executed for a given set of arguments.

myApp.js

const workerpool = require('workerpool');
const pool = workerpool.pool();

function add(a, b) {
  return a + b;
}

pool
  .exec(add, [3, 4])
  .then(function (result) {
    console.log('result', result); // outputs 7
  })
  .catch(function (err) {
    console.error(err);
  })
  .then(function () {
    pool.terminate(); // terminate all workers when done
  });

Note that both function and arguments must be static and stringifiable, as they need to be sent to the worker in a serialized form. In case of large functions or function arguments, the overhead of sending the data to the worker can be significant.

Dedicated workers

A dedicated worker can be created in a separate script, and then used via a worker pool.

myWorker.js

const workerpool = require('workerpool');

// a deliberately inefficient implementation of the fibonacci sequence
function fibonacci(n) {
  if (n < 2) return n;
  return fibonacci(n - 2) + fibonacci(n - 1);
}

// create a worker and register public functions
workerpool.worker({
  fibonacci: fibonacci,
});

This worker can be used by a worker pool:

myApp.js

const workerpool = require('workerpool');

// create a worker pool using an external worker script
const pool = workerpool.pool(__dirname + '/myWorker.js');

// run registered functions on the worker via exec
pool
  .exec('fibonacci', [10])
  .then(function (result) {
    console.log('Result: ' + result); // outputs 55
  })
  .catch(function (err) {
    console.error(err);
  })
  .then(function () {
    pool.terminate(); // terminate all workers when done
  });

// or run registered functions on the worker via a proxy:
pool
  .proxy()
  .then(function (worker) {
    return worker.fibonacci(10);
  })
  .then(function (result) {
    console.log('Result: ' + result); // outputs 55
  })
  .catch(function (err) {
    console.error(err);
  })
  .then(function () {
    pool.terminate(); // terminate all workers when done
  });

Worker can also initialize asynchronously:

myAsyncWorker.js

define(['workerpool/dist/workerpool'], function (workerpool) {
  // a deliberately inefficient implementation of the fibonacci sequence
  function fibonacci(n) {
    if (n < 2) return n;
    return fibonacci(n - 2) + fibonacci(n - 1);
  }

  // create a worker and register public functions
  workerpool.worker({
    fibonacci: fibonacci,
  });
});

Examples

Examples are available in the examples directory:

https://github.com/josdejong/workerpool/tree/master/examples

API

The API of workerpool consists of two parts: a function workerpool.pool to create a worker pool, and a function workerpool.worker to create a worker.

pool

A workerpool can be created using the function workerpool.pool:

workerpool.pool([script: string] [, options: Object]) : Pool

When a script argument is provided, the provided script will be started as a dedicated worker. When no script argument is provided, a default worker is started which can be used to offload functions dynamically via Pool.exec. Note that on node.js, script must be an absolute file path like __dirname + '/myWorker.js'. In a browser environment, script can also be a data URL like 'data:application/javascript;base64,...'. This allows embedding the bundled code of a worker in your main application. See examples/embeddedWorker for a demo.

The following options are available:

  • minWorkers: number | 'max'. The minimum number of workers that must be initialized and kept available. Setting this to 'max' will create maxWorkers default workers (see below).
  • maxWorkers: number. The default number of maxWorkers is the number of CPU's minus one. When the number of CPU's could not be determined (for example in older browsers), maxWorkers is set to 3.
  • maxQueueSize: number. The maximum number of tasks allowed to be queued. Can be used to prevent running out of memory. If the maximum is exceeded, adding a new task will throw an error. The default value is Infinity.
  • workerType: 'auto' | 'web' | 'process' | 'thread'.
    • In case of 'auto' (default), workerpool will automatically pick a suitable type of worker: when in a browser environment, 'web' will be used. When in a node.js environment, worker_threads will be used if available (Node.js >= 11.7.0), else child_process will be used.
    • In case of 'web', a Web Worker will be used. Only available in a browser environment.
    • In case of 'process', child_process will be used. Only available in a node.js environment.
    • In case of 'thread', worker_threads will be used. If worker_threads are not available, an error is thrown. Only available in a node.js environment.
  • workerTerminateTimeout: number. The timeout in milliseconds to wait for a worker to cleanup it's resources on termination before stopping it forcefully. Default value is 1000.
  • abortListenerTimeout: number. The timeout in milliseconds to wait for abort listener's before stopping it forcefully, triggering cleanup. Default value is 1000.
  • forkArgs: String[]. For process worker type. An array passed as args to child_process.fork
  • forkOpts: Object. For process worker type. An object passed as options to child_process.fork. See nodejs documentation for available options.
  • workerOpts: Object. For web worker type. An object passed to the constructor of the web worker. See WorkerOptions specification for available options.
  • workerThreadOpts: Object. For worker worker type. An object passed to worker_threads.options. See nodejs documentation for available options.
  • onCreateWorker: Function. A callback that is called whenever a worker is being created. It can be used to allocate resources for each worker for example. The callback is passed as argument an object with the following properties:
    • forkArgs: String[]: the forkArgs option of this pool
    • forkOpts: Object: the forkOpts option of this pool
    • workerOpts: Object: the workerOpts option of this pool
    • script: string: the script option of this pool Optionally, this callback can return an object containing one or more of the above properties. The provided properties will be used to override the Pool properties for the worker being created.
  • onTerminateWorker: Function. A callback that is called whenever a worker is being terminated. It can be used to release resources that might have been allocated for this specific worker. The callback is passed as argument an object as described for onCreateWorker, with each property sets with the value for the worker being terminated.
  • emitStdStreams: boolean. For process or thread worker type. If true, the worker will emit stdout and stderr events instead of passing it through to the parent streams. Default value is false.

Important note on 'workerType': when sending and receiving primitive data types (plain JSON) from and to a worker, the different worker types ('web', 'process', 'thread') can be used interchangeably. However, when using more advanced data types like buffers, the API and returned results can vary. In these cases, it is best not to use the 'auto' setting but have a fixed 'workerType' and good unit testing in place.

A worker pool contains the following functions:

  • Pool.exec(method: Function | string, params: Array | null [, options: Object]) : Promise<any, Error>
    Execute a function on a worker with given arguments.

    • When method is a string, a method with this name must exist at the worker and must be registered to make it accessible via the pool. The function will be executed on the worker with given parameters.
    • When method is a function, the provided function fn will be stringified, send to the worker, and executed there with the provided parameters. The provided function must be static, it must not depend on variables in a surrounding scope.
    • The following options are available:
      • on: (payload: any) => void. An event listener, to handle events sent by the worker for this execution. See Events for more details.
      • transfer: Object[]. A list of transferable objects to send to the worker. Not supported by process worker type. See example for usage.
  • Pool.proxy() : Promise<Object, Error>
    Create a proxy for the worker pool. The proxy contains a proxy for all methods available on the worker. All methods return promises resolving the methods result.

  • Pool.stats() : Object
    Retrieve statistics on workers, and active and pending tasks.

    Returns an object containing the following properties:

    {
      totalWorkers: 0,
      busyWorkers: 0,
      idleWorkers: 0,
      pendingTasks: 0,
      activeTasks: 0
    }
    
  • Pool.terminate([force: boolean [, timeout: number]]) : Promise<void, Error>

    If parameter force is false (default), workers will finish the tasks they are working on before terminating themselves. Any pending tasks will be rejected with an error 'Pool terminated'. When force is true, all workers are terminated immediately without finishing running tasks. If timeout is provided, worker will be forced to terminate when the timeout expires and the worker has not finished.

The function Pool.exec and the proxy functions all return a Promise. The promise has the following functions available:

  • Promise.then(fn: Function<result: any>) : Promise<any, Error>
    Get the result of the promise once resolve.
  • Promise.catch(fn: Function<error: Error>) : Promise<any, Error>
    Get the error of the promise when rejected.
  • Promise.finally(fn: Function<void>)
    Logic to run when the Promise either resolves or rejects
  • Promise.cancel() : Promise<any, Error>
    A running task can be cancelled. The worker executing the task is enforced to terminate immediately. The promise will be rejected with a Promise.CancellationError.
  • Promise.timeout(delay: number) : Promise<any, Error>
    Cancel a running task when it is not resolved or rejected within given delay in milliseconds. The timer will start when the task is actually started, not when the task is created and queued. The worker executing the task is enforced to terminate immediately. The promise will be rejected with a Promise.TimeoutError.

Example usage:

const workerpool = require('workerpool');

function add(a, b) {
  return a + b;
}

const pool1 = workerpool.pool();

// offload a function to a worker
pool1
  .exec(add, [2, 4])
  .then(function (result) {
    console.log(result); // will output 6
  })
  .catch(function (err) {
    console.error(err);
  });

// create a dedicated worker
const pool2 = workerpool.pool(__dirname + '/myWorker.js');

// supposed myWorker.js contains a function 'fibonacci'
pool2
  .exec('fibonacci', [10])
  .then(function (result) {
    console.log(result); // will output 55
  })
  .catch(function (err) {
    console.error(err);
  });

// send a transferable object to the worker
// supposed myWorker.js contains a function 'sum'
const toTransfer = new Uint8Array(2).map((_v, i) => i)
pool2
  .exec('sum', [toTransfer], { transfer: [toTransfer.buffer] })
  .then(function (result) {
    console.log(result); // will output 3
  })
  .catch(function (err) {
    console.error(err);
  });

// create a proxy to myWorker.js
pool2
  .proxy()
  .then(function (myWorker) {
    return myWorker.fibonacci(10);
  })
  .then(function (result) {
    console.log(result); // will output 55
  })
  .catch(function (err) {
    console.error(err);
  });

// create a pool with a specified maximum number of workers
const pool3 = workerpool.pool({ maxWorkers: 7 });

worker

A worker is constructed as:

workerpool.worker([methods: Object<String, Function>] [, options: Object]) : void

Argument methods is optional and can be an object with functions available in the worker. Registered functions will be available via the worker pool.

The following options are available:

  • onTerminate: ([code: number]) => Promise<void> | void. A callback that is called whenever a worker is being terminated. It can be used to release resources that might have been allocated for this specific worker. The difference with pool's onTerminateWorker is that this callback runs in the worker context, while onTerminateWorker is executed on the main thread.

Example usage:

// file myWorker.js
const workerpool = require('workerpool');

function add(a, b) {
  return a + b;
}

function multiply(a, b) {
  return a * b;
}

// create a worker and register functions
workerpool.worker({
  add: add,
  multiply: multiply,
});

Asynchronous results can be handled by returning a Promise from a function in the worker:

// file myWorker.js
const workerpool = require('workerpool');

function timeout(delay) {
  return new Promise(function (resolve, reject) {
    setTimeout(resolve, delay);
  });
}

// create a worker and register functions
workerpool.worker({
  timeout: timeout,
});

Transferable objects can be sent back to the pool using Transfer helper class:

// file myWorker.js
const workerpool = require('workerpool');

function array(size) {
  var array = new Uint8Array(size).map((_v, i) => i);
  return new workerpool.Transfer(array, [array.buffer]);
}

// create a worker and register functions
workerpool.worker({
  array: array,
});

Events

You can send data back from workers to the pool while the task is being executed using the workerEmit function:

workerEmit(payload: any) : unknown

This function only works inside a worker and during a task.

Example:

// file myWorker.js
const workerpool = require('workerpool');

function eventExample(delay) {
  workerpool.workerEmit({
    status: 'in_progress',
  });

  workerpool.workerEmit({
    status: 'complete',
  });

  return true;
}

// create a worker and register functions
workerpool.worker({
  eventExample: eventExample,
});

To receive those events, you can use the on option of the pool exec method:

pool.exec('eventExample', [], {
  on: function (payload) {
    if (payload.status === 'in_progress') {
      console.log('In progress...');
    } else if (payload.status === 'complete') {
      console.log('Done!');
    }
  },
});

Worker API

Workers have access to a worker api which contains the following methods

  • emit: (payload: unknown | Transfer): void
  • addAbortListener: (listener: () => Promise<void>): void

addAbortListener

Worker termination may be recoverable through abort listeners which are registered through worker.addAbortListener. If all registered listeners resolve then the worker will not be terminated, allowing for worker reuse in some cases.

NOTE: For operations to successfully clean up, a worker implementation should be async. If the worker thread is blocked, then the worker will be killed.

function asyncTimeout() {
  var me = this;
  return new Promise(function (resolve) {
    let timeout = setTimeout(() => {
        resolve();
    }, 5000);

    // Register a listener which will resolve before the time out
    // above triggers.
    me.worker.addAbortListener(async function () {
        clearTimeout(timeout);
        resolve();
    });
  });
}

// create a worker and register public functions
workerpool.worker(
  {
    asyncTimeout: asyncTimeout,
  },
  {
    abortListenerTimeout: 1000
  }
);

emit

Events may also be emitted from the worker api through worker.emit

// file myWorker.js
const workerpool = require('workerpool');

function eventExample(delay) {
  this.worker.emit({
    status: "in_progress",
  });
  workerpool.workerEmit({
    status: 'complete',
  });

  return true;
}

// create a worker and register functions
workerpool.worker({
  eventExample: eventExample,
});

Utilities

Following properties are available for convenience:

  • platform: The Javascript platform. Either node or browser
  • isMainThread: Whether the code is running in main thread or not (Workers)
  • cpus: The number of CPUs/cores available

Roadmap

  • Implement functions for parallel processing: map, reduce, forEach, filter, some, every, ...
  • Implement graceful degradation on old browsers not supporting webworkers: fallback to processing tasks in the main application.
  • Implement session support: be able to handle a series of related tasks by a single worker, which can keep a state for the session.

Related libraries

Build

First clone the project from github:

git clone git://github.com/josdejong/workerpool.git
cd workerpool

Install the project dependencies:

npm install

Then, the project can be build by executing the build script via npm:

npm run build

This will build the library workerpool.js and workerpool.min.js from the source files and put them in the folder dist.

Test

To execute tests for the library, install the project dependencies once:

npm install

Then, the tests can be executed:

npm test

To test code coverage of the tests:

npm run coverage

To see the coverage results, open the generated report in your browser:

./coverage/index.html

Publish

  • Describe changes in HISTORY.md.
  • Update version in package.json, run npm install to update it in package-lock.json too.
  • Push to GitHub.
  • Deploy to npm via npm publish.
  • Add a git tag with the version number like:
    git tag v1.2.3
    git push --tags
    

License

Copyright (C) 2014-2025 Jos de Jong wjosdejong@gmail.com

Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.