axios, got, isomorphic-fetch, and node-fetch are all tools for making HTTP requests, but they target different environments and development needs. axios is a promise-based HTTP client that works in both the browser and Node.js, known for its automatic JSON transformation and interceptors. got is a human-friendly HTTP client specifically for Node.js, offering advanced features like pagination and robust retry logic. isomorphic-fetch is a polyfill that enables the native fetch API in older environments, while node-fetch brings the fetch API specifically to Node.js environments prior to native support. Choosing between them depends on whether you need isomorphic code, Node-specific power features, or alignment with web standards.
When architecting data layers for modern web applications, the choice of HTTP client impacts everything from bundle size to error handling strategies. axios, got, isomorphic-fetch, and node-fetch all solve the same fundamental problem β making HTTP requests β but they approach it from different angles. Let's compare how they handle real-world engineering scenarios.
axios is truly isomorphic.
XMLHttpRequest in the browser and the native http module in Node.js.// axios: Works in Browser and Node
import axios from 'axios';
const res = await axios.get('/api/data');
got is strictly for Node.js.
http and tls.// got: Node.js Only
import got from 'got';
const res = await got.get('https://api.example.com/data');
isomorphic-fetch polyfills the global environment.
fetch into the global scope for both Node and older browsers.fetch.// isomorphic-fetch: Global Polyfill
import 'isomorphic-fetch';
const res = await fetch('/api/data'); // Uses injected global
node-fetch is for Node.js environments.
fetch standard specifically for server-side JavaScript.fetch makes this package less critical.// node-fetch: Node.js Implementation
import fetch from 'node-fetch';
const res = await fetch('https://api.example.com/data');
axios wraps the response in an object and parses JSON automatically.
response.data..json() manually, which reduces boilerplate.// axios: Auto-parsed JSON
const response = await axios.get('/api/user');
console.log(response.data); // Direct access to parsed body
got also simplifies response handling but returns a plain object.
Content-Type header matches.response.body.// got: Auto-parsed JSON
const response = await got.get('https://api.example.com/user');
console.log(response.body); // Direct access to parsed body
isomorphic-fetch follows the strict Fetch API standard.
.json() to read the payload.// isomorphic-fetch: Manual JSON parsing
const response = await fetch('/api/user');
const data = await response.json(); // Explicit parsing required
node-fetch mirrors the standard Fetch API behavior.
isomorphic-fetch, it requires manual parsing.fetch but adds extra lines of code.// node-fetch: Manual JSON parsing
const response = await fetch('https://api.example.com/user');
const data = await response.json(); // Explicit parsing required
axios throws an error for any HTTP status outside 2xx.
catch block.// axios: Throws on 4xx/5xx
try {
await axios.get('/api/protected');
} catch (error) {
if (error.response) console.log(error.response.status);
}
got provides detailed error classes for different failure types.
RequestError, HTTPError, and TimeoutError.// got: Detailed Error Classes
try {
await got.get('https://api.example.com/protected');
} catch (error) {
if (error instanceof got.HTTPError) console.log(error.response.statusCode);
}
isomorphic-fetch does not throw on HTTP errors (4xx/5xx).
response.ok to handle HTTP errors.// isomorphic-fetch: Manual Error Check
const response = await fetch('/api/protected');
if (!response.ok) {
throw new Error(`HTTP error! status: ${response.status}`);
}
node-fetch behaves identically to the standard Fetch API.
// node-fetch: Manual Error Check
const response = await fetch('https://api.example.com/protected');
if (!response.ok) {
throw new Error(`HTTP error! status: ${response.status}`);
}
axios uses interceptors to modify requests or responses globally.
// axios: Request Interceptors
axios.interceptors.request.use(config => {
config.headers.Authorization = 'Bearer token';
return config;
});
got uses hooks to achieve similar goals with more control.
// got: Hooks
const api = got.extend({
hooks: {
beforeRequest: [
options => { options.headers.authorization = 'Bearer token'; }
]
}
});
isomorphic-fetch has no built-in interception mechanism.
fetch in a custom function.// isomorphic-fetch: Custom Wrapper
const originalFetch = fetch;
const fetchWithAuth = (url, opts = {}) => {
opts.headers = { ...opts.headers, Authorization: 'Bearer token' };
return originalFetch(url, opts);
};
node-fetch also lacks native interceptors.
isomorphic-fetch, you need to create a wrapper function for global logic.// node-fetch: Custom Wrapper
const originalFetch = fetch;
const fetchWithAuth = (url, opts = {}) => {
opts.headers = { ...opts.headers, Authorization: 'Bearer token' };
return originalFetch(url, opts);
};
axios remains actively maintained with a large community.
got is actively developed for the Node ecosystem.
isomorphic-fetch is considered legacy in modern stacks.
fetch is now available in all major browsers and Node 18+.node-fetch is stable but less critical with Node 18+.
fetch in new Node projects where possible.| Feature | axios | got | isomorphic-fetch | node-fetch |
|---|---|---|---|---|
| Environment | π Browser & Node | π₯οΈ Node Only | π Browser & Node (Polyfill) | π₯οΈ Node Only |
| JSON Parsing | β Automatic | β Automatic | β Manual (.json()) | β Manual (.json()) |
| Error Handling | β οΈ Throws on 4xx/5xx | β οΈ Throws on 4xx/5xx | β OK Status Check | β OK Status Check |
| Interceptors | π οΈ Built-in | π οΈ Hooks | β Custom Wrapper | β Custom Wrapper |
| Status | π’ Active | π’ Active | π‘ Legacy | π’ Stable |
axios is the pragmatic choice for full-stack teams π§°. It reduces boilerplate with auto-parsing and interceptors, making it ideal for dashboards, admin panels, and apps where client/server code sharing matters.
got is the power user's tool for Node.js π§. If you are building backend services, CLI tools, or scripts, its advanced retry logic and error details save hours of debugging.
isomorphic-fetch and node-fetch are for standard-compliance purists π. They are useful if you want to match browser fetch behavior exactly, but be aware that isomorphic-fetch is fading in favor of native implementations.
Final Thought: For most modern frontend architectures, native fetch (with a small wrapper for error handling) or axios (for interceptors) are the top contenders. Reserve got for dedicated Node.js services where its specific strengths shine.
Choose node-fetch if you are on Node.js versions prior to 18 and want to use the standard fetch API syntax without the extra features of axios. Note that version 3 is ESM-only, so ensure your project supports ES modules, or stick to version 2 for CommonJS.
Choose axios if you need a single library that works identically in both the browser and Node.js without configuration. It is ideal for projects that rely heavily on interceptors for authentication or logging, and for teams that prefer automatic JSON parsing over manual response handling.
Choose got if you are building a backend or CLI tool strictly in Node.js. It offers superior developer experience for server-side tasks, including built-in pagination, advanced retry strategies, and detailed error reporting that axios or fetch do not provide out of the box.
Choose isomorphic-fetch only if you must support legacy environments that lack native fetch and you want to write standard fetch syntax. However, for new projects, prefer native fetch or axios, as this package is considered legacy and adds unnecessary global polyfills.
A light-weight module that brings Fetch API to Node.js.
You might be looking for the v2 docs
Instead of implementing XMLHttpRequest in Node.js to run browser-specific Fetch polyfill, why not go from native http to fetch API directly? Hence, node-fetch, minimal code for a window.fetch compatible API on Node.js runtime.
See Jason Miller's isomorphic-unfetch or Leonardo Quixada's cross-fetch for isomorphic usage (exports node-fetch for server-side, whatwg-fetch for client-side).
window.fetch API.res.text() and res.json()) to UTF-8 automatically.window.fetch offers, feel free to open an issue.Current stable release (3.x) requires at least Node.js 12.20.0.
npm install node-fetch
import fetch from 'node-fetch';
node-fetch from v3 is an ESM-only module - you are not able to import it with require().
If you cannot switch to ESM, please use v2 which remains compatible with CommonJS. Critical bug fixes will continue to be published for v2.
npm install node-fetch@2
Alternatively, you can use the async import() function from CommonJS to load node-fetch asynchronously:
// mod.cjs
const fetch = (...args) => import('node-fetch').then(({default: fetch}) => fetch(...args));
To use fetch() without importing it, you can patch the global object in node:
// fetch-polyfill.js
import fetch, {
Blob,
blobFrom,
blobFromSync,
File,
fileFrom,
fileFromSync,
FormData,
Headers,
Request,
Response,
} from 'node-fetch'
if (!globalThis.fetch) {
globalThis.fetch = fetch
globalThis.Headers = Headers
globalThis.Request = Request
globalThis.Response = Response
}
// index.js
import './fetch-polyfill'
// ...
Using an old version of node-fetch? Check out the following files:
NOTE: The documentation below is up-to-date with 3.x releases, if you are using an older version, please check how to upgrade.
import fetch from 'node-fetch';
const response = await fetch('https://github.com/');
const body = await response.text();
console.log(body);
import fetch from 'node-fetch';
const response = await fetch('https://api.github.com/users/github');
const data = await response.json();
console.log(data);
import fetch from 'node-fetch';
const response = await fetch('https://httpbin.org/post', {method: 'POST', body: 'a=1'});
const data = await response.json();
console.log(data);
import fetch from 'node-fetch';
const body = {a: 1};
const response = await fetch('https://httpbin.org/post', {
method: 'post',
body: JSON.stringify(body),
headers: {'Content-Type': 'application/json'}
});
const data = await response.json();
console.log(data);
URLSearchParams is available on the global object in Node.js as of v10.0.0. See official documentation for more usage methods.
NOTE: The Content-Type header is only set automatically to x-www-form-urlencoded when an instance of URLSearchParams is given as such:
import fetch from 'node-fetch';
const params = new URLSearchParams();
params.append('a', 1);
const response = await fetch('https://httpbin.org/post', {method: 'POST', body: params});
const data = await response.json();
console.log(data);
NOTE: 3xx-5xx responses are NOT exceptions, and should be handled in then(), see the next section.
Wrapping the fetch function into a try/catch block will catch all exceptions, such as errors originating from node core libraries, like network errors, and operational errors which are instances of FetchError. See the error handling document for more details.
import fetch from 'node-fetch';
try {
await fetch('https://domain.invalid/');
} catch (error) {
console.log(error);
}
It is common to create a helper function to check that the response contains no client (4xx) or server (5xx) error responses:
import fetch from 'node-fetch';
class HTTPResponseError extends Error {
constructor(response) {
super(`HTTP Error Response: ${response.status} ${response.statusText}`);
this.response = response;
}
}
const checkStatus = response => {
if (response.ok) {
// response.status >= 200 && response.status < 300
return response;
} else {
throw new HTTPResponseError(response);
}
}
const response = await fetch('https://httpbin.org/status/400');
try {
checkStatus(response);
} catch (error) {
console.error(error);
const errorBody = await error.response.text();
console.error(`Error body: ${errorBody}`);
}
Cookies are not stored by default. However, cookies can be extracted and passed by manipulating request and response headers. See Extract Set-Cookie Header for details.
The "Node.js way" is to use streams when possible. You can pipe res.body to another stream. This example uses stream.pipeline to attach stream error handlers and wait for the download to complete.
import {createWriteStream} from 'node:fs';
import {pipeline} from 'node:stream';
import {promisify} from 'node:util'
import fetch from 'node-fetch';
const streamPipeline = promisify(pipeline);
const response = await fetch('https://github.githubassets.com/images/modules/logos_page/Octocat.png');
if (!response.ok) throw new Error(`unexpected response ${response.statusText}`);
await streamPipeline(response.body, createWriteStream('./octocat.png'));
In Node.js 14 you can also use async iterators to read body; however, be careful to catch
errors -- the longer a response runs, the more likely it is to encounter an error.
import fetch from 'node-fetch';
const response = await fetch('https://httpbin.org/stream/3');
try {
for await (const chunk of response.body) {
console.dir(JSON.parse(chunk.toString()));
}
} catch (err) {
console.error(err.stack);
}
In Node.js 12 you can also use async iterators to read body; however, async iterators with streams
did not mature until Node.js 14, so you need to do some extra work to ensure you handle errors
directly from the stream and wait on it response to fully close.
import fetch from 'node-fetch';
const read = async body => {
let error;
body.on('error', err => {
error = err;
});
for await (const chunk of body) {
console.dir(JSON.parse(chunk.toString()));
}
return new Promise((resolve, reject) => {
body.on('close', () => {
error ? reject(error) : resolve();
});
});
};
try {
const response = await fetch('https://httpbin.org/stream/3');
await read(response.body);
} catch (err) {
console.error(err.stack);
}
import fetch from 'node-fetch';
const response = await fetch('https://github.com/');
console.log(response.ok);
console.log(response.status);
console.log(response.statusText);
console.log(response.headers.raw());
console.log(response.headers.get('content-type'));
Unlike browsers, you can access raw Set-Cookie headers manually using Headers.raw(). This is a node-fetch only API.
import fetch from 'node-fetch';
const response = await fetch('https://example.com');
// Returns an array of values, instead of a string of comma-separated values
console.log(response.headers.raw()['set-cookie']);
import fetch, {
Blob,
blobFrom,
blobFromSync,
File,
fileFrom,
fileFromSync,
} from 'node-fetch'
const mimetype = 'text/plain'
const blob = fileFromSync('./input.txt', mimetype)
const url = 'https://httpbin.org/post'
const response = await fetch(url, { method: 'POST', body: blob })
const data = await response.json()
console.log(data)
node-fetch comes with a spec-compliant FormData implementations for posting multipart/form-data payloads
import fetch, { FormData, File, fileFrom } from 'node-fetch'
const httpbin = 'https://httpbin.org/post'
const formData = new FormData()
const binary = new Uint8Array([ 97, 98, 99 ])
const abc = new File([binary], 'abc.txt', { type: 'text/plain' })
formData.set('greeting', 'Hello, world!')
formData.set('file-upload', abc, 'new name.txt')
const response = await fetch(httpbin, { method: 'POST', body: formData })
const data = await response.json()
console.log(data)
If you for some reason need to post a stream coming from any arbitrary place, then you can append a Blob or a File look-a-like item.
The minimum requirement is that it has:
Symbol.toStringTag getter or property that is either Blob or Filestream() method or a arrayBuffer() method that returns a ArrayBuffer.The stream() must return any async iterable object as long as it yields Uint8Array (or Buffer)
so Node.Readable streams and whatwg streams works just fine.
formData.append('upload', {
[Symbol.toStringTag]: 'Blob',
size: 3,
*stream() {
yield new Uint8Array([97, 98, 99])
},
arrayBuffer() {
return new Uint8Array([97, 98, 99]).buffer
}
}, 'abc.txt')
You may cancel requests with AbortController. A suggested implementation is abort-controller.
An example of timing out a request after 150ms could be achieved as the following:
import fetch, { AbortError } from 'node-fetch';
// AbortController was added in node v14.17.0 globally
const AbortController = globalThis.AbortController || await import('abort-controller')
const controller = new AbortController();
const timeout = setTimeout(() => {
controller.abort();
}, 150);
try {
const response = await fetch('https://example.com', {signal: controller.signal});
const data = await response.json();
} catch (error) {
if (error instanceof AbortError) {
console.log('request was aborted');
}
} finally {
clearTimeout(timeout);
}
See test cases for more examples.
url A string representing the URL for fetchingoptions Options for the HTTP(S) requestPromise<Response>Perform an HTTP(S) fetch.
url should be an absolute URL, such as https://example.com/. A path-relative URL (/file/under/root) or protocol-relative URL (//can-be-http-or-https.com/) will result in a rejected Promise.
The default values are shown after each option key.
{
// These properties are part of the Fetch Standard
method: 'GET',
headers: {}, // Request headers. format is the identical to that accepted by the Headers constructor (see below)
body: null, // Request body. can be null, or a Node.js Readable stream
redirect: 'follow', // Set to `manual` to extract redirect headers, `error` to reject redirect
signal: null, // Pass an instance of AbortSignal to optionally abort requests
// The following properties are node-fetch extensions
follow: 20, // maximum redirect count. 0 to not follow redirect
compress: true, // support gzip/deflate content encoding. false to disable
size: 0, // maximum response body size in bytes. 0 to disable
agent: null, // http(s).Agent instance or function that returns an instance (see below)
highWaterMark: 16384, // the maximum number of bytes to store in the internal buffer before ceasing to read from the underlying resource.
insecureHTTPParser: false // Use an insecure HTTP parser that accepts invalid HTTP headers when `true`.
}
If no values are set, the following request headers will be sent automatically:
| Header | Value |
|---|---|
Accept-Encoding | gzip, deflate, br (when options.compress === true) |
Accept | */* |
Content-Length | (automatically calculated, if possible) |
Host | (host and port information from the target URI) |
Transfer-Encoding | chunked (when req.body is a stream) |
User-Agent | node-fetch |
Note: when body is a Stream, Content-Length is not set automatically.
The agent option allows you to specify networking related options which are out of the scope of Fetch, including and not limited to the following:
See http.Agent for more information.
If no agent is specified, the default agent provided by Node.js is used. Note that this changed in Node.js 19 to have keepalive true by default. If you wish to enable keepalive in an earlier version of Node.js, you can override the agent as per the following code sample.
In addition, the agent option accepts a function that returns http(s).Agent instance given current URL, this is useful during a redirection chain across HTTP and HTTPS protocol.
import http from 'node:http';
import https from 'node:https';
const httpAgent = new http.Agent({
keepAlive: true
});
const httpsAgent = new https.Agent({
keepAlive: true
});
const options = {
agent: function(_parsedURL) {
if (_parsedURL.protocol == 'http:') {
return httpAgent;
} else {
return httpsAgent;
}
}
};
Stream on Node.js have a smaller internal buffer size (16kB, aka highWaterMark) from client-side browsers (>1MB, not consistent across browsers). Because of that, when you are writing an isomorphic app and using res.clone(), it will hang with large response in Node.
The recommended way to fix this problem is to resolve cloned response in parallel:
import fetch from 'node-fetch';
const response = await fetch('https://example.com');
const r1 = response.clone();
const results = await Promise.all([response.json(), r1.text()]);
console.log(results[0]);
console.log(results[1]);
If for some reason you don't like the solution above, since 3.x you are able to modify the highWaterMark option:
import fetch from 'node-fetch';
const response = await fetch('https://example.com', {
// About 1MB
highWaterMark: 1024 * 1024
});
const result = await res.clone().arrayBuffer();
console.dir(result);
Passed through to the insecureHTTPParser option on http(s).request. See http.request for more information.
The redirect: 'manual' option for node-fetch is different from the browser & specification, which
results in an opaque-redirect filtered response.
node-fetch gives you the typical basic filtered response instead.
import fetch from 'node-fetch';
const response = await fetch('https://httpbin.org/status/301', { redirect: 'manual' });
if (response.status === 301 || response.status === 302) {
const locationURL = new URL(response.headers.get('location'), response.url);
const response2 = await fetch(locationURL, { redirect: 'manual' });
console.dir(response2);
}
An HTTP(S) request containing information about URL, method, headers, and the body. This class implements the Body interface.
Due to the nature of Node.js, the following properties are not implemented at this moment:
typedestinationmodecredentialscacheintegritykeepaliveThe following node-fetch extension properties are provided:
followcompresscounteragenthighWaterMarkSee options for exact meaning of these extensions.
(spec-compliant)
input A string representing a URL, or another Request (which will be cloned)options Options for the HTTP(S) requestConstructs a new Request object. The constructor is identical to that in the browser.
In most cases, directly fetch(url, options) is simpler than creating a Request object.
An HTTP(S) response. This class implements the Body interface.
The following properties are not implemented in node-fetch at this moment:
trailer(spec-compliant)
body A String or Readable streamoptions A ResponseInit options dictionaryConstructs a new Response object. The constructor is identical to that in the browser.
Because Node.js does not implement service workers (for which this class was designed), one rarely has to construct a Response directly.
(spec-compliant)
Convenience property representing if the request ended normally. Will evaluate to true if the response status was greater than or equal to 200 but smaller than 300.
(spec-compliant)
Convenience property representing if the request has been redirected at least once. Will evaluate to true if the internal redirect counter is greater than 0.
(deviation from spec)
Convenience property representing the response's type. node-fetch only supports 'default' and 'error' and does not make use of filtered responses.
This class allows manipulating and iterating over a set of HTTP headers. All methods specified in the Fetch Standard are implemented.
(spec-compliant)
init Optional argument to pre-fill the Headers objectConstruct a new Headers object. init can be either null, a Headers object, an key-value map object or any iterable object.
// Example adapted from https://fetch.spec.whatwg.org/#example-headers-class
import {Headers} from 'node-fetch';
const meta = {
'Content-Type': 'text/xml'
};
const headers = new Headers(meta);
// The above is equivalent to
const meta = [['Content-Type', 'text/xml']];
const headers = new Headers(meta);
// You can in fact use any iterable objects, like a Map or even another Headers
const meta = new Map();
meta.set('Content-Type', 'text/xml');
const headers = new Headers(meta);
const copyOfHeaders = new Headers(headers);
Body is an abstract interface with methods that are applicable to both Request and Response classes.
(deviation from spec)
Readable streamData are encapsulated in the Body object. Note that while the Fetch Standard requires the property to always be a WHATWG ReadableStream, in node-fetch it is a Node.js Readable stream.
(spec-compliant)
BooleanA boolean property for if this body has been consumed. Per the specs, a consumed body cannot be used again.
fetch comes with methods to parse multipart/form-data payloads as well as
x-www-form-urlencoded bodies using .formData() this comes from the idea that
Service Worker can intercept such messages before it's sent to the server to
alter them. This is useful for anybody building a server so you can use it to
parse & consume payloads.
import http from 'node:http'
import { Response } from 'node-fetch'
http.createServer(async function (req, res) {
const formData = await new Response(req, {
headers: req.headers // Pass along the boundary value
}).formData()
const allFields = [...formData]
const file = formData.get('uploaded-files')
const arrayBuffer = await file.arrayBuffer()
const text = await file.text()
const whatwgReadableStream = file.stream()
// other was to consume the request could be to do:
const json = await new Response(req).json()
const text = await new Response(req).text()
const arrayBuffer = await new Response(req).arrayBuffer()
const blob = await new Response(req, {
headers: req.headers // So that `type` inherits `Content-Type`
}.blob()
})
(node-fetch extension)
An operational error in the fetching process. See ERROR-HANDLING.md for more info.
(node-fetch extension)
An Error thrown when the request is aborted in response to an AbortSignal's abort event. It has a name property of AbortError. See ERROR-HANDLING.MD for more info.
Since 3.x types are bundled with node-fetch, so you don't need to install any additional packages.
For older versions please use the type definitions from DefinitelyTyped:
npm install --save-dev @types/node-fetch@2.x
Thanks to github/fetch for providing a solid implementation reference.
![]() | ![]() | ![]() | ![]() | ![]() |
|---|---|---|---|---|
| David Frank | Jimmy WΓ€rting | Antoni Kepinski | Richie Bendall | Gregor Martynus |