main-thread-scheduling
Fast and consistently responsive apps using a single function call
Install
npm install main-thread-scheduling
Overview
The library lets you run computationally heavy tasks on the main thread while ensuring:
- Your app's UI doesn't freeze.
- Your users' computer fans don't spin.
- It's easy to plug it into your existing codebase.
A real-world showcase of searching in 10k files and getting results instantly — https://twitter.com/antoniostoilkov/status/1539576912498118656.
Use Cases
- You want to turn a synchronous function into a non-blocking asynchronous function. Avoids UI freezes.
- You want to render important elements first and less urgent ones second. Improves perceived performance.
- You want to run a long background task that doesn't spin the fans after a while. Avoids bad reputation.
- You want to run multiple backgrounds tasks that don't degrade your app performance with time. Prevents death by a thousand cuts.
How It Works
- Uses
requestIdleCallback()
and requestAfterFrame()
for scheduling. - Stops task execution when user interacts with the UI (if
navigator.scheduling.isInputPending()
API is available). - Global queue. Multiple tasks are executed one by one so increasing the number of tasks doesn't degrade performance linearly.
- Sorts tasks by importance. Sorts by strategy and gives priority to tasks requested
later.
- Considerate about your existing code. Tasks with
idle
strategy are executed last so there
isn't some unexpected work that slows down the main thread after the background task is finished.
Why
- Simple. 90% of the time you only need the
yieldOrContinue(strategy)
function. The API has two more functions for more advanced cases. - Not a weekend project. Actively maintained for three years — see contributors page. I've been using it in my own products for over four years — Nota and iBar. Flux.ai are also using it in their product (software for designing hardware circuits using web technologies).
- This is the future. Some browsers have already implemented support for scheduling tasks on the main thread. This library tries even harder to improve user perceived performance — see explanation for details.
- High quality. Aiming for high-quality with my open-source principles.
Example
You can see the library in action in this CodeSandbox. Try removing the call to yieldToContinue()
and then type in the input to see the difference.
API
yieldOrContinue(strategy: 'interactive' | 'smooth' | 'idle')
The complexity of the entire library is hidden behind this method. You can have great app performance by calling a single method.
async function findInFiles(query: string) {
for (const file of files) {
await yieldOrContinue('interactive')
for (const line of file.lines) {
fuzzySearchLine(line, query)
}
}
}
requestAfterFrame(callback)
This is a utility function, most people don't need to use it. The same way requestAnimationFrame()
queues a callback
to be executed just before a frame is rendered requestAfterFrame()
is called just after a frame is rendered.
queueTask(callback)
This is a utility function, most people don't need to use it. The same way queueMicrotask()
queues a callback
to be executed in the end of the current microtask, queueTask()
queues the task for the next task. You learn more at here.
More complex scenarios
The library has two more functions available:
yieldControl(strategy: 'interactive' | 'smooth' | 'idle')
isTimeToYield(strategy: 'interactive' | 'smooth' | 'idle')
These two functions are used together to handle more advanced use cases.
A simple use case where you will need those two functions is when you want to render your view before yielding back control to the browser to continue its work:
async function doHeavyWork() {
for (const value of values) {
if (isTimeToYield('interactive')) {
render()
await yieldControl('interactive')
}
computeHeavyWorkOnValue(value)
}
}
Scheduling strategies
There are three scheduling strategies available. You can think about them more easily by completing the sentence with one of the three words: "Scheduling the task keeps the page interactive
/smooth
/idle
."
interactive
– use this for things that need to display to the user as fast as possible. Every interactive
task is run for 83ms – this gives you a nice cycle of doing heavy work and letting the browser render pending changes.smooth
— use this for things you want to display to the user quickly but you still want for animations to run smoothly for example. smooth
runs for 13ms and then gives around 3ms to render the frame.idle
– use this for background tasks. Every idle task is run for 5ms.
Alternatives
Web Workers
Web Workers are a great fit if you have: 1) heavy algorithm (e.g. image processing), 2) heavy process (runs for a long time, big part of the app lifecycle). However, in reality, it's rare to see people using them. That's because they require significant investment of time due to the complexity that can't be avoided when working with CPU threads regardless of the programming language. This library can be used as a gateway before transitioning to Web Workers. In most cases, you would discover the doing it on the main thread is good enough.
scheduler.postTask()
scheduler.postTask()
is available in some browsers today. postTask
is a great alternative, you just need to have a better understanding on its inner workings. main-thread-scheduling
aims to be easier to use. For example, main-thread-scheduling
uses the isInputPending()
API to ensure the UI doesn't freeze when the user interacts with the page (if you use scheduler.postTask()
you will need to do that manually). Also, if you have running animations while running your tasks, you will see main-thread-scheduling
perform better.
If you want the benefits of main-thread-scheduling
, but you prefer the callback based postTask()
API/thinking model, then here is an implementation of postTask()
using yieldOrContinue()
:
async function postTask(callback: () => void | Promise<void>) {
await yieldOrContinue('interactive')
await callback()
}