ONNX.js
ONNX.js is a Javascript library for running ONNX models on browsers and on Node.js.
ONNX.js has adopted WebAssembly and WebGL technologies for providing an optimized ONNX model inference runtime for both CPUs and GPUs.
Why ONNX models
The Open Neural Network Exchange (ONNX) is an open standard for representing machine learning models. The biggest advantage of ONNX is that it allows interoperability across different open source AI frameworks, which itself offers more flexibility for AI frameworks adoption. See Getting ONNX Models.
Why ONNX.js
With ONNX.js, web developers can score pre-trained ONNX models directly on browsers with various benefits of reducing server-client communication and protecting user privacy, as well as offering install-free and cross-platform in-browser ML experience.
ONNX.js can run on both CPU and GPU. For running on CPU, WebAssembly is adopted to execute model at near-native speed. Furthermore, ONNX.js utilizes Web Workers to provide a "multi-threaded" environment to parallelize data processing. Empirical evaluation shows very promising performance gains on CPU by taking full advantage of WebAssembly and Web Workers. For running on GPUs, a popular standard for accessing GPU capabilities - WebGL is adopted. ONNX.js has further adopted several novel optimization techniques for reducing data transfer between CPU and GPU, as well as some techniques to reduce GPU processing cycles to further push the performance to the maximum.
See Compatibility and Operators Supported for a list of platforms and operators ONNX.js currently supports.
Benchmarks
Benchmarks have been run against the most prominent open source solutions in the same market. Below are the results collected for Chrome and Edge browsers on one sample machine (computations run on both CPU and GPU):
NOTE:
- Keras.js doesn't support WebGL usage on Edge
- Keras.js and TensorFlow.js doesn't support WebAssembly usage on any browser
The specs of the machine that was used to perform the benchmarking is listed below:
- OS: Microsoft Windows 10 Enterprise Insider Preview
- Model: HP Z240 Tower Workstation
- Processor: Intel(R) Core(TM) i7-6700 CPU @ 3.40GHz, 3401 Mhz, 4 Core(s), 8 Logical Processor(s)
- Installed Physical Memory (RAM): 32.0 GB
- GPU make / Chip type: AMD FirePro W2100 / AMD FirePro SDI (0x6608)
- GPU Memory (approx.): 18.0 GB
Demo
ONNX.js demo website shows the capabilities of ONNX.js. Check the code.
Getting Started
There are multiple ways to use ONNX.js in a project:
Using <script>
tag
This is the most straightforward way to use ONNX.js. The following HTML example shows how to use it:
<html>
<head>
</head>
<body>
<script src="https://cdn.jsdelivr.net/npm/onnxjs/dist/onnx.min.js"></script>
<script>
const myOnnxSession = new onnx.InferenceSession();
myOnnxSession.loadModel("./my-model.onnx").then(()=>{
const inferenceInputs = getInputs();
session.run(inferenceInputs).then(output=>{
const outputTensor = output.values().next().value;
console.log(`model output tensor: ${outputTensor.data}.`);
});
})
</script>
</body>
</html>
Refer to browser/Add for an example.
Using NPM and bundling tools
Modern browser based applications are usually built by frameworks like Angular, React, Vue.js and so on. This solution usually builds the source code into one or more bundle file(s). The following TypeScript example shows how to use ONNX.js in an async context:
- Import
Tensor
and InferenceSession
.
import {Tensor, InferenceSession} from 'onnxjs';
- Create an instance of
InferenceSession
.
const session = new InferenceSession();
- Load the ONNX.js model
const url = './data/models/resnet/model.onnx';
await session.loadModel(url);
- Create your input Tensor(s) similar to the example below. You need to do any pre-processing required by
your model at this stage. For that refer to the documentation of the model you have:
const inputs = [new Tensor(new Float32Array([1.0,2.0,3.0,4.0]), 'float32', [2,2])];
- Run the model with the input Tensors. The output Tensor(s) are available once the run operation is complete:
const outputMap = await session.run(inputs);
const outputTensor = outputMap.values().next().value;
More verbose examples on how to use ONNX.js are located under the examples
folder. For further info see Examples
Running in Node.js
ONNX.js can run in Node.js as well. This is usually for testing purpose. Use the require()
function to load ONNX.js:
require('onnxjs');
Refer to node/Add for a detailed example.
Documents
Developers
For information on development ONNX.js, please check Development
For API reference, please check API.
Getting ONNX models
You could get ONNX models easily in multiple ways:
Learn more about ONNX
Compatibility
Desktop Platforms
OS/Browser | Chrome | Edge | FireFox | Safari | Opera | Electron | Node.js |
---|
Windows 10 | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | - | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: |
macOS | :heavy_check_mark: | - | :heavy_check_mark: | Coming soon | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: |
Ubuntu LTS 18.04 | :heavy_check_mark: | - | :heavy_check_mark: | - | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: |
Mobile Platforms
OS/Browser | Chrome | Edge | FireFox | Safari | Opera |
---|
iOS | Coming soon | - | Coming soon | Coming soon | Coming soon |
Android | :heavy_check_mark: | :heavy_check_mark: | Coming soon | - | :heavy_check_mark: |
Operators
ONNX.js currently supports most operators in ai.onnx operator set v7 (opset v7). See operators.md for a complete, detailed list of which ONNX operators are supported by the 3 available builtin backends (cpu, wasm, and webgl).
Support for ai.onnx.ml operators is coming soon. operators-ml.md has the most recent status of ai.onnx.ml operators.
Contribute
We’d love to embrace your contribution to ONNX.js. Please refer to CONTRIBUTING.md.
Thanks
Thanks to BrowserStack for providing cross browser testing support.
License
Copyright (c) Microsoft Corporation. All rights reserved.
Licensed under the MIT License.