Security News
PyPI’s New Archival Feature Closes a Major Security Gap
PyPI now allows maintainers to archive projects, improving security and helping users make informed decisions about their dependencies.
Web Augmented Reality. A-Frame Component. Currently supports gesture recognition
:pushpin: Gesture Recognition A-Frame Component: Capable of recognizing 18 gestures.
:pushpin: Custom Gesture Training: Train personalized models and pass them to the FernAR engine for customized gesture recognition.
Place the library into a scene to initiate an AR scene using the device camera. This allows you to use gestures for controlling A-Frame elements. If the library's predefined gestures are insufficient, you can explore the Train Your Own Model section to train a custom model.
Live demo: Enter AR
<html>
<head>
<meta name="viewport" content="width=device-width, initial-scale=1" />
<script src="https://aframe.io/releases/1.5.0/aframe.min.js"></script>
<script src="https://unpkg.com/aframe-event-set-component@5.0.0/dist/aframe-event-set-component.min.js"></script>
<script src="https://cdn.jsdelivr.net/npm/fern-ar@1.0.0/dist/fernar-gesture.prod.js"></script>
</head>
<body>
<a-scene fernar-gesture="drawLandmarker:true">
<a-camera active="true" position="0 0 0"></a-camera>
<a-box
id="coolbox"
fernar-gesture-target="gesture: 0,1,2"
event-set__red="_event: fernar-gesture-event-0; material.color: red"
event-set__blue="_event: fernar-gesture-event-1; material.color: blue"
event-set__green="_event: fernar-gesture-event-2; material.color: green"
color="yellow"
position="0 0 -10"
></a-box>
</a-scene>
</body>
</html>
<script src="https://cdn.jsdelivr.net/npm/fern-ar@latest/dist/fernar-gesture.prod.js"></script>
$ npm install fern-ar@latest
fernar-gesture
to <a-scene>
<a-scene fernar-gesture>
FERNAR.updateModel
function with JSON (string), weights (array), and weight bin file name.async function initializeScene() {
const jsonModelPath = "../test-model/model.json";
const binModelPath = "../test-model/model.weights.bin";
// Fetching JSON model file
const jsonResponse = await fetch(jsonModelPath);
const jsonModelFile = await jsonResponse.text();
// Fetching binary model file
const binResponse = await fetch(binModelPath);
const binModelFile = new Uint8Array(await binResponse.arrayBuffer());
await FERNAR.updateModel(jsonModelFile, binModelFile, binModelPath);
}
initializeScene();
FERNAR.setDebugMode
to enable debug messagesFERNAR.setDebugMode(true);
fernar-gesture
Property | Type | default |
---|---|---|
drawLandmarker | boolean | true |
threshold | int | 10 |
confidence | number | 0.7 |
planePosition | array | -6 3 -7 |
planeWidth | int | 5 |
planeHeight | int | 5 |
<a-scene>
<a-scene fernar-gesture="drawLandmarker:true">
<a-camera active="true" position="0 0 0"></a-camera>
</a-scene>
Instruct FernAR to register the entity for gesture events by adding fernar-gesture-target
to the entity and specifying an array of gesture IDs the entity should listen to. The event name would be fernar-gesture-event-${gesture-id}
.
For example, by using <a-box fernar-gesture-target="gesture: 1,2,5"></a-box>
, the FernAR engine will send fernar-gesture-event-1, fernar-gesture-event-2, and fernar-gesture-event-5 events to the <a-box>
every time the engine detects the corresponding gesture.
entity.addEventListener("fernar-gesture-event-0", function (event) {
...
});
Refer to this Git page Train Model to train a customized model. Provide a CSV file (you can use the one in data/label.csv), enter the new gesture ID you want to recognize, and perferm the gesture in front of the camera. The web pages will record the gesture landmark every time you click Store Gesture
.
After labeling, you can either click the Download CSV File
button to download the new labeling CSV file or specify the total number of gestures you want to train (ex: gestures 0,1,2,3 would be total 4 gestures) and click Start Training
to train the model using the CSV file you just labeled.
Once the training is complete, the web pages will automatically save the model in your device's Download folder. You should find two files: model.json and model.weight.bin.
Any support or questions are welcome! If you encounter any bugs or have suggestions for improvements, please feel free to open issues, and let's discuss them together.
:star: The gesture recognition heavily relies on the repo hand-gesture-recognition-using-mediapipe
:star: A-Frame: https://github.com/aframevr/aframe
:star: MediaPipe: https://developers.google.com/mediapipe/solutions/vision/hand_landmarker
:star: Mind AR: https://github.com/hiukim/mind-ar-js
FAQs
augmented reality in web. A-Frame. Gesture Recognition
We found that fern-ar demonstrated a healthy version release cadence and project activity because the last version was released less than a year ago. It has 1 open source maintainer collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Security News
PyPI now allows maintainers to archive projects, improving security and helping users make informed decisions about their dependencies.
Research
Security News
Malicious npm package postcss-optimizer delivers BeaverTail malware, targeting developer systems; similarities to past campaigns suggest a North Korean connection.
Security News
CISA's KEV data is now on GitHub, offering easier access, API integration, commit history tracking, and automated updates for security teams and researchers.