Research
Security News
Malicious npm Packages Inject SSH Backdoors via Typosquatted Libraries
Socket’s threat research team has detected six malicious npm packages typosquatting popular libraries to insert SSH backdoors.
Light weight face detector hight level client with multiple detection techniques.
This repo contains various types of face detection techniques. All the face detection techniques are fine tunned and optimized out of the box to work the best with any resolution images and takes no time to get started
Key features:
Detectors:
( More on the way...)
Like said setup and usage is very simple and easy.
Example
from face_detectors import Ultralight320Detector
from face_detectors.utils import annotate_image
detector = Ultralight320Detector()
image = cv2.imread("image.png")
faces = detector.detect_faces(image)
image = annotate_image(image, faces, width=3)
cv2.imshow("view", image)
cv2.waitKey(100000)
Every detector has different types of features and can be used for different purposes for example detecting only one face we can use hog with number_of_times_to_upsample=1
or caffemodel, we can also use models but other models like Ultralight models are good for multiple and small face detections.
(The following is test on MacBook Pro 2.3 GHz Quad-Core Intel Core i5 with 8 GB 2133 MHz LPDDR3)
Detector | IMAGE 1 (ms) | IMAGE 2 (ms) | IMAGE 3 (ms) | IMAGE 4 (ms) |
---|---|---|---|---|
Caffe Model | 0.0334 | 0.0327 | 0.0314 | 0.0344 |
CNN | 0.5216 | 0.1371 | 0.4339 | 0.2264 |
Hog | 0.0970 | 0.4521 | 0.0847 | 0.0548 |
UltraLight (320px) | 0.0128 | 0.0203 | 0.0128 | 0.0149 |
UltraLight (640px) | 0.0347 | 0.0391 | 0.0430 | 0.0384 |
The below is IMAGE 2 result
Briefly describing face-detectors package that are all the detectors and utility functions.
Caffemodel is very light weight model that uses less resources to perform detections that is created by caffe (Convolutional Architecture for Fast Feature Embedding).
import cv2
from face_detectors import CaffemodelDetector
from face_detectors.utils import annotate_image
vid = cv2.VideoCapture(0)
detector = CaffemodelDetector()
while True:
rect, frame = vid.read()
if not rect:
break
bbox = detector.detect_faces(frame)
frame = annotate_image(frame, bbox)
cv2.imshow("Caffe Model Detection", frame)
cv2.waitKey(1)
Configurable options for CaffeModel detector.
Syntax: CaffemodelDetector(**options)
Options | Description |
---|---|
convert_color | Takes OpenCV COLOR codes to convert the images. Defaults to cv2.COLOR_BGR2RGB |
confidence | Confidence score is used to refrain from making predictions when it is not above a sufficient threshold. Defaults to 0.5 |
scale | Scales the image for faster output (No need to set this manually, scale will be determined automatically if no value is given) |
mean | Scalar with mean values which are subtracted from channels. Values are intended to be in (mean-R, mean-G, mean-B) order if image has BGR ordering and swapRB is true. Defaults to (104.0, 177.0, 123.0). |
scalefactor | Multiplier for images values. Defaults to 1.0. |
crop | Flag which indicates whether image will be cropped after resize or not. Defaults to False. |
swapRB | Flag which indicates that swap first and last channels in 3-channel image is necessary. Defaults to False. |
transpose | Transpose image. Defaults to False. |
resize | Spatial size for output image. Default is (300, 300) |
Useful methods for this detector:
detect_faces(image)
This method will return coordinates for all the detected faces of the given image
Options | Description |
---|---|
image | image in numpy array format |
detect_faces_keypoints(image, get_all=false)
This method will return coordinates for all the detected faces along with their facial keypoints of the given image. Keypoints are detected using dlib's new shape_predictor_68_face_landmarks_GTX.dat` model.
Note: Generating keypoints might take more time if compared with detect_faces
method
Options | Description |
---|---|
image | Image in numpy array format |
get_all | Weather to get all facial keypoints or the main (chin, nose, eyes, mouth) |
CNN (Convolutional Neural Network) might not be a light weight model but it is good at detecting faces from all angles. This detector is a hight level wrapper around dlib::cnn_face_detection_model_v1
that is fine tuned to improve overall performance and accuracy.
import cv2
from face_detectors import CNNDetector
from face_detectors.utils import annotate_image
vid = cv2.VideoCapture(0)
detector = CNNDetector()
while True:
rect, frame = vid.read()
if not rect:
break
bbox = detector.detect_faces(frame)
frame = annotate_image(frame, bbox)
cv2.imshow("CNN Detection", frame)
cv2.waitKey(1)
Configurable options for CNNDetector detector.
Syntax: CNNDetector(**options)
Options | Description |
---|---|
convert_color | Takes OpenCV COLOR codes to convert the images. Defaults to cv2.COLOR_BGR2RGB |
number_of_times_to_upsample | Up samples the image number_of_times_to_upsample before running the basic detector. By default is 1. |
confidence | Confidence score is used to refrain from making predictions when it is not above a sufficient threshold. Defaults to 0.5 |
scale | Scales the image for faster output (No need to set this manually, scale will be determined automatically if no value is given) |
detect_faces(image)
This method will return coordinates for all the detected faces of the given image
Options | Description |
---|---|
image | image in numpy array format |
detect_faces_keypoints(image, get_all=false)
This method will return coordinates for all the detected faces along with their facial keypoints of the given image. Keypoints are detected using dlib's new shape_predictor_68_face_landmarks_GTX.dat
model.
Note: Generating keypoints might take more time if compared with detect_faces
method
Options | Description |
---|---|
image | Image in numpy array format |
get_all | Weather to get all facial keypoints or the main (chin, nose, eyes, mouth) |
This detector uses Histogram of Oriented Gradients (HOG) and Linear SVM classifier for face detection. It is also combined with an image pyramid and a sliding window detection scheme. HogDetector
is a high level client over dlib's hog face detector and is fine tuned to make it more optimized in both speed and accuracy.
If you want to detect faster with HogDetector
and don't care about number of detections then set number_of_times_to_upsample=1
in the options, it will detect less fasces in less time, mainly used for real time one face detection.
import cv2
from face_detectors import HogDetector
from face_detectors.utils import annotate_image
vid = cv2.VideoCapture(0)
detector = HogDetector()
while True:
rect, frame = vid.read()
if not rect:
break
bbox = detector.detect_faces(frame)
frame = annotate_image(frame, bbox)
cv2.imshow("Hog Detection", frame)
cv2.waitKey(1)
Configurable options for HogDetector detector.
Syntax: HogDetector(**options)
Options | Description |
---|---|
convert_color | Takes OpenCV COLOR codes to convert the images. Defaults to cv2.COLOR_BGR2RGB |
number_of_times_to_upsample | Up samples the image number_of_times_to_upsample before running the basic detector. By default is 2. |
confidence | Confidence score is used to refrain from making predictions when it is not above a sufficient threshold. Defaults to 0.5 |
scale | Scales the image for faster output (No need to set this manually, scale will be determined automatically if no value is given) |
detect_faces(image)
This method will return coordinates for all the detected faces of the given image
Options | Description |
---|---|
image | image in numpy array format |
detect_faces_keypoints(image, get_all=false)
This method will return coordinates for all the detected faces along with their facial keypoints of the given image. Keypoints are detected using dlib's new shape_predictor_68_face_landmarks_GTX.dat
model.
Note: Generating keypoints might take more time if compared with detect_faces
method
Options | Description |
---|---|
image | Image in numpy array format |
get_all | Weather to get all facial keypoints or the main (chin, nose, eyes, mouth) |
Ultra Light detection model is what the name says, it a very light weight, accuracy with impressive speed which is pre-trained on 320x240 sized images and only excepts 320x240 sized images but don't worry Ultralight320Detector
detector will do all for you.
import cv2
from face_detectors import Ultralight320Detector
from face_detectors.utils import annotate_image
vid = cv2.VideoCapture(0)
detector = Ultralight320Detector()
while True:
rect, frame = vid.read()
if not rect:
break
bbox = detector.detect_faces(frame)
frame = annotate_image(frame, bbox)
cv2.imshow("Ultra 320 Detection", frame)
cv2.waitKey(1)
Configurable options for Ultralight320Detector detector.
Syntax: Ultralight320Detector(**options)
Options | Description |
---|---|
convert_color | Takes OpenCV COLOR codes to convert the images. Defaults to cv2.COLOR_BGR2RGB |
mean | Metric used to measure the performance of models doing detection tasks. Defaults to [127, 127, 127]. |
confidence | Confidence score is used to refrain from making predictions when it is not above a sufficient threshold. Defaults to 0.5 |
scale | Scales the image for faster output (No need to set this manually, scale will be determined automatically if no value is given) |
cache | It uses same model for all the created sessions. Default is True |
detect_faces(image)
This method will return coordinates for all the detected faces of the given image
Options | Description |
---|---|
image | image in numpy array format |
detect_faces_keypoints(image, get_all=false)
This method will return coordinates for all the detected faces along with their facial keypoints of the given image. Keypoints are detected using dlib's new shape_predictor_68_face_landmarks_GTX.dat
model.
Note: Generating keypoints might take more time if compared with detect_faces
method
Options | Description |
---|---|
image | Image in numpy array format |
get_all | Weather to get all facial keypoints or the main (chin, nose, eyes, mouth) |
Ultra Light detection model is what the name says, it a very light weight, accuracy with impressive speed which is pre-trained on 640x480 sized images and only excepts 640x480 sized images but don't worry Ultralight640Detector
detector will do all for you.
This detector will be more accurate than 320 sized ultra light model (Ultralight320Detector
) but might take a little more time.
import cv2
from face_detectors import Ultralight640Detector
from face_detectors.utils import annotate_image
vid = cv2.VideoCapture(0)
detector = Ultralight640Detector()
while True:
rect, frame = vid.read()
if not rect:
break
bbox = detector.detect_faces(frame)
frame = annotate_image(frame, bbox)
cv2.imshow("Ultra 640 Detection", frame)
cv2.waitKey(1)
Configurable options for Ultralight640Detector detector.
Syntax: Ultralight640Detector(**options)
Options | Description |
---|---|
convert_color | Takes OpenCV COLOR codes to convert the images. Defaults to cv2.COLOR_BGR2RGB |
mean | Metric used to measure the performance of models doing detection tasks. Defaults to [127, 127, 127]. |
confidence | Confidence score is used to refrain from making predictions when it is not above a sufficient threshold. Defaults to 0.5 |
scale | Scales the image for faster output (No need to set this manually, scale will be determined automatically if no value is given) |
cache | It uses same model for all the created sessions. Default is True |
detect_faces(image)
This method will return coordinates for all the detected faces of the given image
Options | Description |
---|---|
image | image in numpy array format |
detect_faces_keypoints(image, get_all=false)
This method will return coordinates for all the detected faces along with their facial keypoints of the given image. Keypoints are detected using dlib's new shape_predictor_68_face_landmarks_GTX.dat
model.
Note: Generating keypoints might take more time if compared with detect_faces
method
Options | Description |
---|---|
image | Image in numpy array format |
get_all | Weather to get all facial keypoints or the main (chin, nose, eyes, mouth) |
Annotates the given image with the payload returned by any of the detectors and returns a well annotated image with boxes and keypoints on the faces.
Configurable options for annotate_image function.
Syntax: annotate_image(**options)
Options | Description |
---|---|
image | Give image for annotation in numpy.Array format |
faces | Payload returned by detector.detect_faces or detector.detect_faces_keypoints |
box_rgb | RGB color for rectangle to be of. Defaults to (100, 0, 255). |
keypoints_rgb | RGB color for keypoints to be of. Defaults to (150, 0, 255). |
width | Width of annotations. Defaults to 2 |
FAQs
Light weight face detector hight level client with multiple detection techniques.
We found that face-detectors demonstrated a healthy version release cadence and project activity because the last version was released less than a year ago. It has 1 open source maintainer collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Research
Security News
Socket’s threat research team has detected six malicious npm packages typosquatting popular libraries to insert SSH backdoors.
Security News
MITRE's 2024 CWE Top 25 highlights critical software vulnerabilities like XSS, SQL Injection, and CSRF, reflecting shifts due to a refined ranking methodology.
Security News
In this segment of the Risky Business podcast, Feross Aboukhadijeh and Patrick Gray discuss the challenges of tracking malware discovered in open source softare.