Skip to content

Latest commit

 

History

History
420 lines (285 loc) · 20.8 KB

README.md

File metadata and controls

420 lines (285 loc) · 20.8 KB

Face Detectors

PyPI Downloads Language grade: Python CodeFactor

This repo contains various types of face detection techniques. All the face detection techniques are fine tunned and optimized out of the box to work the best with any resolution images and takes no time to get started

Key features:

  • Easy to understand and setup
  • Easy to manage
  • Requires very less or no tuning for any resolution image
  • No need to download models, they're automatically maintained
  • Uses ultralight face detection models that is very fast on CPU alone
  • Get very good speed and accuracy on CPU alone
  • All detectors share same parameters and methods, makes it easier to switch and go

Detectors:

  • Hog detector
  • CNN detector
  • Caffemodel detector
  • UltraLight 320 detector
  • UltraLight 640 detector

( More on the way...)

Installation

Use the package manager pip to install face-detectors with the following command:

pip install face-detectors

If you would like to get the latest master or branch from github, you could also:

pip install git+https://github.com/Saadmairaj/face-detectors

Or even select a specific revision (branch/tag/commit):

pip install git+https://github.com/Saadmairaj/face-detectors@master

Similarly, for tag specify tag with @v0.x.x. For example to download tag v0.1.0 from Git use:

pip install git+https://github.com/Saadmairaj/[email protected]

Quick usage

Like said setup and usage is very simple and easy.

  • Import the detector you want,
  • Initialize it,
  • Get predicts

Example

from face_detectors import Ultralight320Detector
from face_detectors.utils import annotate_image

detector = Ultralight320Detector()

image = cv2.imread("image.png")

faces = detector.detect_faces(image)
image = annotate_image(image, faces, width=3)

cv2.imshow("view", image)
cv2.waitKey(100000)

Performance

Every detector has different types of features and can be used for different purposes for example detecting only one face we can use hog with number_of_times_to_upsample=1 or caffemodel, we can also use models but other models like Ultralight models are good for multiple and small face detections.

(The following is test on MacBook Pro 2.3 GHz Quad-Core Intel Core i5 with 8 GB 2133 MHz LPDDR3)

Detector IMAGE 1 (ms) IMAGE 2 (ms) IMAGE 3 (ms) IMAGE 4 (ms)
Caffe Model 0.0334 0.0327 0.0314 0.0344
CNN 0.5216 0.1371 0.4339 0.2264
Hog 0.0970 0.4521 0.0847 0.0548
UltraLight (320px) 0.0128 0.0203 0.0128 0.0149
UltraLight (640px) 0.0347 0.0391 0.0430 0.0384

The below is IMAGE 2 result

View complete comparison

Documentation

Briefly describing face-detectors package that are all the detectors and utility functions.

CaffeModel Detector

Caffemodel is very light weight model that uses less resources to perform detections that is created by caffe (Convolutional Architecture for Fast Feature Embedding).

import cv2
from face_detectors import CaffemodelDetector
from face_detectors.utils import annotate_image

vid = cv2.VideoCapture(0)
detector = CaffemodelDetector()

while True:
    rect, frame = vid.read()
    if not rect:
        break

    bbox = detector.detect_faces(frame)
    frame = annotate_image(frame, bbox)

    cv2.imshow("Caffe Model Detection", frame)

    cv2.waitKey(1)

Configurable options for CaffeModel detector.

Syntax: CaffemodelDetector(**options)

Options Description
convert_color Takes OpenCV COLOR codes to convert the images. Defaults to cv2.COLOR_BGR2RGB
confidence Confidence score is used to refrain from making predictions when it is not above a sufficient threshold. Defaults to 0.5
scale Scales the image for faster output (No need to set this manually, scale will be determined automatically if no value is given)
mean Scalar with mean values which are subtracted from channels. Values are intended to be in (mean-R, mean-G, mean-B) order if image has BGR ordering and swapRB is true. Defaults to (104.0, 177.0, 123.0).
scalefactor Multiplier for images values. Defaults to 1.0.
crop Flag which indicates whether image will be cropped after resize or not. Defaults to False.
swapRB Flag which indicates that swap first and last channels in 3-channel image is necessary. Defaults to False.
transpose Transpose image. Defaults to False.
resize Spatial size for output image. Default is (300, 300)

Useful methods for this detector:

  • detect_faces(image)

    This method will return coordinates for all the detected faces of the given image

    Options Description
    image image in numpy array format
  • detect_faces_keypoints(image, get_all=false)

    This method will return coordinates for all the detected faces along with their facial keypoints of the given image. Keypoints are detected using dlib's new shape_predictor_68_face_landmarks_GTX.dat` model.

    Note: Generating keypoints might take more time if compared with detect_faces method

    Options Description
    image Image in numpy array format
    get_all Weather to get all facial keypoints or the main (chin, nose, eyes, mouth)

CNN Detector

CNN (Convolutional Neural Network) might not be a light weight model but it is good at detecting faces from all angles. This detector is a hight level wrapper around dlib::cnn_face_detection_model_v1 that is fine tuned to improve overall performance and accuracy.

import cv2
from face_detectors import CNNDetector
from face_detectors.utils import annotate_image

vid = cv2.VideoCapture(0)
detector = CNNDetector()

while True:
    rect, frame = vid.read()
    if not rect:
        break

    bbox = detector.detect_faces(frame)
    frame = annotate_image(frame, bbox)

    cv2.imshow("CNN Detection", frame)

    cv2.waitKey(1)

Configurable options for CNNDetector detector.

Syntax: CNNDetector(**options)

Options Description
convert_color Takes OpenCV COLOR codes to convert the images. Defaults to cv2.COLOR_BGR2RGB
number_of_times_to_upsample Up samples the image number_of_times_to_upsample before running the basic detector. By default is 1.
confidence Confidence score is used to refrain from making predictions when it is not above a sufficient threshold. Defaults to 0.5
scale Scales the image for faster output (No need to set this manually, scale will be determined automatically if no value is given)
  • detect_faces(image)

    This method will return coordinates for all the detected faces of the given image

    Options Description
    image image in numpy array format
  • detect_faces_keypoints(image, get_all=false)

    This method will return coordinates for all the detected faces along with their facial keypoints of the given image. Keypoints are detected using dlib's new shape_predictor_68_face_landmarks_GTX.dat model.

    Note: Generating keypoints might take more time if compared with detect_faces method

    Options Description
    image Image in numpy array format
    get_all Weather to get all facial keypoints or the main (chin, nose, eyes, mouth)

Hog Detector

This detector uses Histogram of Oriented Gradients (HOG) and Linear SVM classifier for face detection. It is also combined with an image pyramid and a sliding window detection scheme. HogDetector is a high level client over dlib's hog face detector and is fine tuned to make it more optimized in both speed and accuracy.

If you want to detect faster with HogDetector and don't care about number of detections then set number_of_times_to_upsample=1 in the options, it will detect less fasces in less time, mainly used for real time one face detection.

import cv2
from face_detectors import HogDetector
from face_detectors.utils import annotate_image

vid = cv2.VideoCapture(0)
detector = HogDetector()

while True:
    rect, frame = vid.read()
    if not rect:
        break

    bbox = detector.detect_faces(frame)
    frame = annotate_image(frame, bbox)

    cv2.imshow("Hog Detection", frame)

    cv2.waitKey(1)

Configurable options for HogDetector detector.

Syntax: HogDetector(**options)

Options Description
convert_color Takes OpenCV COLOR codes to convert the images. Defaults to cv2.COLOR_BGR2RGB
number_of_times_to_upsample Up samples the image number_of_times_to_upsample before running the basic detector. By default is 2.
confidence Confidence score is used to refrain from making predictions when it is not above a sufficient threshold. Defaults to 0.5
scale Scales the image for faster output (No need to set this manually, scale will be determined automatically if no value is given)
  • detect_faces(image)

    This method will return coordinates for all the detected faces of the given image

    Options Description
    image image in numpy array format
  • detect_faces_keypoints(image, get_all=false)

    This method will return coordinates for all the detected faces along with their facial keypoints of the given image. Keypoints are detected using dlib's new shape_predictor_68_face_landmarks_GTX.dat model.

    Note: Generating keypoints might take more time if compared with detect_faces method

    Options Description
    image Image in numpy array format
    get_all Weather to get all facial keypoints or the main (chin, nose, eyes, mouth)

Ultra Light Detection (320px)

Ultra Light detection model is what the name says, it a very light weight, accuracy with impressive speed which is pre-trained on 320x240 sized images and only excepts 320x240 sized images but don't worry Ultralight320Detector detector will do all for you.

import cv2
from face_detectors import Ultralight320Detector
from face_detectors.utils import annotate_image

vid = cv2.VideoCapture(0)
detector = Ultralight320Detector()

while True:
    rect, frame = vid.read()
    if not rect:
        break

    bbox = detector.detect_faces(frame)
    frame = annotate_image(frame, bbox)

    cv2.imshow("Ultra 320 Detection", frame)

    cv2.waitKey(1)

Configurable options for Ultralight320Detector detector.

Syntax: Ultralight320Detector(**options)

Options Description
convert_color Takes OpenCV COLOR codes to convert the images. Defaults to cv2.COLOR_BGR2RGB
mean Metric used to measure the performance of models doing detection tasks. Defaults to [127, 127, 127].
confidence Confidence score is used to refrain from making predictions when it is not above a sufficient threshold. Defaults to 0.5
scale Scales the image for faster output (No need to set this manually, scale will be determined automatically if no value is given)
cache It uses same model for all the created sessions. Default is True
  • detect_faces(image)

    This method will return coordinates for all the detected faces of the given image

    Options Description
    image image in numpy array format
  • detect_faces_keypoints(image, get_all=false)

    This method will return coordinates for all the detected faces along with their facial keypoints of the given image. Keypoints are detected using dlib's new shape_predictor_68_face_landmarks_GTX.dat model.

    Note: Generating keypoints might take more time if compared with detect_faces method

    Options Description
    image Image in numpy array format
    get_all Weather to get all facial keypoints or the main (chin, nose, eyes, mouth)

Ultra Light Detection (640px)

Ultra Light detection model is what the name says, it a very light weight, accuracy with impressive speed which is pre-trained on 640x480 sized images and only excepts 640x480 sized images but don't worry Ultralight640Detector detector will do all for you.

This detector will be more accurate than 320 sized ultra light model (Ultralight320Detector) but might take a little more time.

import cv2
from face_detectors import Ultralight640Detector
from face_detectors.utils import annotate_image

vid = cv2.VideoCapture(0)
detector = Ultralight640Detector()

while True:
    rect, frame = vid.read()
    if not rect:
        break

    bbox = detector.detect_faces(frame)
    frame = annotate_image(frame, bbox)

    cv2.imshow("Ultra 640 Detection", frame)

    cv2.waitKey(1)

Configurable options for Ultralight640Detector detector.

Syntax: Ultralight640Detector(**options)

Options Description
convert_color Takes OpenCV COLOR codes to convert the images. Defaults to cv2.COLOR_BGR2RGB
mean Metric used to measure the performance of models doing detection tasks. Defaults to [127, 127, 127].
confidence Confidence score is used to refrain from making predictions when it is not above a sufficient threshold. Defaults to 0.5
scale Scales the image for faster output (No need to set this manually, scale will be determined automatically if no value is given)
cache It uses same model for all the created sessions. Default is True
  • detect_faces(image)

    This method will return coordinates for all the detected faces of the given image

    Options Description
    image image in numpy array format
  • detect_faces_keypoints(image, get_all=false)

    This method will return coordinates for all the detected faces along with their facial keypoints of the given image. Keypoints are detected using dlib's new shape_predictor_68_face_landmarks_GTX.dat model.

    Note: Generating keypoints might take more time if compared with detect_faces method

    Options Description
    image Image in numpy array format
    get_all Weather to get all facial keypoints or the main (chin, nose, eyes, mouth)

Annotate Image Function

Annotates the given image with the payload returned by any of the detectors and returns a well annotated image with boxes and keypoints on the faces.

Configurable options for annotate_image function.

Syntax: annotate_image(**options)

Options Description
image Give image for annotation in numpy.Array format
faces Payload returned by detector.detect_faces or detector.detect_faces_keypoints
box_rgb RGB color for rectangle to be of. Defaults to (100, 0, 255).
keypoints_rgb RGB color for keypoints to be of. Defaults to (150, 0, 255).
width Width of annotations. Defaults to 2