FaceAR Glossary

AR Technologies

FRX (Face tracking)

The technology to detect and track the presence of a human face in a digital video frame to enable FaceAR camera experiences.

Background separation

Neural network that separates a user on the foreground from the background in a sequence of video frames to remove the background or replace it with another image. As a standalone feature, the background segmentation is used in video conferencing, live streaming and video communication apps. It also can be used as part of face filter together with facial animation in entertainment apps.

Bokeh effect

The algorithm that separates a person on the foreground and blurs the background as in photography.

Skin segmentation

Neural network trained to recognize and segment the skin area for recoloring tasks, e.g. make the skin tone lighter or darker.

Skin smoothing

Neural network trained to segment and blur the skin on the face to create a retouch effect similar to professional photography.

Neck smoothing

Neural network trained to segment and blur the neck area for beautification purposes.

Hair Segmentation

Neural network to detect and segment the image into hair, face and background to allow for real-time hair modification such as change the hair color.

Eye segmentation

Neural network to detect and segment the eye into iris, pupil and eyeball for eye recoloring and virtual contact lens try on.

Lips segmentation

Neural network to detect and segment the lips on the users face for virtual lipstic try on effect.

Full body segmentation

Neural network trained to recognize the human body in full length and separate it from the background in images and videos.

Acne removal (tap)

Acne removal is an algorithm to remove acne in photos by single taps on the selected area.

Acne removal (auto)

The algorithm to automatically remove acne in photos.

Eye bag removal

The algorithm segments and lightens the area under the eyes in photos for beautification purposes.

Hair strands painting

The algorithms to change the hair color with several colors applied simultaneously, e.g. for strands highlight or coloring.

Features

2+ faces detection (Multi-face)

Algorithm allowing for masks application to several people simultaneously for more engaging group AR experiences. For the quality user experience, we generally don’t recommend supporting more than 3 faces on mobile devices due to limited computing capabilities.

Pulse (Heart rate)

Algorithm analyses fine patterns of the facial areas and their color variations within time to detect pulse frequency in real-time.

Ruler (Ruler to phone)

Algorithm analyses face area size to estimate distance from user's face to camera in real-time.

Mask on a picture from Camera Roll

Mask application on pre-recorded images the user uploads from the Camera Roll.

Mask on video from Camera Roll

Mask application on pre-recorded videos the user uploads from the Camera Roll.

Post-processing effects

Graphical camera effects and animations applied on pre-recorded videos.

Continuous photo editing

AR effect is processed in real-time on the image. E.g. beautification slider to control the face modification or "Before/After" slider.

Touches

Small AR scenarios enabled thought the user touches on the screen. AR objects or camera effects can change color and behaviour. Applied in FaceAR games or interactive face filters to increase engagement.

Trigger

Small AR scenarios enabled through user facial expressions. The user can interact with effects or call them opening mouth, smiling, raising eyebrows or frowning. Applied in FaceAR games or interactive face filters to increase engagement.

SFX

Sound effects support in FaceAR experiences, e.g. add music to filters.

Voice changer

The 3rd party software allowing to change the tone of the user's voice, e.g. make the user talk like a robot, kid, male/female, etc.

Glasses detection

The algorithm detects if the user wears glasses and removes them in a virtual mask. Applied in glasses try-on for convenient frame choice and face filters for AR glasses would not overlay the real glasses.

Graphical technologies

Face beautification

The AR filter based on face tracking which automatically retouches the appearance applying skin smooth, morphing, eye makeup, teeth whitening, eye flare and LUT effect.

Morphing

Changing the size and proportions of the face, e.g. slim down the cheeks, nose or modify them for fun masks.

Skinned mesh animation

AR models look not static but moving, animated and transforming.

Physically-based rendering

AR models behave like the real objects in the flow of real-world light and physics, e.g. support gravity or mirror the light with the camera rotates and user tilts.

LUT post-processing

Real-time or offline color correction of pre-recorded images, e.g. Instagram-like filters.

Texture sequences

The digital representation of the surface of an AR object providing the sophisticated and life-like object representation.

Video textures

Infuse a static image with dynamic qualities and explicit action to achieve an enhanced look and feel of the FaceAR video experience.

Action units

The fundamental actions of individual muscles or groups of muscles of the face that enable the AR mask to support user facial expressions, e.g. in emojis, avatars or full-face AR masks.

Last updated on