FRX (Face tracking)
The technology to detect and track the presence of a human face in a digital video frame to enable FaceAR camera experiences.
Neural network that separates a user on the foreground from the background in a sequence of video frames to remove the background or replace it with another image. As a standalone feature, the background segmentation is used in video conferencing, live streaming and video communication apps. It also can be used as part of face filter together with facial animation in entertainment apps.
The algorithm that separates a person on the foreground and blurs the background as in photography.
Neural network trained to recognize and segment the skin area for recoloring tasks, e.g. make the skin tone lighter or darker.
Neural network trained to segment and blur the skin on the face to create a retouch effect similar to professional photography.
Neural network trained to segment and blur the neck area for beautification purposes.
Neural network to detect and segment the image into hair, face and background to allow for real-time hair modification such as change the hair color.
Neural network to detect and segment the eye into iris, pupil and eyeball for eye recoloring and virtual contact lens try on.
Neural network to detect and segment the lips on the users face for virtual lipstic try on effect.
Full body segmentation
Neural network trained to recognize the human body in full length and separate it from the background in images and videos.
Acne removal (tap)
Acne removal is an algorithm to remove acne in photos by single taps on the selected area.
Acne removal (auto)
The algorithm to automatically remove acne in photos.
Eye bag removal
The algorithm segments and lightens the area under the eyes in photos for beautification purposes.
Hair strands painting
The algorithms to change the hair color with several colors applied simultaneously, e.g. for strands highlight or coloring.
2+ faces detection (Multi-face)
Algorithm allowing for masks application to several people simultaneously for more engaging group AR experiences. For the quality user experience, we generally don’t recommend supporting more than 3 faces on mobile devices due to limited computing capabilities.
Pulse (Heart rate)
Algorithm analyses fine patterns of the facial areas and their color variations within time to detect pulse frequency in real-time.
Ruler (Distance to phone)
Algorithm analyses face area size to estimate distance from user's face to camera in real-time.
Text Texture (on mask)
Allows to write text as texture on any 2D or 3D model in the effect..
Mask on a picture from Camera Roll
Mask application on pre-recorded images the user uploads from the Camera Roll.
Mask on video from Camera Roll
Mask application on pre-recorded videos the user uploads from the Camera Roll.
Graphical camera effects and animations applied on pre-recorded videos.
Continuous photo editing
AR effect is processed in real-time on the image. E.g. beautification slider to control the face modification or "Before/After" slider.
Small AR scenarios enabled thought the user touches on the screen. AR objects or camera effects can change color and behaviour. Applied in FaceAR games or interactive face filters to increase engagement.
Small AR scenarios enabled through user facial expressions. The user can interact with effects or call them opening mouth, smiling, raising eyebrows or frowning. Applied in FaceAR games or interactive face filters to increase engagement.
Sound effects support in FaceAR experiences, e.g. add music to filters.
The 3rd party software allowing to change the tone of the user's voice, e.g. make the user talk like a robot, kid, male/female, etc.
The algorithm detects if the user wears glasses and removes them in a virtual mask. Applied in glasses try-on for convenient frame choice and face filters for AR glasses would not overlay the real glasses.
The AR filter based on face tracking which automatically retouches the appearance applying skin smooth, morphing, eye makeup, teeth whitening, eye flare and LUT effect.
Changing the size and proportions of the face, e.g. slim down the cheeks, nose or modify them for fun masks.
Skinned mesh animation
AR models look not static but moving, animated and transforming.
AR models behave like the real objects in the flow of real-world light and physics, e.g. support gravity or mirror the light with the camera rotates and user tilts.
Real-time or offline color correction of pre-recorded images, e.g. Instagram-like filters.
The digital representation of the surface of an AR object providing the sophisticated and life-like object representation.
Infuse a static image with dynamic qualities and explicit action to achieve an enhanced look and feel of the FaceAR video experience.
The fundamental actions of individual muscles or groups of muscles of the face that enable the AR mask to support user facial expressions, e.g. in emojis, avatars or full-face AR masks.