( 1 of 512 )
United States Patent Application 20110317872
Kind Code A1
Free; Robert Mikio December 29, 2011
Low Threshold Face Recognition
Abstract
Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, are disclosed for reducing the impact of lighting conditions and biometric distortions, while providing a low-computation solution for reasonably effective (low threshold) face recognition. In one aspect, the methods include processing a captured image of a face of a user seeking to access a resource by conforming a subset of the captured face image to a reference model. The reference model corresponds to a high information portion of human faces. The methods further include comparing the processed captured image to at least one target profile corresponding to a user associated with the resource, and selectively recognizing the user seeking access to the resource based on a result of said comparing.
Inventors: Free; Robert Mikio; (San Jose, CA)
Assignee: APPLE INC.
Cupertino
CA
Serial No.: 826581
Series Code: 12
Filed: June 29, 2010
Current U.S. Class: 382/103; 382/118
Class at Publication: 382/103; 382/118
International Class: G06K 9/00 20060101 G06K009/00
Claims
1. A method performed by an image processor, the method comprising: processing a captured image of a face of a user seeking to access a resource by conforming a subset of the captured face image to a reference model, the reference model corresponding to a high information portion of human faces; comparing the processed captured image to at least one target profile corresponding to a user associated with the resource; and selectively recognizing the user seeking access to the resource based on a result of said comparing.
2. The method of claim 1, wherein the high information portion includes eyes and a mouth.
3. The method of claim 2, wherein the high information portion further includes a tip of a nose.
4. The method of claim 2, wherein said processing the captured image comprises detecting a face within the captured image by identifying the eyes in an upper one third of the captured image and the mouth in the lower third of the captured image.
5. The method of claim 4, wherein the reference model includes a reference image of a face, and wherein said processing the captured image further comprises matching the eyes of the detected face with eyes of the face in the reference image to obtain a normalized image of the detected face.
6. The method of claim 5, wherein said processing the captured image further comprises vertically scaling a distance between an eyes-line and the mouth of the detected face to equal a corresponding distance for the face in the reference image in order to obtain the normalized image of the detected face.
7. The method of claim 6, wherein said processing the captured image further comprises matching the mouth of the detected face to the mouth of the face in the reference image in order to obtain the normalized image of the detected face.
8. The method of claim 5, wherein said comparing the processed captured image comprises: obtaining a difference image of the detected face by subtracting the normalized image of the detected face from a normalized image of a target face associated with a target profile; and calculating scores of respective pixels of the difference image based on a weight defined according to proximity of the respective pixels to high information portions of the human faces.
9. The method of claim 8, where the weight decreases with a distance from the high information portions of the human faces.
10. The method of claim 9, wherein the weight decreases continuously with the distance from the high information portions of the human faces.
11. The method of claim 9, wherein the weight decreases discretely with the distance from the high information portions of the human faces.
12. The method of claim 9, wherein the weight decreases from a maximum weight value at a mouth-level to a minimum value at an eyes-line.
13. The method of claim 1, wherein said selectively recognizing the user comprises presenting to the user a predetermined indication according to a user's profile.
14. The method of claim 1, wherein the resource represents an appliance, the method further comprising capturing the image using an image capture device of the appliance.
15. The method of claim 14, wherein said selectively recognizing the user comprises turning on a display of the appliance, wherein the display had been off prior to said comparing.
16. The method of claim 14, wherein said processing the captured image comprises: applying an orange-distance filter to the captured image; and segmenting a skin-tone orange portion of the orange-distance filtered image to represent a likely presence of a face in front of the image capture device.
17. The method of claim 16, wherein said processing the captured image further comprises determining changes in area and in location of the skin-tone orange portion of the captured image relative to a previously captured image to represent likely movement of the face in front of the image capture device.
18. The method of claim 17, wherein said processing the captured image further comprises detecting a face within the skin-tone orange portion of the orange-distance filtered image when the determined changes are less than predetermined respective variations.
19. An appliance comprising: a data storage device configured to store profiles of users associated with the appliance; an image capture device configured to acquire color frames; one or more data processors configured to perform operations including: apply an orange-distance filter to a frame acquired by the image capture device; determine respective changes in area and location of a skin-tone orange portion of the acquired frame relative to a previously acquired frame; infer, based on the determined changes, a presence of a face substantially at rest when the frame was acquired; detect a face corresponding to the skin-tone orange portion of the acquired frame in response to the inference, the detection including finding eyes and a mouth within the skin-tone orange portion; normalize the detected face based on locations of eyes and a mouth of a face in a reference image; analyze weighted differences between normalized target faces and the normalized detected face, the analysis including weighting portions of a face based on information content corresponding to the portions, and the target faces being associated with respective users of the appliance; match the face detected in the acquired frame with one of the target faces based on a result of the analysis; and acknowledge the match of the detected face in accordance with a profile stored on the data storage device and associated with the matched user of the appliance.
20. The appliance of claim 19, wherein the data storage device is further configured to store: rules for analyzing the weighted differences including weighting rules and scoring rules; and rules for matching the detected face against target faces.
Description
BACKGROUND
[0001] This specification relates to low threshold face recognition, e.g., a face recognition system that can tolerate a certain level of false positives in making face recognition determinations.
[0002] Most face recognition systems fall into one of two categories. A first category system tends to be robust and can tackle various lighting conditions, orientations, scale and the like, and tends to be computationally expensive. A second category system is specialized for security-type applications and can work under controlled lighting conditions. Adopting the first category systems for face recognition on consumer operated portable appliances that are equipped with a camera would unnecessarily use an appliance's computing resources and drain its power. Moreover, as the consumer portable appliances tend to be used both indoor and outdoor, the second category systems for face recognition may be ineffective. Such ineffectiveness may be further exacerbated by the proximity of the user to the camera, i.e., small changes in distance to and tilt of the appliance's camera dramatically distort features, making traditional biometrics used in security-type face recognition ineffective.
SUMMARY
[0003] This specification describes technologies relating to reducing the impact of lighting conditions and biometric distortions, while providing a low-computation solution for reasonably effective (low threshold) face recognition that can be implemented on camera-equipped consumer portable appliances.
[0004] In general, one aspect of the subject matter described in this specification can be implemented in methods performed by an image processor that include the actions of processing a captured image of a face of a user seeking to access a resource by conforming a subset of the captured face image to a reference model. The reference model corresponds to a high information portion of human faces. The methods further include comparing the processed captured image to at least one target profile corresponding to a user associated with the resource, and selectively recognizing the user seeking access to the resource based on a result of said comparing.
[0005] These and other implementations can include one or more of the following features. In some cases, the high information portion includes eyes and a mouth. In some other cases, the high information portion further includes a tip of a nose. Processing the captured image can include detecting a face within the captured image by identifying the eyes in an upper one third of the captured image and the mouth in the lower third of the captured image. The reference model includes a reference image of a face, and processing the captured image further can include matching the eyes of the detected face with eyes of the face in the reference image to obtain a normalized image of the detected face. Additionally, processing the captured image can further include vertically scaling a distance between an eyes-line and the mouth of the detected face to equal a corresponding distance for the face in the reference image in order to obtain the normalized image of the detected face. In addition, processing the captured image can further include matching the mouth of the detected face to the mouth of the face in the reference image in order to obtain the normalized image of the detected face.
[0006] In some implementations, comparing the processed captured image can include obtaining a difference image of the detected face by subtracting the normalized image of the detected face from a normalized image of a target face associated with a target profile. Comparing can further include calculating scores of respective pixels of the difference image based on a weight defined according to proximity of the respective pixels to high information portions of the human faces. The weight decreases with a distance from the high information portions of the human faces. For example, the weight decreases continuously with the distance from the high information portions of the human faces. As another example, the weight decreases discretely with the distance from the high information portions of the human faces. As yet another example, the weight decreases from a maximum weight value at a mouth-level to a minimum value at an eyes-line.
[0007] In some implementations, selectively recognizing the user can include presenting to the user a predetermined indication according to a user's profile. The resource can represent an appliance, and the methods can further include capturing the image using an image capture device of the appliance. Selectively recognizing the user can include turning on a display of the appliance, if the display had been off prior to the comparison.
[0008] In some implementations, processing the captured image can include applying an orange-distance filter to the captured image, and segmenting a skin-tone orange portion of the orange-distance filtered image to represent a likely presence of a face in front of the image capture device. Processing the captured image can further include determining changes in area and in location of the skin-tone orange portion of the captured image relative to a previously captured image to represent likely movement of the face in front of the image capture device. Also, processing the captured image further can include detecting a face within the skin-tone orange portion of the orange-distance filtered image when the determined changes are less than predetermined respective variations.
[0009] According to another aspect, the described subject matter can also be implemented in an appliance including a data storage device configured to store profiles of users associated with the appliance. The appliance further includes an image capture device configured to acquire color frames. Further, the appliance includes one or more data processors configured to apply an orange-distance filter to a frame acquired by the image capture device. The one or more data processors are further configured to determine respective changes in area and location of a skin-tone orange portion of the acquired frame relative to a previously acquired frame, and to infer, based on the determined changes, a presence of a face substantially at rest when the frame was acquired. Further, the one or more data processors is configured to detect a face corresponding to the skin-tone orange portion of the acquired frame in response to the inference, the detection including finding eyes and a mouth within the skin-tone orange portion. Furthermore, the one or more data processors are configured to normalize the detected face based on locations of eyes and a mouth of a face in a reference image. In addition, the one or more data processors are configured to analyze weighted differences between normalized target faces and the normalized detected face. The analysis includes weighting portions of a face based on information content corresponding to the portions. The target faces are associated with respective users of the appliance. Additionally, the one or more data processors are configured to match the face detected in the acquired frame with one of the target faces based on a result of the analysis, and to acknowledge the match of the detected face in accordance with a profile stored on the data storage device and associated with the matched user of the appliance.
[0010] These and other implementations can include one or more of the following features. The data storage device is further configured to store rules for analyzing the weighted differences including weighting rules and scoring rules, and rules for matching the detected face against target faces.
[0011] Particular implementations of the subject matter described in this specification can be configured to realize one or more of the following potential advantages. The techniques and systems disclosed in this specification can reduce the impact of lighting and emphasize skin variance. By acquiring images with the appliance's own image capture device, the approximate location and orientation of face features can be pre-assumed and can avoid the overhead of other face recognition systems. The disclosed methods can ignore face biometrics, and rather use feature locations to normalize an image of a test face. Further, the face recognition techniques are based on a simple, weighted difference map, rather than traditional (and computationally expensive) correlation matching.
[0012] The details of one or more implementations of the subject matter described in this specification are set forth in the accompanying drawings and the description below. Other features, aspects, and advantages of the subject matter will become apparent from the description, the drawings, and the claims.