WO2007101269A1 - Approaches and apparatus for eye detection in a digital image - Google Patents
Approaches and apparatus for eye detection in a digital image Download PDFInfo
- Publication number
- WO2007101269A1 WO2007101269A1 PCT/US2007/063029 US2007063029W WO2007101269A1 WO 2007101269 A1 WO2007101269 A1 WO 2007101269A1 US 2007063029 W US2007063029 W US 2007063029W WO 2007101269 A1 WO2007101269 A1 WO 2007101269A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- eye
- profile
- candidate
- selector
- candidates
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/193—Preprocessing; Feature extraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
Definitions
- the invention is an approach and apparatus for localizing eyes of a human in a digital image to be processed for iris recognition.
- FIG. 1 a is a diagram of an overall illustrative structure of an eye finding system
- Figure Ib is a diagram with a group structure of the eye finding system
- Figure 2a is a diagram of an approach for determining a profile of an eye as provided by a measure of profile metrics
- Figure 2b is a diagram of a structure of a profiler
- Figures 3a, 3b and 3c show an image of a selected eye, a pupil image 42 and a binary image 43 of a pupil, respectively;
- Figure 4 is a diagram of an overall iris recognition system
- Figure 5 shows a diagram of a kernel having a box representing the diameter of a pupil and a box representing a pupil reflection
- Figure 6 shows a box representing a region of interest having areas which are relatively light, dark and lighter than the relatively light area situated in the dark area representing a pupil model
- Figure 7 is a histogram of the contrast or intensity values of areas of Figure 6;
- Figure 8 is like Figure 6 but does not have a lighter (modeling a typical reflection) than the relatively light area situated in the dark area representing a pupil;
- Figure 9 is a scale of marks representing pixels of a region of interest ranked according to lightness and darkness.
- Figures 10a and 10b relate to eye finding using reflection and/or non-reflection measures.
- Eye detection may be the first step toward building a reliable automated iris recognition system in a natural context. Some iris recognition systems rely heavily on predetermined eye locations to properly zoom on the input eye prior to iris segmentation. In addition to biometrics, eye detection (also known as eye finding or eye localization) may support other new technology areas such as eye tracking and human computer interaction or driver drowsiness monitoring systems. Eye detection may serve social learning to identify eye directions like pointing gesture using eye directions. The present approach and apparatus may be used for finding eyes within a digital image. A local contrast change profiling may be in eye finding.
- the present approach may be based on a system engineering approach to construct the illumination scheme during eye image acquisition to shine the surface of the pupil surface and result into a high reflection point preferably within the pupil of the eye image or close to the pupil of the eye image.
- This specular reflection point may be used as a reference for an eye search in the digital image.
- the search may be limited to a simplified localization scheme of the highest value pixels associated with these specular reflection pixels and analyze the very local features surrounding these hot spots to confirm an eye profile.
- the eye finding approach may be implemented as a cascade process that divides the local features of an eye into a primary feature of contrast profile associated with high pixel values, depict only potential eye pairs within a specified range, and then test the resulting valid pairs against a feature vector of two or more variables that includes a predefined regular shape fitting with multiple curve fitting measures.
- the present approach may be for quickly and robustly localizing the eyes of a human eye in close-up or face images.
- the approach is based on sensing reflection points within the pupil region as a precursor to the analysis.
- the approach is formulated to work for cases where reflection is not present within the pupil.
- the technical approach locates eyes whether there is or there is no reflection.
- the detection may hence be simplified to search for these specific reflection points surrounded with dark contrast that represent the pupil.
- the region of interest centered at these potential locations may be processed to find an eye profile. Two valid eyes may be extracted that are within an expected range of eye positioning.
- the approach for finding eyes decomposes into the following steps. There may be a contrast filter to detect specular reflection pixels.
- the eye pair may be defined as a valid pair if the two potential eyes are spaced within a predefined range.
- An adaptive threshold may be applied to detect a central blob.
- There may be curve fitting of the blob boundaries into a shape. Curve fitness and shape area coverage of the blob surface may be measured for validation.
- the approach described here may be part of a preprocessing technique used to locate the eyes of a human in a digital image to be processed for iris recognition.
- Eye detection may be the first stage for any automated iris recognition analysis system and may be critical for consistent iris segmentation.
- Several eye detection algorithms may be developed as a basis for face detection. Eye finding approaches may be classified into several categories based upon knowledge based approaches, template matching, and eye socket corner detection.
- the present approach may address real-time operational requirements.
- One solution may be to cascade localized features of the eye to speed up the process.
- Appearance based approaches using Eigenspace supervised classification technique may be used to capture the most representative variability of eye appearance.
- Template matching can be regarded as a brute force approach which may include constructing a kernel that is representative of a typical eye socket and convolve the image with the kernel template to identify the highest values of the convolution indicating a match of the eye the identified locations.
- Knowledge based approaches may be based on specific rules that are captured by an expert that discriminate the eye local features from any other features. These sets of rules may then be tested against virtually all possible combination to identify the eye locations.
- the present approach may provide for quickly and robustly localizing the eyes of a human eye in close-up or face images.
- the approach may be based on sensing reflection points within the pupil region as a precursor to the analysis.
- the approach may be also based on sensing the pupil profile in case of no reflection. If reflection is present, the detection may then be simplified to search for these specific reflection points surrounded with dark contrast that represent the pupil.
- the region of interest centered at these potential locations may then be processed to find an eye profile. Two valid pairs may be extracted that are within an expected range of eye positioning.
- the present approach for finding eyes may decompose into the following.
- the eye may be illuminated to generate a reflection reference point on the pupil surface.
- the captured wide-field image may be filtered using reflection detection contrast changes to find potential eye locations.
- the local contrast change between the central point and its surrounding pixels may be computed and results may be prioritized to extract valid eye pair with maximum local contrast change.
- the eye pair may be defined as a valid pair if the two potential eyes are spaced within a predefined range.
- an adaptive threshold may be executed on a cropped image of the central region of the potential eye to extract a blob of the pupil.
- a single blob may be depicted based on size, its distance to the central point of the cropped image, and how good it fits to a predefined fitting model (e.g., circular shape).
- a predefined fitting model e.g., circular shape
- the blob edges may be fitted into the pupil fitting model. Curve fitness and model shape area coverage of the blob surface may be measured for validation.
- a preprocessing technique may locate the eyes of a human in a digital image to be processed for iris recognition.
- An overall illustrative structure of an eye finding system 10 is shown in Figure Ia.
- the system engineering for eye illumination is not necessarily shown in this system.
- a digital camera may be used for acquiring 9 images of candidates.
- the image may be put through a contrast filter 11 to detect relevant high contrast areas of the image.
- There may a prioritization (i.e., ranking) of the significant spots in the image in block 12.
- N candidates may be extracted in block 13.
- a candidate may have a coordinate Ci(x,y).
- An output of block 13 may go to block 14 where a new first eye may be selected from the candidate list. From block 14, the eye image candidate may go to a block 15 for a measurement of profile metrics.
- a profile of the eye may go to a diamond 17 where a determination of the validity of the profile is made. If the profile is not valid, then that first eye may be deleted from the list at block 17. Then at a diamond 18, a count is checked to note whether it is greater than zero. If not, the then this approach is stopped at place 19. If so, then a new first eye may be selected at block 14 from the list from block 13. The profile metrics of this new first eye may be measured at block 15 and passed on to diamond 16 to determine the validity of the profile. If the profile is valid, then the selected first eye may go to place 20, and a second eye is selected at block 21 from the list of candidates from block 13 having a coordinate c 2 (x,y).
- the spacing of the first and second eyes may be determined at block 22 as D(ci(x,y),c 2 (x,y)). The spacing may be checked to see whether it is within an appropriate range at a diamond 23. If not, then the second eye may be deleted from the list at block 24. If so, then metrics of the profile of the second eye may be measured at block 25. The profile metric may be forwarded to a diamond 26 for a determination of the validity of the profile. If the profile is not valid, then the second eye may be deleted from the list at block 24 and at diamond 27, a question of whether the count is greater than zero. If so, then another second eye may be selected from the list at block 21, and the approach via the blocks 21, 23 and 25, and diamonds 23, 26 and 27 may be repeated. If not, then the approach for the second eye may end at place 19. If the profile is valid at diamond 26, then the selected second eye may go to place 20.
- a higher level approach to system 10 in Figure Ia may include an output of the contrast filtering 11 going a select candidate block 111.
- An output from block 111 may go to a validate profile block 112.
- Outputs from block 112 may go to a select candidate block 114 and a result block 20, or eliminate candidate block 113.
- An output of block 113 may go to the select candidate block 111 and/or to the stop place 19.
- An output from block 114 may go to a validate pair block 115.
- Block 115 may provide an output to a validate profile block 116.
- Outputs from block 116 may go to an eliminate candidate 117 and/or to the result block 20.
- Outputs of block 117 may go the select candidate 114 and the stop place 19.
- the processing in system 10 may be digital, although it may be analog, or it may be partially digital and analog.
- Figure Ib is a diagram with a group structure of the eye finding system 10.
- the corresponding components (according to reference numbers) of Figure Ia may have additional description.
- the candidates noted herein may refer to various images of eyes.
- a camera 9 may be connected to the contrast filter 11.
- An output of the filter 11 may go to a ranking mechanism 12, which in turn is connected to the candidates extractor 13.
- the output of extractor 13 may go to a candidate determiner 14 for selecting a new first candidate.
- Mechanism 12, extractor 13 and determiner 14 constitute a candidate selector 111.
- An output of determiner 14 may go to a metric profiler 15 which in turn has an output connected to a profile evaluator 16.
- Profiler 15 and evaluator 16 may constitute profile validator 112.
- Outputs of evaluator 16 may go to candidate determiner 21, resulter 20 and candidate remover 17.
- Remover may have an output that goes to a counter 18.
- Candidate remover 17 and counter 18 may constitute a candidate eliminator 113. If counter 18 has a count of greater than zero, an output may go to the candidate determiner 14 for selection of a new candidate. If the output is not greater than zero, then an output may go to the stopper 19.
- a candidate determiner 21 for selecting a 2nd candidate may have an output to a space measurer 22.
- the candidate Space measurer 22 may have an output to the range indicator 23 which may indicate whether the two candidates are at an appropriate distance from each other for validation.
- Measure 22 and indicator 23 may constitute a pair validator 115.
- Candidate determiner 21 and previously noted ranking mechanism 12 and candidates extractor 13 may constitute a candidate selector 114. If the pair of candidates is valid then an output from validator 115 may go to a profiler 25, or if the pair is not valid then an output from validator 115 may go to a candidate remover 24. An output of profiler 25 may go to a profile evaluator 26 which may determine whether the profile of the second candidate is valid or not.
- an output of evaluator 26 may provide second candidate information to the resulter 20. If invalid, then an output of evaluator 26 may provide a signal to the candidate remover 24. Profiler 25 and profiler evaluator 26 may constitute a profile validator 116. An output of candidate remover may go to a counter 27. If the counter 27 indicates a value greater than zero then an output may go to the candidate determiner 21 for selecting a second candidate. If the counter 27 indicates a value not greater than zero, then an output may go to a stopper 19. The candidate remover 24 and counter 27 may constitute a candidate eliminator 117.
- Figure 2a shows the approach for determining a profile of an eye as provided by a measure profile metrics or eye profiling block 15, 25.
- An image 41 of a selected eye ( Figure 3a) may go to an extract pupil region block 31.
- the block dimension is determined based on the maximum expected value of the pupil diameter.
- a maximum pupil input 44 may be provided to block 31.
- An output from block may be a pupil image 42 ( Figure 3b) which goes to an adaptive thresholding block 32.
- a percent input 45 may be provided to block 32.
- the pixel distribution to compute the intensity histogram may be provided to block 32.
- An output of block 32 may be a binary image 43 ( Figure 3c) of the pupil which effectively covers a region of interest.
- the output of block 32 may go to a find contours block 33.
- the found contours of image 43 may go to a select n (two or more) most centralized contours block 34.
- the selected most centralized contours may go to a curve fitting block 35 to curve fit the boundary of the pupil blob to a circle, ellipse or the like.
- the circle may be adequate for virtually all cases.
- the output of the curve fitting block may go to a diamond 36 to indicate the level of curve fitness and its' adequacy.
- the approach is to loop through the n depicted contours to pick the contour that fits the most or best to the model based on the perimeter and coverage fitting.
- An output 46 from diamond 36 may provide pupil information such as the curve fitting, whether the item is an eye, based upon the fitness measures, the percent of pixels within the curve that fit well the model, the radius and center of the pupil model, and the proportion of the blob that is contained within the pupil model.
- Figure 2b is a structural version of Figure 2a.
- a pupil region extractor 31 of profiler 15, 25 may be connected to an output of the candidate selector 111 or 114 of Figure Ib.
- An image 41 and a maximum pupil signal 44 may be input to extractor 31.
- An output of the extractor 31 may be connected to an adaptive thresholder 32.
- a percent input 45 may be provided to the thresholder 32.
- the output 43 (e.g., binary image) may go to a contours finder 33.
- An input to a most centralized contour picker 34 may be from contours finder 33.
- An output of the picker 34 may go to a curve fitter 35.
- An input to the selector of the best curve to fit the model diamond 36 may be from the curve fitter 35.
- An output 46 may provide pupil information 46 to a profile evaluator 16 or 26.
- the threshold may be adaptively set based upon the histogram distribution of the intensities of the pixel within the region of interest.
- a minimum threshold is based upon the coverage of the object of interest (pupil) in pixels with respect to the size of the ROI image (i.e., region of interest).
- the percentage of the blob size with respect to the ROI is assumed to be at least the ratio of the minimum expected size of a pupil blob (i.e., pupil surface) with respect to the ROI surface (chosen to be the same size of the maximum expected pupil diameter).
- the percentage ratio, ⁇ may be computed with the following equation.
- R 1n and R M represent the minimum and maximum possible values of expected radius of the pupil
- S p is the minimum surface of the pupil
- S ROI is a surface that is a region of interest
- E[] is an expected value operator.
- Fitness metrics may be used within the eye profiling procedure. At least two metrics can be detected to measure how good the estimated regular shape fits the detected curve at the boundary of the pupil blob.
- the first curve fitting metric may incorporate the following formula.
- the curve f(x,y) represents the boundary of the blob
- F(x,y) is the border curve of estimated fitting shape
- F c (x,y) is the moment center of the model shape.
- N in the above equation represents the length of the curve f(x,y) .
- the operator u() is the step function and ⁇ « 1 is a tolerance factor.
- a fitting metrics may be basically the ratio of the estimated shape surface coverage or intersection of the surface of the model and the blob over the blob surface.
- S blob is the surface of the blob.
- a rectilinear image rotation angle may be noted.
- An iris image capture system that captures both eyes simultaneously may provide a way to measure a head tilt angle. By detecting pupil regions of both eyes during an eye finding procedure, one may calculate the angle of the line passing through both pupil center masses and the horizontal axis of the camera. The eye finder system 10 may then extract both eye images at the estimated orientation axis of the eyes. A misalignment in line detection may be further addressed using the nature of the matching approach which accounts for any nonsignificant eye orientation.
- FIG. 2b is a diagram of a structure of a profiler.
- the overall eye detection system is shown in Figure 4. It shows a camera 61 that may provide an image with a face in it to the eye finder 10 as noted herein.
- the eyefinder 10, 62 may provide an image of one or two eyes that go to the iris segmentation block 63.
- a polar segmentation (POSE) system in block 63 may be used to perform the segmentation.
- POSE may be based on the assumption that image (e.g., 320 x 240 pixels) has a visible pupil where iris can be partially visible. There may be pupil segmentation at the inner border between the iris and pupil and segmentation at the outer border between the iris and the sclera and iris and eyelids.
- An output having a segmented image may go to a block 64 for mapping / normalization and feature extraction.
- An output from block 64 may go to an encoding block 65 which may provide an output, such as a barcode of the images to block put in terms of ones and zeros.
- the coding of the images may provide a basis for storage in block 66 of the eye information which may be used for enrolling, indexing, matching, and so on, at block 67, of the eye information, such as that of the iris and pupil, related to the eye.
- Figure 5 shows a diagram of a kernel 70 of a candidate which may be one of several candidates. Box 71 maybe selected to fit within a circular shape that would represent the minimum possible diameter of the pupil.
- Box 72 may be selected to fit within a circular shape that might represent the maximum size of the reflection.
- the actual circular shapes in Figure 5 may be used instead of the boxes 70 and 71; however, the circular shape requires much computation and the square shape or box may be regarded as being an adequate approximation. This mechanism may be used to locate pupil location candidates.
- a blob suspected of being a pupil may be profiled with a fitness curve on its outer portion. If the curve fits a predefined model like a circle, then one may give it a score of a certain percent of fitness. A second part of the fitness check is to determine what percentage of the pixels of the pupil is contained within the model curve. If the fitness percentages are significant enough to a predefined level, then one might assume that the object scrutinized is a pupil. If so, then the object is checked relative to a range of distance between two eyes.
- a threshold level, ⁇ may be adaptive based on contrast, illumination, and other information. The threshold may be determined with the equation noted herein for ⁇ min .
- Figure 6 shows a box 73 which may be a region of interest. An area 74 may be of a first color which is relatively light. An area 75 may be of a second color that is dark. An area 76 may be of a third color that is lighter than the first color.
- a histogram may be taken of the contents of box or region 73. The histogram may look like the graph of Figure 7.
- the ordinate axis represents the number of pixels having a contrast or intensity (i.e., lightness / darkness) value of the values represented on the abscissa axis, which range from 0 to 255, i.e., from dark to light, respectively.
- the result is two peaks 78 and 79 with a middle point 77 which may be associated with the ⁇ min .
- the plot 81 appears normal.
- Other plots having one peak, a flat peak or peaks, peaks having a large separation, or other appearance that appear abnormal relative to plot generally indicate an unacceptable situation.
- the present approach utilizes adaptive thresholding which has a threshold that is not fixed or arbitrary. The depicted threshold is limited with the minimum value of that defined by equation (1).
- Figure 8 is like Figure 6 which has an area 76 of reflection on pupil 75 which Figure 8 does not have. However, an area 86 of reflection may be assumed for the pupil 85 in Figure 8.
- the pixels of a region of interest or kernel 87 may be ranked according to lightness and darkness as represented by marks on a scale 95 of diagram 90 as shown in Figure 9.
- An arrow 96 represents a direction of increasingly lighter pixels.
- An arrow 97 represents a direction of increasingly darker pixels.
- each mark may represent a pixel; although each mark could represent any number of pixels or a fraction of a pixel or pixels.
- the kernel 87 size may be N pixels. N pixels may be
- the reflection 86 may be represented by "Nr f c".
- “Nr f c” may refer to the reflection 86 pixels.
- the “Nr fc " pixels may be represented by a group 91 of marks on scale 95.
- “N-Nr fc” may represent the dark area 85.
- the “N-N rfc " pixels may be represented by a group 92 of marks on scale 95.
- the measure may be defined as the argument of the maximum difference between the reflection pixel measure (local maxima) within the reflection spot and the average mean of the dark pixels that represent the pupil profile.
- the vector is t ⁇ iWernel elements sorted in a descending order based on the intensity values as shown in Figure 9. An average value of intensity may be calculated for each group of pixels.
- Figures 10a and 10b relate to eye finding using reflection and/or non-reflection measures relative to eyes 101 and 104, respectively. For a situation of no actual reflection on the pupil, then there may be a representative value of the dark pixels in the bottom scale that maximize the argument l/ ⁇ o . This may be true for either condition whether there is reflection or no reflection. Hence, the formulas may be combined into one to work for both situations as indicated by the following equation,
Abstract
A system for finding and providing images of eyes acceptable for review, recordation, analysis, segmentation, mapping, normalization, feature extraction, encoding, storage, enrollment, indexing, matching, and/or the like. The system may acquire images of the candidates run them through a contrast filter. The images may be ranked and a number of candidates may be extracted for a list from where a candidate may be selected. Metrics of the eyes may be measured and their profiles evaluated. Also, the spacing between a pair of eyes may be evaluated to confirm the pair's validity. Eye images that do not measure up to certain standards may be discarded and new ones may be selected.
Description
APPROACHES AND APPARATUS FOR EYE DETECTION IN A DIGITAL IMAGE
This application is a continuation-in-part of U.S. Patent Application No. 11/275,703, filed January 25, 2006, which claims the benefit of U.S. Provisional Application No. 60/647,270, filed January 26, 2005.
This application is a continuation-in-part of U.S. Patent Application No. 11/043,366, filed January 26, 2005.
This application is a continuation-in-part of U.S. Patent Application No. 11/372,854, filed March 10, 2006; This application claims the benefit of U.S. Provisional Application No.
60/778,770, filed March 3, 2006.
The government may have rights in the present invention.
Background
Related applications may include U.S. Patent Application No. 10/979,129, filed November 3, 2004, which is a continuation-in-part of U.S. Patent Application
10/655,124, filed September 5, 2003, which are hereby incorporated by reference, and U.S. Patent Application 11/382,373, filed May 9, 2006, which is hereby incorporated by reference.
U.S. Patent Application No. 11/275,703, filed January 25, 2006, is hereby incorporated by reference.
U.S. Provisional Application No. 60/647,270, filed January 26, 2005, is hereby incorporated by reference.
U.S. Patent Application No. 11/043,366, filed January 26, 2005, is hereby incorporated by reference. U.S. Patent Application No. 11/372,854, filed March 10, 2006, is hereby incorporated by reference.
U.S. Provisional Application No. 60/778,770, filed March 3, 2006, is hereby incorporated by reference.
Summary
The invention is an approach and apparatus for localizing eyes of a human in a digital image to be processed for iris recognition.
Brief Description of the Drawing
Figure 1 a is a diagram of an overall illustrative structure of an eye finding system;
Figure Ib is a diagram with a group structure of the eye finding system;
Figure 2a is a diagram of an approach for determining a profile of an eye as provided by a measure of profile metrics; Figure 2b is a diagram of a structure of a profiler;
Figures 3a, 3b and 3c show an image of a selected eye, a pupil image 42 and a binary image 43 of a pupil, respectively;
Figure 4 is a diagram of an overall iris recognition system;
Figure 5 shows a diagram of a kernel having a box representing the diameter of a pupil and a box representing a pupil reflection;
Figure 6 shows a box representing a region of interest having areas which are relatively light, dark and lighter than the relatively light area situated in the dark area representing a pupil model;
Figure 7 is a histogram of the contrast or intensity values of areas of Figure 6; Figure 8 is like Figure 6 but does not have a lighter (modeling a typical reflection) than the relatively light area situated in the dark area representing a pupil;
Figure 9 is a scale of marks representing pixels of a region of interest ranked according to lightness and darkness; and
Figures 10a and 10b relate to eye finding using reflection and/or non-reflection measures.
Description
Eye detection may be the first step toward building a reliable automated iris recognition system in a natural context. Some iris recognition systems rely heavily on predetermined eye locations to properly zoom on the input eye prior to iris segmentation. In addition to biometrics, eye detection (also known as eye finding or eye localization) may support other new technology areas such as eye tracking and human computer interaction or driver drowsiness monitoring systems. Eye detection may serve social learning to identify eye directions like pointing gesture using eye directions. The present approach and apparatus may be used for finding eyes within a digital image. A local contrast change profiling may be in eye finding. Instead of extracting multiple local features and search globally in the image as many COTS (commercial off- the-shelf) facial recognition packages are based on, the present approach may be based on a system engineering approach to construct the illumination scheme during eye image
acquisition to shine the surface of the pupil surface and result into a high reflection point preferably within the pupil of the eye image or close to the pupil of the eye image. This specular reflection point may be used as a reference for an eye search in the digital image. Thus, during the image analysis, the search may be limited to a simplified localization scheme of the highest value pixels associated with these specular reflection pixels and analyze the very local features surrounding these hot spots to confirm an eye profile. To meet the requirements of a real-time system, the eye finding approach may be implemented as a cascade process that divides the local features of an eye into a primary feature of contrast profile associated with high pixel values, depict only potential eye pairs within a specified range, and then test the resulting valid pairs against a feature vector of two or more variables that includes a predefined regular shape fitting with multiple curve fitting measures.
The present approach may be for quickly and robustly localizing the eyes of a human eye in close-up or face images. The approach is based on sensing reflection points within the pupil region as a precursor to the analysis. The approach is formulated to work for cases where reflection is not present within the pupil. The technical approach locates eyes whether there is or there is no reflection. However, in case of the reflection, The detection may hence be simplified to search for these specific reflection points surrounded with dark contrast that represent the pupil. Then the region of interest centered at these potential locations may be processed to find an eye profile. Two valid eyes may be extracted that are within an expected range of eye positioning. The approach for finding eyes decomposes into the following steps. There may be a contrast filter to detect specular reflection pixels. There may be results prioritization to extract valid eye pair with maximum local contrast change. The eye pair may be defined as a valid pair if the two potential eyes are spaced within a predefined range. An adaptive threshold may be applied to detect a central blob. There may be curve fitting of the blob boundaries into a shape. Curve fitness and shape area coverage of the blob surface may be measured for validation. The approach described here may be part of a preprocessing technique used to locate the eyes of a human in a digital image to be processed for iris recognition. Eye detection may be the first stage for any automated iris recognition analysis system and may be critical for consistent iris segmentation. Several eye detection algorithms may be developed as a basis for face detection. Eye finding approaches may be classified into several categories based upon knowledge based approaches, template matching, and eye socket corner detection. The present approach may address real-time
operational requirements. One solution may be to cascade localized features of the eye to speed up the process.
Appearance based approaches using Eigenspace supervised classification technique that is based on learning from a set of training images may be used to capture the most representative variability of eye appearance. Template matching can be regarded as a brute force approach which may include constructing a kernel that is representative of a typical eye socket and convolve the image with the kernel template to identify the highest values of the convolution indicating a match of the eye the identified locations. Knowledge based approaches may be based on specific rules that are captured by an expert that discriminate the eye local features from any other features. These sets of rules may then be tested against virtually all possible combination to identify the eye locations.
The present approach may provide for quickly and robustly localizing the eyes of a human eye in close-up or face images. The approach may be based on sensing reflection points within the pupil region as a precursor to the analysis. The approach may be also based on sensing the pupil profile in case of no reflection. If reflection is present, the detection may then be simplified to search for these specific reflection points surrounded with dark contrast that represent the pupil. The region of interest centered at these potential locations may then be processed to find an eye profile. Two valid pairs may be extracted that are within an expected range of eye positioning.
The present approach for finding eyes may decompose into the following. To start, the eye may be illuminated to generate a reflection reference point on the pupil surface. The captured wide-field image may be filtered using reflection detection contrast changes to find potential eye locations. For each potential eye location, the local contrast change between the central point and its surrounding pixels may be computed and results may be prioritized to extract valid eye pair with maximum local contrast change. The eye pair may be defined as a valid pair if the two potential eyes are spaced within a predefined range. For each eye of valid eye pair, an adaptive threshold may be executed on a cropped image of the central region of the potential eye to extract a blob of the pupil. Just a single blob may be depicted based on size, its distance to the central point of the cropped image, and how good it fits to a predefined fitting model (e.g., circular shape). With a predefined model shape, such as a circle or an ellipse, the blob edges may be fitted
into the pupil fitting model. Curve fitness and model shape area coverage of the blob surface may be measured for validation.
A preprocessing technique may locate the eyes of a human in a digital image to be processed for iris recognition. An overall illustrative structure of an eye finding system 10 is shown in Figure Ia. The system engineering for eye illumination is not necessarily shown in this system. A digital camera may be used for acquiring 9 images of candidates. The image may be put through a contrast filter 11 to detect relevant high contrast areas of the image. There may a prioritization (i.e., ranking) of the significant spots in the image in block 12. Of these, N candidates may be extracted in block 13. A candidate may have a coordinate Ci(x,y). An output of block 13 may go to block 14 where a new first eye may be selected from the candidate list. From block 14, the eye image candidate may go to a block 15 for a measurement of profile metrics. A profile of the eye may go to a diamond 17 where a determination of the validity of the profile is made. If the profile is not valid, then that first eye may be deleted from the list at block 17. Then at a diamond 18, a count is checked to note whether it is greater than zero. If not, the then this approach is stopped at place 19. If so, then a new first eye may be selected at block 14 from the list from block 13. The profile metrics of this new first eye may be measured at block 15 and passed on to diamond 16 to determine the validity of the profile. If the profile is valid, then the selected first eye may go to place 20, and a second eye is selected at block 21 from the list of candidates from block 13 having a coordinate c2(x,y). The spacing of the first and second eyes may be determined at block 22 as D(ci(x,y),c2(x,y)). The spacing may be checked to see whether it is within an appropriate range at a diamond 23. If not, then the second eye may be deleted from the list at block 24. If so, then metrics of the profile of the second eye may be measured at block 25. The profile metric may be forwarded to a diamond 26 for a determination of the validity of the profile. If the profile is not valid, then the second eye may be deleted from the list at block 24 and at diamond 27, a question of whether the count is greater than zero. If so, then another second eye may be selected from the list at block 21, and the approach via the blocks 21, 23 and 25, and diamonds 23, 26 and 27 may be repeated. If not, then the approach for the second eye may end at place 19. If the profile is valid at diamond 26, then the selected second eye may go to place 20.
A higher level approach to system 10 in Figure Ia may include an output of the contrast filtering 11 going a select candidate block 111. An output from block 111 may go to a validate profile block 112. Outputs from block 112 may go to a select candidate
block 114 and a result block 20, or eliminate candidate block 113. An output of block 113 may go to the select candidate block 111 and/or to the stop place 19. An output from block 114 may go to a validate pair block 115. Block 115 may provide an output to a validate profile block 116. Outputs from block 116 may go to an eliminate candidate 117 and/or to the result block 20. Outputs of block 117 may go the select candidate 114 and the stop place 19. The processing in system 10 may be digital, although it may be analog, or it may be partially digital and analog.
Figure Ib is a diagram with a group structure of the eye finding system 10. The corresponding components (according to reference numbers) of Figure Ia may have additional description. The candidates noted herein may refer to various images of eyes. A camera 9 may be connected to the contrast filter 11. An output of the filter 11 may go to a ranking mechanism 12, which in turn is connected to the candidates extractor 13. The output of extractor 13 may go to a candidate determiner 14 for selecting a new first candidate. Mechanism 12, extractor 13 and determiner 14 constitute a candidate selector 111.
An output of determiner 14 may go to a metric profiler 15 which in turn has an output connected to a profile evaluator 16. Profiler 15 and evaluator 16 may constitute profile validator 112. Outputs of evaluator 16 may go to candidate determiner 21, resulter 20 and candidate remover 17. Remover may have an output that goes to a counter 18. Candidate remover 17 and counter 18 may constitute a candidate eliminator 113. If counter 18 has a count of greater than zero, an output may go to the candidate determiner 14 for selection of a new candidate. If the output is not greater than zero, then an output may go to the stopper 19.
A candidate determiner 21 for selecting a 2nd candidate may have an output to a space measurer 22. The candidate Space measurer 22 may have an output to the range indicator 23 which may indicate whether the two candidates are at an appropriate distance from each other for validation. Measure 22 and indicator 23 may constitute a pair validator 115. Candidate determiner 21 and previously noted ranking mechanism 12 and candidates extractor 13 may constitute a candidate selector 114. If the pair of candidates is valid then an output from validator 115 may go to a profiler 25, or if the pair is not valid then an output from validator 115 may go to a candidate remover 24. An output of profiler 25 may go to a profile evaluator 26 which may determine whether the profile of the second candidate is valid or not. If valid, then an output of evaluator 26 may provide second candidate information to the resulter 20. If invalid, then an output of evaluator 26
may provide a signal to the candidate remover 24. Profiler 25 and profiler evaluator 26 may constitute a profile validator 116. An output of candidate remover may go to a counter 27. If the counter 27 indicates a value greater than zero then an output may go to the candidate determiner 21 for selecting a second candidate. If the counter 27 indicates a value not greater than zero, then an output may go to a stopper 19. The candidate remover 24 and counter 27 may constitute a candidate eliminator 117.
Figure 2a shows the approach for determining a profile of an eye as provided by a measure profile metrics or eye profiling block 15, 25. An image 41 of a selected eye (Figure 3a) may go to an extract pupil region block 31. The block dimension is determined based on the maximum expected value of the pupil diameter. A maximum pupil input 44 may be provided to block 31. An output from block may be a pupil image 42 (Figure 3b) which goes to an adaptive thresholding block 32. A percent input 45 may be provided to block 32. The pixel distribution to compute the intensity histogram may be provided to block 32. An output of block 32 may be a binary image 43 (Figure 3c) of the pupil which effectively covers a region of interest. The output of block 32 may go to a find contours block 33. The found contours of image 43 may go to a select n (two or more) most centralized contours block 34. The selected most centralized contours may go to a curve fitting block 35 to curve fit the boundary of the pupil blob to a circle, ellipse or the like. The circle may be adequate for virtually all cases. The output of the curve fitting block may go to a diamond 36 to indicate the level of curve fitness and its' adequacy. The approach is to loop through the n depicted contours to pick the contour that fits the most or best to the model based on the perimeter and coverage fitting. An output 46 from diamond 36 may provide pupil information such as the curve fitting, whether the item is an eye, based upon the fitness measures, the percent of pixels within the curve that fit well the model, the radius and center of the pupil model, and the proportion of the blob that is contained within the pupil model.
Figure 2b is a structural version of Figure 2a. A pupil region extractor 31 of profiler 15, 25 may be connected to an output of the candidate selector 111 or 114 of Figure Ib. An image 41 and a maximum pupil signal 44 may be input to extractor 31. An output of the extractor 31 may be connected to an adaptive thresholder 32. A percent input 45 may be provided to the thresholder 32. The output 43 (e.g., binary image) may go to a contours finder 33. An input to a most centralized contour picker 34 may be from contours finder 33. An output of the picker 34 may go to a curve fitter 35. An input to
the selector of the best curve to fit the model diamond 36 may be from the curve fitter 35. An output 46 may provide pupil information 46 to a profile evaluator 16 or 26.
For the thresholding of block 32, the threshold may be adaptively set based upon the histogram distribution of the intensities of the pixel within the region of interest. A minimum threshold is based upon the coverage of the object of interest (pupil) in pixels with respect to the size of the ROI image (i.e., region of interest). The percentage of the blob size with respect to the ROI is assumed to be at least the ratio of the minimum expected size of a pupil blob (i.e., pupil surface) with respect to the ROI surface (chosen to be the same size of the maximum expected pupil diameter). Hence, the percentage ratio, λ , may be computed with the following equation.
Where R1n and RM represent the minimum and maximum possible values of expected radius of the pupil, Sp is the minimum surface of the pupil, SROI is a surface that is a region of interest, and E[] is an expected value operator. Fitness metrics may be used within the eye profiling procedure. At least two metrics can be detected to measure how good the estimated regular shape fits the detected curve at the boundary of the pupil blob. The first curve fitting metric may incorporate the following formula.
F(x,y) is the border curve of estimated fitting shape, and Fc (x,y) is the moment center of the model shape. N in the above equation represents the length of the curve f(x,y) .the operator u() is the step function and ε « 1 is a tolerance factor.
Another consideration may be given to measuring the proportion of the blob within the estimated model curve. A fitting metrics may be basically the ratio of the estimated shape surface coverage or intersection of the surface of the model and the blob over the blob surface.
_ Surfacejblob fl F(x,y)) ''2 ~ Sblob where Sblob is the surface of the blob.
A rectilinear image rotation angle may be noted. An iris image capture system that captures both eyes simultaneously may provide a way to measure a head tilt angle. By detecting pupil regions of both eyes during an eye finding procedure, one may calculate the angle of the line passing through both pupil center masses and the horizontal axis of the camera. The eye finder system 10 may then extract both eye images at the estimated orientation axis of the eyes. A misalignment in line detection may be further addressed using the nature of the matching approach which accounts for any nonsignificant eye orientation. The advantage of the present preprocessing approach is that one may reduce the amount of shifting of bits during the matching process to a few bits thus yielding to faster time response of the system. If rotation correction is not performed, the matching uncertainty may be set to maximum and thus the barcode bit shifting is set to its maximum. On the other hand, if such correction is performed, the matching process may be limited to just a few bits shifted to account for any misalignments of eyes with images in the database. Figure 2b is a diagram of a structure of a profiler.
The overall eye detection system is shown in Figure 4. It shows a camera 61 that may provide an image with a face in it to the eye finder 10 as noted herein. The eyefinder 10, 62 may provide an image of one or two eyes that go to the iris segmentation block 63. A polar segmentation (POSE) system in block 63 may be used to perform the segmentation. POSE may be based on the assumption that image (e.g., 320 x 240 pixels) has a visible pupil where iris can be partially visible. There may be pupil segmentation at the inner border between the iris and pupil and segmentation at the outer border between the iris and the sclera and iris and eyelids. An output having a segmented image may go to a block 64 for mapping / normalization and feature extraction. An output from block 64 may go to an encoding block 65 which may provide an output, such as a barcode of the images to block put in terms of ones and zeros. The coding of the images may provide a basis for storage in block 66 of the eye information which may be used for enrolling, indexing, matching, and so on, at block 67, of the eye information, such as that of the iris and pupil, related to the eye. Figure 5 shows a diagram of a kernel 70 of a candidate which may be one of several candidates. Box 71 maybe selected to fit within a circular shape that would represent the minimum possible diameter of the pupil. Box 72 may be selected to fit within a circular shape that might represent the maximum size of the reflection. The actual circular shapes in Figure 5 may be used instead of the boxes 70 and 71; however,
the circular shape requires much computation and the square shape or box may be regarded as being an adequate approximation. This mechanism may be used to locate pupil location candidates.
A blob suspected of being a pupil may be profiled with a fitness curve on its outer portion. If the curve fits a predefined model like a circle, then one may give it a score of a certain percent of fitness. A second part of the fitness check is to determine what percentage of the pixels of the pupil is contained within the model curve. If the fitness percentages are significant enough to a predefined level, then one might assume that the object scrutinized is a pupil. If so, then the object is checked relative to a range of distance between two eyes.
A threshold level, λ, may be adaptive based on contrast, illumination, and other information. The threshold may be determined with the equation noted herein for λmin. Figure 6 shows a box 73 which may be a region of interest. An area 74 may be of a first color which is relatively light. An area 75 may be of a second color that is dark. An area 76 may be of a third color that is lighter than the first color. A histogram may be taken of the contents of box or region 73. The histogram may look like the graph of Figure 7. The ordinate axis represents the number of pixels having a contrast or intensity (i.e., lightness / darkness) value of the values represented on the abscissa axis, which range from 0 to 255, i.e., from dark to light, respectively. The result is two peaks 78 and 79 with a middle point 77 which may be associated with the λmin. The plot 81 appears normal. Other plots having one peak, a flat peak or peaks, peaks having a large separation, or other appearance that appear abnormal relative to plot generally indicate an unacceptable situation. One may note that the present approach utilizes adaptive thresholding which has a threshold that is not fixed or arbitrary. The depicted threshold is limited with the minimum value of that defined by equation (1).
There may be a situation where there is no reflection to be found on a pupil. Figure 8 is like Figure 6 which has an area 76 of reflection on pupil 75 which Figure 8 does not have. However, an area 86 of reflection may be assumed for the pupil 85 in Figure 8. The pixels of a region of interest or kernel 87 may be ranked according to lightness and darkness as represented by marks on a scale 95 of diagram 90 as shown in Figure 9. An arrow 96 represents a direction of increasingly lighter pixels. An arrow 97 represents a direction of increasingly darker pixels. For illustrative purposes, each mark may represent a pixel; although each mark could represent any number of pixels or a fraction of a pixel or pixels. The kernel 87 size may be N pixels. N pixels may be
10 of 18
represented by a group 93 of marks on a scale 95. The reflection 86 may be represented by "Nrfc". "Nrfc" may refer to the reflection 86 pixels. The "Nrfc" pixels may be represented by a group 91 of marks on scale 95. "N-Nrfc" may represent the dark area 85. The "N-Nrfc" pixels may be represented by a group 92 of marks on scale 95. In cases where we have reflections on the pupil, the measure may be defined as the argument of the maximum difference between the reflection pixel measure (local maxima) within the reflection spot and the average mean of the dark pixels that represent the pupil profile. Hence,
Cpupύ O, y) = arg max(vmax - μo ) x'y
The vector is tϊiWernel elements sorted in a descending order based on the intensity values as shown in Figure 9. An average value of intensity may be calculated for each group of pixels.
For the "Nrfc" group 91, one may have the local maxima of the reflection spot vmax estimated as the average mean of only the first K elements of the reflection pixels. K may be selected to be such as K « Nrfc . For the "N-Nrfc" group 92, one may have "μo",
Figures 10a and 10b relate to eye finding using reflection and/or non-reflection measures relative to eyes 101 and 104, respectively. For a situation of no actual reflection on the pupil, then there may be a representative value of the dark pixels in the bottom scale that maximize the argument l/μo. This may be true for either condition whether there is reflection or no reflection. Hence, the formulas may be combined into one to work for both situations as indicated by the following equation,
In the present specification, some of the matter may be of a hypothetical or prophetic nature although stated in another manner or tense.
Although the invention has been described with respect to at least one illustrative example, many variations and modifications will become apparent to those skilled in the art upon reading the present specification. It is therefore the intention that the appended claims be interpreted as broadly as possible in view of the prior art to include all such variations and modifications.
11 of 18
Claims
1. A candidate finder system comprising: a filter; a first candidate selector connected to the filter; a first profile validator connected to the candidate selector; a first candidate eliminator connected to the second profile validator and the first candidate selector; a second candidate selector connected to the first profile validator; a pair validator connected to the second candidate determiner; a second profile validator connected to the pair validator; and a second candidate eliminator connected to the first profile validator and the second candidate determiner.
2. The system of claim 1, further comprising: a resulter connected to the first and second profile validators; and a stopper connected to the first and second candidate eliminators.
3. The system of claim 1, wherein a candidate is an eye.
4. The system of claim 1, wherein: the first candidate selector comprises: a ranking mechanism connected to the filter; a candidates extractor connected to the ranking mechanism; and a first candidate determiner connected to the candidate extractor, the first profiler and the first candidate eliminator; and the second candidate selector comprises: the ranking mechanism connected to the filter; the candidates extractor connected to the ranking mechanism; and a second candidate determiner connected to the first profile evaluator, the pair validator and the second candidate eliminator.
5. The system of claim 1, wherein: the first profile validator comprises:
12 of 18 a first profiler connected to the first candidate selector; and a first profile evaluator connected to the first profiler, the first candidate eliminator and the second candidate selector; and the second profile validator comprises: a second profiler connected to the pair validator; and a second profile evaluator connected to the second profiler and the second candidate eliminator.
6. The system of claim 5, wherein: a candidate is an eye having a pupil; the first profiler comprises: a pupil region extractor connected to the first candidate selector; an adaptive thresholder connected to the pupil region extractor; a contours finder connected to the adaptive thresholder; a contour picker connected to the contours finder; a curve fitter connected to the contour picker; and a curve selector connected to the curve fitter and the first profile evaluator; and the second profiler comprises: a pupil region extractor connected to the pair validator candidate selector; an adaptive thresholder connected to the pupil region extractor; a contours finder connected to the adaptive thresholder; a contour picker connected to the contours finder; a curve fitter connected to the contour picker; and a curve selector connected to the curve fitter and the second profile evaluator.
7. The system of claim 6, wherein an adaptive threshold is computed adaptively based upon a pupil size and a contrast distribution of an image region having a pupil and its surrounding area.
8. The system of claim 1, wherein the pair validator comprises: a space measurer connected to the second candidate selector; and
13 of 18 a range indicator connected to the space measurer and the second profile validator.
9. The system of claim 1, further comprising: a resulter connected to the first and second profile validators; and wherein the resulter is for receiving a valid pair of candidates.
10. The system of claim 1, wherein: the filter is a contrast filter; and the candidates are eyes.
11. The system of claim 10, wherein the contrast filter is based upon a pupil dark profile with respect to its surroundings.
12. The system of claim 10, wherein the contrast filter is based upon a light reflection profile with respect to its surroundings
13. A method for eye finding, comprising: providing candidates; contrast filtering the candidates; selecting a first eye from the candidates; validating a profile of the first eye; selecting a second eye from the candidates; validating an amount of space between the first and second eyes; and validating a profile of the second eye.
14. The method of claim 13, further comprising: eliminating the first eye if the first eye has an invalid profile; eliminating the second eye if the amount of space between the first and second eyes is invalid; and eliminating the second eye if the amount of space between the first and second eyes is valid and the second eye has an invalid profile.
15. The method of claim 14, wherein:
14 of 18 the selecting a first eye comprises: prioritizing the candidates according to contrast into a first list of candidates; extracting a number of candidates from the first list of candidates into a second list of candidates; and selecting a first eye from the second list of candidates; and the selecting a second eye comprises selecting a second eye from the second list of candidates.
16. The method of claim 15, wherein: the validating a profile of first eye comprises: measuring profile metrics of the first eye; and determining whether the profile metrics of the first eye are valid; and the validating a profile of second eye comprises: measuring profile metrics of the second eye; and determining whether the profile metrics of the second eye are valid.
17. The method of claim 16, wherein the validating an amount of space between the first and second eyes comprises: measuring a spacing between the first and second eyes; and determining whether the spacing is within a valid range.
18. An eye finder system comprising: a camera; a filter connected to the camera; a candidate lister connected to the filter; an eye selector connected to the candidate lister; and an eye profile evaluator connected to the eye selector.
19. The system of claim 18, further comprising an eye eliminator connected to the eye profile evaluator.
20. The system of claim 19, wherein:
15 of 18 the eye eliminator comprises: a deleter, connected to the eye profile evaluator, for removing the eye from a list provided by the candidate lister if the eye profile evaluator indicates a profile of the eye to be invalid; and a counter connected to the deleter and to the eye selector; if the counter has a count greater than zero, then the eye selector may select a new eye from the list provided by the candidate lister; or if the counter has a count not greater than zero, then the system may stop.
21. The system of claim 18, further comprising: a second eye selector connected to the eye profile evaluator; an eye pair evaluator connected to the second eye selector; and a second profile evaluator connected to the eye pair evaluator.
22. The system of claim 18, wherein the eye profile evaluator comprises: a pupil region extractor connected to the eye selector; an adaptive thresholder connected to the pupil region extractor; a contours finder connected to the adaptive thresholder; a contour picker connected to the contours finder; a curve fitter connected to the contour picker; and a best curve selector connected to the curve fitter.
23. The system of claim 21, wherein if the profile evaluator and the second profile evaluator indicate that a first eye and a second eye have appropriate metrics, and the eye pair evaluator determines that the first eye and second eye are appropriately spaced from each other, then the first and second eyes or portions of them are acceptable for review, recordation, analysis, segmentation, mapping, normalization, feature extraction, encoding, storage, enrollment, indexing, matching, and/or the like.
24. The system of claim 21, wherein if just one eye is indicated by one or more profile evaluators to have appropriate metrics, then the eye or a portion of it is acceptable for review, recordation, analysis, segmentation, mapping,
16 of 18 normalization, feature extraction, encoding, storage, enrollment, indexing, matching, and/or the like.
17 of 18
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP07757683A EP1991946B1 (en) | 2006-03-03 | 2007-03-01 | Approaches and apparatus for eye detection in a digital image |
JP2008558464A JP5155886B2 (en) | 2006-03-03 | 2007-03-01 | Method and apparatus for detecting eyes in digital images |
KR1020087024199A KR101343901B1 (en) | 2006-03-03 | 2008-10-02 | Approaches and apparatus for eye detection in a digital image |
Applications Claiming Priority (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US77877006P | 2006-03-03 | 2006-03-03 | |
US60/778,770 | 2006-03-03 | ||
US11/372,854 | 2006-03-10 | ||
US11/372,854 US8442276B2 (en) | 2006-03-03 | 2006-03-10 | Invariant radial iris segmentation |
US11/672,108 | 2007-02-07 | ||
US11/672,108 US8090157B2 (en) | 2005-01-26 | 2007-02-07 | Approaches and apparatus for eye detection in a digital image |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2007101269A1 true WO2007101269A1 (en) | 2007-09-07 |
Family
ID=38141292
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2007/063029 WO2007101269A1 (en) | 2006-03-03 | 2007-03-01 | Approaches and apparatus for eye detection in a digital image |
Country Status (5)
Country | Link |
---|---|
US (1) | US8090157B2 (en) |
EP (1) | EP1991946B1 (en) |
JP (1) | JP5155886B2 (en) |
KR (1) | KR101343901B1 (en) |
WO (1) | WO2007101269A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2465881A (en) * | 2008-12-05 | 2010-06-09 | Honeywell Int Inc | A system for iris recognition using a set of quality metrics |
Families Citing this family (51)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7088872B1 (en) * | 2002-02-14 | 2006-08-08 | Cogent Systems, Inc. | Method and apparatus for two dimensional image processing |
US8064647B2 (en) * | 2006-03-03 | 2011-11-22 | Honeywell International Inc. | System for iris detection tracking and recognition at a distance |
US7593550B2 (en) | 2005-01-26 | 2009-09-22 | Honeywell International Inc. | Distance iris recognition |
US8442276B2 (en) | 2006-03-03 | 2013-05-14 | Honeywell International Inc. | Invariant radial iris segmentation |
US7933507B2 (en) | 2006-03-03 | 2011-04-26 | Honeywell International Inc. | Single lens splitter camera |
US8045764B2 (en) | 2005-01-26 | 2011-10-25 | Honeywell International Inc. | Expedient encoding system |
US8705808B2 (en) * | 2003-09-05 | 2014-04-22 | Honeywell International Inc. | Combined face and iris recognition system |
US8098901B2 (en) | 2005-01-26 | 2012-01-17 | Honeywell International Inc. | Standoff iris recognition system |
US8085993B2 (en) * | 2006-03-03 | 2011-12-27 | Honeywell International Inc. | Modular biometrics collection system architecture |
US8049812B2 (en) * | 2006-03-03 | 2011-11-01 | Honeywell International Inc. | Camera with auto focus capability |
US8131477B2 (en) * | 2005-11-16 | 2012-03-06 | 3M Cogent, Inc. | Method and device for image-based biological data quantification |
KR101308368B1 (en) | 2006-03-03 | 2013-09-16 | 허니웰 인터내셔널 인코포레이티드 | An iris recognition system having image quality metrics |
WO2008131201A1 (en) | 2007-04-19 | 2008-10-30 | Global Rainmakers, Inc. | Method and system for biometric recognition |
US8063889B2 (en) | 2007-04-25 | 2011-11-22 | Honeywell International Inc. | Biometric data collection system |
US8275179B2 (en) * | 2007-05-01 | 2012-09-25 | 3M Cogent, Inc. | Apparatus for capturing a high quality image of a moist finger |
US8411916B2 (en) * | 2007-06-11 | 2013-04-02 | 3M Cogent, Inc. | Bio-reader device with ticket identification |
US20090092283A1 (en) * | 2007-10-09 | 2009-04-09 | Honeywell International Inc. | Surveillance and monitoring system |
US8436907B2 (en) | 2008-05-09 | 2013-05-07 | Honeywell International Inc. | Heterogeneous video capturing system |
US9332191B2 (en) * | 2009-03-02 | 2016-05-03 | Honeywell International Inc. | Method and system for determining shutter fluttering sequence |
TWI395144B (en) * | 2008-06-25 | 2013-05-01 | Altek Corp | Method of eye sight judgment |
US20100014755A1 (en) * | 2008-07-21 | 2010-01-21 | Charles Lee Wilson | System and method for grid-based image segmentation and matching |
US8644565B2 (en) * | 2008-07-23 | 2014-02-04 | Indiana University Research And Technology Corp. | System and method for non-cooperative iris image acquisition |
US8213782B2 (en) * | 2008-08-07 | 2012-07-03 | Honeywell International Inc. | Predictive autofocusing system |
US8090246B2 (en) | 2008-08-08 | 2012-01-03 | Honeywell International Inc. | Image acquisition system |
US8630464B2 (en) * | 2009-06-15 | 2014-01-14 | Honeywell International Inc. | Adaptive iris matching using database indexing |
US8472681B2 (en) | 2009-06-15 | 2013-06-25 | Honeywell International Inc. | Iris and ocular recognition system using trace transforms |
US8878773B1 (en) | 2010-05-24 | 2014-11-04 | Amazon Technologies, Inc. | Determining relative motion as input |
US8948467B2 (en) * | 2010-08-06 | 2015-02-03 | Honeywell International Inc. | Ocular and iris processing system and method |
US8742887B2 (en) | 2010-09-03 | 2014-06-03 | Honeywell International Inc. | Biometric visitor check system |
US8947351B1 (en) | 2011-09-27 | 2015-02-03 | Amazon Technologies, Inc. | Point of view determinations for finger tracking |
US8942434B1 (en) * | 2011-12-20 | 2015-01-27 | Amazon Technologies, Inc. | Conflict resolution for pupil detection |
US9317113B1 (en) | 2012-05-31 | 2016-04-19 | Amazon Technologies, Inc. | Gaze assisted object recognition |
US9265458B2 (en) | 2012-12-04 | 2016-02-23 | Sync-Think, Inc. | Application of smooth pursuit cognitive testing paradigms to clinical drug development |
US9380976B2 (en) | 2013-03-11 | 2016-07-05 | Sync-Think, Inc. | Optical neuroinformatics |
US9094576B1 (en) | 2013-03-12 | 2015-07-28 | Amazon Technologies, Inc. | Rendered audiovisual communication |
US9094388B2 (en) | 2013-05-01 | 2015-07-28 | Dmitri Tkachev | Methods and systems for identifying, verifying, and authenticating an identity |
US9269012B2 (en) | 2013-08-22 | 2016-02-23 | Amazon Technologies, Inc. | Multi-tracker object tracking |
US9672649B2 (en) | 2013-11-04 | 2017-06-06 | At&T Intellectual Property I, Lp | System and method for enabling mirror video chat using a wearable display device |
US9547798B2 (en) | 2014-05-20 | 2017-01-17 | State Farm Mutual Automobile Insurance Company | Gaze tracking for a vehicle operator |
GB2526380A (en) * | 2014-05-24 | 2015-11-25 | Iriss Medical Technologies Ltd | Contact lens assessment |
US9282237B2 (en) * | 2014-07-17 | 2016-03-08 | Schlage Lock Company Llc | Multifocal iris recognition device |
US10425814B2 (en) | 2014-09-24 | 2019-09-24 | Princeton Identity, Inc. | Control of wireless communication device capability in a mobile device with a biometric key |
WO2016089592A1 (en) | 2014-12-03 | 2016-06-09 | Sri Internaitonal | System and method for mobile device biometric add-on |
CA2996039A1 (en) | 2015-08-21 | 2017-03-02 | Magic Leap, Inc. | Eyelid shape estimation using eye pose measurement |
CN114140867A (en) | 2015-10-16 | 2022-03-04 | 奇跃公司 | Eye pose recognition using eye features |
JP2019506694A (en) * | 2016-01-12 | 2019-03-07 | プリンストン・アイデンティティー・インコーポレーテッド | Biometric analysis system and method |
WO2017172695A1 (en) | 2016-03-31 | 2017-10-05 | Princeton Identity, Inc. | Systems and methods of biometric anaysis with adaptive trigger |
US10366296B2 (en) | 2016-03-31 | 2019-07-30 | Princeton Identity, Inc. | Biometric enrollment systems and methods |
JP6751324B2 (en) * | 2016-09-14 | 2020-09-02 | 株式会社デンソーアイティーラボラトリ | Iris detection device, iris detection method, and program |
US10607096B2 (en) | 2017-04-04 | 2020-03-31 | Princeton Identity, Inc. | Z-dimension user feedback biometric system |
WO2019023032A1 (en) | 2017-07-26 | 2019-01-31 | Princeton Identity, Inc. | Biometric security systems and methods |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0899680A2 (en) * | 1997-08-29 | 1999-03-03 | Eastman Kodak Company | Method for automatic detection of human eyes in digital images |
EP1139270A2 (en) * | 2000-03-30 | 2001-10-04 | Nec Corporation | Method for computing the location and orientation of an object in three-dimensional space |
WO2003003910A1 (en) * | 2001-07-02 | 2003-01-16 | Smart Eye Ab | Method for image analysis |
Family Cites Families (478)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7467809B2 (en) | 1992-05-05 | 2008-12-23 | Automotive Technologies International, Inc. | Vehicular occupant characteristic determination system and method |
US4641349A (en) | 1985-02-20 | 1987-02-03 | Leonard Flom | Iris recognition system |
US4836670A (en) | 1987-08-19 | 1989-06-06 | Center For Innovative Technology | Eye movement detector |
US7512254B2 (en) | 2001-11-07 | 2009-03-31 | Symbol Technologies, Inc. | System and method for mobile biometric authentication |
US6091899A (en) | 1988-09-16 | 2000-07-18 | Canon Kabushiki Kaisha | Apparatus for detecting the direction of visual axis and information selecting apparatus utilizing the same |
US5231674A (en) | 1989-06-09 | 1993-07-27 | Lc Technologies, Inc. | Eye tracking method and apparatus |
CA2054344C (en) | 1990-10-29 | 1997-04-15 | Kazuhiro Itsumi | Video camera having focusing and image-processing function |
GB2250828B (en) | 1990-11-29 | 1994-08-17 | Asahi Optical Co Ltd | Automatic focusing device |
JP2522859B2 (en) | 1990-12-14 | 1996-08-07 | 日産自動車株式会社 | Eye position detection device |
GB2255870B (en) | 1991-05-15 | 1995-11-01 | Asahi Optical Co Ltd | Automatic focusing apparatus |
US5291560A (en) | 1991-07-15 | 1994-03-01 | Iri Scan Incorporated | Biometric personal identification system based on iris analysis |
JP3206980B2 (en) | 1992-09-11 | 2001-09-10 | 旭光学工業株式会社 | Automatic focus control device |
EP0593386A3 (en) | 1992-10-16 | 1996-07-31 | Ibm | Method and apparatus for accessing touch screen desktop objects via fingerprint recognition |
US5608472A (en) | 1992-11-06 | 1997-03-04 | Research Development Foundation | Eye imaging system |
JP2583010B2 (en) | 1993-01-07 | 1997-02-19 | インターナショナル・ビジネス・マシーンズ・コーポレイション | Method of maintaining consistency between local index table and global index table in multi-tier index structure |
JP3500539B2 (en) | 1993-02-25 | 2004-02-23 | 富士通株式会社 | Automatic focus adjustment method for infrared camera |
US6325765B1 (en) | 1993-07-20 | 2001-12-04 | S. Hutson Hay | Methods for analyzing eye |
JPH07128579A (en) | 1993-10-29 | 1995-05-19 | Canon Inc | Detecting method for light of sight, detecting means for line of sight and video device having detecting means for line of sight |
US5572596A (en) | 1994-09-02 | 1996-11-05 | David Sarnoff Research Center, Inc. | Automated, non-invasive iris recognition system and method |
US6714665B1 (en) * | 1994-09-02 | 2004-03-30 | Sarnoff Corporation | Fully automated iris recognition system utilizing wide and narrow fields of view |
JP3635687B2 (en) | 1994-09-07 | 2005-04-06 | 株式会社ニコン | Automatic focusing device |
US6012376A (en) | 1995-04-28 | 2000-01-11 | Raytheon Company | Gun sight system for a military vehicle |
US6363164B1 (en) | 1996-05-13 | 2002-03-26 | Cummins-Allison Corp. | Automated document processing system using full image scanning |
US6493363B1 (en) | 1995-11-09 | 2002-12-10 | The United States Of America As Represented By The Secretary Of Agricultural | Automated counting insect electrocutor |
JPH09212644A (en) | 1996-02-07 | 1997-08-15 | Oki Electric Ind Co Ltd | Iris recognition device and iris recognition method |
US6433818B1 (en) | 1998-11-06 | 2002-08-13 | Fotonation, Inc. | Digital camera with biometric security |
US6503163B1 (en) | 1996-05-15 | 2003-01-07 | Sensar, Inc. | Precision cable drive |
US5717512A (en) | 1996-05-15 | 1998-02-10 | Chmielewski, Jr.; Thomas A. | Compact image steering and focusing device |
US6400835B1 (en) | 1996-05-15 | 2002-06-04 | Jerome H. Lemelson | Taillight mounted vehicle security system employing facial recognition using a reflected image |
US6320610B1 (en) | 1998-12-31 | 2001-11-20 | Sensar, Inc. | Compact imaging device incorporating rotatably mounted cameras |
US6718665B2 (en) | 1996-05-17 | 2004-04-13 | Dimplex North America Limited | Flame simulating assembly |
WO1997046978A1 (en) | 1996-06-06 | 1997-12-11 | British Telecommunications Public Limited Company | Personal identification |
JP3436293B2 (en) | 1996-07-25 | 2003-08-11 | 沖電気工業株式会社 | Animal individual identification device and individual identification system |
US6837436B2 (en) | 1996-09-05 | 2005-01-04 | Symbol Technologies, Inc. | Consumer interactive shopping system |
US6108636A (en) | 1996-10-15 | 2000-08-22 | Iris Corporation Berhad | Luggage handling and reconciliation system using an improved security identification document including contactless communication insert unit |
JP3533308B2 (en) | 1997-02-10 | 2004-05-31 | 株式会社ニデック | Ophthalmic equipment |
US6229905B1 (en) | 1997-03-26 | 2001-05-08 | Oki Electric Industry Co., Ltd. | Animal identification based on irial granule analysis |
US6285780B1 (en) | 1997-03-28 | 2001-09-04 | Oki Electric Industry Co., Ltd. | Apparatus for identifying individual animals and image processing method |
US6144754A (en) | 1997-03-28 | 2000-11-07 | Oki Electric Industry Co., Ltd. | Method and apparatus for identifying individuals |
GB9709883D0 (en) | 1997-05-16 | 1997-07-09 | Ncr Int Inc | User verification system |
US5859686A (en) | 1997-05-19 | 1999-01-12 | Northrop Grumman Corporation | Eye finding and tracking system |
US6516416B2 (en) | 1997-06-11 | 2003-02-04 | Prism Resources | Subscription access system for use with an untrusted network |
US6119096A (en) | 1997-07-31 | 2000-09-12 | Eyeticket Corporation | System and method for aircraft passenger check-in and boarding using iris recognition |
US6246751B1 (en) | 1997-08-11 | 2001-06-12 | International Business Machines Corporation | Apparatus and methods for user identification to deny access or service to unauthorized users |
US7071971B2 (en) | 1997-08-25 | 2006-07-04 | Elbex Video Ltd. | Apparatus for identifying the scene location viewed via remotely operated television camera |
US6007202A (en) | 1997-10-23 | 1999-12-28 | Lasersight Technologies, Inc. | Eye illumination system and method |
EP0910986A1 (en) | 1997-10-24 | 1999-04-28 | BRITISH TELECOMMUNICATIONS public limited company | Imaging apparatus |
US6064752A (en) | 1997-11-04 | 2000-05-16 | Sensar, Inc. | Method and apparatus for positioning subjects before a single camera |
US6069967A (en) | 1997-11-04 | 2000-05-30 | Sensar, Inc. | Method and apparatus for illuminating and imaging eyes through eyeglasses |
US6021210A (en) | 1997-12-01 | 2000-02-01 | Sensar, Inc. | Image subtraction to remove ambient illumination |
US6055322A (en) | 1997-12-01 | 2000-04-25 | Sensor, Inc. | Method and apparatus for illuminating and imaging eyes through eyeglasses using multiple sources of illumination |
US6028949A (en) | 1997-12-02 | 2000-02-22 | Mckendall; Raymond A. | Method of verifying the presence of an eye in a close-up image |
US5953440A (en) | 1997-12-02 | 1999-09-14 | Sensar, Inc. | Method of measuring the focus of close-up images of eyes |
US6154321A (en) | 1998-01-20 | 2000-11-28 | University Of Washington | Virtual retinal display with eye tracking |
US6101477A (en) | 1998-01-23 | 2000-08-08 | American Express Travel Related Services Company, Inc. | Methods and apparatus for a travel-related multi-function smartcard |
US6088470A (en) | 1998-01-27 | 2000-07-11 | Sensar, Inc. | Method and apparatus for removal of bright or dark spots by the fusion of multiple images |
US5978494A (en) | 1998-03-04 | 1999-11-02 | Sensar, Inc. | Method of selecting the best enroll image for personal identification |
JP3271750B2 (en) | 1998-03-05 | 2002-04-08 | 沖電気工業株式会社 | Iris identification code extraction method and device, iris recognition method and device, data encryption device |
US6508397B1 (en) | 1998-03-30 | 2003-01-21 | Citicorp Development Center, Inc. | Self-defense ATM |
US8108307B1 (en) | 1998-03-30 | 2012-01-31 | Citicorp Development Center, Inc. | System, method and apparatus for value exchange utilizing value-storing applications |
US6320612B1 (en) | 1998-05-12 | 2001-11-20 | Jan J. Young | Vehicular camera system with plural perspectives |
US6594399B1 (en) | 1998-05-14 | 2003-07-15 | Sensar, Inc. | Method and apparatus for integrating multiple 1-D filters into a digital image stream interface |
US6957341B2 (en) | 1998-05-14 | 2005-10-18 | Purdue Research Foundation | Method and system for secure computational outsourcing and disguise |
GB9811586D0 (en) | 1998-05-30 | 1998-07-29 | Stevenson Neil J | A vehicle entry/exit control system |
JP2000011163A (en) | 1998-06-18 | 2000-01-14 | Matsushita Electric Ind Co Ltd | Iris image pickup device and method therefor |
JP3296420B2 (en) | 1998-06-18 | 2002-07-02 | 松下電器産業株式会社 | Iris imaging apparatus and iris imaging method thereof |
JP2000005150A (en) | 1998-06-18 | 2000-01-11 | Matsushita Electric Ind Co Ltd | Device and method for iris image pickup |
US6424845B1 (en) | 1998-06-19 | 2002-07-23 | Ncr Corporation | Portable communication device |
US5956122A (en) | 1998-06-26 | 1999-09-21 | Litton Systems, Inc | Iris recognition apparatus and method |
DE69943199D1 (en) | 1998-07-09 | 2011-03-31 | Univ Colorado State Res Found | APPARATUS AND METHOD FOR IMAGING THE NETWORK SKIN VACCULATE |
JP2000023946A (en) | 1998-07-10 | 2000-01-25 | Matsushita Electric Ind Co Ltd | Iris imaging instrument and method therefor |
JP3610234B2 (en) | 1998-07-17 | 2005-01-12 | 株式会社メディア・テクノロジー | Iris information acquisition device and iris identification device |
JP3315648B2 (en) | 1998-07-17 | 2002-08-19 | 沖電気工業株式会社 | Iris code generation device and iris recognition system |
JP2000067237A (en) * | 1998-08-17 | 2000-03-03 | Victor Co Of Japan Ltd | Person identifying device |
JP3562970B2 (en) | 1998-09-10 | 2004-09-08 | 沖電気工業株式会社 | Biological identification device |
JP2000092046A (en) | 1998-09-11 | 2000-03-31 | Mitsubishi Electric Corp | Remote authentication system |
US6134339A (en) | 1998-09-17 | 2000-10-17 | Eastman Kodak Company | Method and apparatus for determining the position of eyes and for correcting eye-defects in a captured frame |
EP1116191B1 (en) | 1998-09-22 | 2002-08-14 | Siemens Aktiengesellschaft | Method and device for verifying the authorization to log onto a system |
JP2000105830A (en) | 1998-09-29 | 2000-04-11 | Oki Electric Ind Co Ltd | Individual identifying device |
JP3383594B2 (en) | 1998-09-29 | 2003-03-04 | 沖電気工業株式会社 | Eye opening measurement device |
JP3337988B2 (en) | 1998-09-29 | 2002-10-28 | 沖電気工業株式会社 | Individual identification device |
US6330674B1 (en) | 1998-09-30 | 2001-12-11 | Compaq Computer Corporation | Use of biometrics as a methodology for defining components for ECC encryption |
US6522772B1 (en) | 1998-09-30 | 2003-02-18 | Ncr Corporation | Self-service checkout terminal having a biometric sensing device for verifying identity of a user and associated method |
JP2000107156A (en) | 1998-10-08 | 2000-04-18 | Hoan Denshi Tsushin Gijutsu Kyokai | Authentication device |
JP3212952B2 (en) | 1998-10-26 | 2001-09-25 | 沖電気工業株式会社 | Iris pattern input device |
US6580356B1 (en) | 1998-11-05 | 2003-06-17 | Eckhard Alt | Advanced personal identification systems and techniques |
US6196972B1 (en) | 1998-11-11 | 2001-03-06 | Spentech, Inc. | Doppler ultrasound method and apparatus for monitoring blood flow |
GB9824697D0 (en) | 1998-11-11 | 1999-01-06 | Ncr Int Inc | Terminal |
JP4258868B2 (en) | 1998-11-16 | 2009-04-30 | 沖電気工業株式会社 | Iris pattern recognition device |
JP3523795B2 (en) | 1998-11-19 | 2004-04-26 | 沖電気工業株式会社 | Access control system |
US6289113B1 (en) | 1998-11-25 | 2001-09-11 | Iridian Technologies, Inc. | Handheld iris imaging apparatus and method |
US6532298B1 (en) | 1998-11-25 | 2003-03-11 | Iridian Technologies, Inc. | Portable authentication device and method using iris patterns |
US6753919B1 (en) | 1998-11-25 | 2004-06-22 | Iridian Technologies, Inc. | Fast focus assessment system and method for imaging |
US6377699B1 (en) | 1998-11-25 | 2002-04-23 | Iridian Technologies, Inc. | Iris imaging telephone security module and method |
US6424727B1 (en) | 1998-11-25 | 2002-07-23 | Iridian Technologies, Inc. | System and method of animal identification and animal transaction authorization using iris patterns |
JP2000182050A (en) | 1998-12-11 | 2000-06-30 | Oki Electric Ind Co Ltd | Individual identifying device for animal |
JP2000185031A (en) | 1998-12-22 | 2000-07-04 | Oki Electric Ind Co Ltd | Individual identification device |
JP3296423B2 (en) | 1998-12-25 | 2002-07-02 | 松下電器産業株式会社 | Access control system |
JP2000189403A (en) | 1998-12-25 | 2000-07-11 | Oki Electric Ind Co Ltd | Iris region extraction and individual identifying device |
US6393136B1 (en) | 1999-01-04 | 2002-05-21 | International Business Machines Corporation | Method and apparatus for determining eye contact |
NL1011509C2 (en) | 1999-01-05 | 2000-07-06 | Nedap Nv | Method for biometric identification of animals. |
KR100320465B1 (en) | 1999-01-11 | 2002-01-16 | 구자홍 | Iris recognition system |
US6332193B1 (en) | 1999-01-18 | 2001-12-18 | Sensar, Inc. | Method and apparatus for securely transmitting and authenticating biometric data over a network |
US6950139B2 (en) | 1999-01-22 | 2005-09-27 | Nikon Corporation | Image reading device and storage medium storing control procedure for image reading device |
TW468123B (en) | 1999-01-29 | 2001-12-11 | Ibm | Method and apparatus for automatic gas pump charging cost of gasoline to auto computer using pan technology |
JP2000237167A (en) | 1999-02-19 | 2000-09-05 | Oki Electric Ind Co Ltd | Iris recognizing device |
JP2000242788A (en) | 1999-02-22 | 2000-09-08 | Matsushita Electric Ind Co Ltd | Iris recognition system |
JP3728386B2 (en) | 1999-03-08 | 2005-12-21 | 沖電気工業株式会社 | Iris recognition device |
KR100319608B1 (en) | 1999-03-09 | 2002-01-05 | 김영환 | A stacked semiconductor package and the fabricating method thereof |
KR100320188B1 (en) | 1999-03-23 | 2002-01-10 | 구자홍 | Forgery judgment method for iris recognition system |
GB9907514D0 (en) | 1999-04-01 | 1999-05-26 | Ncr Int Inc | Self service terminal |
GB9907513D0 (en) | 1999-04-01 | 1999-05-26 | Ncr Int Inc | Transaction recordal and validation |
US6247813B1 (en) | 1999-04-09 | 2001-06-19 | Iritech, Inc. | Iris identification system and method of identifying a person through iris recognition |
JP4037009B2 (en) | 1999-06-16 | 2008-01-23 | 沖電気工業株式会社 | Iris recognition device and identification system |
JP2001005948A (en) | 1999-06-17 | 2001-01-12 | Matsushita Electric Ind Co Ltd | Iris imaging device |
JP3317928B2 (en) | 1999-06-17 | 2002-08-26 | 松下電器産業株式会社 | Hotel management equipment |
JP2001004909A (en) | 1999-06-18 | 2001-01-12 | Olympus Optical Co Ltd | Camera having automatic focusing device |
US6438752B1 (en) | 1999-06-22 | 2002-08-20 | Mediaone Group, Inc. | Method and system for selecting television programs based on the past selection history of an identified user |
WO2001006448A1 (en) | 1999-07-14 | 2001-01-25 | Veridicom, Inc. | Ultra-rugged i.c. sensor and method of making the same |
JP2001034754A (en) | 1999-07-19 | 2001-02-09 | Sony Corp | Iris authentication device |
US6553494B1 (en) | 1999-07-21 | 2003-04-22 | Sensar, Inc. | Method and apparatus for applying and verifying a biometric-based digital signature to an electronic document |
US6516078B1 (en) | 1999-07-29 | 2003-02-04 | Hewlett-Packard Company | Multi-level detection and deterrence of counterfeiting of documents with reduced false detection |
US6120461A (en) | 1999-08-09 | 2000-09-19 | The United States Of America As Represented By The Secretary Of The Army | Apparatus for tracking the human eye with a retinal scanning display, and method thereof |
JP2001067399A (en) | 1999-08-25 | 2001-03-16 | Oki Electric Ind Co Ltd | Electronic money transaction system |
AU7346800A (en) | 1999-09-02 | 2001-03-26 | Automated Business Companies | Communication and proximity authorization systems |
US6370260B1 (en) | 1999-09-03 | 2002-04-09 | Honeywell International Inc. | Near-IR human detector |
US7076088B2 (en) | 1999-09-03 | 2006-07-11 | Honeywell International Inc. | Near-infrared disguise detection |
US6282475B1 (en) | 1999-09-20 | 2001-08-28 | Valdemar L. Washington | System for automatically adjustable devices in an automotive vehicle |
US7391865B2 (en) | 1999-09-20 | 2008-06-24 | Security First Corporation | Secure data parser method and system |
JP2001101429A (en) | 1999-09-28 | 2001-04-13 | Omron Corp | Method and device for observing face, and recording medium for face observing processing |
US6674367B2 (en) | 1999-09-28 | 2004-01-06 | Clifford Sweatte | Method and system for airport and building security |
JP2001118103A (en) | 1999-10-15 | 2001-04-27 | Oki Electric Ind Co Ltd | Gate managing device |
JP2001135422A (en) | 1999-11-08 | 2001-05-18 | Yazaki Corp | Electric shield connector directly attached to apparatus |
CA2392229C (en) | 1999-11-30 | 2016-08-30 | Transforming Technologies, Inc. | Methods, systems, and apparatuses for secure interactions |
US6711562B1 (en) | 1999-12-01 | 2004-03-23 | The Trustees Of Columbia University In The City Of New York | Cache sensitive search (CSS) tree indexing system and method |
US6505193B1 (en) | 1999-12-01 | 2003-01-07 | Iridian Technologies, Inc. | System and method of fast biometric database searching using digital certificates |
US6775774B1 (en) | 1999-12-06 | 2004-08-10 | Bsi 2000, Inc. | Optical card based system for individualized tracking and record keeping |
JP2001167275A (en) | 1999-12-13 | 2001-06-22 | Oki Electric Ind Co Ltd | Individual identifying device |
US6446045B1 (en) | 2000-01-10 | 2002-09-03 | Lucinda Stone | Method for using computers to facilitate and control the creating of a plurality of functions |
US6516087B1 (en) | 2000-01-10 | 2003-02-04 | Sensar, Inc. | Method for real time correlation of stereo images |
US6494363B1 (en) | 2000-01-13 | 2002-12-17 | Ncr Corporation | Self-service terminal |
JP2001222661A (en) | 2000-02-08 | 2001-08-17 | Oki Software Kk | Automatic transaction device |
GB0004287D0 (en) | 2000-02-23 | 2000-04-12 | Leeper Kim | System and method for authenticating electronic documents |
JP2001236324A (en) | 2000-02-24 | 2001-08-31 | Fujitsu Ltd | Portable electronic device with individual authenticating function by biometric information |
EP1297478A2 (en) | 2000-03-15 | 2003-04-02 | Emedicalfiles, Inc. | Web-hosted healthcare medical information management system |
JP3631655B2 (en) | 2000-03-22 | 2005-03-23 | シャープ株式会社 | Solid-state imaging device |
JP3825222B2 (en) | 2000-03-24 | 2006-09-27 | 松下電器産業株式会社 | Personal authentication device, personal authentication system, and electronic payment system |
US20030159051A1 (en) | 2000-03-27 | 2003-08-21 | Wilhelm Hollnagel | Method for generating electronic signatures |
GB0007360D0 (en) | 2000-03-28 | 2000-05-17 | Ncr Int Inc | Electronic wallet |
US6540392B1 (en) | 2000-03-31 | 2003-04-01 | Sensar, Inc. | Micro-illuminator for use with image recognition system |
US6299306B1 (en) | 2000-03-31 | 2001-10-09 | Sensar, Inc. | Method and apparatus for positioning subjects using a holographic optical element |
US6968457B2 (en) | 2000-03-31 | 2005-11-22 | Joseph Wing On Tam | Method for making secured personal identity card and procedures for validation and obtaining secure personal information |
US6441482B1 (en) | 2000-04-11 | 2002-08-27 | Omnivision Technologies, Inc. | Biometric device with integrated CMOS image sensor |
JP2001292981A (en) | 2000-04-12 | 2001-10-23 | Oki Electric Ind Co Ltd | Iris recognition apparatus |
JP2001297177A (en) | 2000-04-17 | 2001-10-26 | Oki Electric Ind Co Ltd | Accounting system |
DE60125759T2 (en) | 2000-04-28 | 2007-10-31 | Precise Biometrics Ab | BIOMETRIC IDENTITY TEST |
US7280984B2 (en) | 2000-05-08 | 2007-10-09 | Phelan Iii Frank | Money card system, method and apparatus |
US20010051924A1 (en) | 2000-05-09 | 2001-12-13 | James Uberti | On-line based financial services method and system utilizing biometrically secured transactions for issuing credit |
EP1285326B1 (en) | 2000-05-16 | 2006-03-08 | Swisscom Mobile AG | Method and terminal for inputting instructions |
US6493669B1 (en) | 2000-05-16 | 2002-12-10 | Delphi Technologies, Inc. | Speech recognition driven system with selectable speech models |
DK1285409T3 (en) | 2000-05-16 | 2005-08-22 | Swisscom Mobile Ag | Process of biometric identification and authentication |
GB0012840D0 (en) | 2000-05-25 | 2000-07-19 | Thirdphase Limited | Method and system for collection and verification of data from plural sites |
EP1290534A2 (en) | 2000-06-02 | 2003-03-12 | Kinetic Sciences Inc. | Method for biometric encryption of e-mail |
US6323761B1 (en) | 2000-06-03 | 2001-11-27 | Sam Mog Son | Vehicular security access system |
US7599847B2 (en) | 2000-06-09 | 2009-10-06 | Airport America | Automated internet based interactive travel planning and management system |
JP4228520B2 (en) | 2000-06-12 | 2009-02-25 | 沖電気工業株式会社 | Iris photography device |
US7120607B2 (en) | 2000-06-16 | 2006-10-10 | Lenovo (Singapore) Pte. Ltd. | Business system and method using a distorted biometrics |
US6836554B1 (en) | 2000-06-16 | 2004-12-28 | International Business Machines Corporation | System and method for distorting a biometric for transactions with enhanced security and privacy |
US20020194131A1 (en) | 2001-06-18 | 2002-12-19 | Dick Richard S. | Method and system for electronically transmitting authorization to release medical information |
US7313529B2 (en) | 2000-06-23 | 2007-12-25 | Medtronic, Inc. | Portable extender for data transmission within a medical device communication system |
MY134895A (en) | 2000-06-29 | 2007-12-31 | Multimedia Glory Sdn Bhd | Biometric verification for electronic transactions over the web |
KR100373850B1 (en) | 2000-10-07 | 2003-02-26 | 주식회사 큐리텍 | Identification system and method using iris, and media that can record computer program sources thereof |
FR2811843B1 (en) | 2000-07-13 | 2002-12-06 | France Telecom | ACTIVATION OF AN INTERACTIVE MULTIMEDIA TERMINAL |
DE50012605D1 (en) | 2000-07-14 | 2006-05-24 | Voice Trust Ag | Method and system for authorizing a commercial transaction |
AUPQ896000A0 (en) | 2000-07-24 | 2000-08-17 | Seeing Machines Pty Ltd | Facial image processing system |
US7552333B2 (en) | 2000-08-04 | 2009-06-23 | First Data Corporation | Trusted authentication digital signature (tads) system |
WO2002013130A1 (en) | 2000-08-07 | 2002-02-14 | Skidata Ag | Reader device for bar-coded authorisation cards |
BE1013637A6 (en) | 2000-08-07 | 2002-05-07 | Smet Francis De | Method for searching for information on the Internet |
JP4469476B2 (en) | 2000-08-09 | 2010-05-26 | パナソニック株式会社 | Eye position detection method and eye position detection apparatus |
EP1410137A2 (en) | 2000-08-09 | 2004-04-21 | Datawipe Management Services Limited | Personal data device and protection system and method for storing and protecting personal data |
FR2829272A1 (en) | 2000-08-22 | 2003-03-07 | France Telecom | DEVICE FOR AUTOMATED CONTROL OF ELECTRICAL EQUIPMENT BY A PERSON |
JP3751806B2 (en) * | 2000-08-24 | 2006-03-01 | 大日本スクリーン製造株式会社 | Image processing apparatus and recording medium |
WO2002019124A1 (en) | 2000-08-30 | 2002-03-07 | Matsushita Electric Industrial Co.,Ltd. | Authentication system, authentication request device, validating device, and service medium |
AU2002212968A1 (en) | 2000-09-12 | 2002-03-26 | Viaken Systems, Inc. | Techniques for providing and obtaining research and development information technology on remote computing resources |
JP4529263B2 (en) | 2000-09-18 | 2010-08-25 | 沖電気工業株式会社 | Iris recognition device |
WO2002025345A2 (en) | 2000-09-22 | 2002-03-28 | Applied Science Fiction | Lens focusing device, system and method for use with multiple light wavelengths |
GB0024274D0 (en) | 2000-10-04 | 2000-11-15 | Harris John W | Vehicle security and occupant safety system |
US7277561B2 (en) * | 2000-10-07 | 2007-10-02 | Qritek Co., Ltd. | Iris identification |
US6819219B1 (en) | 2000-10-13 | 2004-11-16 | International Business Machines Corporation | Method for biometric-based authentication in wireless communication for access control |
JP2002125142A (en) | 2000-10-16 | 2002-04-26 | Matsushita Electric Ind Co Ltd | Iris image pickup device |
JP4487405B2 (en) | 2000-10-16 | 2010-06-23 | 沖電気工業株式会社 | Iris circle detection device and iris pattern encoding device using the same |
JP2002122899A (en) | 2000-10-16 | 2002-04-26 | Matsushita Electric Ind Co Ltd | Iris image pickup apparatus |
JP4649035B2 (en) | 2000-10-18 | 2011-03-09 | 株式会社トプコン | Eye characteristics measuring device |
JP2002122778A (en) | 2000-10-19 | 2002-04-26 | Fuji Electric Co Ltd | Automatic focusing unit and electronic imaging unit |
JP2002133415A (en) | 2000-10-27 | 2002-05-10 | Oki Electric Ind Co Ltd | Individual identifying device |
US6754640B2 (en) | 2000-10-30 | 2004-06-22 | William O. Bozeman | Universal positive pay match, authentication, authorization, settlement and clearing system |
AU2002239758A1 (en) | 2000-11-10 | 2002-06-11 | Medidata Solutions, Inc. | Method and apparatus of assuring informed consent while conducting secure clinical trials |
KR100649303B1 (en) | 2000-11-16 | 2006-11-24 | 엘지전자 주식회사 | Apparatus of taking pictures in iris recognition system based on both of eyes's images |
GB2369205B (en) | 2000-11-17 | 2005-02-02 | Personal Data Prot System Ltd | Personal data device and protection system and method for storing and protecting personal data |
AU2002220402A1 (en) | 2000-11-21 | 2002-06-03 | Margaret Hurley | System and method for transmitting goods, remuneration, and information |
JP2002153445A (en) | 2000-11-21 | 2002-05-28 | Oki Electric Ind Co Ltd | Iris recognition device |
JP2002153444A (en) | 2000-11-21 | 2002-05-28 | Oki Electric Ind Co Ltd | Individual authentication system |
EP1213638A1 (en) | 2000-12-06 | 2002-06-12 | Siemens Aktiengesellschaft | Enabling devices |
US7125335B2 (en) | 2000-12-08 | 2006-10-24 | Igt | Casino gambling system with biometric access control |
US7114080B2 (en) | 2000-12-14 | 2006-09-26 | Matsushita Electric Industrial Co., Ltd. | Architecture for secure remote access and transmission using a generalized password scheme with biometric features |
US20030182182A1 (en) | 2000-12-18 | 2003-09-25 | Kocher Robert W. | Biometrics-based voting |
US6792134B2 (en) | 2000-12-19 | 2004-09-14 | Eastman Kodak Company | Multi-mode digital image processing method for detecting eyes |
US6920237B2 (en) | 2000-12-19 | 2005-07-19 | Eastman Kodak Company | Digital image processing method and computer program product for detecting human irises in an image |
US6930707B2 (en) | 2000-12-22 | 2005-08-16 | International Business Machines Corporation | Digital camera apparatus with biometric capability |
US6867683B2 (en) | 2000-12-28 | 2005-03-15 | Unisys Corporation | High security identification system for entry to multiple zones |
US7028009B2 (en) | 2001-01-17 | 2006-04-11 | Contentguardiholdings, Inc. | Method and apparatus for distributing enforceable property rights |
US7215797B2 (en) | 2001-02-02 | 2007-05-08 | Lg Electronics Inc. | Iris recognition system |
US20020142844A1 (en) | 2001-02-06 | 2002-10-03 | Kerr Michael A. | Biometric broadband gaming system and method |
US20020112177A1 (en) | 2001-02-12 | 2002-08-15 | Voltmer William H. | Anonymous biometric authentication |
US6732278B2 (en) | 2001-02-12 | 2004-05-04 | Baird, Iii Leemon C. | Apparatus and method for authenticating access to a network resource |
JP4453209B2 (en) | 2001-02-27 | 2010-04-21 | 沖電気工業株式会社 | Automatic transaction equipment |
JP3586431B2 (en) | 2001-02-28 | 2004-11-10 | 松下電器産業株式会社 | Personal authentication method and device |
KR100374707B1 (en) | 2001-03-06 | 2003-03-04 | 에버미디어 주식회사 | Method of recognizing human iris using daubechies wavelet transform |
KR100374708B1 (en) | 2001-03-06 | 2003-03-04 | 에버미디어 주식회사 | Non-contact type human iris recognition method by correction of rotated iris image |
JP2002271689A (en) | 2001-03-12 | 2002-09-20 | Sony Corp | Imaging device, and iris control method for the imaging device |
US6845479B2 (en) | 2001-03-14 | 2005-01-18 | Tality Uk Limited | Method for testing for the presence of faults in digital circuits |
US7095901B2 (en) * | 2001-03-15 | 2006-08-22 | Lg Electronics, Inc. | Apparatus and method for adjusting focus position in iris recognition system |
US7271839B2 (en) | 2001-03-15 | 2007-09-18 | Lg Electronics Inc. | Display device of focal angle and focal distance in iris recognition system |
JP2002286650A (en) | 2001-03-27 | 2002-10-03 | Nippon Steel Corp | Apparatus and method for acquiring data for adjustment of formation as well as computer program and computer- readable storage medium |
GB2373943A (en) | 2001-03-28 | 2002-10-02 | Hewlett Packard Co | Visible and infrared imaging camera |
GB0107689D0 (en) | 2001-03-28 | 2001-05-16 | Ncr Int Inc | Self service terminal |
US20020186131A1 (en) | 2001-04-03 | 2002-12-12 | Brad Fettis | Card security device |
JP2002312772A (en) | 2001-04-13 | 2002-10-25 | Oki Electric Ind Co Ltd | Individual identification device and eye forgery judgment method |
JP2002329204A (en) | 2001-04-27 | 2002-11-15 | Oki Electric Ind Co Ltd | System and program for individual authentication |
DE60135879D1 (en) | 2001-04-27 | 2008-10-30 | Bausch & Lomb | IRIS PATTERN RECOGNITION AND ORIENTATION |
US20020158750A1 (en) | 2001-04-30 | 2002-10-31 | Almalik Mansour Saleh | System, method and portable device for biometric identification |
JP2002334325A (en) | 2001-05-11 | 2002-11-22 | Matsushita Electric Ind Co Ltd | Method and device for picking up image to be authenticated |
JP2002341406A (en) | 2001-05-11 | 2002-11-27 | Matsushita Electric Ind Co Ltd | Method and device for imaging object to be authenticated |
US20040193893A1 (en) | 2001-05-18 | 2004-09-30 | Michael Braithwaite | Application-specific biometric templates |
US20020175182A1 (en) | 2001-05-23 | 2002-11-28 | Matthews Shaun Kerry | Self contained dispenser incorporating a user monitoring system |
US20020194128A1 (en) | 2001-06-14 | 2002-12-19 | Michael Maritzen | System and method for secure reverse payment |
JP2003006628A (en) | 2001-06-20 | 2003-01-10 | Oki Electric Ind Co Ltd | Iris recognizing device |
US20020198731A1 (en) | 2001-06-26 | 2002-12-26 | Barnes Jessica M. | Method and apparatus for processing an international passenger |
JP2003016434A (en) | 2001-06-27 | 2003-01-17 | Matsushita Electric Ind Co Ltd | Individual authenticating device |
JP2003016435A (en) | 2001-06-27 | 2003-01-17 | Matsushita Electric Ind Co Ltd | Device and method for individual authentication |
US20040233038A1 (en) | 2001-07-10 | 2004-11-25 | American Express Travel Related Services Company, Inc. | Method and system for retinal scan recognition biometrics on a fob |
US6523165B2 (en) | 2001-07-13 | 2003-02-18 | Numerical Technologies, Inc. | Alternating phase shift mask design conflict resolution |
JP2003030659A (en) | 2001-07-16 | 2003-01-31 | Matsushita Electric Ind Co Ltd | Iris authentication device and iris image pickup device |
JP2003036434A (en) | 2001-07-23 | 2003-02-07 | Nagano Kogaku Kenkyusho:Kk | Optical system for portable individual authenticating device |
JP2003037765A (en) | 2001-07-24 | 2003-02-07 | Matsushita Electric Ind Co Ltd | Iris imager device |
JP2003037766A (en) | 2001-07-24 | 2003-02-07 | Matsushita Electric Ind Co Ltd | Iris imager |
CA2354614C (en) | 2001-07-30 | 2002-12-10 | Silvano Pregara | Autofocus sensor |
US20030169334A1 (en) | 2001-08-06 | 2003-09-11 | Michael Braithwaite | Iris capture device having expanded capture volume |
US7190825B2 (en) | 2001-08-17 | 2007-03-13 | Geo-Rae Co., Ltd. | Portable communication device for stereoscopic image display and transmission |
RU2273051C2 (en) | 2001-08-21 | 2006-03-27 | Дайболд, Инкорпорейтед | System and method for checking deposits in banking machines |
US20030046228A1 (en) | 2001-08-28 | 2003-03-06 | Jean-Marc Berney | User-wearable functional jewelry with biometrics and smartcard to remotely sign and/or authenticate to e-services |
JP2003085084A (en) | 2001-09-12 | 2003-03-20 | Sony Corp | Contents delivery system and method, portable terminal, delivery server, and recording medium |
US6690997B2 (en) | 2001-09-13 | 2004-02-10 | M.A. Rivalto, Inc. | System for automated package-pick up and delivery |
JP2003099693A (en) | 2001-09-20 | 2003-04-04 | Fujitsu Ltd | Electronic settlement method |
US7058209B2 (en) | 2001-09-20 | 2006-06-06 | Eastman Kodak Company | Method and computer program product for locating facial features |
US7203343B2 (en) | 2001-09-21 | 2007-04-10 | Hewlett-Packard Development Company, L.P. | System and method for determining likely identity in a biometric database |
US7269737B2 (en) | 2001-09-21 | 2007-09-11 | Pay By Touch Checking Resources, Inc. | System and method for biometric authorization for financial transactions |
JP2003108720A (en) | 2001-09-26 | 2003-04-11 | Ricoh Co Ltd | Work flow support system, its method, work flow support program, and computer readable recording medium with the program recorded thereon |
JP2003108983A (en) | 2001-09-28 | 2003-04-11 | Matsushita Electric Ind Co Ltd | Eye image pickup device, iris authentication device, and portable terminal device with iris authentication function |
US20030065626A1 (en) | 2001-09-28 | 2003-04-03 | Allen Karl H. | User verification for conducting health-related transactions |
US20030073499A1 (en) | 2001-10-02 | 2003-04-17 | Kenneth Reece | Network gaming device and method for allowing a player to participate in a live game over a network |
JP2003111749A (en) | 2001-10-09 | 2003-04-15 | Bmf:Kk | Device for discriminating human |
US20030071743A1 (en) | 2001-10-12 | 2003-04-17 | Singapore Technologies Electronics Limited | Aircraft monitoring and incident management system |
US20030074317A1 (en) | 2001-10-15 | 2003-04-17 | Eyal Hofi | Device, method and system for authorizing transactions |
WO2003034652A1 (en) | 2001-10-17 | 2003-04-24 | Byers James T | Method and apparatus for providing biometric information as a signature to a contract |
US6708176B2 (en) | 2001-10-19 | 2004-03-16 | Bank Of America Corporation | System and method for interactive advertising |
FR2831302A1 (en) | 2001-10-19 | 2003-04-25 | St Microelectronics Sa | CODING OF CONCENTRIC INFORMATION |
FR2831351A1 (en) | 2001-10-19 | 2003-04-25 | St Microelectronics Sa | Bandpass filter for detecting texture of iris, has passband which is real, bi-dimensional, oriented along phase axis and resulting from product of pair of identical one-dimensional Hamming windows having specified transfer function |
JP2003132355A (en) | 2001-10-25 | 2003-05-09 | Matsushita Electric Ind Co Ltd | Iris authenticating method and its device |
US20030080194A1 (en) | 2001-10-25 | 2003-05-01 | O'hara Sean M. | Biometric water mixing valve |
AU2002365224A1 (en) | 2001-11-06 | 2003-09-02 | Radian Inc. | Physiomagnetometric inspection and surveillance system and method |
US6598971B2 (en) | 2001-11-08 | 2003-07-29 | Lc Technologies, Inc. | Method and system for accommodating pupil non-concentricity in eyetracker systems |
US20030092489A1 (en) | 2001-11-09 | 2003-05-15 | Veradej Annusorn Andy | Interactive gaming with biometric verification |
JP2003150942A (en) | 2001-11-16 | 2003-05-23 | Kiyomi Nakamura | Eye position tracing method |
JP3904903B2 (en) | 2001-11-21 | 2007-04-11 | 三菱電機株式会社 | Personal authentication device and personal authentication method |
US20030099379A1 (en) | 2001-11-26 | 2003-05-29 | Monk Bruce C. | Validation and verification apparatus and method |
JP4068334B2 (en) | 2001-11-26 | 2008-03-26 | 日本電気株式会社 | Fingerprint authentication method, fingerprint authentication system, and biometric authentication system |
CN1592911A (en) | 2001-11-26 | 2005-03-09 | 罗纳德·H·鲍尔 | Portable messaging device adapted to perform financial transactions |
SG124246A1 (en) | 2001-11-26 | 2006-08-30 | Inventio Ag | System for security control and/or transportation of persons with an elevator installation, method of operating this system, and method of retro-fitting an elevator installation with this system |
US20030210139A1 (en) | 2001-12-03 | 2003-11-13 | Stephen Brooks | Method and system for improved security |
KR100453943B1 (en) | 2001-12-03 | 2004-10-20 | 주식회사 세넥스테크놀로지 | Iris image processing recognizing method and system for personal identification |
KR100456619B1 (en) | 2001-12-05 | 2004-11-10 | 한국전자통신연구원 | A system for registering and authenticating human face using support vector machines and method thereof |
US7239726B2 (en) | 2001-12-12 | 2007-07-03 | Sony Corporation | System and method for effectively extracting facial feature information |
US20030115148A1 (en) | 2001-12-13 | 2003-06-19 | Takhar Harinder Singh | Method and apparatus for processing a secure transaction |
US7003669B2 (en) | 2001-12-17 | 2006-02-21 | Monk Bruce C | Document and bearer verification system |
KR100464081B1 (en) | 2001-12-20 | 2004-12-30 | 엘지전자 주식회사 | Automatic upgrade method of iris code database for iris recognition system |
US20030116630A1 (en) | 2001-12-21 | 2003-06-26 | Kba-Giori S.A. | Encrypted biometric encoded security documents |
US20030125057A1 (en) | 2001-12-27 | 2003-07-03 | Pesola Troy Raymond | System and method for automatic synchronization of managed data |
KR100854890B1 (en) | 2001-12-28 | 2008-08-28 | 엘지전자 주식회사 | Iris recording and recognition method using of several led for iris recognition system |
US20030126560A1 (en) | 2001-12-28 | 2003-07-03 | Koninklijke Philips Electronics N.V. | Adaptive bookmarking of often-visited web sites |
AU2003201231A1 (en) | 2002-01-04 | 2003-07-30 | Lab 7 Networks, Inc. | Communication security system |
US7506172B2 (en) | 2002-01-07 | 2009-03-17 | Igt | Gaming device with biometric system |
AU2003235641A1 (en) | 2002-01-16 | 2003-07-30 | Iritech, Inc. | System and method for iris identification using stereoscopic face recognition |
US20030158762A1 (en) | 2002-01-18 | 2003-08-21 | Jiang Wu | System and method for airplane security / service / maintenance management |
JP4275344B2 (en) | 2002-01-22 | 2009-06-10 | 富士フイルム株式会社 | Imaging apparatus, imaging method, and program |
US20030140928A1 (en) | 2002-01-29 | 2003-07-31 | Tuan Bui | Medical treatment verification system and method |
US20030149881A1 (en) | 2002-01-31 | 2003-08-07 | Digital Security Inc. | Apparatus and method for securing information transmitted on computer networks |
US20030141411A1 (en) | 2002-01-31 | 2003-07-31 | Ashish Pandya | Novel method to secure airline travel |
KR100954640B1 (en) | 2002-02-05 | 2010-04-27 | 파나소닉 주식회사 | Personal authentication method and device |
EP1495386A2 (en) | 2002-02-07 | 2005-01-12 | Eyeticket Corporation | System and method for automated biometric data collection |
US20030158821A1 (en) | 2002-02-15 | 2003-08-21 | Maia Regis Pires | Equivalence and data verification system |
JP2003242125A (en) | 2002-02-18 | 2003-08-29 | Canon Inc | Portable information terminal, authentication auxiliary terminal and individual authentication method |
US20030225711A1 (en) | 2002-02-20 | 2003-12-04 | Martin Paping | Method and apparatus for postal user identification and billing |
KR100842501B1 (en) | 2002-02-21 | 2008-07-01 | 엘지전자 주식회사 | Indicated device of iris location for iris recognition system |
US7206431B2 (en) | 2002-02-22 | 2007-04-17 | Symbol Technologies, Inc. | System and method for generating and verifying a self-authenticating document |
US20030182151A1 (en) | 2002-02-26 | 2003-09-25 | Neal Taslitz | Method of using biometric measurements as a legal seal for authenticating real estate deeds and mortgages |
US6905411B2 (en) | 2002-02-27 | 2005-06-14 | Igt | Player authentication for cashless gaming machine instruments |
US20030163739A1 (en) | 2002-02-28 | 2003-08-28 | Armington John Phillip | Robust multi-factor authentication for secure application environments |
JP2003271940A (en) | 2002-03-13 | 2003-09-26 | Oki Electric Ind Co Ltd | Iris recognition device |
JP2003271565A (en) | 2002-03-15 | 2003-09-26 | Matsushita Electric Ind Co Ltd | Individual authentication system, individual authentication terminal, reader and individual authentication method |
US20030174049A1 (en) | 2002-03-18 | 2003-09-18 | Precision Dynamics Corporation | Wearable identification appliance that communicates with a wireless communications network such as bluetooth |
US7204425B2 (en) | 2002-03-18 | 2007-04-17 | Precision Dynamics Corporation | Enhanced identification appliance |
US8086867B2 (en) | 2002-03-26 | 2011-12-27 | Northrop Grumman Systems Corporation | Secure identity and privilege system |
AUPS140502A0 (en) * | 2002-03-27 | 2002-05-09 | Seeing Machines Pty Ltd | Method for automatic detection of facial features |
US20030189480A1 (en) | 2002-04-04 | 2003-10-09 | Laurence Hamid | Remote actuation system, device and method |
US20030189481A1 (en) | 2002-04-04 | 2003-10-09 | Laurence Hamid | Remote actuation system, device and method |
US20040052418A1 (en) | 2002-04-05 | 2004-03-18 | Bruno Delean | Method and apparatus for probabilistic image analysis |
JP4154917B2 (en) | 2002-04-15 | 2008-09-24 | 沖電気工業株式会社 | Iris photography device |
JP2003308522A (en) | 2002-04-15 | 2003-10-31 | Oki Electric Ind Co Ltd | Iris recognizer |
KR20030082128A (en) | 2002-04-16 | 2003-10-22 | 엘지전자 주식회사 | System of mouse include iris recognition of pc |
US7145457B2 (en) | 2002-04-18 | 2006-12-05 | Computer Associates Think, Inc. | Integrated visualization of security information for an individual |
US20030236120A1 (en) | 2002-04-19 | 2003-12-25 | Kenneth Reece | Method and device for determining the physical location and identity of a user |
JP2003317084A (en) * | 2002-04-19 | 2003-11-07 | Nec Corp | System, method and program for detecting gaze from face image |
KR100438841B1 (en) | 2002-04-23 | 2004-07-05 | 삼성전자주식회사 | Method for verifying users and updating the data base, and face verification system using thereof |
JP4008282B2 (en) | 2002-04-24 | 2007-11-14 | 沖電気工業株式会社 | Pupil / iris circle detector |
JP2003323607A (en) | 2002-04-30 | 2003-11-14 | Matsushita Electric Ind Co Ltd | Iris image pickup device |
US6745520B2 (en) | 2002-05-10 | 2004-06-08 | John L. Puskaric | Integrated rapid access entry/egress system |
JP2003331265A (en) | 2002-05-14 | 2003-11-21 | Matsushita Electric Ind Co Ltd | Eye image imaging device and iris authentication device |
AUPS254302A0 (en) | 2002-05-24 | 2002-06-13 | Resmed Limited | A sleepiness test |
US7519819B2 (en) | 2002-05-29 | 2009-04-14 | Digimarc Corporatino | Layered security in digital watermarking |
US7472283B2 (en) | 2002-05-30 | 2008-12-30 | Hewlett-Packard Development Company, L.P. | Method and apparatus for secured digital video and access tracking |
JP2004005167A (en) | 2002-05-31 | 2004-01-08 | Matsushita Electric Ind Co Ltd | Eye position specification method and device |
JP2004021406A (en) | 2002-06-13 | 2004-01-22 | Matsushita Electric Ind Co Ltd | Eye position specifying method, and authentication apparatus |
JP2004023733A (en) | 2002-06-20 | 2004-01-22 | Canon Inc | Image photographing device and its control method |
JP4174244B2 (en) | 2002-06-20 | 2008-10-29 | キヤノン株式会社 | Image capturing apparatus and control method thereof |
US20040005078A1 (en) | 2002-06-21 | 2004-01-08 | Spectra Systems Corporation | Method and apparatus for digitally watermarking images created with a mobile imaging device |
JP3963789B2 (en) * | 2002-06-24 | 2007-08-22 | デジタルファッション株式会社 | Eye detection device, eye detection program, recording medium for recording the program, and eye detection method |
US7177449B2 (en) | 2002-06-26 | 2007-02-13 | Hewlett-Packard Development Company, L.P. | Image correction system and method |
US7898385B2 (en) | 2002-06-26 | 2011-03-01 | Robert William Kocher | Personnel and vehicle identification system using three factors of authentication |
JP2004030334A (en) | 2002-06-26 | 2004-01-29 | Nec Soft Ltd | Method, system and program for biometrics authentication service |
JP2004038305A (en) | 2002-06-28 | 2004-02-05 | Matsushita Electric Ind Co Ltd | Individual identification device |
US7406184B2 (en) | 2002-07-03 | 2008-07-29 | Equinox Corporation | Method and apparatus for using thermal infrared for face recognition |
US8412623B2 (en) | 2002-07-15 | 2013-04-02 | Citicorp Credit Services, Inc. | Method and system for a multi-purpose transactional platform |
TW588243B (en) | 2002-07-31 | 2004-05-21 | Trek 2000 Int Ltd | System and method for authentication |
GB2391681B (en) | 2002-08-01 | 2005-09-21 | Ncr Int Inc | Self-service terminal |
US7169052B2 (en) | 2002-08-05 | 2007-01-30 | Igt | Personalized gaming apparatus and gaming method |
JP3779247B2 (en) | 2002-08-08 | 2006-05-24 | 株式会社リコー | Imaging device |
US7333798B2 (en) | 2002-08-08 | 2008-02-19 | Value Added Communications, Inc. | Telecommunication call management and monitoring system |
DE10237132A1 (en) | 2002-08-13 | 2004-02-26 | BSH Bosch und Siemens Hausgeräte GmbH | Household appliance with biometric identification for control of access by activation and deactivation of a locking mechanism for the appliance door |
US20040037450A1 (en) | 2002-08-22 | 2004-02-26 | Bradski Gary R. | Method, apparatus and system for using computer vision to identify facial characteristics |
KR20040017978A (en) | 2002-08-23 | 2004-03-02 | 삼성전자주식회사 | Refrigerator and control method thereof |
JP2004094575A (en) | 2002-08-30 | 2004-03-25 | Megane Center:Kk | Customer identification system by iris |
US20040042641A1 (en) | 2002-08-30 | 2004-03-04 | Jakubowski Peter Joel | Personnel identity verification system |
US6853444B2 (en) | 2002-08-30 | 2005-02-08 | Waleed S. Haddad | Non-contact optical imaging system for biometric identification |
US7486806B2 (en) | 2002-09-13 | 2009-02-03 | Panasonic Corporation | Iris encoding method, individual authentication method, iris code registration device, iris authentication device, and iris authentication program |
US20040059590A1 (en) | 2002-09-13 | 2004-03-25 | Dwayne Mercredi | Credential promotion |
US20040050930A1 (en) | 2002-09-17 | 2004-03-18 | Bernard Rowe | Smart card with onboard authentication facility |
JP4272863B2 (en) | 2002-09-20 | 2009-06-03 | キヤノン株式会社 | Camera and camera system |
US20040059953A1 (en) | 2002-09-24 | 2004-03-25 | Arinc | Methods and systems for identity management |
US20040117636A1 (en) | 2002-09-25 | 2004-06-17 | David Cheng | System, method and apparatus for secure two-tier backup and retrieval of authentication information |
US7084904B2 (en) | 2002-09-30 | 2006-08-01 | Microsoft Corporation | Foveated wide-angle imaging system and method for capturing and viewing wide-angle images in real time |
AU2003282943A1 (en) | 2002-10-11 | 2004-05-04 | Digimarc Corporation | Systems and methods for recognition of individuals using multiple biometric searches |
JP2004152046A (en) | 2002-10-31 | 2004-05-27 | Oki Electric Ind Co Ltd | User authentication method and biological information recording device, user authentication device, user authentication system, and ticket issuing device |
WO2004042658A1 (en) | 2002-11-07 | 2004-05-21 | Matsushita Electric Industrial Co., Ltd. | Method for cerficating individual, iris registering device, system for certificating iris, and program for cerficating individual |
JP2004163356A (en) | 2002-11-15 | 2004-06-10 | Matsushita Electric Ind Co Ltd | Eyeglasses detecting apparatus and authentication system |
JP2004164483A (en) | 2002-11-15 | 2004-06-10 | Matsushita Electric Ind Co Ltd | Eye image certification device, and access control system and information processor using it |
JP2004171350A (en) | 2002-11-21 | 2004-06-17 | Matsushita Electric Ind Co Ltd | Eye imaging device and information device using the same |
US6820059B2 (en) | 2003-04-08 | 2004-11-16 | Richard Glee Wood | Method for reducing fraud in government benefit programs using a smart card |
US7804982B2 (en) | 2002-11-26 | 2010-09-28 | L-1 Secure Credentialing, Inc. | Systems and methods for managing and detecting fraud in image databases used with identification documents |
JP2004178141A (en) | 2002-11-26 | 2004-06-24 | Hitachi Ltd | Ic card with illicit use preventing function |
US7130452B2 (en) | 2002-12-03 | 2006-10-31 | International Business Machines Corporation | System and method for multi-party validation, authentication and/or authorization via biometrics |
JP2004206444A (en) | 2002-12-25 | 2004-07-22 | Matsushita Electric Ind Co Ltd | Individual authentication method and iris authentication device |
JP2004212431A (en) | 2002-12-27 | 2004-07-29 | Casio Comput Co Ltd | Autofocus apparatus and autofocus method |
KR20040062247A (en) | 2003-01-02 | 2004-07-07 | 엘지전자 주식회사 | Structure of iris recognition camera and using method of the same |
TWI349204B (en) | 2003-01-10 | 2011-09-21 | Panasonic Corp | Group admission system and server and client therefor |
JP2004220376A (en) | 2003-01-15 | 2004-08-05 | Sanyo Electric Co Ltd | Security management method and system, program, and recording medium |
US7542945B2 (en) | 2003-01-15 | 2009-06-02 | Sanmina-Sci Corporation | Authentication device, system and methods |
KR100543699B1 (en) | 2003-01-21 | 2006-01-20 | 삼성전자주식회사 | Method and Apparatus for user authentication |
TWI224287B (en) | 2003-01-23 | 2004-11-21 | Ind Tech Res Inst | Iris extraction method |
US7404086B2 (en) | 2003-01-24 | 2008-07-22 | Ac Technology, Inc. | Method and apparatus for biometric authentication |
JP4128570B2 (en) | 2003-01-28 | 2008-07-30 | 富士通株式会社 | Biometric information verification device |
US7173348B2 (en) | 2003-03-03 | 2007-02-06 | Startech Automotive Anti Theft Systems Ltd. | Device, system and method for preventing vehicle theft |
JP2004261515A (en) | 2003-03-04 | 2004-09-24 | Matsushita Electric Ind Co Ltd | Iris image pickup device |
JP3598109B2 (en) | 2003-03-13 | 2004-12-08 | 松下電器産業株式会社 | Iris code generation method, personal authentication method, iris code registration device, and personal authentication device |
US20030177051A1 (en) | 2003-03-13 | 2003-09-18 | Robin Driscoll | Method and system for managing worker resources |
US7184577B2 (en) | 2003-03-14 | 2007-02-27 | Intelitrac, Inc. | Image indexing search system and method |
JP2004280547A (en) | 2003-03-17 | 2004-10-07 | Matsushita Electric Ind Co Ltd | Individual identification device |
JP2004287621A (en) | 2003-03-19 | 2004-10-14 | Matsushita Electric Ind Co Ltd | Iris recognition system |
US7436986B2 (en) | 2003-03-25 | 2008-10-14 | Bausch & Lomb Incorporated | Positive patient identification |
JP2004318248A (en) | 2003-04-11 | 2004-11-11 | Matsushita Electric Ind Co Ltd | Iris authentication system, iris authentication method, and iris authentication program |
JP2004315127A (en) | 2003-04-14 | 2004-11-11 | Mitsubishi Electric Corp | Elevator controller |
TWI314304B (en) | 2003-05-05 | 2009-09-01 | Inventio Ag | System for security checking or transport of persons by a lift installation and a method for operating this system |
US8171304B2 (en) | 2003-05-15 | 2012-05-01 | Activcard Ireland Limited | Method, system and computer program product for multiple biometric template screening |
JP2005004181A (en) | 2003-05-21 | 2005-01-06 | Fujinon Corp | Visible light/infrared light photographing lens system |
US7421097B2 (en) | 2003-05-27 | 2008-09-02 | Honeywell International Inc. | Face identification verification using 3 dimensional modeling |
US20040252866A1 (en) | 2003-06-10 | 2004-12-16 | Christel-Loic Tisse | Generation of a typical image restored from a set of several images showing the same element |
GB2402840A (en) | 2003-06-10 | 2004-12-15 | Guy Frank Howard Walker | Mobile with wireless key entry system |
US6992562B2 (en) | 2003-06-10 | 2006-01-31 | Visteon Global Technologies, Inc. | Biometric keyless entry system |
JP2005004524A (en) | 2003-06-12 | 2005-01-06 | Oki Electric Ind Co Ltd | Identifying system, and personal authenticating system |
JP2005010826A (en) | 2003-06-16 | 2005-01-13 | Fujitsu Ltd | Authentication terminal device, biometrics information authentication system and biometrics information acquisition system |
JP2005011207A (en) | 2003-06-20 | 2005-01-13 | Matsushita Electric Ind Co Ltd | Ic card, biometrics authentication system and method for authenticating biometrics |
JP2005025577A (en) | 2003-07-03 | 2005-01-27 | Matsushita Electric Ind Co Ltd | Ic card, biometrics authentication system, and biometrics authentication method |
US20050012817A1 (en) | 2003-07-15 | 2005-01-20 | International Business Machines Corporation | Selective surveillance system with active sensor management policies |
JP4462863B2 (en) | 2003-07-16 | 2010-05-12 | 株式会社エヌ・ティ・ティ・データ | Personal authentication device, biometric feature information update method, and program |
US7731093B2 (en) | 2003-08-04 | 2010-06-08 | Canon Kabushiki Kaisha | Image reading/forming apparatus and method |
JP3918788B2 (en) | 2003-08-06 | 2007-05-23 | コニカミノルタフォトイメージング株式会社 | Imaging apparatus and program |
JP3802892B2 (en) | 2003-08-08 | 2006-07-26 | 株式会社シゲマツ | Iris authentication device |
US20050206502A1 (en) | 2003-08-27 | 2005-09-22 | Georg Bernitz | Method and apparatus for releasing a vehicle for a user |
JP2005096744A (en) | 2003-09-01 | 2005-04-14 | Matsushita Electric Ind Co Ltd | Occupant certifying system |
EP1671258A4 (en) | 2003-09-04 | 2008-03-19 | Sarnoff Corp | Method and apparatus for performing iris recognition from an image |
US7933507B2 (en) | 2006-03-03 | 2011-04-26 | Honeywell International Inc. | Single lens splitter camera |
US8098901B2 (en) | 2005-01-26 | 2012-01-17 | Honeywell International Inc. | Standoff iris recognition system |
US7593550B2 (en) | 2005-01-26 | 2009-09-22 | Honeywell International Inc. | Distance iris recognition |
US8442276B2 (en) | 2006-03-03 | 2013-05-14 | Honeywell International Inc. | Invariant radial iris segmentation |
US8045764B2 (en) | 2005-01-26 | 2011-10-25 | Honeywell International Inc. | Expedient encoding system |
US8085993B2 (en) | 2006-03-03 | 2011-12-27 | Honeywell International Inc. | Modular biometrics collection system architecture |
US8064647B2 (en) | 2006-03-03 | 2011-11-22 | Honeywell International Inc. | System for iris detection tracking and recognition at a distance |
US7183895B2 (en) | 2003-09-05 | 2007-02-27 | Honeywell International Inc. | System and method for dynamic stand-off biometric verification |
US8705808B2 (en) | 2003-09-05 | 2014-04-22 | Honeywell International Inc. | Combined face and iris recognition system |
US7362210B2 (en) | 2003-09-05 | 2008-04-22 | Honeywell International Inc. | System and method for gate access control |
US7756301B2 (en) | 2005-01-26 | 2010-07-13 | Honeywell International Inc. | Iris recognition system and method |
KR20050025927A (en) | 2003-09-08 | 2005-03-14 | 유웅덕 | The pupil detection method and shape descriptor extraction method for a iris recognition, iris feature extraction apparatus and method, and iris recognition system and method using its |
JP2005100063A (en) | 2003-09-24 | 2005-04-14 | Sanyo Electric Co Ltd | Authentication device and method |
NO319858B1 (en) | 2003-09-26 | 2005-09-26 | Tandberg Telecom As | Identification procedure |
JP4615272B2 (en) | 2003-09-29 | 2011-01-19 | 富士フイルム株式会社 | Authentication system, program, and building |
JP2005148883A (en) | 2003-11-12 | 2005-06-09 | Dainippon Printing Co Ltd | Pin change and closure release method for ic card |
US7447911B2 (en) | 2003-11-28 | 2008-11-04 | Lightuning Tech. Inc. | Electronic identification key with portable application programs and identified by biometrics authentication |
US7338167B2 (en) | 2003-12-10 | 2008-03-04 | Joslin Diabetes Center, Inc. | Retinal imaging system |
EP1542182A3 (en) | 2003-12-11 | 2006-01-25 | NCR International, Inc. | Automated teller machine |
US20050129286A1 (en) | 2003-12-16 | 2005-06-16 | Hekimian Christopher D. | Technique using eye position and state of closure for increasing the effectiveness of iris recognition authentication systems |
US20050138387A1 (en) | 2003-12-19 | 2005-06-23 | Lam Wai T. | System and method for authorizing software use |
US7443441B2 (en) | 2004-01-05 | 2008-10-28 | Canon Kabushiki Kaisha | Lens apparatus and image-taking system with multiple focus modes |
JP2005195893A (en) | 2004-01-07 | 2005-07-21 | Canon Inc | Imaging apparatus, its control method and program |
JP2005242677A (en) | 2004-02-26 | 2005-09-08 | Ntt Comware Corp | Composite authentication system and method, and program for the same |
JP2005248445A (en) | 2004-03-01 | 2005-09-15 | Matsushita Electric Ind Co Ltd | Coordination authenticating device |
US20050199703A1 (en) | 2004-03-09 | 2005-09-15 | Friedman Lawrence J. | Method and system for a host based smart card |
GB2411980A (en) | 2004-03-10 | 2005-09-14 | Giga Byte Tech Co Ltd | Computer booting using biometrics |
JP3727323B2 (en) | 2004-03-11 | 2005-12-14 | 沖電気工業株式会社 | Passbook handling device |
US20050206501A1 (en) | 2004-03-16 | 2005-09-22 | Michael Farhat | Labor management system and method using a biometric sensing device |
US20050210267A1 (en) | 2004-03-18 | 2005-09-22 | Jun Sugano | User authentication method and system, information terminal device and service providing server, subject identification method and system, correspondence confirmation method and system, object confirmation method and system, and program products for them |
US20050210270A1 (en) | 2004-03-19 | 2005-09-22 | Ceelox, Inc. | Method for authenticating a user profile for providing user access to restricted information based upon biometric confirmation |
US7336806B2 (en) | 2004-03-22 | 2008-02-26 | Microsoft Corporation | Iris-based biometric identification |
JP2005304809A (en) | 2004-04-22 | 2005-11-04 | Matsushita Electric Ind Co Ltd | Eye image pickup device with lighting system |
US8918900B2 (en) | 2004-04-26 | 2014-12-23 | Ivi Holdings Ltd. | Smart card for passport, electronic passport, and method, system, and apparatus for authenticating person holding smart card or electronic passport |
WO2005109344A1 (en) | 2004-05-10 | 2005-11-17 | Matsushita Electric Industrial Co., Ltd. | Method for registering iris, device for registering iris and program for registering iris |
US20050255840A1 (en) | 2004-05-13 | 2005-11-17 | Markham Thomas R | Authenticating wireless phone system |
US7518651B2 (en) | 2004-05-28 | 2009-04-14 | Aptina Imaging Corporation | Multiple image autofocus |
JP3966309B2 (en) * | 2004-07-14 | 2007-08-29 | 松下電器産業株式会社 | Pupil detection device and iris authentication device |
JP4686153B2 (en) | 2004-09-10 | 2011-05-18 | 日立オムロンターミナルソリューションズ株式会社 | Information processing apparatus, fraud detection method, and automatic teller machine |
US7956890B2 (en) | 2004-09-17 | 2011-06-07 | Proximex Corporation | Adaptive multi-modal integrated biometric identification detection and surveillance systems |
JP4225501B2 (en) | 2004-11-15 | 2009-02-18 | 高司 澤口 | Portable personal authentication device and electronic system to which access is permitted by the device |
US7298873B2 (en) | 2004-11-16 | 2007-11-20 | Imageware Systems, Inc. | Multimodal biometric platform |
CN101111748B (en) | 2004-12-03 | 2014-12-17 | 弗卢克公司 | Visible light and ir combined image camera with a laser pointer |
IL165586A0 (en) | 2004-12-06 | 2006-01-15 | Daphna Palti Wasserman | Multivariate dynamic biometrics system |
US7418115B2 (en) | 2004-12-07 | 2008-08-26 | Aoptix Technologies, Inc. | Iris imaging using reflection from the eye |
CN101233459B (en) | 2005-03-17 | 2011-01-26 | 成像器件系统股份有限公司 | Multimodal biometric analysis |
WO2006132686A2 (en) | 2005-06-03 | 2006-12-14 | Sarnoff Corporation | Method and apparatus for designing iris biometric systems for use in minimally |
ATE526212T1 (en) | 2005-07-11 | 2011-10-15 | Volvo Technology Corp | METHOD AND ARRANGEMENT FOR CARRYING OUT DRIVER IDENTITY CHECKS |
US7751598B2 (en) | 2005-08-25 | 2010-07-06 | Sarnoff Corporation | Methods and systems for biometric identification |
US7471451B2 (en) | 2005-10-14 | 2008-12-30 | Flir Systems, Inc. | Multiple field of view optical system |
US20070160266A1 (en) | 2006-01-11 | 2007-07-12 | Jones Michael J | Method for extracting features of irises in images using difference of sum filters |
WO2007103834A1 (en) | 2006-03-03 | 2007-09-13 | Honeywell International, Inc. | Indexing and database search system |
KR101308368B1 (en) | 2006-03-03 | 2013-09-16 | 허니웰 인터내셔널 인코포레이티드 | An iris recognition system having image quality metrics |
US7580620B2 (en) | 2006-05-08 | 2009-08-25 | Mitsubishi Electric Research Laboratories, Inc. | Method for deblurring images using optimized temporal coding patterns |
US7756407B2 (en) | 2006-05-08 | 2010-07-13 | Mitsubishi Electric Research Laboratories, Inc. | Method and apparatus for deblurring images |
JP4207980B2 (en) | 2006-06-09 | 2009-01-14 | ソニー株式会社 | IMAGING DEVICE, IMAGING DEVICE CONTROL METHOD, AND COMPUTER PROGRAM |
US20080005578A1 (en) | 2006-06-29 | 2008-01-03 | Innovya Research & Development Ltd. | System and method for traceless biometric identification |
US7722461B2 (en) | 2006-07-12 | 2010-05-25 | Igt | Method and system for time gaming with skill wagering opportunities |
US20080148030A1 (en) | 2006-12-14 | 2008-06-19 | General Instrument Corporation | Method and System for Configuring Electronic Communication Device |
GB0700468D0 (en) | 2007-01-10 | 2007-02-21 | Mitsubishi Electric Inf Tech | Improved image identification |
US8025399B2 (en) | 2007-01-26 | 2011-09-27 | Aoptix Technologies, Inc. | Combined iris imager and wavefront sensor |
US20080211347A1 (en) | 2007-03-02 | 2008-09-04 | Joshua Isaac Wright | Circuit System With Supply Voltage For Driving An Electromechanical Switch |
US8063889B2 (en) | 2007-04-25 | 2011-11-22 | Honeywell International Inc. | Biometric data collection system |
US20090092283A1 (en) | 2007-10-09 | 2009-04-09 | Honeywell International Inc. | Surveillance and monitoring system |
US8436907B2 (en) | 2008-05-09 | 2013-05-07 | Honeywell International Inc. | Heterogeneous video capturing system |
US8213782B2 (en) | 2008-08-07 | 2012-07-03 | Honeywell International Inc. | Predictive autofocusing system |
US8090246B2 (en) | 2008-08-08 | 2012-01-03 | Honeywell International Inc. | Image acquisition system |
US8280119B2 (en) | 2008-12-05 | 2012-10-02 | Honeywell International Inc. | Iris recognition system using quality metrics |
-
2007
- 2007-02-07 US US11/672,108 patent/US8090157B2/en active Active
- 2007-03-01 WO PCT/US2007/063029 patent/WO2007101269A1/en active Application Filing
- 2007-03-01 JP JP2008558464A patent/JP5155886B2/en active Active
- 2007-03-01 EP EP07757683A patent/EP1991946B1/en active Active
-
2008
- 2008-10-02 KR KR1020087024199A patent/KR101343901B1/en active IP Right Grant
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0899680A2 (en) * | 1997-08-29 | 1999-03-03 | Eastman Kodak Company | Method for automatic detection of human eyes in digital images |
EP1139270A2 (en) * | 2000-03-30 | 2001-10-04 | Nec Corporation | Method for computing the location and orientation of an object in three-dimensional space |
WO2003003910A1 (en) * | 2001-07-02 | 2003-01-16 | Smart Eye Ab | Method for image analysis |
Non-Patent Citations (2)
Title |
---|
BONNEY B ET AL: "Iris pattern extraction using bit planes and standard deviations", SIGNALS, SYSTEMS AND COMPUTERS, 2004. CONFERENCE RECORD OF THE THIRTY-EIGHTH ASILOMAR CONFERENCE ON PACIFIC GROVE, CA, USA NOV. 7-10, 2004, PISCATAWAY, NJ, USA,IEEE, 7 November 2004 (2004-11-07), pages 582 - 586, XP010779896, ISBN: 0-7803-8622-1 * |
CHOW G ET AL: "TOWARDS A SYSTEM FOR AUTOMATIC FACIAL FEATURE DETECTION", PATTERN RECOGNITION, ELSEVIER, KIDLINGTON, GB, vol. 26, no. 12, 1 December 1993 (1993-12-01), pages 1739 - 1755, XP000420368, ISSN: 0031-3203 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2465881A (en) * | 2008-12-05 | 2010-06-09 | Honeywell Int Inc | A system for iris recognition using a set of quality metrics |
GB2465881B (en) * | 2008-12-05 | 2011-06-22 | Honeywell Int Inc | An iris recognition system using quality metrics |
US8280119B2 (en) | 2008-12-05 | 2012-10-02 | Honeywell International Inc. | Iris recognition system using quality metrics |
Also Published As
Publication number | Publication date |
---|---|
EP1991946A1 (en) | 2008-11-19 |
JP5155886B2 (en) | 2013-03-06 |
JP2009529196A (en) | 2009-08-13 |
KR101343901B1 (en) | 2013-12-20 |
EP1991946B1 (en) | 2011-12-07 |
US8090157B2 (en) | 2012-01-03 |
KR20080108116A (en) | 2008-12-11 |
US20070189582A1 (en) | 2007-08-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8090157B2 (en) | Approaches and apparatus for eye detection in a digital image | |
CN109816644B (en) | Bearing defect automatic detection system based on multi-angle light source image | |
Doyle et al. | Robust detection of textured contact lenses in iris recognition using BSIF | |
JP5107045B2 (en) | Method for identifying a pixel representing an iris in an image acquired for the eye | |
CN105067638B (en) | Tire fetal membrane face character defect inspection method based on machine vision | |
Haindl et al. | Unsupervised detection of non-iris occlusions | |
JP2009523265A (en) | Method for extracting iris features in an image | |
EP3440593B1 (en) | Method and apparatus for iris recognition | |
JP2007188504A (en) | Method for filtering pixel intensity in image | |
KR101315646B1 (en) | Method and apparatus for finger vein pattern extraction using guided gabor filter | |
Abate et al. | BIRD: Watershed based iris detection for mobile devices | |
Short et al. | Latent fingerprint segmentation using ridge template correlation | |
CN111968098A (en) | Strip steel surface defect detection method, device and equipment | |
CN110555875A (en) | Pupil radius detection method and device, computer equipment and storage medium | |
CN110889355A (en) | Face recognition verification method, system and storage medium | |
US11375133B2 (en) | Automatic exposure module for an image acquisition system | |
CN111353992B (en) | Agricultural product defect detection method and system based on textural features | |
Ives et al. | Iris recognition using histogram analysis | |
Alonso-Fernandez et al. | Quality factors affecting iris segmentation and matching | |
Lin et al. | Detection of surface flaws on textured LED lenses using wavelet packet transform based partial least squares techniques | |
Joshi et al. | A novel approach implementation of eyelid detection in biometric applications | |
Wang et al. | Liveness detection of dorsal hand vein based on the analysis of Fourier spectral | |
KR101704717B1 (en) | Apparatus for recognizing iris and operating method thereof | |
Moravcik et al. | Detection of determined eye features in digital image | |
El Sayed et al. | A method for contactless palm ROI extraction |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2007757683 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2008558464 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020087024199 Country of ref document: KR |