US20060274949A1 - Using photographer identity to classify images - Google Patents
Using photographer identity to classify images Download PDFInfo
- Publication number
- US20060274949A1 US20060274949A1 US11/143,541 US14354105A US2006274949A1 US 20060274949 A1 US20060274949 A1 US 20060274949A1 US 14354105 A US14354105 A US 14354105A US 2006274949 A1 US2006274949 A1 US 2006274949A1
- Authority
- US
- United States
- Prior art keywords
- photographer
- image
- interest
- person
- digital
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/35—Categorising the entire scene, e.g. birthday party or wedding scene
- G06V20/38—Outdoor scenes
Definitions
- the invention relates to image processing and more particularly relates to the identification or recognition of a person of interest in a collection of digital images or videos.
- Descriptors of a photographer can be detected and/or recorded in an image capture device such as a digital camera.
- U.S. Patent Application Publication US 2003/235411A1 discloses an imaging apparatus wherein a biological recording means records biological information of a photographer and subsequently uses this information to control the imaging apparatus.
- U.S. Pat. No. 6,433,818 describes a camera with built-in apparatus for acquiring measured biometric data of a potential user, comparing it with corresponding biometric signature data entered in the camera, and allowing camera operation only if there is a match. This function provides security and prevents unauthorized access to the data stored on the camera, thereby discouraging theft.
- U.S. Patent Application Publication US 2002/0080256A1 describes a digital camera with built-in apparatus for biometric identification of the photographer and for recording the identification of the photographer of an image. Furthermore, this feature can be used to discourage theft or piracy of the camera. Preferably, an image of the photographer's iris is captured to determine the identity of the photographer.
- This object is achieved by a method of identifying at least one particular person of interest in a collection of digital images taken over time, comprising:
- FIG. 1 is a schematic diagram of computer system that can implement the present invention
- FIG. 2 is a flow chart of an embodiment of the present invention
- FIGS. 3A and 3B are appearance models of a person of interest having different associated times
- FIG. 4 is a more detailed flow chart of an embodiment of the present invention.
- FIG. 5 is flow chart of an alternative embodiment of the present invention.
- FIGS. 6A-6D illustrate four different image orientations.
- the present invention can be implemented in computer hardware and computerized equipment.
- the method can be performed in a digital camera or other digital image capture devices, a digital printer, on an internet server, on a kiosk, and on a personal computer.
- FIG. 1 there is illustrated a computer system for implementing the present invention.
- the computer system is shown for the purpose of illustrating a preferred embodiment, the present invention is not limited to the computer system shown, but can be used on any electronic processing system such as found in digital cameras, cell phones, portable computing devices, home computers, kiosks, retail or wholesale photofinishing, or any other system for the processing of digital images.
- the computer system includes a microprocessor-based unit 20 (also referred to herein as a digital image processor) for receiving and processing software programs and for performing other processing functions.
- the digital image processor 20 processes images from image capture devices 10 such as cameras, scanners, or computer image generation software.
- the digital image processor 20 can be used to process digital images to make adjustments for overall brightness, tone scale, image structure, etc. of digital images in a manner such that a pleasing looking image is produced by an image output device.
- the digital image processor 20 interfaces with the general control computer 40 (also a microprocessor based unit) for exchanging data and commands.
- the general control computer 40 and the digital image processor 20 can be two different microprocessors, or the functions of each can be performs by a single physical microprocessor.
- the digital image processor 20 often outputs an image to an image output device 30 for example a printer for displaying the image.
- a display device 50 is electrically connected to the digital image processor 20 for displaying user-related information associated with the software, e.g., by means of a graphical user interface.
- a keyboard 60 is also connected to the microprocessor based unit 20 via the general control computer 40 for permitting a user to input information to the software.
- a mouse can be used for moving a selector on the display device 50 and for selecting an item on which the selector overlays, as is well known in the art.
- a compact disk-read only memory which typically includes software programs, is inserted into the general control computer 40 for providing a means of inputting the software programs and other information to the general control computer 40 and the digital image processor 20 .
- a floppy disk can also include a software program, and is inserted into the general control computer 40 for inputting the software program.
- the general control computer 40 can be programmed, as is well known in the art, for storing the software program internally.
- the general control computer 40 can have a network connection, such as a telephone line or wireless connection, to an external network, such as a local area network or the Internet.
- Images can also be displayed on the display device 50 via a personal computer card (PC Card), such as, as it was formerly known, a PCMCIA card (based on the specifications of the Personal Computer Memory Card International Association) which contains digitized images electronically embodied in the card.
- PC Card personal computer card
- PCMCIA card based on the specifications of the Personal Computer Memory Card International Association
- the image output device 30 provides a final image.
- the output device 30 can be a printer or other output device that provides a paper or other hard copy final image.
- the output device 30 can also be an output device that provides the final image as a digital file.
- the output device 30 can also include combinations of output, such as a printed image and a digital file on a memory unit, such as a CD or DVD.
- Digital images and other data are stored on an offline memory device 70 such as an external hard drive, flash media, a drive that writes to CD-ROM or DVD media, or the like.
- a digital image includes one or more digital image channels or color components.
- Each digital image channel is a two-dimensional array of pixels.
- Each pixel value relates to the amount of light received by the imaging capture device corresponding to the physical region of pixel.
- a digital image will often consist of red, green, and blue digital image channels.
- Motion imaging applications can be thought of as a sequence of digital images.
- a digital image channel is described as a two dimensional array of pixel values arranged by rows and columns, those skilled in the art will recognize that the present invention can be applied to non rectilinear arrays with equal effect.
- the present invention can be implemented in a combination of software and/or hardware and is not limited to devices which are physically connected and/or located within the same physical location.
- One or more of the devices illustrated in FIG. 1 can be located remotely and can be connected via a network.
- One or more of the devices can be connected wirelessly, such as by a radio-frequency link, either directly or via a network.
- the present invention can be employed in a variety of user contexts and environments.
- Exemplary contexts and environments include, without limitation, wholesale digital photofinishing (which involves exemplary process steps or stages such as film in, digital processing, prints out), retail digital photofinishing (film in, digital processing, prints out), home printing (home scanned film or digital images, digital processing, prints out), desktop software (software that applies algorithms to digital prints to make them better—or even just to change them), digital fulfillment (digital images in—from media or over the web, digital processing, with images out—in digital form on media, digital form over the web, or printed on hard-copy prints), kiosks (digital or scanned input, digital processing, digital or hard copy output), mobile devices (e.g., PDA or cell phone that can be used as a processing unit, a display unit, or a unit to give processing instructions), and as a service offered via the World Wide Web.
- wholesale digital photofinishing which involves exemplary process steps or stages such as film in, digital processing, prints out
- retail digital photofinishing film in,
- the invention can stand alone or can be a component of a larger system solution.
- human interfaces e.g., the scanning or input, the digital processing, the display to a user (if needed), the input of user requests or processing instructions (if needed), the output, can each be on the same or different devices and physical locations, and communication between the devices and locations can be via public or private network connections, or media based communication.
- the method of the invention can be fully automatic, can have user input (be fully or partially manual), can have user or operator review to accept/reject the result, or can be assisted by metadata (metadata that can be user supplied, supplied by a measuring device (e.g. in a camera), or determined by an algorithm).
- the algorithm(s) can interface with a variety of workflow user interface schemes.
- a first embodiment of this method is illustrated in FIG. 2 .
- a digital image and video collection 104 is analyzed to recognize the identities of the people therein. Recognition of the people in a collection of the digital images or videos 104 allows for the creation of a database 120 relating digital images or videos 104 and the identities of the people in the digital images or videos 104 . Consequently, the database 120 can be queried 122 to find those images and videos containing the query person of interest.
- the database 120 can be in any form.
- the database 120 can be distributed across many files, by for example, inserting a person classification 118 into the file header of the digital image or video 104 .
- a query 122 for images of “Jim” returns query results 124 containing the set of digital images or videos 104 containing the person of interest “Jim”. Then the query results 124 can be returned.
- the query results 124 are the set of digital images and videos 104 containing the query person of interest.
- the digital images and videos 104 of the query result 124 can be displayed on a display device 50 as shown in FIG. 1 for review by the user.
- the digital image or video 104 is passed to the capture information extractor 108 .
- the capture information extractor 108 determines the capture information 112 related to the capture of the digital image and videos 104 of the collection, and outputs the resulting information, where capture information 112 related to a particular digital image or video 104 is associated with that particular digital image or video 104 .
- the capture information 112 can be the time a particular digital image or video 104 was captured, the location the image or video was captured, the focal length, the subject distance, the number of rows and columns of pixels, the magnification, whether the flash was fired, whether the self-timer was activated, etc.
- the capture information 112 for a particular image or video is determined by one of several methods by the capture information extractor 108 .
- capture information 112 is embedded in the file header of the digital image or video 104 .
- the EXIF image format (described at www.exif.org) allows the image or video capture device to store information associated with the image or video in the file header.
- the “Date ⁇ Time” entry is associated with the date and time the image was captured.
- the digital image or video results from scanning film and the image capture time is determined by detection of the date printed into the image (as is often done at capture time) area, usually in the lower left corner of the image.
- the date a photograph is printed is often printed on the back of the print.
- some film systems contain a magnetic layer in the film for storing information such as the capture date.
- the capture information extractor 108 uses the most appropriate method for extracting the capture information 112 of the image.
- the source of the digital images and videos is a digital camera, and the capture information extractor 108 extracts the capture information from the image file information.
- the images and videos of the collection are also analyzed by the photographer determiner 138 to determine the identity of the particular photographer for each image and video.
- the identity of the photographer can be stored in the “Camera Owner”, “Image Creator”, “Photographer”, or “Copyright” tags for example.
- the identity of the photographer of an image or video can be entered manually before, during, or after capturing the video.
- several cameras e.g. in U.S. Patent Application Publication US 2002/0080256A1 have been described that contain means for extracting biometric information from the photographer, identifying the photographer, and then annotating the image with the identity of the photographer.
- the photographer determiner 138 discovers the identity of the photographer 132 and passes that information to the individual recognition classifier 114 .
- the photographer 132 can not be able to be identified by the photographer determiner 138 .
- the photographer 132 is “unknown”. For example, this situation can occur when a person who owns the camera is on vacation and asks a stranger to use her (the vacationing camera owner) camera to capture an image of her in front of a landmark.
- a camera such as described in U.S. Patent Application US 2002/0080256A1 can only feasibly identify the photographer from a small set of potential camera users (e.g.
- the primary user is probably the camera owner, and secondary users are friends and family of the camera owner) whose profiles are known by the camera.
- an image captured by a stranger using the camera would simply be identified by the photographer determiner 138 as having an “unknown” photographer 132 .
- Image capture time can be included in the capture information 112 and can be a precise minute in time, e.g. Mar. 27, 2004 at 10:17 AM. Or the image capture time can be less precise, e.g. 2004 or March 2004.
- the image capture time can be in the form of a probability distribution function e.g. Mar. 27, 2004+/ ⁇ 2 days with 95% confidence.
- the image capture time is input to the individual recognition classifier 114 .
- the person finder 106 identifies one or more persons in the digital video or digital image.
- people are found by face detection.
- Methods for detecting human faces are well known in the art of digital image processing.
- a face detection method for finding human faces in images is described in the following article: Jones, M. J.; Viola, P., “Fast Multi-view Face Detection”, IEEE Conference on Computer Vision and Pattern Recognition ( CVPR ), June 2003.
- the facial features e.g. eyes, nose, mouth, etc.
- facial features can also be used.
- an embodiment can be based upon the facial similarity metric described by M. Turk and A. Pentland in “Eigenfaces for Recognition” Journal of Cognitive Neuroscience . Vol 3, No. 1. 71-86, 1991.
- Facial descriptors are obtained by projecting the image of a face onto a set of principal component functions that describe the variability of facial appearance. The similarity between any two faces is measured by computing the Euclidean distance of the features obtained by projecting each face onto the same set of functions. In this case, an appearance model can describe the mean and covariance of the projection of multiple images of the same face onto these principal component functions.
- the similarity metric (for example computed by comparer 140 of FIG. 4 ) can calculate the Mahalonobis of a new face to the mean of the model.
- the appearance model and similarity measurement could include a combination of several disparate feature types such as Eigenfaces, facial measurements, color/texture information, wavelet features etc.
- a person finder 106 outputs personal features 110 that are quantitative descriptions of a person detected by the person detector 106 .
- the person finder 106 outputs one set of personal features 110 for each detected person.
- the personal features are the locations of 82 feature points associated with specific facial features, found using a method similar to the aforementioned active appearance model of Cootes et al.
- a visual representation of the facial feature points for an image of a face is shown in FIG. 3A and FIG. 3B as an illustration.
- the personal features 110 can additionally be represented with quantifiable descriptors such as eye color, skin color, face shape, etc.
- the personal features 110 can also be distances between specific feature points or angles formed by lines connecting sets of specific feature points. Color cues are easily extracted from the digital image or video 104 once the person and facial features are located by the person finder 106 .
- the personal features 110 are input to the individual recognition classifier 114 .
- a set of appearance models 116 l - 116 M of M persons of interest from a stored digital database is input to the individual recognition classifier 114 for determining the identity of each of the persons detected by the person finder 106 and represented by the personal features 110 .
- the appearance models 116 l - 116 M contain a set of features associated with a particular person of interest.
- Each appearance model for an individual of interest can include a number of sub-models, for example models describing the appearance of the person of interest at different ages as described in above-cited co-pending application U.S. patent application Ser. No. 11/116,729. Or the sub-models could describe the appearance of the person of interest with different appearances (e.g.
- a person of interest is a person that is important to identify in a collection of digital images and videos 104 .
- each child would likely be a person of interest because it would be useful and valuable for a head of the household to identify occurrences of each child in each photo and video 104 of the digital media collection.
- the aforementioned camera owner as primary photographer, can be a person of interest as can be any secondary photographers (i.e. friends and family of the camera owner).
- the purpose of the individual recognition classifier 114 is to use one or more of the appearance models 116 l - 116 M , the image capture information 112 , and the personal features 110 to determine if a person detected by the person finder 106 is a particular person of interest. This is accomplished by comparing the personal features 110 of the detected person with the features of the appearance model 116 of the particular person of interest.
- the individual recognition classifier 114 outputs the person classification 118 .
- the individual recognition classifier outputs a high probability that a person detected by the person finder 106 is a particular person of interest m when the personal features 110 associated with the detected person are similar to the appearance model 116 m associated with the particular person of interest m.
- the person classification 118 is the likelihood (probability) that the person detected by the person finder 106 is the person of interest.
- the person classification 118 can be a binary indicator (e.g. true of false) indicating the classifier's decision on whether the person associated with the personal features 110 is the person of interest.
- the person classification 118 can then be associated with digital image or video containing the detected person. For example, an image can be tagged or annotated in the database 120 as containing the person of interest “Mary”. Furthermore, the person classification 118 information can be associated with the personal features 110 . Furthermore, the image can also be tagged with the location within the digital image or video of the person detected by the person finder 106 that the individual recognition classifier 114 determined was the person of interest. Furthermore, the likelihood or probability of the person classification 118 can be added to the annotation in the database 120 as well.
- the query 122 is made for digital images or videos 104 containing a particular person of interest
- the database 120 is searched and the appropriate query results 124 are returned to the user on a user interface such as the display device 50 of FIG. 1 .
- the individual recognition classifier uses the photographer 132 information as follows: When the image capture information 112 indicates that the camera's self-timer was used to capture a particular image or video, then the individual recognition classifier 114 outputs the person classification 118 indicating a high likelihood (probability) that the digital image or video 104 contains the photographer 132 . For example, the probability is in the range of 0.75 to 1.0, depending on the particular photographer 132 . This results from the fact that most times, the camera self-timer is used to allow time for the photographer 132 to activate the camera then run to be included in the scene captured by the camera in the particular digital image or video 104 .
- the individual recognition classifier 114 outputs a person classification 118 indicating a medium likelihood (probability) that the digital image or video 104 contains the photographer 132 .
- a person classification 118 indicating a medium likelihood (probability) that the digital image or video 104 contains the photographer 132 .
- the camera's image or video capture is activated using remote control to allow time the photographer 132 to be in the scene captured by the camera in the particular digital image or video 104 .
- remote activation of the camera's capture sequence can also be used to avoid shaking the camera. For this reason, the probability associated with the person classification 118 indicating that the photographer 132 is in the digital image or video 104 is in the range of 0.5 to 0.9, depending on the particular photographer 132 .
- FIG. 4 shows a more detailed view of the individual recognition classifier 114 .
- a comparer 140 compares the personal features 110 with each of the M appearance models 116 l-M representing M individuals of interest and outputs a score set 142 of M scores indicating the likelihood or probability that the particular person found by the person finder is the m th person of interest.
- the comparer 140 can for example calculate a Euclidean distance between the personal features 110 and each of the M appearance models.
- the comparer 140 can compute a Mahalanobis distance from each of the M appearance models to produce the score set 142 .
- the comparer 140 can be any classifier, such as but not limited to, Gaussian Maximum Likelihood, a neural network, a support vector machine or the like.
- the m th entry of the score set 142 indicates the probability that the particular face corresponding to the set of personal features 110 is the m th person of interest.
- the score set 142 is passed to a score adjuster 144 for adjusting the score according to information related to the photographer 132 and the capture information 112 , producing an adjusted score set.
- the score adjuster 144 considers the score set likelihoods or probabilities and refines the scores as follows.
- the scores in the score set 142 are adjusted according to the capture information 112 .
- the capture information 112 indicates that the self-timer flag was used to capture the digital image or video 104
- the score for the m th person of interest is increased if the m th person is the photographer 132 of the digital image or video 104 .
- the score adjuster 144 considers the prior probability (i.e.
- the prior probability of an event A is the probability that A occurs in the absence of any other data) of the photographer 132 appearing in the digital image or video 104 when the self-timer is user to capture the digital image or video 104 .
- the score for the m th person of interest is increased if the m th person is the photographer 132 of the digital image or video 104 .
- the score adjuster 144 outputs a person classification 118 for the person detected by the person finder 106 .
- the person classification 118 can be the score set 142 after the adjustment by the score adjuster 144 .
- the person classification 118 can be the identity of the person of interest with the highest corresponding score in the adjusted score set, providing that that score is greater than a threshold T0.
- the individual recognition classifier 114 preferably computes:
- P ( x m
- f ) P ( f
- the individual recognition classifier 114 knows the identity of the photographer 132 , a photographer content profile 130 associated with that photographer 132 can be used by the score adjuster 144 to properly adjust the score set 142 in an appropriate fashion.
- the photographer content profile 130 contains the prior probabilities of persons of interest appearing in images and videos for a particular photographer.
- the score set 142 contains the probabilities P(f
- the photographer content profiles 130 for a particular photographer are generated by having a user manually verify the identities of people in a set of images or videos captured by the particular photographer. Then the prior probabilities are computed and stored in the photographer content profile 130 .
- the prior probabilities can consider the capture information 112 .
- a photographer 132 has one set of prior probabilities that describe the probability of a person being the m th person of interest when the self-timer function is on as indicated by the capture information 112 , and a second set of prior probabilities with the self-timer function is off.
- the following table gives the prior probabilities that a given person in an image or video is the m th person of interest for two different photographers and when the self-timer function is off and on.
- Photographer is person of Photographer is person of interest 0 interest 1 Mth Person Self-timer Self-timer Self-timer Self-timer of interest on off on off 0 0.3 0.0 0.3 0.2 1 0.3 0.3 0.3 0.0 2 0.2 0.4 0.3 0.6 unknown 0.2 0.3 0.1 0.2
- the photographer 132 of the digital image or video 104 is identified along with capture information 112 related to the digital image or video 104 .
- the photographer content profile 130 corresponding to the photographer 132 is used to determine the identities of persons in the digital image or video 104 .
- FIG. 5 shows an alternative embodiment for using the identify of the photographer 132 to identify objects within the digital image or video 104 or to classify the digital image or video itself.
- Digital images or videos 104 from a collection are passed to a feature extractor 152 for extracting features 160 that are passed to an object classifier 164 .
- the appearance models for objects of interest 117 l-M are passed to the object classifier.
- a comparer 140 compares the features 160 extracted from the images or videos with the appearance models of the objects of interest 117 l-M , producing the score set 142 .
- the score set 142 includes a probability that the features 160 relate to a particular object m of interest.
- the identity of the photographer 132 and capture information 112 is passed to the score adjuster 144 .
- the score adjuster 144 loads a photographer content profile 130 that describes the prior probability of the particular object m is in the image or video and adjusts the score set 142 accordingly to produce an adjusted score set, resulting in a classification 118 that is stored in a database 120 , ready for querying as before.
- the photographer content profile 130 describes the prior probability that the image is an object of interest.
- the photographer content profile 130 associated with the photographer 132 describes the probability that the image represents the object of interest.
- the object classifier 164 can have the objective of determining whether an image is an “indoor” or “outdoor” image.
- different photographers capture different proportions of their images indoor versus outdoor. For example, photographer 0 captures 30% indoor images and 70% outdoor images but photographer 1 captures 30% indoor images and 70% outdoor images. This information is provided by the photographer content profile 130 corresponding to the identify of the photographer 132 .
- knowledge of the prior probability for a particular photographer allows the score adjuster 144 to adequately consider the prior probability and improve the overall performance of the object classifier 164 .
- the object classifier 164 in this example classifies the state of the image as either an indoor or an outdoor image.
- Other object classifiers 164 can be used to determine whether an image is a natural scene or a manmade scene.
- an object classifier 164 can be used to determine if an image is a sunset image or not a sunset image.
- the object classifier 164 can also be used to find specific objects within an image.
- the individual recognition classifier 114 is an example of an object classifier used to recognize individual objects within a scene.
- the object classifier 164 is used to detect the presence of a redeye defect (which may occur 0, 1 or many times in a single image). The object classifier 164 determines whether an image represents the object of interest.
- the photographer content profile 130 can be dependent on the capture information 112 in addition to the identify of the photographer 132 .
- the following table is a photographer content profile 130 and shows the probability that a given image or video captured by a given photographer is “indoor” or “outdoor”, as based on the month of the image capture that is contained in the capture information 112 .
- “indoor” and “outdoor” are objects of interest that are described with an appearance model used by the object classifier 164 to classify an image as an indoor image or an outdoor image.
- the photographer content profile 130 has been shown to include prior probability information for the classification of a single image or video. Additionally, the photographer content profile 130 can contain information related to the probability of a particular image (or video) being outdoors given that a previous image was outdoors for example. This information is stored as a state transition matrix as are commonly used in the study of Markov Chains and hidden Markov Models. The state transition matrices are dependent on the identification of the photographer, because they are related to the photographic behavior of that individual. For example, a state transition matrix for photographer 2 : For photographer 2 Probability image or When video n + 1 is image n is: indoor Outdoor indoor 0.8 0.2 outdoor 0.3 0.7
- the transition matrix is a special case of a prior probability.
- the transition matrix can depend on the capture information 112 , for example the time interval between the image n and the image n+1. For example, if the time interval between the two images is large (e.g. 3 days), then the state transition matrix is: For photographer 2 Probability image or When video n + 1 is image n is: indoor Outdoor indoor 0.6 0.4 outdoor 0.43 0.57
- hidden Markov models can be used along with the state transition matrix to determine the state of the image (in this case, whether the image is indoor or outdoor.)
- the capture information 112 and identify of the photographer 132 are used to select the correct photographer content profile 130 for inputting to the object classifier 164 .
- knowing the identity of the photographer further enables a more accurate image orientation detection.
- a rectangular image is properly viewed in one of four possible viewing orientations.
- Luo, in U.S. Patent Application Publication US 2003/0152289 A1 teaches a method to automatically determination which of these four orientations is correct for an image based on semantic object and scene layout detections and employing an arbitration method to estimate of image orientation.
- the prior probabilities of horizontal, vertical 1, vertical 2, and upside down horizontal are noted to be 70%-14%-14%-2%.
- FIG. 6A shows an example of an image in horizontal orientation.
- FIGS. 6B-6D show image having orientations of vertical 1, vertical 2, and upside down horizontal, respectively.
- Having improved image orientation determination performance improves the performance of subsequent image understanding algorithms that utilize a priori knowledge of orientation.
- Knowledge of other photographer specific capture traits provides additional information that can improve the performance of subsequent image enhancement, image organization, and image management algorithms. For example, knowing if a photographer does not have a steady hand (resulting in excessive camera motion during image capture), could signal a subsequent image enhancement algorithm to more aggressively attempt to correct for image blur. Yet another example is if a particular photographer tends to not keep the camera level with the horizon, knowing the photographer identity could signal an image enhancement algorithm that levels the image.
Landscapes
- Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Oral & Maxillofacial Surgery (AREA)
- General Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
- Television Signal Processing For Recording (AREA)
- Studio Devices (AREA)
Abstract
Description
- Reference is made to commonly assigned, co-pending U.S. patent application Ser. No. 11/116,729 filed Apr. 28, 2005, entitled “Using Time in Recognizing Persons in Images” by Gallagher et al, the disclosure of which is incorporated herein by reference.
- The invention relates to image processing and more particularly relates to the identification or recognition of a person of interest in a collection of digital images or videos.
- With the advent of digital photography, consumers are amassing large collections of digital images and videos. The average number of images captures with digital cameras per photographer is still increasing each year. As a consequence, the organization and retrieval of images and videos is already a problem for the typical consumer. Currently, the length of time spanned by a typical consumer's digital image collection is only a few years. The organization and retrieval problem and its complexity will continue to grow as the length of time spanned by the average digital image and video collection increases and the sources for image content also increases.
- Descriptors of a photographer, such as biometric data, can be detected and/or recorded in an image capture device such as a digital camera.
- U.S. Patent Application Publication US 2003/235411A1 discloses an imaging apparatus wherein a biological recording means records biological information of a photographer and subsequently uses this information to control the imaging apparatus.
- U.S. Pat. No. 6,433,818 describes a camera with built-in apparatus for acquiring measured biometric data of a potential user, comparing it with corresponding biometric signature data entered in the camera, and allowing camera operation only if there is a match. This function provides security and prevents unauthorized access to the data stored on the camera, thereby discouraging theft.
- U.S. Patent Application Publication US 2002/0080256A1 describes a digital camera with built-in apparatus for biometric identification of the photographer and for recording the identification of the photographer of an image. Furthermore, this feature can be used to discourage theft or piracy of the camera. Preferably, an image of the photographer's iris is captured to determine the identity of the photographer.
- One of the most prevalent image content subjects of photography is people. Furthermore, in a given collection of digital images and videos, certain people tend to occur frequently. For example, it is common for most images captured by new parents to contain their new baby. Consumers desire the ability to find all images from their collection containing a specific person.
- It is an object of the present invention to make use of the fact that a particular photographer has captured the images in a collection to find objects of interest in the collection.
- This object is achieved by a method of identifying at least one particular person of interest in a collection of digital images taken over time, comprising:
- providing the collection of digital images with each digital image containing one or more persons;
- storing in a digital database an appearance model having a set of features associated with the particular person of interest and a prior probability of the person of interest appearing in images by a particular photographer; and
- comparing the appearance model with features extracted from the digital images in the collection of digital images and using the prior probability and the appearance model to determine if the particular person of interest is in one or more digital images in the collection of digital images.
- It is an advantage of the present invention that when a known photographer has taken images in a collection, that fact can be used to find images of interest. This is made possible because the likelihood that an object of interest appears in an image is related to the identity of the photographer.
- It is a further advantage of the invention that knowledge of the identity of the particular photographer of an image can be used in determining the orientation of the captured image and whether the image was captured indoors or outdoors.
- The above-mentioned and other features and objects of this invention and the manner of attaining them will become more apparent and the invention itself will be better understood by reference to the following description of an embodiment of the invention taken in conjunction with the accompanying figures wherein:
-
FIG. 1 is a schematic diagram of computer system that can implement the present invention; -
FIG. 2 is a flow chart of an embodiment of the present invention; -
FIGS. 3A and 3B are appearance models of a person of interest having different associated times; -
FIG. 4 is a more detailed flow chart of an embodiment of the present invention; -
FIG. 5 is flow chart of an alternative embodiment of the present invention; and -
FIGS. 6A-6D illustrate four different image orientations. - In the following description, some embodiments of the present invention will be described as software programs and/or algorithmic flow diagrams. Those skilled in the art will readily recognize that the equivalent of such software can also be constructed in hardware. Because image manipulation algorithms and systems are well known, the present description will be directed in particular to algorithms and systems forming part of, or cooperating more directly with, the method in accordance with the present invention. Other aspects of such algorithms and systems, and hardware and/or software for producing and otherwise processing the image signals involved therewith, not specifically shown or described herein can be selected from such systems, algorithms, components, and elements known in the art. Given the description as set forth in the following specification, all software implementation thereof is conventional and within the ordinary skill in such arts.
- The present invention can be implemented in computer hardware and computerized equipment. For example, the method can be performed in a digital camera or other digital image capture devices, a digital printer, on an internet server, on a kiosk, and on a personal computer. Referring to
FIG. 1 , there is illustrated a computer system for implementing the present invention. Although the computer system is shown for the purpose of illustrating a preferred embodiment, the present invention is not limited to the computer system shown, but can be used on any electronic processing system such as found in digital cameras, cell phones, portable computing devices, home computers, kiosks, retail or wholesale photofinishing, or any other system for the processing of digital images. The computer system includes a microprocessor-based unit 20 (also referred to herein as a digital image processor) for receiving and processing software programs and for performing other processing functions. Thedigital image processor 20 processes images fromimage capture devices 10 such as cameras, scanners, or computer image generation software. Thedigital image processor 20 can be used to process digital images to make adjustments for overall brightness, tone scale, image structure, etc. of digital images in a manner such that a pleasing looking image is produced by an image output device. Thedigital image processor 20 interfaces with the general control computer 40 (also a microprocessor based unit) for exchanging data and commands. Thegeneral control computer 40 and thedigital image processor 20 can be two different microprocessors, or the functions of each can be performs by a single physical microprocessor. Thedigital image processor 20 often outputs an image to animage output device 30 for example a printer for displaying the image. Adisplay device 50 is electrically connected to thedigital image processor 20 for displaying user-related information associated with the software, e.g., by means of a graphical user interface. Akeyboard 60 is also connected to the microprocessor basedunit 20 via thegeneral control computer 40 for permitting a user to input information to the software. As an alternative to using thekeyboard 60 for input, a mouse can be used for moving a selector on thedisplay device 50 and for selecting an item on which the selector overlays, as is well known in the art. - A compact disk-read only memory (CD-ROM) which typically includes software programs, is inserted into the
general control computer 40 for providing a means of inputting the software programs and other information to thegeneral control computer 40 and thedigital image processor 20. In addition, a floppy disk can also include a software program, and is inserted into thegeneral control computer 40 for inputting the software program. Still further, thegeneral control computer 40 can be programmed, as is well known in the art, for storing the software program internally. Thegeneral control computer 40 can have a network connection, such as a telephone line or wireless connection, to an external network, such as a local area network or the Internet. - Images can also be displayed on the
display device 50 via a personal computer card (PC Card), such as, as it was formerly known, a PCMCIA card (based on the specifications of the Personal Computer Memory Card International Association) which contains digitized images electronically embodied in the card. - The
image output device 30 provides a final image. Theoutput device 30 can be a printer or other output device that provides a paper or other hard copy final image. Theoutput device 30 can also be an output device that provides the final image as a digital file. Theoutput device 30 can also include combinations of output, such as a printed image and a digital file on a memory unit, such as a CD or DVD. Digital images and other data are stored on anoffline memory device 70 such as an external hard drive, flash media, a drive that writes to CD-ROM or DVD media, or the like. - A digital image includes one or more digital image channels or color components. Each digital image channel is a two-dimensional array of pixels. Each pixel value relates to the amount of light received by the imaging capture device corresponding to the physical region of pixel. For color imaging applications, a digital image will often consist of red, green, and blue digital image channels. Motion imaging applications can be thought of as a sequence of digital images. Those skilled in the art will recognize that the present invention can be applied to, but is not limited to, a digital image channel for any of the herein-mentioned applications. Although a digital image channel is described as a two dimensional array of pixel values arranged by rows and columns, those skilled in the art will recognize that the present invention can be applied to non rectilinear arrays with equal effect. Those skilled in the art will also recognize that for digital image processing steps described hereinbelow as replacing original pixel values with processed pixel values is functionally equivalent to describing the same processing steps as generating a new digital image with the processed pixel values while retaining the original pixel values.
- It should also be noted that the present invention can be implemented in a combination of software and/or hardware and is not limited to devices which are physically connected and/or located within the same physical location. One or more of the devices illustrated in
FIG. 1 can be located remotely and can be connected via a network. One or more of the devices can be connected wirelessly, such as by a radio-frequency link, either directly or via a network. - The present invention can be employed in a variety of user contexts and environments. Exemplary contexts and environments include, without limitation, wholesale digital photofinishing (which involves exemplary process steps or stages such as film in, digital processing, prints out), retail digital photofinishing (film in, digital processing, prints out), home printing (home scanned film or digital images, digital processing, prints out), desktop software (software that applies algorithms to digital prints to make them better—or even just to change them), digital fulfillment (digital images in—from media or over the web, digital processing, with images out—in digital form on media, digital form over the web, or printed on hard-copy prints), kiosks (digital or scanned input, digital processing, digital or hard copy output), mobile devices (e.g., PDA or cell phone that can be used as a processing unit, a display unit, or a unit to give processing instructions), and as a service offered via the World Wide Web.
- In each case, the invention can stand alone or can be a component of a larger system solution. Furthermore, human interfaces, e.g., the scanning or input, the digital processing, the display to a user (if needed), the input of user requests or processing instructions (if needed), the output, can each be on the same or different devices and physical locations, and communication between the devices and locations can be via public or private network connections, or media based communication. Where consistent with the foregoing disclosure of the present invention, the method of the invention can be fully automatic, can have user input (be fully or partially manual), can have user or operator review to accept/reject the result, or can be assisted by metadata (metadata that can be user supplied, supplied by a measuring device (e.g. in a camera), or determined by an algorithm). Moreover, the algorithm(s) can interface with a variety of workflow user interface schemes.
- The invention is inclusive of combinations of the embodiments described herein. References to “a particular embodiment” and the like refer to features that are present in at least one embodiment of the invention. Separate references to “an embodiment” or “particular embodiments” or the like do not necessarily refer to the same embodiment or embodiments; however, such embodiments are not mutually exclusive, unless so indicated or as are readily apparent to one of skill in the art.
- A first embodiment of this method is illustrated in
FIG. 2 . A digital image andvideo collection 104 is analyzed to recognize the identities of the people therein. Recognition of the people in a collection of the digital images orvideos 104 allows for the creation of adatabase 120 relating digital images orvideos 104 and the identities of the people in the digital images orvideos 104. Consequently, thedatabase 120 can be queried 122 to find those images and videos containing the query person of interest. Thedatabase 120 can be in any form. Thedatabase 120 can be distributed across many files, by for example, inserting aperson classification 118 into the file header of the digital image orvideo 104. For example, aquery 122 for images of “Jim” returns query results 124 containing the set of digital images orvideos 104 containing the person of interest “Jim”. Then the query results 124 can be returned. The query results 124 are the set of digital images andvideos 104 containing the query person of interest. The digital images andvideos 104 of thequery result 124 can be displayed on adisplay device 50 as shown inFIG. 1 for review by the user. - In the system, the digital image or
video 104 is passed to thecapture information extractor 108. Thecapture information extractor 108 determines thecapture information 112 related to the capture of the digital image andvideos 104 of the collection, and outputs the resulting information, wherecapture information 112 related to a particular digital image orvideo 104 is associated with that particular digital image orvideo 104. Thecapture information 112 can be the time a particular digital image orvideo 104 was captured, the location the image or video was captured, the focal length, the subject distance, the number of rows and columns of pixels, the magnification, whether the flash was fired, whether the self-timer was activated, etc. Thecapture information 112 for a particular image or video is determined by one of several methods by thecapture information extractor 108. Often times, captureinformation 112 is embedded in the file header of the digital image orvideo 104. For example, the EXIF image format (described at www.exif.org) allows the image or video capture device to store information associated with the image or video in the file header. The “Date\Time” entry is associated with the date and time the image was captured. In some cases, the digital image or video results from scanning film and the image capture time is determined by detection of the date printed into the image (as is often done at capture time) area, usually in the lower left corner of the image. The date a photograph is printed is often printed on the back of the print. Alternatively, some film systems contain a magnetic layer in the film for storing information such as the capture date. Thecapture information extractor 108 uses the most appropriate method for extracting thecapture information 112 of the image. Preferably, the source of the digital images and videos is a digital camera, and thecapture information extractor 108 extracts the capture information from the image file information. - The images and videos of the collection are also analyzed by the
photographer determiner 138 to determine the identity of the particular photographer for each image and video. When the EXIF file format is used, the identity of the photographer can be stored in the “Camera Owner”, “Image Creator”, “Photographer”, or “Copyright” tags for example. The identity of the photographer of an image or video can be entered manually before, during, or after capturing the video. Furthermore, several cameras (e.g. in U.S. Patent Application Publication US 2002/0080256A1) have been described that contain means for extracting biometric information from the photographer, identifying the photographer, and then annotating the image with the identity of the photographer. In any case, thephotographer determiner 138 discovers the identity of thephotographer 132 and passes that information to theindividual recognition classifier 114. In some cases, thephotographer 132 can not be able to be identified by thephotographer determiner 138. In this case, thephotographer 132 is “unknown”. For example, this situation can occur when a person who owns the camera is on vacation and asks a stranger to use her (the vacationing camera owner) camera to capture an image of her in front of a landmark. A camera such as described in U.S. Patent Application US 2002/0080256A1 can only feasibly identify the photographer from a small set of potential camera users (e.g. the primary user is probably the camera owner, and secondary users are friends and family of the camera owner) whose profiles are known by the camera. In this case, an image captured by a stranger using the camera would simply be identified by thephotographer determiner 138 as having an “unknown”photographer 132. - Image capture time can be included in the
capture information 112 and can be a precise minute in time, e.g. Mar. 27, 2004 at 10:17 AM. Or the image capture time can be less precise, e.g. 2004 or March 2004. The image capture time can be in the form of a probability distribution function e.g. Mar. 27, 2004+/−2 days with 95% confidence. The image capture time is input to theindividual recognition classifier 114. - The
person finder 106 identifies one or more persons in the digital video or digital image. Preferably, people are found by face detection. Methods for detecting human faces are well known in the art of digital image processing. For example, a face detection method for finding human faces in images is described in the following article: Jones, M. J.; Viola, P., “Fast Multi-view Face Detection”, IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2003. Furthermore, once a face is detected, the facial features (e.g. eyes, nose, mouth, etc.) can also be localized using well known methods such as described by Yuille et al. in, “Feature Extraction from Faces Using Deformable Templates,” Int. Journal of Comp. Vis., Vol. 8, Iss. 2, 1992, pp. 99-111. The authors describe a method of using energy minimization with template matching for locating the mouth, eye and iris/sclera boundary. Facial features can also be found using active appearance models as described by T. F. Cootes and C. J. Taylor “Constrained active appearance models”, 8th International Conference on Computer Vision,volume 1, pages 748-754. IEEE Computer Society Press, July 2001. - Alternatively, different facial features can also be used. For example, an embodiment can be based upon the facial similarity metric described by M. Turk and A. Pentland in “Eigenfaces for Recognition” Journal of Cognitive Neuroscience. Vol 3, No. 1. 71-86, 1991. Facial descriptors are obtained by projecting the image of a face onto a set of principal component functions that describe the variability of facial appearance. The similarity between any two faces is measured by computing the Euclidean distance of the features obtained by projecting each face onto the same set of functions. In this case, an appearance model can describe the mean and covariance of the projection of multiple images of the same face onto these principal component functions. The similarity metric (for example computed by
comparer 140 ofFIG. 4 ) can calculate the Mahalonobis of a new face to the mean of the model. - The appearance model and similarity measurement could include a combination of several disparate feature types such as Eigenfaces, facial measurements, color/texture information, wavelet features etc.
- A
person finder 106 outputspersonal features 110 that are quantitative descriptions of a person detected by theperson detector 106. Preferably, theperson finder 106 outputs one set ofpersonal features 110 for each detected person. Preferably the personal features are the locations of 82 feature points associated with specific facial features, found using a method similar to the aforementioned active appearance model of Cootes et al. A visual representation of the facial feature points for an image of a face is shown inFIG. 3A andFIG. 3B as an illustration. Alternatively, thepersonal features 110 can additionally be represented with quantifiable descriptors such as eye color, skin color, face shape, etc. Thepersonal features 110 can also be distances between specific feature points or angles formed by lines connecting sets of specific feature points. Color cues are easily extracted from the digital image orvideo 104 once the person and facial features are located by theperson finder 106. Thepersonal features 110 are input to theindividual recognition classifier 114. - A set of appearance models 116 l-116 M of M persons of interest from a stored digital database is input to the
individual recognition classifier 114 for determining the identity of each of the persons detected by theperson finder 106 and represented by the personal features 110. The appearance models 116 l-116 M contain a set of features associated with a particular person of interest. Each appearance model for an individual of interest can include a number of sub-models, for example models describing the appearance of the person of interest at different ages as described in above-cited co-pending application U.S. patent application Ser. No. 11/116,729. Or the sub-models could describe the appearance of the person of interest with different appearances (e.g. one sub-model represents the appearance of the person of interest with facial hair, another sub-model represents the appearance of the person of interest without facial hair). The appearance model will be discussed in greater detail hereinbelow. A person of interest is a person that is important to identify in a collection of digital images andvideos 104. For example, in a household with small children, each child would likely be a person of interest because it would be useful and valuable for a head of the household to identify occurrences of each child in each photo andvideo 104 of the digital media collection. The aforementioned camera owner, as primary photographer, can be a person of interest as can be any secondary photographers (i.e. friends and family of the camera owner). The purpose of theindividual recognition classifier 114 is to use one or more of the appearance models 116 l-116 M, theimage capture information 112, and thepersonal features 110 to determine if a person detected by theperson finder 106 is a particular person of interest. This is accomplished by comparing thepersonal features 110 of the detected person with the features of theappearance model 116 of the particular person of interest. Theindividual recognition classifier 114 outputs theperson classification 118. Typically, the individual recognition classifier outputs a high probability that a person detected by theperson finder 106 is a particular person of interest m when thepersonal features 110 associated with the detected person are similar to the appearance model 116 m associated with the particular person of interest m. Preferably, theperson classification 118 is the likelihood (probability) that the person detected by theperson finder 106 is the person of interest. Theperson classification 118 can be a binary indicator (e.g. true of false) indicating the classifier's decision on whether the person associated with thepersonal features 110 is the person of interest. - By repeating the application of the
individual recognition classifier 114 to each person detected in each digital image orvideo 104 of a collection, all persons detected by theperson finder 106 are classified as to the likelihood that each is the particular person of interest. - The
person classification 118 can then be associated with digital image or video containing the detected person. For example, an image can be tagged or annotated in thedatabase 120 as containing the person of interest “Mary”. Furthermore, theperson classification 118 information can be associated with the personal features 110. Furthermore, the image can also be tagged with the location within the digital image or video of the person detected by theperson finder 106 that theindividual recognition classifier 114 determined was the person of interest. Furthermore, the likelihood or probability of theperson classification 118 can be added to the annotation in thedatabase 120 as well. When thequery 122 is made for digital images orvideos 104 containing a particular person of interest, thedatabase 120 is searched and the appropriate query results 124 are returned to the user on a user interface such as thedisplay device 50 ofFIG. 1 . - The individual recognition classifier uses the
photographer 132 information as follows: When theimage capture information 112 indicates that the camera's self-timer was used to capture a particular image or video, then theindividual recognition classifier 114 outputs theperson classification 118 indicating a high likelihood (probability) that the digital image orvideo 104 contains thephotographer 132. For example, the probability is in the range of 0.75 to 1.0, depending on theparticular photographer 132. This results from the fact that most times, the camera self-timer is used to allow time for thephotographer 132 to activate the camera then run to be included in the scene captured by the camera in the particular digital image orvideo 104. - Likewise, when the
image capture information 112 indicates that the camera's image or video capture was activated via remote control, then theindividual recognition classifier 114 outputs aperson classification 118 indicating a medium likelihood (probability) that the digital image orvideo 104 contains thephotographer 132. This results from the fact that often times, the camera's image or video capture is activated using remote control to allow time thephotographer 132 to be in the scene captured by the camera in the particular digital image orvideo 104. However, remote activation of the camera's capture sequence can also be used to avoid shaking the camera. For this reason, the probability associated with theperson classification 118 indicating that thephotographer 132 is in the digital image orvideo 104 is in the range of 0.5 to 0.9, depending on theparticular photographer 132. -
FIG. 4 shows a more detailed view of theindividual recognition classifier 114. For each set ofpersonal features 110 extracted by the person finder 106 acomparer 140 compares thepersonal features 110 with each of theM appearance models 116 l-M representing M individuals of interest and outputs a score set 142 of M scores indicating the likelihood or probability that the particular person found by the person finder is the mth person of interest. Thecomparer 140 can for example calculate a Euclidean distance between thepersonal features 110 and each of the M appearance models. Or thecomparer 140 can compute a Mahalanobis distance from each of the M appearance models to produce the score set 142. Further, thecomparer 140 can be any classifier, such as but not limited to, Gaussian Maximum Likelihood, a neural network, a support vector machine or the like. Preferably, the mth entry of the score set 142 indicates the probability that the particular face corresponding to the set ofpersonal features 110 is the mth person of interest. - The score set 142 is passed to a
score adjuster 144 for adjusting the score according to information related to thephotographer 132 and thecapture information 112, producing an adjusted score set. Thescore adjuster 144 considers the score set likelihoods or probabilities and refines the scores as follows. The scores in the score set 142 are adjusted according to thecapture information 112. When thecapture information 112 indicates that the self-timer flag was used to capture the digital image orvideo 104, then the score for the mth person of interest is increased if the mth person is thephotographer 132 of the digital image orvideo 104. In this manner, thescore adjuster 144 considers the prior probability (i.e. the prior probability of an event A is the probability that A occurs in the absence of any other data) of thephotographer 132 appearing in the digital image orvideo 104 when the self-timer is user to capture the digital image orvideo 104. Likewise, when thecapture information 112 indicated that the digital image orvideo 104 was captured using a remote signal to activate the camera's capture mechanism, the score for the mth person of interest is increased if the mth person is thephotographer 132 of the digital image orvideo 104. Thescore adjuster 144 outputs aperson classification 118 for the person detected by theperson finder 106. Theperson classification 118 can be the score set 142 after the adjustment by thescore adjuster 144. Or theperson classification 118 can be the identity of the person of interest with the highest corresponding score in the adjusted score set, providing that that score is greater than a threshold T0. - Conceptually, to produce the
person classification 118, theindividual recognition classifier 114 preferably computes: - P(x=m|f) , i.e. the probability that the person detected by the
person finder 106 is person m, given the evidence of the personal features 110 f By Bayes rule, - This quantity can be computed for each of the m persons of interest by:
P(x=m|f)=P(f|x=m)P(x=m)/P(f)
The term P(f|x=m) is the probability of observation of the personal features f when it is known the person is the mth person of interest. This can be found by consideringappearance model 116 for the mth person of interest which describes the variation in appearance for the mth person of interest. The term P(x=m) is the prior probability and is the probability that a person in an image or video is the person of interest m. The prior probability varies as a function of the identity of thephotographer 132. For example, a wife and husband share a camera. Of the people appearing in images where the wife is the photographer, 60% are the baby, 20% are the husband, and 20% are others. When the husband is the photographer, 40% of the people appearing in photos are the baby, 20% are the wife, 5% are the husband, and 35% are others. Therefore, when theindividual recognition classifier 114 knows the identity of thephotographer 132, aphotographer content profile 130 associated with thatphotographer 132 can be used by thescore adjuster 144 to properly adjust the score set 142 in an appropriate fashion. In the preferred embodiment, thephotographer content profile 130 contains the prior probabilities of persons of interest appearing in images and videos for a particular photographer. Preferably, the score set 142 contains the probabilities P(f|x=m) and thescore adjuster 144 adjusts the scores by multiplicatively weighting with the prior probability P(x=m) specified in thephotographer content profile 130. - The
photographer content profiles 130 for a particular photographer are generated by having a user manually verify the identities of people in a set of images or videos captured by the particular photographer. Then the prior probabilities are computed and stored in thephotographer content profile 130. The prior probabilities can consider thecapture information 112. For example, aphotographer 132 has one set of prior probabilities that describe the probability of a person being the mth person of interest when the self-timer function is on as indicated by thecapture information 112, and a second set of prior probabilities with the self-timer function is off. For example, the following table gives the prior probabilities that a given person in an image or video is the mth person of interest for two different photographers and when the self-timer function is off and on.Photographer is person of Photographer is person of interest 0interest 1Mth Person Self-timer Self-timer Self-timer Self-timer of interest on off on off 0 0.3 0.0 0.3 0.2 1 0.3 0.3 0.3 0.0 2 0.2 0.4 0.3 0.6 unknown 0.2 0.3 0.1 0.2 - Therefore, the
photographer 132 of the digital image orvideo 104 is identified along withcapture information 112 related to the digital image orvideo 104. Thephotographer content profile 130 corresponding to thephotographer 132 is used to determine the identities of persons in the digital image orvideo 104. -
FIG. 5 shows an alternative embodiment for using the identify of thephotographer 132 to identify objects within the digital image orvideo 104 or to classify the digital image or video itself. Digital images orvideos 104 from a collection are passed to afeature extractor 152 for extractingfeatures 160 that are passed to anobject classifier 164. - Additionally, the appearance models for objects of
interest 117 l-M are passed to the object classifier. Acomparer 140 compares thefeatures 160 extracted from the images or videos with the appearance models of the objects ofinterest 117 l-M, producing the score set 142. Preferably, the score set 142 includes a probability that thefeatures 160 relate to a particular object m of interest. - The identity of the
photographer 132 and captureinformation 112 is passed to thescore adjuster 144. Thescore adjuster 144 loads aphotographer content profile 130 that describes the prior probability of the particular object m is in the image or video and adjusts the score set 142 accordingly to produce an adjusted score set, resulting in aclassification 118 that is stored in adatabase 120, ready for querying as before. Alternatively, thephotographer content profile 130 describes the prior probability that the image is an object of interest. Thephotographer content profile 130 associated with thephotographer 132 describes the probability that the image represents the object of interest. - For example, the
object classifier 164 can have the objective of determining whether an image is an “indoor” or “outdoor” image. However, different photographers capture different proportions of their images indoor versus outdoor. For example,photographer 0captures 30% indoor images and 70% outdoor images butphotographer 1captures 30% indoor images and 70% outdoor images. This information is provided by thephotographer content profile 130 corresponding to the identify of thephotographer 132. Thus, knowledge of the prior probability for a particular photographer allows thescore adjuster 144 to adequately consider the prior probability and improve the overall performance of theobject classifier 164. - The
object classifier 164 in this example classifies the state of the image as either an indoor or an outdoor image.Other object classifiers 164 can be used to determine whether an image is a natural scene or a manmade scene. Furthermore, anobject classifier 164 can be used to determine if an image is a sunset image or not a sunset image. - The
object classifier 164 can also be used to find specific objects within an image. For example, theindividual recognition classifier 114 is an example of an object classifier used to recognize individual objects within a scene. - In another embodiment, the
object classifier 164 is used to detect the presence of a redeye defect (which may occur 0, 1 or many times in a single image). Theobject classifier 164 determines whether an image represents the object of interest. - Furthermore, the
photographer content profile 130 can be dependent on thecapture information 112 in addition to the identify of thephotographer 132. For example, the following table is aphotographer content profile 130 and shows the probability that a given image or video captured by a given photographer is “indoor” or “outdoor”, as based on the month of the image capture that is contained in thecapture information 112. In this case, “indoor” and “outdoor” are objects of interest that are described with an appearance model used by theobject classifier 164 to classify an image as an indoor image or an outdoor image.Photographer 0Photographer 1indoor Outdoor indoor outdoor Jan-March 0.3 0.7 0.5 0.5 April-June 0.5 0.5 0.3 0.7 July-Sept. 0.6 0.4 0.2 0.8 Oct.-Dec 0.4 0.6 0.8 0.2 - The
photographer content profile 130 has been shown to include prior probability information for the classification of a single image or video. Additionally, thephotographer content profile 130 can contain information related to the probability of a particular image (or video) being outdoors given that a previous image was outdoors for example. This information is stored as a state transition matrix as are commonly used in the study of Markov Chains and hidden Markov Models. The state transition matrices are dependent on the identification of the photographer, because they are related to the photographic behavior of that individual. For example, a state transition matrix for photographer 2:For photographer 2 Probability image or When video n + 1 is image n is: indoor Outdoor indoor 0.8 0.2 outdoor 0.3 0.7 - The transition matrix is a special case of a prior probability. The transition matrix can depend on the
capture information 112, for example the time interval between the image n and theimage n+ 1. For example, if the time interval between the two images is large (e.g. 3 days), then the state transition matrix is:For photographer 2 Probability image or When video n + 1 is image n is: indoor Outdoor indoor 0.6 0.4 outdoor 0.43 0.57 - Once the correction
photographer content profile 130 is known, hidden Markov models can be used along with the state transition matrix to determine the state of the image (in this case, whether the image is indoor or outdoor.) - The
capture information 112 and identify of thephotographer 132 are used to select the correctphotographer content profile 130 for inputting to theobject classifier 164. In a further embodiment, knowing the identity of the photographer further enables a more accurate image orientation detection. In general, a rectangular image is properly viewed in one of four possible viewing orientations. Luo, in U.S. Patent Application Publication US 2003/0152289 A1 teaches a method to automatically determination which of these four orientations is correct for an image based on semantic object and scene layout detections and employing an arbitration method to estimate of image orientation. The prior probabilities of horizontal, vertical 1, vertical 2, and upside down horizontal are noted to be 70%-14%-14%-2%.FIG. 6A shows an example of an image in horizontal orientation.FIGS. 6B-6D show image having orientations of vertical 1, vertical 2, and upside down horizontal, respectively. - Generally, photographers capture landscape scenes in the horizontal orientation (i.e. photographers rarely hold the camera upside down when photographing a scene). However, for vertical or portrait scenes with an aspect ratio having larger height than width, there are two possible camera orientations. Most photographers take “vertical” photographs using only one of these positions. Knowing the photographer's identity establishes that photographer's individual behavior, thus improving its performance in determining image orientation.
- An example orientation probability versus photographer matrix (the
photographer probability profile 130 ofFIG. 5 ) is shown below:Photographer 0Photographer 1Horizontal 0.850 0.650 Upside Down 0.005 0.010 Vertical 10.125 0.030 Vertical 2 0.020 0.310 - Having improved image orientation determination performance improves the performance of subsequent image understanding algorithms that utilize a priori knowledge of orientation. Knowledge of other photographer specific capture traits (types of scenes, camera settings, typical location and time of capture) provides additional information that can improve the performance of subsequent image enhancement, image organization, and image management algorithms. For example, knowing if a photographer does not have a steady hand (resulting in excessive camera motion during image capture), could signal a subsequent image enhancement algorithm to more aggressively attempt to correct for image blur. Yet another example is if a particular photographer tends to not keep the camera level with the horizon, knowing the photographer identity could signal an image enhancement algorithm that levels the image.
- The invention has been described in detail with particular reference to certain preferred embodiments thereof, but it will be understood that variations and modifications can be effected within the spirit and scope of the invention.
-
- 10 image capture devices
- 20 microprocessor-based unit (aka digital image processor)
- 30 image output device
- 40 general control computer
- 50 display device
- 60 keyboard
- 70 offline memory device
- 104 digital images or videos
- 106 person finder
- 108 information extractor
- 110 personal features
- 112 capture information
- 114 individual recognition classifier
- 116 appearance model of a person of interest
- 117 objects of interest
- 118 person classification
- 120 database
- 122 query or queried
- 124 query results
- 130 photographer content profile
- 132 photographer
- 138 photographer determiner
- 140 comparer
- 142 score set
- 144 score adjuster
- 160 features
- 164 object classifier
Claims (12)
Priority Applications (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/143,541 US20060274949A1 (en) | 2005-06-02 | 2005-06-02 | Using photographer identity to classify images |
JP2008514746A JP4908505B2 (en) | 2005-06-02 | 2006-05-30 | Image classification using photographers |
EP06771462A EP1886255B1 (en) | 2005-06-02 | 2006-05-30 | Using photographer identity to classify images |
DE602006007468T DE602006007468D1 (en) | 2005-06-02 | 2006-05-30 | USING THE IDENTITY OF PHOTOGRAPH FOR IMAGE CLASSIFICATION |
CN2006800192912A CN101189621B (en) | 2005-06-02 | 2006-05-30 | Using photographer identity to classify images |
PCT/US2006/020712 WO2006130542A1 (en) | 2005-06-02 | 2006-05-30 | Using photographer identity to classify images |
US12/258,387 US7574054B2 (en) | 2005-06-02 | 2008-10-25 | Using photographer identity to classify images |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/143,541 US20060274949A1 (en) | 2005-06-02 | 2005-06-02 | Using photographer identity to classify images |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/258,387 Division US7574054B2 (en) | 2005-06-02 | 2008-10-25 | Using photographer identity to classify images |
Publications (1)
Publication Number | Publication Date |
---|---|
US20060274949A1 true US20060274949A1 (en) | 2006-12-07 |
Family
ID=36778284
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/143,541 Abandoned US20060274949A1 (en) | 2005-06-02 | 2005-06-02 | Using photographer identity to classify images |
US12/258,387 Expired - Fee Related US7574054B2 (en) | 2005-06-02 | 2008-10-25 | Using photographer identity to classify images |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/258,387 Expired - Fee Related US7574054B2 (en) | 2005-06-02 | 2008-10-25 | Using photographer identity to classify images |
Country Status (6)
Country | Link |
---|---|
US (2) | US20060274949A1 (en) |
EP (1) | EP1886255B1 (en) |
JP (1) | JP4908505B2 (en) |
CN (1) | CN101189621B (en) |
DE (1) | DE602006007468D1 (en) |
WO (1) | WO2006130542A1 (en) |
Cited By (42)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070121625A1 (en) * | 2005-11-30 | 2007-05-31 | Samsung Electronics Co., Ltd. | Method and device for creating/reproducing multimedia file, and information recording medium storing multimedia file |
US20090046954A1 (en) * | 2007-08-14 | 2009-02-19 | Kensuke Ishii | Image sharing system and method |
US20090141993A1 (en) * | 2007-12-03 | 2009-06-04 | Honeywell International Inc. | System for finding archived objects in video data |
US20100002940A1 (en) * | 2008-07-03 | 2010-01-07 | Sony Corporation | Image data processing apparatus and image data processing method |
US20100054601A1 (en) * | 2008-08-28 | 2010-03-04 | Microsoft Corporation | Image Tagging User Interface |
US20100054600A1 (en) * | 2008-08-28 | 2010-03-04 | Microsoft Corporation | Tagging Images With Labels |
US20110026840A1 (en) * | 2009-07-28 | 2011-02-03 | Samsung Electronics Co., Ltd. | System and method for indoor-outdoor scene classification |
US20110033115A1 (en) * | 2005-12-05 | 2011-02-10 | Masao Shiraishi | Method of detecting feature images |
US20110064396A1 (en) * | 2009-09-14 | 2011-03-17 | Canon Kabushiki Kaisha | Image capturing apparatus and method for controlling the same |
US20120041919A1 (en) * | 2010-08-10 | 2012-02-16 | Fujitsu Limited | Information processing apparatus and information processing method |
US8311337B2 (en) | 2010-06-15 | 2012-11-13 | Cyberlink Corp. | Systems and methods for organizing and accessing feature vectors in digital images |
US20130170541A1 (en) * | 2004-07-30 | 2013-07-04 | Euclid Discoveries, Llc | Video Compression Repository and Model Reuse |
US20140140583A1 (en) * | 2012-08-22 | 2014-05-22 | Canon Kabushiki Kaisha | Image recognition apparatus and image recognition method for identifying object |
US8842154B2 (en) | 2007-01-23 | 2014-09-23 | Euclid Discoveries, Llc | Systems and methods for providing personal video services |
US8908766B2 (en) | 2005-03-31 | 2014-12-09 | Euclid Discoveries, Llc | Computer method and apparatus for processing image data |
US8929669B2 (en) | 2010-06-23 | 2015-01-06 | Panasonic Intellectual Property Corporation Of America | Image evaluation apparatus that calculates an importance degree of each of a plurality of images |
US8942283B2 (en) | 2005-03-31 | 2015-01-27 | Euclid Discoveries, Llc | Feature-based hybrid video codec comparing compression efficiency of encodings |
US9106977B2 (en) | 2006-06-08 | 2015-08-11 | Euclid Discoveries, Llc | Object archival systems and methods |
US20160371305A1 (en) * | 2014-08-01 | 2016-12-22 | Baidu Online Network Technology (Beijing) Co., Ltd. | Method, device and apparatus for generating picture search library, and picture search method, device and apparatus |
US9532069B2 (en) | 2004-07-30 | 2016-12-27 | Euclid Discoveries, Llc | Video compression repository and model reuse |
US9578345B2 (en) | 2005-03-31 | 2017-02-21 | Euclid Discoveries, Llc | Model-based video encoding and decoding |
US9621917B2 (en) | 2014-03-10 | 2017-04-11 | Euclid Discoveries, Llc | Continuous block tracking for temporal prediction in video encoding |
US20170124400A1 (en) * | 2015-10-28 | 2017-05-04 | Raanan Y. Yehezkel Rohekar | Automatic video summarization |
US20170161591A1 (en) * | 2015-12-04 | 2017-06-08 | Pilot Ai Labs, Inc. | System and method for deep-learning based object tracking |
US9743078B2 (en) | 2004-07-30 | 2017-08-22 | Euclid Discoveries, Llc | Standards-compliant model-based video encoding and decoding |
US20180018536A1 (en) * | 2016-07-18 | 2018-01-18 | Beijing Xiaomi Mobile Software Co., Ltd. | Method, Device and Computer-Readable Medium for Enhancing Readability |
US10091507B2 (en) | 2014-03-10 | 2018-10-02 | Euclid Discoveries, Llc | Perceptual optimization for model-based video encoding |
US10097851B2 (en) | 2014-03-10 | 2018-10-09 | Euclid Discoveries, Llc | Perceptual optimization for model-based video encoding |
US10410086B2 (en) * | 2017-05-30 | 2019-09-10 | Google Llc | Systems and methods of person recognition in video streams |
US10536816B1 (en) | 2018-08-17 | 2020-01-14 | International Business Machines Corporation | Mobile device that creates a communication group based on the mobile device identifying people currently located at a particular location |
WO2020026112A1 (en) * | 2018-07-31 | 2020-02-06 | Marvell World Trade Ltd. | Metadata generation for multiple object types |
US10664688B2 (en) | 2017-09-20 | 2020-05-26 | Google Llc | Systems and methods of detecting and responding to a visitor to a smart home environment |
CN111326183A (en) * | 2014-02-07 | 2020-06-23 | 高通科技公司 | System and method for processing a temporal image sequence |
US10957171B2 (en) | 2016-07-11 | 2021-03-23 | Google Llc | Methods and systems for providing event alerts |
US11256951B2 (en) | 2017-05-30 | 2022-02-22 | Google Llc | Systems and methods of person recognition in video streams |
US11315400B1 (en) * | 2017-09-19 | 2022-04-26 | Alarm.Com Incorporated | Appearance based access verification |
US11356643B2 (en) | 2017-09-20 | 2022-06-07 | Google Llc | Systems and methods of presenting appropriate actions for responding to a visitor to a smart home environment |
US20220215202A1 (en) * | 2021-01-05 | 2022-07-07 | Applied Research Associates, Inc. | System and method for determining the geographic location in an image |
US20220415035A1 (en) * | 2021-06-23 | 2022-12-29 | Celebr8 Life, Inc. | Machine learning model and neural network to predict data anomalies and content enrichment of digital images for use in video generation |
US11587320B2 (en) | 2016-07-11 | 2023-02-21 | Google Llc | Methods and systems for person detection in a video feed |
US11783010B2 (en) | 2017-05-30 | 2023-10-10 | Google Llc | Systems and methods of person recognition in video streams |
US11893795B2 (en) | 2019-12-09 | 2024-02-06 | Google Llc | Interacting with visitors of a connected home environment |
Families Citing this family (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8326775B2 (en) * | 2005-10-26 | 2012-12-04 | Cortica Ltd. | Signature generation for multimedia deep-content-classification by a large-scale matching system and method thereof |
GB2449412B (en) * | 2007-03-29 | 2012-04-25 | Hewlett Packard Development Co | Integrating object detectors |
US9721148B2 (en) | 2007-12-31 | 2017-08-01 | Applied Recognition Inc. | Face detection and recognition |
CN104866553A (en) | 2007-12-31 | 2015-08-26 | 应用识别公司 | Method, system, and computer program for identification and sharing of digital images with face signatures |
US9639740B2 (en) | 2007-12-31 | 2017-05-02 | Applied Recognition Inc. | Face detection and recognition |
FR2935498B1 (en) * | 2008-08-27 | 2010-10-15 | Eads Europ Aeronautic Defence | METHOD FOR IDENTIFYING AN OBJECT IN A VIDEO ARCHIVE |
US20100106573A1 (en) * | 2008-10-25 | 2010-04-29 | Gallagher Andrew C | Action suggestions based on inferred social relationships |
NO331287B1 (en) * | 2008-12-15 | 2011-11-14 | Cisco Systems Int Sarl | Method and apparatus for recognizing faces in a video stream |
TWI410265B (en) * | 2009-04-03 | 2013-10-01 | Academia Sinica | Multimedia game automatic recording and configuration of the image method |
KR101598632B1 (en) | 2009-10-01 | 2016-02-29 | 마이크로소프트 테크놀로지 라이센싱, 엘엘씨 | Mobile terminal and method for editing tag thereof |
US8416997B2 (en) * | 2010-01-27 | 2013-04-09 | Apple Inc. | Method of person identification using social connections |
US8180209B2 (en) * | 2010-05-19 | 2012-05-15 | Eastman Kodak Company | Determining camera activity from a steadiness signal |
US8200076B2 (en) * | 2010-05-19 | 2012-06-12 | Eastman Kodak Company | Estimating gender or age of a photographer |
US8180208B2 (en) * | 2010-05-19 | 2012-05-15 | Eastman Kodak Company | Identifying a photographer |
US20120013640A1 (en) * | 2010-07-15 | 2012-01-19 | Academia Sinica | Graphical representation of events |
CN102231156A (en) * | 2011-06-07 | 2011-11-02 | 盛乐信息技术(上海)有限公司 | Method and system for searching user according to shape characteristic |
US8548207B2 (en) | 2011-08-15 | 2013-10-01 | Daon Holdings Limited | Method of host-directed illumination and system for conducting host-directed illumination |
US9202105B1 (en) | 2012-01-13 | 2015-12-01 | Amazon Technologies, Inc. | Image analysis for user authentication |
JP5885569B2 (en) * | 2012-04-12 | 2016-03-15 | 京セラドキュメントソリューションズ株式会社 | Image processing apparatus and image processing program |
US9058806B2 (en) | 2012-09-10 | 2015-06-16 | Cisco Technology, Inc. | Speaker segmentation and recognition based on list of speakers |
WO2014042646A1 (en) * | 2012-09-14 | 2014-03-20 | Empire Technology Development Llc | Associating an identity to a creator of a set of visual files |
US8886011B2 (en) | 2012-12-07 | 2014-11-11 | Cisco Technology, Inc. | System and method for question detection based video segmentation, search and collaboration in a video processing environment |
US9524282B2 (en) * | 2013-02-07 | 2016-12-20 | Cherif Algreatly | Data augmentation with real-time annotations |
US10038740B2 (en) | 2013-08-19 | 2018-07-31 | Nant Holdings Ip, Llc | Camera-to-camera interactions, systems and methods |
US10043112B2 (en) * | 2014-03-07 | 2018-08-07 | Qualcomm Incorporated | Photo management |
US10698995B2 (en) | 2014-08-28 | 2020-06-30 | Facetec, Inc. | Method to verify identity using a previously collected biometric image/data |
US10803160B2 (en) | 2014-08-28 | 2020-10-13 | Facetec, Inc. | Method to verify and identify blockchain with user question data |
US10614204B2 (en) | 2014-08-28 | 2020-04-07 | Facetec, Inc. | Facial recognition authentication system including path parameters |
US10915618B2 (en) | 2014-08-28 | 2021-02-09 | Facetec, Inc. | Method to add remotely collected biometric images / templates to a database record of personal information |
CA2902093C (en) | 2014-08-28 | 2023-03-07 | Kevin Alan Tussy | Facial recognition authentication system including path parameters |
US12130900B2 (en) | 2014-08-28 | 2024-10-29 | Facetec, Inc. | Method and apparatus to dynamically control facial illumination |
US11256792B2 (en) | 2014-08-28 | 2022-02-22 | Facetec, Inc. | Method and apparatus for creation and use of digital identification |
USD987653S1 (en) | 2016-04-26 | 2023-05-30 | Facetec, Inc. | Display screen or portion thereof with graphical user interface |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020080256A1 (en) * | 2000-12-22 | 2002-06-27 | International Business Machines Corporation | Digital camera apparatus with biometric capability |
US6433818B1 (en) * | 1998-11-06 | 2002-08-13 | Fotonation, Inc. | Digital camera with biometric security |
US6591005B1 (en) * | 2000-03-27 | 2003-07-08 | Eastman Kodak Company | Method of estimating image format and orientation based upon vanishing point location |
US20030152289A1 (en) * | 2002-02-13 | 2003-08-14 | Eastman Kodak Company | Method and system for determining image orientation |
US20030235411A1 (en) * | 2002-06-20 | 2003-12-25 | Canon Kabushiki Kaisha | Imaging apparatus and method of controlling same |
US20040174434A1 (en) * | 2002-12-18 | 2004-09-09 | Walker Jay S. | Systems and methods for suggesting meta-information to a camera user |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6977679B2 (en) | 2001-04-03 | 2005-12-20 | Hewlett-Packard Development Company, L.P. | Camera meta-data for content categorization |
GB0112773D0 (en) * | 2001-05-25 | 2001-07-18 | Univ Manchester | Object identification |
US7313268B2 (en) | 2002-10-31 | 2007-12-25 | Eastman Kodak Company | Method for using effective spatio-temporal image recomposition to improve scene classification |
CN1306456C (en) * | 2002-12-13 | 2007-03-21 | 佳能株式会社 | Image processing method and apparatus |
JP2004304585A (en) * | 2003-03-31 | 2004-10-28 | Ntt Docomo Inc | Device, method, and program for image management |
JP2004318603A (en) * | 2003-04-17 | 2004-11-11 | Nippon Telegr & Teleph Corp <Ntt> | Image retrieval method, self photographing estimation device and method, and self photographing estimation program |
JP2005072949A (en) * | 2003-08-25 | 2005-03-17 | Casio Comput Co Ltd | Image photographing device and program |
-
2005
- 2005-06-02 US US11/143,541 patent/US20060274949A1/en not_active Abandoned
-
2006
- 2006-05-30 WO PCT/US2006/020712 patent/WO2006130542A1/en active Application Filing
- 2006-05-30 JP JP2008514746A patent/JP4908505B2/en not_active Expired - Fee Related
- 2006-05-30 CN CN2006800192912A patent/CN101189621B/en not_active Expired - Fee Related
- 2006-05-30 DE DE602006007468T patent/DE602006007468D1/en active Active
- 2006-05-30 EP EP06771462A patent/EP1886255B1/en not_active Not-in-force
-
2008
- 2008-10-25 US US12/258,387 patent/US7574054B2/en not_active Expired - Fee Related
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6433818B1 (en) * | 1998-11-06 | 2002-08-13 | Fotonation, Inc. | Digital camera with biometric security |
US6591005B1 (en) * | 2000-03-27 | 2003-07-08 | Eastman Kodak Company | Method of estimating image format and orientation based upon vanishing point location |
US20020080256A1 (en) * | 2000-12-22 | 2002-06-27 | International Business Machines Corporation | Digital camera apparatus with biometric capability |
US20030152289A1 (en) * | 2002-02-13 | 2003-08-14 | Eastman Kodak Company | Method and system for determining image orientation |
US20030235411A1 (en) * | 2002-06-20 | 2003-12-25 | Canon Kabushiki Kaisha | Imaging apparatus and method of controlling same |
US20040174434A1 (en) * | 2002-12-18 | 2004-09-09 | Walker Jay S. | Systems and methods for suggesting meta-information to a camera user |
Cited By (72)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130170541A1 (en) * | 2004-07-30 | 2013-07-04 | Euclid Discoveries, Llc | Video Compression Repository and Model Reuse |
US9532069B2 (en) | 2004-07-30 | 2016-12-27 | Euclid Discoveries, Llc | Video compression repository and model reuse |
US9743078B2 (en) | 2004-07-30 | 2017-08-22 | Euclid Discoveries, Llc | Standards-compliant model-based video encoding and decoding |
US8902971B2 (en) * | 2004-07-30 | 2014-12-02 | Euclid Discoveries, Llc | Video compression repository and model reuse |
US8964835B2 (en) | 2005-03-31 | 2015-02-24 | Euclid Discoveries, Llc | Feature-based video compression |
US8942283B2 (en) | 2005-03-31 | 2015-01-27 | Euclid Discoveries, Llc | Feature-based hybrid video codec comparing compression efficiency of encodings |
US8908766B2 (en) | 2005-03-31 | 2014-12-09 | Euclid Discoveries, Llc | Computer method and apparatus for processing image data |
US9578345B2 (en) | 2005-03-31 | 2017-02-21 | Euclid Discoveries, Llc | Model-based video encoding and decoding |
US20070121625A1 (en) * | 2005-11-30 | 2007-05-31 | Samsung Electronics Co., Ltd. | Method and device for creating/reproducing multimedia file, and information recording medium storing multimedia file |
US20110033115A1 (en) * | 2005-12-05 | 2011-02-10 | Masao Shiraishi | Method of detecting feature images |
US9106977B2 (en) | 2006-06-08 | 2015-08-11 | Euclid Discoveries, Llc | Object archival systems and methods |
US8842154B2 (en) | 2007-01-23 | 2014-09-23 | Euclid Discoveries, Llc | Systems and methods for providing personal video services |
US8144944B2 (en) | 2007-08-14 | 2012-03-27 | Olympus Corporation | Image sharing system and method |
US20090046954A1 (en) * | 2007-08-14 | 2009-02-19 | Kensuke Ishii | Image sharing system and method |
US20090141993A1 (en) * | 2007-12-03 | 2009-06-04 | Honeywell International Inc. | System for finding archived objects in video data |
US8160371B2 (en) * | 2007-12-03 | 2012-04-17 | Honeywell International Inc. | System for finding archived objects in video data |
US8331691B2 (en) * | 2008-07-03 | 2012-12-11 | Sony Corporation | Image data processing apparatus and image data processing method |
US20100002940A1 (en) * | 2008-07-03 | 2010-01-07 | Sony Corporation | Image data processing apparatus and image data processing method |
US20150016691A1 (en) * | 2008-08-28 | 2015-01-15 | Microsoft Corporation | Image Tagging User Interface |
US8396246B2 (en) | 2008-08-28 | 2013-03-12 | Microsoft Corporation | Tagging images with labels |
US9020183B2 (en) | 2008-08-28 | 2015-04-28 | Microsoft Technology Licensing, Llc | Tagging images with labels |
US8867779B2 (en) * | 2008-08-28 | 2014-10-21 | Microsoft Corporation | Image tagging user interface |
US20100054601A1 (en) * | 2008-08-28 | 2010-03-04 | Microsoft Corporation | Image Tagging User Interface |
US20100054600A1 (en) * | 2008-08-28 | 2010-03-04 | Microsoft Corporation | Tagging Images With Labels |
US20110026840A1 (en) * | 2009-07-28 | 2011-02-03 | Samsung Electronics Co., Ltd. | System and method for indoor-outdoor scene classification |
US8644624B2 (en) * | 2009-07-28 | 2014-02-04 | Samsung Electronics Co., Ltd. | System and method for indoor-outdoor scene classification |
US20110064396A1 (en) * | 2009-09-14 | 2011-03-17 | Canon Kabushiki Kaisha | Image capturing apparatus and method for controlling the same |
US8244119B2 (en) * | 2009-09-14 | 2012-08-14 | Canon Kabushiki Kaisha | Image capturing apparatus and method for controlling the same |
US8311337B2 (en) | 2010-06-15 | 2012-11-13 | Cyberlink Corp. | Systems and methods for organizing and accessing feature vectors in digital images |
US8929669B2 (en) | 2010-06-23 | 2015-01-06 | Panasonic Intellectual Property Corporation Of America | Image evaluation apparatus that calculates an importance degree of each of a plurality of images |
US20120041919A1 (en) * | 2010-08-10 | 2012-02-16 | Fujitsu Limited | Information processing apparatus and information processing method |
US8682834B2 (en) * | 2010-08-10 | 2014-03-25 | Fujitsu Limited | Information processing apparatus and information processing method |
US9489566B2 (en) * | 2012-08-22 | 2016-11-08 | Canon Kabushiki Kaisha | Image recognition apparatus and image recognition method for identifying object |
US20140140583A1 (en) * | 2012-08-22 | 2014-05-22 | Canon Kabushiki Kaisha | Image recognition apparatus and image recognition method for identifying object |
CN111326183A (en) * | 2014-02-07 | 2020-06-23 | 高通科技公司 | System and method for processing a temporal image sequence |
US9621917B2 (en) | 2014-03-10 | 2017-04-11 | Euclid Discoveries, Llc | Continuous block tracking for temporal prediction in video encoding |
US10091507B2 (en) | 2014-03-10 | 2018-10-02 | Euclid Discoveries, Llc | Perceptual optimization for model-based video encoding |
US10097851B2 (en) | 2014-03-10 | 2018-10-09 | Euclid Discoveries, Llc | Perceptual optimization for model-based video encoding |
US20160371305A1 (en) * | 2014-08-01 | 2016-12-22 | Baidu Online Network Technology (Beijing) Co., Ltd. | Method, device and apparatus for generating picture search library, and picture search method, device and apparatus |
US20170124400A1 (en) * | 2015-10-28 | 2017-05-04 | Raanan Y. Yehezkel Rohekar | Automatic video summarization |
US9818032B2 (en) * | 2015-10-28 | 2017-11-14 | Intel Corporation | Automatic video summarization |
US20170161591A1 (en) * | 2015-12-04 | 2017-06-08 | Pilot Ai Labs, Inc. | System and method for deep-learning based object tracking |
US11587320B2 (en) | 2016-07-11 | 2023-02-21 | Google Llc | Methods and systems for person detection in a video feed |
US10957171B2 (en) | 2016-07-11 | 2021-03-23 | Google Llc | Methods and systems for providing event alerts |
US20180018536A1 (en) * | 2016-07-18 | 2018-01-18 | Beijing Xiaomi Mobile Software Co., Ltd. | Method, Device and Computer-Readable Medium for Enhancing Readability |
US10068151B2 (en) * | 2016-07-18 | 2018-09-04 | Beijing Xiaomi Mobile Software Co., Ltd. | Method, device and computer-readable medium for enhancing readability |
US11256951B2 (en) | 2017-05-30 | 2022-02-22 | Google Llc | Systems and methods of person recognition in video streams |
US10410086B2 (en) * | 2017-05-30 | 2019-09-10 | Google Llc | Systems and methods of person recognition in video streams |
US11783010B2 (en) | 2017-05-30 | 2023-10-10 | Google Llc | Systems and methods of person recognition in video streams |
US10599950B2 (en) | 2017-05-30 | 2020-03-24 | Google Llc | Systems and methods for person recognition data management |
US11386285B2 (en) * | 2017-05-30 | 2022-07-12 | Google Llc | Systems and methods of person recognition in video streams |
US10685257B2 (en) * | 2017-05-30 | 2020-06-16 | Google Llc | Systems and methods of person recognition in video streams |
US11315400B1 (en) * | 2017-09-19 | 2022-04-26 | Alarm.Com Incorporated | Appearance based access verification |
US10664688B2 (en) | 2017-09-20 | 2020-05-26 | Google Llc | Systems and methods of detecting and responding to a visitor to a smart home environment |
US11256908B2 (en) | 2017-09-20 | 2022-02-22 | Google Llc | Systems and methods of detecting and responding to a visitor to a smart home environment |
US12125369B2 (en) | 2017-09-20 | 2024-10-22 | Google Llc | Systems and methods of detecting and responding to a visitor to a smart home environment |
US11710387B2 (en) | 2017-09-20 | 2023-07-25 | Google Llc | Systems and methods of detecting and responding to a visitor to a smart home environment |
US11356643B2 (en) | 2017-09-20 | 2022-06-07 | Google Llc | Systems and methods of presenting appropriate actions for responding to a visitor to a smart home environment |
US11734363B2 (en) | 2018-07-31 | 2023-08-22 | Marvell Asia Pte, Ltd. | Storage edge controller with a metadata computational engine |
US11036807B2 (en) | 2018-07-31 | 2021-06-15 | Marvell Asia Pte Ltd | Metadata generation at the storage edge |
US11294965B2 (en) | 2018-07-31 | 2022-04-05 | Marvell Asia Pte Ltd | Metadata generation for multiple object types |
EP4266194A1 (en) * | 2018-07-31 | 2023-10-25 | Marvell Asia Pte, Ltd. | Metadata generation for multiple object types |
WO2020026112A1 (en) * | 2018-07-31 | 2020-02-06 | Marvell World Trade Ltd. | Metadata generation for multiple object types |
CN112513834A (en) * | 2018-07-31 | 2021-03-16 | 马维尔亚洲私人有限公司 | Metadata generation for multiple object types |
EP4220437A1 (en) * | 2018-07-31 | 2023-08-02 | Marvell World Trade Ltd. | Systems and methods for generating metadata describing unstructured data objects at the storage edge |
US11748418B2 (en) | 2018-07-31 | 2023-09-05 | Marvell Asia Pte, Ltd. | Storage aggregator controller with metadata computation control |
US10536816B1 (en) | 2018-08-17 | 2020-01-14 | International Business Machines Corporation | Mobile device that creates a communication group based on the mobile device identifying people currently located at a particular location |
US11893795B2 (en) | 2019-12-09 | 2024-02-06 | Google Llc | Interacting with visitors of a connected home environment |
US20220215202A1 (en) * | 2021-01-05 | 2022-07-07 | Applied Research Associates, Inc. | System and method for determining the geographic location in an image |
US11461993B2 (en) * | 2021-01-05 | 2022-10-04 | Applied Research Associates, Inc. | System and method for determining the geographic location in an image |
US20220415035A1 (en) * | 2021-06-23 | 2022-12-29 | Celebr8 Life, Inc. | Machine learning model and neural network to predict data anomalies and content enrichment of digital images for use in video generation |
US12087047B2 (en) * | 2021-06-23 | 2024-09-10 | Celebr8 Life | Machine learning model and neural network to predict data anomalies and content enrichment of digital images for use in video generation |
Also Published As
Publication number | Publication date |
---|---|
CN101189621A (en) | 2008-05-28 |
EP1886255B1 (en) | 2009-06-24 |
WO2006130542A1 (en) | 2006-12-07 |
JP4908505B2 (en) | 2012-04-04 |
CN101189621B (en) | 2010-11-24 |
US20090046933A1 (en) | 2009-02-19 |
US7574054B2 (en) | 2009-08-11 |
DE602006007468D1 (en) | 2009-08-06 |
JP2008543224A (en) | 2008-11-27 |
EP1886255A1 (en) | 2008-02-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7574054B2 (en) | Using photographer identity to classify images | |
US10346677B2 (en) | Classification and organization of consumer digital images using workflow, and face detection and recognition | |
US7522773B2 (en) | Using time in recognizing persons in images | |
US8199979B2 (en) | Classification system for consumer digital images using automatic workflow and face detection and recognition | |
US7551755B1 (en) | Classification and organization of consumer digital images using workflow, and face detection and recognition | |
US7555148B1 (en) | Classification system for consumer digital images using workflow, face detection, normalization, and face recognition | |
US7558408B1 (en) | Classification system for consumer digital images using workflow and user interface modules, and face detection and recognition | |
US7587068B1 (en) | Classification database for consumer digital images | |
US20070098303A1 (en) | Determining a particular person from a collection | |
US10528795B2 (en) | Imaging workflow using facial and non-facial features | |
US7869658B2 (en) | Representative image selection based on hierarchical clustering | |
JP5517952B2 (en) | Permanent recognition of individuals in images | |
US20120148118A1 (en) | Method for classifying images and apparatus for the same | |
US8189916B2 (en) | Image processing method, system, and computer readable medium | |
Manyam et al. | Two faces are better than one: Face recognition in group photographs | |
JP3962517B2 (en) | Face detection method and apparatus, and computer-readable medium | |
JP4859057B2 (en) | Image processing apparatus, image processing method, program, and recording medium | |
WO2009096524A1 (en) | Method, device, and program for judging image orientation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: EASTMAN KODAK COMPANY, NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GALLAGHER, ANDREW C.;WELDY, JOHN A.;REEL/FRAME:016656/0895 Effective date: 20050525 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |
|
AS | Assignment |
Owner name: EASTMAN KODAK COMPANY, NEW YORK Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: KODAK AVIATION LEASING LLC, NEW YORK Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: FAR EAST DEVELOPMENT LTD., NEW YORK Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: KODAK PHILIPPINES, LTD., NEW YORK Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: EASTMAN KODAK INTERNATIONAL CAPITAL COMPANY, INC., Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: KODAK PORTUGUESA LIMITED, NEW YORK Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: QUALEX INC., NORTH CAROLINA Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: FPC INC., CALIFORNIA Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: KODAK (NEAR EAST), INC., NEW YORK Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: KODAK AMERICAS, LTD., NEW YORK Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: CREO MANUFACTURING AMERICA LLC, WYOMING Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: KODAK IMAGING NETWORK, INC., CALIFORNIA Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: LASER-PACIFIC MEDIA CORPORATION, NEW YORK Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: NPEC INC., NEW YORK Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: PAKON, INC., INDIANA Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 Owner name: KODAK REALTY, INC., NEW YORK Free format text: PATENT RELEASE;ASSIGNORS:CITICORP NORTH AMERICA, INC.;WILMINGTON TRUST, NATIONAL ASSOCIATION;REEL/FRAME:029913/0001 Effective date: 20130201 |
|
AS | Assignment |
Owner name: MONUMENT PEAK VENTURES, LLC, TEXAS Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:INTELLECTUAL VENTURES FUND 83 LLC;REEL/FRAME:064599/0304 Effective date: 20230728 |