JP-7857337-B2 - Information processing device, information processing method, and program
Inventors
- 奥野 泰弘
- 児玉 壮飛
Assignees
- キヤノン株式会社
Dates
- Publication Date
- 20260512
- Application Date
- 20240423
Claims (14)
- A first detection means for detecting a subject region corresponding to the tracking target from a first image and a second image that chronologically follows the first image, A second detection means for detecting a partial region indicating a specific part from the first image and the second image, An acquisition means for acquiring the number of first subregions detected in the first image and the number of second subregions detected in the second image, The system includes a means for associating the aforementioned partial region with the aforementioned subject region, The association means is an information processing device that, when the number of second subregions is less than the number of first subregions, does not associate a specific part shown in the subregion detected in the second image with the subject region .
- A calculation means for calculating an evaluation value for determining association from the second sub- region, The system further includes setting means for setting a threshold value for the evaluation value in accordance with the change in the number of the second subregion and the number of the first subregion, The information processing apparatus according to claim 1, characterized in that it controls whether or not to associate a specific part of the second subregion with the subject region based on the evaluation value and the threshold value.
- The information processing apparatus according to claim 2, characterized in that, when the number of the second subregions is less than the number of the first subregions and the evaluation value is less than the threshold, control is performed so as not to associate a specific part of the second subregion with the subject region .
- The information processing apparatus according to claim 2, characterized in that the evaluation value is a score that evaluates the tracking accuracy of the subject area in the second image.
- The information processing apparatus according to claim 2, characterized in that the threshold is set based on the evaluation value in the partial region with the highest evaluation value, calculated by the calculation means in the first image.
- The information processing apparatus according to claim 5, characterized in that the threshold is set based on the evaluation value in the sub-region with the second highest evaluation value, calculated by the calculation means in the first image.
- The information processing apparatus according to claim 6, characterized in that the threshold is set based on the number of elapsed frames in the second image relative to the first image.
- The information processing apparatus according to claim 2 , characterized in that the evaluation value is the similarity between the feature quantity of the first subregion associated with the subject region in the first image and the feature quantity of the second subregion .
- The information processing apparatus according to claim 1, characterized in that, when the number of the second subregions is less than the number of the first subregions , it suppresses associating a specific part of the second subregion with the subject region .
- The information processing apparatus according to claim 9, characterized in that, when the number of the second subregions is less than the number of the first subregions , additional conditions are set to associate a specific part of the second subregion with the subject region .
- The information processing apparatus according to claim 1, characterized in that the display control indicating the second sub- region is performed according to the content of the association control.
- The information processing apparatus according to claim 11, characterized in that, in the second image, if the second partial region is associated with the subject region , a frame indicating the second partial region is displayed, and if the second partial region is not associated with the subject region, a frame indicating the second partial region is not displayed.
- A step of detecting a subject region corresponding to the tracking target from a first image and a second image that chronologically follows the first image, A step of detecting a partial region indicating a specific part from the first image and the second image, A step of obtaining the number of first subregions detected in the first image and the second subregions detected in the second image, The process includes relating the aforementioned partial region to the aforementioned subject region, An information processing method wherein, if the number of the second subregions is less than the number of the first subregions, a specific part shown in the subregion detected in the second image is not associated with the subject region .
- A program for causing a computer to function as one of the means of an information processing device described in any one of claims 1 to 12.
Description
This invention relates to an information processing device, an information processing method, and a program. The system detects and tracks a specific subject within a series of images. Tracking involves detecting the desired subject area within an image and continuously tracking that same subject area across consecutive images. Based on the tracking results, the camera's autofocus process is performed. Patent Document 1 discloses a method for tracking a subject while associating its whole and its parts. The whole and parts of a subject refer to, for example, in the case of a person, the entire human body as the whole and the face as the part. In Patent Document 1, the association between the whole and the parts is based on the positional relationship (e.g., proximity) between the body and the parts of the subject. Japanese Patent Publication No. 2021-152578 A block diagram showing an example of the configuration of the information processing device according to Embodiment 1.A flowchart illustrating an example of the overall processing performed by an information processing device.A flowchart showing an example of processing by the state acquisition unit.A flowchart illustrating an example of processing by the threshold setting unit.A flowchart illustrating an example of processing performed by the association unit.A flowchart illustrating an example of processing by the display control unit.A diagram illustrating the state in which the number of local detections changes. The embodiments will be described in detail below with reference to the attached drawings. Note that the following embodiments do not limit the invention as defined in the claims. While multiple features are described in the embodiments, not all of these features are essential to the invention, and the features may be combined in any way. Furthermore, in the attached drawings, identical or similar configurations are given the same reference numerals, and redundant descriptions are omitted. [Embodiment 1] The following describes the information processing device according to Embodiment 1. The information processing device according to this embodiment takes a series of images as input, tracks subjects detected in each image, tracks the entire subject to be tracked, and can also detect specific parts of the subject and associate them with the subject. In the following, when simply referred to as "specific part," it refers to a specific part of the subject to be processed that is detected by the information processing device according to this embodiment. In the following, such a specific part may also be referred to as "local" or "local part." In this embodiment, we will describe an example where the human body (entire body) is used as the subject to be processed, and the head of the human body is used as the specific part of the subject. However, this is not particularly limited as long as similar processing can be performed. For example, the subject and its specific part may be the head and pupils of a human body, an animal (entire body) and its face, or a vehicle and its license plate. Furthermore, the specific part does not necessarily have to be part of the subject, as long as it is an object that is expected to be imaged in conjunction with the subject. For example, the subject and its specific part may be a vehicle or animal and the head of a person riding in that vehicle. Although the person's head is not part of the vehicle, it moves in conjunction with the tracking vehicle, so it can be considered a specific part of the subject. Figure 1 shows an example of the configuration of the information processing device 100 according to this embodiment. As an example of hardware configuration, the information processing device 100 includes a CPU 101, a computer bus 102, a first memory 103, a second memory 104, an input unit 105, a display unit 106, and a communication unit 107. The CPU 101 controls the entire information processing device 100. The first memory 103 and the second memory 104 are memories that store control programs for executing the processing according to this embodiment, as well as various data. In Figure 1, the first memory 103 is shown mainly to store control programs, and the second memory 104 is shown mainly to store various data; however, this is not particularly limited as long as the entire information processing device 100 can store similar data. The input unit 105 consists of a keyboard or touch panel, etc., and accepts input from the user. The display unit 106 consists of a display device such as a liquid crystal display, and can display processing results to the user. The communication unit 107 can communicate with external devices to exchange data. The computer bus 102 connects the various functional units of the information processing device 100. The information processing device 100 according to this embodiment may be implemented as a computer equipped with a program that performs the processes described below. I