KR-102965316-B1 - METHOD AND APPARATUS FOR ESTIMATING POSE, COMPUTER-READABLE STORAGE MEDIUM AND COMPUTER PROGRAM FOR CONTROLLING THE HOLDER DEVICE
Abstract
The pose estimation method of the embodiment may include the steps of: generating a plurality of images based on an input image, including a first image of a first reference size and a second image of a second reference size larger than the size of the first image; extracting feature points of the first image and the second image; estimating the pose of the first image by matching the first image with a plurality of DB images; estimating the pose of the second image by matching the second image of the first image with the plurality of DB images; and determining one pose among the estimated poses of the plurality of second images according to a preset criterion. The embodiment has the effect of improving processing speed by changing the size of the input image, measuring the pose for the first image with a smaller size, and estimating the pose for the second image for the first image in which the pose was measured.
Inventors
- 임승욱
- 유연걸
- 윤찬민
- 정준영
Assignees
- 에스케이텔레콤 주식회사
Dates
- Publication Date
- 20260513
- Application Date
- 20191029
Claims (10)
- A method performed in a pose estimation device for estimating the pose of a camera equipped in a terminal, A step of generating a plurality of images based on an input image collected from the camera, comprising a first image of a first reference size smaller than the size of the input image and a second image of a second reference size larger than the size of the first image and smaller than or equal to the size of the input image; A step of extracting feature points of the first image and the second image; A step of estimating the pose of the first image by matching feature points of the first image with feature points of a plurality of DB images; A step of estimating the pose of the second image by matching the feature points of the second image with the feature points of the plurality of DB images for the first image in which the pose is estimated; and A step of determining one pose among the multiple estimated poses of the second images according to a preset criterion. Includes, A pose estimation method in which the number of feature points of the first image is less than the number of feature points of the second image.
- In paragraph 1, The step of generating the first image and the second image is A pose estimation method for generating the first image and the second image by changing the size of the input image along the major axis.
- In paragraph 2, The step of generating the first image and the second image is A pose estimation method in which, if the size of the input image is larger than the second reference size, the size of the input image is changed to a size between the first image and the second image, and if the size of the input image is smaller than the second reference size, the input image is used as the second image.
- In paragraph 1, The step of estimating the pose of the second image above is, A step of matching feature points of the second image with feature points of the plurality of DB images; A step of removing unnecessary feature points among the above-mentioned matched plurality of feature points; and Step of estimating the pose of the second image above A pose estimation method including
- In paragraph 4, The step of estimating the pose of the second image above is, A pose estimation method for estimating the pose of the second image using 3D coordinate values of the DB image at feature points of the second image.
- In paragraph 1, The step of determining one pose according to the above-mentioned preset criteria is a pose estimation method that determines one pose by considering the number of matched feature points between the second image and the plurality of DB images and the error rate of the average distance between the matched feature points.
- In paragraph 6, The step of determining a pose according to the above-mentioned preset criteria is a pose estimation method that determines the pose by considering the variance value of the 3D coordinates of the feature points.
- In Paragraph 7, The above average value and variance value are obtained by a pose estimation method that projects 3D coordinate values onto the second image.
- As a computer-readable recording medium storing a computer program, The above computer program is, A computer-readable recording medium comprising instructions for a processor to perform a method according to any one of claims 1 through 8.
- As a computer program stored on a computer-readable recording medium, The above computer program is, A computer program comprising instructions for a processor to perform a method according to any one of paragraphs 1 through 8.
Description
Method and apparatus for estimating pose, computer-readable storage medium and computer program The embodiment relates to a pose estimation method for effectively estimating the pose of a camera equipped in a terminal. Augmented Reality (AR) is a technology that uses location and orientation information to determine an approximate location, identifies the services desired by the user through a comparison between facility information—such as surrounding buildings—and real-world video information input based on camera movement, and provides relevant information. More specifically, augmented reality is a field of virtual reality (VR) that uses computer graphics techniques to superimpose virtual objects onto a real environment, making them appear as if they exist within the original environment. Unlike conventional virtual reality, which focuses solely on virtual spaces and objects, augmented reality is a technology that superimposes virtual objects onto a foundation of the real world to provide supplementary information that is difficult to obtain from the real world alone. With the commercialization of 5G communication, such augmented reality technology is gaining prominence in the field of mobile AR technology used in communication terminals, and marker-based mobile AR technology or sensor-based mobile AR technology is generally used in current mobile AR technology applications. Marker-based mobile AR technology is a technology that recognizes a real object by recognizing a marker corresponding to the real object when photographing a real object to be augmented using a virtual object, and sensor-based mobile AR technology is a technology that infers the current location and the direction the terminal is looking by using GPS and a digital compass installed in the terminal, and overlays POI (Point of Interests) information corresponding to the image in the inferred direction. However, marker-based mobile AR technology has the problem that it cannot augment virtual objects without markers, and sensor-based mobile AR technology has the problem of being unable to accurately augment virtual objects onto specific objects due to errors in the current location and orientation of the detected device. FIG. 1 is a block diagram schematically showing an apparatus for performing a pose estimation method according to an embodiment. FIG. 2 is a flowchart illustrating a pose estimation method according to an embodiment. FIG. 3 is a diagram illustrating the step of generating an image according to an embodiment. FIG. 4 is a diagram illustrating the step of extracting feature points of an image according to an embodiment. FIG. 5 is a flowchart illustrating the steps for estimating the pose of an image according to an embodiment. FIGS. 6 to 9 are drawings for explaining the step of estimating the pose of an image according to an embodiment. FIG. 10 is a drawing illustrating the steps for determining an optimal pose according to an embodiment. Hereinafter, embodiments will be described in detail with reference to the drawings. FIG. 1 is a block diagram schematically showing an apparatus for performing a pose estimation method according to an embodiment. Referring to FIG. 1, the pose estimation method according to the embodiment can be performed in a pose estimation device (100). The pose estimation device (100) may receive an input image from a terminal (200). The terminal (200) may include various terminals such as mobile phones and computers. The terminal (200) may be equipped with a camera or a sensor. The pose estimation device (100) may obtain an input image from the camera or sensor of the terminal (200). The pose estimation device (100) can receive DB images from DB (300). DB (300) may be an area where accurate image information is stored. DB (300) may be created internally by collecting actual images and may be an area of a portal site such as Naver, Daum, or Google. The pose estimation device (100) can estimate the position of the terminal (200) using the input image and DB image, and can establish a foundation for providing AR services using this. The pose estimation device (100) can generate input images into multiple images. The pose estimation device (100) can generate an input image having a first image having a first reference size and a second image having a second reference size larger than the first reference size. Alternatively, the pose estimation device (100) may generate three or more images. The sizes of the first and second images may be changed to be smaller than the size of the input image. The second image may be the same size as the input image. The pose estimation device (100) can extract feature points for a first image. The pose estimation device (100) can extract feature points for a second image. Since the size of the first image is smaller than the size of the second image, the number of feature points in the first image may be less than the number of feature points in the second image. The pose