Search

JP-2026074524-A - Information processing device, information processing method, and program

JP2026074524AJP 2026074524 AJP2026074524 AJP 2026074524AJP-2026074524-A

Abstract

[Challenge] To output user representations (avatars, etc.) that feel natural and natural. [Solution] The information processing device 102 of the present disclosure is an information processing device that processes a user's expression in at least one of CG and live-action images, and comprises a determination means 114 for determining whether the user's expression is a target for processing, and a processing means 112 for processing the user's expression based on the determination result of the determination means 114, the processing means 112 processing the user's expression based on past user expressions and target user expressions. [Selection Diagram] Figure 1

Inventors

  • 大川 修一

Assignees

  • キヤノン株式会社

Dates

Publication Date
20260507
Application Date
20241021

Claims (9)

  1. An information processing device for processing user representations in at least one of CG and live-action images, A determination means for determining whether the user's expression is subject to processing, The system includes a processing means for processing the user's expression based on the determination result of the determination means, The processing means is characterized by processing the user's expression based on the user's past expression and the user's target expression.
  2. The information processing apparatus according to claim 1, characterized in that the target user representation is a representation based on the current user representation.
  3. The information processing apparatus according to claim 1, characterized in that the user's expression includes at least one of visual information and auditory information.
  4. The information processing apparatus according to claim 1, characterized in that it has an output means for outputting the user's expression processed by the processing means.
  5. The information processing apparatus according to claim 4, characterized in that the output means is a display means for displaying the user's expression.
  6. The information processing apparatus according to claim 4, characterized in that the output means is a recording means for recording the user's expression.
  7. The information processing apparatus according to claim 4, characterized in that the output means is a communication means for transmitting the user's expression via communication.
  8. An information processing method for processing user representations in at least one of CG and live-action images, A determination step to determine whether the user's expression is subject to processing, The process includes a processing step for processing the user's expression based on the determination result of the determination step, The processing step is characterized by processing the user's expression based on the user's past expression and the user's target expression.
  9. A program for causing a computer to execute each step of the information processing method described in claim 8.

Description

This disclosure relates to an information processing device, and more particularly to an information processing device for processing user expressions. A technology for controlling the display of a user's avatar is known (Patent Document 1). Japanese Patent Publication No. 2024-77887 This is a block diagram showing the configuration of an information processing system.This flowchart shows the controls used when processing user expressions.This is a settings screen for when to start and stop processing the user's expression.This is a conceptual diagram illustrating how to process user expressions.This is a conceptual diagram illustrating how to process user expressions. (First embodiment) (composition) Figure 1 is a block diagram showing an example of the system configuration according to the first embodiment. The system according to the first embodiment is an image processing system for presenting a mixed reality (MR) space, which fuses the real space and the virtual space, to the system user. In the first embodiment, we will describe a case in which the MR space is presented to the user by displaying a composite image that combines an image of the virtual space drawn by computer graphics (CG) and an image of the real space (real-life image). The system according to the first embodiment includes a display device 101, an information processing device 102, and an operating device 103. The information processing device 102 is composed of the following components. The functions of each component are configured by one or more CPUs (central processing units) that function as the control unit of the information processing device 102 executing programs. The components of the information processing device 102 may be composed of integrated circuits or the like, as long as they perform similar functions. The information processing device 102 combines an image of the real world captured from the display device 101 with an image of the virtual world generated in the information processing device 102 to produce a composite image. The information processing device 102 outputs this composite image as a mixed reality (MR) image to the display device 101. Note that the first embodiment relates to an information processing device for displaying images of the virtual world, and the system according to the first embodiment is not limited to an MR system that displays an MR image (an image created by combining an image of the real world with an image of the virtual world). In other words, the system according to the first embodiment may be a so-called XR system, such as a VR (virtual reality) system that presents only images of the virtual world to the user, or an AR (augmented reality) system that presents images of the virtual world to the user by making the real world transparent. The display device 101 includes a recording unit 104, an audio output unit 105, an imaging unit 106, and a display unit 107. The recording unit 104 captures sounds such as those around the display unit 101 and the user's own voice as audio data and outputs it to the information processing device 102. The recording unit 104 may be a directional microphone or a movable lavalier microphone, depending on the application. Multiple such recording devices may also be provided. The audio output unit 105 outputs audio data from the information processing device 102. The audio output unit 105 may be a speaker, earphones, or headphones. Furthermore, multiple such audio output devices may be provided, and each audio output device may be wired or wireless. Alternatively, for example, a device equipped with both a recording unit 104 and an audio output unit 105, such as wireless earphones with a microphone function, may communicate with the display device 101 or the information processing device 102. The imaging unit 106 continuously images the real space in a time series and outputs the captured images of the real space (captured images) to the information processing device 102. The imaging unit 106 may include a stereo camera consisting of two cameras fixed to each other so that it can capture the real space in the direction of the user's line of sight. In addition to a camera that captures the outside of the display device 101 as a reference, it may also be equipped with a camera that captures the inside so that it can acquire the user's expressions such as facial expressions and body language. Furthermore, a camera other than the display device 101 may be used in conjunction to capture the user's expressions. For example, a camera fixed to a tripod or the wall of a building may be used in conjunction. In that case, the captured images from the camera other than the display device 101 may be transmitted to the information processing device 102 via the display device 101, or they may be transmitted directly to the information processing device 102. The user's expressions are transmitted to the information processing device 102 as captured images and stored in t