Search

JP-7857475-B2 - Receiver

JP7857475B2JP 7857475 B2JP7857475 B2JP 7857475B2JP-7857475-B2

Inventors

  • 青木 秀一

Assignees

  • 日本放送協会

Dates

Publication Date
20260512
Application Date
20250515
Priority Date
20191227

Claims (5)

  1. A receiving device that receives multiple video signals transmitted by a transmitting device and captured from different locations, along with location information of the imaging device that captured the video signals, The aforementioned video signals are grouped into multiple groups according to the shooting position. A communication interface that simultaneously receives all video signals from the transmitting device that are assigned an ID for identifying the group, The system includes a control unit that decodes the aforementioned video signal to generate a decoded video and acquires the user's viewpoint position, If there is a change in viewpoint position that does not involve a change in the ID, the control unit identifies the location information of the imaging device that captured multiple video signals necessary to generate an image corresponding to the user's viewpoint position, and performs viewpoint interpolation using the decoded image corresponding to the location information. A receiving device that, in the event of a change in viewpoint position accompanied by a change in the aforementioned ID, acquires the default decoded video from the decoded video of the video group corresponding to the new viewpoint position.
  2. The receiving device according to claim 1, wherein the first-level group among the multiple groups is a group to which multiple video signals used to generate images from a viewpoint different from the shooting position belong.
  3. The receiving device according to claim 2, wherein the second-tier group among the plurality of groups is one or more first-tier groups to which video signals captured within the same location belong.
  4. The receiving device according to claim 2 or 3, wherein the video signals belonging to the first hierarchical group are compressed using predictive coding based on the correlation between video signals.
  5. The receiving device according to any one of claims 1 to 4, wherein the ID and the video signal are stored in a file in ISO Base Media File Format.

Description

This invention relates to a receiving device for receiving video signals. As an evolution of 360-degree video, video systems that allow users to view 360-degree video from their preferred viewpoint are being considered. For example, Non-Patent Literature 1 lists use cases such as "3D image messaging" and "immersive 6DoF streaming," which involve viewpoint movement within the range of head movement while the user is seated (3DoF+) or viewpoint movement within the range of free movement (6DoF). Generally, a single camera captures images from a single viewpoint. By using multiple images (multi-viewpoint images) captured by multiple cameras, it's possible to generate images from intermediate positions between the points where those images were taken. This is a known technique known as viewpoint interpolation or intermediate viewpoint image generation (see, for example, Patent Document 1). Furthermore, Non-Patent Document 2 specifies OMAF (Omnidirectional Media Format) as a file format for storing 360-degree video signals from a predetermined viewpoint. Patent No. 6306952 3GPP TR 26.928, "3rd Generation Partnership Project; Technical Specification Group SA WG4 Extended Reality in 5G; (Release 16)", ver 0.5.0, 2019-07ISO/IEC 23090-2: 2019, “Information technology - Coded representation of immersive media - Part 2: Omnidirectional media format” This is a block diagram showing an example of a video transmission system according to the first embodiment.This is a diagram illustrating the first group ID according to the first embodiment.This is a diagram illustrating the second group ID according to the first embodiment.This diagram illustrates the hierarchical structure of a video signal according to the first embodiment.This figure shows an example of the configuration of a file that stores a video signal according to the first embodiment.This figure shows an example of the configuration of a file that stores a video signal according to the first embodiment.This figure shows an example of the display of a receiving device according to the first embodiment.This is a sequence diagram showing an example of operation of the video transmission system according to the first embodiment.This is a block diagram showing an example of a video transmission system according to the second embodiment.This is a sequence diagram showing an example of operation of the video transmission system according to the second embodiment.This is a block diagram showing an example of a video transmission system according to the third embodiment.This is a sequence diagram showing an example of operation of the video transmission system according to the third embodiment.This is a block diagram illustrating the general configuration of a computer. The embodiments will be described in detail below with reference to the drawings. <First Embodiment> (Video transmission system) First, a video transmission system according to the first embodiment will be described. Figure 1 is a diagram showing a video transmission system 1 according to the first embodiment, illustrating the outline of the transmission of multiple video signals captured at multiple locations. The video transmission system 1 comprises multiple imaging devices (cameras) 10, multiple first transmitting devices 20, a second transmitting device 30, and a receiving device 40. In this specification, "location" refers to a predetermined area established for a specific purpose, such as a venue, stadium, facility, or park. In this embodiment, the video transmission system 1 comprises three first transmitting devices 20, with the first transmitting device 20A transmitting a video signal captured at location A, the first transmitting device 20B transmitting a video signal captured at location B, and the first transmitting device 20C transmitting a video signal captured at location C, as will be described below. The imaging device 10 adds metadata to the video signal captured at a predetermined position and transmits it to the first transmitting device 20. Here, the metadata transmitted by the imaging device 10 is position information indicating the imaging position of the imaging device 10, and may also include information indicating the imaging direction. Furthermore, the metadata may also include distance information indicating the distance from the imaging device 10 to the captured object. In other words, multiple imaging devices 10A installed at location A each attach metadata to the video signals captured at predetermined locations in location A and transmit them to the first transmitting device 20A. Similarly, multiple imaging devices 10B installed at location B each attach metadata to the video signals captured at predetermined locations in location B and transmit them to the first transmitting device 20B. Similarly, multiple imaging devices 10C installed at location C each attach metadata to the video signals captured at predetermined locations in location C and transmit them to the first