(19)
(11)EP 3 393 130 B1

(12)EUROPEAN PATENT SPECIFICATION

(45)Mention of the grant of the patent:
29.04.2020 Bulletin 2020/18

(21)Application number: 16875499.2

(22)Date of filing:  07.12.2016
(51)International Patent Classification (IPC): 
H04N 21/236(2011.01)
H04N 21/431(2011.01)
H04H 20/95(2008.01)
G10L 19/16(2013.01)
H04N 21/439(2011.01)
H04H 60/07(2008.01)
H04N 21/488(2011.01)
(86)International application number:
PCT/JP2016/086462
(87)International publication number:
WO 2017/104519 (22.06.2017 Gazette  2017/25)

(54)

TRANSMISSION DEVICE, TRANSMISSION METHOD, RECEIVING DEVICE AND RECEIVING METHOD FOR ASSOCIATING SUBTITLE DATA WITH CORRESPONDING AUDIO DATA

SENDEVORRICHTUNG, SENDEVERFAHREN, EMPFANGSVORRICHTUNG UND EMPFANGSVERFAHREN ZUM ZUORDNEN VON UNTERTITELDATEN ZU ENTSPRECHENDEN AUDIODATEN

DISPOSITIF ET PROCÉDÉ D'ÉMISSION, DISPOSITIF ET PROCÉDÉ DE RÉCEPTION POUR ASSOCIER DES DONNÉES DE SOUS-TITRES AUX DONNÉES AUDIO CORRESPONDANTES


(84)Designated Contracting States:
AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

(30)Priority: 18.12.2015 JP 2015247266

(43)Date of publication of application:
24.10.2018 Bulletin 2018/43

(73)Proprietor: Sony Corporation
Tokyo 108-0075 (JP)

(72)Inventor:
  • TSUKAGOSHI, Ikuo
    Minato-Ku Tokyo 108-0075 (JP)

(74)Representative: 2SPL Patentanwälte PartG mbB 
Postfach 15 17 23
80050 München
80050 München (DE)


(56)References cited: : 
WO-A1-2015/126117
JP-A- 2007 300 323
JP-A- 2015 159 365
JP-A- 2002 354 366
JP-A- 2009 171 393
US-A1- 2013 002 833
  
  • YONGTAO HU ET AL: "Speaker-Following Video Subtitles", ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS ANDAPPLICATIONS, ASSOCIATION FOR COMPUTING MACHINERY, US, vol. 11, no. 2, 7 January 2015 (2015-01-07), pages 1-17, XP058064903, ISSN: 1551-6857, DOI: 10.1145/2632111
  • SIMONE FÜG ET AL: "Metadata Updates to MPEG-H 3D audio", 112. MPEG MEETING; 22-6-2015 - 26-6-2015; WARSAW; (MOTION PICTURE EXPERT GROUP OR ISO/IEC JTC1/SC29/WG11),, no. m36586, 17 June 2015 (2015-06-17), XP030064954,
  • MEHTA AND T ZIEGLER S: "Personalized and immersive broadcast audio", IBC 2014 CONFERENCE, 11-15 SEPTEMBER 2014, AMSTERDAM,, 11 September 2014 (2014-09-11), XP030082501,
  
Note: Within nine months from the publication of the mention of the grant of the European patent, any person may give notice to the European Patent Office of opposition to the European patent granted. Notice of opposition shall be filed in a written reasoned statement. It shall not be deemed to have been filed until the opposition fee has been paid. (Art. 99(1) European Patent Convention).


Description

[Technical Field]



[0001] The present technology relates to a transmission apparatus, a transmission method, a reception apparatus, and a reception method, and more particularly, to a transmission apparatus and so on for sending subtitle information together with video information.

[Background Art]



[0002] It has been common practice up to now in DVB (Digital Video Broadcasting)-based broadcasting and so on to send subtitle information using bitmap data. In recent years, proposals have been made to send subtitle information using text character codes, i.e., in a text-based form. In this case, font expansion is conducted to match the resolution of the receiving side.

[0003] Also, in the case where subtitle information is sent in a text-based form, a proposal has been made to carry timing information in text information. As this text information, for example, TTML (Timed Text Markup Language) has been advocated by W3C (World Wide Web Consortium) (refer to PTL 1).

[0004] Also, for example, PTL 2 describes, as three-dimensional (3D) acoustic technology, a 3D audio rendering technology that maps audio data of an object sound source onto a speaker existing at an arbitrary position on the basis of position information thereof.

[0005] In the case where 3D audio is delivered together with video, object information (position information) can be configured such that it is possible for a viewer to recognize that a voice produced by a speaker, one of video image objects, comes from the same direction as the speaker irrespective of a speaker position.

[Citation List]


[Patent Literature]



[0006] 

[PTL 1]
JP 2012-169885A

[PTL 2]
JP-T-2014-520491



[0007] In document US 2017/180767 A1 an apparatus and a method are shown for transceiving a broadcast signal including a subtitling service. Accordingly, a method for transmitting a broadcast signal is shown, comprising encoding video data to generate a video stream, generating a broadcast signal including the generated video stream and a subtitle stream, and transmitting the generated broadcast signal. Accordingly, a transmission stream which provides a digital broadcast subtitling service using an XML subtitle can be transmitted.

[0008] In document "speaker-following video subtitles" by Yongtao Hu et al., ACM Transactions on multimedia computing communications and applications, association for computing machinery, US, vol. 11, no. 2, 7 January 2015, a method for improving the presentation of subtitles in video is proposed. An identification algorithms to detect the speakers based on audio and visual information is proposed. The placement of the subtitles is determined using global optimization.

[0009] In document US 2013 002833 (A1) systems and methods are presented for processing three-dimensional (3D) programming. The programming includes closed caption (CC) information that includes caption data and a location identifier that specifies a location for the caption data within the 3D programming. The programming information is processed to render the caption data at the specified location and to present the programming on the display. By encoding location identification information into the three-dimensional programming, a high level of configurability can be provided and the 3D experience can be preserved while captions are displayed.

[0010] In document "metadata updates to MPEG-H 3D audio", by Simone Füg et al., 112. MPEG meeting, 22-6-2015 to 26-6-2015, Warsaw, a range of metadata extensions is proposed for MPEG-H 3D audio. It is proposed to enable to convey metadata for rendering objects that are relevant in upcoming content formats.

[0011] In document "personalized and immersive broadcast audio" by S. Mehta et al., IBC 2014 conference, 11-15 September 2014, Amsterdam, 11 September 20014, the use of Dolby AC-4 format is proposed.

[Summary]


[Technical Problem]



[0012] It is an object of the present technology to effectively display subtitles on a receiving side.

[Solution to Problem]



[0013] A concept of the present technology lies in a transmission apparatus that includes a video encoding section, a subtitle encoding section, an audio encoding section, a transmission section, and an information insertion section. The video encoding section generates a video stream having video coded data. The subtitle encoding section generates a subtitle stream having subtitle data corresponding to a speech of a speaker. The audio encoding section generates an audio stream having object coded data including audio data whose sound source is a speech of the speaker and position information regarding a position of the sound source in an image of the video stream. The transmission section sends a container stream in a given format including the video stream, the subtitle stream, and the audio stream. The information insertion section inserts, into a layer of a container stream, metainformation for associating subtitle data corresponding to each speech included in the subtitle stream with audio data corresponding to each speech included in the audio stream and with position information associated with said audio data.

[0014]  In the present technology, a video stream having video coded data is generated by the video encoding section. A subtitle stream having subtitle data corresponding to a speech of a speaker is generated by the subtitle encoding section. For example, subtitle data may be TTML data or data in a format derived from the TTML.

[0015] An audio stream is generated by the audio encoding section. The audio stream has object coded data including audio data whose sound source is a speech of a speaker and position information regarding the sound source. A container stream in a given format is sent by the transmission section. The container stream includes the video stream, the subtitle stream, and the audio stream.

[0016] Metainformation is inserted into a layer of a container stream. The metainformation associates subtitle data corresponding to each speech included in the subtitle stream and audio data corresponding to each speech included in the audio stream.

[0017] For example, identification information may be added to each piece of subtitle data corresponding to each speech, and identification information may be added to each piece of audio data corresponding to each speech. Metainformation may be table information that associates identification information added to the subtitle data corresponding to each speech and identification information added to the audio data corresponding to each speech. Then, in this case, tag information having identification information may be inserted into an audio frame including audio data corresponding to each speech.

[0018] Thus, in the present technology, metainformation for associating subtitle data corresponding to each speech included in the subtitle stream and audio data corresponding to each speech included in the audio stream is inserted into a layer of a container stream. For this reason, in the case where a subtitle display position is controlled on a receiving side using position information regarding a sound source included in object coded data such that the display position is located at a speaker's position of a video image, it is possible to readily use adequate position information by using this metainformation.

[0019] Also, another concept of the present technology lies in a reception apparatus that includes a reception section that receives a container stream in a given format including a video stream, a subtitle stream, and an audio stream. The video stream has video coded data. The subtitle stream has subtitle data corresponding to a speech of a speaker. The audio stream has object coded data including audio data whose sound source is a speech of a speaker and position information regarding the sound source. The reception apparatus further includes a control section that controls a video decoding process, a subtitle decoding process, a video superimposition process, an audio decoding process, and a superimposition position process. The video decoding process acquires video data through a decoding process of the video stream. The subtitle decoding process acquires bitmap data of a subtitle corresponding to the speaker's speech through a decoding process of the subtitle stream. The video superimposition process acquires video data for display through superimposition of the bitmap data of the subtitle on the video data. The audio decoding process acquires channel data corresponding to the speech through a decoding process of the audio stream. The superimposition position process finds display coordinates of a subtitle corresponding to each of the speeches on a monitor screen on the basis of position information regarding the sound source corresponding to each of the speeches and performs control such that the bitmap data of the subtitle corresponding to each of the speeches is superimposed on the video data at the position corresponding to the display coordinates found.

[0020] In the present technology, the reception section receives a container stream in a given format including a video stream, a subtitle stream, and an audio stream. Here, the video stream has video coded data. The subtitle stream has subtitle data corresponding to a speech of a speaker. The audio stream has object coded data including audio data whose sound source is a speech of a speaker and position information regarding the sound source.

[0021] Video data is acquired by a video decoding section through a decoding process of the video stream. Bitmap data of a subtitle corresponding to a speech of a speaker is acquired by a subtitle decoding section through a decoding process of the subtitle stream. Video data for display is acquired by a video superimposition section through superimposition of bitmap data of a subtitle on the video data.

[0022] Channel data corresponding to a speech is acquired by an audio decoding section through a decoding process of the audio stream. Display coordinates of a subtitle corresponding to each speech on a monitor screen are found by a control section on the basis of position information regarding a sound source corresponding to each of the speeches, and superimposition of bitmap data of the subtitle corresponding to each of the speeches is controlled by the control section such that the bitmap data is superimposed on the video data at the position corresponding to the display coordinates found.

[0023] Thus, the present technology finds a position on a monitor screen where a subtitle corresponding to each speech is superimposed on the basis of position information regarding a sound source corresponding to each of the speeches and performs control such that the bitmap data of the subtitle corresponding to each of the speeches is superimposed on the video data at the position corresponding to the display coordinates found. For this reason, it is possible to display a subtitle corresponding to a speech of a speaker at the speaker's position in a video image.

[0024] It should be noted, for example, that, in the present technology, metainformation is inserted, into a container stream, that associates subtitle data corresponding to each speech included in a subtitle stream and audio data corresponding to each speech included in an audio stream, and that the control section may recognize, on the basis of this metainformation, correspondence between subtitle data corresponding to each speech and audio data corresponding to each speech.

[Advantageous Effect of Invention]



[0025] According to the present technology, it is possible to effectively display subtitles on a receiving side. It should be noted that the effect described in the present specification is merely illustrative and is not limited, and that there may be additional effects.

[Brief Description of Drawings]



[0026] 

[Fig. 1]
Fig. 1 is a block diagram illustrating a configuration example of a transmission/reception system as an embodiment.

[Fig. 2]
Fig. 2 is a diagram illustrating an example of a video image displayed on a monitor screen.

[Fig. 3]
Fig. 3 is a block diagram illustrating a configuration example of a stream generation section included in a service transmitter.

[Fig. 4]
Fig. 4 is a diagram illustrating a TTML structure.

[Fig. 5]
Fig. 5 is a diagram illustrating a structural example of metadata, styling, and layout, each of which is a component existing in a TTML structure head.

[Fig. 6]
Fig. 6 is a diagram illustrating an example of a TTML structure body.

[Fig. 7]
Fig. 7 is a diagram illustrating a configuration example of MPEG-H 3D Audio transport data.

[Fig. 8]
Fig. 8 is a diagram illustrating a structural example of an audio frame in MPEG-H 3D Audio transport data.

[Fig. 9]
Fig. 9 is a diagram for describing a packet type that is defined anew.

[Fig. 10]
Fig. 10 is a diagram illustrating an example of correspondence between an audio stream having consecutive audio frames including audio compressed data and sound units.

[Fig. 11]
Fig. 11 is a set of diagrams illustrating, in a simplified manner, examples of details of tag information 'Speech_tag_information()' included in an audio frame.

[Fig. 12]
Fig. 12 is a diagram illustrating an example of insertion of tag information 'Speech_tag_information()' and table information 'Tag_table()' into an audio stream.

[Fig. 13]
Fig. 13 is a diagram illustrating an example of details of table information 'Tag Table 1' inserted into the audio stream.

[Fig. 14]
Fig. 14 is a diagram illustrating a structural example of tag information 'Speech_tag_information().'

[Fig. 15]
Fig. 15 is a diagram illustrating a structural example of table information 'Tag table().'

[Fig. 16]
Fig. 16 is a diagram illustrating details of main pieces of information in the structural examples of tag information 'Speech_tag_information()' and table information 'Tag table().'

[Fig. 17]
Fig. 17 is a diagram illustrating a structural example of a link table descriptor (Link_table_descriptor()).

[Fig. 18]
Fig. 18 is a set of diagrams illustrating association between 'speech_id' and 'subtitle_speech_region_id' established by the link table descriptor.

[Fig. 19]
Fig. 19 is a diagram illustrating a configuration example of a transport stream TS.

[Fig. 20]
Fig. 20 is a block diagram illustrating a configuration example of a service receiver.

[Fig. 21]
Fig. 21 is a block diagram illustrating a configuration example of an audio decoding section.

[Fig. 22]
Fig. 22 is a diagram illustrating an overview of a rendering process in an object renderer.

[Fig. 23]
Fig. 23 is a diagram for describing a calculation example of display coordinates.

[Fig. 24]
Fig. 24 is a diagram illustrating an example of multi-region display in which a plurality of speakers exist in a video image and in which subtitles, each corresponding to a speech of each of the speakers, are superimposed on the video image.

[Fig. 25]
Fig. 25 is a diagram illustrating a configuration example of an MMT stream.

[Fig. 26]
Fig. 26 is a pair of diagrams illustrating a configuration example of an MP4 stream (file) including an audio track data and a subtitle track data.

[Fig. 27]
Fig. 27 is a diagram illustrating an example of describing an MPD file.

[Fig. 28]
Fig. 28 is a diagram illustrating details of main pieces of information in the example of describing an MPD file.

[Fig. 29]
Fig. 29 is a pair of diagrams illustrating a structure of an AC4 Simple Transport layer.

[Fig. 30]
Fig. 30 is a pair of diagrams illustrating schematic configurations of a TOC (ac4_toc()) and a substream (ac4_substream_data()).


[Description of Embodiments]



[0027] A description will be given below of a mode for carrying out the present invention (hereinafter referred to as an embodiment). It should be noted that the description will be given in the following order: 1. Embodiment

2. Modification example


<1. Embodiment>


[Configuration Example of the Transmission/Reception System]



[0028] Fig. 1 illustrates a configuration example of a transmission/reception system 10 as an embodiment. The transmission/reception system 10 includes a service transmitter 100 and a service receiver 200. The service transmitter 100 sends a MPEG-2 transport stream (hereinafter simply referred to as a 'transport stream TS') as a container stream (multiplexed stream) over broadcast waves or in a network packet.

[0029] The transport stream TS has a video stream, a subtitle stream, and an audio stream. The video stream has video coded data. The subtitle stream has subtitle data corresponding to a speech of a speaker. The audio stream has object coded data including audio data whose sound source is a speech of a speaker and position information regarding the sound source.

[0030] In the present embodiment, the subtitle data is TTML (Timed Text Markup Language) data having text information regarding subtitles. The audio stream has audio frames that are provided consecutively. The audio frames include audio compressed data acquired through an encoding process of audio data. In the present embodiment, MPEG-H 3D Audio is used to code audio streams.

[0031] The service transmitter 100 inserts, into an audio frame including audio data corresponding to each speech, tag information having identification information for identifying the audio data. Also, the service transmitter 100 inserts, into subtitle data corresponding to each speech, identification information for identifying the subtitle data.

[0032] Also, the service transmitter 100 inserts, into a layer of the transport stream TS as a container stream, metainformation for associating subtitle data corresponding to each speech included in the subtitle stream and audio data corresponding to each speech included in the audio stream. In the present embodiment, metainformation is table information that associates identification information added to the subtitle data corresponding to each speech and identification information added to the audio data corresponding to each speech.

[0033] The service receiver 200 receives the transport stream TS sent from the service transmitter 100 over broadcast waves or in a network packet. This transport stream TS includes a subtitle stream and an audio stream in addition to a video stream as described above.

[0034] The service receiver 200 acquires video data for display by superimposing bitmap data of a subtitle corresponding to each speech acquired through a decoding process of the subtitle stream on video data acquired through a decoding process of the video stream. Also, the service receiver 200 acquires channel data corresponding to the speech through a decoding process of the audio stream. A voice resulting from this channel data is recognized by a viewer as if the voice were heard from a direction of a speaker included in one of video image objects.

[0035] The service receiver 200 finds display coordinates of a subtitle corresponding to each of the speeches on a monitor screen on the basis of position information regarding the sound source corresponding to each of the speeches and performs control such that the bitmap data of the subtitle corresponding to each of the speeches is superimposed on the video data at the position corresponding to the display coordinates found. This makes it possible to display a subtitle corresponding to each speech at the speaker's position in a video image.

[0036] Here, the service receiver recognizes correspondence between subtitle data corresponding to each speech and audio data corresponding to each speech, for example, on the basis of metainformation for associating subtitle data corresponding to each speech included in a subtitle stream and audio data corresponding to each speech included in an audio stream that are sent inserted in the layer of the transport stream TS as a container stream.

[0037] Fig. 2 illustrates an example of a video image displayed on a monitor screen. In the example illustrated, there is a speaker (depicted by an arrow A) . The speaker is included in one of objects in a video image. A voice of the speaker's speech is recognized by a viewer as if the voice were heard from the speaker's direction because of channel data acquired from rendering using audio data whose sound source is the speaker and position information regarding the sound source. Here, (θ, φ, r) are position information. θ is the azimuth, φ is the elevation, and r is the radius. In the example illustrated, although the direction 'm' is used as a reference direction, the reference direction is not limited thereto.

[0038] Also, a position where a subtitle corresponding to the speaker's speech is superimposed on the monitor screen is found, for example, as depicted by a dotted line rectangular frame (depicted by an arrow B), and the subtitle is displayed here. That is, the subtitle corresponding to the speaker's speech is displayed at the speaker's position in the video image. It should be noted that details of calculations for finding, from sound source position information, the position where a subtitle is superimposed on the monitor screen will be described later.

[Stream Generation Section of the Service Transmitter]



[0039] Fig. 3 illustrates a configuration example of a stream generation section 110 included in the service transmitter 100. This stream generation section 110 has a control section 111, a video encoder 112, a text format conversion section 113, a subtitle encoder 114, an audio encoder 115, and a multiplexer 116. The control section 111 includes a CPU 111a and controls the operation of each section of the stream generation section 110.

[0040] The video encoder 112 inputs video data SV and generates a video stream (video elementary stream) having video coded data by coding the video data SV.

[0041] The text format conversion section 113 inputs text data (character codes) as subtitle information and converts the data into subtitle information, and, in this embodiment, TTML data.

[0042] Fig. 4 illustrates a TTML structure. TTML is written in XML Base. Metadata, styling, and layout, each of which is an element, exist in a head. Fig. 5(a) illustrates a structural example of metadata (TTM: TTML Metadata). This metadata includes title information regarding the metadata and copyright information.

[0043] Fig. 5(b) illustrates a structural example of styling (TTS: TTML Styling). This styling includes, in addition to identification information (id), information such as color (color), font (fontFamily), size (fontSize), and alignment (textAlign).

[0044] Fig. 5(c) illustrates a structural example of layout (region: TTML layout). This layout includes, in addition to identification information (id) regarding a region where a subtitle is provided, information such as a start position (origin) of the region, the region's extent (extent), padding (padding), background color (backgroundColor), and alignment (displayAlign). 'subtitle_speech_region_id' is identification information regarding a subtitle speech region.

[0045] Fig. 6 illustrates a structural example of a body. In the example illustrated, three pieces of subtitle information, subtitle 1, subtitle 2, and subtitle 3 are included. Display start and end timings are written for each subtitle, and identification information regarding a subtitle speech region (subtitle_speech_region_id) and text data are also written. As for subtitle 1, for example, the display start timing is "0.76s," and the display end timing is "3.45s," and "subtitle_speech_region_id = id1," and text data is 'It seems a paradox, does it not,.'

[0046] Referring back to Fig. 3, the subtitle encoder 114 converts TTML data, acquired by the text format conversion section 113, into a variety of segments and generates a subtitle stream that includes a PES packet having these segments provided in a payload. This subtitle stream has subtitle data corresponding to a speech of a speaker.

[0047] The audio encoder 115 inputs, as audio data SA, object data of a given number of content groups together with channel data. One or a plurality of pieces of object content belong to each content group.

[0048] The audio encoder 115 acquires MPEG-H 3D Audio transport data by coding the audio data SA and generates an audio stream (audio elementary stream) including this transport data. 3D Audio transport data includes object coded data of a given number of content groups together with channel coded data.

[0049] The audio encoder 113 inserts, under control of the control section 111 and into an audio frame including audio compressed data of a sound unit corresponding to a speech of a speaker, tag information 'Speech_tag_information()' indicating that audio compressed data of this sound unit is included. Also, the audio decoder 113 inserts, under control of the control section 111 and into an arbitrary audio frame, table information 'Tag_table()' having sound unit information in an audio stream.

[0050] Fig. 7 illustrates a configuration example of MPEG-H 3D Audio transport data. In this configuration example, transport data includes a piece of channel coded data and three pieces of object coded data. The piece of channel coded data is 5.1-channel channel coded data (CD) and includes SCE1, CPE1.1, CPE1.2, LFE1, each of which is a piece of coded sample data.

[0051] The three pieces of object coded data belong to coded data (DOD) of a dialog object content group. These three pieces of object coded data are coded data of dialog objects corresponding to first, second, and third languages.

[0052] The respective pieces of coded data of the dialog objects corresponding to the first, second, and third languages include coded sample data SCE2, SCE3, and SCE4 and object metadata for rendering these pieces of coded sample data by mapping the sample data onto speakers existing at arbitrary positions. In the present embodiment, coded sample data in this object coded data is included in audio data whose sound source is a speech of a speaker.

[0053] Coded data is differentiated by type by the concept, group. In the present configuration example, 5.1-channel channel coded data is group 1 (Group 1). Also, the pieces of coded data of the dialog objects corresponding to the first, second, and third languages are group 2 (Group 2), group 3 (Group 3), and group 4 (Group 4), respectively. Also, that which can be selected between groups on the receiving side is registered as a switch group and coded. In this configuration example, group 2 (Group 2), group 3 (Group 3), and group 4 (Group 4) that belong to the dialog object content group are switch group 1 (SW Group 1) .

[0054] Fig. 8 is a diagram illustrating a structural example of an audio frame in MPEG-H 3D Audio transport data. This audio frame includes a plurality of MPEG audio frame packets (mpeg Audio Stream Packets). Each audio stream packet includes a header and a payload.

[0055] The header has information such as packet type, packet label, and packet length. Information defined in the header's packet type is provided in the payload. "SYNC" equivalent to a synchronous start code, "Frame," actual data of 3D Audio transport data, and "Config" indicating the configuration of this "Frame," are available as payload information.

[0056] "Frame" includes channel coded data and object coded data included in 3D Audio transport data. Here, channel coded data includes coded sample data of an SCE (Single Channel Element), a CPE (Channel Pair Element), an LFE (Low Frequency Element), and so on. Also, object coded data includes coded sample data of an SCE (Single Channel Element) and metadata for rendering the coded sample data by mapping the data onto a speaker existing at an arbitrary position. This metadata is included as an extension element (Ext_element).

[0057] Also, "TagTable" and "SpeechTagInformation" defined anew in the present embodiment exist in payload information. "SpeechTagInformation" includes tag information 'Speech_tag_information()' described above, and "TagTable" includes table information 'Tag_table()' described above. Fig. 9 illustrates that 'SpeechTagInformation' for including tag information 'Speech_tag_information()' and 'TagTable' for including table information 'Tag_table()' are defined anew.

[0058] Fig. 10 illustrates an example of correspondence between an audio stream having audio frames including audio compressed data provided consecutively and sound units. In this example, a sound unit 1 having voice details 1 and a sound unit 2 having voice details 2 exist.

[0059] The sound unit 1 has a start position in an audio frame f1 and an end position in an audio frame f2. A sample-by-sample offset of the audio frame f1 from the beginning of the frame to the start position is "xs1," and a sample-by-sample offset of the audio frame f2 from the beginning of the frame to the end position is "xe1." Also, the sound unit 2 has a start position in an audio frame f3 and an end position in an audio frame f4. A sample-by-sample offset of the audio frame f3 from the beginning of the frame to the start position is "xs2," and a sample-by-sample offset of the audio frame f4 from the beginning of the frame to the end position is "xe2."

[0060] Fig. 11 illustrating, in a simplified manner, examples of details of tag information 'Speech_tag_information()' included in an audio frame. It should be noted that frame count information, group identification information, switch group information, and so on are omitted in this example.

[0061] Fig. 11(a) is an example in which an audio frame includes a start position of a sound unit (refer to the audio frames f1 and f3 in Fig. 10). In this case, information such as sound unit identification information 'speech_id,' originator identification information 'speaker_id,' type information 'sted_type,' and offset information 'sample_offset' exists.

[0062] "speech_id = 1" indicates that identification information regarding the corresponding sound unit is "1." "speaker_id = 1" indicates that identification information regarding the originator (speaker) of the corresponding sound unit is "1." "sted_type = 1" indicates that type information is "1," that is, that a sound unit start position is included. Then, "sample_offset = xs" indicates that the sample-by-sample offset from the beginning of the frame to the start position is "xs."

[0063] Fig. 11(b) is an example in which an audio frame includes neither a start position nor an end position (refer to the audio frames starting from the one following the audio frame f1 to the one previous to the audio frame f2 and the audio frames starting from the one following the audio frame f3 to the one previous to the audio frame f4 in Fig. 10). In this case, information such as sound unit identification information 'speech_id,' originator identification information 'speaker-id,' and type information 'sted_type' exists.

[0064] "speech_id = 1" indicates that identification information regarding the corresponding sound unit is "1." "speaker_id = 1" indicates that identification information regarding the originator (speaker) of the corresponding sound unit is "1." "sted_type = 3" indicates that type information is "3," that is, that a sound unit continues with no start position or end position included.

[0065] Fig. 11(c) is an example in which an audio frame includes an end position of a sound unit (refer to the audio frames f2 and f4 in Fig. 10). In this case, information such as sound unit identification information 'speech_id,' originator identification information 'speaker_id,' type information 'sted_type,' and offset information 'sample_offset' exists.

[0066]  "speech_id = 1" indicates that identification information regarding the corresponding sound unit is "1." "speaker_id = 1" indicates that identification information regarding the originator (speaker) of the corresponding sound unit is "1." "sted_type = 2" indicates that type information is "2," that is, that a sound unit end position is included. Then, "sample_offset = xe" indicates that the sample-by-sample offset from the beginning of the frame to the end position is "xe."

[0067] Fig. 12 illustrates an example of insertion of tag information 'Speech_tag_information()' and table information 'Tag_table()' into an audio stream. This example focuses on group 2 (Group_id = 2) of switch group 1 (SW_group_id = 1) and illustrates a case in which speaker 1 (speaker_id = 1), speaker 2 (speaker_id = 2), and speaker 3 (speaker_id = 3) exist as originators of sound units.

[0068] Tag information 'Speech_tag_information()' regarding four sound units is inserted into the audio stream at frame positions C1, C2, C3, and C4. Each piece of sound unit tag information is normally inserted over a plurality of frames. C1, C2, C3, and C4 individually illustrate frame counts indicating frame positions from the beginning of a service.

[0069] Tag information inserted at the C1 frame position corresponds to the sound unit (speech id=1) relating to the voice of the speaker 2 (speaker_id = 2). Tag information inserted at the C2 frame position corresponds to the sound unit (speech id=2) relating to the voice of the speaker 2 (speaker_id = 2). Tag information inserted at the C3 frame position corresponds to the sound unit (speech id=3) relating to the voice of the speaker 1 (speaker_id = 1). Tag information inserted at the C4 frame position corresponds to the sound unit (speech id=4) relating to the voice of the speaker 3 (speaker_id = 3).

[0070] Table information 'Tag Table 1' having information regarding the four sound units is inserted into the audio stream at a frame position prior to the frame positions C1, C2, C3, and C4 at which the four sound units are inserted. Fig. 13 illustrates an example of details of the table information 'Tag Table 1.' In addition to switch group identification information 'SW_group_id' and group identification information 'Group_id' that are common to each of the sound units, sound unit information 'Speech_id,' originator identification information 'Speaker_id,' and frame count information 'Audio_frame_count' exist for each sound unit.

[0071] Fig. 14 illustrates a structural example (syntax) of tag information 'Speech_tag_information().' Fig. 15 illustrates a structural example (syntax) of table information 'Tag table().' Fig. 16 illustrates details (semantics) of main pieces of information in these structural examples.

[0072] A description will be given first of a structural example of tag information 'Speech_tag_information().' The 16-bit field of 'audio_frame_count' indicates the frame position of the audio frame, into which this tag information is inserted, from the beginning of the service as a frame count value. The 8-bit field of 'num_of_speeches' indicates the number of sound units. Each of the fields, the 8-bit field of 'speech_id,' the 8-bit field of 'group_id,' the 8-bit field of 'sw_group_id,' the 8-bit field of 'speaker_id,' the 2-bit field of 'sted_type,' and the 14-bit field of 'sample_offset' repeatedly exists as many times as the number of sound units.

[0073] The field of 'speech_id' indicates sound unit identification information. The field of 'group_id' indicates group identification information. The field of 'sw_group_id' indicates switch group identification information. The field of 'speaker_id' indicates identification information regarding the originator (speaker). The field of 'sted_type' indicates that a start position or an end position of a sound unit is included or that neither a start position nor an end position are included (a sound unit continues). For example, "01" indicates that a start position is included, "10" indicates that an end position is included, and "11" indicates that neither a start position nor an end position are included (a sound unit continues).

[0074] The field of 'sample_offset' indicates the sample-by-sample offset from the beginning of the frame to the sound unit start position/end position. When 'sted_type' is "01," this indicates the sample-by-sample offset from the beginning of the frame to the start position. When 'sted_type' is "10," this indicates the sample-by-sample offset from the beginning of the frame to the end position.

[0075] A description will be given next of a structural example of table information 'Tag_table().' The 8-bit field of 'group_id' indicates group identification information. The 8-bit field of 'sw_group_id' indicates switch group identification information. The 8-bit field of 'num_of_speeches' indicates the number of sound units. Each of the 8-bit field of 'speech_id,' the 8-bit field of 'speaker_id,' and the 16-bit field of 'audio_frame_count' repeatedly exists as many times as the number of sound units.

[0076] The field of 'speech_id' indicates sound unit identification information. The field of 'speaker_id' indicates identification information regarding the originator (speaker). The field of 'audio_frame_count' indicates the frame position of the audio frame, in which a sound unit start position is included, from the beginning of the service as a frame count value.

[0077] Referring back to Fig. 3, the multiplexer 116 individually packetizes a video stream output from the video encoder 112, a subtitle stream output from the subtitle encoder 114, and an audio stream output from the audio encoder 115 into a PES packet, and further into a transport packet, followed by multiplexing of the packets, thereby acquiring the transport stream TS as a container stream (multiplexed stream).

[0078] The multiplexer 116 inserts, under control of the control section 111 and into the layer of the transport stream TS, metainformation for associating subtitle data corresponding to each speech included in the subtitle stream and audio data corresponding to each speech included in the audio stream. In the present embodiment, a link table descriptor (Link_table_descriptor) is inserted into an audio elementary stream loop and/or a subtitle elementary stream loop existing under a PMT (Program Map Table).

[0079] Fig. 17 illustrates a structural example (syntax) of the link table descriptor. The 8-bit field of 'link table_descriptor_tag' indicates the descriptor type. Here, the field indicates that the descriptor type is link table descriptor. The 8-bit field of 'Link_table_descriptor_length' indicates the descriptor length (size), and the subsequent byte count is indicated as a descriptor length.

[0080] The 8-bit field of 'num_of_groups' indicates the number of groups. The following elements are repeated as many times as this number. The 8-bit field of 'stream_id' is used to identify an audio stream in the case where a plurality of streams are included. A transport stream can be identified by using this information and 'stream ID' indicated by "MPEG-H 3D Audio Multi-Stream Descriptor." The field of 'group_id' indicates group identification information. The field of 'sw_group_id' indicates switch group identification information. The 8-bit field of 'target_subtitle_language' indicates the language of the subtitle to be linked.

[0081] The 8-bit field of 'num_of_speeches' indicates the number of sound units. The 8-bit field of 'speech_id,' the 8-bit field of 'speaker_id,' and the 8-bit field of 'subtitle_speech_region_id' repeatedly exist as many times as the number of sound units. The field of 'speech_id' indicates sound unit identification information. The field of 'speaker_id' indicates identification information regarding the originator (speaker). The field of 'subtitle_speech_region_id' indicates identification information regarding the subtitle speech region.

[0082] Tag information is inserted into an audio frame including audio data corresponding to each speech of the speaker, and identification information 'speech_id' for identifying audio data is included in the tag information (refer to Fig. 18(a) and Fig. 14). Also, identification information 'subtitle_speech_region_id' for identifying the subtitle data is included in subtitle data corresponding to each speech of the speaker (refer to Fig. 18(b) and Fig. 5(c)). Therefore, as 'speech_id' and 'subtitle_speech_region_id' are associated by a link table descriptor (refer to Fig. 18(c)), subtitle data corresponding to each speech included in a subtitle stream and audio data corresponding to each speech included in an audio stream are associated.

[0083] The operation of the stream generation section 110 illustrated in Fig. 3 will be described briefly. The video data SV is supplied to the video encoder 112. The video data SV is coded by this video encoder 112, generating a video stream including coded video data. This video stream is supplied to the multiplexer 116.

[0084] Also, text data (character codes) DT as subtitle information is supplied to the text format conversion section 113. TTML data is acquired by this text format conversion section 113 as text information. In this case, identification information 'subtitle_speech_region_id' of the subtitle speech region is included in the subtitle data corresponding to each speech of the speaker (refer to Fig. 5(c)).

[0085] TTML data acquired by the text format conversion section 113 is supplied to the subtitle encoder 114. The TTML data is converted by the subtitle encoder 114 into a variety of segments, generating a subtitle stream included in a PES packet having these segments provided in a payload. This subtitle stream is supplied to the multiplexer 116.

[0086] Also, the audio data SA is supplied to the audio encoder 115. Object data of a given number of content groups is included in this audio data SA together with channel data. Here, one or a plurality of pieces of object content belong to each content group.

[0087] The audio encoder 115 acquires 3D Audio transport data by coding the audio data SA. Object coded data of a given number of content groups is included in this 3D Audio transport data together with channel coded data. Then, an audio stream including this 3D Audio transport data is generated by the audio encoder 115.

[0088] In this case, tag information is inserted into an audio frame including audio data corresponding to each speech of the speaker. The tag information has identification information for identifying the audio data and includes 'speech_id' (refer to Fig. 14). Also, table information 'Tag_table()' having sound unit information in the audio stream is inserted into an arbitrary audio frame by the audio encoder 115 under control of the control section 111.

[0089] The video stream generated by the video encoder 112 is supplied to the multiplexer 116. Also, the subtitle stream generated by the subtitle encoder 114 is supplied to the multiplexer 116. Further, the audio stream generated by the audio encoder 115 is supplied to the multiplexer 116. The stream supplied from each encoder is packetized into a PES packet, and further into a transport packet by the multiplexer 116, followed by multiplexing, allowing the transport stream TS to be acquired.

[0090] At this time, metainformation is inserted into the layer of the transport stream TS by the multiplexer 116. The metainformation is used to associate subtitle data corresponding to each speech included in the subtitle stream and audio data corresponding to each speech included in the audio stream. Specifically, a link table descriptor (refer to Fig. 17) is inserted. This descriptor has description for associating 'speech_id' and 'subtitle_speech_region_id.'

[Configuration of the Transport Stream TS]



[0091] Fig. 19 illustrates a configuration example of the transport stream TS. In this configuration example, a PES packet 'video PES' of the video stream identified by PID1, a PES packet 'audio PES' of the audio stream identified by PID2, and a PES packet 'subtitle PES' of the subtitle stream identified by PID3, exist.

[0092] An audio stream (Audio coded stream) is inserted into a PES payload of the PES packet of the audio stream. In this audio stream, tag information 'Speech_tag_information()' is inserted into an audio frame including audio data corresponding to each speech of the speaker. Table information 'Tag_table()' is also inserted into the audio frame. Identification information 'speech_id' for identifying the audio data is included in the tag information (refer to Fig. 14).

[0093] A subtitle stream (Subtitle coded stream) is inserted into a PES payload of the PES packet of the subtitle stream. Identification information 'subtitle_speech_region_id' for identifying the subtitle data is included in the subtitle information corresponding to each speech of the speaker (refer to Fig. 5(c)).

[0094] Also, a PMT (Program Map Table) is included in the transport stream TS as PSI (Program Specific Information). PSI is information stating to which program each elementary stream included in the transport stream belongs. A program loop that describes information relating to the program as a whole exists in the PMT.

[0095] Also, elementary stream loops, each having information relating to each elementary stream, exist in the PMT. In this configuration example, a video elementary stream loop (video ES loop) corresponding to the video stream, an audio elementary stream loop (audio ES loop) corresponding to the audio stream, and a subtitle elementary stream loop (subtitle ES loop) corresponding to the subtitle stream exist.

[0096] In the video elementary stream loop (video ES loop), information such as stream type and PID (packet identifier) is provided to correspond to the video stream, and a descriptor is also provided that states information relating to the video stream. The value of 'Stream_type' of the video stream is set to '0×24,' and PID information indicates PID1 assigned to the PES packet 'video PES' of the video stream as described above. An HEVC descriptor is provided as a descriptor.

[0097] Also, in the audio elementary stream loop (audio ES loop), information such as stream type and PID (packet identifier) is provided to correspond to the audio stream, and a descriptor is also provided that states information relating to the audio stream. The value of 'Stream_type' of the audio stream is set to '0×2C,' and PID information indicates PID2 assigned to the PES packet 'audio PES' of the audio stream as described above. A link table descriptor (refer to Fig. 17) is provided as a descriptor. A link table descriptor is used to indicate correspondence between subtitle data corresponding to each speech included in the subtitle stream and audio data corresponding to each speech included in the audio stream by associating 'speech_id' and 'subtitle_speech_region_id' (refer to Fig. 17).

[0098] Also, information such as stream type and PID (packet identifier) is provided in the subtitle elementary stream loop (subtitle ES loop) to correspond to the subtitle stream, and a descriptor for describing information relating to the audio stream thereof is also provided. 'Stream_type' of this subtitle stream is "private stream," and PID information indicates PID3 assigned to the PES packet 'audio PES' of the video stream as described above. A link table descriptor (refer to Fig. 17) is provided as a descriptor.

[0099] It should be noted that although provided in both the audio elementary stream loop and the subtitle elementary stream loop in the example illustrated, a link table descriptor may be provided only in one of the two loops.

[Configuration Example of the Service Receiver]



[0100] Fig. 20 illustrates a configuration example of the service receiver 200. The service receiver 200 has a reception section 201, a demultiplexer 202, a video decoding section 203, a video superimposition section 204, a panel drive circuit 205, and a display panel 206. The service receiver 200 also has an audio decoding section 214, a voice output processing circuit 215, a speaker system 216, a subtitle decoding section 217, a display coordinates mapping section 218, and a superimposition position correction section 219. The service receiver 200 also has a CPU 221, a flash ROM 222, a DRAM 223, an internal bus 224, a remote control reception section 225, and a remote control transmitter 226.

[0101] The CPU 221 controls the operation of each section of the service receiver 200. The flash ROM 222 stores control software and retains data. The DRAM 223 is included in a work area of the CPU 221. The CPU 221 controls each section of the service receiver 200 by loading the software and data read out from the flash ROM 222 into the DRAM 223 and starting the software.

[0102] The remote control reception section 225 receives a remote control signal (remote control code) sent from the remote control transmitter 226, supplying the signal to the CPU 221. The CPU 221 controls each section of the service receiver 200 on the basis of this remote control code. The CPU 221, the flash ROM 222, and the DRAM 223 are connected to the internal bus 224.

[0103] The reception section 201 receives the transport stream TS as a container stream (multiplexed stream) sent over broadcast waves or in a network packet from the service transmitter 100. This transport stream TS has a subtitle stream and an audio stream in addition to a video stream. The subtitle stream has subtitle data corresponding to a speech of a speaker. The audio stream has object coded data including audio data whose sound source is a speech of a speaker and position information regarding the sound source.

[0104] In the audio stream, tag information 'Speech_tag_information()' is inserted into an audio frame including audio data corresponding to each speech of the speaker, and table information 'Tag_table()' is inserted. Identification information 'speech_id' for identifying the audio data is included in the tag information (refer to Fig. 14). Also, identification information 'subtitle_speech_region_id' for identifying the subtitle data is included in subtitle data corresponding to each speech of the speaker (refer to Fig. 5(c)).

[0105] Also, a link table descriptor is provided in the layer of the transport stream TS. A link table descriptor is used to indicate correspondence between subtitle data corresponding to each speech included in the subtitle stream and audio data corresponding to each speech included in the audio stream by associating 'speech_id' and 'subtitle_speech_region_id' (refer to Fig. 17).

[0106] The demultiplexer 202 extracts the video stream, the subtitle stream, and the audio stream from the transport stream TS. The demultiplexer 202 also extracts various pieces of information such as descriptor information from the transport stream TS and sends the information to the CPU 221. The link table descriptor described above is included in these various pieces of information. The CPU 221 associates 'speech_id' and 'subtitle_speech_region_id' using this descriptor and therefore recognizes correspondence between subtitle data corresponding to each speech included in the subtitle stream and audio data corresponding to each speech included in the audio stream.

[0107] The video decoding section 203 acquires non-compressed video data through a decoding process of the video stream extracted by the demultiplexer 202. The video superimposition section 204 superimposes subtitle bitmap data (display data) on the video data acquired by the video decoding section 203. The panel drive circuit 205 drives the display panel 206 on the basis of image data for display acquired by the video superimposition section 204. The display panel 206 includes, for example, an LCD (Liquid Crystal Display), an organic EL display (organic electroluminescence display), or other type of display.

[0108] The audio decoding section 214 acquires audio data for driving each of the speakers included in the speaker system 216 through a decoding process of the audio stream extracted by the demultiplexer 202.

[0109] In this case, as for coded data of a plurality of pieces of object content included in the switch group, the audio decoding section 214 decodes, under control of the CPU 221 and of coded data of a given pieces of object content included in the audio stream, coded data of only one of the pieces of object content relating to user selection.

[0110] Also, the audio decoding section 214 extracts a variety of pieces of information inserted in the audio stream and sends them to the CPU 221. Tag information 'Speech_tag_information()' and table information 'Tag_table()' described above are also included in these pieces of information.

[0111] The CPU 221 recognizes, using this tag information and for each audio frame, whether the audio frame includes audio data of a sound unit corresponding to the speaker's speech (coded sample data), sound unit identification information regarding the sound unit, originator identification information regarding the sound unit, frame count information, type information indicating whether a start position or an end position is included, offset information from the beginning of the frame to the start position or the end position, and so on.

[0112] Also, the CPU 221 recognizes, using this table information, information regarding a given sound unit within the audio stream, that is, frame count information indicating the frame position of each sound unit from the beginning of the service, sound unit identification information regarding each sound unit, originator identification information regarding each sound unit, and so on.

[0113] Fig. 21 illustrates a configuration example of the audio decoding section 214. The audio decoding section 214 has a decoder 251, an object renderer 252, and a mixer 253.

[0114] The decoder 251 acquires object data of a given number of pieces of object content together with channel data through a decoding process of the object stream extracted by the demultiplexer 202. This decoder 251 performs a process nearly opposite to that performed by the audio encoder 115 of the stream generation section 110 illustrated in Fig. 3. It should be noted that, as for a plurality of pieces of object content included in the switch group, the decoder 251 acquires, under control of the CPU 221, object data of only one piece of object content relating to user selection.

[0115] Also, the decoder 251 extracts various pieces of information inserted in the audio stream and sends them to the CPU 221. Tag information 'Speech_tag_information()' and table information 'Tag_table()' described above are also included in these pieces of information.

[0116] The object renderer 252 acquires channel data of a given number of pieces of object content through a rendering process of object data of the given number of pieces of object content (audio data, position information).

[0117] Fig. 22 illustrates an overview of a rendering process in the object renderer 252. In this rendering process, a sound image position p is found by using position information (θ, φ, r) included in object data in each of triangular areas (LT-LB-RB) and (RT-LB-RB) in which three speakers are included, and channel data is acquired by mapping audio data onto each speaker from the sound image position p. It should be noted that the number and layout of speakers are not limited to those in the example illustrated.

[0118] This rendering process is also similarly performed on object data relating to audio data whose sound source is the speaker's speech. Because of channel data acquired by such a rendering process, the speaker's speech is recognized by a viewer as if the speech were heard from the sound image position p, and therefore, from the speaker's direction in a video image (refer to Fig. 2).

[0119] The mixer 253 acquires final channel data for driving each of the speakers included in the speaker system 216 (audio data for driving each speaker) by combining channel data acquired by the decoder 251 and channel data of each piece of object content acquired by the object renderer 252.

[0120] Referring back to Fig. 20, the voice output processing circuit 215 performs various processes such as D/A conversion and amplification on audio data for driving each speaker acquired by the audio decoding section 214, supplying the data to the speaker system 216. The speaker system 216 includes a plurality of speakers with a plurality of channels such as 2 channels, 5.1 channels, 7.1 channels, and 22.2 channels.

[0121] The subtitle decoding section 217 acquires bitmap data of a subtitle corresponding to the speaker's speech by processing subtitle data of each region included in the subtitle stream extracted by the demultiplexer 202.

[0122] The display coordinates mapping section 218 finds, in relation to the bitmap data of the subtitle acquired by the subtitle decoding section 217 described above, display coordinates (x1, y1) on the monitor screen of the subtitle using, of position information (θ, φ, r) in the object data relating to the audio data corresponding to this subtitle, information regarding the azimuth θ and the elevation φ. Here, correspondence between the subtitle and the audio data is established, for example, on the basis of the description regarding association between 'speech_id' and 'subtitle_speech_region_id' in the of the link table descriptor.

[0123] A description will be given of an example of calculating x1 and y1 at display coordinates (x1, y1) with reference to Fig. 23. Here, we assume that the monitor is sized MW (horizontally) and MH (vertically) and that the monitor viewing distance is D. A point C is taken such that AC = AB/2 = MW/2.

[0124] At this time, each of the following equations holds:





[0125] Also, each of the following equations holds:







[0126] Further, each of the following equations holds:







[0127] For this reason, letting the horizontal display resolution of the monitor be denoted by W (pixels) and the vertical display resolution of the monitor be denoted by H (pixels), x1 and y1 at the display coordinates (x1, y1) equivalent to the sound image position p can be found by the following equations (1) and (2):





[0128] It should be noted that although, in the above calculation example, a direction A is a reference direction, the display coordinates (x1, y1) can be found in a similar manner even in the case where other direction is a reference direction. For example, in the case where a direction C is a reference direction, the above θ is replaced with γ-θ in the equations. Fig. 23 illustrates, with a dashed line, an example in which the subtitle display position is taken at point P thus found at the display coordinates (x1, y1). It should be noted that various options, including whether to take P at the end position or at the center of the region area or whether to keep it within the monitor display area, can be taken.

[0129] Referring back to Fig. 20, the superimposition position correction section 219 performs a position conversion process on the bitmap data of the subtitle corresponding to the speaker's speech acquired by the subtitle decoding section 217 such that the superimposition on the video data takes place at the position corresponding to the display coordinates (x1, y1) found by the display coordinates mapping section 218, supplying the data to the video superimposition section 204 thereafter. Such a position conversion process allows the subtitle corresponding to the speaker's speech to be displayed at the speaker's position in the video image (refer to Fig. 2).

[0130] A brief description will be given of the operation of the service receiver 200 illustrated in Fig. 20. The service receiver 200 receives the transport stream TS sent from the service transmitter 100 over broadcast waves or in a network packet. This transport stream TS includes a subtitle stream and an audio stream in addition to a video stream as described above.

[0131] The transport stream TS received by the reception section 201 is supplied to the demultiplexer 202. The demultiplexer 202 extracts the video stream, the subtitle stream, and the audio stream from the transport stream TS.

[0132] Also, the demultiplexer 202 extracts various pieces of information such as descriptor information from the transport stream TS and sends them to the CPU 221. The link table descriptor described above is included in these various pieces of information. The CPU 221 associates 'speech_id' and 'subtitle_speech_region_id' using this descriptor and therefore recognizes correspondence between subtitle data corresponding to each speech included in the subtitle stream and audio data corresponding to each speech included in the audio stream.

[0133] The video stream extracted from the transport stream TS is supplied to the video decoding section 203. The video decoding section 203 acquires non-compressed video data through a decoding process of the video stream. This video data is supplied to the video superimposition section 204.

[0134] Also, the audio stream extracted by the demultiplexer 202 is supplied to the audio decoding section 214. The audio decoding section 214 acquires audio data for driving each of the speakers included in the speaker system 216 through a decoding process of the audio stream.

[0135]  Also, the audio decoding section 214 extracts a variety of pieces of information inserted in the audio stream and sends them to the CPU 221. Tag information 'Speech_tag_information(), and table information 'Tag_table (), are also included in these pieces of information.

[0136] The CPU 221 recognizes, using tag information and for each audio frame, whether the audio frame includes audio compressed data of a given sound unit, sound unit identification information regarding the sound unit, originator identification information regarding the sound unit, frame count information, type information indicating whether a start position or an end position is included, offset information from the beginning of the frame to the start position or end position, and so on.

[0137] Also, the CPU 221 recognizes, using table information, information regarding a given sound unit within the audio stream, that is, frame count information indicating the frame position of each sound unit from the beginning of the service, sound unit identification information regarding each sound unit, originator identification information regarding each sound unit, and so on.

[0138] Audio data for driving each of the speakers acquired by the audio decoding section 214 is supplied to the voice output processing circuit 215. The voice output processing circuit 215 performs necessary processes such as D/A conversion and amplification on this audio data.

[0139] The audio data after being processed by the voice output processing circuit 215 is supplied to the speaker system 216. As a result, acoustic output corresponding to the display image on the display panel 206 is acquired from the speaker system 216. In this case, as a result of the rendering process by the object renderer 252 described above, the speaker's speech becomes recognized by the viewer as if the speech were heard from the direction of the speaker in the video image (refer to Fig. 2) .

[0140] Also, the subtitle stream extracted by the demultiplexer 202 is supplied to the subtitle decoding section 217. The subtitle decoding section 217 acquires bitmap data of the subtitle corresponding to the speaker's speech through processing of the subtitle data of each region included in the subtitle stream extracted by the demultiplexer 202. This bitmap data is supplied to the superimposition position correction section 219.

[0141] Also, of position information (θ, φ, r) in the object data relating to the audio data corresponding to this subtitle, information regarding the azimuth θ and the elevation φ is supplied from the audio decoding section 214 to the display coordinates mapping section 218. Here, correspondence between the subtitle and the audio data is established, for example, on the basis of the description regarding association between 'speech_id' and 'subtitle_speech_region_id' in the link table descriptor.

[0142] The display coordinates mapping section 218 finds, in relation to the bitmap data of the subtitle acquired by the subtitle decoding section 217, display coordinates (x1, y1) on the monitor screen of the subtitle using information regarding the azimuth θ and the elevation φ. These display coordinates (x1, y1) are supplied to the superimposition position correction section 219.

[0143] The superimposition position correction section 219 performs a position conversion process on the bitmap data of the subtitle corresponding to the speaker's speech acquired by the subtitle decoding section 217 such that the superimposition takes place at the position corresponding to the display coordinates (x1, y1). The bitmap data of the subtitle after the position conversion is supplied to the video superimposition section 204.

[0144] The video superimposition section 204 acquires video data for display through superimposition of the subtitle bitmap data, subjected to the position conversion process by the superimposition position correction section 219, on the video data acquired by the video decoding section 203. This video data for display is supplied to the panel drive circuit 205.

[0145] The panel drive circuit 205 drives the display panel 206 on the basis of the video data for display. This allows a video image corresponding to the video data for display to be displayed on the display panel 206. In this case, as a result of the position conversion process by the superimposition position correction section 219, the subtitle corresponding to the speaker's speech is displayed at the speaker's position in the video image (refer to Fig. 2).

[0146] As described above, in the transmission/reception system 10 illustrated in Fig. 1, the service transmitter 100 inserts, into the layer of the transport stream TS, metainformation for associating subtitle data corresponding to each speech included in the subtitle stream and audio data corresponding to each speech included in the audio stream. For this reason, in the case where the subtitle display position is controlled on the receiving side using position information regarding the sound source included in object coded data such that the subtitle is displayed at the speaker's position in the video image, it becomes easy to use appropriate position information by using this metainformation.

[0147] Also, in the transmission/reception system 10 illustrated in Fig. 1, the service receiver 200 finds the display coordinates of the subtitle corresponding to each of the speeches on the monitor screen on the basis of position information regarding the sound source corresponding to each of the speeches and performs control such that the bitmap data of the subtitle corresponding to each of the speeches is superimposed on the video data at the position corresponding to the display coordinates found. For this reason, it is possible to display a subtitle corresponding to a speech of a speaker at the speaker's position in a video image.

<2. Modification Example>



[0148] It should be noted that an example of single region display has described above in which one speaker existed in a video image and in which a subtitle corresponding to the speaker is superimposed on the video image (refer to Fig. 2). However, the present technology functions similarly in multi-region display in which a plurality of speakers exist in a video image and in which subtitles, each corresponding to a speech of each of the speakers, are superimposed on the video image.

[0149] That is, because 'speech_id' and 'subtitle_speech_region_id' are associated with each other, it is possible to display a subtitle corresponding to a speech of each speaker at the corresponding speaker's position by individually finding the display coordinates of the subtitle corresponding to the speech of each of the speakers. In the example illustrated, subtitles corresponding to speeches of speakers in the video image depicted by arrows A1, A2, and A3 appear at positions of dotted line rectangular frames depicted by arrows B1, B2, and B3, respectively.

[0150] Also, the embodiment described above has illustrated an example in which a container stream (multiplexed stream) is an MPEG-2 transport stream (transport stream TS). However, the present technology is similarly applicable to a system that makes deliveries using container streams in MP4 format or in other formats. Among examples thereof are an MPEG-DASH-based stream delivery system and a transmission/reception system that handles MMT (MPEG Media Transport) structure transport stream.

[0151] Fig. 25 illustrates a configuration example of an MMT stream in the case where media access information (data to be encapsulated into a container) is sent inserted in an audio stream. MMT packets of respective assets such as video, audio, and subtitle exist. In this configuration example, an MMT packet for an audio asset identified by ID2 and an MMT packet for a subtitle asset identified by ID3 exist together with an MMT packet for a video asset identified by ID1.

[0152] In the audio asset (audio stream), tag information 'Speech_tag_information(), is inserted into an audio frame that includes audio data corresponding to each speech of the speaker, and table information 'Tag_table (), is inserted. Identification information 'speech_id' for identifying audio data is included in the tag information (refer to Fig. 14). In the subtitle asset (subtitle stream), identification information 'subtitle_speech_region_id' for identifying subtitle data is included in the subtitle data corresponding to each speech of the speaker (refer to Fig. 5(c)).

[0153] Also, message packets such as PA (Packet Access) message packet exist in the MMT stream. The PA message packet includes tables such as MMT packet table. The MP table includes information for each asset. In this case, a link table descriptor (refer to Fig. 17) is also included as asset information regarding an audio and/or subtitle asset.

[0154] Fig. 26(a) illustrates a configuration example of an MP4 stream (file) including audio track (track A) data. Fig. 26(b) illustrates a configuration example of an MP4 stream (file) including subtitle track (track B) data. The example illustrated is an example in the case of a fragmented MP4. A given number of movie fragments are provided in an MP4 stream. A "moof" box and a "mdat" box are included in a movie fragment. Control information is included in the "moof" box. The main body of media data is included in the "mdat" box. Because a fragment acquired by fragmentation of track data is placed in the "mdat" box, control information placed in the "moof" box is control information regarding these fragments.

[0155] In the MP4 stream 'audio bitstream' corresponding to the audio track illustrated in Fig. 26(a), a given number of audio frames are provided in the "mdat" box of each movie fragment. Also, in this MP4 stream 'audio bitstream,' a "traf" box exists in the "moof" box of each movie fragment, and a "tfdt" box exists in the "traf" box. Decoding time "baseMediaDecodeTime" of the first access unit after the "moof" box is written in this "tfdt" box.

[0156] Also, a "tfdt" box exists in the "moof" box, and "sgpd" box exists in that "tfdt" box, and further, a "tscl" box exists in the "sgpd" box. 'Audiostreamtype' and 'Attribute' parameters are written in this "tscl" box. 'Audiostreamtype = AudioCodec' indicates an audio compressed format. 'Attribute = speechtag' indicates that tag information 'Speech_tag_information(), is inserted in the audio stream.

[0157] In an MP4 stream 'subtitle bitstream' corresponding to the subtitle track illustrated in Fig. 26(b), a given number of TTML documents are provided in the "mdat" box of each of the movie fragments. Also, in this MP4 stream 'subtitle bitstream,' a "traf" box exists in the "moof" box of each of the movie fragments, and a "tfdt" box exists in the "traf" box. Presentation time "baseMediaPresentationTime" of the first subtitle after the "moof" box is written in this "tfdt" box.

[0158] Further, a "tfdt" box exists in the "moof" box, and a "sgpd" box exists in that "tfdt" box, and further, a "tscl" box exists in the "sgpd" box. 'subtitlestreamtype' and 'Attribute' parameters are written in this "tscl" box. 'subtitlestreamtype = TTML' indicates that the subtitle data format is TTML. 'Attribute = speechregionid' indicates that speech region identification information 'subtitle_speech_region_id' is inserted in the subtitle data.

[0159] Fig. 27 illustrates an example of describing an MPD file. Also, Fig. 28 illustrates details of main pieces of information in the example of describing an MPD file. As has been previously known, a media stream (MP4 stream) and an MPD file as a metafile are sent to the receiving side via a communication network transport channel in an MPEG-DASH-based stream delivery system.

[0160] '<AdaptationSet mimeType= "audio/mp4" group= "1">' is written to indicate that an adaptation set (AdaptationSet) exists for an audio stream, that an audio stream in MP4 file structure is supplied, and that group 1 is assigned.

[0161] Also, in this adaptation set (AdaptationSet) for the audio stream, '<SupplementaryDescriptor schemeIdUri= "urn:brdcst:codecType" value= "Audio Codec Stream"/>' is written to indicate an audio stream codec type. MPEG-H, AAC, AC3, AC4, and so on are available as codecs.

[0162] Also, in this adaptation set (AdaptationSet) for the audio stream, '<SupplementaryDescriptor schemeIdUri= "urn:brdcst:speechtag" value= "true"/>' is written to indicate that tag information 'Speech_tag_information(), is included.

[0163] '<AdaptiveSet mimeType= "subtitle/mp4" group= "2">' is written to indicate that an adaptation set (AdaptationSet) exists for a subtitle stream, that a subtitle stream in MP4 file structure is supplied, and that group 2 is assigned.

[0164] Also, in this adaptation set (AdaptationSet) for the subtitle stream, '<SupplementaryDescriptor schemeIdUri= "urn:brdcst:codecType" value= "Subtitle Codec Stream"/>' is written to indicate a subtitle stream codec type. TTML and other codecs are available.

[0165] Also, in this adaptation set (AdaptationSet) for the subtitle stream, '<SupplementaryDescriptor schemeIdUri= "urn:brdcst:speechregionid" value= "true"/>' is written to indicate that speech region identification information 'subtitle_speech_region_id' is included.

[0166] Also, in the embodiment described above, an example has illustrated in which the audio compression format is MPEG-H 3D Audio. However, the present technology is similarly applicable to a case in which other audio compression format such as AAC, AC3, and AC4 is used.

[0167] Fig. 29(a) illustrates a structure of an AC4 Simple Transport layer. A syncWord field, a frame Length field, a 'RawAc4Frame' field as a coded data field, and a CRC field exist. In the 'RawAc4Frame' field, a TOC (Table Of Content) field exists at the beginning, followed by a given number of Substream fields as illustrated in Fig. 29(b) .

[0168] As illustrated in Fig. 30(b), a metadata area exists in the substream (ac4_substream_data()), and a 'umd_payloads_substream()' field is provided therein. Tag information 'Speech_tag_information(), (refer to Fig. 14) and table information 'Tag_table (), (refer to Fig. 15) are placed in an 'umd_payload_byte' field inside this 'umd_payloads_substream()' field.

[0169] It should be noted that, as illustrated in Fig. 30(a), it is indicated that an 'ac4_presentation_info()' field exists in the TOC (ac4_toc()), that further an 'umd_info()' field exists therein, and that tag information and so on is inserted in the above 'umd_payloads_substream()' field in the 'umd_info()' field.

[0170] Also, in the embodiment described above, an example has illustrated in which a TTML is used as text information regarding subtitles in a given format having display timing information. However, the present technology is not limited thereto, and it is possible to use other text information having information equivalent to TTML. For example, a TTML-derived format may be used.

[0171] Also, although, in the embodiment described above, the transmission/reception system 10 including the service transmitter 100 and the service receiver 200 has illustrated, the transmission/reception system to which the present technology is applicable is not limited thereto. For example, the portion of the service receiver 200 may include a set top box and a monitor that are connected through a digital interface such as HDMI (High-Definition Multimedia Interface). It should be noted that 'HDMI' is a registered trademark.

[0172] A main feature of the present technology lies in facilitating use of appropriate position information through use of metainformation by inserting, into a layer of a container stream, the metainformation for associating subtitle data corresponding to each speech included in a subtitle stream and audio data corresponding to each speech included in an audio stream during control performed on a receiving side to display a subtitle at a speaker's position in a video image through use of position information of a sound source included in object coded data (refer to Figs. 18 and 19).

[0173] Also, a main feature of the present technology lies in allowing a subtitle corresponding to a speech of a speaker to be displayed at the speaker's position by finding display coordinates of a subtitle corresponding to each speech on a monitor screen on the basis of position information regarding the sound source corresponding to each of the speeches and performing control such that bitmap data of the subtitle corresponding to each of the speeches is superimposed on the video data at the position corresponding to the display coordinates found (refer to Figs. 2 and 20).

[Reference Signs List]



[0174] 
10
Transmission/reception system
100
Service transmitter
110
Stream generation section
111
Control section
111a
CPU
112
Video encoder
113
Text format conversion section
114
Subtitle encoder
115
Audio encoder
116
Multiplexer
200
Service receiver
201
Reception section
202
Demultiplexer
203
Video decoding section
204
Video superimposition section
205
Panel drive circuit
206
Display panel
214
Audio decoding section
215
Voice output processing circuit
216
Speaker system
217
Subtitle decoding section
218
Display coordinates mapping section
219
Superimposition position correction section
221
CPU
222
Flash ROM
223
DRAM
224
Internal bus
225
Remote control reception section
226
Remote control transmitter
251
Decoder
252
Object renderer
253
Mixer



Claims

1. A transmission apparatus (100) comprising:

a video encoding section (112) adapted to generate a video stream having video coded data;

a subtitle encoding section (114) adapted to generate a subtitle stream having subtitle data corresponding to a speech of a speaker;

an audio encoding section (115) adapted to generate an audio stream having object coded data including audio data whose sound source is a speech of the speaker and position information regarding a position of the sound source in an image of the video stream;

a transmission section adapted to send a container stream in a given format including the video stream, the subtitle stream, and the audio stream; and

an information insertion section (111, 116) adapted to insert, into a layer of the container stream, metainformation for associating subtitle data corresponding to each speech included in the subtitle stream with audio data corresponding to each speech included in the audio stream and with the position information associated with said audio data.


 
2. The transmission apparatus of claim 1, wherein
identification information is added to each piece of subtitle data corresponding to each of the speeches, and identification information is added to each piece of audio data corresponding to each of the speeches, and
the metainformation is table information that associates the identification information added to the subtitle data corresponding to each of the speeches and the identification information added to the audio data corresponding to each of the speeches.
 
3. The transmission apparatus (100) of claim 2, wherein
tag information having the identification information is inserted into an audio frame including audio data corresponding to each of the speeches.
 
4. The transmission apparatus (100) of claim 1, wherein
the subtitle data is TTML data or data in a format derived from the TTML.
 
5. The transmission apparatus (100) of claim 1, wherein the container stream is a transport stream (TS), wherein the metainformation is inserted into the layer of the transport stream (TS) by a multiplexer (116) which also multiplexes the audio, video, and subtitle streams.
 
6. A transmission method comprising:

a video encoding (112) step of generating a video stream that has video coded data;

a subtitle encoding (114) step of generating a subtitle stream that has subtitle data corresponding to a speech of a speaker;

an audio encoding (115) step of generating an audio stream that has object coded data including audio data whose sound source is a speech of the speaker and position information regarding a position of the sound source in an image of the video stream;

a transmission step of sending a container stream in a given format that includes the video stream, the subtitle stream, and the audio stream by a transmission section; and

an information insertion (111, 116) step of inserting metainformation, into a layer of the container stream, that associates subtitle data corresponding to each speech included in the subtitle stream with audio data corresponding to each speech included in the audio stream and with the position information associated with said audio data.


 
7.  A reception apparatus (200) comprising:

a reception section (201) adapted to receive a container stream in a given format including a video stream, a subtitle stream, an audio stream, and metainformation wherein

the video stream has video coded data,

the subtitle stream has subtitle data corresponding to a speech of a speaker,

the audio stream has object coded data including audio data whose sound source is the speech of the speaker and position information regarding a position of the sound source in an image of the video stream, and the metainformation associates subtitle data corresponding to each speech included in the subtitle stream with audio data corresponding to each speech included in the audio stream and with the position information associated with said audio data,

the reception apparatus further comprising:
a control section adapted to control

a video decoding (203) process whereby video data is acquired through a decoding process of the video stream,

a subtitle decoding (217) process whereby bitmap data of a subtitle corresponding to the speech of the speaker is acquired through a decoding process of the subtitle stream,

a video superimposition (204) process whereby video data for display is acquired through superimposition of the bitmap data of the subtitle on the video data,

an audio decoding (214) process whereby channel data corresponding to the speech is acquired through a decoding process of the audio stream, and

a superimposition position process whereby display coordinates of a subtitle corresponding to each of the speeches on a monitor screen are found on the basis of position information regarding the sound source corresponding to each of the speeches as indicated in metainformation inserted in the container stream and control is performed such that the bitmap data of the subtitle corresponding to each of the speeches is superimposed on the video data at the position corresponding to the display coordinates found.


 
8. A reception method comprising:

a reception (201) step of receiving a container stream in a given format including a video stream, a subtitle stream, an audio stream by a reception section, and metainformation, wherein

the video stream has video coded data,

the subtitle stream has subtitle data corresponding to a speech of a speaker,

the audio stream has object coded data including audio data whose sound source is the speech of the speaker and position information regarding a position of the sound source in an image of the video stream, and the metainformation associates subtitle data corresponding to each speech included in the subtitle stream with audio data corresponding to each speech included in the audio stream and with the position information associated with said audio data,

the reception method further comprising

a video decoding (203) step of acquiring video data through a decoding process of the video stream;

a subtitle decoding (217) step of acquiring bitmap data of a subtitle corresponding to the speech of the speaker through a decoding process of the subtitle stream;

a video superimposition (204) step of acquiring video data for display through superimposition of the bitmap data of the subtitle on the video data;

an audio decoding (214) step of acquiring channel data corresponding to the speech through a decoding process of the audio stream; and

a control step of finding display coordinates of a subtitle corresponding to each of the speeches on a monitor screen on the basis of position information regarding the sound source corresponding to each of the speeches as indicated in metainformation inserted in the container stream and controlling such that the bitmap data of the subtitle corresponding to each of the speeches is superimposed on the video data at the position corresponding to the display coordinates found.


 


Ansprüche

1. Sendevorrichtung (100), Folgendes umfassend:

einen Videocodierabschnitt (112), der eingerichtet ist, um einen Videostrom mit videocodierten Daten zu erzeugen;

einen Untertitelcodierabschnitt (114), der eingerichtet ist, um einen Untertitelstrom mit Untertiteldaten zu erzeugen, die einer Rede eines Sprechers entsprechen;

einen Audiocodierabschnitt (115), der eingerichtet ist, um einen Audiostrom mit objektcodierten Daten zu erzeugen, die Audiodaten, deren Schallquelle eine Rede des Sprechers ist, und Positionsinformationen hinsichtlich einer Position der Schallquelle in einem Bild des Videostroms umfassen;

einen Sendeabschnitt, der eingerichtet ist, um einen Containerstrom in einem gegebenen Format zu senden, das den Videostrom, den Untertitelstrom und den Audiostrom umfasst; und

einen Informationeneinfügungsabschnitt (111, 116), der eingerichtet ist, um Metainformationen zum Zuordnen von Untertiteldaten, die jeder Rede entsprechen, die in dem Untertitelstrom enthalten ist, zu Audiodaten, die jeder Rede entsprechen, die in dem Audiostrom enthalten ist, und zu den Positionsinformationen, die den Audiodaten zugeordnet sind, in eine Schicht des Containerstroms einzufügen.


 
2. Sendevorrichtung nach Anspruch 1, wobei
Identifikationsinformationen zu jedem Stück Untertiteldaten hinzugefügt werden, die jeder der Reden entsprechen, und Identifikationsinformationen zu jedem Stück Audiodaten hinzugefügt werden, die jeder der Reden entsprechen, und
die Metainformationen Tabelleninformationen sind, welche die Identifikationsinformationen, die zu den Untertiteldaten hinzugefügt wurden, die jeder der Reden entsprechen, und die Identifikationsinformationen, die zu den Audiodaten hinzugefügt wurden, die jeder der Reden entsprechen, zuordnen.
 
3. Sendevorrichtung (100) nach Anspruch 2, wobei
Kennungsinformationen mit den Identifikationsinformationen in einen Audio-Frame eingefügt werden, der Audiodaten umfasst, die jeder der Reden entsprechen.
 
4. Sendevorrichtung (100) nach Anspruch 1, wobei
die Untertiteldaten TTML-Daten oder Daten in einem Format sind, das aus dem TTML abgeleitet ist.
 
5. Sendevorrichtung (100) nach Anspruch 1, wobei der Containerstrom ein Transportstrom (TS) ist, wobei die Metainformationen durch einen Multiplexer (116) in die Schicht des Transportstroms (TS) eingefügt werden, der auch die Audio-, Video- und Untertitelströme multiplexiert.
 
6. Sendeverfahren, Folgendes umfassend:

einen Videocodierschritt (112) des Erzeugens eines Videostroms, der videocodierte Daten aufweist;

einen Untertitelcodierschritt (114) des Erzeugens eines Untertitelstroms, der Untertiteldaten aufweist, die einer Rede eines Sprechers entsprechen;

einen Audiocodierschritt (115) des Erzeugens eines Audiostroms, der objektcodierte Daten aufweist, die Audiodaten, deren Schallquelle eine Rede des Sprechers ist, und Positionsinformationen hinsichtlich einer Position der Schallquelle in einem Bild des Videostroms umfassen;

einen Sendeschritt des Sendens eines Containerstroms in einem gegebenen Format, das den Videostrom, den Untertitelstrom und den Audiostrom umfasst, durch einen Sendeabschnitt; und

einen Informationeneinfügungsschritt (111, 116) des Einfügens von Metainformationen in eine Schicht des Containerstroms, die Untertiteldaten, die jeder Rede entsprechen, die in dem Untertitelstrom enthalten ist, zu Audiodaten, die jeder Rede entsprechen, die in dem Audiostrom enthalten ist, und zu den Positionsinformationen zuordnen, die den Audiodaten zugeordnet sind.


 
7. Empfangsvorrichtung (200), Folgendes umfassend:

einen Empfangsabschnitt (201), der eingerichtet ist, um einen Containerstrom in einem gegebenen Format zu empfangen, das einen Videostrom, einen Untertitelstrom, einen Audiostrom und Metainformationen umfasst, wobei

der Videostrom videocodierte Daten aufweist,

der Untertitelstrom Untertiteldaten aufweist, die einer Rede eines Sprechers entsprechen,

der Audiostrom objektcodierte Daten aufweist, die Audiodaten, deren Schallquelle die Rede des Sprechers ist, und Positionsinformationen hinsichtlich einer Position der Schallquelle in einem Bild des Videostroms umfassen, und die Metainformationen Untertiteldaten, die jeder Rede entsprechen, die in dem Untertitelstrom enthalten ist, zu Audiodaten, die jeder Rede entsprechen, die in dem Audiostrom enthalten ist, und zu den Positionsinformationen zuordnen, die den Audiodaten zugeordnet sind,

die Empfangsvorrichtung weiterhin Folgendes umfassend:

einen Steuerungsabschnitt, der eingerichtet ist zum Steuern

eines Videodecodierprozesses (203), wodurch Videodaten durch einen Decodierprozess des Videostroms erfasst werden,

eines Untertiteldecodierprozesses (217), wodurch Bitmap-Daten eines Untertitels, welcher der Rede des Sprechers entspricht, durch einen Decodierprozess des Untertitelstroms erfasst wird,

eines Videoüberlagerungsprozesses (204), wodurch Videodaten zur Anzeige durch Überlagerung der Bitmap-Daten des Untertitels auf den Videodaten erfasst werden,

eines Audiodecodierprozesses (214), wodurch Kanaldaten, die der Rede entsprechen, durch einen Decodierprozess des Audiostroms erfasst werden, und

eines Überlagerungspositionsprozesses, wodurch Anzeigekoordinaten eines Untertitels, der jeder der Reden entspricht, auf einem Bildschirm auf der Grundlage von Positionsinformationen hinsichtlich der Schallquelle ermittelt werden, die jeder der Reden entspricht, wie in Metainformationen angegeben, die in den Containerstrom eingefügt wurden, und eine Steuerung derartig durchgeführt wird, dass die Bitmap-Daten des Untertitels, der jeder der Reden entspricht, auf den Videodaten an der Position überlagert werden, die den ermittelten Anzeigekoordinaten entspricht.


 
8. Empfangsverfahren, Folgendes umfassend:

einen Empfangsschritt (201) des Empfangens eines Containerstroms in einem gegebenen Format, das einen Videostrom, einen Untertitelstrom, einen Audiostrom durch einen Empfangsabschnitt und Metainformationen umfasst, wobei

der Videostrom videocodierte Daten aufweist, der Untertitelstrom Untertiteldaten aufweist, die einer Rede eines Sprechers entsprechen,

der Audiostrom objektcodierte Daten aufweist, die Audiodaten, deren Schallquelle die Rede des Sprechers ist, und Positionsinformationen hinsichtlich einer Position der Schallquelle in einem Bild des Videostroms umfassen, und die Metainformationen Untertiteldaten, die jeder Rede entsprechen, die in dem Untertitelstrom enthalten ist, zu Audiodaten, die jeder Rede entsprechen, die in dem Audiostrom enthalten ist, und zu den Positionsinformationen zuordnen, die den Audiodaten zugeordnet sind,

das Empfangsverfahren weiterhin Folgendes umfassend:

einen Videodecodierschritt (203) des Erfassens von Videodaten durch einen Decodierprozess des Videostroms; einen Untertiteldecodierschritt (217) des Erfassens von Bitmap-Daten eines Untertitels, welcher der Rede des Sprechers entspricht, durch einen Decodierprozess des Untertitelstroms;

einen Videoüberlagerungsschritt (204) des Erfassens von Videodaten zur Anzeige durch Überlagerung der Bitmap-Daten des Untertitels auf den Videodaten,

einen Audiodecodierschritt (214) des Erfassens von Kanaldaten, die der Rede entsprechen, durch einen Decodierprozess des Audiostroms; und

einen Steuerunsschritt des Ermittelns von Anzeigekoordinaten eines Untertitels, der jeder der Reden entspricht, auf einem Bildschirm auf der Grundlage von Positionsinformationen hinsichtlich der Schallquelle, die jeder der Reden entspricht, wie in Metainformationen angegeben, die in den Containerstrom eingefügt wurden, und des Steuerns, so dass die Bitmap-Daten des Untertitels, der jeder der Reden entspricht, auf den Videodaten an der Position überlagert werden, die den ermittelten Anzeigekoordinaten entspricht.


 


Revendications

1. Appareil d'émission (100) comprenant :

une section de codage vidéo (112) adaptée pour générer un flux vidéo ayant des données vidéo codées ;

une section de codage de sous-titres (114) adaptée pour générer un flux de sous-titres ayant des données de sous-titres correspondant à un propos d'un orateur ;

une section de codage audio (115) adaptée pour générer un flux audio ayant des données objets codées comportant des données audio dont la source sonore est un propos de l'orateur et des informations de position concernant une position de la source sonore dans une image du flux vidéo ;

une section d'émission adaptée pour envoyer un flux conteneur dans un format donné comportant le flux vidéo, le flux de sous-titres et le flux audio ; et

une section d'insertion d'informations (111, 116) adaptée pour insérer, dans une couche du flux conteneur, des méta-informations destinées à associer des données de sous-titres correspondant à chaque propos inclus dans le flux de sous-titres avec des données audio correspondant à chaque propos inclus dans le flux audio et avec les informations de position associées auxdites données audio.


 
2. Appareil d'émission de la revendication 1, dans lequel
des informations d'identification sont ajoutées à chaque morceau de données de sous-titres correspondant à chacun des propos, et des informations d'identification sont ajoutées à chaque morceau de données audio correspondant à chacun des propos, et
les méta-informations sont des informations tabulaires qui associent les informations d'identification ajoutées aux données de sous-titres correspondant à chacun des propos et les informations d'identification ajoutées aux données audio correspondant à chacun des propos.
 
3. Appareil d'émission (100) de la revendication 2, dans lequel
des informations balises ayant les informations d'identification sont insérées dans une trame audio comportant des données audio correspondant à chacun des propos.
 
4. Appareil d'émission (100) de la revendication 1, dans lequel
les données de sous-titres sont des données TTML ou des données dans un format dérivé du TTML.
 
5. Appareil d'émission (100) de la revendication 1,
dans lequel le flux conteneur est un flux de transport (TS),
dans lequel les méta-informations sont insérées dans la couche du flux de transport (TS) par un multiplexeur (116) qui multiplexe également les flux audio, vidéo et de sous-titres.
 
6. Procédé d'émission comprenant :

une étape de codage vidéo (112) consistant à générer un flux vidéo qui a des données vidéo codées ;

une étape de codage de sous-titres (114) consistant à générer un flux de sous-titres qui a des données de sous-titres correspondant à un propos d'un orateur ;

une étape de codage audio (115) consistant à générer un flux audio ayant des données objets codées comportant des données audio dont la source sonore est un propos de l'orateur et des informations de position concernant une position de la source sonore dans une image du flux vidéo ;

une étape d'émission consistant en l'envoi d'un flux conteneur dans un format donné qui comporte le flux vidéo, le flux de sous-titres et le flux audio, par une section d'émission ; et

une étape d'insertion d'informations (111, 116) consistant à insérer, dans une couche du flux conteneur, des méta-informations qui associent des données de sous-titres correspondant à chaque propos contenu dans le flux de sous-titres avec des données audio correspondant à chaque propos contenu dans le flux audio et avec les informations de position associées auxdites données audio.


 
7. Appareil de réception (200) comprenant :

une section de réception (201) adaptée pour recevoir un flux conteneur dans un format donné comportant un flux vidéo, un flux de sous-titres, un flux audio et des méta-informations, dans lequel

le flux vidéo a des données vidéo codées,

le flux de sous-titres a des données de sous-titres correspondant à un propos d'un orateur,

le flux audio a des données objets codées comportant des données audio dont la source sonore est le propos de l'orateur et des informations de position concernant une position de la source sonore dans une image du flux vidéo, et les méta-informations associent des données de sous-titres correspondant à chaque propos inclus dans le flux de sous-titres avec des données audio correspondant à chaque propos inclus dans le flux audio et avec les informations de position associées auxdites données audio,

l'appareil de réception comprenant en outre :

une section de commande adaptée pour commander

un processus de décodage vidéo (203) par lequel des données vidéo sont acquises par un processus de décodage du flux vidéo,

un processus de décodage de sous-titres (217) par lequel des données matricielles d'un sous-titre correspondant au propos de l'orateur sont acquises par un processus de décodage du flux de sous-titres,

un processus de superposition vidéo (204) par lequel des données vidéo à afficher sont acquises par superposition des données matricielles des sous-titres sur les données vidéo,

un processus de décodage audio (214) par lequel des données de voie correspondant au propos sont acquises par un processus de décodage du flux audio, et

un processus de position de superposition par lequel des coordonnées d'affichage d'un sous-titre correspondant à chacun des propos sur un écran de contrôle sont trouvées sur la base d'informations de position concernant la source sonore correspondant à chacun des propos telles qu'indiquées dans des méta-informations insérées dans le flux conteneur et une commande est exécutée de telle sorte que les données matricielles du sous-titre correspondant à chacun des propos sont superposées sur les données vidéo à la position correspondant aux coordonnées d'affichage trouvées.


 
8. Procédé de réception comprenant :

une étape de réception (201) consistant en la réception d'un flux conteneur dans un format donné comportant un flux vidéo, un flux de sous-titres, un flux audio par une section de réception, et de méta-informations, dans lequel

le flux vidéo a des données vidéo codées,

le flux de sous-titres a des données de sous-titres correspondant à un propos d'un orateur,

le flux audio a des données objets codées comportant des données audio dont la source sonore est le propos de l'orateur et des informations de position concernant une position de la source sonore dans une image du flux vidéo, et les méta-informations associent des données de sous-titres correspondant à chaque propos inclus dans le flux de sous-titres avec des données audio correspondant à chaque propos inclus dans le flux audio et avec les informations de position associées auxdites données audio,

le procédé de réception comprenant en outre

une étape de décodage vidéo (203) consistant à acquérir des données vidéo par un processus de décodage du flux vidéo ;

une étape de décodage de sous-titres (217) consistant à acquérir des données matricielles d'un sous-titre correspondant au propos de l'orateur par un processus de décodage du flux de sous-titres ;

une étape de superposition vidéo (204) consistant à acquérir des données vidéo à afficher par superposition des données matricielles du sous-titre sur les données vidéo ;

une étape de décodage audio (214) consistant à acquérir des données de voie correspondant au propos par un processus de décodage du flux audio ; et

une étape de commande consistant à trouver des coordonnées d'affichage d'un sous-titre correspondant à chacun des propos sur un écran de contrôle sur la base d'informations de position concernant la source sonore correspondant à chacun des propos telles qu'indiquées dans des méta-informations insérées dans le flux conteneur et exécuter une commande de telle sorte que les données matricielles du sous-titre correspondant à chacun des propos soient superposées sur les données vidéo à la position correspondant aux coordonnées d'affichage trouvées.


 




Drawing

























































































Cited references

REFERENCES CITED IN THE DESCRIPTION



This list of references cited by the applicant is for the reader's convenience only. It does not form part of the European patent document. Even though great care has been taken in compiling the references, errors or omissions cannot be excluded and the EPO disclaims all liability in this regard.

Patent documents cited in the description




Non-patent literature cited in the description