U.S. patent application number 16/922671 was filed with the patent office on 2020-10-22 for reception apparatus, reception method, transmission apparatus, and transmission method.
This patent application is currently assigned to SONY CORPORATION. The applicant listed for this patent is SONY CORPORATION. Invention is credited to Jun KITAHARA, Naohisa KITAZATO, Yasuaki YAMAGISHI.
Application Number | 20200336797 16/922671 |
Document ID | / |
Family ID | 1000004942717 |
Filed Date | 2020-10-22 |
View All Diagrams
United States Patent
Application |
20200336797 |
Kind Code |
A1 |
KITAZATO; Naohisa ; et
al. |
October 22, 2020 |
RECEPTION APPARATUS, RECEPTION METHOD, TRANSMISSION APPARATUS, AND
TRANSMISSION METHOD
Abstract
The present technology relates to a reception apparatus, a
reception method, a transmission apparatus, and a transmission
method, by which an increase in data size of signaling information
transmitted by broadcasting can be suppressed. Provided is a
reception apparatus including: a reception unit that receives a
broadcast wave of digital broadcasting using an IP (Internet
Protocol) transmission system; and a control unit that controls, on
the basis of information for managing only a broadcast component
transmitted by a broadcast wave of the digital broadcasting, which
is first signaling information transmitted by a broadcast wave of
the digital broadcasting, or information for managing at least one
component of the broadcast component and a communication component
transmitted by communication, which is second signaling information
transmitted by communication, operations of respective units for
acquiring the at least one component of the broadcast component and
the communication component. The present technology is applicable
to a television receiver, for example.
Inventors: |
KITAZATO; Naohisa; (Tokyo,
JP) ; YAMAGISHI; Yasuaki; (Kanagawa, JP) ;
KITAHARA; Jun; (Shizuoka, JP) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
SONY CORPORATION |
Tokyo |
|
JP |
|
|
Assignee: |
SONY CORPORATION
Tokyo
JP
|
Family ID: |
1000004942717 |
Appl. No.: |
16/922671 |
Filed: |
July 7, 2020 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
15300564 |
Sep 29, 2016 |
10743082 |
|
|
PCT/JP2015/059587 |
Mar 27, 2015 |
|
|
|
16922671 |
|
|
|
|
Current U.S.
Class: |
1/1 |
Current CPC
Class: |
H04L 67/06 20130101;
H04N 21/64322 20130101; H04L 67/02 20130101; H04N 21/2362 20130101;
H04N 21/4345 20130101; H04H 2201/16 20130101; H04N 21/8586
20130101; H04L 65/608 20130101; H04L 65/4076 20130101; H04H 20/93
20130101; H04N 21/65 20130101 |
International
Class: |
H04N 21/643 20060101
H04N021/643; H04H 20/93 20060101 H04H020/93; H04L 29/06 20060101
H04L029/06; H04L 29/08 20060101 H04L029/08; H04N 21/2362 20060101
H04N021/2362; H04N 21/434 20060101 H04N021/434; H04N 21/65 20060101
H04N021/65; H04N 21/858 20060101 H04N021/858 |
Foreign Application Data
Date |
Code |
Application Number |
Apr 11, 2014 |
JP |
2014-081536 |
Claims
1. (canceled)
2. A reception apparatus, comprising: a tuner configured to receive
a digital broadcast signal based on an Internet Protocol (IP)
transmission system; and one or more processors configured to:
acquire service configuration information transmitted via the
digital broadcast signal, the service configuration information
including at least one of bootstrap information or communication
delivery information corresponding to signaling information, the
bootstrap information including IP address information and port
information about the signaling information transmitted via the
digital broadcast signal, the communication delivery information
including information which indicates communication delivery of the
signaling information, receive a user selection for a service
corresponding to the service configuration information, the user
selection being performed using a remote controller, acquire, based
on the service configuration information, the signaling information
corresponding to the service selected by the user selection,
acquire, based on the signaling information, at least one of a
broadcast component transmitted via the digital broadcast signal or
a communication component transmitted via a communication other
than the digital broadcast signal, and output the at least one of
the broadcast component or the communication component.
3. The reception apparatus according to claim 2, wherein the
information in the communication delivery information indicating
the communication delivery of the signaling information includes
location information including a piece of a uniform resource
locator (URL) corresponding to a signaling server.
4. The reception apparatus according to claim 2, wherein the one or
more processors are configured to determine acquisition of the
signaling information transmitted via the communication other than
the digital broadcast signal based on the service configuration
information and capability of the reception apparatus.
5. The reception apparatus according to claim 2, wherein the one or
more processors are configured to acquire the broadcast component
based on the signaling information transmitted via the digital
broadcast signal.
6. The reception apparatus according to claim 2, wherein the one or
more processors are configured to acquire the communication
component based on the signaling information transmitted via the
communication other than the digital broadcast signal.
7. The reception apparatus according to claim 2, wherein the one or
more processors are configured to receive input from a
microphone.
8. A reception method performed in a reception apparatus, the
reception method comprising: receiving, by a tuner, a digital
broadcast signal based on an Internet Protocol (IP) transmission
system, acquiring, by one or more processors, service configuration
information transmitted via the digital broadcast signal, the
service configuration information including at least one of
bootstrap information or communication delivery information
corresponding to signaling information, the bootstrap information
including IP address information and port information about the
signaling information transmitted via the digital broadcast signal,
the communication delivery information including information which
indicates communication delivery of the signaling information;
receiving, by the one or more processors, a user selection for a
service corresponding to the service configuration information, the
user selection being performed using a remote controller;
acquiring, by the one or more processors, based on the service
configuration information, the signaling information corresponding
to the service selected by the user selection; acquiring, by the
one or more processors, based on the signaling information, at
least one of a broadcast component transmitted via the digital
broadcast signal or a communication component transmitted via a
communication other than the digital broadcast signal; and output,
by the one or more processors, the at least one of the broadcast
component or the communication component.
9. The reception method according to claim 8, wherein the
information in the communication delivery information indicating
that the communication delivery of the signaling information
includes location information including a piece of a uniform
resource locator (URL) corresponding to a signaling server.
10. The reception method according to claim 8, comprising:
determining, by the one or more processors, to acquire the
signaling information transmitted via the communication other than
the digital broadcast signal based on the service configuration
information and capability of the reception apparatus.
11. The reception method according to claim 8, wherein the
acquiring, by the one or more processors, the broadcast component
is based on the signaling information transmitted via the digital
broadcast signal.
12. The reception method according to claim 8, further comprising:
acquiring, by the one or more processors, the communication
component based on the signaling information transmitted via the
communication other than the digital broadcast signal.
13. The reception method according to claim 8, further comprising:
receiving, by the one or more processors, input from a microphone.
Description
CROSS REFERENCE TO RELATED APPLICATIONS
[0001] This application is a continuation of U.S. patent
application Ser. No. 15/300,564, filed on Sep. 29, 2016, which is a
U.S. National Phase of International Patent Application No.
PCT/JP2015/059587, filed on Mar. 27, 2015, which claims priority
benefit of Japanese Patent Application No. JP 2014-081536, filed in
the Japan Patent Office on Apr. 11, 2014. The benefit of priority
is claimed to each of the foregoing, and the entire contents of
each of the foregoing are incorporated herein by reference.
TECHNICAL FIELD
[0002] The present technology relates to a reception apparatus, a
reception method, a transmission apparatus, and a transmission
method and more particularly to a reception apparatus, a reception
method, a transmission apparatus, and a transmission method, by
which an increase in data size of signaling information transmitted
by broadcasting can be suppressed.
BACKGROUND ART
[0003] In recent years, in the fields of digital broadcasting, in
addition to services utilizing broadcasting, hybrid services
cooperating with communication have been introduced (e.g., see
Patent Document 1). In such hybrid services, components such as
video, audio, and closed captions for providing those services are
transmitted by broadcasting or communication as streams. [0004]
Patent Document 1: Japanese Patent Application Laid-open No.
2011-66556
SUMMARY OF INVENTION
Problem to be Solved by the Invention
[0005] By the way, if the hybrid services are introduced, it is
necessary to describe information relating to broadcasting and
communication in signaling information. Thus, an increase in data
size of the signaling information transmitted by broadcasting is
assumed. Therefore, the data size of the signaling information
transmitted by broadcasting is required to be reduced.
[0006] The present technology has been made in view of the
above-mentioned circumstances to be capable of suppressing an
increase in data size of signaling information transmitted by
broadcasting.
Means for Solving the Problem
[0007] A reception apparatus according to a first aspect of the
present technology is a reception apparatus including: a reception
unit that receives a broadcast wave of digital broadcasting using
an IP (Internet Protocol) transmission system; and a control unit
that controls, on the basis of information for managing only a
broadcast component transmitted by a broadcast wave of the digital
broadcasting, which is first signaling information transmitted by a
broadcast wave of the digital broadcasting, or information for
managing at least one component of the broadcast component and a
communication component transmitted by communication, which is
second signaling information transmitted by communication,
operations of respective units for acquiring the at least one
component of the broadcast component and the communication
component.
[0008] The first signaling information may be information in units
of services and may include a plurality of management information
items for acquiring the broadcast component transmitted through a
FLUTE (File Delivery over Unidirectional Transport) session.
[0009] The first signaling information may be information in units
of services and may include one management information item
obtained by integrating a plurality of management information items
for acquiring the broadcast component transmitted through a FLUTE
session as a parameter defined at a component level.
[0010] The first signaling information and the second signaling
information may include management information defining information
relating to the second signaling information as a parameter at a
service level.
[0011] The management information may include information
indicating a range of signaling information, a version information,
and a URL (Uniform Resource Locator) indicating an acquisition
source, as the information relating to the second signaling
information.
[0012] The management information may further include information
indicating an update interval of the second signaling information,
as the information relating to the second signaling
information.
[0013] The management information may further include information
indicating a timing to terminate acquisition of the second
signaling information, as the information relating to the second
signaling information.
[0014] The second signaling information may be information in units
of services and may include a plurality of management information
items for acquiring the broadcast component and an MPD (Media
Presentation Description) complying with a standard of MPEG-DASH
(Moving Picture Expert Group-Dynamic Adaptive Streaming over HTTP)
as management information for acquiring the communication
component.
[0015] The first signaling information may be transmitted in an
upper layer than an IP layer in a hierarchy of a protocol in the IP
transmission system, and a common IP address may be assigned to the
broadcast component configuring a particular service and the first
signaling information.
[0016] The reception apparatus may be an independent apparatus or
may be an internal block configuring a single apparatus.
[0017] A reception method according to the first aspect of the
present technology is a reception method corresponding to the
reception apparatus according to the first aspect of the present
technology.
[0018] In the reception apparatus according to the first aspect of
the present technology, and a reception method, a broadcast wave of
digital broadcasting using an IP transmission system is received,
and, on the basis of information for managing only a broadcast
component transmitted by a broadcast wave of the digital
broadcasting, which is first signaling information transmitted by a
broadcast wave of the digital broadcasting, or information for
managing at least one component of the broadcast component and a
communication component transmitted by communication, which is
second signaling information transmitted by communication,
operations of respective units for acquiring the at least one
component of the broadcast component and the communication
component are controlled.
[0019] A transmission apparatus according to a second aspect of the
present technology is a transmission apparatus including: a first
acquisition unit that acquires first signaling information for
managing only a broadcast component transmitted by a broadcast wave
of digital broadcasting using an IP transmission system; a second
acquisition unit that acquires one or more broadcast components
configuring a service; and a transmission unit that transmits the
first signaling information together with the broadcast component
by a broadcast wave of the digital broadcasting using the IP
transmission system.
[0020] The first signaling information may be information in units
of services and may include a plurality of management information
items for acquiring the broadcast component transmitted through a
FLUTE session.
[0021] The first signaling information may be information in units
of services and may includes one management information item
obtained by integrating a plurality of management information items
for acquiring the broadcast component transmitted through a FLUTE
session as a parameter defined at a component level.
[0022] A receiver that receives a broadcast wave of the digital
broadcasting using the IP transmission system may be capable of
acquiring information for managing the at least one component of
the broadcast component and a communication component transmitted
by communication, which is second signaling information transmitted
by communication, and the first signaling information and the
second signaling information may include management information
defining information relating to the second signaling information
as a parameter at a service level.
[0023] The management information may include information
indicating a range of signaling information, version information,
and a URL of an acquisition source, as the information relating to
the second signaling information.
[0024] The management information may further include information
indicating an update interval of the second signaling information,
as the information relating to the second signaling
information.
[0025] The management information may further include information
indicating a timing to terminate acquisition of the second
signaling information, as the information relating to the second
signaling information.
[0026] The second signaling information may be information in units
of services and may include an MPD complying with a standard of
MPEG-DASH as the plurality of management information items for
acquiring the broadcast component and the management information
for acquiring the communication component.
[0027] The first signaling information may be transmitted in an
upper layer than an IP layer in a hierarchy of a protocol in the IP
transmission system, and a common IP address may be assigned to the
broadcast component configuring a particular service and the first
signaling information.
[0028] The transmission apparatus may be an independent apparatus
or may be an internal block configuring a single apparatus.
[0029] A transmission method according to the second aspect of the
present technology is a transmission method corresponding to the
transmission apparatus according to the second aspect of the
present technology.
[0030] In the transmission apparatus according to the second aspect
of the present technology and a transmission method, first
signaling information for managing only a broadcast component
transmitted by a broadcast wave of digital broadcasting using an IP
transmission system is acquired, one or more broadcast components
configuring the service is acquired, and the first signaling
information is transmitted together with the broadcast component by
a broadcast wave of the digital broadcasting using the IP
transmission system.
Effects of the Invention
[0031] In accordance with the first aspect and the second aspect of
the present technology, it is possible to suppress an increase in
data size of signaling information transmitted by broadcasting.
[0032] It should be noted that the effect described here is not
necessarily limitative and may be any effect described in the
present disclosure.
BRIEF DESCRIPTION OF DRAWINGS
[0033] FIG. 1 A diagram showing a protocol stack of digital
broadcasting in an IP transmission system.
[0034] FIG. 2 A diagram showing a configuration of a broadcast wave
of the digital broadcasting in the IP transmission system.
[0035] FIG. 3 A diagram showing a structure of signaling
information according to Solution 1.
[0036] FIG. 4 A diagram explaining a channel selection scenario of
a basic service according to Solution 1.
[0037] FIG. 5 A diagram explaining a channel selection scenario of
a hybrid service according to Solution 1.
[0038] FIG. 6 A diagram showing a structure of signaling
information according to Solution 2.
[0039] FIG. 7 A diagram explaining a channel selection scenario of
a basic service according to Solution 2.
[0040] FIG. 8 A diagram explaining a channel selection scenario of
a hybrid service according to Solution 2.
[0041] FIG. 9 A diagram for explaining a scenario in the case where
a transition from the basic service to the hybrid service is
made.
[0042] FIG. 10 A diagram for explaining a scenario in the case
where an acquisition source of streams is changed in the hybrid
service.
[0043] FIG. 11 A diagram for explaining another scenario in the
case where the acquisition source of the streams is changed in the
hybrid service.
[0044] FIG. 12 A diagram for explaining a scenario in the case
where a transition from the hybrid service to the basic service is
made.
[0045] FIG. 13 A diagram for explaining another scenario in the
case where a transition from the hybrid service to the basic
service is made.
[0046] FIG. 14 A diagram showing a syntax of an SCD common to
Solutions 1 and 2.
[0047] FIG. 15 A diagram showing an SPD according to Solution
1.
[0048] FIG. 16 A diagram showing an SPD according to Solution
2.
[0049] FIG. 17 A diagram showing details of a ComponentLocation
element.
[0050] FIG. 18 A diagram showing an SPD common to Solutions 1 and
2.
[0051] FIG. 19 A diagram showing a syntax of a Protocol Version
Descriptor.
[0052] FIG. 20 A diagram showing a syntax of an NRT Service
Descriptor.
[0053] FIG. 21 A diagram showing a syntax of a Capability
Descriptor.
[0054] FIG. 22 A diagram showing a syntax of an Icon
Descriptor.
[0055] FIG. 23 A diagram showing a syntax of an ISO-639 Language
Descriptor.
[0056] FIG. 24 A diagram showing a syntax of a Receiver Targeting
Descriptor.
[0057] FIG. 25 A diagram showing a syntax of an Associated Service
Descriptor.
[0058] FIG. 26 A diagram showing a syntax of a Content Advisory
Descriptor.
[0059] FIG. 27 A diagram showing a syntax of an AVC Video
Descriptor.
[0060] FIG. 28 A diagram showing a syntax of an HEVC Video
Descriptor.
[0061] FIG. 29 A diagram showing a syntax of an MPEG4 AAC Audio
Descriptor.
[0062] FIG. 30 A diagram showing a syntax of an AC3 Audio
Descriptor.
[0063] FIG. 31 A diagram showing a syntax of Caption
Parameters.
[0064] FIG. 32 A diagram showing a configuration example of a
broadcast communication system.
[0065] FIG. 33 A diagram showing a configuration example of a
transmission apparatus.
[0066] FIG. 34 A diagram showing a configuration example of an
Internet server.
[0067] FIG. 35 A diagram showing a configuration example of a
reception apparatus.
[0068] FIG. 36 A flowchart explaining transmission processing.
[0069] FIG. 37 A flowchart explaining streaming delivery
processing.
[0070] FIG. 38 A flowchart explaining signaling
information-providing processing.
[0071] FIG. 39 A flowchart explaining channel selection
processing.
[0072] FIG. 40 A diagram showing a configuration example of a
computer.
MODE(S) FOR CARRYING OUT THE INVENTION
[0073] Hereinafter, embodiments of the present technology will be
described with reference to the drawings. Note that descriptions
will be made in the following order.
[0074] 1. Outline of Digital Broadcasting in IP Transmission
System
[0075] 2. Solution 1
[0076] (1) Structure of Signaling Information
[0077] (2) Specific Application Example
[0078] 3. Solution 2
[0079] (1) Structure of Signaling Information
[0080] (2) Specific Application Example
[0081] 4. Common to Solutions 1 and 2
[0082] (1) Structure of Signaling Information
[0083] (2) Specific Application Example
[0084] 5. Syntax
[0085] (1) Syntax of SCD
[0086] (2) Syntax of SPD
[0087] 6. System Configuration
[0088] 7. Flows of Processing Executed by Apparatuses
[0089] 8. Configuration of Computer
[0090] <1. Outline of Digital Broadcasting in IP Transmission
System>
[0091] (Protocol Stack)
[0092] FIG. 1 is a diagram showing a protocol stack of digital
broadcasting in an IP transmission system.
[0093] As shown in FIG. 1, a lowest layer is a physical layer. The
frequency band of broadcast waves assigned for a service (channel)
corresponds to this. An upper layer that is adjacent to the
physical layer is an IP layer sandwiching a BBP stream (Base Band
Packet Stream) therebetween. The BBP stream is a stream including
packets storing various types of data of the IP transmission
system.
[0094] The IP layer corresponds to an IP (Internet Protocol) in the
TCP/IP protocol stack. IP packets are identified by IP addresses.
An upper layer adjacent to the IP layer is a UDP layer. In a
further upper layer, RTP and FLUTE/ALS are shown. Thus, in the
digital broadcasting in the IP transmission system, packets in
which a port number of a UDP (User Datagram Protocol) is specified
are transmitted and an RTP (Real time Transport Protocol) session
or a FLUTE (File Delivery over Unidirectional Transport) session is
established, for example.
[0095] In an upper layer adjacent to the FLUTE/ALS, fMP4
(Fragmented MP4) is shown. In addition, in an upper layer adjacent
to the RTP and the fMP4, video data (Video), audio data (Audio),
and closed caption data (Closed Caption) are shown, for example.
Thus, the RTP session is used when the video data and the audio
data are transmitted as streams, and the FLUTE session is used when
the video data and the audio data are transmitted as files.
[0096] Further, in an upper layer of the FLUTE/ALS, NRT content,
ESG, and SCS are shown. The NRT content, ESG, and SCS are
transmitted through the FLUTE session. The NRT content is content
transmitted by NRT (Non-Real Time) broadcasting, stored in storage
of a receiver, and then reproduced. Note that the NRT content is an
example of the content and other content files may be transmitted
through the FLUTE session. The ESG is an electronic service
guide.
[0097] The SCS (Service Channel Signaling) is signaling information
in units of services and transmitted through the FLUTE session. For
example, USD (User Service Description), MPD (Media Presentation
Description), SDP (Session Description Protocol), FDD (File
Delivery Description), SPD (Service Parameter Description), and IS
(Initialization Segment) are transmitted as the SCS.
[0098] LLS (Low Layer Signaling) is low-layer signaling information
and transmitted in the BBP stream. For example, service
configuration information items such as SCD (Service Configuration
Description), EAD (Emergency Alerting Description), and RRD (Region
Rating Description) are transmitted as the LLS.
[0099] (Configuration of Broadcast Wave in IP Transmission
System)
[0100] FIG. 2 is a diagram showing a configuration of a broadcast
wave of the digital broadcasting in the IP transmission system.
[0101] As shown in FIG. 2, a plurality of BBP streams are
transmitted in a broadcast wave (RF Channel) having a predetermined
frequency band. Further, each of the BBP streams includes an NTP
(Network Time Protocol), a plurality of service channels, an
electronic service guide (ESG Service), and an LLS. Note that the
NTP, the service channels, and the electronic service guide are
transmitted according to a UDP/IP protocol while the LLS is
transmitted in the BBP stream. Further, the NTP is time information
and can be common to the plurality of service channels.
[0102] The service channels (hereinafter, referred to as
"services") include components such as video, audio, and closed
captions, which are information items configuring a program, and an
SCS such as USD and SPD. A common IP address is added to the
services. Using this IP address, the components, the SCS, and the
like can be packaged for one or more services.
[0103] An RF channel ID (RF channel id) is assigned to a broadcast
wave (RF Channel) having a predetermined frequency band. Further, a
BBP stream ID (BBP stream id) is assigned to one or more BBP
streams transmitted by each broadcast wave. In addition, a service
ID (service id) is assigned to one or more services transmitted by
each of the BBP streams.
[0104] Such a configuration corresponding to a combination of
network ID (network id), transport stream ID (transport stream id),
and service ID (service id) used in the MPEG2-TS (Moving Picture
Expert Group 2-Transport Stream) system (hereinafter, referred to
as "triplet") is employed as the ID system of the IP transmission
system. This triplet indicates a BBP stream configuration and a
service configuration in a broadcasting network.
[0105] The use of this ID system can achieve compatibility with the
currently widely spread MPEG2-TS system, and hence it is possible
to easily perform simulcast during transition from the MPEG2-TS
system to the IP transmission system, for example. It should be
noted that the RF channel ID and the BBP stream ID in the ID system
of the IP transmission system correspond to the network ID and the
transport stream ID in the MPEG2-TS system, respectively.
[0106] <2. Solution 1>
[0107] By the way, if a hybrid service is introduced into the
digital broadcasting in the IP transmission system, it is necessary
to describe information relating to broadcasting and communication
in signaling information. Thus, an increase in data size of the
signaling information transmitted by broadcasting is assumed.
Therefore, the data size is required to be reduced. In view of
this, as a method for addressing this requirement, two solutions,
Solution 1 and Solution 2 will be proposed in the present
technology. Solution 1 will be first described.
(1) Structure of Signaling Information
[0108] FIG. 3 is a diagram showing a structure of signaling
information according to Solution 1. In FIG. 3, different
acquisition sources for signaling information and components are in
a left-hand region and a right-hand region with respect to a
boundary line shown as the bold line extending in a vertical
direction of the figure. The left-hand region shows "broadcasting
acquisition" in which the acquisition source is broadcasting and
the right-hand region shows "communication acquisition" in which
the acquisition source is communication.
[0109] As the signaling information in the broadcasting acquisition
in the left-hand region of the figure, LLS (Low Layer Signaling)
and SCS (Service Channel Signaling) are provided. The LLS is
acquired when the receiver performs initial scanning, for example.
SCD, EAD, and RRD are acquired as the LLS. The SCD (Service
Configuration Description) employs the triplet used in the MPEG2-TS
system. This triplet indicates the BBP stream configuration and the
service configuration in the broadcasting network. The SCD further
includes information on an IP address and the like serving as
attribute/setting information in units of services, bootstrap
information for acquiring the SCS and ESG, and the like. The SCD
further includes, if the signaling information is transmitted by
communication, information relating to the signaling information
transmitted by communication.
[0110] The EAD (Emergency Alerting Description) includes
information relating to emergency notice. The RRD (Region Rating
Description) includes rating information. Note that the SCD, EAD,
and RRD are described with a markup language, for example, an XML
(Extensible Markup Language).
[0111] The SCS is transmitted through the FLUTE session, and hence
acquired according to IP address, port number, and TSI described in
SCS bootstrap information of the SCD. USD, MPD, SDP, FDD, SPD, and
IS are acquired as the SCS. The USD (User Service Description)
includes link information for referring to the MPD, SDP, and FDD.
Note that the USD is in some cases called USBD (User Service Bundle
Description). The MPD (Media Presentation Description) includes
information on URLs (Uniform Resource Locators) and the like for
streams (components) transmitted in units of services. Note that
the MPD complies with the MPEG-DASH (Moving Picture Expert
Group-Dynamic Adaptive Streaming over HTTP) standard.
[0112] The SDP (Session Description Protocol) includes a service
attribute in units of services, component configuration
information, a component attribute, component filter information,
component location information, and the like. The FDD (File
Delivery Description) includes, as index information for each TSI
(Transport Session Identifier), location information (e.g., URL)
and information on TOI (Transport Object Identifier) and the like.
Note that the FDD may be included as an element in the USD.
[0113] The SPD (Service Parameter Description) includes various
parameters defined at a service level and a component level. The
SPD further includes, if the signaling information is transmitted
by communication, information relating to the signaling information
transmitted by communication. The IS (Initialization Segment) is
control information that is, if (files of) the components are
segmented and transmitted as segments complying with the ISO Base
Media File Format standard, transmitted together with media
segments storing segment data. Note that the IS transmitted in
units of components such as video and audio.
[0114] If the components such as the video and audio are
transmitted as streams in units of segments by broadcasting through
the FLUTE, a segment transmitted through the FLUTE session is
determined by using the signaling information such as the SDP, FDD,
and IS and a broadcast component is acquired.
[0115] On the other hand, the USD, MPD, SDP, FDD, SPD, and IS are
acquired as the signaling information in the communication
acquisition on the right-hand side of the figure. That is, the
signaling information transmitted by communication has a signaling
system identical to that of the SCS in the broadcasting
acquisition. This signaling information is acquired according to a
URL of a signaling server that is described in the SCD transmitted
by broadcasting. If the signaling information is transmitted by
communication, the URL of the signaling server that provides the
signaling information is described in the SPD acquired by
broadcasting, and hence the signaling information may be acquired
according to that URL. That is, the SCD is acquired during initial
scanning or the like, and hence the URL described in the SCD is a
stationary URL (Fixed URL). On the other hand, the SPD is acquired
during channel selection or the like, and hence it can be said that
the URL described in the SPD is a dynamic URL.
[0116] Then, if the components such as the video and audio are
transmitted as streams by broadcasting through the FLUTE in units
of segments, a segment is determined by using the signaling
information such as the SDP, FDD, and IS, and a broadcast component
(Component) is acquired. If the components are, as the streams,
transmitted by communication in units of segments, a segment is
determined by using the signaling information such as the MPD and a
communication component (Component) is acquired. Note that the
files of the USD, MPD, SDP, FDD, SPD, IS, and the like acquired by
communication can be handled as a single file in the ZIP file
format, for example.
[0117] As described above, in the signaling information acquired by
broadcasting, only the information relating to the broadcast
component (stream) transmitted by broadcasting is described and the
information relating to the communication component (stream)
transmitted by communication is not described. Therefore, the
increase in data size of signaling information transmitted by
broadcasting can be suppressed. In contrast, in the signaling
information acquired by communication, the information relating to
the broadcast component (stream) transmitted by broadcasting and
the information relating to the communication component (stream)
transmitted by communication are both described. The signaling
information transmitted by communication is acquired via the
Internet, and hence there is less need to worry about the data size
in comparison with the signaling information transmitted by
broadcasting. Thus, this is not an issue.
(2) Specific Application Example
[0118] Next, a specific application example according to Solution 1
will be described. Here, a service formed of only streams
transmitted by broadcasting (hereinafter, referred to as "basic
service") and a service formed of streams transmitted by
broadcasting and communication (hereinafter, referred to as "hybrid
service") will be described in the stated order.
[0119] (2-1) Basic Service
[0120] FIG. 4 is a diagram for explaining a channel selection
scenario of the basic service according to Solution 1.
[0121] In FIG. 4, a transmitter of a broadcasting station
(broadcaster) transmits, through a broadcast wave (RF Channel) of
the digital broadcasting using the IP transmission system,
components and signaling information configuring each service in a
BBP stream. It should be noted that the above-mentioned ID system
is employed in this digital broadcasting. Further, files of the
components and signaling information are transmitted through the
FLUTE session. In the basic service, only streams transmitted by
broadcasting are acquired and streams and signaling information
transmitted by communication from an Internet server such as a
streaming server and a signaling server are not acquired.
[0122] As shown in FIG. 4, a receiver placed in each house or the
like acquires an SCD transmitted in LLS by initial scanning and
records it on an NVRAM (Procedure 1). This SCD includes SCS
Bootstrap information in which IP address, port number, and TSI for
acquiring the SCS are described. When the user selects a particular
service (basic service) (Procedure 2), the receiver reads out the
SCD from the NVRAM (Procedure 3), and is connected to the SCS
transmitted by a broadcast wave according to the SCS Bootstrap
information and acquire signaling information (Procedures 4 and
5).
[0123] The file of the SCS is being transmitted through the FLUTE
session, and hence the signaling information such as the USD, MPD,
SDP, and FDD is acquired by parsing data stored in LCT packets
(Procedure 6). Although a deliveryMethod element is considered as a
child element of the USD, for the sake of description, it is shown
separately from the USD in the figure. Further, link information is
described in the USD and this link information is used for
acquiring the MPD, SDP, and FDD. Those signaling information items
are all included in the SCS, and hence all of them can also be
acquired from it.
[0124] In AdaptationSet elements of the MPD, Representation
elements are provided and components transmitted as streams by
broadcasting or communication are enumerated. In the Representation
elements, segment URLs indicating acquisition sources of the
components are enumerated other than representation IDs. In the
example of the MPD of FIG. 4, video and audio components are
enumerated in the Representation elements within the AdaptationSet
elements. Further, in the deliveryMethod element of the USD,
information for identifying a delivery mode for the components is
specified.
[0125] In the channel selection scenario of FIG. 4, the basic
service, that is, the component is transmitted only by
broadcasting, and hence a broadcastAppService element is provided
in the deliveryMethod element and the URL of the component
transmitted by broadcasting is specified in a basepattern element.
By matching the segment URLs described in the MPD with the URL
described in the deliveryMethod element, it is determined that the
video and audio components enumerated in the MPD are being
transmitted by broadcasting (Procedure 7).
[0126] Further, tsi attribute, contentLocation attribute, and toi
attribute are described in the FDD. In the tsi attribute, a TSI
(Transport Session Identifier) that is identification information
of each FLUTE session are specified. Further, in the toi attribute,
an TOI (Transport Object Identifier) that is identification
information of a plurality of objects transmitted in each FLUTE
session is specified. In the contentLocation attribute, the URL of
the file is specified. By matching the segment URLs described in
the MPD with the URLs described in the FDD, the TSI and TOI for
acquiring the components enumerated in the MPD are determined
(Procedure 8). In addition, by referring to the SPD, IP addresses
and port numbers for acquiring those video and audio components are
determined (Procedure 8).
[0127] In this manner, the IP address, port number, TSI, and TOI
for acquiring the video component and the IP address, port number,
TSI, and TOI for acquiring the audio component are acquired. Using
the IP addresses, port numbers, TSIs, and TOIs for the video and
audio components, the receiver is connected to the video and audio
streams being transmitted through the FLUTE session, and acquires
the LCT packets (Procedures 9 and 10). Then, the receiver extracts
segment data (media segment) stored in the LCT packets and
temporarily stores it in a buffer for buffering (Procedures 11 and
12), and performs rendering (Procedure 13). With this, at the
receiver, a picture of a program corresponding to the particular
service (basic service) selected by the user is displayed and at
the same time sound synchronized with that picture is output.
[0128] As described above, in the basic service, the video and
audio components (streams) are delivered only by broadcasting, and
hence the information relating to those components is described in
the signaling information (SCS) transmitted by broadcasting.
Therefore, the receiver acquires the video and audio components
(streams) on the basis of the signaling information transmitted by
broadcasting. At this time, only the information relating to the
components being transmitted by broadcasting is described in this
signaling information, and hence the increase in data size can be
suppressed. Note that the receiver not adapted for the hybrid
service basically cannot be connected to the Internet, and hence
cannot acquire the signaling information transmitted by
communication. Even such a receiver can acquire the signaling
information transmitted by broadcasting, and hence it is possible
to view the program of the basic service.
[0129] (2-2) Hybrid Service
[0130] FIG. 5 is a diagram for explaining a channel selection
scenario of a hybrid service according to Solution 1.
[0131] In FIG. 5, as in FIG. 4, the transmitter of the broadcasting
station is transmitting, through a broadcast wave of the digital
broadcasting using the IP transmission system, components and
signaling information configuring each service in a BBP stream.
Further, in the hybrid service, at the receiver, streams
transmitted by communication are also acquired, and hence a stream
of audio (A2: Audio2) is transmitted from the streaming server.
Further, signaling information (e.g., ZIP file of USD) is
transmitted from the signaling server.
[0132] As shown in FIG. 5, a receiver placed in each house or the
like acquires an SCD transmitted in LLS by initial scanning and
records it on an NVRAM (Procedure 1). In this SCD, a
SignalingOverinternet element, and a hybrid attribute and an url
attribute as descendants thereof are described. In the hybrid
attribute, "basic" is specified in the case of the basic service
and "hybrid" is specified in the case of the hybrid service.
Therefore, "hybrid" is specified here. In the url attribute, a URL
of a signaling server, for example, is specified as an acquisition
source of the signaling information.
[0133] Note that the initial scanning is performed at the start of
the use of the receiver, for example. Thus, it is not frequently
performed. Therefore, even if, at the receiver adapted for the
hybrid service, "hybrid" is specified as the hybrid attribute in
the SignalingOverinternet element of the SCD acquired in the
initial scanning, it is assumed that the streams of the basic
service are received. That is, when "hybrid" is specified in this
hybrid attribute, it can be said that it indicates that streams of
the hybrid service may be received.
[0134] When the user selects a particular service (hybrid service)
(Procedure 2), the receiver reads out the SCD from the NVRAM
(Procedure 3), and accesses the signaling server via the Internet
according to the URL specified as the url attribute of the
SignalingOverinternet element, such that the signaling information
such as the USD, MPD, SDP, and FDD is acquired (Procedures 4, 5,
and 6).
[0135] In the example of the MPD of FIG. 5, video and audio
components are enumerated in the Representation elements within the
AdaptationSet elements. Further, in the channel selection scenario
of FIG. 5, the hybrid service, that is, the component is
transmitted by broadcasting and communication, and hence a
broadcastAppService element and a unicastAppService element are
provided in a deliveryMethod element of the USD. In a basepattern
element of the broadcastAppService element, a URL of a component
transmitted by broadcasting is specified. In a basepattern element
of the unicastAppService element, a URL of a component transmitted
by communication is specified.
[0136] By matching the segment URLs described in the MPD with the
URLs described in the deliveryMethod element of the USD, it is
determined that, out of the components enumerated in the MPD, the
video component is being transmitted by broadcasting and the audio
component is being transmitted by communication (Procedure 8).
Further, by matching the segment URLs described in the MPD with the
URLs described in the FDD, the TSI and TOI for acquiring the video
component transmitted by broadcasting are determined (Procedure 9).
In addition, by referring to the SPD, the IP address and port
number for acquiring the video component are determined (Procedure
9). Note that the audio component is transmitted by communication,
and hence the segment URL of the MPD corresponding to this audio
component is the URL of the streaming server delivering the
component (stream) of the audio (A2: Audio2).
[0137] In this manner, the IP address, port number, TSI, and TOI
for acquiring the video component and the URL of the streaming
server for acquiring the audio component are acquired. Using the IP
address, port number, TSI, and TOI of the video component, the
receiver is connected to the video stream being transmitted through
the FLUTE session, and acquires LCT packets (Procedure 10-1). Then,
the receiver extracts segment data (media segment) stored in the
LCT packets (Procedure 12). Further, the receiver accesses the
streaming server via the Internet according to the segment URL of
the MPD, and is connected to the audio stream (Procedure 10-2).
[0138] As a result, the video and audio streams are acquired
(Procedure 11), and hence the receiver temporarily stores data of
them in a buffer for buffering (Procedure 13), and performs
rendering (Procedure 14). With this, at the receiver, a picture of
a program corresponding to the particular service (hybrid service)
selected by the user is displayed and at the same time sound
synchronized with that picture is output.
[0139] As described above, in the hybrid service, the video and
audio components (streams) are delivered by broadcasting and
communication, and hence the information relating to those
components is described in the signaling information transmitted by
communication. Therefore, the receiver acquires the video and audio
components (streams) on the basis of the signaling information
transmitted by communication. At this time, in this signaling
information, the information items relating to the components
transmitted by broadcasting and communication are both described.
However, in the case of the signaling information transmitted by
communication, there is less need to worry about the data size in
comparison with the signaling information transmitted by
broadcasting. Thus, this is not an issue. Note that the receiver
adapted for the hybrid service can be connected to the Internet as
a premise, and hence even if the signaling information is
transmitted by communication, this is not an issue.
[0140] <3. Solution 2>
[0141] Next, Solution 2 will be described. In Solution 2, the
signaling information acquired by broadcasting is simplified in
comparison with Solution 1 described above.
(1) Structure of Signaling Information
[0142] FIG. 6 is a diagram showing a structure of signaling
information according to Solution 2. In FIG. 6, as in FIG. 3, the
left-hand region shows the "broadcasting acquisition" and the
right-hand region shows the "communication acquisition." Further,
out of the signaling information in broadcasting acquisition in the
left-hand region of the figure, the LLS is the same as that in FIG.
3 and the SCS is simplified in comparison with FIG. 3. That is, in
FIG. 6, the SPD and the IS are acquired as the SCS.
[0143] Here, the SPD includes various parameters defined at the
service level and the component level. As this parameter at the
component level, information (e.g., port number, TSI, and TOI) for
acquiring the component transmitted by broadcasting is described.
Thus, a segment transmitted through the FLUTE session is determined
and a broadcast component (Component) is acquired.
[0144] On the other hand, as the signaling information in the
communication acquisition on the right-hand side of the figure, the
USD, MPD, SDP, FDD, SPD, and IS are acquired as in FIG. 3. Then, by
using the signaling information such as the SDP, FDD, and IS, a
segment transmitted through the FLUTE session is determined and a
broadcast component (Component) is acquired. Further, by using the
signaling information such as the MPD, a segment is determined and
a communication component (Component) is acquired.
[0145] As described above, information items for acquiring the
broadcast components (streams) transmitted by broadcasting are
collected in the SPD, and hence the signaling information can be
simplified without using the USD, MPD, SDP, and FDD. Further, in
the structure of the signaling information according to Solution 2,
as in Solution 1, in the signaling information acquired by the
broadcasting, only the information relating to the broadcast
component (stream) transmitted by broadcasting is described and the
information relating to the communication component transmitted by
communication (stream) is not described, and hence it is possible
to suppress the increase in data size of signaling information
transmitted by broadcasting.
(2) Specific Application Example
[0146] Next, a specific application example according to Solution 2
will be described, and, as in Solution 1 described above, channel
selection scenarios in the basic service and the hybrid service
will be described.
[0147] (2-1) Basic Service
[0148] FIG. 7 is a diagram for explaining a channel selection
scenario of a basic service according to Solution 2.
[0149] In FIG. 7, as in FIG. 4, the transmitter of the broadcasting
station is transmitting, through a broadcast wave of the digital
broadcasting using the IP transmission system, components and
signaling information configuring each service in a BBP stream.
Further, in the basic service, streams transmitted by broadcasting
are only acquired and streams and signaling information transmitted
by communication from the Internet server are not acquired.
[0150] As shown in FIG. 7, a receiver placed in each house or the
like acquires an SCD transmitted in LLS by initial scanning and
records it on an NVRAM (Procedure 1). This SCD includes SCS
Bootstrap information in which IP address, port number, and TSI for
acquiring the SCS are described. When the user selects a particular
service (basic service) (Procedure 2), the receiver reads out the
SCD from the NVRAM (Procedure 3), and is connected to the SCS
transmitted by a broadcast wave according to the SCS Bootstrap
information and acquire the signaling information (Procedures 4 and
5).
[0151] The file of the SCS is being transmitted through the FLUTE
session, and hence the signaling information such as the SPD is
acquired by parsing data stored in LCT packets (Procedure 6). Here,
in the SPD, componentId attribute, componentType attribute, and
ComponentLocation element are described as the parameters at the
level of the components such as the video and audio. In the
componentId attribute, a component ID is specified. In the
componentType attribute, type information of the component is
specified.
[0152] Further, in the ComponentLocation element, portNum
attribute, tsi attribute, startToi attribute, and endToi attribute
are described as location information for each component. That is,
port number, TSI, and TOI are specified as information for
acquiring the components (streams) transmitted by broadcasting.
Note that, in the startToi attribute, a start value of the TOI if
the TOI changes in time series is specified. Further, in the endToi
attribute, an end value of the TOI if the TOI changes in time
series is specified. That is, by specifying the startToi attribute
and the endToi attribute, that value is sequentially incremented
from the start value to the end value of the TOI.
[0153] In this manner, the IP address, port number, TSI, and TOI
for acquiring the video component and the IP address, port number,
TSI, and TOI for acquiring the audio component are acquired
(Procedure 7). Using the IP addresses, port numbers, TSIs, and TOIs
for the video and audio components, the receiver is connected to
the stream being transmitted through the FLUTE session, and
acquires the LCT packets (Procedures 8 and 9). Then, the receiver
extracts segment data (media segment) stored in the LCT packets and
temporarily stores it in a buffer for buffering (Procedures 10 and
11), and performs rendering (Procedure 12). With this, at the
receiver, a picture of a program corresponding to the particular
service (basic service) selected by the user is displayed and at
the same time sound synchronized with that picture is output.
[0154] As described above, in the basic service, the video and
audio components (streams) are delivered only by broadcasting, and
hence the information relating to those components is described in
the signaling information (SCS) transmitted by broadcasting.
Therefore, the receiver acquires the video and audio components
(streams) on the basis of the signaling information transmitted by
broadcasting. At this time, only the information relating to the
components being transmitted by broadcasting is described in this
signaling information, and hence the increase in data size can be
suppressed.
[0155] (2-2) Hybrid Service
[0156] FIG. 8 is a diagram for explaining a channel selection
scenario of a hybrid service according to Solution 2.
[0157] In FIG. 8, as in FIG. 4, the transmitter of the broadcasting
station is transmitting, through a broadcast wave of the digital
broadcasting using the IP transmission system, components and
signaling information configuring each service in a BBP stream.
Further, in the hybrid service, at the receiver, streams
transmitted by communication are also acquired, and hence a stream
of audio (A2: Audio2) is transmitted from the streaming server.
Further, signaling information (e.g., ZIP file including USD and
the like) is being transmitted from the signaling server.
[0158] As shown in FIG. 8, a receiver placed in each house or the
like acquires an SCD transmitted in LLS by initial scanning, and
records it on an NVRAM (Procedure 1). In this SCD, a
SignalingOverinternet element, and a hybrid attribute and an url
attribute as descendants thereof are described. In the hybrid
attribute, "hybrid" is specified. Further, the URL of the signaling
server, for example, is specified in the url attribute. Note that,
as described above, even when "hybrid" is specified as the hybrid
attribute in the SignalingOverinternet element, it is assumed that
the streams of the basic service are received. Therefore, when
"hybrid" is specified in this hybrid attribute, it can be said that
it indicates that streams of the hybrid service may be
received.
[0159] When the user selects a particular service (hybrid service)
(Procedure 2), the receiver reads out the SCD from the NVRAM
(Procedure 3), and accesses the signaling server via the Internet
according to the URL specified as the url attribute of the
SignalingOverinternet element, and accesses the signaling server
via the Internet, such that the signaling information such as the
USD, MPD, SDP, and FDD is acquired (Procedures 4, 5, and 6).
[0160] In the example of the MPD of FIG. 8, video and audio
components are enumerated in the Representation elements within the
AdaptationSet elements. Further, in the channel selection scenario
of FIG. 8, the hybrid service, that is, the component is
transmitted by broadcasting and communication, and hence the
broadcastAppService element and the unicastAppService element are
provided in the deliveryMethod element of the USD.
[0161] Then, by matching the segment URLs described in the MPD with
the URLs described in the deliveryMethod element of the USD, it is
determined that, out of the components enumerated in the MPD, the
video component is being transmitted by broadcasting and the audio
component is being transmitted by communication (Procedure 8).
Further, by matching the segment URLs described in the MPD with the
URLs described in the FDD, the TSI and TOI for acquiring the video
component transmitted by broadcasting are determined (Procedure
9).
[0162] In addition, by referring to the SPD, the IP address and
port number for acquiring the video component are determined
(Procedure 9). Note that the audio component is transmitted by
communication, and hence the segment URL of the MPD corresponding
to this audio component is the URL of the streaming server
delivering the component (stream) of the audio (A2: Audio2).
[0163] In this manner, the IP address, port number, TSI, and TOI
for acquiring the video component and the URL of the streaming
server for acquiring the audio component are acquired. Using the IP
address, port number, TSI, and TOI of the video component, the
receiver is connected to the video stream being transmitted through
the FLUTE session, and acquires LCT packets (Procedure 10-1). Then,
the receiver extracts segment data (media segment) stored in the
LCT packets (Procedure 12). Further, the receiver accesses the
streaming server via the Internet according to the segment URL of
the MPD, and is connected to the audio stream (Procedure 10-2).
[0164] As a result, the video and audio streams are acquired
(Procedure 11), and hence the receiver temporarily stores data of
them in a buffer for buffering (Procedure 13), and performs
rendering (Procedure 14). With this, in the receiver, a picture of
a program corresponding to the particular service (hybrid service)
selected by the user is displayed and at the same time sound
synchronized with that picture is output.
[0165] As described above, in the hybrid service, the video and
audio components (streams) are delivered by broadcasting and
communication, and hence the information relating to those
components is described in the signaling information transmitted by
communication. Therefore, the receiver acquires the video and audio
components (streams) on the basis of the signaling information
transmitted by communication. At this time, in this signaling
information, the information items relating to the components
transmitted by broadcasting and communication are both described.
In the case of the signaling information transmitted by
communication, there is less need to worry about the data size in
comparison with the signaling information transmitted by
broadcasting. Thus, this is not an issue.
[0166] <4. Solutions 1 and 2>
[0167] By the way, at the receiver, the selected particular service
is continued in the hybrid service. However, when the program is
changed, for example, a case where a transition from the basic
service to the hybrid service is made, a case where a transition
from the hybrid service to the basic service is made, and the like
are conceivable other than a case where the acquisition source of
streams is changed. Hereinafter, those cases will be described. It
should be noted that the following descriptions are common to
Solution 1 and Solution 2, and hence descriptions of the both
solutions will be made together, not separately.
(1) Structure of Signaling Information
[0168] As the structure of the signaling information, either one
structure of the structure of the signaling information according
to Solution 1 (FIG. 3) and the structure of the signaling
information according to Solution 2 (FIG. 6) is used.
(2) Specific Application Example
[0169] (2-1) Transition from Basic Service to Hybrid Service
[0170] FIG. 9 is a diagram for explaining a scenario in the case
where a transition from the basic service to the hybrid service is
made.
[0171] In FIG. 9, as in FIG. 4, the transmitter of the broadcasting
station is transmitting, through a broadcast wave of the digital
broadcasting using the IP transmission system, components and
signaling information configuring each service in a BBP stream.
Further, in the hybrid service, at the receiver, streams
transmitted by communication are also acquired, and hence a stream
of audio (A2: Audio2) is being transmitted from the streaming
server. Further, the signaling information is being transmitted
from the signaling server.
[0172] As shown in FIG. 9, at the receiver placed in each house or
the like, the particular service selected by the user is a basic
service, and hence video and audio streams transmitted by
broadcasting are acquired by using signaling information
transmitted by broadcasting. With this, at the receiver, a picture
of a program corresponding to the basic service is displayed and at
the same time sound synchronized with that picture is output
(Procedure 1). Note that this Procedure 1 corresponds to, in the
case of Solution 1, the procedures in the channel selection
scenario of FIG. 4 and corresponds to, in the case of Solution 2,
the procedures in the channel selection scenario of FIG. 7.
[0173] At the receiver, the signaling information (SCS) transmitted
by broadcasting is acquired according to the SCS Bootstrap
information of the SCD, and the contents of the
SignalingOverinternet element described in the SPD are constantly
monitored (Procedure 2). Here, in the SignalingOverinternet
element, the hybrid attribute and url attribute are described as
the descendants thereof. In the hybrid attribute, "basic" is
specified in the case of the basic service and "hybrid" is
specified in the case of the hybrid service. In the url attribute,
the URL of the signaling server, for example, is specified as the
acquisition source of the signaling information.
[0174] That is, if the service selected by the user is a basic
service, "basic" is specified in the hybrid attribute. However, if
this service is transitioned from the basic service to the hybrid
service, the value of the hybrid attribute of the
SignalingOverinternet element is changed from "basic" to "hybrid."
In this case, the receiver accesses the signaling server via the
Internet according to the URL specified as the url attribute of the
SignalingOverinternet element and acquires the signaling
information (Procedures 3 and 4).
[0175] Then, at the receiver, the video component transmitted by
broadcasting and the audio component transmitted by communication
are acquired by using the signaling information transmitted by
communication. With this, at the receiver, a picture of a program
corresponding to the hybrid service is displayed and at the same
time sound synchronized with that picture is output (Procedures 5
and 6). Note that Procedures 4 to 6 corresponds to, in the case of
Solution 1, the procedures in the channel selection scenario of
FIG. 5 and corresponds to, in the case of Solution 2, the
procedures in the channel selection scenario of FIG. 8.
[0176] As described above, in the case where a transition from the
basic service to the hybrid service is made, in the basic service
after transition, the video and audio components are delivered only
by broadcasting, and hence the information relating to those
components is being described in the signaling information (SCS)
transmitted by broadcasting. Therefore, the receiver acquires the
video and audio components (streams) on the basis of the signaling
information transmitted by broadcasting. At this time, only the
information relating to the components being transmitted by
broadcasting is described in this signaling information, and hence
the increase in data size can be suppressed.
[0177] On the other hand, in the hybrid service after transition,
the video and audio components are being delivered by broadcasting
and communication, and hence the information relating to those
streams is described in the signaling information transmitted by
communication. Therefore, the receiver acquires the video and audio
components (streams) on the basis of the signaling information
transmitted by communication. At this time, in this signaling
information, the information items relating to the components
transmitted by broadcasting and communication are both described.
In the case of the signaling information transmitted by
communication, there is less need to worry about the data size in
comparison with the signaling information transmitted by
broadcasting. Thus, this is not an issue. By the way, even in the
case where the transition is made to the hybrid service, if the
user wishes to view the basic service, the receiver corresponds
only to the basic service, or the receiver is not connected to the
Internet, the reception of the basic service is still continued
without referring to the SignalingOverinternet element.
[0178] (2-2) Change of Stream Acquisition Source in Hybrid
Service
[0179] FIG. 10 is a diagram for explaining a scenario in the case
where the acquisition source of the streams is changed in the
hybrid service.
[0180] In FIG. 10, as in FIG. 4, the transmitter of the
broadcasting station is transmitting, through a broadcast wave of
the digital broadcasting using the IP transmission system,
components and signaling information configuring each service in a
BBP stream. Further, in the hybrid service, at the receiver,
streams transmitted by communication are also acquired, and hence
an audio stream is being transmitted from the streaming server. It
should be noted that, in this hybrid service, different types of
audio of programs are provided, and hence there are provided a
streaming server (Streaming Server1) that provides audio 2 (A2) and
a streaming server (Streaming Server2) that provides audio 3 (A3).
Further, signaling information (e.g., ZIP file including USD and
the like) is being transmitted from the signaling server.
[0181] As shown in FIG. 10, at the receiver placed in each house or
the like, using the signaling information transmitted by
communication, a video component transmitted by broadcasting and an
audio-2 component transmitted by communication from the streaming
server (Streaming Server1) are acquired. With this, at the
receiver, a picture of a program corresponding to the hybrid
service is displayed and at the same time sound corresponding to
the audio 2 synchronized with that picture is output (Procedure 1).
Note that this Procedure 1 corresponds to, in the case of Solution
1, the procedures in the channel selection scenario of FIG. 5 and
corresponds to, in the case of Solution 2, the procedures in the
channel selection scenario of FIG. 8.
[0182] Further, in the SignalingOverinternet element of the SPD in
the signaling information transmitted by communication, version
attribute and minUpdatePeriod attribute are described other than
the hybrid attribute and url attribute. In the version attribute,
version information of the signaling information is specified. In
the minUpdatePeriod attribute, an update interval of the signaling
information is specified. Therefore, the receiver accesses the
signaling server at the update interval of the signaling
information according to the URL of the signaling server, and
acquires the signaling information (Procedure 2). Then, the
receiver checks the attribute values of the hybrid attribute and
version attribute of the SPD, to thereby check whether or not the
contents of the signaling information have been updated (Procedure
3).
[0183] In the scenario of FIG. 10, at a timing when the program of
the hybrid service is changed, the contents of the signaling
information are updated and the audio is changed from the audio 2
to the audio 3. Therefore, at the receiver, using the updated
signaling information, a video component transmitted by
broadcasting and a component of the audio 3 transmitted by
communication from the streaming server (Streaming Server2) are
acquired (Procedures 4 and 5). With this, at the receiver, a
picture of a program corresponding to the hybrid service is
displayed and at the same time sound corresponding to the audio 3
synchronized with that picture is output. Note that those
Procedures 4 and 5 correspond to, in the case of Solution 1, the
procedures in the channel selection scenario of FIG. 5 and
correspond to, in the case of Solution 2, the procedures in the
channel selection scenario of FIG. 8.
[0184] FIG. 11 is a diagram for explaining another scenario in the
case where the acquisition source of the streams is changed in the
hybrid service.
[0185] FIG. 11 shows, as in FIG. 10, a scenario in the case where
the acquisition source of the streams is changed in the hybrid
service. They are different from each other in that the update of
the signaling information is checked using the signaling
information transmitted by communication in FIG. 10 while the
update of the signaling information is checked using the signaling
information transmitted by broadcasting in FIG. 11.
[0186] Specifically, as shown in FIG. 11, at the receiver placed in
each house or the like, using the signaling information transmitted
by communication, a video component transmitted by broadcasting and
an audio-2 component transmitted by communication from the
streaming server (Streaming Server1) are acquired. With this, at
the receiver, a picture of a program corresponding to the hybrid
service is displayed and at the same time sound corresponding to
the audio 2 synchronized with that picture is output (Procedure 1).
Note that this Procedure 1 corresponds to, in the case of Solution
1, the procedures in the channel selection scenario of FIG. 5 and
corresponds to, in the case of Solution 2, the procedures in the
channel selection scenario of FIG. 8.
[0187] The receiver is capable of acquiring the signaling
information (SCS) transmitted by broadcasting according to the SCS
Bootstrap information of the SCD, and the contents of the
SignalingOverinternet element described in the SPD are constantly
monitored (Procedure 2). Then, the receiver checks the attribute
values of the hybrid attribute and version attribute of the SPD, to
thereby check whether or not the contents of the signaling
information transmitted by communication have been updated
(Procedure 3). If the signaling information transmitted by
communication has been updated, the receiver accesses the signaling
server according to the URL of the signaling server, and acquires
the updated signaling information (Procedure 4).
[0188] In the scenario of FIG. 11, as in the scenario of FIG. 10,
at a timing when the program of the hybrid service is changed, the
contents of the signaling information are updated and the audio is
changed from the audio 2 to the audio 3, and hence, at the
receiver, a video component transmitted by broadcasting and a
component of the audio 3 transmitted by communication from the
streaming server (Streaming Server2) are acquired by using the
updated signaling information (Procedures 5 and 6). With this, at
the receiver, a picture of a program corresponding to the hybrid
service is displayed and at the same time sound corresponding to
the audio 3 synchronized with that picture is output. Note that
those Procedures 5 and 6 correspond to, in the case of Solution 1,
the procedures in the channel selection scenario of FIG. 5 and
correspond to, in the case of Solution 2, the procedures in the
channel selection scenario of FIG. 8.
[0189] As described above, in the case where the acquisition source
of the streams is changed in the hybrid service, the video and
audio components (streams) are delivered by broadcasting and
communication, and hence the information relating to those
components is described in the signaling information transmitted by
communication. Therefore, the receiver acquires the video and audio
components (streams) on the basis of the signaling information
transmitted by communication. At this time, in this signaling
information, the information items relating to the components
transmitted by broadcasting and communication are both described.
In the case of the signaling information transmitted by
communication, there is less need to worry about the data size in
comparison with the signaling information transmitted by
broadcasting. Thus, this is not an issue.
[0190] (2-3) Transition from Hybrid Service to Basic Service
[0191] FIG. 12 is a diagram for explaining a scenario in the case
where a transition from the hybrid service to the basic service is
made.
[0192] In FIG. 12, as in FIG. 4, the transmitter of the
broadcasting station is transmitting, through a broadcast wave of
the digital broadcasting using the IP transmission system,
components and signaling information configuring each service in a
BBP stream. Further, in the hybrid service, at the receiver,
streams transmitted by communication are also acquired, and hence
an audio stream is being transmitted from the streaming server. It
should be noted that, in this hybrid service, audio (A2) and audio
3 (A3) are provided, and hence a plurality of streaming servers
(Streaming Servers 1 and 2) are provided. Further, signaling
information (e.g., ZIP file including USD and the like) is being
transmitted from the signaling server.
[0193] As shown in FIG. 12, at the receiver placed in each house or
the like, the particular service selected by the user is a hybrid
service, and hence a video component transmitted by broadcasting
and a component of the audio 3 transmitted by communication from
the streaming server (Streaming Server2) is acquired by using the
signaling information transmitted by communication. With this, at
the receiver, a picture of a program corresponding to the hybrid
service is displayed and at the same time sound corresponding to
the audio 3 synchronized with that picture is output (Procedure 1).
Note that this Procedure 1 corresponds to, in the case of Solution
1, the procedures in the channel selection scenario of FIG. 5 and
corresponds to, in the case of Solution 2, the procedures in the
channel selection scenario of FIG. 8.
[0194] Further, in the signaling information transmitted by
communication, other than the hybrid attribute, the version
attribute, and the url attribute, the minUpdatePeriod attribute and
the endtime attribute are described in the SignalingOverinternet
element of the SPD. In the minUpdatePeriod attribute, an update
interval of the signaling information is specified. In the endtime
attribute, the point of time of the end of the communication
acquisition of the signaling information is specified. Therefore,
the receiver accesses the signaling server at the update interval
of the signaling information according to the URL of the signaling
server, and acquires the signaling information (Procedure 2). Then,
the receiver checks the attribute value of the endtime attribute of
the SPD, to thereby check whether or not to terminate the
communication acquisition of the signaling information (Procedure
3).
[0195] When the communication acquisition of the signaling
information is terminated, at the receiver, the signaling
information (SCS) transmitted by broadcasting is acquired according
to the SCS Bootstrap information of the SCD (Procedure 4). That is,
the signaling information is changed from the communication
acquisition to the broadcasting acquisition and a transition from
the hybrid service to the basic service is made.
[0196] Then, at the receiver, using signaling information (SCS)
transmitted by broadcasting, the video and audio components
transmitted by broadcasting are acquired (Procedures 5 and 6). With
this, at the receiver, a picture of a program corresponding to the
basic service is displayed and at the same time sound synchronized
with that picture is output. Note that those Procedures 5 and 6
correspond to, in the case of Solution 1, the procedures in the
channel selection scenario of FIG. 4 and correspond to, in the case
of Solution 2, the procedures in the channel selection scenario of
FIG. 7.
[0197] FIG. 13 is a diagram for explaining another scenario in the
case where a transition from the hybrid service to the basic
service is made.
[0198] FIG. 13 shows, as in FIG. 12, the scenario in the case where
a transition from the hybrid service to the basic service is made.
They are different in that the end of the communication acquisition
of the signaling information is checked using the signaling
information transmitted by communication in FIG. 12 while the end
of the communication acquisition of the signaling information is
checked using the signaling information transmitted by broadcasting
in FIG. 13.
[0199] Specifically, at the receiver placed in each house or the
like, the particular service selected by the user is a hybrid
service, and hence a video component transmitted by broadcasting
and a component of the audio 3 transmitted by communication from
the streaming server (Streaming Server2) are acquired by using the
signaling information transmitted by communication. With this, at
the receiver, a picture of a program corresponding to the hybrid
service is displayed and at the same time sound corresponding to
the audio 3 synchronized with that picture is output (Procedure 1).
Note that this Procedure 1 corresponds to, in the case of Solution
1, the procedures in the channel selection scenario of FIG. 5 and
corresponds to, in the case of Solution 2, the procedures in the
channel selection scenario of FIG. 8.
[0200] The receiver is capable of acquiring the signaling
information (SCS) transmitted by broadcasting according to the SCS
Bootstrap information of the SCD, and the contents of the
SignalingOverinternet element described in the SPD are constantly
monitored (Procedure 2). Then, the receiver checks the attribute
value of the endtime attribute of the SPD, to thereby check whether
or not to terminate the communication acquisition of the signaling
information (Procedure 3). At the receiver, if the communication
acquisition of the signaling information is terminated, the
signaling information (SCS) transmitted by broadcasting is acquired
according to the SCS Bootstrap information of the SCD (Procedure
4). That is, the signaling information is changed from the
communication acquisition to the broadcasting acquisition and a
transition from the hybrid service to the basic service is
made.
[0201] Then, at the receiver, the video and audio components
transmitted by broadcasting are acquired by using signaling
information (SCS) transmitted by broadcasting (Procedures 5 and 6).
With this, at the receiver, a picture of a program corresponding to
the basic service is displayed and at the same time sound
synchronized with that picture is output. Note that those
Procedures 5 and 6 correspond to, in the case of Solution 1, the
procedures in the channel selection scenario of FIG. 4 and
correspond to, in the case of Solution 2, the procedures in the
channel selection scenario of FIG. 7.
[0202] As described above, in the case where a transition from the
hybrid service to the basic service is made, in the hybrid service
before transition, the video and audio components are being
delivered by broadcasting and communication, and hence the
information relating to those components is described in the
signaling information transmitted by communication. Therefore, the
receiver acquires the video and audio components (streams) on the
basis of the signaling information transmitted by communication. At
this time, in this signaling information, the information items
relating to the components transmitted by broadcasting and
communication are both described. In the case of the signaling
information transmitted by communication, there is less need to
worry about the data size in comparison with the signaling
information transmitted by broadcasting. Thus, this is not an
issue.
[0203] On the other hand, in the basic service after transition,
the video and audio components are delivered only by broadcasting,
and hence the information relating to those components is being
described in the signaling information (SCS) transmitted by
broadcasting. Therefore, the receiver acquires the video and audio
components (streams) on the basis of the signaling information
transmitted by broadcasting. At this time, only the information
relating to the components being transmitted by broadcasting is
described in this signaling information, and hence the increase in
data size can be suppressed.
[0204] <5. Syntax>
[0205] (1) Syntax of SCD
[0206] FIG. 14 is a diagram showing a syntax of the SCD. Note that
the SCD of FIG. 14 is common to Solution 1 and Solution 2.
[0207] The SCD is described with a markup language, for example, an
XML. Note that, in FIG. 14, "@" is added to the attribute out of
the element and the attribute. Further, the indented elements and
attributes are those specified with respect to upper-level elements
thereof. The same relationship between them applies to other
syntaxes to be described later.
[0208] As shown in FIG. 14, the Scd element is an upper-level
element of majorProtocolversion attribute, minorProtocolversion
attribute, RFchannelId attribute, name attribute, Tuning_RF
element, and BBPStream element.
[0209] In the majorProtocolversion attribute and the
minorProtocolversion attribute, protocol version information is
specified. In the RFchannelId attribute, an RF channel ID of a
broadcasting station in units of physical channels is specified. In
the name attribute, the name of the broadcasting station in units
of physical channels is specified.
[0210] In the Tuning_RF element, information relating to the
channel selection is specified. The Tuning RF element is an
upper-level element of frequency attribute and PreambleL1Pre
attribute. In the frequency attribute, a frequency in selecting a
predetermined bandwidth is specified. In the PreambleL1Pre
attribute, control information of a physical layer is
specified.
[0211] In the BBPStream element, information relating to one or
more BBP streams is specified. The BBPStream element is an
upper-level element of bbpStreamId attribute, payloadType
attribute, name attribute, ESGBootstrap element,
ClockReferenceInformation element, Tuning_BBPS element, and Service
element.
[0212] In the bbpStreamId attribute, a BBP stream ID is specified.
If a plurality of BBP streams are provided, they are identified by
bbpStreamId attributes. In the payloadType attribute, a payload
type of the BBP stream is specified. For example, "ipv4," "ipv6,"
or "ts" is specified as this payload type. "Ipv4" indicates IPv4
(Internet Protocol version 4). "Ipv6" indicates IPv6 (Internet
Protocol Version 6). "Ts" indicates a TS (Transport Stream). In the
name attribute, the name of the BBP stream is specified.
[0213] In the ESGBootstrap element, information on access to the
ESG is specified. The ESGBootstrap element is an upper-level
element of an ESGProvider element. In the ESGProvider element,
information relating to the ESG is specified for each ESG provider.
The ESGProvider element is an upper-level element of providerName
attribute, ESGBroadcastLocation element, and ESGBroadbandLocation
element.
[0214] In the providerName attribute, the name of the ESG provider
is specified. If an ESG is transmitted by broadcasting, the
ESGBroadcastLocation element specifies an ESG service, using
RFchannelId attribute, BBPStreamId attribute, and ESGServiceId
attribute (triplet). In the RFchannelId attribute, an RF channel ID
of a broadcasting station that transmits the ESG service is
specified. In the BBPStreamId attribute, an BBP stream ID of a BBP
stream that transmits the ESG service is specified. In the
ESGServiceId attribute, a service ID of the ESG service is
specified.
[0215] If an ESG is transmitted by communication, the
ESGBroadbandLocation element specifies a URL for accessing a file
of that ESG, using a ESG url attribute.
[0216] In the ClockReferenceInformation element, information
relating to time information (e.g., NTP) is specified. The
ClockReferenceInformation element is an upper-level element of
sourceIPAddress attribute, destinationIPAddress attribute, portNum
attribute, and clockReferenceFormat attribute.
[0217] In the sourceIPAddress attribute and the
destinationIPAddress attribute, IP addresses of source and
destination for transmitting the time information are specified. In
the portNum attribute, a port number for transmitting the time
information is specified. In the clockReferenceFormat attribute,
type information of the time information is specified. For example,
"NTPnormal" or "NTP27M" is specified as this type information.
"NTPnormal" indicates a normal NTP. Further, "NTP27M" indicates
adaptability for a reference clock of 27 MHz of a PCR (Program
Clock Reference).
[0218] In the Tuning_BBPS element, the information relating to the
channel selection for each BBP stream is specified. The Tuning_BBPS
element is an upper-level element of plpId attribute and
PreambleL1post element. In the plpId attribute, an PLP ID for
identifying the BBP stream is specified. Note that the PLP ID
corresponds to the BBP stream ID. In the PreambleL1post element,
control information of the physical layer is specified.
[0219] In the Service element, information relating to one or more
services is specified. The Service element is an upper-level
element of serviceId attribute, serviceType attribute, hidden
attribute, hiddenGuide attribute, shortName attribute, longName
attribute, accesControl attribute, SourceOrigin element, SCS
bootstrap element, SignalingOverinternet element, and Associated
Service element.
[0220] In the serviceId attribute, the service ID is specified. If
a plurality of services are provided, they are identified by
serviceId attributes. In the serviceType attribute, type
information of the service is specified. For example, "continuous,"
"scripted," or "esg" is specified as this type information.
"Continuous" indicates a video/audio service. "Scripted" indicates
an NRT service. "Esg" indicates an ESG service.
[0221] In the hidden attribute and the hiddenGuide attribute,
whether or not the service identified by the service ID is a hidden
service is specified. If "on" is specified as those attribute
values, that service is set not to be displayed.
[0222] Further, if "off" is specified as those attribute values,
that service is displayed. For example, if "on" is specified as the
hidden attribute, that service is set not to be selected by an
operation of a remote controller. Further, for example, if "on" is
specified as the hiddenGuide attribute, that service is set not to
be displayed in the ESG.
[0223] In the shortName attribute and the longName attribute, the
service identified by the name of the service ID is specified. It
should be noted that, in the shortName attribute, the name of the
name of the service has to be specified with at most seven letters.
In the accesControl attribute, whether or not the service
identified by the service ID has been encrypted is specified. If
"on" is specified as the accesControl attribute, it indicates that
that service has been encrypted and, if "off" is specified, it
indicates that that service has not been encrypted.
[0224] In the SourceOrigin element, information for identifying the
service is specified. The SourceOrigin element is an upper-level
element of country attribute, original RFchannelId attribute,
bbpStreamId attribute, and serviceId attribute. In the country
attribute, a country code is specified. In the original RFchannelId
attribute, an original RF channel ID is specified. The original RF
channel ID is an ID for identifying the broadcasting network and
the same value is used therefor also when that service is
re-transmitted. In the bbpStreamId attribute, the BBP stream ID is
specified. In the serviceId attribute, the service ID is specified.
That is, using the country code, original RF channel ID, BBP stream
ID, and service ID, a specific ID can be assigned to each
service.
[0225] In an SCS Bootstrap element, information on access to the
service is specified. The SCS Bootstrap element is an upper-level
element of hybrid attribute, sourceIPAddress attribute,
destinationIPAddress attribute, portNum attribute, and tsi
attribute. In the hybrid attribute, information indicating whether
or not the signaling information adapted for the hybrid service is
specified. For example, if "basic" is specified as the hybrid
attribute, it indicates adaptability for the basic service and, if
"hybrid" is specified as the hybrid attribute, it indicates
adaptability for the hybrid service. In the sourceIPAddress
attribute and the destinationIPAddress attribute, IP addresses of
source and destination for transmitting the time information are
specified. In the portNum attribute, the port number for
transmitting the SCS is specified. In the tsi attribute, the TSI in
the FLUTE session that transmits the SCS is specified.
[0226] In the SignalingOverinternet element, information relating
to the signaling information transmitted by communication is
specified. The SignalingOverinternet element is an upper-level
element of hybrid attribute and url attribute. In the hybrid
attribute, information indicating whether or not the signaling
information adapted for the hybrid service is specified. For
example, "basic" is specified as the hybrid attribute, it indicates
adaptability for the basic service and, if "hybrid" is specified as
the hybrid attribute, it indicates adaptability for the hybrid
service. In the url attribute, a URL indicating an acquisition
source of the signaling information is specified. For example, in
the url attribute, a URL of the signaling server is specified.
[0227] In the Associated Service element, information relating to
an associated slave service is specified. The Associated Service
element is an upper-level element of RFchannelId attribute,
bbpStreamId attribute, and serviceId attribute. In the RFchannelId
attribute, an RF channel ID of the associated slave service is
specified. In the bbpStreamId attribute, a BBP stream ID of the
associated slave service is specified. In the serviceId attribute,
a service ID of the associated slave service is specified.
[0228] In FIG. 14, regarding cardinality, when "1" is specified,
only one element or attribute is necessarily specified. When "0 . .
. 1" is specified, it is optional whether to specify an element or
attribute. When "1 . . . n" is specified, one or more elements or
attributes are specified. When "0 . . . n" is specified, it is
optional whether to specify one or more elements or attributes. The
same meanings of the cardinality as described above apply to other
syntaxes to be described later.
[0229] (2) Syntax of SPD
[0230] (Syntax of SPD According to Solution 1)
[0231] FIG. 15 is a diagram showing a syntax of the SPD according
to Solution 1. That is, the SPD of FIG. 15 can be used in the
channel selection scenario of the basic service of FIG. 4 or the
channel selection scenario of the hybrid service of FIG. 5 that is
described above. Note that the SCD is described with a markup
language, for example, an XML.
[0232] As shown in FIG. 15, an Spd element is an upper-level
element of serviceId attribute, spindicator attribute, Protocol
VersionDescriptor element, NRTServiceDescriptor element,
CapabilityDescriptor element, IconDescriptor element,
ISO639LanguageDescriptor element, ReceiverTargetingDescriptor
element, AssociatedServiceDescriptor element,
ContentAdvisoryDescriptor element, and SignalingOverinternet
element.
[0233] In the serviceId attribute, the service ID is specified. In
the spindicator attribute, whether or not each service identified
by the service ID has been encrypted is specified. If "on" is
specified as the spindicator attribute, it indicates that that
service has been encrypted. If "off" is specified, it indicates
that that service has not been encrypted.
[0234] In the Protocol VersionDescriptor element, information for
indicating what kind of service the data service is specified. In
the NRTServiceDescriptor element, information relating to the NRT
service is specified. In the CapabilityDescriptor element,
information relating to a function (capability) required of the
receiver that receives the provided NRT service is specified.
[0235] In the IconDescriptor element, information indicating an
acquisition source of an icon used in the NRT service is specified.
In the ISO639LanguageDescriptor element, a language code of the NRT
service is specified. In the ReceiverTargetingDescriptor element,
target information on the NRT service is specified.
[0236] In the AssociatedServiceDescriptor element, information
relating to the associated slave service is specified. In the
ContentAdvisoryDescriptor element, information relating to the
rating region is specified.
[0237] In the SignalingOverinternet element, information relating
to the signaling information transmitted by communication is
specified. The SignalingOverinternet element is an upper-level
element of hybrid attribute, version attribute, and url attribute.
In the hybrid attribute, information indicating whether or not it
is the signaling information adapted for the hybrid service is
specified. For example, "basic" is specified as the hybrid
attribute, it indicates adaptability for the basic service and, if
"hybrid" is specified as the hybrid attribute, it indicates
adaptability for the hybrid service. In the version attribute,
version information of the signaling information is specified. In
the url attribute, a
[0238] URL indicating an acquisition source of the signaling
information is specified. For example, in the url attribute, a URL
of the signaling server is specified.
[0239] In the SPD, the various parameters at the service level are
defined by the above-mentioned Descriptor elements. Note that, in
FIG. 15, the Protocol VersionDescriptor element,
NRTServiceDescriptor element, CapabilityDescriptor element,
IconDescriptor element, ISO639LanguageDescriptor element, and
ReceiverTargetingDescriptor element are those defined for the NRT
service.
[0240] (Syntax of SPD According to Solution 2)
[0241] FIG. 16 is a diagram showing a syntax of the SPD according
to Solution 2. That is, the SPD of FIG. 16 can be used in the
channel selection scenario of the basic service of FIG. 7 or the
channel selection scenario of the hybrid service of FIG. 8 that is
described above.
[0242] As shown in FIG. 16, an Spd element is an upper-level
element of serviceId attribute, spindicator attribute, Protocol
VersionDescriptor element, NRTServiceDescriptor element,
CapabilityDescriptor element, IconDescriptor element,
ISO639LanguageDescriptor element, ReceiverTargetingDescriptor
element, AssociatedServiceDescriptor element,
ContentAdvisoryDescriptor element, SignalingOverinternet element,
and Component element.
[0243] In the SPD of FIG. 16, the various parameters at the service
level and the component level are defined. The parameter at the
service level is the same as the SPD of FIG. 15.
[0244] For example, in the SignalingOverinternet element,
information relating to the signaling information transmitted by
communication is specified. The SignalingOverinternet element is an
upper-level element of hybrid attribute, version attribute, and url
attribute. In the hybrid attribute, information indicating whether
or not it is the signaling information adapted for the hybrid
service is specified. For example, if "basic" is specified as the
hybrid attribute, it indicates adaptability for the basic service
and, if "hybrid" is specified as the hybrid attribute, it indicates
adaptability for the hybrid service. In the version attribute,
version information of the signaling information is specified. In
the url attribute, a URL indicating an acquisition source of the
signaling information is specified. For example, in the url
attribute, a URL of the signaling server is specified.
[0245] Note that the elements other than the SignalingOverinternet
element are also the same as the SPD of FIG. 15, and hence
duplicated descriptions thereof will be omitted. In the SPD of In
FIG. 16, the various parameters at the component level are defined
by the Component element.
[0246] The Component element is an upper-level element of
componentId attribute, componentType attribute, componentEncription
attribute, ComponentLocation element, TargetedDeviceDescriptor
element, ContentAdvisoryDescriptor element, VideoParameters
element, AudioParameters element, and CaptionParameters
element.
[0247] In the componentId attribute, a component ID used for
associating a component with another component in another table is
specified. In the componentType attribute, type information of the
component is specified. In the componentEncription attribute,
whether or not each component identified by the component ID has
been encrypted is specified. If "on" is specified as the
componentEncription attribute, it indicates that that component has
been encrypted. If "off" is specified, it indicates that that
component has not been encrypted.
[0248] In the ComponentLocation element, component location
information is specified. Note that the detailed contents of the
ComponentLocation element will be described later with reference to
FIG. 17. In the TargetedDeviceDescriptor element, information
relating to the display of the device as a target is specified. In
the ContentAdvisoryDescriptor element, rating information in units
of components is specified.
[0249] In the VideoParameters element, video parameters are
specified. The VideoParameters element is an upper-level element of
AVCVideoDescriptor element and HEVCVideoDescriptor element. That
is, if AVC (Advanced Video Coding) is used as the encoding method
for the video data, the AVCVideoDescriptor element is specified,
and, if HEVC (High Efficiency Video Coding) is used as the encoding
method for the video data, the HEVCVideoDescriptor element is
specified. Note that the AVC and
[0250] HEVC are examples of the encoding method for the video data,
and, if another encoding method is used, the corresponding Video
Descriptor element will be specified.
[0251] In the AudioParameters element, audio parameters are
specified. The AudioParameters element is an upper-level element of
MPEG4AACAudioDescriptor element and AC3AudioDescriptor element.
That is, if MPEG4AAC (Advanced Audio Coding) is used as the
encoding method for the audio data, the MPEG4AACAudioDescriptor
element is specified, and, if AC3 (Audio Code number 3) is used as
the encoding method for the audio data, the AC3AudioDescriptor
element is specified. Note that the MPEG4AAC and AC3 are examples
of the encoding method for the audio data, and, if another encoding
method is used, the corresponding AudioDescriptor element will be
specified.
[0252] In the CaptionParameters element, closed caption parameters
are specified.
[0253] (Detailed Contents of ComponentLocation element)
[0254] FIG. 17 is a diagram showing detailed contents of the
ComponentLocation element of FIG. 16.
[0255] In the ComponentLocation element, component location
information is specified. The ComponentLocation element is an
upper-level element of portNumber attribute, tsi attribute,
startToi attribute, and endToi attribute. In the portNumber
attribute, a port number of a targeted component is specified. In
the tsi attribute, an TSI of a FLUTE session in which a targeted
component is transmitted is specified. In the startToi attribute,
the start value of the TOI in the case where the TOI is changed in
a time sequence is specified. In the endToi attribute, the end
value of the TOI in the case where the TOI is changed in a time
sequence is specified. That is, by specifying the startToi
attribute and the endToi attribute, the value thereof is
sequentially incremented from the start value to the end value of
the TOI.
[0256] (Syntax of SPD Common to Solutions 1 and 2)
[0257] FIG. 18 is a diagram showing a syntax of the SPD common to
Solutions 1 and 2. That is, the SPD of FIG. 18 can be used in the
scenario of FIG. 10, the scenario of FIG. 12, or the like that is
described above.
[0258] As shown in FIG. 18, the Spd element is an upper-level
element of serviceId attribute, spindicator attribute, Protocol
VersionDescriptor element, NRTServiceDescriptor element,
CapabilityDescriptor element, IconDescriptor element,
ISO639LanguageDescriptor element, ReceiverTargetingDescriptor
element, AssociatedServiceDescriptor element,
ContentAdvisoryDescriptor element, and SignalingOverinternet
element.
[0259] Although the various parameters at the service level are
defined in the SPD of FIG. 18, descriptions of the parameters at
the service level, which have the same contents as the SPD of FIG.
15 for Solution 1, will be duplicated. Therefore, descriptions
thereof will be appropriately omitted. That is, the SPD of FIG. 18
is different from the contents of the SignalingOverinternet element
in comparison with the SPD of FIG. 15. Further, the description
portion before the Component element in the SPD of FIG. 16 for
Solution 2 will be replaced by the syntax of the SPD of FIG. 18.
That is, the SPD of FIG. 18 is different from the contents of the
SignalingOverinternet element in comparison with the SPD of FIG.
16.
[0260] In the SignalingOverinternet element, information relating
to the signaling information transmitted by communication is
specified. The SignalingOverinternet element is an upper-level
element of hybrid attribute, version attribute, url attribute,
minUpdatePeriod attribute, and endtime attribute. In the hybrid
attribute, information indicating whether or not it is the
signaling information adapted for the hybrid service is specified.
For example, if "basic" is specified as the hybrid attribute, it
indicates adaptability for the basic service and, if "hybrid" is
specified as the hybrid attribute, it indicates adaptability for
the hybrid service.
[0261] In the version attribute, version information of the
signaling information is specified. In the url attribute, a URL
indicating an acquisition source of the signaling information is
specified. For example, in the url attribute, a URL of the
signaling server is specified. In the minUpdatePeriod attribute,
information indicating an update interval of the signaling
information is specified. In the endtime attribute, information
indicating the point of time of the end of the communication
acquisition of the signaling information is specified.
[0262] Next, referring to FIGS. 19 to 31, a detailed structure of
the Descriptor element described in the SPD will be described. Note
that each Descriptor element is described with a markup language,
for example, an XML. Further, in FIGS. 19 to 31, "@" is added to
the attribute out of the element and the attribute. Further, the
indented elements and attributes are those specified with respect
to upper-level elements thereof.
[0263] (Protocol Version Descriptor)
[0264] FIG. 19 is a diagram showing a syntax of a Protocol Version
Descriptor.
[0265] In a Protocol Version Descriptor element, information for
indicating what kind of service the data service is specified. The
Protocol Version Descriptor element is an upper-level element of
protocolIdentifier attribute, majorProtocolVersion attribute, and
minorProtocolVersion attribute.
[0266] In the protocolIdentifier attribute, type information of the
format of the data service is specified. For example, "A/90," "NRT"
is specified as this type information. "A/90" indicates a method of
transmitting universal data. Further, "NRT" indicates a method of
transmitting NRT
[0267] (Non-Real Time).
[0268] In the majorProtocolVersion attribute and the
minorProtocolVersion attribute, the version of the data service is
specified. In the majorProtocolVersion attribute, a major version
is specified. In the minorProtocolVersion attribute, a minor
version is specified.
[0269] (NRT Service Descriptor)
[0270] FIG. 20 is a diagram showing a syntax of an NRT Service
Descriptor.
[0271] In the NRTServiceDescriptor element, information relating to
the NRT service is specified. The NRTServiceDescriptor element is
an upper-level element of ConsumptionModel attribute, autoUpdate
attribute, storageReservarion attribute, and defaultContentSize
attribute.
[0272] In the ConsumptionModel attribute, a transmission mode of
the NRT service is specified. For example, "B&D," "push,"
"portal," "triggered" is specified as this transmission mode.
"B&D" is the abbreviation of Browse and Download and a mode for
downloading file data of an NRT content selected by the user.
"Push" is a mode for providing a contracted NRT service as a push
type. "Portal" is a mode for transmitting and immediately
displaying an HTML file or the like. "Triggered" is a mode for
providing the application.
[0273] In the autoUpdate attribute, whether or not the NRT service
has been automatically updated is specified. If "on" is specified
as the autoUpdate attribute, it indicates that that NRT service is
automatically updated. If "off" is specified, it indicates that NRT
service is not automatically updated. In the storageReservarion
attribute, a necessary storage volume is specified. The size per
NRT content is specified in the defaultContentSize attribute.
[0274] (Capability Descriptor)
[0275] FIG. 21 is a diagram showing a syntax of a Capability
Descriptor.
[0276] In the Capability Descriptor element, information relating
to the function (capability) required of the receiver that receives
the provided NRT service is specified. The Capability Descriptor
element is an upper-level element of IndivisualCapabilityCodes
element, IndivisualCapabilityString element, and CapabilityOrSets
element.
[0277] The IndivisualCapabilityCodes element is an upper-level
element of essentialIndicator attribute, capabilityCode attribute,
and formatIdentifier attribute. In the essentialIndicator
attribute, information indicating whether or not the capability is
essential is specified. In the capabilityCode attribute, the code
of the capability determined in advance is specified. That is,
whether or not the capability specified by the code of the
capability is essential is specified by the essentialIndicator
attribute and capabilityCode attribute. In the formatIdentifier
attribute, if a code of the capability is arbitrarily specified, a
function (capability) that should be assessed is specified.
[0278] The IndivisualCapabilityString element is an upper-level
element of essentialIndicator attribute, capabilityCategoryCode
attribute, and capabilityString attribute. In the
essentialIndicator attribute, information indicating whether or not
the capability is essential is specified. In the
capabilityCategoryCode attribute, a code for each category of the
capability is specified. That is, by the essentialIndicator
attribute and the capabilityCategoryCode attribute, whether or not
the capability specified by the code for each category of the
capability is essential is specified. In the capabilityString
attribute, the function (capability) that should be assessed is
specified for each category of the capability.
[0279] The CapabilityOrSets element is specified in the case of
specifying the assessment for each code of the capability by the
above-mentioned IndivisualCapabilityCodes element and the
assessment for each code of the category of the capability by the
IndivisualCapabilityString element with OR logic. Therefore,
although the CapabilityOrSets element is an upper-level element of
essentialIndicator attribute, CapabilityCodesInSets element, and
CapabilityStringsInSets element, the essentialIndicator attribute
corresponds to the above-mentioned essentialIndicator
attribute.
[0280] Further, capabilityCode attribute and formatIdentifier
attribute in the CapabilityCodesInSets element correspond to the
capabilityCode attribute and the formatIdentifier attribute in the
above-mentioned IndivisualCapabilityCodes element, respectively. In
addition, capabilityCategoryCode attribute and capabilityString
attribute in the CapabilityStringsInSets element correspond to the
capabilityCategoryCode attribute and the capabilityString attribute
in the above-mentioned IndivisualCapabilityString element,
respectively.
[0281] (Icon Descriptor)
[0282] FIG. 22 is a diagram showing a syntax of an Icon
Descriptor.
[0283] In the IconDescriptor element, the information indicating
the acquisition source of the icon used in the NRT service is
specified. The IconDescriptor element is an upper-level element of
a content linkage attribute. In the content linkage attribute, a
URL indicating the acquisition source of the icon is specified.
[0284] (ISO-639 Language Descriptor)
[0285] FIG. 23 is a diagram showing a syntax of an ISO-639 Language
Descriptor.
[0286] In the ISO639LanguageDescriptor element, a language code of
the NRT service is specified. The ISO639LanguageDescriptor element
is an upper-level element of a languageCode attribute. In the
languageCode attribute, a language code defined by ISO 639 is
specified.
[0287] (Receiver Targeting Descriptor)
[0288] FIG. 24 is a diagram showing a syntax of a Receiver
Targeting Descriptor.
[0289] In the ReceiverTargetingDescriptor element, target
information on the NRT service is specified. The
ReceiverTargetingDescriptor element is an upper-level element of a
TargetEntry element. The TargetEntry element is an upper-level
element of geoLocation attribute, postalCode attribute, and
demographic category attribute.
[0290] In the geoLocation attribute, a geographical position as a
target of the NRT service is specified. In the postalCode
attribute, a postal code of a region as the target of the NRT
service is specified. In the demographic category attribute, a
category of users as targets of the NRT service is specified. For
example, "males," "females," or "Ages 12-17" is specified as this
category. "Males" indicates that the targets of the NRT service are
males. "Females" indicates that the targets of the NRT service are
females. "Ages 12-17" indicates that the targets of the NRT service
are 12 to 17 years old.
[0291] (Associated Service Descriptor)
[0292] FIG. 25 is a diagram showing a syntax of an Associated
Service Descriptor.
[0293] In the AssociatedServiceDescriptor element, the information
relating to the associated slave service is specified. The
AssociatedServiceDescriptor element is the upper-level element of
the RFchannelId attribute, BBPStreamId attribute, and serviceId
attribute. In the RFchannelId attribute, an RF channel ID is
specified. In the BBPStreamId attribute, a BBP stream ID is
specified. In the serviceId attribute, a service ID is specified.
That is, the associated slave service is specified by the
triplet.
[0294] (Content Advisory Descriptor)
[0295] FIG. 26 is a diagram showing a syntax of a Content Advisory
Descriptor.
[0296] In a Content Advisory Descriptor element, information
relating to the rating region is specified. The Content Advisory
Descriptor element is the upper-level element of the version
attribute and RatingRegion element. In the version attribute, RRT
version information is specified.
[0297] The RatingRegion element is an upper-level element of the
ratingRegionId attribute and RatingDimension element. In the
ratingRegionId attribute, a rating region ID is specified. The
RatingDimension element is an upper-level element of dimensionIndex
attribute, ratingValue attribute, and ratingTag attribute. Rating
information, for example, how to set the age limit is specified by
those attributes.
[0298] (AVC Video Descriptor)
[0299] FIG. 27 is a diagram showing a syntax of an AVC Video
Descriptor.
[0300] If AVC is used as the encoding method for the video data, in
the AVC Video Descriptor element, information relating to the
contents of a AVC codec is specified. An AVCVideo Descriptor
element is an upper-level element of profileIdc attribute,
constraintSet0 attribute, constraintSet1 attribute, constraintSet2
attribute, AVCCompatibleFlags attribute, levelIdc attribute,
stillPresent attribute, and 24HourPicture attribute.
[0301] Information relating to the contents of the AVC codec is
specified by those attributes.
[0302] (HEVC Video Descriptor)
[0303] FIG. 28 is a diagram showing a syntax of an HEVC Video
Descriptor.
[0304] If HEVC is used as the encoding method for the video data,
information relating to the contents of an HEVC codec is specified
in the HEVCVideoDescriptor element. The HEVCVideoDescriptor element
is an upper-level element of profileSpace attribute, tierFlag
attribute, profileIdc attribute, profileCompatibilityIndication
attribute, progressiveSourceFlag attribute, nonPackedConstraintFlag
attribute, frameOnlyConstraintFlag attribute, levelIdc attribute,
temporalLayerSubsetFlag attribute, stillPresent attribute,
24HourPicture attribute, temporalIdMin attribute, and temporalIdMax
attribute. Information relating to the contents of the HEVC codec
is specified as those attributes.
[0305] (MPEG4 AAC Audio Descriptor)
[0306] FIG. 29 is a diagram showing a syntax of an MPEG4 AAC Audio
Descriptor.
[0307] If MPEG4AAC is used as the encoding method for the audio
data, information relating to the contents of an AAC codec is
specified in the MPEG4AACAudioDescriptor element. The
MPEG4AACAudioDescriptor element is an upper-level attribute of
profile attribute, level attribute, channelConfig attribute,
AACServiceType attribute, receiverMixRqd attribute, mainId
attribute, asvc attribute, language attribute, and componentName
attribute. Information relating to the contents of the AAC codec is
specified by those attributes.
[0308] (AC3 Audio Descriptor)
[0309] FIG. 30 is a diagram showing a syntax of an AC3 Audio
Descriptor.
[0310] If AC3 is used as the encoding method for the audio data, in
the AC3AudioDescriptor element, information relating to the
contents of an AC3 codec is specified. The AC3AudioDescriptor
element is an upper-level element of sampleRateCode attribute, bsId
attribute, bitRateCode attribute, bsMod attribute, numChannels
attribute, fullSvc attribute, langcod attribute, mainId attribute,
priority attribute, textCod attribute, and language attribute.
Information relating to the contents of the AC3 codec is specified
by those attributes.
[0311] (Caption Parameters)
[0312] FIG. 31 is a diagram showing a syntax of Caption
Parameters.
[0313] In a Caption Parameters element, information relating to the
contents of the closed caption is specified. The Caption Parameters
element is an upper-level element of captionServiceNumber
attribute, language attribute, easyReader attribute, and
wideAspectRatio attribute. Information relating to the contents of
the closed caption is specified by those attributes.
[0314] Note that the syntaxes of the SCD, the SPD, and the
Descriptor elements of the SPD that are described above with
reference to FIGS. 14 to 31 are examples and other syntaxes can be
employed.
[0315] <6. System Configuration>
[0316] (Configuration of Broadcast Communication System)
[0317] FIG. 32 is a diagram showing a configuration example of the
broadcast communication system.
[0318] As shown in FIG. 32, a broadcast communication system 1 is
constituted of a data providing server 10, a transmission apparatus
20, a streaming server 30, a signaling server 40, and a reception
apparatus 60. In FIG. 32, the reception apparatus 60 is mutually
connected to the streaming server 30 and the signaling server 40
via the Internet 90. Note that, in FIG. 32, the data providing
server 10, the streaming server 30, and the signaling server 40
constitute the Internet server.
[0319] The data providing server 10 provides components such as
video data and audio data to the transmission apparatus 20 and the
streaming server 30. Further, the data providing server 10 provides
original data of signaling information to the transmission
apparatus 20 and the signaling server 40.
[0320] The transmission apparatus 20 transmits the component
provided from the data providing server 10, through a broadcast
wave of the digital broadcasting using the IP transmission system.
Further, the transmission apparatus 20 generates signaling
information using the original data of the signaling information
provided from the data providing server 10, and transmits it
together with the component through a broadcast wave of the digital
broadcasting using the IP transmission system. Note that the
transmission apparatus 20 corresponds to the above-mentioned
transmitter (e.g., FIG. 4) and is provided by, for example, the
broadcaster.
[0321] In response to a request from the reception apparatus 60,
the streaming server 30 streams and delivers the component provided
from the data providing server 10, as a stream via the Internet 90.
Note that the streaming server 30 corresponds to the
above-mentioned streaming server (e.g., FIG. 5) and is provided by,
for example, the broadcaster. Further, a plurality of streaming
servers 30 can be placed depending on the application form.
[0322] The signaling server 40 generates signaling information
using the original data of the signaling information provided from
the data providing server 10. The signaling server 40 requests, in
response to a request from the reception apparatus 60, the
signaling information via the Internet 90. Note that the signaling
server 40 corresponds to the above-mentioned signaling server
(e.g., FIG. 5) and provided by, for example, the broadcaster.
Further, a plurality of signaling servers 40 can be placed
depending on the application form.
[0323] The reception apparatus 60 receives a broadcast wave of the
digital broadcasting that is transmitted from the transmission
apparatus 20, and acquires the signaling information transmitted by
that broadcast wave of the digital broadcasting. Further, the
reception apparatus 60 accesses the signaling server 40 via the
Internet 90, and acquires the signaling information provided from
the signaling server 40.
[0324] On the basis of the signaling information acquired by
broadcasting or communication, the reception apparatus 60 acquires
the component transmitted by a broadcast wave of the digital
broadcasting transmitted from the transmission apparatus 20 or the
component streamed/delivered from the streaming server 30 via the
Internet 90. The reception apparatus 60 displays, on the basis of
the component acquired by broadcasting or communication, a picture
on the display and outputs sound synchronized with that picture
from the speaker.
[0325] Note that the reception apparatus 60 corresponds to the
above-mentioned receiver (e.g., FIG. 4) and is placed in each
house, for example. Further, the reception apparatus 60 may include
a display and a speaker or may be incorporated in a television
receiver, a video recorder, or the like.
[0326] The broadcast communication system 1 is thus configured.
Next, detailed configurations of the respective apparatuses
constituting the broadcast communication system 1 in FIG. 32 will
be described.
[0327] (Configuration of Transmission Apparatus)
[0328] FIG. 33 is a diagram showing a configuration example of the
transmission apparatus of FIG. 32.
[0329] As shown in FIG. 33, the transmission apparatus 20 is
constituted of a communication unit 201, a segment data generator
202, a signaling information generator 203, a Mux 204, and a
transmission unit 205.
[0330] The communication unit 201 is constituted of a component
acquisition unit 211 and a signaling information acquisition unit
212. The component acquisition unit 211 acquires video data and
audio data provided from the data providing server 10, and supplies
them to the segment data generator 202. Further, the signaling
information acquisition unit 212 acquires the original data of the
signaling information provided from the data providing server 10,
and supplies it to the signaling information generator 203.
[0331] The segment data generator 202 generates, on the basis of
the video data and the audio data supplied from the component
acquisition unit 211, segment data, and supplies it to the Mux 204.
The signaling information generator 203 generates, on the basis of
the original data of the signaling information supplied from the
signaling information acquisition unit 212, signaling information,
and supplies it to the Mux 204. Note that the data providing server
10 may provide the signaling information itself rather than the
original data of the signaling information. In this case, the
signaling information acquired by the signaling information
acquisition unit 212 is supplied to the Mux 204 as it is.
[0332] The Mux 204 multiplexes the segment data supplied from the
segment data generator 202 and the signaling information supplied
from the signaling information generator 203 to generate a BBP
stream, and supplies it to the transmission unit 205. The
transmission unit 205 modulates the BBP stream supplied from the
Mux 204, and transmits it as a broadcast wave of the digital
broadcasting using the IP transmission system via an antenna 221.
Note that, at this time, the segment data and the signaling
information (SCS) are, for example, transmitted through the FLUTE
session.
[0333] (Configuration of Internet Server)
[0334] FIG. 34 is a diagram showing a configuration example of the
Internet server of FIG. 32. As shown in FIG. 34, the Internet
server is constituted of the data providing server 10, the
streaming server 30, and the signaling server 40.
[0335] (Configuration of Data Providing Server)
[0336] The data providing server 10 is constituted of a control
unit 101, component storage 102, and a communication unit 103. The
control unit 101 controls operations of the respective units of the
data providing server 10. The component storage 102 stores various
components such as video data and audio data and the original data
of the signaling information.
[0337] The communication unit 103 provides, under the control of
the control unit 101, the components and the original data of the
signaling information that are stored in the component storage 102,
to the transmission apparatus 20. Further, the communication unit
103 provides, under the control of the control unit 101, the
components stored in the component storage 102 to the streaming
server 30. In addition, the communication unit 103 provides, under
the control of the control unit 101, the original data of the
signaling information stored in the component storage 102 to the
signaling server 40.
[0338] (Configuration of Streaming Server)
[0339] The streaming server 30 is constituted of a control unit
301, a communication unit 302, and a segment data generator 303.
The control unit 301 controls operations of the respective units of
the streaming server 30. The communication unit 302 supplies, under
the control of the control unit 301, the video data and the audio
data provided from the data providing server 10 to the segment data
generator 303.
[0340] The segment data generator 303 generates a segment data on
the basis of the video data and the audio data supplied from the
communication unit 302. The segment data generator 303 supplies,
under the control of the control unit 301, the segment data to the
communication unit 302. The communication unit 302 streams and
delivers, in response to a request from the reception apparatus 60,
the segment data supplied from the segment data generator 303 to
the reception apparatus 60 via the Internet 90.
[0341] (Configuration of Signaling Server)
[0342] The signaling server 40 is constituted of a control unit
401, a communication unit 402, and a signaling information
generator 403. The control unit 401 controls operations of the
respective units of the signaling server 40. The communication unit
402 supplies, under the control of the control unit 401, the
original data of the signaling information provided from the data
providing server 10, to the signaling information generator
403.
[0343] The signaling information generator 403 generates signaling
information on the basis of the original data of the signaling
information supplied from the communication unit 402. The signaling
information generator 403 supplies, under the control of the
control unit 401, the signaling information to the communication
unit 402. The communication unit 402 provides, in response to a
request from the reception apparatus 60, the signaling information
supplied from the signaling information generator 403, to the
reception apparatus 60 via the Internet 90.
[0344] Note that, in FIGS. 32 and 34, for the sake of description,
the data providing server 10, the streaming server 30, and the
signaling server 40 are separate apparatuses as the Internet
server. However, the Internet server only needs to include the
functional configurations shown in FIG. 34, and, for example, the
data providing server 10, the streaming server 30, and the
signaling server 40 may be considered as a single apparatus. In
this case, for example, the duplicated functions of the control
unit, the communication unit, and the like can get together.
[0345] (Configuration of Reception Apparatus)
[0346] FIG. 35 is a diagram showing a configuration example of the
reception apparatus of FIG. 32.
[0347] As shown in FIG. 35, the reception apparatus 60 is
constituted of a control unit 601, an NVRAM 602, an input unit 603,
a tuner 604, a Demux 605, a video selector 606, an audio selector
607, a communication unit 608, a Demux 609, a video decoder 610, a
video output unit 611, an audio decoder 612, and an audio output
unit 613.
[0348] The control unit 601 controls operations of the respective
units of the reception apparatus 60. The NVRAM 602 is a nonvolatile
memory and records various types of data under the control of the
control unit 601. The input unit 603 supplies, according to user's
operation, an operation signal to the control unit 601. The control
unit 601 controls, on the basis of the operation signal supplied
from the input unit 603, operations of the respective units of the
reception apparatus 60.
[0349] Under the control of the control unit 601, the tuner 604
extracts, from a broadcast wave of the digital broadcasting using
the IP transmission system received via an antenna 621, a
broadcasting signal of a particular service, with respect to which
a selection instruction has been made, and demodulates it, and
supplies the resulting BBP stream to the Demux 605.
[0350] Under the control of the control unit 601, the Demux 605
demultiplexes the BBP stream supplied from the tuner 604 into the
video data and the audio data and the signaling information and
supplies the video data to the video selector 606 and the audio
data to the audio selector 607. Further, the Demux 605 supplies the
signaling information to the control unit 601. The control unit 601
controls, on the basis of the signaling information supplied from
the Demux 605, operations of the respective units for acquiring the
component transmitted by broadcasting.
[0351] Note that, if the segment data and the signaling information
are being transmitted through the FLUTE session, the Demux 605
performs, under the control of the control unit 601, filtering
processing using IP address, port number, TSI, TOI, and the like,
to thereby obtain the video data and the audio data and the
signaling information.
[0352] Under the control of the control unit 601, the communication
unit 608 requests delivery of the stream from the streaming server
30, via the Internet 90. The communication unit 608 receives the
stream streamed/delivered from the streaming server 30 via the
Internet 90 and supplies it to the Demux 609.
[0353] The Demux 609 demultiplexes, under the control of the
control unit 601, the stream supplied from the communication unit
608 into the video data and the audio data, and supplies the video
data to the video selector 606 and supplies the audio data to the
audio selector 607. Note that, for example, if the stream delivered
from the streaming server 30 is only the audio, the Demux 609
supplies that audio data to the audio selector 607 without
demultiplexing the component.
[0354] Further, under the control of the control unit 601, the
communication unit 608 requests the signaling information from the
signaling server 40 via the Internet 90. The communication unit 608
receives the signaling information transmitted from the signaling
server 40 via the Internet 90 and supplies it to the control unit
601. On the basis of the signaling information supplied from the
communication unit 608, the control unit 601 controls operations of
the respective units for acquiring the component transmitted by
broadcasting or communication.
[0355] Under the control of the control unit 601, the video
selector 606 supplies either one of the video data supplied from
the Demux 605 and the video data supplied from the Demux 609 to the
video decoder 610.
[0356] The video decoder 610 decodes the video data supplied from
the video selector 606 and supplies it to the video output unit
611. The video output unit 611 supplies the video data supplied
from the video decoder 610 to a display (not shown) at the
subsequent stage. With this, for example, the picture of the
program is displayed on the display.
[0357] The audio selector 607 supplies, under the control of the
control unit 601, either one of the audio data supplied from the
Demux 605 and the audio data supplied from the Demux 609 to the
audio decoder 612.
[0358] The audio decoder 612 decodes the audio data supplied from
the audio selector 607 and supplies it to the audio output unit
613. The audio output unit 613 supplies the audio data supplied
from the audio decoder 612 to a speaker (not shown) at the
subsequent stage. With this, sound corresponding to the picture of
the program, for example, is output from the speaker.
[0359] <7. Flows of Processing Executed by Apparatuses>
[0360] Next, a flow of processing executed by the respective
apparatuses constituting the broadcast communication system 1 in
FIG. 32 will be described with reference to the flowcharts in FIGS.
36 to 39.
[0361] (Transmission Processing)
[0362] First, referring to a flowchart in FIG. 36, transmission
processing executed by the transmission apparatus 20 of FIG. 32
will be described.
[0363] In Step S201, the component acquisition unit 211 acquires a
component provided from the data providing server 10. Here, for
example, video data and audio data are acquired as the component
and supplied to the segment data generator 202. In Step S202, the
segment data generator 202 generates segment data on the basis of
the video data and the audio data supplied from the component
acquisition unit 211, and supplies it to the Mux 204.
[0364] In Step S203, the signaling information acquisition unit 212
acquires the original data of the signaling information provided
from the data providing server 10, and supplies it to the signaling
information generator 203. In Step S204, the signaling information
generator 203 generates the signaling information on the basis of
the original data of the signaling information supplied from the
signaling information acquisition unit 212, and supplies it to the
Mux 204. Note that, if the signaling information is provided from
the data providing server 10, the signaling information acquired by
the signaling information acquisition unit 212 will be supplied to
the Mux 204.
[0365] In Step S205, the Mux 204 multiplies the segment data
supplied from the segment data generator 202 with the signaling
information supplied from the signaling information generator 203,
generates a BBP stream, and supplies it to the transmission unit
205. In Step S206, the transmission unit 205 modulates the BBP
stream supplied from the Mux 204, and transmits it as a broadcast
wave of the digital broadcasting using the IP transmission system,
via the antenna 221.
[0366] Note that, at this time, the segment data and the signaling
information (SCS) are, for example, transmitted through the FLUTE
session. When the processing of Step S206 is terminated, the
transmission processing of FIG. 36 is terminated.
[0367] In the above, the transmission processing has been
described.
[0368] (Streaming Delivery Processing)
[0369] Next, referring to a flowchart in FIG. 37, streaming
delivery processing executed by the streaming server 30 of FIG. 32
will be described.
[0370] In Step S301, the communication unit 302 acquires, under the
control of the control unit 301, the component provided from the
data providing server 10. Here, for example, video data and audio
data are acquired as the component and supplied to the segment data
generator 303. In Step S302, under the control of the control unit
301, the segment data generator 303 generates segment data on the
basis of the video data and the audio data supplied from the
communication unit 302.
[0371] In Step S303, the control unit 301 monitors the
communication condition of the communication unit 302, and
determines whether or not a request of the streaming delivery has
been received from the reception apparatus 60. In Step S303, after
the request of the streaming delivery is received from the
reception apparatus 60, the processing proceeds to Step S304.
[0372] In Step S304, under the control of the control unit 301, the
communication unit 302 streams and delivers the segment data
supplied from the segment data generator 303, as a stream via the
Internet 90 to the reception apparatus 60. When the processing of
Step S304 is terminated, the streaming delivery processing of FIG.
37 is terminated.
[0373] In the above, the streaming delivery processing has been
described.
[0374] (Signaling Information-Providing Processing)
[0375] Next, referring to a flowchart in FIG. 38, the signaling
information-providing processing executed by the signaling server
40 in FIG. 32 will be described.
[0376] In Step S401, the communication unit 402 acquires, under the
control of the control unit 401, the original data of the signaling
information provided from the data providing server 10, and
supplies it to the signaling information generator 403. In Step
S402, under the control of the control unit 401, the signaling
information generator 403 generates signaling information on the
basis of the original data of the signaling information supplied
from the communication unit 402.
[0377] In Step S403, the control unit 401 monitors the
communication condition of the communication unit 402, and
determines whether or not a request of the signaling information
has been received from the reception apparatus 60. In Step S403,
after the request of the signaling information is received from the
reception apparatus 60, the processing proceeds to Step S404.
[0378] In Step S404, under the control of the control unit 401, the
communication unit 402 provides the signaling information supplied
from the signaling information generator 403, to the reception
apparatus 60 via the Internet 90. When the processing of Step S404
is terminated, the signaling information-providing processing of
FIG. 38 is terminated.
[0379] In the above, the signaling information-providing processing
has been described.
[0380] (Channel Selection Processing)
[0381] Finally, referring to a flowchart in FIG. 39, channel
selection processing executed by the reception apparatus 60 of FIG.
32 will be described. Note that, it is assumed that the initial
scanning is performed preceding this channel selection processing
in the reception apparatus 60, and channel selection information
such as the SCD is recorded on the NVRAM 602.
[0382] In Step S601, on the basis of the operation signal supplied
from the input unit 603, the control unit 601 determines whether or
not a channel selection operation has been made by the user. In
Step S601, for example, after the channel selection operation is
made by user's operation with respect to the remote controller, the
processing proceeds to Step S602. In Step S602, the control unit
601 determines whether or not to acquire the signaling information
from the broadcast.
[0383] If it is in Step S602 determined to acquire the signaling
information from the broadcast, the processing proceeds to Step
S603. In Step S603, the control unit 601 controls the Demux 605
according to the SCS Bootstrap information of the SCD read out from
the NVRAM 602, to be connected to the SCS being transmitted by
broadcasting, and acquires the signaling information (SCS). Here,
for example, in the case of the basic service (FIG. 4, etc.), the
component transmitted by broadcasting is only acquired, and hence
the signaling information transmitted by broadcasting is
acquired.
[0384] On the other hand, if it is in Step S602 determined that the
signaling information is acquired from the communication, the
processing proceeds to Step S604. In Step S604, the control unit
601 controls the communication unit 608 according to the URL
specified as the url attribute of the SignalingOverinternet element
of the SCD read out from the NVRAM 602, and accesses the signaling
server 40 via the Internet 90 to acquire the signaling information.
Here, for example, in the case of the hybrid service (FIG. 5,
etc.), the components transmitted by broadcasting and communication
are acquired, and hence the signaling information transmitted by
communication is acquired.
[0385] When the signaling information transmitted by broadcasting
or communication is acquired by the processing of Step S603 or Step
S604, the processing proceeds to Step S605. In Step S605, the
control unit 601 parses the signaling information acquired in the
processing of Step S603 or Step S604. By this parsing processing,
an acquisition source of the video and audio components (streams)
is determined.
[0386] In Step S606, according to the result of the parsing
processing of Step S605, the control unit 601 determines whether or
not the stream is being transmitted by broadcasting. If it is in
Step S606 determined that the stream is being transmitted by
broadcasting, the processing proceeds to Step S607.
[0387] In Step S607, the control unit 601 controls the tuner 604,
the Demux 605, or the like to acquire the video and audio
components transmitted by broadcasting. The thus acquired video
data and audio data are supplied to the video selector 606 or the
audio selector 607 at the subsequent stage. For example, in the
basic service (FIG. 4, etc.) or the hybrid service (FIG. 5, etc.),
the video and audio components are being transmitted by
broadcasting, and hence the IP address, port number, TSI, and TOI
for acquiring those components are determined. By being connected
to the stream being transmitted through the FLUTE session, the
segment data is acquired and the video data and the audio data are
obtained.
[0388] Note that, if it is in Step S606 determined that the stream
is not being transmitted by broadcasting, Step S607 is skipped and
the processing proceeds to Step S608. In Step S608, according to
the result of the parsing processing of Step S605, the control unit
601 determines whether or not the stream is being transmitted by
communication. If it is in Step S608 determined that the stream is
being transmitted by communication, the processing proceeds to Step
S609.
[0389] In Step S609, the control unit 601 controls the
communication unit 608, the Demux 609, or the like to acquire the
video and audio components being transmitted by communication. The
thus acquired video data and audio data are supplied to the video
selector 606 or the audio selector 607 at the subsequent stage. For
example, in the hybrid service (FIG. 5, etc.), the video and audio
components are transmitted by communication. Therefore, by
accessing the streaming server 30 via the Internet 90 according to
the segment URL of the MPD and being connected to the video and the
audio stream, the segment data is acquired and the video data and
the audio data are obtained.
[0390] Note that, if it is in Step S608 determined that the stream
is not transmitted by communication, Step S609 is skipped and the
processing proceeds to Step S610. That is, by the processing of
Steps S606 to S609, the component transmitted by broadcasting or
communication is acquired and the video data and the audio data are
obtained. Then, the video data is supplied to the video decoder 610
via the video selector 606, and the audio data is supplied to the
audio decoder 612 via the audio selector 607.
[0391] In Step S610, the video decoder 610 decodes the video data
supplied from the video selector 606, and supplies it to the video
output unit 611. Further, the audio decoder 612 decodes the audio
data supplied from the audio selector 607, and supplies it to the
audio output unit 613.
[0392] In Step S611, the video output unit 611 supplies the video
data supplied from the video decoder 610 to a display (not shown)
at the subsequent stage. Further, the audio output unit 613
supplies the audio data supplied from the audio decoder 612 to a
speaker (not shown) at the subsequent stage. With this, the picture
such as the program is displayed on the display and sound
synchronized with that picture is output from the speaker. When the
processing of Step S611 is terminated, the channel selection
processing of FIG. 39 is terminated.
[0393] In the above, the channel selection processing has been
described.
[0394] Note that, although "D" that is the abbreviation of
Description is used as the name of the signaling information in the
above description, "T" that is the abbreviation of Table may be
used. For example, the SCD (Service Configuration Description) may
be described as an SCT (Service Configuration Table). Further, for
example, the SPD (Service Parameter Description) may be described
as an SPT (Service Parameter Table). It should be noted that the
difference of those names is a formal difference between
"Description" and "Table" and the substantial contents of the
signaling information items are not different.
[0395] <8. Configuration of Computer>
[0396] The above-mentioned series of processing may be executed by
hardware or may be executed by software. If the series of
processing is executed by software, programs configuring that
software are installed into a computer. FIG. 40 is a diagram
showing a configuration example of hardware of a computer that
executes the above-mentioned series of processing according to the
programs.
[0397] In a computer 900, a CPU (Central Processing Unit) 901, a
ROM (Read Only Memory) 902, and a RAM (Random Access Memory) 903
are connected to one another via a bus 904. An input/output
interface 905 is further connected to the bus 904. An input unit
906, an output unit 907, a recording unit 908, a communication unit
909, and a drive 910 are connected to the input/output interface
905.
[0398] The input unit 906 is constituted of a keyboard, a mouse, a
microphone, and the like. The output unit 907 is constituted of a
display, a speaker, and the like. The recording unit 908 is
constituted of a hard disk, a nonvolatile memory, and the like. The
communication unit 909 is constituted of a network interface and
the like. The drive 910 drives a removable medium 911 such as a
magnetic disk, an optical disc, a magneto-optical disk, and a
semiconductor memory.
[0399] In the thus configured computer 900, the above-mentioned
series of processing is performed by the CPU 901 loading programs
stored in the ROM 902 and the recording unit 908 into the RAM 903
via the input/output interface 905 and the bus 904 and executing
them.
[0400] The programs executed by the computer 900 (CPU 901) can be
recorded and provided on the removable medium 911 as a package
medium, for example. Further, the programs can be provided via a
wired or wireless transmission medium such as a local-area network,
the Internet, and digital satellite broadcasting.
[0401] In the computer 900, the programs can be installed into the
recording unit 908 via the input/output interface 905 by the
removable medium 911 being mounted on the drive 910. Further, the
programs can be received by the communication unit 909 via the
wired or wireless transmission medium and installed into the
recording unit 908. Otherwise, the programs can be installed into
the ROM 902 or the recording unit 908 in advance.
[0402] In the present specification, the processing executed by the
computer according to the programs does not necessarily need to be
performed in a time sequence in the order described as the
flowchart. That is, the processing executed by the computer
according to the programs includes processes executed in parallel
or individually (e.g., parallel processing or processing by
objects). Further, the programs may be processed by a single
computer (processor) or may be processed by a plurality of
computers in a distributed manner.
[0403] Note that embodiments of the present technology are not
limited to the above-mentioned embodiments and various
modifications can be made without departing from the gist of the
present technology.
[0404] It should be noted that the present technology may take the
following configurations.
[0405] (1)
[0406] A reception apparatus, including:
[0407] a reception unit that receives a broadcast wave of digital
broadcasting using an IP (Internet Protocol) transmission system;
and
[0408] a control unit that controls, on the basis of information
for managing only a broadcast component transmitted by a broadcast
wave of the digital broadcasting, which is first signaling
information transmitted by a broadcast wave of the digital
broadcasting, or information for managing at least one component of
the broadcast component and a communication component transmitted
by communication, which is second signaling information transmitted
by communication, operations of respective units for acquiring the
at least one component of the broadcast component and the
communication component.
[0409] (2)
[0410] The reception apparatus according to (1), in which, the
first signaling information is information in units of services and
includes a plurality of management information items for acquiring
the broadcast component transmitted through a FLUTE (File Delivery
over Unidirectional Transport) session.
[0411] (3)
[0412] The reception apparatus according to (1), in which
[0413] the first signaling information is information in units of
services and includes one management information item obtained by
integrating a plurality of management information items for
acquiring the broadcast component transmitted through a FLUTE
session as a parameter defined at a component level.
[0414] (4)
[0415] The reception apparatus according to any one of (1) to (3),
in which
[0416] the first signaling information and the second signaling
information include management information defining information
relating to the second signaling information as a parameter at a
service level.
[0417] (5)
[0418] The reception apparatus according to (4), in which
[0419] the management information includes information indicating a
range of signaling information, a version information, and a URL
(Uniform Resource Locator) indicating an acquisition source, as the
information relating to the second signaling information.
[0420] (6)
[0421] The reception apparatus according to (5), in which
[0422] the management information further includes information
indicating an update interval of the second signaling information,
as the information relating to the second signaling
information.
[0423] (7)
[0424] The reception apparatus according to (5) or (6), in
which
[0425] the management information further includes information
indicating a timing to terminate acquisition of the second
signaling information, as the information relating to the second
signaling information.
[0426] (8)
[0427] The reception apparatus according to any one of (1) to (7),
in which
[0428] the second signaling information is information in units of
services and includes a plurality of management information items
for acquiring the broadcast component and an MPD (Media
Presentation Description) complying with a standard of MPEG-DASH
(Moving Picture
[0429] Expert Group-Dynamic Adaptive Streaming over HTTP) as
management information for acquiring the communication
component.
[0430] (9)
[0431] The reception apparatus according to any one of (1) to (8),
in which
[0432] the first signaling information is transmitted in an upper
layer than an IP layer in a hierarchy of a protocol in the IP
transmission system, and
[0433] a common IP address is assigned to the broadcast component
configuring a particular service and the first signaling
information.
[0434] (10)
[0435] A reception method for a reception apparatus, including the
steps of:
[0436] receiving, by the reception apparatus, a broadcast wave of
digital broadcasting using an IP transmission system; and
[0437] controlling, on the basis of information for managing only a
broadcast component transmitted by a broadcast wave of the digital
broadcasting, which is first signaling information transmitted by a
broadcast wave of the digital broadcasting, or information for
managing at least one component of the broadcast component and a
communication component transmitted by communication, which is
second signaling information transmitted by communication,
operations of respective units for acquiring the at least one
component of the broadcast component and the communication
component.
[0438] (11)
[0439] A transmission apparatus, including:
[0440] a first acquisition unit that acquires first signaling
information for managing only a broadcast component transmitted by
a broadcast wave of digital broadcasting using an IP transmission
system;
[0441] a second acquisition unit that acquires one or more
broadcast components configuring a service; and
[0442] a transmission unit that transmits the first signaling
information together with the broadcast component by a broadcast
wave of the digital broadcasting using the IP transmission
system.
[0443] (12)
[0444] The transmission apparatus according to (11), in which
[0445] the first signaling information is information in units of
services and includes a plurality of management information items
for acquiring the broadcast component transmitted through a FLUTE
session.
[0446] (13)
[0447] The transmission apparatus according to (11), in which
[0448] the first signaling information is information in units of
services and includes one management information item obtained by
integrating a plurality of management information items for
acquiring the broadcast component transmitted through a FLUTE
session as a parameter defined at a component level.
[0449] (14)
[0450] The transmission apparatus according to any one of (11) to
(13), in which
[0451] a receiver that receives a broadcast wave of the digital
broadcasting using the IP transmission system is capable of
acquiring information for managing the at least one component of
the broadcast component and a communication component transmitted
by communication, which is second signaling information transmitted
by communication, and
[0452] the first signaling information and the second signaling
information include management information defining information
relating to the second signaling information as a parameter at a
service level.
[0453] (15)
[0454] The transmission apparatus according to (14), in which
[0455] the management information includes information indicating a
range of signaling information, version information, and a URL of
an acquisition source, as the information relating to the second
signaling information.
[0456] (16)
[0457] The transmission apparatus according to (15), in which
[0458] the management information further includes information
indicating an update interval of the second signaling information,
as the information relating to the second signaling
information.
[0459] (17)
[0460] The transmission apparatus according to (15) or (16), in
which
[0461] the management information further includes information
indicating a timing to terminate acquisition of the second
signaling information, as the information relating to the second
signaling information.
[0462] (18)
[0463] The transmission apparatus according to any one of (14) to
(17), in which
[0464] the second signaling information is information in units of
services and includes an MPD complying with a standard of MPEG-DASH
as the plurality of management information items for acquiring the
broadcast component and the management information for acquiring
the communication component.
[0465] (19)
[0466] The transmission apparatus according to any one of (11) to
(18), in which
[0467] the first signaling information is transmitted in an upper
layer than an IP layer in a hierarchy of a protocol in the IP
transmission system, and
[0468] a common IP address is assigned to the broadcast component
configuring a particular service and the first signaling
information.
[0469] (20)
[0470] A transmission method for a transmission apparatus,
including the steps of:
[0471] acquiring, by the transmission apparatus, first signaling
information for managing only a broadcast component transmitted by
a broadcast wave of digital broadcasting using an IP transmission
system;
[0472] acquiring one or more broadcast components configuring the
service; and
[0473] transmitting the first signaling information together with
the broadcast component by a broadcast wave of the digital
broadcasting using the IP transmission system.
DESCRIPTION OF REFERENCE NUMERALS
[0474] 1 broadcast communication system, 10 data providing server,
20 transmission apparatus, 30 streaming server, 40 signaling
server, 60 reception apparatus, 90 Internet, 201 communication
unit, 202 segment data generator, 203 signaling information
generator, 204 Mux, 205 transmission unit, 601 control unit, 602
NVRAM, 604 tuner, 605 Demux, 606 video selector, 607 audio
selector, 608 communication unit, 609 Demux, 610 video decoder, 611
video output unit, 612 audio decoder, 613 audio output unit, 900
computer, 901 CPU
* * * * *