U.S. patent application number 17/290103 was filed with the patent office on 2021-12-23 for method for allocating resources for relay node in next generation communication system, and device therefor.
This patent application is currently assigned to LG ELECTRONICS INC.. The applicant listed for this patent is LG ELECTRONICS INC.. Invention is credited to Youngtae KIM, Yunjung YI.
Application Number | 20210400540 17/290103 |
Document ID | / |
Family ID | 1000005826887 |
Filed Date | 2021-12-23 |
United States Patent
Application |
20210400540 |
Kind Code |
A1 |
KIM; Youngtae ; et
al. |
December 23, 2021 |
METHOD FOR ALLOCATING RESOURCES FOR RELAY NODE IN NEXT GENERATION
COMMUNICATION SYSTEM, AND DEVICE THEREFOR
Abstract
Disclosed in the present application is a method by which a
child node transmits and receives a signal in a next generation
wireless communication system. Particularly, the method comprises
the steps of: transmitting, to a parent node, a resource request
message for requesting either a downlink resource or an uplink
resource; receiving, from the parent node, a resource allocation
grant message indicating either the requested resource or a
flexible resource; transmitting and receiving a signal by using the
indicated resource, wherein the flexible resource is used as the
requested resource when the resource allocation grant message
indicates the flexible resource.
Inventors: |
KIM; Youngtae; (Seoul,
KR) ; YI; Yunjung; (Seoul, KR) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
LG ELECTRONICS INC. |
Seoul |
|
KR |
|
|
Assignee: |
LG ELECTRONICS INC.
Seoul
KR
|
Family ID: |
1000005826887 |
Appl. No.: |
17/290103 |
Filed: |
July 9, 2019 |
PCT Filed: |
July 9, 2019 |
PCT NO: |
PCT/KR2019/008396 |
371 Date: |
April 29, 2021 |
Current U.S.
Class: |
1/1 |
Current CPC
Class: |
H04W 72/044 20130101;
H04W 28/26 20130101; H04W 72/042 20130101 |
International
Class: |
H04W 28/26 20060101
H04W028/26; H04W 72/04 20060101 H04W072/04 |
Foreign Application Data
Date |
Code |
Application Number |
Nov 23, 2018 |
KR |
10-2018-0146721 |
Claims
1. A method of transmitting and receiving signals by a child node
in a next-generation wireless communication system, the method
comprising: transmitting a resource request message for requesting
for one of a downlink (DL) resource and an uplink (UL) resource to
a parent node; receiving a resource allocation grant message
indicating one of the requested resource and a flexible resource
from the parent node; and transmitting and receiving a signal using
the indicated resource, wherein, based on the resource allocation
grant message indicating the flexible resource, the flexible
resource is used as the requested resource.
2. The method of claim 1, wherein the resource request message
includes information about at which timing the requested resource
is located after the resource request message is transmitted.
3. The method of claim 1, wherein, upon failing to receive the
resource allocation grant message until a predetermined time
elapses after the resource requested message is transmitted, the
child node regards allocation of the requested resource as being
rejected by the parent node.
4. The method of claim 1, further comprising receiving, from the
parent node, a resource allocation rejection message indicating
that the requested resource is not valid until a predetermined time
elapses after the resource requested message is transmitted.
5. The method of claim 4, wherein the resource allocation rejection
message includes information about another resource rather than the
requested resource among the DL resource and the UL resource.
6. A relay node in a wireless communication system, the relay node
comprising: a wireless communication module; at least one
processor; and at least one memory operably connected to the at
least one processor and configured to store instructions for
causing the at least one processor to perform a specific operation
based on execution of the instructions, wherein the specific
operation comprises transmitting a resource request message for
requesting for one of a downlink (DL) resource and an uplink (UL)
resource to a parent node, receiving a resource allocation grant
message indicating one of the requested resource and a flexible
resource from the parent node, and transmitting and receiving a
signal using the indicated resource, and wherein, based on the
resource allocation grant message indicating the flexible resource,
the flexible resource is used as the requested resource.
7. The relay node of claim 6, wherein the resource request message
includes information about at which timing the requested resource
is located after the resource request message is transmitted.
8. The relay node of claim 6, wherein, upon failing to receive the
resource allocation grant message until a predetermined time
elapses after the resource requested message is transmitted, the at
least one processor regards allocation of the requested resource as
being rejected by the parent node.
9. The relay node of claim 6, wherein the specific operation
comprises receiving, from the parent node, a resource allocation
rejection message indicating that the requested resource is not
valid until a predetermined time elapses after the resource
requested message is transmitted.
10. The relay node of claim 9, wherein the resource allocation
rejection message includes information about another resource
rather than the requested resource among the DL resource and the UL
resource.
Description
TECHNICAL FIELD
[0001] The present disclosure relates to a wireless communication
system and, more particularly, to a method of allocating resources
for a relay node in a next-generation communication system and an
apparatus therefor.
BACKGROUND ART
[0002] As more and more communication devices demand larger
communication traffic along with the current trends, a
future-generation 5.sup.th generation (5G) system is required to
provide an enhanced wireless broadband communication, compared to
the legacy LTE system. In the future-generation 5G system,
communication scenarios are divided into enhanced mobile broadband
(eMBB), ultra-reliability and low-latency communication (URLLC),
massive machine-type communication (mMTC), and so on.
[0003] Herein, eMBB is a future-generation mobile communication
scenario characterized by high spectral efficiency, high user
experienced data rate, and high peak data rate, URLLC is a
future-generation mobile communication scenario characterized by
ultra-high reliability, ultra-low latency, and ultra-high
availability (e.g., vehicle to everything (V2X), emergency service,
and remote control), and mMTC is a future-generation mobile
communication scenario characterized by low cost, low energy, short
packet, and massive connectivity (e.g., Internet of things
(IoT)).
DETAILED DESCRIPTION OF THE DISCLOSURE
Technical Problems
[0004] Hereinbelow, a method of allocating resources for a relay
node in a next-generation communication system and an apparatus
therefor, based on the above discussion, will be proposed.
Technical Solutions
[0005] According to an aspect of the present disclosure, provided
herein is a method of transmitting and receiving signals by a child
node in a next-generation wireless communication system, including
transmitting a resource request message for requesting for one of a
downlink (DL) resource and an uplink (UL) resource to a parent
node; receiving a resource allocation grant message indicating one
of the requested resource and a flexible resource from the parent
node; and transmitting and receiving a signal using the indicated
resource, wherein, based on the resource allocation grant message
indicating the flexible resource, the flexible resource is used as
the requested resource.
[0006] In another aspect of the present disclosure, provided herein
is a relay node in a wireless communication system, including a
wireless communication module; at least one processor; and at least
one memory operably connected to the at least one processor and
configured to store instructions for causing the at least one
processor to perform a specific operation based on execution of the
instructions, wherein the specific operation includes transmitting
a resource request message for requesting for one of a downlink
(DL) resource and an uplink (UL) resource to a parent node,
receiving a resource allocation grant message indicating one of the
requested resource and a flexible resource from the parent node,
and transmitting and receiving a signal using the indicated
resource, and wherein, based on the resource allocation grant
message indicating the flexible resource, the flexible resource is
used as the requested resource.
[0007] The resource request message may include information about
at which timing the requested resource is located after the
resource request message is transmitted.
[0008] Upon failing to receive the resource allocation grant
message until a predetermined time elapses after the resource
requested message is transmitted, allocation of the requested
resource may be regarded as being rejected by the parent node.
[0009] A resource allocation rejection message indicating that the
requested resource is not valid until a predetermined time elapses
after the resource requested message is transmitted may be received
from the parent node. The resource allocation rejection message may
include information about another resource rather than the
requested resource among the DL resource and the UL resource.
Advantageous Effects
[0010] According to an embodiment of the present disclosure, a
relay node may be more efficiently allocated resources to transmit
and receive signals in a next-generation communication system.
[0011] It will be appreciated by persons skilled in the art that
the effects that could be achieved with the present disclosure are
not limited to what has been particularly described hereinabove and
other advantages of the present disclosure will be more clearly
understood from the following detailed description.
DESCRIPTION OF DRAWINGS
[0012] FIG. 1 is a diagram illustrating the control-plane and
user-plane architecture of radio interface protocols between a user
equipment (UE) and an evolved UMTS terrestrial radio access network
(E-UTRAN) in conformance to a 3rd generation partnership project
(3GPP) radio access network standard.
[0013] FIG. 2 is a diagram illustrating physical channels and a
general signal transmission method using the physical channels in a
3GPP system.
[0014] FIG. 3 illustrates a structure of a radio frame in a Long
Term Evolution (LTE) system.
[0015] FIGS. 4, 5 and 6 are diagrams illustrating structures of a
radio frame and slots used in a new RAT (NR) system.
[0016] FIG. 7 abstractly illustrates a hybrid beamforming structure
in terms of TXRUs and physical antennas.
[0017] FIG. 8 illustrates a beam sweeping operation for an SS and
system information during DL transmission.
[0018] FIG. 9 illustrates a cell in a new radio access technology
(NR) system.
[0019] FIG. 10 is a flowchart illustrating resource allocation
according to an embodiment of the present disclosure.
[0020] FIG. 11 is a block diagram illustrating elements of a device
for implementing embodiments of the present disclosure.
[0021] FIGS. 12 to 14 are diagrams illustrating an artificial
intelligence (AI) system and device for implementing embodiments of
the present disclosure.
BEST MODE FOR CARRYING OUT THE DISCLOSURE
[0022] The configuration, operation, and other features of the
present disclosure will readily be understood with embodiments of
the present disclosure described with reference to the attached
drawings. Embodiments of the present disclosure as set forth herein
are examples in which the technical features of the present
disclosure are applied to a 3rd generation partnership project
(3GPP) system.
[0023] While embodiments of the present disclosure are described in
the context of long term evolution (LTE) and LTE-advanced (LTE-A)
systems, they are purely exemplary. Therefore, the embodiments of
the present disclosure are applicable to any other communication
system as long as the above definitions are valid for the
communication system.
[0024] The term, base station (BS) may be used to cover the
meanings of terms including remote radio head (RRH), evolved Node B
(eNB or eNode B), transmission point (TP), reception point (RP),
relay, and so on.
[0025] <Artificial Intelligence (AI)>
[0026] AI refers to the field of studying AI or methodology for
making the same, and machine learning refers to the field of
defining various issues dealt with in the AI field and studying
methodology for solving the various issues. The machine learning is
defined as an algorithm that enhances the performance of a certain
task through consistent experiences with the task.
[0027] An artificial neural network (ANN) is a model used in the
machine learning and may mean a whole model of problem-solving
ability which is composed of artificial neurons (nodes) that form a
network by synaptic connections. The ANN may be defined by a
connection pattern between neurons in different layers, a learning
process for updating model parameters, and an activation function
for generating an output value.
[0028] The ANN may include an input layer, an output layer, and
optionally one or more hidden layers. Each layer includes one or
more neurons, and the ANN may include a synapse that links neurons.
In the ANN, each neuron may output the function value of the
activation function for input signals, weights, and bias input
through the synapse.
[0029] The model parameter refers to a parameter determined through
learning and includes the weight value of a synaptic connection and
the bias of a neuron. A hyperparameter means a parameter to be set
in the machine learning algorithm before learning and includes a
learning rate, a repetition number, a mini-batch size, and an
initialization function.
[0030] The purpose of the learning of the ANN may be to determine
the model parameter that minimizes a loss function. The loss
function may be used as an index to determine the optimal model
parameter in the learning process of the ANN.
[0031] Machine learning may be classified into supervised learning,
unsupervised learning, and reinforcement learning according to
learning mechanisms.
[0032] The supervised learning may refer to a method of training
the ANN in a state that labels for learning data are given, and the
label may mean a correct answer (or result value) that the ANN must
infer when the learning data is input to the ANN. The unsupervised
learning may refer to a method of training the ANN in a state that
labels for learning data are not given. The reinforcement learning
may refer to a method of learning an agent defined in a certain
environment to select a behavior or a behavior sequence that
maximizes cumulative compensation in each state.
[0033] Machine learning implemented with a deep neural network
(DNN) including a plurality of hidden layers among ANNs is referred
to as deep learning. The deep running is part of the machine
running. The machine learning used herein includes the deep
running.
[0034] <Robot>
[0035] A robot may refer to a machine that automatically processes
or operates a given task based on its own ability. In particular, a
robot having a function of recognizing an environment and making a
self-determination may be referred to as an intelligent robot.
[0036] Robots may be classified into industrial robots, medical
robots, home robots, military robots, etc. according to use
purposes or fields.
[0037] The robot may include a driving unit having an actuator or a
motor and perform various physical operations such as moving a
robot joint. In addition, a movable robot may include a driving
unit having a wheel, a brake, a propeller, etc. and may travel on
the ground or fly in the air through the driving unit.
[0038] <Autonomous Driving (Self-Driving)>
[0039] Autonomous driving refers to a technique of driving by
itself. An autonomous driving vehicle refers to a vehicle moving
with no user manipulation or with minimum user manipulation.
[0040] For example, the autonomous driving may include a technology
for maintaining a current lane, a technology for automatically
adjusting a speed such as adaptive cruise control, a technique for
automatically moving along a predetermined route, and a technology
for automatically setting a route and traveling along the route
when a destination is determined.
[0041] The vehicle may include a vehicle having only an internal
combustion engine, a hybrid vehicle having an internal combustion
engine and an electric motor together, and an electric vehicle
having only an electric motor. Further, the vehicle may include not
only an automobile but also a train, a motorcycle, etc.
[0042] The autonomous driving vehicle may be regarded as a robot
having the autonomous driving function.
[0043] <Extended Reality (XR)>
[0044] Extended reality is collectively referred to as virtual
reality (VR), augmented reality (AR), and mixed reality (MR). The
VR technology provides real-world objects and backgrounds as CG
images, the AR technology provides virtual CG images on real object
images, and the MR technology is a computer graphic technology of
mixing and combining virtual objects with the real world.
[0045] The MR technology is similar to the AR technology in that
real and virtual objects are shown together. However, the MR
technology is different from the AR technology in that the AR
technology uses virtual objects to complement real objects, whereas
the MR technology deal with virtual and real objects in the same
way.
[0046] The XR technology may be applied to a HMD, a head-up display
(HUD), a mobile phone, a tablet PC, a laptop computer, a desktop
computer, a TV, a digital signage, etc. A device to which the XR
technology is applied may be referred to as an XR device.
[0047] 5G communication involving a new radio access technology
(NR) system will be described below.
[0048] Three key requirement areas of 5G are (1) enhanced mobile
broadband (eMBB), (2) massive machine type communication (mMTC),
and (3) ultra-reliable and low latency communications (URLLC).
[0049] Some use cases may require multiple dimensions for
optimization, while others may focus only on one key performance
indicator (KPI). 5G supports such diverse use cases in a flexible
and reliable way.
[0050] eMBB goes far beyond basic mobile Internet access and covers
rich interactive work, media and entertainment applications in the
cloud or augmented reality (AR). Data is one of the key drivers for
5G and in the 5G era, we may for the first time see no dedicated
voice service. In 5G, voice is expected to be handled as an
application program, simply using data connectivity provided by a
communication system. The main drivers for an increased traffic
volume are the increase in the size of content and the number of
applications requiring high data rates. Streaming services (audio
and video), interactive video, and mobile Internet connectivity
will continue to be used more broadly as more devices connect to
the Internet. Many of these applications require always-on
connectivity to push real time information and notifications to
users. Cloud storage and applications are rapidly increasing for
mobile communication platforms. This is applicable for both work
and entertainment. Cloud storage is one particular use case driving
the growth of uplink data rates. 5G will also be used for remote
work in the cloud which, when done with tactile interfaces,
requires much lower end-to-end latencies in order to maintain a
good user experience. Entertainment, for example, cloud gaming and
video streaming, is another key driver for the increasing need for
mobile broadband capacity. Entertainment will be very essential on
smart phones and tablets everywhere, including high mobility
environments such as trains, cars and airplanes. Another use case
is AR for entertainment and information search, which requires very
low latencies and significant instant data volumes.
[0051] One of the most expected 5G use cases is the functionality
of actively connecting embedded sensors in every field, that is,
mMTC. It is expected that there will be 20.4 billion potential
Internet of things (IoT) devices by 2020. In industrial IoT, 5G is
one of areas that play key roles in enabling smart city, asset
tracking, smart utility, agriculture, and security
infrastructure.
[0052] URLLC includes services which will transform industries with
ultra-reliable/available, low latency links such as remote control
of critical infrastructure and self-driving vehicles. The level of
reliability and latency are vital to smart-grid control, industrial
automation, robotics, drone control and coordination, and so
on.
[0053] 5G communication involving a new radio access technology
(NR) system will be described below.
[0054] 5G may complement fiber-to-the home (FTTH) and cable-based
broadband (or data-over-cable service interface specifications (DOC
SIS)) as a means of providing streams at data rates of hundreds of
megabits per second to giga bits per second. Such a high speed is
required for TV broadcasts at or above a resolution of 4K (6K, 8K,
and higher) as well as virtual reality (VR) and AR. VR and AR
applications mostly include immersive sport games. A special
network configuration may be required for a specific application
program. For VR games, for example, game companies may have to
integrate a core server with an edge network server of a network
operator in order to minimize latency.
[0055] The automotive sector is expected to be a very important new
driver for 5G, with many use cases for mobile communications for
vehicles. For example, entertainment for passengers requires
simultaneous high capacity and high mobility mobile broadband,
because future users will expect to continue their good quality
connection independent of their location and speed. Other use cases
for the automotive sector are AR dashboards. These display overlay
information on top of what a driver is seeing through the front
window, identifying objects in the dark and telling the driver
about the distances and movements of the objects. In the future,
wireless modules will enable communication between vehicles
themselves, information exchange between vehicles and supporting
infrastructure and between vehicles and other connected devices
(e.g., those carried by pedestrians). Safety systems may guide
drivers on alternative courses of action to allow them to drive
more safely and lower the risks of accidents. The next stage will
be remote-controlled or self-driving vehicles. These require very
reliable, very fast communication between different self-driving
vehicles and between vehicles and infrastructure. In the future,
self-driving vehicles will execute all driving activities, while
drivers are focusing on traffic abnormality elusive to the vehicles
themselves. The technical requirements for self-driving vehicles
call for ultra-low latencies and ultra-high reliability, increasing
traffic safety to levels humans cannot achieve.
[0056] Smart cities and smart homes, often referred to as smart
society, will be embedded with dense wireless sensor networks.
Distributed networks of intelligent sensors will identify
conditions for cost- and energy-efficient maintenance of the city
or home. A similar setup can be done for each home, where
temperature sensors, window and heating controllers, burglar
alarms, and home appliances are all connected wirelessly. Many of
these sensors are typically characterized by low data rate, low
power, and low cost, but for example, real time high definition
(HD) video may be required in some types of devices for
surveillance.
[0057] The consumption and distribution of energy, including heat
or gas, is becoming highly decentralized, creating the need for
automated control of a very distributed sensor network. A smart
grid interconnects such sensors, using digital information and
communications technology to gather and act on information. This
information may include information about the behaviors of
suppliers and consumers, allowing the smart grid to improve the
efficiency, reliability, economics and sustainability of the
production and distribution of fuels such as electricity in an
automated fashion. A smart grid may be seen as another sensor
network with low delays.
[0058] The health sector has many applications that may benefit
from mobile communications. Communications systems enable
telemedicine, which provides clinical health care at a distance. It
helps eliminate distance barriers and may improve access to medical
services that would often not be consistently available in distant
rural communities. It is also used to save lives in critical care
and emergency situations. Wireless sensor networks based on mobile
communication may provide remote monitoring and sensors for
parameters such as heart rate and blood pressure.
[0059] Wireless and mobile communications are becoming increasingly
important for industrial applications. Wires are expensive to
install and maintain, and the possibility of replacing cables with
reconfigurable wireless links is a tempting opportunity for many
industries. However, achieving this requires that the wireless
connection works with a similar delay, reliability and capacity as
cables and that its management is simplified. Low delays and very
low error probabilities are new requirements that need to be
addressed with 5G.
[0060] Finally, logistics and freight tracking are important use
cases for mobile communications that enable the tracking of
inventory and packages wherever they are by using location-based
information systems. The logistics and freight tracking use cases
typically require lower data rates but need wide coverage and
reliable location information.
[0061] The 3GPP communication standards define downlink (DL)
physical channels corresponding to resource elements (REs) carrying
information originated from a higher layer, and DL physical signals
which are used in the physical layer and correspond to REs which do
not carry information originated from a higher layer. For example,
physical downlink shared channel (PDSCH), physical broadcast
channel (PBCH), physical multicast channel (PMCH), physical control
format indicator channel (PCFICH), physical downlink control
channel (PDCCH), and physical hybrid ARQ indicator channel (PHICH)
are defined as DL physical channels, and reference signals (RSs)
and synchronization signals (SSs) are defined as DL physical
signals. An RS, also called a pilot signal, is a signal with a
predefined special waveform known to both a gNode B (gNB) and a
user equipment (UE). For example, cell specific RS, UE-specific RS
(UE-RS), positioning RS (PRS), and channel state information RS
(CSI-RS) are defined as DL RSs. The 3GPP LTE/LTE-A standards define
uplink (UL) physical channels corresponding to REs carrying
information originated from a higher layer, and UL physical signals
which are used in the physical layer and correspond to REs which do
not carry information originated from a higher layer. For example,
physical uplink shared channel (PUSCH), physical uplink control
channel (PUCCH), and physical random access channel (PRACH) are
defined as UL physical channels, and a demodulation reference
signal (DMRS) for a UL control/data signal, and a sounding
reference signal (SRS) used for UL channel measurement are defined
as UL physical signals.
[0062] In the present disclosure, the PDCCH/PCFICH/PHICH/PDSCH
refers to a set of time-frequency resources or a set of REs, which
carry downlink control information (DCI)/a control format indicator
(CFI)/a DL acknowledgement/negative acknowledgement (ACK/NACK)/DL
data. Further, the PUCCH/PUSCH/PRACH refers to a set of
time-frequency resources or a set of REs, which carry UL control
information (UCI)/UL data/a random access signal. In the present
disclosure, particularly a time-frequency resource or an RE which
is allocated to or belongs to the
PDCCH/PCFICH/PHICH/PDSCH/PUCCH/PUSCH/PRACH is referred to as a
PDCCH RE/PCFICH RE/PHICH RE/PDSCH RE/PUCCH RE/PUSCH RE/PRACH RE or
a PDCCH resource/PCFICH resource/PHICH resource/PDSCH
resource/PUCCH resource/PUSCH resource/PRACH resource. Hereinbelow,
if it is said that a UE transmits a PUCCH/PUSCH/PRACH, this means
that UCI/UL data/a random access signal is transmitted on or
through the PUCCH/PUSCH/PRACH. Further, if it is said that a gNB
transmits a PDCCH/PCFICH/PHICH/PDSCH, this means that DCI/control
information is transmitted on or through the
PDCCH/PCFICH/PHICH/PDSCH.
[0063] Hereinbelow, an orthogonal frequency division multiplexing
(OFDM) symbol/carrier/subcarrier/RE to which a
CRS/DMRS/CSI-RS/SRS/UE-RS is allocated to or for which the
CRS/DMRS/CSI-RS/SRS/UE-RS is configured is referred to as a
CRS/DMRS/CSI-RS/SRS/UE-RS symbol/carrier/subcarrier/RE. For
example, an OFDM symbol to which a tracking RS (TRS) is allocated
or for which the TRS is configured is referred to as a TRS symbol,
a subcarrier to which a TRS is allocated or for which the TRS is
configured is referred to as a TRS subcarrier, and an RE to which a
TRS is allocated or for which the TRS is configured is referred to
as a TRS RE. Further, a subframe configured to transmit a TRS is
referred to as a TRS subframe. Further, a subframe carrying a
broadcast signal is referred to as a broadcast subframe or a PBCH
subframe, and a subframe carrying a synchronization signal (SS)
(e.g., a primary synchronization signal (PSS) and/or a secondary
synchronization signal (SSS)) is referred to as an SS subframe or a
PSS/SSS subframe. An OFDM symbol/subcarrier/RE to which a PSS/SSS
is allocated or for which the PSS/SSS is configured is referred to
as a PSS/SSS symbol/subcarrier/RE.
[0064] In the present disclosure, a CRS port, a UE-RS port, a
CSI-RS port, and a TRS port refer to an antenna port configured to
transmit a CRS, an antenna port configured to transmit a UE-RS, an
antenna port configured to transmit a CSI-RS, and an antenna port
configured to transmit a TRS, respectively. Antenna port configured
to transmit CRSs may be distinguished from each other by the
positions of REs occupied by the CRSs according to CRS ports,
antenna ports configured to transmit UE-RSs may be distinguished
from each other by the positions of REs occupied by the UE-RSs
according to UE-RS ports, and antenna ports configured to transmit
CSI-RSs may be distinguished from each other by the positions of
REs occupied by the CSI-RSs according to CSI-RS ports. Therefore,
the term CRS/UE-RS/CSI-RS/TRS port is also used to refer to a
pattern of REs occupied by a CRS/UE-RS/CSI-RS/TRS in a
predetermined resource area.
[0065] FIG. 1 illustrates control-plane and user-plane protocol
stacks in a radio interface protocol architecture conforming to a
3GPP wireless access network standard between a UE and an evolved
UMTS terrestrial radio access network (E-UTRAN). The control plane
is a path in which the UE and the E-UTRAN transmit control messages
to manage calls, and the user plane is a path in which data
generated from an application layer, for example, voice data or
Internet packet data is transmitted.
[0066] A physical (PHY) layer at layer 1 (L1) provides information
transfer service to its higher layer, a medium access control (MAC)
layer. The PHY layer is connected to the MAC layer via transport
channels. The transport channels deliver data between the MAC layer
and the PHY layer. Data is transmitted on physical channels between
the PHY layers of a transmitter and a receiver. The physical
channels use time and frequency as radio resources. Specifically,
the physical channels are modulated in orthogonal frequency
division multiple access (OFDMA) for downlink (DL) and in single
carrier frequency division multiple access (SC-FDMA) for uplink
(UL).
[0067] The MAC layer at layer 2 (L2) provides service to its higher
layer, a radio link control (RLC) layer via logical channels. The
RLC layer at L2 supports reliable data transmission. RLC
functionality may be implemented in a function block of the MAC
layer. A packet data convergence protocol (PDCP) layer at L2
performs header compression to reduce the amount of unnecessary
control information and thus efficiently transmit Internet protocol
(IP) packets such as IP version 4 (IPv4) or IP version 6 (IPv6)
packets via an air interface having a narrow bandwidth.
[0068] A radio resource control (RRC) layer at the lowest part of
layer 3 (or L3) is defined only on the control plane. The RRC layer
controls logical channels, transport channels, and physical
channels in relation to configuration, reconfiguration, and release
of radio bearers. A radio bearer refers to a service provided at
L2, for data transmission between the UE and the E-UTRAN. For this
purpose, the RRC layers of the UE and the E-UTRAN exchange RRC
messages with each other. If an RRC connection is established
between the UE and the E-UTRAN, the UE is in RRC Connected mode and
otherwise, the UE is in RRC Idle mode. A Non-Access Stratum (NAS)
layer above the RRC layer performs functions including session
management and mobility management.
[0069] DL transport channels used to deliver data from the E-UTRAN
to UEs include a broadcast channel (BCH) carrying system
information, a paging channel (PCH) carrying a paging message, and
a shared channel (SCH) carrying user traffic or a control message.
DL multicast traffic or control messages or DL broadcast traffic or
control messages may be transmitted on a DL SCH or a separately
defined DL multicast channel (MCH). UL transport channels used to
deliver data from a UE to the E-UTRAN include a random access
channel (RACH) carrying an initial control message and a UL SCH
carrying user traffic or a control message. Logical channels that
are defined above transport channels and mapped to the transport
channels include a broadcast control channel (BCCH), a paging
control channel (PCCH), a Common Control Channel (CCCH), a
multicast control channel (MCCH), a multicast traffic channel
(MTCH), etc.
[0070] FIG. 2 illustrates physical channels and a general method
for transmitting signals on the physical channels in the 3GPP
system.
[0071] Referring to FIG. 2, when a UE is powered on or enters a new
cell, the UE performs initial cell search (S201). The initial cell
search involves acquisition of synchronization to an eNB.
Specifically, the UE synchronizes its timing to the eNB and
acquires a cell identifier (ID) and other information by receiving
a primary synchronization channel (P-SCH) and a secondary
synchronization channel (S-SCH) from the eNB. Then the UE may
acquire information broadcast in the cell by receiving a physical
broadcast channel (PBCH) from the eNB. During the initial cell
search, the UE may monitor a DL channel state by receiving a
downlink reference signal (DL RS).
[0072] After the initial cell search, the UE may acquire detailed
system information by receiving a physical downlink control channel
(PDCCH) and receiving a physical downlink shared channel (PDSCH)
based on information included in the PDCCH (S202).
[0073] If the UE initially accesses the eNB or has no radio
resources for signal transmission to the eNB, the UE may perform a
random access procedure with the eNB (S203 to S206). In the random
access procedure, the UE may transmit a predetermined sequence as a
preamble on a physical random access channel (PRACH) (S203 and
S205) and may receive a response message to the preamble on a PDCCH
and a PDSCH associated with the PDCCH (S204 and S206). In the case
of a contention-based RACH, the UE may additionally perform a
contention resolution procedure.
[0074] After the above procedure, the UE may receive a PDCCH and/or
a PDSCH from the eNB (S207) and transmit a physical uplink shared
channel (PUSCH) and/or a physical uplink control channel (PUCCH) to
the eNB (S208), which is a general DL and UL signal transmission
procedure. Particularly, the UE receives downlink control
information (DCI) on a PDCCH. Herein, the DCI includes control
information such as resource allocation information for the UE.
Different DCI formats are defined according to different usages of
DCI.
[0075] Control information that the UE transmits to the eNB on the
UL or receives from the eNB on the DL includes a DL/UL
acknowledgment/negative acknowledgment (ACK/NACK) signal, a channel
quality indicator (CQI), a precoding matrix index (PMI), a rank
indicator (RI), etc. In the 3GPP LTE system, the UE may transmit
control information such as a CQI, a PMI, an RI, etc. on a PUSCH
and/or a PUCCH.
[0076] FIG. 3 illustrates a structure of a radio frame used in the
LTE system.
[0077] Referring to FIG. 3, a radio frame is 10 ms
(327200.times.Ts) long and divided into 10 equal-sized subframes.
Each subframe is 1 ms long and further divided into two slots. Each
time slot is 0.5 ms (15360.times.Ts) long. Herein, Ts represents a
sampling time and Ts=1/(15 kHz.times.2048)=3.2552.times.10.sup.-8
(about 33 ns). A slot includes a plurality of Orthogonal Frequency
Division Multiplexing (OFDM) symbols or SC-FDMA symbols in the time
domain by a plurality of Resource Blocks (RBs) in the frequency
domain. In the LTE system, one RB includes 12 subcarriers by 7 (or
6) OFDM symbols. A unit time during which data is transmitted is
defined as a Transmission Time Interval (TTI). The TTI may be
defined in units of one or more subframes. The above-described
radio frame structure is purely exemplary and thus the number of
subframes in a radio frame, the number of slots in a subframe, or
the number of OFDM symbols in a slot may vary.
[0078] FIG. 4 illustrates a structure of a radio frame used in
NR.
[0079] In NR, UL and DL transmissions are configured in frames. The
radio frame has a length of 10 ms and is defined as two 5-ms
half-frames (HF). The half-frame is defined as five 1 ms subframes
(SF). A subframe is divided into one or more slots, and the number
of slots in a subframe depends on subcarrier spacing (SCS). Each
slot includes 12 or 14 OFDM(A) symbols according to a cyclic prefix
(CP). When a normal CP is used, each slot includes 14 symbols. When
an extended CP is used, each slot includes 12 symbols. Here, the
symbols may include OFDM symbols (or CP-OFDM symbols) and SC-FDMA
symbols (or DFT-s-OFDM symbols).
[0080] In the NR system, the OFDM(A) numerology (e.g., SCS, CP
length, etc.) may be configured differently among a plurality of
cells merged for one UE. Thus, the (absolute time) duration of a
time resource (e.g., SF, slot or TTI) (referred to as a time unit
(TU) for simplicity) composed of the same number of symbols may be
set differently among the merged cells.
[0081] FIG. 5 illustrates a slot structure of an NR frame. A slot
includes a plurality of symbols in the time domain. For example, in
the case of the normal CP, one slot includes seven symbols. On the
other hand, in the case of the extended CP, one slot includes six
symbols. A carrier includes a plurality of subcarriers in the
frequency domain. A resource block (RB) is defined as a plurality
of consecutive subcarriers (e.g., 12 consecutive subcarriers) in
the frequency domain. A bandwidth part (BWP) is defined as a
plurality of consecutive (P)RBs in the frequency domain and may
correspond to one numerology (e.g., SCS, CP length, etc.). A
carrier may include up to N (e.g., five) BWPs. Data communication
is performed through an activated BWP, and only one BWP may be
activated for one UE. In the resource grid, each element is
referred to as a resource element (RE), and one complex symbol may
be mapped thereto.
[0082] FIG. 6 illustrates a structure of a self-contained slot. In
the NR system, a frame has a self-contained structure in which a DL
control channel, DL or UL data, a UL control channel, and the like
may all be contained in one slot. For example, the first N symbols
(hereinafter, DL control region) in the slot may be used to
transmit a DL control channel, and the last M symbols (hereinafter,
UL control region) in the slot may be used to transmit a UL control
channel. N and M are integers greater than or equal to 0. A
resource region (hereinafter, a data region) that is between the DL
control region and the UL control region may be used for DL data
transmission or UL data transmission. For example, the following
configuration may be considered. Respective sections are listed in
a temporal order.
[0083] 1. DL only configuration
[0084] 2. UL only configuration
[0085] 3. Mixed UL-DL configuration [0086] DL region+Guard period
(GP)+UL control region [0087] DL control region+GP+UL region [0088]
DL region: (i) DL data region, (ii) DL control region+DL data
region [0089] UL region: (i) UL data region, (ii) UL data region+UL
control region
[0090] The PDCCH may be transmitted in the DL control region, and
the PDSCH may be transmitted in the DL data region. The PUCCH may
be transmitted in the UL control region, and the PUSCH may be
transmitted in the UL data region. Downlink control information
(DCI), for example, DL data scheduling information, UL data
scheduling information, and the like, may be transmitted on the
PDCCH. Uplink control information (UCI), for example, ACK/NACK
information about DL data, channel state information (CSI), and a
scheduling request (SR), may be transmitted on the PUCCH. The GP
provides a time gap in the process of the UE switching from the
transmission mode to the reception mode or from the reception mode
to the transmission mode. Some symbols at the time of switching
from DL to UL within a subframe may be configured as the GP.
[0091] In an NR system, a technique of using an ultra-high
frequency band, that is, a millimeter frequency band at or above 6
GHz is considered in order to transmit data to a plurality of users
at a high transmission rate in a wide frequency band. In 3GPP, this
technique is called NR and will be referred to as an NR system in
the present disclosure. However, the millimeter frequency band has
the frequency property that a signal is attenuated too rapidly
according to distance due to the use of too high a frequency band.
Accordingly, the NR system using a frequency band at or above at
least 6 GHz employs a narrow beam transmission scheme in which a
signal is transmitted with concentrated energy in a specific
direction, not omni-directionally, to thereby compensate for rapid
propagation attenuation and thus overcome decrease of coverage
caused by the rapid propagation attenuation. However, if a service
is provided by using only one narrow beam, the service coverage of
one gNB becomes narrow, and thus the gNB provides a service in a
wide band by collecting a plurality of narrow beams.
[0092] As a wavelength becomes short in the millimeter frequency
band, that is, millimeter wave (mmW) band, it is possible to
install a plurality of antenna elements in the same area. For
example, a total of 100 antenna elements may be installed at
(wavelength) intervals of 0.5 lambda in a 30-GHz band with a
wavelength of about 1 cm in a two-dimensional (2D) array on a 5 cm
by 5 cm panel. Therefore, it is considered to increase coverage or
throughput by increasing beamforming gain through use of a
plurality of antenna elements in mmW.
[0093] To form a narrow beam in the millimeter frequency band, a
beamforming scheme is mainly considered, in which a gNB or a UE
transmits the same signals with appropriate phase differences
through multiple antennas, to thereby increase energy only in a
specific direction. Such beamforming schemes include digital
beamforming for generating a phase difference between digital
baseband signals, analog beamforming for generating a phase
difference between modulated analog signals by using time delay
(i.e., a cyclic shift), and hybrid beamforming using both digital
beamforming and analog beamforming. If a transceiver unit (TXRU) is
provided to enable control of transmission power and a phase per
antenna element, independent beamforming per frequency resource is
possible. However, installation of TXRUs for all of about 100
antenna elements is not feasible in terms of cost. That is, to
compensate for rapid propagation attenuation in the millimeter
frequency band, multiple antennas should be used, and digital
beamforming requires as many radio frequency (RF) components (e.g.,
digital to analog converters (DACs), mixers, power amplifiers, and
linear amplifiers) as the number of antennas. Accordingly,
implementation of digital beamforming in the millimeter frequency
band faces the problem of increased cost of communication devices.
Therefore, in the case in which a large number of antennas is
required as in the millimeter frequency band, analog beamforming or
hybrid beamforming is considered. In analog beamforming, a
plurality of antenna elements is mapped to one TXRU, and the
direction of a beam is controlled by an analog phase shifter. A
shortcoming of this analog beamforming scheme is that frequency
selective beamforming (BF) cannot be provided because only one beam
direction can be produced in a total band. Hybrid BF stands between
digital BF and analog BF, in which B TXRUs fewer than Q antenna
elements are used. In hybrid BF, the directions of beams
transmittable at the same time are limited to B or below although
the number of beam directions is different according to connections
between B TXRUs and Q antenna elements.
[0094] Digital BF performs signal processing on a digital baseband
signal that is to be transmitted or is received as mentioned above,
and therefore digital BF may transmit or receive signals in
multiple directions at the same time using multiple beams. In
contrast, analog BF performs beamforming with a received analog
signal or an analog signal to be transmitted in a modulated state,
and therefore analog BF may not simultaneously transmit or receive
signals in multiple directions beyond the range covered by one
beam. In general, a gNB communicates with multiple users at the
same time using broadband transmission or multi-antenna
characteristics. When the gNB uses analog or hybrid BF and forms an
analog beam in one beam direction, the gNB is allowed to
communicate only with users included in the same analog beam
direction due to the characteristics of analog BF. An RACH resource
allocation scheme and a scheme of resource utilization in the gNB
according to the present disclosure to be described later are
proposed in consideration of constraints resulting from the
characteristics of analog BF or hybrid BF.
[0095] FIG. 7 abstractly illustrates a hybrid beamforming structure
in terms of TXRUs and physical antennas.
[0096] For the case in which multiple antennas are used, hybrid BF
with digital BF and analog BF in combination has emerged. Analog BF
(or RF BF) is an operation of performing precoding (or combining)
in a transceiver (RF unit). Due to precoding (combining) in each of
a baseband unit and a transceiver (or an RF unit), hybrid BF offers
the benefit of performance close to the performance of digital BF,
while reducing the number of RF chains and the number of DACs (or
analog to digital converters (ADCs). For convenience, a hybrid BF
structure may be represented by N TXRUs and M physical antennas.
Digital BF for L data layers to be transmitted by a transmission
end may be represented as an N-by-L matrix, and then N converted
digital signals are converted into analog signals through TXRUs and
subjected to analog BF represented as an M-by-N matrix.
[0097] In FIG. 7, the number of digital beams is L, and the number
of analog beams is N. Further, it is considered in the NR system
that a gNB is configured to change analog BF on a symbol basis so
as to more efficiently support BF for a UE located in a specific
area. Further, when one antenna panel is defined by N TXRUs and M
RF antennas, introduction of a plurality of antenna panels to which
independent hybrid BF is applicable is also considered. As such, in
the case in which a gNB uses a plurality of analog beams, a
different analog beam may be preferred for signal reception at each
UE. Therefore, a beam sweeping operation is under consideration, in
which for at least an SS, system information, and paging, a gNB
changes a plurality of analog beams on a symbol basis in a specific
slot or SF to allow all UEs to have reception opportunities.
[0098] FIG. 8 illustrates a beam sweeping operation for an SS and
system information during DL transmission.
[0099] In FIG. 8, physical resources or a physical channel which
broadcasts system information of the NR system is referred to as an
xPBCH. Analog beams from different antenna panels may be
transmitted simultaneously in one symbol, and introduction of a
beam reference signal (BRS) transmitted for a single analog beam
corresponding to a specific antenna panel as illustrated in FIG. 8
is under discussion in order to measure a channel per analog beam.
BRSs may be defined for a plurality of antenna ports, and each
antenna port of the BRSs may correspond to a single analog beam.
Unlike the BRSs, the SS or the xPBCH may be transmitted for all
analog beams included in an analog beam group so that any UE may
receive the SS or the xPBCH successfully.
[0100] FIG. 9 illustrates a cell in an NR system.
[0101] Referring to FIG. 9, compared to a wireless communication
system such as legacy LTE in which one eNB forms one cell,
configuration of one cell by a plurality of transmission/reception
points (TRPs) is under discussion in the NR system. If a plurality
of TRPs forms one cell, even though a TRP serving a UE is changed,
seamless communication is advantageously possible, thereby
facilitating mobility management for UEs.
[0102] Compared to the LTE/LTE-A system in which a PSS/SSS is
transmitted omnidirectionally, a method of transmitting a signal
such as a PSS/SSS/PBCH through BF performed by sequentially
switching a beam direction to all directions at a gNB applying
mmWave is considered. Signal transmission/reception performed by
switching a beam direction is referred to as beam sweeping or beam
scanning. In the present disclosure, "beam sweeping" is a behavior
of a transmission side, and "beam scanning" is a behavior of a
reception side. For example, if up to N beam directions are
available to the gNB, the gNB transmits a signal such as a
PSS/SSS/PBCH in the N beam directions. That is, the gNB transmits
an SS such as the PSS/SSS/PBCH in each direction by sweeping a beam
in directions available to or supported by the gNB. Alternatively,
if the gNB is capable of forming N beams, some beams may be grouped
into one beam group, and the PSS/SSS/PBCH may be
transmitted/received on a group basis. One beam group includes one
or more beams. Signals such as the PSS/SSS/PBCH transmitted in the
same direction may be defined as one SS block (SSB), and a
plurality of SSBs may exist in one cell. If a plurality of SSBs
exist, an SSB index may be used to identify each SSB. For example,
if the PSS/SSS/PBCH is transmitted in 10 beam directions in one
system, the PSS/SSS/PBCH transmitted in the same direction may form
an SSB, and it may be understood that 10 SSBs exist in the system.
In the present disclosure, a beam index may be interpreted as an
SSB index
[0103] Currently, in 3GPP Release 16, i.e., standardization of an
NR system, a relay gNB is under discussion for the purpose of
reducing wired connection between gNBs while compensating for a
coverage hole. This is implemented through integrated access and
backhaul (IAB). A donor gNB (DgNB) transmits a signal to a UE via a
relay gNB. IAB includes a wireless backhaul link for communication
between a DgNB and a relay gNB or between relay gNBs and an access
link for communication between a DgNB and a UE or between a relay
gNB and a UE.
[0104] Signal transmission through IAB is broadly categorized into
two scenarios. The first one is an in-band scenario in which a
wireless backhaul link and an access link use the same frequency
band, and the second one is an out-band scenario in which the
wireless backhaul link and the access link use different frequency
bands. The first scenario should also deal with interference
between the wireless backhaul link and the access link compared to
the second scenario, so that the first scenario may be lower than
the second scenario in terms of feasibility of implementation.
[0105] The present disclosure relates to how to regulate resources
between a parent node and a child node when the parent node informs
the child node of resources of an IAB node.
[0106] In standardization for a current NR system, it is assumed
that nodes on the backhaul link transmit SSBs or CSI-RSs in order
to perform a discovery procedure. Each IAB node measures or
discovers an SSB or a CSI-RS to feed back the measured or
discovered SSB or CSI-RS to a parent node or a donor node. A
network or a middle node determines route selection based on the
feedback value. When the middle node is in charge of route
selection, the parent node may relay the discovered or measured
feedback value up to the middle node. When the network is
responsible for route selection for nodes that the network manages,
the parent node may relay the discovered or measured feedback value
up to the donor node.
[0107] Since this discovery operation is based on the assumption
that the IAB node operates in a half-duplex scheme which does not
allow simultaneous transmission and reception, there is a problem
in that, while transmitting an SSB or a CSI-RS for discovery, the
IAB node is incapable of measuring or discovering SSBs or CSI-RSs
transmitted by other nodes. To solve this problem, it is necessary
to perform time division multiplexing (TDM) on SSBs or CSI-RSs
transmitted between nodes. To this end, a transmission pattern for
transmission of the SSBs or the CSI-RSs or a muting pattern for
discontinuing ongoing transmission and discovering or measuring
discovery signals from other nodes may be needed.
[0108] Hereinbelow, for convenience of description, when RN1 and
RN2 connected via a backhaul link are present and RN1 relays
transmitted and received data to RN2, RN1 will be referred to as a
parent node of RN2, and RN2 will be referred to as a child node RN
of RN1.
[0109] In an IAB scenario of the current NR system, the IAB node is
allocated resources categorized into DL, UL, and flexible link (FL)
resources by the parent node in a UE mode and is allocated
resources categorized into DL, UL, FL, and not available (NA)
resources by the parent node in a gNB mode. In the gNB mode of the
IAB node, the DL, UL, and FL resources are divided into two types,
i.e., soft resources and hard resources, as described below. The
hard resources may always be used for the IAB node of the gNB mode,
whereas availability of the soft resources by the IAB node of the
gNB mode may be indicated by the parent node.
[0110] Soft resources that the parent node implicitly or explicitly
indicates for use are used for a UE mode of the child node and are
used as DL, UL, or FL resources between the parent node and the
child node. Soft resources that the parent node does not implicitly
or explicitly indicate for use are used for a gNB mode of the child
node and the child node uses the soft resources as DL, UE, or FL
resources for a child node thereof or for a UE. That is, whether
the parent node will directly use the soft resources or will cause
the child node thereof to use the soft resources may be
determined.
[0111] However, there may be a situation in which the child node
needs to urgently use the soft resources. For example, when the
child node should support a UE having a packet with less latency, a
situation in which the child node should allocate resources on the
soft resources may occur. Then, the child node may request that the
parent node allocate the soft resources that the child node desires
to use.
First Embodiment
[0112] First, the child node may make a request to the parent node
for the soft resources to use the soft resources allocated by the
parent node from the perspective of a distributed unit (DU).
Hereinbelow, a detailed description will be given.
[0113] 1. The child node makes a request to the parent node for one
of the soft DL, UL, and FL resources selected thereby for use. The
parent node properly selects the requested resource and informs the
child node of the selected resource. Particularly, the parent node
informs the child node of the selected resource through DCI in
order to dynamically allocate the resource. Additionally, the
parent node may inform the child node in which slot (or in which
symbol of the slot) the selected resource is located from a slot in
which the DCI is included.
[0114] When the parent node indicates a slot, if the requested
resource in the slot is DL, it is assumed that the requested
resource means a DL resource in the indicated slot. If the
requested resource in the slot is UL, it is assumed that the
requested resource means a UL resource in the indicated slot. If
the requested resource in the slot is FL, it is assumed that the
requested resource means an FL resource in the indicated slot.
Alternatively, if the requested resource in the slot is DL, it is
assumed that the requested resource means a DL resource or an FL
resource in the indicated slot. If the requested resource in the
slot is UL, it is assumed that the requested resource means a UL
resource or an FL resource in the indicated slot.
[0115] When the child node makes a request for an FL resource, the
parent node may also inform the child node of whether an FL
resource should be used as a DL resource or as a UL resource.
Alternatively, if the child node makes a request for the DL
resource or the UL resource and the parent node informs the child
node of the FL resource, the child node uses the FL resource as the
resource requested thereby, i.e., the DL resource or the UL
resource. Since cross-link interference (CLI) may differ according
to whether the FL resource is used as the DL resource or as the UL
resource, the FL resource may be helpful when the parent node
selects a soft resource.
[0116] Obviously, the parent node may reject the request made by
the child node through the DCI. Assuming that the requested
resource is not valid within a predetermined time duration (e.g., X
ms) from a requested timing, the parent node may inform the child
node that the request is rejected through the DCI. The
predetermined time duration may be predefined according to a band,
a combination of bands, and/or a numerology or may be indicated
through RRC signaling.
[0117] Alternatively, when the child node fails to receive a
special response from the parent node during a predetermined time
duration (e.g., Y ms), the child node may determine that the parent
node has rejected the request. When the parent node switches a DL
resource and a UL resource several times, a switching time
increases. Therefore, if the request for the UL resource is
rejected, the parent node may transmit a rejection message by
allocating the DL resource. This is advantageous in that the
switching time is not necessary. The predetermined time duration
may be predefined according to a band, a combination of bands,
and/or a numerology or may be indicated through RRC signaling.
[0118] 2. The child node may make a request for one of the soft DL,
UL, and FL resources selected thereby together with the location of
the requested resource from a requested timing of the requested
resource. After determining whether the requested resource is
valid, the parent node may transmit a confirmation or rejection
message. Particularly, when the parent node informs the child node
of the selected resource, the parent node may indicate the selected
resource through the DCI in order to dynamically allocate
resources. When the child node makes a request for the FL resource,
the parent node may also inform the child node of information as to
whether the FL resource should be used as the DL resource or as the
UL resource. Alternatively, when the child node makes a request for
the DL resource or the UL resource and the parent node informs the
child node of the FL resource, the child node uses the FL resource
as the resource requested thereby, i.e., the DL resource or the UL
resource. Since CLI may differ according to whether the FL resource
is used as the DL resource or as the UL resource, the FL resource
may be helpful when the parent node selects the soft resource.
[0119] When the child node fails to receive a special response from
the parent node during a predetermined time duration (e.g., Y ms),
the child node may determine that the parent node has rejected the
request. When the parent node switches a DL resource and a UL
resource several times, a switching time increases. Therefore, if
the request for the UL resource is rejected, the parent node may
transmit a rejection message by allocating the DL resource. This is
advantageous in that the switching time is not necessary. The
predetermined time duration may be predefined according to a band,
a combination of bands, and/or a numerology or may be indicated
through RRC signaling.
[0120] When the parent node determines that the resource requested
by the child node is not valid, the parent node may inform the
child node of a new soft resource selected thereby without
transmitting the rejection message. If the requested resource is a
DL resource, the parent node may inform the child node of a
resource determined to be proper among DL resources (and FL
resources) through the DCI. If the requested resource is a UL
resource, the parent node may inform the child node of a resource
determined to be proper among UL resources (and FL resources)
through the DCI.
[0121] Alternatively, if the requested resource is an FL resource,
the parent node may inform the child node of a resource determined
to be proper among FL resources through the DCI. If the requested
resource is a DL or UL resource but the parent node informs the
child node of an FL resource, it may be assumed that the FL
resource is used according to whether the requested resource is a
DL resource or a UL resource. In this case, the parent node may
indicate, through the DCI, in which DL/UL/FL slot (or in which
symbol of the slot) the indicated resource is located from a slot
in which a PDCCH including the DCI is included. In this case, the
parent node may inform the child node of only a location of a slot
(or a location of a symbol of the slot) without distinguishing
between DL/UL/FL slots.
[0122] 3. The child node may make a request for one of the soft
DL/UL resources selected thereby together with the location of the
DL or UL resource from a requested timing. In particular, the child
node may make a request for the location of the DL or UL resource
including an FL resource. For example, upon requesting the DL
resource, the child node requests location of the DL resource among
DL resources and FL resources and, upon requesting a UL resource,
the child node requests location of the UL resource among UL
resources and FL resources. The parent node may determine whether
the requested resource is valid and transmit a confirmation message
or a rejection message. The parent node informs the child node of
the requested message through the DCI in order to dynamically
allocate resources.
[0123] When the child node fails to receive a special response from
the parent node during a predetermined time duration (e.g., Y ms),
the child node may determine that the parent node has rejected the
request. When the parent node switches a DL resource and a UL
resource several times, a switching time increases. Therefore, if
the request for the UL resource is rejected, the parent node may
transmit a rejection message by allocating the DL resource. This is
advantageous in that the switching time is not necessary. The
predetermined time duration may be predefined according to a band,
a combination of bands, and/or a numerology or may be indicated
through RRC signaling.
[0124] When the parent node determines that the resource requested
by the child node is not valid, the parent node may inform the
child node of a new soft resource selected thereby without
transmitting the rejection message. If the requested resource is a
DL resource, the parent node may inform the child node of a
resource determined to be proper among DL resources (and FL
resources) through the DCI. If the requested resource is a UL
resource, the parent node may inform the child node of a resource
determined to be proper among UL resources (and FL resources)
through the DCI.
[0125] Although the child node makes a request for a DL/UL
resource, if the parent node informs the child node of an FL
resource, the child node uses the FL resource as the resource
requested thereby, i.e., the DL resource or the UL resource. In
this case, the parent node may indicate, through the DCI, in which
DL/UL/FL slot (or in which symbol of the slot) the indicated
resource is located from a slot in which the DCI (or a PDCCH
including the DCI) is included. The parent node may inform the
child node of only a location of a slot (or a location of a symbol
of the slot) without distinguishing between DL/UL/FL slots.
[0126] 4. The child node may make a request only for use of a
resource and then the parent node may inform the child node of a
soft resource selected thereby. It is apparent that the parent node
rejects the request made by the child node through the DCI.
Assuming that the requested resource is not valid within a
predetermined time duration (e.g., X ms) from a requested timing,
the parent node may inform the child node that the request is
rejected through the DCI. The predetermined time duration may be
predefined according to a band, a combination of bands, and/or a
numerology or may be indicated through RRC signaling.
[0127] When the child node fails to receive a special response from
the parent node during a predetermined time duration (e.g., Y ms),
the child node may determine that the parent node has rejected the
request. When the parent node switches a DL resource and a UL
resource several times, a switching time increases. Therefore, if
the request for the UL resource is rejected, the parent node may
transmit a rejection message by allocating the DL resource. This is
advantageous in that the switching time is not necessary. The
predetermined time duration may be predefined according to a band,
a combination of bands, and/or a numerology or may be indicated
through RRC signaling.
Second Embodiment
[0128] Hereinafter, a soft resource will be described in terms of
CLI. The soft resource may be basically mapped to one resource
among DL/UL/FL resources, determined by the parent node. The soft
resource has many relations to CLI. For example, when neighbor
nodes need to transmit and receive a DL signal and a UL signal at
the same time, the UL signal may be subjected to strong CLI by the
DL signal. Generally, the parent node informs the child node of a
DL/UL/FL resource of the child node selected thereby in
consideration of occurrence of CLI. However, if the child node uses
a soft DL resource as a UL resource through arbitrary change or
uses a soft UL resource as a DL resource through arbitrary change,
CLI described above may occur. Therefore, the following is
proposed. [0129] If the parent node permits the child node to use
the soft resource, the child node always uses the soft DL resource
only as the DL resource. Likewise, if the parent node permits the
child node to use the soft resource, the child node always uses the
soft UL resource only as the UL resource.
[0130] The parent node may indicate DL beams or UL beams to be used
to the child node in terms of CLI. This may be regarded as an
attempt to previously coordinate beams by the parent node in order
to mitigate interference between nodes.
[0131] Specifically, the DL beam may be indicated using an SSB
index to which each beam is mapped. The UL beam may be indicated
using an SRS resource index to which each beam is mapped. The
contents of indication may be a beam set indicating multiple beams.
It is assumed that the child node uses beams only within the
indicated beams. Beams may be indicated only for the soft resource.
This is because the soft resource is a resource which is not
statically used and thus it is more difficult for the soft resource
to handle interference.
[0132] Beams may be indicated such that the child node informs the
parent node of information about beams that the child node desires
to use and the parent node checks the beam information and informs
the child node of beams. The child node may make a request for DL
beams using SSB indexes on a specific resource or make a request
for UL beams using SRS resource indexes. The parent node may
confirm or reject beams requested by the child node. Particularly,
when the parent node rejects the requested beams, the parent beam
may again inform the child node of beams (or beam set) selected
thereby. Beams may be indicated only for the soft resource. This is
because the soft resource is a resource which is not statically
used and thus it is more difficult for the soft resource to handle
interference.
[0133] Transmission power may also be considered in terms of CLI.
Hereinbelow, the transmission power may mean power spectral density
or direct transmission power.
[0134] The parent node may indicate DL power or UL power to be used
by the child node. Particularly, the parent node may cope with CLI
by semi-statically configuring the DL power (or power range) to be
used by the child node. The parent node may cope with CLI by
semi-statically configuring a UL power value (an allowable UL
reception power, a power range, a maximum allowable transmission
power, or a maximum allowable transmission power range) to be used
by the child node. Such power information may be indicated only for
the soft resource. This is because the soft resource is a resource
which is not statically used and thus it is more difficult for the
soft resource to handle interference.
[0135] It is apparent that the child node may make a request to the
parent node for power desired to be used thereby.
[0136] The child node may request DL transmission power or UL
reception power (or UL transmission power thereof) on a specific
resource. The parent node may confirm or reject power requested by
the child node. Particularly, when the parent node rejects power,
the parent node may again inform the child node of another power
(or power range). Power may be indicated only for the soft
resource. This is because the soft resource is a resource which is
not statically used and thus it is more difficult for the soft
resource to handle interference.
[0137] Since CLI differs according to whether an FL resource will
be used as a DL resource or as a UL resource, the child node may
inform the parent node of information as to whether FL resource is
used as the DL resource or as the UL resource.
[0138] The child node may inform the parent node of whether the
hard FL resource or the soft FL resource is used as the DL resource
or as the UL resource after determination. The parent node may
again inform the child node that the indicated resource should be
used for another purpose based on information indicated by the
child node. For example, although the child node indicates that the
FL resource is used as the DL resource to the parent node, the
parent node may command the child node to use the FL resource as
the UL resource. Alternatively, although the child node indicates
that the FL resource is used as the UL resource to the parent node,
the parent node may command the child not to use the FL resource as
the DL resource.
[0139] FIG. 10 is a flowchart illustrating resource allocation
according to an embodiment of the present disclosure.
[0140] Referring to FIG. 10, the child node transmits a resource
request message for requesting for one of a DL resource and a UL
resource to the parent node in step 1001. The resource request
message may include information about at which timing the requested
resource is located after the resource request message is
transmitted.
[0141] Next, in step 1003, the child node receives a resource
allocation grant message indicating one of the requested resource
and a flexible resource from the parent node. Particularly, when
the resource allocation grant message indicates the flexible
resource, the flexible resource is used as the requested
resource.
[0142] Finally, in step 1005, the child node transmits and receives
a signal to and from the parent node or another child node
controlled thereby, using the indicated resource.
[0143] Additionally, upon failing to receive the resource
allocation grant message until a predetermined time elapses after
the resource requested message is transmitted, the child node
regards allocation of the requested resource as being rejected by
the parent node. Alternatively, the child node may receive, from
the parent node, a resource allocation rejection message indicating
that the requested resource is not valid until a predetermined time
elapses after the resource requested message is transmitted. In
this case, the resource allocation rejection message may include
information about a resource other than the requested resource
among the DL resource and the UL resource.
[0144] FIG. 11 illustrates an example of a wireless communication
device according to an embodiment of the present disclosure.
[0145] The wireless communication device illustrated in FIG. 11 may
represent a UE and/or a BS according to an embodiment of the
present disclosure. However, the wireless communication device of
FIG. 11 may be replaced with any of various types of devices such
as a vehicle communication system or device, a wearable device, and
a laptop, not limited to the UE and/or the BS according to the
embodiment of the present disclosure. More specifically, the above
device may be a BS, a network node, a Tx UE, an Rx UE, a wireless
device, a wireless communication device, a vehicle, a vehicle
having a self-driving function, an unmanned aerial vehicle (UAV),
an artificial intelligence (AI) module, a robot, an augmented
reality (AR) device, a virtual reality (VR) device, a machine-type
communication (MTC) device, an Internet of things (IoT) device, a
medical device, a FinTech device (or a financial device), a
security device, a weather/environment device, or a device related
to the fourth industrial revolution or a 5G service. The UAV may
be, for example, an aircraft without a human being onboard, which
aviates by a wireless control signal. The MTC device and the IoT
device may be, for example, devices that do not require direct
human intervention or manipulation and may include smartmeters,
vending machines, thermometers, smartbulbs, door locks, or various
sensors. The medical device may be, for example, a device used for
the purpose of diagnosing, treating, relieving, curing, or
preventing disease or a device used for the purpose of inspecting,
replacing, or modifying a structure or a function and may include a
device for treatment, a device for operation, a device for (in
vitro) diagnosis, a hearing aid, or an operation device. The
security device may be, for example, a device installed to prevent
a danger that may arise and to maintain safety and may include a
camera, a CCTV, or a black box. The FinTech device may be, for
example, a device capable of providing a financial service such as
mobile payment and may include a payment device or a point of sale
(POS) system. The weather/environment device may be, for example, a
device for monitoring or predicting a weather/environment.
[0146] The Tx UE or the Rx UE may include, for example, a cellular
phone, a smartphone, a laptop computer, a digital broadcast
terminal, a personal digital assistant (PDA), a portable multimedia
player (PMP), a navigation system, a slate PC, a tablet PC, an
ultrabook, a wearable device (e.g., a smartwatch, smartglasses, or
a head mounted display (HMD)), or a foldable device. The HMD may
be, for example, a type of display device that is worn on the head
and may be used to implement VR or AR.
[0147] In the example of FIG. 11, the UE and/or the BS according to
the embodiment of the present disclosure includes at least one
processor 10 such as a digital signal processor or a
microprocessor, a transceiver 35, a power management module 5, an
antenna 40, a battery 55, a display 15, a keypad 20, a memory 30, a
subscriber identity module (SIM) card 25, a speaker 45, and a
microphone 50. In addition, the UE and/or the BS may include one or
more antennas. The transceiver 35 may be also referred to as an RF
module.
[0148] The at least one processor 10 may be configured to implement
the functions, procedures and/or methods. In at least some of the
embodiments, the at least one processor 10 may implement one or
more protocols, such as layers of radio interface protocols (e.g.,
functional layers).
[0149] The memory 30 is coupled to the at least one processor 10
and stores information related to the operations of the at least
one processor 10. The memory 30 may be located inside or outside
the at least one processor 10 and may be coupled to the at least
one processor 10 by various techniques such as wired or wireless
communication.
[0150] A user may input various types of information (e.g.,
indication information such as a telephone number) by various
techniques such as pressing a button on the keypad 20 or activating
voice using the microphone 50. The at least one processor 10
executes appropriate functions such as receiving and/or processing
information of the user and dialing a telephone number.
[0151] It is also possible to retrieve data (e.g., operational
data) from the SIM card 25 or the memory 30 to execute the
appropriate functions. In addition, the at least one processor 10
may receive and process global positioning system (GPS) information
from a GPS chip to obtain location information about the UE and/or
the BS such as in vehicle navigation, map service, or the like, or
execute functions related to the location information. Further, the
at least one processor 10 may display these various types of
information and data on the display 15 for reference and user
convenience.
[0152] The transceiver 35 is coupled to the at least one processor
10 to transmit and/or receive wireless signals such as RF signals.
The at least one processor 10 may control the transceiver 35 to
initiate communication and transmit wireless signals including
various types of information or data, such as voice communication
data. The transceiver 35 may include a receiver for receiving a
wireless signal and a transmitter for transmitting a wireless
signal. The antenna 40 facilitates the transmission and reception
of wireless signals. In some embodiments, upon receipt of a
wireless signal, the transceiver 35 may forward and convert the
signal to a baseband frequency for processing by the at least one
processor 10. The processed signal may be processed according to
various techniques, such as being converted into audible or
readable information, and output through the speaker 45.
[0153] In some embodiments, a sensor may also be coupled to the at
least one processor 10. The sensor may include one or more sensing
devices configured to detect various types of information,
including velocity, acceleration, light, vibration, and the like.
The at least one processor 10 receives and processes sensor
information obtained from the sensor, such as proximity, position,
image, and the like, thereby executing various functions such as
collision avoidance and autonomous driving.
[0154] Various components such as a camera and a universal serial
bus (USB) port may further be included in the UE and/or the BS. For
example, a camera may further be coupled to the at least one
processor 10, for use in various services including autonomous
driving and vehicle safety services.
[0155] FIG. 11 merely illustrates one example of devices included
in a UE and/or a BS, not limiting the present disclosure. For
example, some components, such as the keypad 20, the GPS chip, the
sensor, the speaker 45 and/or the microphone 50 may be excluded
from UE and/or BS implementation in some embodiments.
[0156] FIG. 12 illustrates an AI device 100 according to an
embodiment of the present disclosure.
[0157] The AI device 100 may be implemented by a stationary or
mobile device, for example, a TV, a projector, a mobile phone, a
smartphone, a desktop computer, a laptop computer, a digital
broadcasting terminal, a personal digital assistant (PDA), a
portable multimedia player (PMP), a navigation device, a tablet PC,
a wearable device, a set-top box (STB), a digital multimedia
broadcasting (DMB) receiver, a radio, a washing machine, a
refrigerator, a desktop computer, a digital signage, a robot, a
vehicle, etc.
[0158] Referring to FIG. 12, the AI device 100 may include a
communication unit 110, an input unit 120, a learning processor
130, a sensing unit 140, an output unit 150, a memory 170, and a
processor 180.
[0159] The communication unit 110 may transmit and receive data to
and from external devices such as an AI server 200 and other AI
devices 100a to 100e based on wired or wireless communication
technology. For example, the communication unit 110 may transmit
and receive sensor information, user inputs, learning models, and
control signals to and from the external devices.
[0160] The communication technology used by the communication unit
110 includes Global System for Mobile communication (GSM), Code
Division Multiple Access (CDM), Long Term Evolution (LTE), 5G,
Wireless Local Area Network (WLAN), Wireless Fidelity (Wi-Fi),
Bluetooth.TM., Radio Frequency Identification (RFID), Infrared Data
Association (IrDA), ZigBee, Near Field Communication (NFC),
etc.
[0161] The input unit 120 may obtain various types of data.
[0162] The input unit 120 may include a camera for inputting a
video signal, a microphone for receiving an audio signal, and a
user input unit for receiving information from a user. The camera
or microphone may be treated as a sensor, and the signal obtained
from the camera or microphone may be considered as sensing data or
sensor information.
[0163] The input unit 120 may obtain learning data for a learning
model and input data to be used when an output is obtained based on
the learning model. The input unit 120 may obtain raw input data.
In this case, the processor 180 or learning processor 130 may
extract an input feature by preprocessing the input data.
[0164] The learning processor 130 may train a model configured with
an ANN based on the learning data. Here, the trained ANN may be
referred to as the learning model. The learning model may be used
to infer a result value for new input data rather than the learning
data, and the inferred value may be used as a basis for determining
whether to perform a certain operation.
[0165] In this case, the learning processor 130 may perform AI
processing together with a learning processor 240 of the AI server
200.
[0166] The learning processor 130 may include a memory integrated
with or implemented in the AI device 100. Alternatively, the
learning processor 130 may be implemented with the memory 170, an
external memory directly coupled to the AI device 100, or a memory
in an external device.
[0167] The sensing unit 140 may obtain at least one of internal
information of the AI device 100, surrounding environment
information of the AI device 100, and user information using
various sensors.
[0168] The sensor included in the sensing unit 140 may include a
proximity sensor, an illumination sensor, an acceleration sensor, a
magnetic sensor, a gyro sensor, an inertial sensor, an RGB sensor,
an IR sensor, a fingerprint recognition sensor, an ultrasonic
sensor, an optical sensor, a microphone, a LIDAR, a radar, and the
like.
[0169] The output unit 150 may generate an output related to
visual, audible, or tactile sense.
[0170] The output unit 150 may include a display unit for
outputting visual information, a speaker for outputting audible
information, a haptic module for outputting tactile information,
and the like.
[0171] The memory 170 may store data supporting various functions
of the Al device 100. For example, the memory 170 may store input
data, learning data, learning models, learning histories, etc.
obtained by the input unit 120.
[0172] The processor 180 may determine at least one executable
operation of the AI device 100 based on information determined or
generated by a data analysis algorithm or machine learning
algorithm. The processor 180 may control the components of the AI
device 100 to perform the determined operation.
[0173] To this end, the processor 180 may request, search for,
receive, or employ data of the learning processor 130 or memory 170
and control the components of the AI device 100 to execute an
expected or preferable operation or among the one or more
executable operations.
[0174] If the processor 180 requires association with an external
device to perform the determined operation, the processor 180 may
generate a control signal for controlling the corresponding
external device and transmit the generated control signal to the
external device.
[0175] The processor 180 may obtain intention information from a
user input and determine the intention of the user based on the
obtained intention information.
[0176] In this case, the processor 180 may obtain the intention
information corresponding to the user input using at least one of a
speech-to-text (STT) engine for converting a voice input into a
character string or a natural language processing (NLP) engine for
obtaining intention information from a natural language.
[0177] At least one of the STT engine and the NLP engine may be
configured with the ANN of which at least a part is trained
according to the machine learning algorithm. At least one of the
STT engine and the NLP engine may be trained by the learning
processor 130, by the learning processor 240 of the AI server 200,
or by distributed processing thereof.
[0178] The processor 180 may collect history information including
user feedback on the operation of the AI device 100 and details
thereof. The processor 180 may store the history information in the
memory 170 or learning processor 130 or transmit the history
information to an external device such as the AI server 200. The
collected history information may be used to update the learning
model.
[0179] The processor 180 may control at least some of the
components of the AI device 100 to drive an application program
stored in the memory 170. Further, the processor 180 may operate
two or more of the components included in the AI device 100 in
combination to drive the application program.
[0180] FIG. 13 illustrates the AI server 200 according to an
embodiment of the present disclosure.
[0181] Referring to FIG. 13, the AI server 200 may mean a device
for training an ANN based on a machine learning algorithm or a
device for using a trained ANN. Here, the AI server 200 may include
a plurality of servers to perform distributed processing or may be
defined as a 5G network. The AI server 200 may be included as a
part of the AI device 100 to perform at least part of AI processing
together.
[0182] The AI server 200 may include a communication unit 210, a
memory 230, the learning processor 240, a processor 260, and the
like.
[0183] The communication unit 210 may transmit and receive data to
and from an external device such as the AI device 100.
[0184] The memory 230 may include a model storage unit 231. The
model storage unit 231 may store a model being trained or trained
(or an ANN 231a) through the learning processor 240.
[0185] The learning processor 240 may train the ANN 231a based on
learning data. The ANN, i.e., a learning model may be included in
the AI server 200 or in an external device such as the AI device
100.
[0186] The learning model may be implemented by hardware, software
or a combination thereof. If a part or the entirety of the learning
model is implemented with software, one or more instructions for
the learning model may be stored in the memory 230.
[0187] The processor 260 may infer a result value for new input
data based on the learning model and generate a response or control
command based on the inferred result value.
[0188] FIG. 14 illustrates an AI system 1 according to an
embodiment of the present disclosure.
[0189] Referring to FIG. 14, at least one of the AI server 200, a
robot 100a, an autonomous driving vehicle 100b, an XR device 100c,
a smartphone 100d, and a home appliance 100e is connected to a
cloud server 10 in the AI system 1. Here, the robot 100a, the
autonomous vehicle 100b, the XR device 100c, the smartphone 100d,
or the home appliance 100e, to which the AI technology is applied,
may be referred to as an AI device 100a to 100e.
[0190] The cloud network 10 may refer to a network configuring part
of a cloud computing infrastructure or a network existing in the
cloud computing infrastructure. Here, the cloud network 10 may be
configured with a 3G network, a 4G or LTE network, or a 5G
network.
[0191] That is, each of the devices 100a to 100e and 200 included
in the AI system 1 may be connected to each other through the cloud
network 10. In particular, the devices 100a to 100e and 200 may
communicate with each other through a BS or may communicate with
each other directly without the BS.
[0192] The AI server 200 may include a server in charge of AI
processing and a server in charge of big data computation.
[0193] The AI server 200 may be connected to at least one of the
robot 100a, the autonomous vehicle 100b, the XR device 100c, the
smartphone 100d, or the home appliance 100e included in the AI
system 1 via the cloud network 10 and help at least part of AI
processing of the connected AI devices 100a to 100e.
[0194] In this case, the AI server 200 may train an ANN according
to a machine learning algorithm on behalf of the AI devices 100a to
100e and directly store or transmit a learning model to the AI
devices 100a to 100e.
[0195] The AI server 200 may receive input data from the AI devices
100a to 100e, infer a result value for the received input data
based on the learning model, generate a response or control command
based on the inferred result value, and transmit the response or
control command to the AI devices 100a to 100e.
[0196] Alternatively, the AI devices 100a to 100e may directly
infer the result value for the input data based on the learning
model and generate the response or control command based on the
inferred result value.
[0197] Hereinafter, various embodiments of the AI devices 100a to
100e to which the above-described technology is applied will be
described. The AI devices 100a to 100e illustrated in FIG. 14 may
be considered as a specific example of the AI device 100
illustrated in FIG. 13.
[0198] <AI+Robot>
[0199] If the AI technology is applied to the robot 100a, the robot
100a may be implemented as a guide robot, a transport robot, a
cleaning robot, a wearable robot, an entertainment robot, a pet
robot, an unmanned flying robot, etc.
[0200] The robot 100a may include a robot control module for
controlling an operation, and the robot control module may refer to
a software module or a chip implemented by hardware.
[0201] The robot 100a may obtain state information of the robot
100a, detect (recognize) a surrounding environment and objects,
generate map data, determine a travel route or driving plan, or
determine a response or action to user interaction by using sensor
information obtained from various types of sensors.
[0202] To determine the travel route or driving plan, the robot
100a may use sensor information obtained from at least one of the
following sensors: a LIDAR, a radar, and a camera to determine a
movement route and a travel plan.
[0203] The robot 100a may perform the above-described operations
based on a learning model configured with at least one ANN. For
example, the robot 100a may recognize the surrounding environment
and objects based on the learning model and determine an operation
based on the recognized surrounding environment or object. Here,
the learning model may be directly trained by the robot 100a or by
an external device such as the AI server 200.
[0204] The robot 100a may operate by directly generating a result
based on the learning model. Alternatively, the robot 100a may
transmit sensor information to the external device such as the AI
server 200 and receive a result generated based on the sensor
information.
[0205] The robot 100a may determine the travel route and driving
plan based on at least one of the map data, the object information
detected from the sensor information, or the object information
obtained from the external device. Then, the robot 100a may move
according to the determined travel path and driving plan under
control of its driving unit.
[0206] The map data may include object identification information
about various objects placed in a space in which the robot 100a
moves. For example, the map data may include object identification
information about fixed objects such as walls and doors and movable
objects such as flower pots and desks. The object identification
information may include a name, a type, a distance, a position,
etc.
[0207] The robot 100a may operate and move by controlling the
driving unit based on the user control/interaction. In this case,
the robot 100a may obtain intention information from the motion or
speech of the user and determine a response based on the obtained
intention information.
[0208] <AI+Autonomous Driving>
[0209] If the AI technology is applied to the autonomous driving
vehicle 100b, the autonomous driving vehicle 100b may be
implemented as a mobile robot, a vehicle, an unmanned flying
vehicle, etc.
[0210] The autonomous driving vehicle 100b may include an
autonomous driving control module for controlling the autonomous
driving function, and the autonomous driving control module may
refer to a software module or a chip implemented by hardware. The
autonomous driving control module may be included in the autonomous
driving vehicle 100b as a component thereof, but it may be
implemented with separate hardware and connected to the outside of
the autonomous driving vehicle 100b.
[0211] The autonomous driving vehicle 100b may obtain state
information about the autonomous driving vehicle 100b based on
sensor information acquired from various types of sensors, detect
(recognize) a surrounding environment and objects, generate map
data, determine a travel route and driving plan, or determine an
operation.
[0212] Similarly to the robot 100a, the autonomous driving vehicle
100b may use the sensor information obtained from at least one of
the following sensors: a LIDAR, a radar, and a camera so as to
determine the travel route and driving plan.
[0213] In particular, the autonomous driving vehicle 100b may
recognize an environment and objects in an area hidden from view or
an area over a certain distance by receiving the sensor information
from external devices. Alternatively, the autonomous driving
vehicle 100b may receive information, which is recognized by the
external devices.
[0214] The autonomous driving vehicle 100b may perform the
above-described operations based on a learning model configured
with at least one ANN. For example, the autonomous driving vehicle
100b may recognize the surrounding environment and objects based on
the learning model and determine the driving path based on the
recognized surrounding environment and objects. The learning model
may be trained by the autonomous driving vehicle 100a or an
external device such as the AI server 200.
[0215] The autonomous driving vehicle 100b may operate by directly
generating a result based on the learning model. Alternatively, the
autonomous driving vehicle 100b may transmit sensor information to
the external device such as the AI server 200 and receive a result
generated based on the sensor information.
[0216] The autonomous driving vehicle 100b may determine the travel
route and driving plan based on at least one of the map data, the
object information detected from the sensor information, or the
object information obtained from the external device. Then, the
autonomous driving vehicle 100b may move according to the
determined travel path and driving plan under control of its
driving unit.
[0217] The map data may include object identification information
about various objects placed in a space (e.g., road) in which the
autonomous driving vehicle 100b moves. For example, the map data
may include object identification information about fixed objects
such as street lamps, rocks, and buildings and movable objects such
as vehicles and pedestrians. The object identification information
may include a name, a type, a distance, a position, etc.
[0218] The autonomous driving vehicle 100b may operate and move by
controlling the driving unit based on the user control/interaction.
In this case, the autonomous driving vehicle 100b may obtain
intention information from the motion or speech of a user and
determine a response based on the obtained intention
information.
[0219] <AI+XR>
[0220] When the AI technology is applied to the XR device 100c, the
XR device 100c may be implemented as a HMD, a HUD mounted in
vehicles, a TV, a mobile phone, a smartphone, a computer, a
wearable device, a home appliance, a digital signage, a vehicle, a
fixed robot, a mobile robot, etc.
[0221] The XR device 100c may analyze three-dimensional point cloud
data or image data obtained from various sensors or external
devices, generate position data and attribute data for
three-dimensional points, obtain information about a surrounding
environment or information about a real object, perform rendering
to on an XR object, and then output the XR object. For example, the
XR device 100c may output an XR object including information about
a recognized object, that is, by matching the XR object with the
recognized object.
[0222] The XR device 100c may perform the above-described
operations based on a learning model configured with at least one
ANN. For example, the XR device 100c may recognize the real object
from the three-dimensional point cloud data or image data based on
the learning model and provide information corresponding to the
recognized real object. The learning model may be directly trained
by the XR device 100c or an external device such as the AI server
200.
[0223] The XR device 100c may operate by directly generating a
result based on the learning model. Alternatively, the XR device
100c may transmit sensor information to the external device such as
the AI server 200 and receive a result generated based on the
sensor information.
[0224] <AI+Robot+Autonomous Driving>
[0225] When the AI technology and the autonomous driving technology
are applied to the robot 100a, the robot 100a may be implemented as
a guide robot, a transport robot, a cleaning robot, a wearable
robot, an entertainment robot, a pet robot, an unmanned flying
robot, etc.
[0226] The robot 100a to which the AI technology and the autonomous
driving technology are applied may refer to the robot 100a with the
autonomous driving function or the robot 100a interacting with the
autonomous driving vehicle 100b.
[0227] The robot 100a having the autonomous driving function may be
collectively referred to as a device that move along a given
movement path without human control or a device that moves by
autonomously determining its movement path.
[0228] The robot 100a having the autonomous driving function and
the autonomous driving vehicle 100b may use a common sensing method
to determine either a travel route or a driving plan. For example,
the robot 100a having the autonomous driving function and the
autonomous driving vehicle 100b may determine either the travel
route or the driving plan based on information sensed through a
LIDAR, a radar, and a camera.
[0229] The robot 100a interacting with the autonomous driving
vehicle 100b may exist separately from with the autonomous driving
vehicle 100b. That is, the robot 100a may perform operations
associated with the autonomous driving function inside or outside
the autonomous driving vehicle 100b or interwork with a user on the
autonomous driving vehicle 100b.
[0230] The robot 100a interacting with the autonomous driving
vehicle 100b may control or assist the autonomous driving function
of the autonomous driving vehicle 100b by obtaining sensor
information on behalf of the autonomous driving vehicle 100b and
providing the sensor information to the autonomous driving vehicle
100b or by obtaining sensor information, generating environment
information or object information, and providing the information to
the autonomous driving vehicle 100b.
[0231] Alternatively, the robot 100a interacting with the
autonomous driving vehicle 100b may monitor the user on the
autonomous driving vehicle 100b or control the autonomous driving
vehicle 100b through the interaction with the user. For example,
when it is determined that the driver is in a drowsy state, the
robot 100a may activate the autonomous driving function of the
autonomous driving vehicle 100b or assist the control of the
driving unit of the autonomous driving vehicle 100b. The function
of the autonomous driving vehicle 100b controlled by the robot 100a
may include not only the autonomous driving function but also
functions installed in the navigation system or audio system
provided in the autonomous driving vehicle 100b.
[0232] Alternatively, the robot 100a interacting with the
autonomous driving vehicle 100b may provide information to the
autonomous driving vehicle 100b outside the autonomous driving
vehicle 100b or assist the autonomous driving vehicle 100b outside
the autonomous driving vehicle 100b. For example, the robot 100a
may provide traffic information including signal information such
as smart traffic lights to the autonomous driving vehicle 100b or
automatically connect an electric charger to a charging port by
interacting with the autonomous driving vehicle 100b like an
automatic electric charger installed in an electric vehicle.
[0233] <AI+Robot+XR>
[0234] When the AI technology and the XR technology are applied to
the robot 100a, the robot 100a may be implemented as a guide robot,
a transport robot, a cleaning robot, a wearable robot, an
entertainment robot, a pet robot, an unmanned flying robot, a
drone, etc.
[0235] The robot 100a to which the XR technology is applied may
refer to a robot subjected to control/interaction in an XR image.
In this case, the robot 100a may be separated from the XR device
100c but interact with the XR device 100c.
[0236] When the robot 100a subjected to control/interaction in the
XR image obtains sensor information from sensors including a
camera, the robot 100a or XR device 100c may generate the XR image
based on the sensor information, and then the XR device 100c may
output the generated XR image. The robot 100a may operate based on
a control signal input through the XR device 100c or user
interaction.
[0237] For example, a user may confirm the XR image corresponding
to the perspective of the robot 100a remotely controlled through an
external device such as the XR device 100c. Then, the user may
adjust the autonomous driving path of the robot 100a or control the
operation or movement of the robot 100a through interaction
therewith or check information about surrounding objects.
[0238] <AI+Autonomous Driving+XR>
[0239] When the AI technology and the XR technology is applied to
the autonomous driving vehicle 100b, the autonomous driving vehicle
100b may be implemented as a mobile robot, a vehicle, an unmanned
flying vehicle, etc.
[0240] The autonomous driving vehicle 100b to which the XR
technology is applied may refer to an autonomous driving vehicle
capable of providing an XR image or an autonomous driving vehicle
subjected to control/interaction in an XR image. In particular, the
autonomous driving vehicle 100b subjected to control/interaction in
the XR image may be separated from the XR device 100c but interact
with the XR device 100c.
[0241] The autonomous driving vehicle 100b capable of providing the
XR image may obtain sensor information from sensors including a
camera and output the generated XR image based on the obtained
sensor information. For example, the autonomous driving vehicle
100b may include an HUD for outputting an XR image, thereby
providing a user with an XR object corresponding to an object in
the screen together with a real object.
[0242] When the XR object is displayed on the HUD, at least part of
the XR object may overlap with the real object which the user looks
at. On the other hand, when the XR object is displayed on a display
provided in the autonomous driving vehicle 100b, at least part of
the XR object may overlap with the object in the screen. For
example, the autonomous driving vehicle 100b may output XR objects
corresponding to objects such as a lane, another vehicle, a traffic
light, a traffic sign, a two-wheeled vehicle, a pedestrian, a
building, etc.
[0243] When the autonomous driving vehicle 100b subjected to
control/interaction in the XR image may obtain the sensor
information from the sensors including the camera, the autonomous
driving vehicle 100b or the XR device 100c may generate the XR
image based on the sensor information, and then the XR device 100c
may output the generated XR image. The autonomous driving vehicle
100b may operate based on a control signal input through an
external device such as the XR device 100c or user interaction.
[0244] The embodiments of the present disclosure described herein
below are combinations of elements and features of the present
disclosure. The elements or features may be considered selective
unless otherwise mentioned. Each element or feature may be
practiced without being combined with other elements or features.
Further, an embodiment of the present disclosure may be constructed
by combining parts of the elements and/or features. Operation
orders described in embodiments of the present disclosure may be
rearranged. Some constructions of any one embodiment may be
included in another embodiment and may be replaced with
corresponding constructions of another embodiment. It will be
obvious to those skilled in the art that claims that are not
explicitly cited in each other in the appended claims may be
presented in combination as an embodiment of the present disclosure
or included as a new claim by a subsequent amendment after the
application is filed.
[0245] In the embodiments of the present disclosure, a description
is made centering on a data transmission and reception relationship
among a BS, a relay, and an MS. In some cases, a specific operation
described as performed by the BS may be performed by an upper node
of the BS. Namely, it is apparent that, in a network comprised of a
plurality of network nodes including a BS, various operations
performed for communication with an MS may be performed by the BS,
or network nodes other than the BS. The term `BS` may be replaced
with the term `fixed station`, `Node B`, `enhanced Node B (eNode B
or eNB)`, `access point`, etc. The term `UE` may be replaced with
the term `mobile station (MS)`, `mobile subscriber station (MSS)`,
`mobile terminal`, etc.
[0246] The embodiments of the present disclosure may be achieved by
various means, for example, hardware, firmware, software, or a
combination thereof. In a hardware configuration, the methods
according to the embodiments of the present disclosure may be
achieved by one or more application specific integrated circuits
(ASICs), digital signal processors (DSPs), digital signal
processing devices (DSPDs), programmable logic devices (PLDs),
field programmable gate arrays (FPGAs), processors, controllers,
microcontrollers, microprocessors, etc.
[0247] In a firmware or software configuration, the embodiments of
the present disclosure may be implemented in the form of a module,
a procedure, a function, etc. For example, software code may be
stored in a memory unit and executed by a processor. The memory
unit is located at the interior or exterior of the processor and
may transmit and receive data to and from the processor via various
known means.
[0248] Those skilled in the art will appreciate that the present
disclosure may be carried out in other specific ways than those set
forth herein without departing from the spirit and essential
characteristics of the present disclosure. The above embodiments
are therefore to be construed in all aspects as illustrative and
not restrictive. The scope of the disclosure should be determined
by the appended claims and their legal equivalents, not by the
above description, and all changes coming within the meaning and
equivalency range of the appended claims are intended to be
embraced therein.
* * * * *