U.S. patent application number 15/185483 was filed with the patent office on 2017-07-06 for augmented reality device based on recognition of spatial structure and method thereof.
The applicant listed for this patent is ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE. Invention is credited to Hyun Woo CHO, Sung Uk JUNG, Young Woon LEE.
Application Number | 20170193299 15/185483 |
Document ID | / |
Family ID | 59235637 |
Filed Date | 2017-07-06 |
United States Patent
Application |
20170193299 |
Kind Code |
A1 |
CHO; Hyun Woo ; et
al. |
July 6, 2017 |
AUGMENTED REALITY DEVICE BASED ON RECOGNITION OF SPATIAL STRUCTURE
AND METHOD THEREOF
Abstract
An augmented reality device based on recognition of a spatial
structure includes: a point cloud normal vector extracting unit
extracting a normal vector for a point cloud from image data input
from a camera; a plane object segmenting unit segmenting a plane in
the image data by using the extracted normal vector; a
representative plane selecting unit selecting a representative
plane among the segmented planes; a spatial structure extracting
unit recognizing a spatial structure by using the representative
plane; and a virtual object matching unit matching a virtual object
in the recognized spatial structure.
Inventors: |
CHO; Hyun Woo; (Daejeon,
KR) ; LEE; Young Woon; (Daejeon, KR) ; JUNG;
Sung Uk; (Daejeon, KR) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE |
Daejeon |
|
KR |
|
|
Family ID: |
59235637 |
Appl. No.: |
15/185483 |
Filed: |
June 17, 2016 |
Current U.S.
Class: |
1/1 |
Current CPC
Class: |
G06T 19/006 20130101;
G06T 2207/10012 20130101; G06K 9/6202 20130101; G06K 9/00 20130101;
G06K 9/00201 20130101; G06T 2207/10028 20130101; G06T 2207/30244
20130101; G06K 9/00671 20130101 |
International
Class: |
G06K 9/00 20060101
G06K009/00; G06T 7/00 20060101 G06T007/00; G06K 9/62 20060101
G06K009/62 |
Foreign Application Data
Date |
Code |
Application Number |
Jan 5, 2016 |
KR |
10-2016-0001034 |
Claims
1. An augmented reality device based on recognition of a spatial
structure, the augmented reality device comprising: a point cloud
normal vector extracting unit extracting a normal vector for a
point cloud from image data input from a camera; a plane object
segmenting unit segmenting a plane in the image data by using the
extracted normal vector; a representative plane selecting unit
selecting a representative plane among the segmented planes; a
spatial structure extracting unit recognizing a spatial structure
by using the representative plane; and a virtual object matching
unit matching a virtual object in the recognized spatial
structure.
2. The augmented reality device of claim 1, wherein the plane
object segmenting unit segments the plane for each direction in
which the point cloud is crowded by using the normal vector.
3. The augmented reality device of claim 1, wherein the
representative plane includes at least one of a bottom plane, a
rear wall, and a ceiling.
4. The augmented reality device of claim 1, wherein the
representative plane selecting unit selects the representative
plane by using a contextual knowledge of the camera and an indoor
space configuration.
5. The augmented reality device of claim 4, wherein in the
contextual knowledge for selecting the bottom plane among the
representative planes, the bottom plane is positioned on the bottom
of the camera and the camera vector is vertical to normal vectors
of tire ceiling plane and the bottom plane among the representative
planes.
6. The augmented reality device of claim 4, wherein in the
contextual knowledge for selecting the rear wall plane among the
representative planes, the normal vectors of the bottom plane an
the wall plane among the representative planes are vertical to each
other, the normal vectors of the ceiling plane and the wall plane
among the representative planes are vertical to each other, and the
rear wall plane is spaced apart from the camera by a predetermined
interval.
7. The augmented reality device of claim 4, wherein in the
contextual knowledge for selecting the ceiling plane among the
representative planes, the normal vectors of the ceiling plane and
the wall plane are vertical to each other and the normal vectors of
the bottom plane and the ceiling plane among the representative
planes are parallel to each other.
8. The augmented reality device of claim 1, wherein the spatial
structure extracting unit extracts a rectangle having a minimum
area while including all point clouds of the selected
representative plane.
9. The augmented reality device of claim 8, wherein the spatial
structure extracting unit fits the rectangle to an intersection by
acquiring the interaction of the selected representative plane to
extract the spatial structure.
10. The augmented reality device of claim 9, wherein the spatial
structure extracting unit matches a real object to the extracted
spatial structure by using the representative plane and 3D posture
information of a real object.
11. The augmented reality device of claim 10, wherein the 3D
posture information includes each corner information and normal
vectors of the representative plane and the real object.
12. The augmented reality device of claim 1, wherein the virtual
object matching unit matches the virtual object in the recognized
spatial structure by using the normal vector of the representative
plane and internal angle component information of a corner.
13. The augmented reality device of claim 1, wherein the camera is
a depth camera.
14. An augmented reality method based on recognition of a spatial
structure, the augmented reality method comprising: segmenting a
plane from input image data; selecting a representative plane among
the segmented planes; recognizing a spatial structure by using the
representative plane; and matching a virtual object in the
recognized spatial structure.
15. The augmented reality method of claim 14, wherein: the
segmenting of the plane includes extracting a normal vector for a
point cloud from the input image data; and segmenting the plane for
each direction in which a point cloud is crowded by using the
extracted normal vector.
16. The augmented reality method of claim 14, wherein in the
selecting of the representative plane, the representative plane is
selected by using a contextual knowledge of a camera and an indoor
space configuration.
17. The augmented reality method of claim 14, wherein the
recognizing of the spatial structure includes extracting a
rectangle having a minimum area while including all point clouds of
the selected representative plane, and fitting the rectangle to an
intersection by acquiring the interaction of the selected
representative plane to extract the spatial structure.
18. The augmented reality method of claim 17, wherein the
recognizing of the spatial structure further includes matching a
real object with the extracted spatial structure by using the
representative plane and 3D posture information of a real
object.
19. The augmented reality method of claim 17, wherein in the
matching of the virtual object, the virtual object is matched in
the recognized spatial structure by using the normal vector of the
representative plane and internal angle component information of a
corner.
28. The augmented reality method of claim 17, wherein the
representative plane includes at least one of a bottom plane, a
rear wall, and a ceiling.
Description
CROSS-REFERENCE TO RELATE(c) APPLICATIONS
[0001] This application claims priority to and the benefit of
Korean Patent Application No. 10-2016-0001034 fifed in the Korean
Intellectual Property Office on Jan. 5, 2016, the entire contents
of which are incorporated herein by reference.
BACKGROUND OF THE INVENTION
[0002] 1. Field of the Invention
[0003] The present invention relates to an augmented reality device
based on a recognition of spatial structure and a method thereof,
and more particularly, to a technology that reconfigures a 3D
spatial structure based on plane information and provides augmented
reality based on the reconfigured spatial structure.
[0004] 2. Description of Related Art
[0005] An augmented reality technology as a technology field
derived from a virtual reality technology that creates a virtual
space similar to reality or further from the reality by using a
computer graphic technology means a technology that synchronizes a
reality space and the virtual space and synthesizes a virtual
object with the reality to look as if the synthesized virtual
object exists in an original reality space.
[0006] With the development of a graphic technology, since the
virtual space which is more realistic than the reality is created,
but an augmented reality technology synthesizes the virtual object
on the reality space to duplicatively show the synthesized virtual
object unlike the virtual reality technology that makes a sense of
difference be felt, the reality space is reinforced with a space
which is fit for a purpose of contents by augmenting additional
information or the virtual object which is small in sense of
difference and cannot be obtained in the reality space.
[0007] However, in spite of such an advantage, in the related art,
a 3D posture of the space is estimated by using a marker or
markerless in order to implement the augmented reality and the
technology has a problem in that as recognition is performed by
using a pre-learned image feature or when the image feature does
not exist in the space, a space to augment the virtual object
cannot be specified, and as a result, 3D posture information cannot
be obtained, thereby lowering accuracy of the implementation of the
augmented reality.
SUMMARY OF THE INVENTION
[0008] The present invention has been made in an effort to provide
an augmented reality device based on recognition of a spatial
structure and a method thereof which provide an augmented reality
service based on recognition of an indoor spatial structure to
provide an accurate augmented reality service without pre-learning
an indoor space.
[0009] The technical objects of the present invention are not
limited to the aforementioned technical objects, and other
technical objects, which are not mentioned above, will be
apparently appreciated by a person having ordinary skill in the art
from the following description.
[0010] An exemplary embodiment of the present invention provides an
augmented reality device based on recognition of a spatial
structure, including: a point cloud normal vector extracting unit
extracting a normal vector for a point cloud from image data input
from a camera; a plane object segmenting unit segmenting a plane in
the image data by using the extracted normal vector; a
representative plane selecting unit selecting a representative
plane among the segmented planes; a spatial structure extracting
unit recognizing a spatial structure by using the representative
plane; and a virtual object matching unit matching a virtual object
in the recognized spatial structure.
[0011] The plane object segmenting unit may segment the plane for
each direction in which the point cloud is crowded by using the
normal vector.
[0012] The representative plane may include at least one of a
bottom plane, a rear wall, and a ceiling.
[0013] The representative plane selecting unit may select the
representative plane by using a contextual knowledge of the camera
and an indoor space configuration.
[0014] In the contextual knowledge for selecting the bottom plane
among the representative planes, the bottom plane may be positioned
on the bottom of the camera and the camera vector may be vertical
to normal vectors of the ceiling plane and the bottom plane among
the representative planes.
[0015] In the contextual knowledge for selecting the rear wall
plane among the representative planes, the normal vectors of the
bottom plane an the wall plane among the representative planes may
be vertical to each other, the normal vectors of the ceiling plane
and the wall plane among the representative planes may be vertical
to each other, and the rear wall plane may be spaced apart torn the
camera by a predetermined interval.
[0016] In the contextual knowledge for selecting the ceiling plane
among the representative planes, the normal vectors of the ceiling
plane and the wall plane may be vertical to each other and the
normal vectors of the bottom plane and the ceiling plane among the
representative planes may be parallel to each other.
[0017] The spatial structure extracting unit may extract a
rectangle having a minimum area while including all point clouds of
the selected representative plane.
[0018] The spatial structure extracting, unit may fit the rectangle
to an intersection by acquiring the interaction of the selected
representative plane to extract the spatial structure.
[0019] The spatial structure extracting unit may match a real
object to the extracted spatial structure by using the
representative plane and 3D posture information of a real
object.
[0020] The 3D posture information may include each corner
information and normal vectors of the representative plane and the
real object.
[0021] The virtual object matching unit may match the virtual
object in the recognized spatial structure by using the normal
vector of the representative plane and internal angle component
information of a corner.
[0022] The camera may be a depth camera.
[0023] Another exemplary embodiment of tire present invention
provides an augmented reality method based on recognition of a
spatial structure, including: segmenting a plane from input image
data; selecting a representative plane among the segmented planes;
recognizing a spatial structure by using the representative plane;
and matching a virtual object in the recognized spatial
structure.
[0024] The segmenting of the plane may include extracting a normal
vector for a point cloud from the input image data; and segmenting
the plane far each direction in which a point cloud is crowded by
using the extracted normal vector.
[0025] In the selecting of the representative plane, the
representative plane may be selected by using a contextual
knowledge of a camera and an indoor space configuration.
[0026] The recognizing of the spatial structure may include
extracting a rectangle having a minimum area while including all
point clouds of the selected representative plane, and fitting the
rectangle to an intersection by acquiring the interaction of the
selected representative plans to extract the spatial structure.
[0027] The recognizing at the spatial structure may further include
matching a real object with the extracted spatial structure by
using the representative plane and 3D posture information of a real
object.
[0028] In the matching of the virtual object, the virtual object
may be matched in the recognized spatial structure by using the
normal vector of the representative plane and internal angle
component information of a corner.
[0029] The representative plane may include at least one of a
bottom plane, a rear wall, and a ceiling.
[0030] According to exemplary embodiments of the present invention,
a 3D space can be recognized without pre-learning a reality space
and a virtual object is accurately matched to the recognized 3D
space by obtaining 3D posture information to increase accuracy of
augmented reality.
[0031] The exemplary embodiments of the present invention are
illustrative only, and various modifications, changes,
substitutions, and additions may be made without departing from the
technical spirit and scope of the appended claims by those skilled
in the art, and it will be appreciated that the modifications and
changes are included in the appended claims.
BRIEF DESCRIPTION OF THE DRAWINGS
[0032] FIG. 1 is a configuration diagram of an augmented reality
device using recognition of a spatial structure according to an
exemplary embodiment of the present invention.
[0033] FIG. 2 is a flowchart illustrating an augmented reality
method based on spatial recognition of an augmented reality device
according to an exemplary embodiment of the present invention.
[0034] FIG. 3 is a spatial configuration diagram for recognition of
an indoor spatial structure by the augmented reality device
according to the exemplary embodiment of the present invention.
[0035] FIG. 4 is a cross-sectional view of a space configuration of
FIG. 3 viewed from the side.
[0036] FIG. 5 is an exemplary diagram illustrating plane object
segmentation performed by the augmented reality device according to
the exemplary embodiment of the present invention.
[0037] FIG. 6 is an exemplary diagram illustrating selection of a
representative plane from segmented planes in FIG. 5.
[0038] FIG. 7 is an exemplary diagram illustrating recognition of
the spatial structure for the representative plane selected in FIG.
6.
[0039] FIG. 8 is an exemplary diagram illustrating matching a real
object to the spatial structure recognized in FIG. 7.
[0040] FIG. 9 is an exemplary diagram illustrating matching a
virtual object to the spatial structure matched with the real
object in FIG. 8.
[0041] FIG. 10 is a configuration diagram of a computer system to
which an augmented reality device is applied according to an
exemplary embodiment of the present invention.
[0042] It should be understood that the appended drawings are not
necessarily to scale, presenting a somewhat simplified
representation of various features illustrative of the basic
principles of the invention. The specific design features of the
present invention as disclosed herein, including, for example,
specific dimensions, orientations, locations, and shapes will be
determined in part by the particular intended application and use
environment.
[0043] In the figures, reference numbers refer to the same or
equivalent parts of the present invention throughout the several
figures of the drawing.
DETAILED DESCRIPTION
[0044] Hereinafter, some exemplary embodiments of the present
invention will be described in detail with reference to the
exemplary drawings. When reference numerals refer to components of
each drawing, it is noted that although the same components are
illustrated in different drawings, the same components are
designated by the same reference numerals as possible. In
describing the exemplary embodiments of the present invention, when
it is determined that the detailed description of the known
components and functions related to the present invention may
obscure understanding of the exemplary embodiments of the present
invention, the detailed description thereof will be omitted.
[0045] Terms such as first, second, A, B, (a), (b), and the like
may be used in describing the components of the exemplary
embodiments of the present invention. The terms are only used to
distinguish a component from another component, but nature or an
order of the component is not limited by the terms. Further, if it
is not contrarily defined, all terms used herein including
technological or scientific terms have the same meanings as those
generally understood by a person with ordinary skill in the art.
Terms which are defined in a generally used dictionary should be
interpreted to have the same meaning as the meaning in the context
of the related art, and are not interpreted as ideal or excessively
formal meanings unless clearly defined in the present
application.
[0046] According to the present invention, in a space based
augmented reality device using a camera and a display device, a 3D
structure of an indoor space is reconfigured based on plant
information and 3D posture information is extracted to implement
augmented reality, thereby improving accuracy of an augmented
reality technology.
[0047] Hereinafter, exemplary embodiments of the present invention
will be described in detail with inference to FIGS. 1 to 10.
[0048] FIG. 1 is a configuration diagram of an augmented reality
device based on recognition of a spatial structure according to an
exemplary embodiment of the present invention.
[0049] Referring to FIG. 1, the augmented reality device according
to the exemplary embodiment of fee present invention may include a
point cloud normal vector extracting unit 110, a plane object
segmenting unit 120, a representative plane selecting unit 130, a
spatial structure extracting unit 140, a virtual object matching
unit 150, and a camera 200. In this case, as the camera 200, a
depth camera may be provided.
[0050] The point cloud normal vector extracting unit 110 converts
raw data of a depth image input from the camera 200 into 3D point
cloud values x, y, and z. Thereafter, the point cloud normal vector
extracting unit 110 calculates and stores a normal vector for the
converted point cloud. Referring to FIGS. 3 and 4, it can be seen
that a point cloud 210 in which the point cloud values are crowded
is displayed based on the camera 200. FIG. 3 is a spatial
configuration diagram for recognition of an indoor spatial
structure by the augmented reality device according to the
exemplary embodiment of the present invention and FIG. 4 is a
cross-sectional view of a space configuration of FIG. 3 viewed from
the side and displays that a normal vector based point cloud 210 is
crowded in a spatial structure 10.
[0051] The plane object segmenting unit 120 segments plane
information on an indoor space by using normal vectors Vx, Vy, and
Vz calculated by the point cloud normal vector extracting unit 110.
In this case, the plane object segmenting unit 110 segments the
plane information by using RANdom SAmple Consensus (RANSAC), a
region growing segmentation technique that splits an image into
small regions and segments the image into regions which finally
remain while combining similar regions by calculating a difference
in color sense or brightness between adjacent regions, and the
like. In this case, the plans object segmenting unit 120 obtains an
equation of a plane, a point cloud in the plane, and a center point
of the plane through the plane information segmentation. FIG. 5 is
an exemplary diagram illustrating plane object segmentation
performed by the augmented reality device according to the
exemplary embodiment of the present invention.
[0052] Referring to FIG. 5, the plane object segmenting unit 120
may classify the plane according to a crowing status of the point
cloud for each normal vector and classify the plane into a bottom
plane region 310, a real wall plane region 320, and a ceiling
region 330. In this case, since the respective regions 316, 320,
and 330 of FIG. 5 are not classified into rectangular shapes and
the point cloud is not shown at a corner portion, and the like not
to be included in the region, and as a result, it may be difficult
to accurately recognize the spatial structure. This may be
recognized by using the plane equation afterwards.
[0053] The representative plane selecting unit 130 performs an
operation of selecting the representative plane required for
recognizing the spatial structure in the plane information obtained
by the plane object segmenting unit 120. In this case, the
representative plane may include a bottom, a rear wall, a ceiling,
and the like.
[0054] The representative plane selecting unit 130 selects the
representative plane by using a contextual knowledge of the camera
and the indoor space configuration. When the contextual knowledge
of the camera and the indoor space configuration is described with
reference to FIGS. 3 and 4, the contextual knowledge of the indoor
space configuration will be described below.
[0055] {circle around (1)} A bottom plane is positioned on the
bottom of the camera.
[0056] {circle around (2)} The normal vectors of the bottom and the
wall are vertical to each other.
[0057] {circle around (3)} The normal vectors of the wall and the
ceiling are vertical to each other.
[0058] {circle around (4)} The camera vector of FIG. 3 is vertical
to the normal vectors of the ceiling and the bottom.
[0059] {circle around (5)} The rear wall is distant from the camera
by 3 m or more.
[0060] {circle around (6)} The normal vectors of the bottom and the
ceiling are parallel to each other.
[0061] The representative plane selecting unit 130 primarily
classifies the vector of the camera 200 and the normal vector of
the plane by comparing the vector of the camera 200 and the normal
vector of the plane with a threshold designated by a user in order
to find the bottom plane P.sub.floor. Thereafter, the
representative plane selecting unit 130 designates the bottom plane
by using the assumptions of {circle around (1)} and {circle around
(4)} described above, designates the real wall P.sub.wall by using
the assumptions {circle around (2)}, {circle around (3)}, and
{circle around (5)}, and designates the ceiling P.sub.ceiling by
using assumptions of {circle around (3)} and {circle around
(6)}.
[0062] FIG. 6 is an exemplary diagram illustrating selection of a
representative plane from partitioned planes in FIG. 5. Referring
to FIG. 6, it can be seen that a bottom plane 410, a rear wall
plane 420, and a ceiling plane 430 are selected as the
representative plane. However, as illustrated in FIG. 5, the plane
object segmenting unit 120 calculates a distance between a point
and a plane by using the general plane equation and compares the
distance with a threshold with respect to points at a corner region
not selected because the crowding of the point cloud is small at
the time of segmenting the plane region, and the like to classify
the plane region again. That is, the corner region, and the like
are also included to the plane region to become a rectangular plane
region.
[0063] The spatial structure extracting unit 140 extracts a
rectangle having a minimum area while including all point clouds of
the representative plane selected by the representative plan
selecting unit 130. In this case, the spatial structure extracting
unit 140 may extract the quadrangle by using a minimum hounding
rectangle (MBR) algorithm. FIG. 7 is an exemplary diagram
illustrating recognition of the spatial structure for the
representative plane selected in FIG. 6. Referring to FIG. 7, the
spatial structure extracting unit 140 fits the rectangle obtained
from the MBR by acquiring an intersection 600 of each
representative plane to intersection information to become an
accurate rectangular spatial structure in order to obtain a
rectangle closer to a real space. As illustrated in FIG. 7, a
bottom plane 610, a rear wall plane 620, and a ceiling plane 630
are configured in the rectangular shape in the rectangular spatial
structure.
[0064] Thereafter, the spatial structure extracting unit 140
obtains 3D posture information of objects from the representative
plane to match an object 220 in the extracted rectangular spatial
structure. FIG. 8 is an exemplary diagram illustrating matching a
real object 220 to the spatial structure recognized by the
augmented real device according to the exemplary embodiment of the
present invention. In this case, the 3D posture information may
include the representative plane, point information of corners of
the real objects 220, and normal vector information of each
representative plane. Therefore, the spatial structure extracting
unit 140 matches the real objects by considering posture
information of the real objects according to the direction of the
representative plane.
[0065] The virtual object matching unit 150 matches virtual objects
710 and 720 by using the normal vector of the representative plane
and internal angle component information of each corner of the
representative plane. FIG. 9 is an exemplary diagram illustrating
matching virtual objects 710 and 720 to the spatial structure
recognized by the augmented real device according to the exemplary
embodiment of the present invention.
[0066] Hereinafter, a method for providing augmented reality based
on spatial recognition of the augmented reality device according to
an exemplary embodiment of the present invention will be described
in detail with reference to FIG. 2.
[0067] When an image is input from the depth camera 200 (S101), the
point cloud normal vector extracting unit 110 converts raw data of
a depth image input from the depth camera 200 into 3D point cloud
values x, y, and z to extract the normal vectors Vx, Vy, and Vz for
the point cloud as illustrated in FIGS. 3 and 4 (S102).
[0068] Thereafter, the plane object segmenting unit 120 segments
the plane information on the indoor space as illustrated in FIG. 5
by using the normal vectors calculated by the point cloud normal
vector extracting unit 110 (S103).
[0069] Subsequently, the representative plane selecting unit 130
selects the representative plane by using the contextual knowledge
of the camera 200 and the indoor space configuration from the plane
information obtained by the plane object segmenting unit 120
(S104). Referring to FIG. 6, the representative plane may become
the bottom plane 410, the rear wall plane 420, and the ceiling
plane 420.
[0070] Thereafter, the spatial structure extracting unit 140
extracts the quadrangle which is the spatial structure having a
minimum area while including all point clouds of the representative
plane selected by the representative plan selecting unit 130
(S105). In this case, the spatial structure extracting unit 140 may
extract the accurate spatial structure by using the minimum
bounding rectangle (MBR) algorithm and obtaining the intersection
of the respective planes. Referring to FIG. 7, it can be seen that
the respective planes 610, 620, and 630 have an accurate
rectangular structure by the intersection 600.
[0071] Subsequently, the spatial structure extracting unit 140
matches the real object 220 in the spatial structure recognized
through the representative plane as illustrated in FIG. 8 by using
the 3D posture information including corner information and normal
vector information of the representative plane and the real objects
(S106).
[0072] Thereafter, the virtual object matching unit 150 matches the
virtual object to the spatial structure matched by the real object
220 as illustrated in FIG. 9 (S107). In this case, the virtual
object matching unit 150 may match the virtual object by using the
normal vector of the representative plane and internal angle
component information of the rectangular corner.
[0073] FIG. 10 is a configuration diagram of a computer system to
which an augmented reality device is applied according to an
exemplary embodiment of the present invention.
[0074] Referring to FIG. 10 a computing system 1000 may include at
least one processor 1100, one memory 1300, one user interface input
device 1400, one user interface output device 1500, one storage
1600, and one network interface 1700 connected through a bus
1200.
[0075] The processor 1100 may be a semiconductor device that
executes processing of commands stored in a central processing unit
(CPU) or the memory 1300 and/or the storage 1600. The memory 1300
and the storage 1600 may include various types of volatile or
non-volatile storage media. For example, the memory 1300 may
include a read only memory (ROM) and a random access memory
(RAM).
[0076] Therefore, steps of a method or an algorithm described in
association with the embodiments disclosed in the specification may
be directly implemented by hardware and software modules executed
by the processor 1100, or a combination thereof. The software
module may reside in storage media (that is, the memory 1300 and/or
the storage 1600) such as a RAM, a flash memory, a ROM, an EPROM,
an EEPROM, a register, a hard disk, a removable disk, and a
CD-ROM.
[0077] The exemplary storage medium is coupled to the processor
1100 and the processor 1100 may read information from the storage
medium and write the information in the storage medium. As another
method, the storage medium may be integrated with the processor
1100. The processor and the storage medium may reside in an
application specific integrated circuit (ASIC). The ASIC may reside
in a user terminal. As another method, the processor and the
storage medium may reside in the user terminal as individual
components.
[0078] The above description just illustrates the technical spirit
of the present invention and various modifications and
transformations can be made by those skilled in the art without
departing from an essential characteristic of the present
invention.
[0079] Accordingly, the exemplary embodiments disclosed herein are
intended to not limit but describe the technical spirit of the
present invention but the scope of the technical spirit of the
present invention is not limited by the exemplary embodiments. The
scope of the present invention should be interpreted by the
appended claims and all technical spirit in the equivalent range
thereto should be interpreted to be embraced by the claims of the
present invention.
* * * * *