U.S. patent application number 17/476382 was filed with the patent office on 2022-01-06 for systems and methods for augmented reality for disaster simulation.
The applicant listed for this patent is State Farm Mutual Automobile Insurance Company. Invention is credited to Timothy Joel Davis, Amy Engelhorn.
Application Number | 20220004684 17/476382 |
Document ID | / |
Family ID | 1000005839440 |
Filed Date | 2022-01-06 |
United States Patent
Application |
20220004684 |
Kind Code |
A1 |
Davis; Timothy Joel ; et
al. |
January 6, 2022 |
SYSTEMS AND METHODS FOR AUGMENTED REALITY FOR DISASTER
SIMULATION
Abstract
An augmented reality (AR) system for generating and displaying a
pre-disaster Enhanced Situation Visualization (ESV) is provided.
The AR system may include an ESV computing device, a user computing
device operated by a user and a reference database. The user
computing device may transmit a reference request message to the
ESV device, the reference request message including an image and/or
GPS location of a property. The ESV computing device may determine
the subject of the image is the property, retrieve reference
information including peril maps associated with the property from
the reference database, and determine situation information
specific to the subject. The reference and situation information
including a loss estimate may be displayed on the user computing
device to provide an ESV of the property. The ESV may be used for
insurance-related activities, such as handling, adjusting, and/or
generating an insurance policy, premium, and/or discount, and/or
generating insurance-related recommendations.
Inventors: |
Davis; Timothy Joel;
(Chicago, IL) ; Engelhorn; Amy; (Normal,
IL) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
State Farm Mutual Automobile Insurance Company |
Bloomington |
IL |
US |
|
|
Family ID: |
1000005839440 |
Appl. No.: |
17/476382 |
Filed: |
September 15, 2021 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
16748460 |
Jan 21, 2020 |
11132481 |
|
|
17476382 |
|
|
|
|
15184886 |
Jun 16, 2016 |
10579749 |
|
|
16748460 |
|
|
|
|
62185359 |
Jun 26, 2015 |
|
|
|
62193987 |
Jul 17, 2015 |
|
|
|
Current U.S.
Class: |
1/1 |
Current CPC
Class: |
G06T 19/006 20130101;
G06F 30/20 20200101 |
International
Class: |
G06F 30/20 20060101
G06F030/20; G06T 19/00 20060101 G06T019/00 |
Claims
1. An Enhanced Situation Visualization (ESV) computing device
comprising a memory and at least one processor communicatively
coupled to the memory, wherein the at least one processor is
configured to: determine potential damage to a property due to an
occurrence of a future natural event; generate a reference response
message including the determined potential damage to the property;
and in response to receiving the reference response message, cause
to be displayed on a user computing device the potential damage to
the property overlaid upon a current view of the property as viewed
through the user computing device.
2. The ESV computing device of claim 1, wherein the at least one
processor is further configured to: retrieve, from a reference
database, reference information that includes a peril map
indicating a risk level associated with the future natural event;
and determine the potential damage based upon the peril map.
3. The ESV computing device of claim 2, wherein the at least one
processor is further configured to determine, based upon the peril
map and the future natural event, a loss estimate associated with
the potential damage.
4. The ESV computing device of claim 3, wherein the at least one
processor is further configured to cause to be displayed on the
user computing device the loss estimate overlaid upon the current
view of the property as viewed through the user computing
device.
5. The ESV computing device of claim 1, wherein the at least one
processor is further configured to cause to be displayed, using the
reference response message on the user computing device, natural
consequences at the property due to the future natural event, the
natural consequences overlaid upon the current view of the property
as viewed through the user computing device.
6. The ESV computing device of claim 1, wherein the at least one
processor is further configured to receive a reference request
message from the user computing device, the reference request
message including (i) an image representative of a current view of
the property as viewed through the user computing device and (ii)
at least one user-input variable, the user-input variable
associated with a characteristic of the future natural event.
7. The ESV computing device of claim 1, wherein the at least one
processor is further configured to transmit the reference response
message to the user computing device.
8. A computer-implemented method for generating and displaying a
pre-disaster Enhanced Situation Visualization (ESV) on a user
interface of a user computing device, the method comprising:
determining potential damage to a property due to an occurrence of
a future natural event; generating a reference response message
including the determined potential damage to the property; and in
response to receiving the reference response message, causing to be
displayed on the user computing device the potential damage to the
property overlaid upon a current view of the property as viewed
through the user computing device.
9. The computer-implemented method of claim 8 further comprising:
retrieving, from a reference database, reference information that
includes a peril map indicating a risk level associated with the
future natural event; and determining the potential damage based
upon the peril map.
10. The computer-implemented method of claim 9 further comprising
determining, based upon the peril map and the future natural event,
a loss estimate associated with the potential damage.
11. The computer-implemented method of claim 10 further comprising
causing to be displayed on the user computing device the loss
estimate overlaid upon the current view of the property as viewed
through the user computing device.
12. The computer-implemented method of claim 8 further comprising
causing to be displayed, using the reference response message on
the user computing device, natural consequences at the property due
to the future natural event, the natural consequences overlaid upon
the current view of the property as viewed through the user
computing device.
13. The computer-implemented method of claim 8 further comprising
receiving a reference request message from the user computing
device, the reference request message including (i) an image
representative of a current view of the property as viewed through
the user computing device and (ii) at least one user-input
variable, the user-input variable associated with a characteristic
of the future natural event.
14. The computer-implemented method of claim 8 further comprising
transmitting the reference response message to the user computing
device.
15. At least one non-transitory computer-readable storage medium
having computer-executable instructions embodied thereon, wherein
when executed by at least one processor, the computer-executable
instructions cause the at least one processor to: determine
potential damage to a property due to an occurrence of a future
natural event; generate a reference response message including the
determined potential damage to the property; and in response to
receiving the reference response message, cause to be displayed on
a user computing device the potential damage to the property
overlaid upon a current view of the property as viewed through the
user computing device.
16. The computer-readable storage medium of claim 15, wherein the
computer-executable instructions further cause the at least one
processor to: retrieve, from a reference database, reference
information that includes a peril map indicating a risk level
associated with the future natural event; and determine the
potential damage based upon the peril map.
17. The computer-readable storage medium of claim 16, wherein the
computer-executable instructions further cause the at least one
processor to determine, based upon the peril map and the future
natural event, a loss estimate associated with the potential
damage.
18. The computer-readable storage medium of claim 15, wherein the
computer-executable instructions further cause the at least one
processor to cause to be displayed, using the reference response
message on the user computing device, natural consequences at the
property due to the future natural event, the natural consequences
overlaid upon the current view of the property as viewed through
the user computing device.
19. The computer-readable storage medium of claim 15, wherein the
computer-executable instructions further cause the at least one
processor to receive a reference request message from the user
computing device, the reference request message including (i) an
image representative of a current view of the property as viewed
through the user computing device and (ii) at least one user-input
variable, the user-input variable associated with a characteristic
of the future natural event.
20. The computer-readable storage medium of claim 15, wherein the
computer-executable instructions further cause the at least one
processor to transmit the reference response message to the user
computing device.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application is a continuation application of U.S.
patent application Ser. No. 16/748,460, filed Jan. 21, 2020,
entitled "SYSTEMS AND METHODS FOR AUGMENTED REALITY FOR DISASTER
SIMULATION," which is a continuation application of U.S. patent
application Ser. No. 15/184,886, filed Jun. 16, 2016, entitled
"SYSTEMS AND METHODS FOR AUGMENTED REALITY FOR DISASTER
SIMULATION," which claims the benefit of priority to U.S.
Provisional Patent Application No. 62/185,359, filed Jun. 26, 2015,
entitled "SYSTEMS AND METHODS FOR ENHANCED SITUATION
VISUALIZATION," and to U.S. Provisional Patent Application No.
62/193,987, filed Jul. 17, 2015, entitled "SYSTEMS AND METHODS FOR
AUGMENTED REALITY FOR DISASTER SIMULATION," and is related to U.S.
patent application Ser. No. 15/184,916, filed Jun. 16, 2016,
entitled "SYSTEMS AND METHODS FOR ENHANCED SITUATION
VISUALIZATION," and to U.S. patent application Ser. No. 16/457,026,
filed Jun. 28, 2019, entitled "SYSTEMS AND METHODS FOR ENHANCED
SITUATION VISUALIZATION," and to U.S. patent application Ser. No.
16/687,059, filed Nov. 18, 2019, entitled "SYSTEMS AND METHODS FOR
ENHANCED SITUATION VISUALIZATION," and to U.S. patent application
Ser. No. 17/401,042, filed Aug. 12, 2021, entitled "SYSTEMS AND
METHODS FOR ENHANCED SITUATION VISUALIZATION," the entire contents
of which are hereby incorporated by reference in their
entirety.
FIELD OF THE DISCLOSURE
[0002] The present disclosure relates to augmented reality and,
more particularly, to systems and methods for generating and
displaying an Enhanced Situation Visualization on a user computing
device of a disaster simulation.
BACKGROUND
[0003] There are many locations across the world that are plagued
by natural disasters. Very few regions, in fact, are excluded from
the risk of disasters, such as tornadoes, floods, forest fires,
hurricanes, earthquakes, mudslides, or tsunamis. Accordingly, some
consumers attempt to defray those risks by purchasing insurance
policies for their property, vehicle(s), and valuable items. When
disaster strikes, though the property, vehicle, or other item may
be damaged or destroyed, these consumers have an avenue to recoup
some or all of the monetary value associated with such damage or
loss. It may be difficult for a consumer or a representative of an
insurance provider to quantify or characterize potential damage to
a property or other object either before or after the disaster
occurs. For example, underwriting documentation regarding the
property or object may be old or out-of-date, and/or no longer
representative of the value of the property; or an extent of damage
or loss may be difficult to visualize and capture without extensive
research. There is a need for an efficient and effective method for
enhancing a view of a likely disaster situation such that any
damage or loss may be conveniently and accurately quantified, or
such loss may be mitigated or avoided altogether.
[0004] The concept of augmented reality (AR) refers to a real-time
direct or indirect view of a physical, real-world environment with
augmented (or supplemented) elements generated by a computer. The
AR view is generally presented on a display of a user computing
device. The user computing device may use the view from a camera as
a base or foundation view, and then augments or supplements the
base view with the AR. The benefits of AR have been realized in
some industries, including, for example, commerce, in which a user
may scan or otherwise view a product and may be presented with
additional images of the product; gaming, in which users may
experience game play in a real-world environment with virtual
supplements; and medicine, in which technology has been developed
that maps subcutaneous veins and arteries onto a patient's
skin.
BRIEF SUMMARY
[0005] The present embodiments may relate to systems and methods
for providing Enhanced Situation Visualization pre- and
post-natural disaster. An Augmented Reality (AR) system for
generating and displaying such an Enhanced Situation Visualization
(ESV) is provided. The AR system may include an ESV computing
device, a user computing device operated by a user, and/or a
reference database. The user computing device may transmit a
reference request message to the ESV device, the reference request
message including an image representative of a current view of a
user computing device of, for example, a property.
[0006] For example, in a pre-disaster situation, an insurance
underwriter may be interested in the risk to a property of flood
damage, to determine whether to provide an insurance policy
associated with the property and/or to adjust an amount of an
existing insurance policy. In at least one exemplary embodiment,
the underwriter (or any other user) may generate and transmit a
reference request message (using their user computing device) that
includes an image representative of their current view and a
subject (e.g., the property that is potentially at risk) of the
current view.
[0007] The ESV computing device may determine a subject of the
image, retrieve reference information associated with the subject
from the reference database, and/or determine situation information
specific to the subject. The reference information and situation
information may be displayed to the user as an ESV on a display of
the user computing device, for example, overlaid upon the user's
current view through the user computing device. The ESV may be used
for insurance-related activities, such as handling, adjusting,
and/or generating an insurance policy associated with a
property.
[0008] In a further aspect, a computer-implemented method for
generating and displaying an Enhanced Situation Visualization (ESV)
on a user interface of a user computing device may be provided. The
method may include receiving, by an ESV computing device, a
reference request message from the user computing device. The
reference request message may include an image representative of a
current view of the user computing device, the image including a
property in a current state. The reference request message may
further include at least one user-input variable associated with a
characteristic of a natural event. The method may also include
determining a first subject of the image, wherein the first subject
is the property, and/or retrieving, from a reference database,
reference information associated with the first subject of the
image. The reference information may include a peril map indicating
a risk level associated with the natural event for the
property.
[0009] The method may further include determining, based upon the
reference request message and/or the reference information,
situation information specific to a situation of the property. The
situation information may include a loss estimate based upon the
peril map, the at least one user-input variable, and the natural
event. The method may also include generating a reference response
message including the situation information and the reference
information, and/or transmitting the reference response message to
the user computing device, wherein the reference response message
causes an Enhanced Situation Visualization including the situation
information to be displayed on the user computing device as
overlaid upon the current view. The ESV may be used for
insurance-related activities, such as handling, adjusting, and/or
generating an insurance policy associated with a property.
[0010] The method may include additional, less, or alternate
actions, including those discussed elsewhere herein, and/or
implemented via one or more local or remote processors, and/or
computer-executable instructions stored on non-transitory computer
media or medium. For example, in one embodiment, the natural event
may be a flood, and the at least one user-input variable may be one
of rainfall amount, rainfall duration, and/or season. The method
may further include displaying the Enhanced Situation Visualization
as an overlay of predicted water level. In another embodiment, the
natural event may be a wildfire, and the at least one user-input
variable may be one of rainfall amount, season, maximum
temperature, and/or wind speed. The method may further include
displaying the Enhanced Situation Visualization as an overlay of a
wildfire path. In other embodiments, the method may further include
displaying the Enhanced Situation Visualization as a simulated
animation.
[0011] In another aspect an Enhanced Situation Visualization (ESV)
computing device including a memory and at least one processor
communicatively coupled to the memory may be provided. The at least
one processor may be programmed to receive a reference request
message from the user computing device, the reference request
message including an image representative of a current view of the
user computing device, the image including a property in a current
state. The reference request message may further include at least
one user-input variable, the user-input variable associated with a
characteristic of a natural event. The at least one processor may
also be programmed to determine a first subject of the image,
wherein the first subject is the property. The at least one
processor may be further programmed to retrieve, from a reference
database, reference information associated with the first subject
of the image. The reference information may include a peril map
indicating a risk level associated with the natural event for the
property. The at least one processor may also be programmed to
determine, based upon the reference request message and the
reference information, situation information specific to a
situation of the property. The situation information may include a
loss estimate based on the peril map, the at least one user-input
variable, and/or the natural event. The at least one processor may
be further programmed to generate a reference response message
including the situation information and the reference information,
and/or transmit the reference response message to the user
computing device.
[0012] The reference response message may cause an Enhanced
Situation Visualization including the situation information to be
displayed on the user computing device as overlaid upon the current
view. The ESV may be used for insurance-related activities, such as
handling, adjusting, and/or generating an insurance policy
associated with a property. The ESV computing device may include
additional, less, or alternate functionality, including that
discussed elsewhere herein.
[0013] In yet another aspect, at least one non-transitory
computer-readable storage media having computer-executable
instructions embodied thereon may be provided. The
computer-executable instructions, when executed by at least one
processor, may cause the processor to receive a reference request
message from the user computing device. The reference request
message may include an image representative of a current view of
the user computing device, the image including a property in a
current state. The reference request message may also include at
least one user-input variable associated with a characteristic of a
natural event.
[0014] The computer-executable instructions may also cause the
processor to determine a first subject of the image, wherein the
first subject is the property, and/or retrieve, from a reference
database, reference information associated with the first subject
of the image. The reference information may include a peril map
indicating a risk level associated with the natural event for the
property. The computer-executable instructions may further cause
the processor to determine, based upon the reference request
message and the reference information, situation information
specific to a situation of the property. The situation information
may include a loss estimate based upon the peril map, the at least
one user-input variable, and the natural event. The
computer-executable instructions may also cause the processor to
generate a reference response message including the situation
information and the reference information, and/or transmit the
reference response message to the user computing device, wherein
the reference response message causes an Enhanced Situation
Visualization including the situation information to be displayed
on the user computing device as overlaid upon the current view.
Additional, less, or alternate instructions may be provided, such
as instructions directing the functionality discussed elsewhere
herein.
[0015] In another aspect, a computer-implemented method for
generating and displaying an Enhanced Situation Visualization (ESV)
on a user interface of a user computing device may be provided. The
method may include receiving, by an ESV computing device, a
reference request message from the user computing device. The
reference request message may include an image representative of a
current view of the user computing device, the image including a
property in a current state. The reference request message may
further include a risk tolerance threshold associated with a
maximum loss estimate for the property due to a natural event.
[0016] The method may also include determining a first subject of
the image, wherein the first subject is the property, and/or
retrieving, from a reference database, reference information
associated with the first subject of the image. The reference
information may include a peril map indicating a risk level
associated with the natural event for the property. The method may
further includes determining, based upon the reference request
message and the reference information, situation information
specific to a situation of the property, wherein the situation
information includes at least one characteristic of the natural
event that satisfies the risk tolerance threshold. The method may
still further includes generating a reference response message
including the situation information and the reference information,
and/or transmitting the reference response message to the user
computing device, wherein the reference response message causes an
Enhanced Situation Visualization including the situation
information to be displayed on the user computing device as
overlaid upon the current view.
[0017] The method may include additional, less, or alternate
actions, including those discussed elsewhere herein, and/or
implemented via one or more local or remote processors, and/or
computer-executable instructions stored on non-transitory computer
media or medium. For example, in one embodiment, the method may
further include displaying the Enhanced Situation Visualization as
a simulated animation.
[0018] Advantages will become more apparent to those skilled in the
art from the following description of the preferred embodiments
which have been shown and described by way of illustration. As will
be realized, the present embodiments may be capable of other and
different embodiments, and their details are capable of
modification in various respects. Accordingly, the drawings and
description are to be regarded as illustrative in nature and not as
restrictive.
BRIEF DESCRIPTION OF THE DRAWINGS
[0019] The Figures described below depict various aspects of the
systems and methods disclosed therein. It should be understood that
each Figure depicts an embodiment of a particular aspect of the
disclosed systems and methods, and that each of the Figures is
intended to accord with a possible embodiment thereof. Further,
wherever possible, the following description refers to the
reference numerals included in the following Figures, in which
features depicted in multiple Figures are designated with
consistent reference numerals.
[0020] There are shown in the drawings arrangements which are
presently discussed, it being understood, however, that the present
embodiments are not limited to the precise arrangements and are
instrumentalities shown, wherein:
[0021] FIG. 1 illustrates a schematic view of an exemplary reality
augmentation system for providing post-disaster Enhanced Situation
Visualization.
[0022] FIG. 2 illustrates an exemplary historical image of a
property for which Enhanced Situation Visualization may be provided
by the ESV computing device shown in FIG. 1.
[0023] FIG. 3 illustrates an exemplary current view of the property
shown in FIG. 2 in a post-disaster state.
[0024] FIG. 4 illustrates an exemplary current view of the property
shown in FIG. 3 in the post-disaster state, including a user
computing device.
[0025] FIG. 5 depicts a block diagram illustrating the transmission
of an exemplary reference request message and an exemplary
reference response message between various components of the
exemplary reality augmentation system shown in FIG. 1.
[0026] FIG. 6 illustrates a first exemplary view of a post-disaster
Enhanced Situation Visualization provided by the ESV computing
device shown in FIG. 1.
[0027] FIG. 7 illustrates a second exemplary view of a
post-disaster Enhanced Situation Visualization provided by the ESV
computing device shown in FIG. 1.
[0028] FIG. 8 illustrates a third exemplary view of a post-disaster
Enhanced Situation Visualization provided by the ESV computing
device shown in FIG. 1.
[0029] FIG. 9 illustrates an exemplary screenshot of a pre-disaster
Enhanced Situation Visualization (ESV) displayed to a user on a
user computing device.
[0030] FIG. 10 illustrates an exemplary pre-disaster Enhanced
Situation Visualization (ESV) including a simulation.
[0031] FIG. 11 illustrates an exemplary configuration of a user
computing device as shown in FIG. 1, in accordance with one
embodiment of the present disclosure.
[0032] FIG. 12 illustrates an exemplary configuration of a server
system as shown in FIG. 1, in accordance with one embodiment of the
present disclosure.
[0033] FIG. 13 illustrates exemplary Enhanced Situation
Visualization provided by the reality augmentation system shown in
FIG. 1.
[0034] FIG. 14 illustrates exemplary augmented reality for disaster
simulation provided by the reality augmentation system shown in
FIG. 1.
[0035] FIG. 15 illustrates a flowchart of an exemplary
computer-implemented method for providing reality augmentation for
disaster simulation using the reality augmentation system shown in
FIG. 1.
[0036] FIG. 16 depicts a diagram of components of one or more
exemplary computing devices that may be used in the system shown in
FIG. 1.
[0037] The Figures depict preferred embodiments for purposes of
illustration only. One skilled in the art will readily recognize
from the following discussion that alternative embodiments of the
systems and methods illustrated herein may be employed without
departing from the principles of the disclosure described
herein.
DETAILED DESCRIPTION
[0038] The present embodiments may relate to, inter alia, systems
and methods for Enhanced Situation Visualization. The system
described herein includes a reality augmentation computer system
having an Enhanced Situation Visualization (ESV) computing device,
a reference database, one or more beacons, and/or a user computing
device. The ESV computing device may be communicatively coupled to
the reference database, which may store a number of historical
images and/or reference information associated with a number of
subjects, as will be described more fully herein. The ESV computing
device may be communicatively coupled to the user computing device
and the one or more beacons via, for example, a wireless
network.
[0039] In one exemplary embodiment, the ESV computing device may
receive a request from the user computing device for information to
describe, supplement, and/or augment a current view of a user
through the user computing device. The request (a "reference
request message") may include a real-time image of the current view
of the user and/or the user computing device, as well as other
additional information, such as device data (associated with the
user computing device), user account information, and/or subject
identifiers (e.g., a barcode on a subject item). The ESV computing
device may determine one or more subjects of the current view. Such
determination may be made, for example, based upon object
recognition, location analysis, user input, processing of any
subject identifier(s), or any other analysis.
[0040] The ESV computing device may then retrieve from the
reference database at least one of a historical image and reference
information associated with the subject. The ESV computing device
may then, based upon the retrieved historical image and/or
reference information, determine situation information specific to
the situation of the user. The ESV computing device may then
generate and transmit a response, including the historical image
and/or reference information and the situation information, to the
user computing device. The response (a "reference response
message") may further include instructions causing the user
computing device to display the information transmitted in the
reference response message on a display area of the user computing
device.
[0041] The information may be displayed as overlaid on, beside, or
instead of the current view of the user, such that the situation
information is not only generally visible to the user but is
visually near the subject with which the information is associated,
relative to the display area. In one exemplary embodiment, the
information is displayed over a real-time view of the subject from
a camera of the user computing device. Accordingly, the ESV
computing device may facilitate enhanced visualization of the
user's current situation, providing additional information,
instruction, and/or guidance that may otherwise be difficult to
obtain by and/or communicate to the user.
[0042] As described below in more detail, the ESV computing device
may be configured to (i) receive a reference request message from a
user computing device, the reference request message including an
image representative of a current view of the user computing device
and at least one user-input variable, the image including a
property in a current state, and the user-input variable associated
with a characteristic of a natural event (ii) determine a first
subject of the image, wherein the first subject is the property;
(iii) retrieve, from a reference database, reference information
associated with the first subject of the image, wherein the
reference information includes a peril map indicating a risk level
associated with the natural event for the property; (iv) determine,
based upon the reference request message and the reference
information, situation information specific to a situation of the
property, wherein the situation information includes a loss
estimate based upon the peril map, the at least one user-input
variable, and the natural event; (v) generate a reference response
message including the situation information and the reference
information; and/or (vi) transmit the reference response message to
the user computing device, wherein the reference response message
causes an Enhanced Situation Visualization including the situation
information to be displayed on the user computing device as
overlaid upon the current view.
[0043] For example, in the event of a natural disaster, a user
(e.g., a homeowner or an insurance claims adjuster) may wish to
utilize the reality augmentation system described herein to
quantify, characterize, and/or catalog a level of damage done to a
property post-disaster. The user may access their computing device
and may further access the reality augmentation system using, for
example, an application ("app") installed on the user computing
device or a web browser. The user may capture a real-time image
representative of their current view as they survey the damaged
property. The user may then send the real-time image to the ESV
computing device within a reference request message transmitted by
the user computing device. The reference request message may
further include device data (such as, for example, GPS,
accelerometer, or compass data from the user computing device,
which may enable the ESV computing device to determine a location
of the user computing device), account or policy information
associated with the user (such as log-in information, an account
number, a user name, a policy number), and/or subject identifiers
(such as, in this example, a house or building number indicative of
the street address of the property).
[0044] The ESV computing device, based upon the real-time image
and/or any of the additional information included in the reference
request message, may determine that the subject of the real-time
image (and, therefore, the subject of the user's current view) is
the property located at a particular address (e.g., 123 Elm Street,
New York City, N.Y.). The ESV computing device may then retrieve
information, including a historical image of the property before
the damage and other reference information, from the reference
database. The reference information may include, in this example,
specifications (e.g., square footage, appraisal value, layout) of
the property, information regarding an insurance policy associated
with the property, the property's building materials, and/or any
public utilities available and/or accessible at or near the
property.
[0045] The ESV computing device may compile, analyze, summarize,
and/or otherwise process the historical image and the reference
information. In this example, the ESV computing device may be
configured to determine--based upon the real-time image, the
historical image, and the reference information--an estimated
amount of damage, an estimated repair cost, a location of certain
damaged or covered utilities, a virtual grid overlay of pre-damage
images, an estimated claim amount, an estimated policy coverage
amount, and/or an estimated disbursement for repairs. These
determinations may be generally referred to as "situation
information," as they are specific to the situation and view
provided by the user. The ESV computing device may generate a
reference response message including the historical image, the
reference information, and/or the situation information. The ESV
computing device may then transmit the reference response message
to the user computing device, such that the historical image may be
displayed beside, overlaid upon, or instead of the current view of
the user. The reference information and/or the situation
information may additionally or alternatively be displayed beside,
overlaid upon, or displayed instead of the current view of the
user. The current view may be a real-time view captured by a camera
of the user computing device or a previously captured or saved
image from the camera.
[0046] The user may then easily view and/or record any reference
information and/or situation information they determine to be
relevant. In the case that the user is a claims adjuster, the user
may be able to utilize the situation information and the historical
image to more easily make comparisons between the pre- and
post-damage property, as well as receive and record suggestions and
estimations about repair costs and the extent of any damage.
Accordingly, the insurance claims process may be enhanced and
simplified for both a homeowner of a damaged property and the
insurance claims adjuster.
[0047] As another example, the user may wish to utilize the reality
augmentation system to visualize how their property may be affected
in the event of a natural disaster. In other words, the reality
augmentation system may be accessed to provide both pre- and
post-disaster information. More specifically, the user may access
their user computing device and may further access the reality
augmentation system using, for example, an app installed on the
user computing device or a web browser. The user may capture a
real-time image representative of the current state of their
property and/or the surrounding terrain. The user may then send the
real-time image to the ESV computing device within a reference
request message transmitted by the user computing device. The
reference request message may further include device data (such as,
for example, GPS, accelerometer, or compass data from the user
computing device, which may enable the ESV computing device to
determine a location of the user computing device), account or
policy information associated with the user (such as log-in
information, an account number, a user name, a policy number),
and/or subject identifiers (such as, in this example, a house or
building number indicative of the street address of the
property).
[0048] The ESV computing device, based upon the real-time image
and/or any of the additional information included in the reference
request message, may determine that the subject of the real-time
image (and, therefore, the subject of the user's current view) is
the property located at a particular address (e.g., 123 Elm Street,
New York City, N.Y.). The ESV computing device may then retrieve
information, including a peril map of the location and other
reference information, from the reference database. As will be
described more fully herein, a peril map indicates the relative
risk at a particular location for a particular natural event (e.g.,
a natural disaster).
[0049] For example, particular locations in a valley or flood plain
may be more at risk to flood damage, whereas other locations in
"Tornado Alley" may be more at risk for tornado or other storm
damage. The additional reference information may include, in this
example, specifications (e.g., square footage, appraisal value,
layout) of the property, information regarding an insurance policy
associated with the property, the property's building materials,
and/or topographical information associated with the location of
the property at a granular level (e.g., at a lot level or
neighborhood level).
[0050] The methods and system described herein may be implemented
using computer programming or engineering techniques including
computer software, firmware, hardware, or any combination or
subset. At least one of the technical problems addressed by this
system includes: (i) inconvenient and imprecise comparison between
prior and current states of objects, such as damaged or enhanced
properties; (ii) difficulty quantifying and cataloging differences
between states pre- and post-disaster states; and/or (iii)
difficult and imprecise estimation of properties more at risk
during particular natural events.
[0051] The system and methods described herein address these
technical problems. A technical effect of the systems and processes
described herein is achieved by performing at least one of the
following: (a) receiving a reference request message from the user
computing device, the reference request message including an image
representative of a current view of the user computing device and
at least one user-input variable, the image including a property in
a current state, and/or the user-input variable associated with a
characteristic of a natural event; (b) determining a first subject
of the image, wherein the first subject is the property; (c)
retrieving, from a reference database, reference information
associated with the first subject of the image, wherein the
reference information includes a peril map indicating a risk level
associated with the natural event for the property; (d)
determining, based upon the reference request message and the
reference information, situation information specific to a
situation of the property, wherein the situation information
includes a loss estimate based upon the peril map, the at least one
user-input variable, and/or the natural event; (e) generating a
reference response message including the situation information and
the reference information; and/or (f) transmitting the reference
response message to the user computing device, wherein the
reference response message causes an Enhanced Situation
Visualization including the situation information to be displayed
on the user computing device as overlaid upon the current view.
[0052] The technical effect achieved by this system is at least one
of: (i) more convenient and efficient comparison between a prior
state of an object or property and a current state of the object or
property; (ii) simpler quantification and characterization of
differences between a pre- and post-disaster states of an object or
property; and/or (iii) more accurate and precise identification of
properties or objects at risk during a natural event.
[0053] "Real-time," as used herein, may refer to at least one of
the time of occurrence of events, the time of receipt and/or
retrieval predetermined data, the time to process the data, and/or
the time of a system response to the events. In the embodiments
described herein, these activities and events occur substantially
instantaneously. For example, a "real-time image" may refer to an
image that is taken and transmitted substantially
instantaneously.
[0054] "Current view," as used herein, may refer to a visual
perspective or point of view available to a user (or user computing
device, e.g., a camera on a smart phone) at a time that the user
requests Enhanced Situation Visualization. The current view
represents the subject(s) at which the user is looking and about
which the user may be interested in receiving information.
[0055] "Location information," as used herein, may refer to
positional or locational data that may enable location of, for
example, an item, object, user, user computing device, and/or
property (i.e., a subject of the reference request). Location
information may include, for example, GPS data (e.g.,
latitude-longitude data), accelerometer data, compass data, a
street address, IP addresses, and/or cellular tower "ping"
data.
[0056] "Device data" or "device information" as used herein, may
refer to any data collected at the device that is representative or
descriptive of the situation and/or current view of the user. For
example, device data may include location information as described
above, camera images or videos that provide the current view, a
device identifier of the user computing device, and/or a device
descriptor of the type of the user computing device (e.g., wearable
device, tablet, smartphone, etc.)
[0057] "Reference information," as used herein, may refer to any
data (including text, audio, video, and/or image data) that may be
stored and may be descriptive of background or reference details
about, for example, an item, object, user, and/or property (i.e., a
subject of a reference request). Reference information may include
publicly available data about the subject. For example, in the case
of a property, such reference information may include prior MLS
listing information, publicly available historical satellite-
and/or street-view imagery, utility records or plans, tax records,
two- or three-dimensional topographical maps, peril maps, and/or
historical weather data (e.g., historical rainfall levels). A
"historical image," as used herein, is an image of the subject at a
prior time or in a prior state. As another example, reference
information may include instruction manuals, safety warnings, and
nutritional or ingredient information. Reference information may
also include user-specific information (e.g., medication
information, street address, insurance policies, user calendars or
schedules) and/or information specific to one or more vendors
maintaining the ESV computing device (e.g., propriety underwriting
information or loan information associated with a property).
[0058] "Situation information," as used herein, refers to details
specific to the situation of a user and/or a subject (e.g., a
property) that are determined based upon reference information and
information included in a reference request message. In one
example, a user may view a property that has been damaged or
destroyed by a natural disaster, and may transmit a reference
request message including location information and an image of the
damage. The reference information for the property may include a
historical image of the property, public utilities available at the
property (e.g., location of a gas main or storm drain), and/or a
current value of an associated insurance policy. The situation
information may include a location of the property, an extent of
the damage, a remediation suggestion, and/or safety warnings about
remaining dangers in the area (e.g., downed power lines).
[0059] In a second example, a user may be viewing their medication,
and may transmit a reference request message including an image of
the medication container. The reference information may include a
prescription schedule for the user and safety guidelines for the
medication. The situation information may include a status
indicator (e.g., whether a medication has been taken or not) and/or
a schedule reminder (e.g., that it is time to take a third dosage
of a particular medication). Other situation information may
include, for example, various simulations associated with the
situation of a user and/or a subject of a reference request
message.
[0060] "Peril maps," as used herein, refer to risk evaluations
associated with particular locations and may be considered an
example of reference information. More specifically, peril maps
accumulate and analyze various risk variables for a location,
including weather- or other nature-based risk. For example, a
location in the Southwest of the United States may have a high risk
for wildfire, based upon historical wildfire records and/or
accumulated insurance data associated with wildfire loss or damage
claims. A location in the Pacific Northwest or areas of the Midwest
of the United States may have a high risk for floods, based upon
historical rainfall and/or flood records, as well as accumulated
insurance data associated with flood loss or damage claims. Other
examples of high-risk natural events may include mudslides,
landslides, earthquakes, hail, hurricanes, tsunamis, tornadoes, and
lightning strikes.
[0061] Peril maps may use this risk-based analysis to assign
various levels, scores, and/or types of risk to locations (e.g., a
particular state, city, county, neighborhood, property, or region).
Conventional peril maps may be two-dimensional maps, such that the
levels/scores/risk types are overlaid on a street or topographical
map for reference by, for example, an insurance underwriting agent
("underwriter").
[0062] Exemplary Reality Augmentation Computer System
[0063] FIG. 1 depicts a schematic view of an exemplary reality
augmentation system 100 for providing Enhanced Situation
Visualization. In the exemplary embodiment, system 100 may include
an Enhanced Situation Visualization (ESV) computing device 102 and
a client subsystem, also referred to as user computing device 104,
communicatively coupled to ESV computing device 102. User computing
device 104 may be any device capable of interconnecting to the
Internet including a mobile computing device or "mobile device,"
such as a smartphone, a personal digital assistant (PDA), a tablet,
a wearable device (e.g., a "smart watch" or a personal projection
device such as "smart glasses"), a "phablet," or other
web-connectable equipment or mobile devices. Although only one user
computing device 104 is shown in FIG. 1, it should be understood
that any number of user computing devices 104 may be in
communication with ESV computing device 102.
[0064] Additionally, a database server 106 may be connected to a
reference database 108 containing information on a variety of
matters, as described below in greater detail. In one embodiment,
reference database 108 may be stored on ESV computing device 102.
In any alternative embodiment, reference database 108 may be stored
remotely from ESV computing device 102 and may be
non-centralized.
[0065] ESV computing device 102 may be configured to receive a
reference request message from user computing device 104. In one
embodiment, a user of user computing device 104 may transmit such a
reference request message from user computing device 104 using an
app or similar software installed on their user computing device
104. In another embodiment, the user may transmit the reference
request message via a web browser on their user computing device
104. The reference request message may be transmitted to the ESV
computing device 102 via a network (e.g., a wireless cellular or IP
network). As described more fully herein, the reference request
message includes at least an image representative of a current view
of the user of user computing device 104.
[0066] In some cases, the user may choose to capture an image of a
particular view and then transmit the reference request message
including the image at a later time. In such a case, the image may
not be of a "current view" in the sense that a user is still seeing
such a view, but may still be considered a "current view" in the
sense that the view is current (or relatively so) relative to the
situation about which the user is sending the reference request
message. Accordingly, the phrase "current view" should not be
construed to limit the embodiments described herein to only those
reference request messages sent immediately after capturing an
image.
[0067] ESV computing device 102 performs object recognition on the
image received in the reference request message to determine a
subject of the image. In the exemplary embodiment, the reference
request message further includes device information, as described
above, which may facilitate object recognition (e.g., by providing
a location of the subject). Once the subject of the image, and
thereby the subject of the reference request message, has been
determined, ESV computing device 102 retrieves reference
information associated with the subject(s) of the image. In some
cases, there may be one subject of an image; in other cases, there
may be more than one subject(s) of an image. ESV computing device
102 may be configured to determine a primary subject if there is
more than one determined subject, wherein the primary subject is
the subject most likely of interest to the user. ESV computing
device 102 may additionally or alternatively be configured to
consider all determined subjects of an image when retrieving
associated reference information.
[0068] Accordingly, wherein "a subject" is used herein, it should
be understood that one or more subjects may be present. Reference
information may be stored on reference database 108 and may be
indexed by subject, such that upon determination of the subject by
ESV computing device 102, corresponding reference information may
be easily retrieved.
[0069] ESV computing device 102 may further determine situation
information based upon the information received in the reference
request message, including the image and the device information, as
well as on the retrieved reference information. The situation
information may be specific and relevant to the situation of the
user (i.e., the situation depicted in the received image), as
described more fully herein. ESV computing 102 returns a reference
response message including the reference information and the
situation information, and/or may display any or all of the
reference information and/or situation information to the user on a
display of user computing device 104, such that the user may have
an enhanced visualization of their situation. ESV computing device
102 may further be configured to receive indication(s) of user
action(s) at user computing device 104. For example, the user may
turn in a different direction or may focus on a different subject,
which may require the transmission of a new reference request
message and reference response message.
[0070] ESV computing device 102 may be configured to determine a
similarity between past and current views and/or subject(s), and
thus may only adjust a certain aspect of the reference information
retrieved and/or the situation information generated. For example,
if the user directs their current view onto a second property next
door to a first property about which a first reference
request-response pair was already generated, ESV computing device
102 may retain some of the public information (e.g., utility
information, risk of natural disasters, etc.) retrieved for the
previous reference response. Accordingly, ESV computing device 102
may generate the second reference response message more efficiently
and/or may facilitate comparison between the first and second
reference response messages.
[0071] Reality augmentation system 100 may further include beacon
110. Although only one beacon 110 is shown in FIG. 1, it should be
understood that reality augmentation system 100 may include any
number of beacons 110 without departing from the scope of the
present disclosure. Each beacon 110 may be associated with and
coupled to an item of interest (e.g., a valuable item or an easy to
lose item) to facilitate location of the item of interest in the
event that the item is lost (e.g., following a natural disaster).
In one exemplary embodiment, a user (e.g., a homeowner) may go
through a set-up or initiation phase to activate beacons 110. The
user may manually associate each of beacons 110 with a particular
item of interest, for example, by filling out a physical or virtual
form that matches a beacon ID (e.g., an alphanumeric code) to the
item of interest to which that beacon is coupled. Accordingly,
reference database 108 and/or ESV computing device 102 may store a
table (or any other suitable reference object) including those
associations between beacons 110 and the item(s) of interest.
[0072] During use, ESV computing device 102 may be in communication
with beacon 110, which may be an active beacon (e.g., continuously
or periodically emitting and/or transmitting signals to ESV
computing device 102) or a passive beacon (e.g., responding to
interrogation signals transmitted by ESV computing device 102).
When an item of interest is lost, for example in the event of a
natural disaster, the user may send a beacon activation request
with the reference request message (e.g., included within the
reference request message or as an addendum). Such a request may be
sent by default when a reference request message is transmitted, or
the user may select a particular option to send the beacon
activation request. ESV computing device receives signals from
beacon 110 and transmits the received signals to the user of the
user computing device, for example, as directional icons to locate
the associated item of interest, as will be described further
herein.
[0073] In one exemplary embodiment, reality augmentation system 100
may further include an insurance server 112, which may be in
communication with ESV computing device 102, user computing device
104, and/or reference database 108. Insurance server 112 may be
associated with and/or maintained by an insurance provider.
Insurance server 112 may provide reference information to reference
database 108, such as, for example, policy information (e.g.,
policy amount, premium, discount) associated with a particular
subject, historical images of the subject taken for underwriting
purposes, past claims involving the subject or a user associated
with the subject, and/or propriety underwriting information
associated with the subject and/or a corresponding policy.
[0074] Additionally or alternatively, insurance server 112 may
receive either or both of the reference request message generated
by user computing device 104 and the reference response message
generated by ESV computing device 102. For example, for a user
computing device 104 associated with the insurance provider (e.g.,
a user computing device 104 used by an employee of the insurance
provider), the user computing device 104 may also transmit the
reference request message (or a copy thereof) to insurance server
112, such that any records (e.g., policy or claims records)
associated with the subject of the message may be updated with the
information included in the reference request message.
[0075] ESV computing device 102, after determining the subject of
the reference quest message, may identify an insurance policy
associated with the subject. ESV computing device 102 may then also
transmit the reference response message (or a copy thereof) to
insurance server 112 associated with a provider of the policy, such
that any records (e.g., policy or claims records) associated with
the subject of the message may be updated with the information
included in the reference response message. Alternatively, user
computing device 104 may transmit some or all of the information in
the reference response message to insurance server 112 for the same
purpose.
[0076] In addition, insurance server 112 may use any or all of the
received information (from either or both of the reference request
message and the reference response message) to automatically
populate a relevant insurance form. For example, if a claims adjust
generates a reference request message for a damaged property,
insurance server 112 may receive information in the reference
response message including the subject of the request (i.e., the
property), the date and/or time of the request, an estimated amount
of damage, and/or an estimated claim amount. Insurance server 112
may then retrieve any policy information associated with the
property (e.g., a homeowner's insurance policy). Any or all of this
information may be used to populate a claims form automatically,
saving the claims adjuster and a homeowner of the property time and
effort in manually populating such a form. Insurance server 112 may
populate the claims form using the reference and/or situation
information received from ESV computing device 102. Additionally or
alternatively, insurance server 112 may transmit the claims form to
ESV computing device 102, which may populate at least one field of
the claims form with the reference information and/or the situation
information. ESV computing device 102 may then transmit the
populated claims form back to insurance server 112.
[0077] ESV: Post-Disaster Property Damage or Loss
[0078] The following examples are illustrative of various aspects
and embodiments of the disclosure described herein. Accordingly, it
should be understood that such examples are for illustration only
and do not limit the scope of the present disclosure.
[0079] FIGS. 2-8 illustrate various aspects of the disclosure using
an example of a damaged property in a post-disaster state. In
particular, FIG. 2 depicts a historical image 204 of a property
202. The historical image 204 may be representative of a past view
of the property 202 (at a time antedating a current view 300, shown
in FIG. 3). The historical image 204 may be associated with
additional reference information about the property 202 including,
for example, property specifications, the presence of a garage 206,
a vehicle 208 associated with the property 202 and/or with
occupant(s) of the property 202, and/or access to a public utility
in the form of a fire hydrant 210. The historical image 204 and the
additional reference information may be stored in reference
database 108 (shown in FIG. 1).
[0080] Additionally, the historical image 204 and/or the additional
reference information may be indexed by (e.g., have tags associated
with or be searchable by) the type of the property 202 (e.g.,
residential), a street address and/or geographic location of the
property 202, a type and/or amount of an insurance policy
associated with the property 202, and/or risks associated with the
property 202 (e.g., higher risk of a natural disaster due to
geographic location, lower risk of fire damage due to location of
the fire hydrant 210). In other words, the reference information
associated the property 202 (including the historical image 204)
may be indexed such that ESV computing device 102 (shown in FIG. 1)
may retrieve it when the property 202 is a subject of a reference
request message.
[0081] FIG. 3 depicts a current view 300 (e.g., a current view of a
user) of the property 202 (shown in FIG. 2) after the property 202
has been damaged, for example, by a natural disaster, such as a
tornado. The current view 300 may include damage 302 to the main
structure of the property 202 (i.e., the house), damage 306 to the
garage 206, an empty driveway (i.e., no vehicle 208 present), and a
pile of debris 310 located where the fire hydrant 210 was located,
as shown in FIG. 2. Although it is clear that there is damage, it
may be unclear the nature and the extent of the damage.
[0082] FIG. 4 depicts the current view 300 of the property 202 in
the post-disaster damaged state as shown in FIG. 3, including a
user computing device 404. User computing device 404 is illustrated
as a personal projection device (e.g., glasses) but may be any
other user computing device or mobile device as described herein
without departing from the scope of the disclosure. User computing
device 404 may be similar to user computing device 104, shown in
FIG. 1. In the exemplary embodiment of FIG. 4, the user of user
computing device 404 may not yet have activated or called ESV
computing device 102 (shown in FIG. 1). For example, the user may
not have yet activated or opened an app associated with reality
augmentation system 100 (shown in FIG. 1) to prepare and transmit a
reference request message to ESV computing device 102.
Alternatively, user computing device 404 has not yet received a
reference response message from ESV computing device 102, or may be
in various stages of generating and/or transmitting a reference
request message. The view displayed on user computing device 404
may be a real-time view as captured by a camera (not shown) of user
computing device 404 or may be a still image of current view
300.
[0083] FIG. 5 depicts a block diagram illustrating the transmission
of a reference request message and a reference response message
between various components of reality augmentation system 100 shown
in FIG. 1. The reference request and response may be directed to
the situation depicted in FIGS. 2-4, namely the damage of the
property 202. User computing device 104 may (A) transmit a
reference request message to ESV computing device 102. In the
exemplary embodiment, the reference request message may include an
image of the property 202 and device information (e.g., GPS data
and compass data). The reference request message may further
include a beacon activation request, should the user of user
computing device 104 have any beacons 110 associated with items of
interest that were in, around, or otherwise associated with the
property 202. In some embodiments, the reference request message
may further include user account or profile information, which may
identify the user and/or user computing device 104, such that any
proprietary or personal information that may be relevant to the
subject or the situation may be retrieved. The user profile
information may enable ESV computing device 102 to obtain, for
example, insurance policy information associated with the property.
Such policy information may be provided to reference database 108
by insurance server 112 (shown in FIG. 1).
[0084] ESV computing device 102 may receive the reference request
message and may subsequently (B) determine a subject of the image
included therein and determine that the subject of the image is the
property 202. ESV computing device 102 may use device information
to enhance the accuracy of the subject determination. ESV computing
device 102 may then (C) retrieve reference information associated
with the property 202 (i.e., the subject of the reference request
message) from reference database 108. As described above, the
reference information may include the historical image 204 (shown
in FIG. 2) and additional reference information, which may be
indexed by subject. Reference database 108 may also store a lookup
table or other reference object that facilitates identification of
beacons 110 and associated items of interest. ESV computing device
102 may (D) activate (e.g., interrogate or process signals from)
beacons 110 associated with the user.
[0085] ESV computing device 102 may use the reference request
message and the reference to (E) determine situation information
relevant to the user's situation. For example, ESV computing device
102 may compare the historical image 204 of the property 202 with
the received image of the current view 300 to identify areas that
have been damaged and/or items that are missing (e.g., not present
in the image of the current view 300). ESV computing device 102 may
further use the additional reference information, such as property
specifications and underwriting information, to identify various
building materials used in the damaged areas and/or rooms that have
been damaged (e.g., a kitchen or bathroom suffering major damage
may be more costly to repair than a bedroom, attic space, or a
closet). ESV computing device 102 may be further configured to make
damage estimates, based upon the comparisons and using spatial
calculations, and/or may subsequently determine repair cost
estimates or a dollar amount that should be requested in an
insurance claim, based upon the identified building materials and
damage estimates.
[0086] It should be understood that in other situations, such as a
reconstruction or renovation of a property or an addition to a
property, ESV computing device 102 may be configured to determine
situation information in a similar manner but "in reverse." For
example, ESV computing device 102 may perform largely the same
functions in determining a subject, retrieving reference
information, comparing a current state of the property to a past
state, identifying building materials used, and/or using spatial
calculations to determine a volume or area of addition (as opposed
to damage or volume/area missing). ESV computing device 102 may
then be configured to make estimates regarding value added to a
property and/or an appropriate insurance policy modification (e.g.,
increase in coverages or premiums, or increases in discounts (such
as due to risk prevention or mitigation)).
[0087] Returning to the example of FIG. 5, ESV computing device 102
may (E) determine additional situation information associated with
received beacon signals, in embodiments in which any beacons 110
are associated with the user and/or the property. ESV computing
device 102 may use the device information received in the reference
request message to compare the location of user computing device
104 to the location of beacon 110.
[0088] ESV computing device 102 may compile and/or summarize the
reference information and the situation information into the
reference response message. The reference response message may then
(F) be transmitted to user computing device 104, where information
in the reference response message (i.e., any or all of the
reference information and/or the situation information) may be
displayed to the user on a display of user computing device 104 to
provide an Enhanced Situation Visualization. Additionally, ESV
computing device 102 may include the relative location of beacon
110 (to user computing device 104) in the reference response
message and may facilitate the translation or conversion of the
relative location of beacon 110 into a graphical representation,
such as an icon, as will be described more fully with respect to
FIG. 8.
[0089] FIG. 6 depicts a first exemplary view of a post-disaster
Enhanced Situation Visualization 600 on user computing device 404
(as shown in FIG. 4). In the exemplary embodiment, user computing
device 404 has received the reference response message from ESV
computing device 102, as described above with respect to FIG. 5.
The reference information included in the reference response
message (i.e., the historical image 204 of the property 202, in the
exemplary embodiment of FIG. 6) has replaced the current view 300
of the property 202 on the display of user computing device 404.
(Note that the current view 300, outside of user computing device
404, has not changed, as evidenced by the pile of debris 310 and
the missing tree-top visible outside of the boundary/display area
of user computing device 404.)
[0090] Although the Enhanced Situation Visualization 600 is
depicted to be opaque (i.e., show no part of the current view 300
on the display of use computing device 404) for the sake of
clarity, it should be understood that the Enhanced Situation
Visualization 600 may be translucent. Accordingly, the user of user
computing device 404 may easily see the state of the property 202
before it was damaged. The user may be able to see, for example,
the location of the fire hydrant 210 before it was covered by
debris 310, and/or that the property 202 has an associated vehicle
208. The user may also be able to compare the historical image 204
with the current view 300 (e.g., by momentarily deactivating the
Enhanced Situation Visualization 600 and/or user computing device
404, or by moving user computing device 404) to characterize,
quantify, and/or catalog the damage done to the property 202.
Enhanced Situation Visualization 600 may be displayed as an
enhancement of a real-time view as captured by a camera of user
computing device 404 or as an enhancement of a still image of the
current view 300.
[0091] FIG. 7 depicts a second exemplary view of post-disaster
Enhanced Situation Visualization 700 on user computing device 704.
In the exemplary embodiment, user computing device 704 is
illustrated as a tablet or smartphone but may be any other user
computing device or mobile device as described herein without
departing from the scope of the disclosure. User computing device
704 may be similar to user computing device 104 shown in FIG. 1
and/or user computing device 404 shown in FIGS. 4 and 6. As
illustrated, the reference information--the description of the
building materials of the property--and situation information--the
estimated damage and the estimated cost to repair--are overlaid
upon the current view 300 of the property.
[0092] Moreover, in one exemplary embodiment, the situation
information received in the reference response message further
includes a virtual grid 712 of the property 202 in its pre-damage
state overlaid on the current view 300 of the property 202.
Accordingly, a user may not only visualize an extent of any damage
by comparing the damaged property 202 to the pre-damage virtual
grid 712, but may also quantify and catalog the damage and cost to
repair the damage using the other situation information provided.
In other embodiments, more or less reference information and/or
situation information may be displayed on the display area of user
computing device 704.
[0093] User computing device 704 may be configured to capture user
actions associated with the content displayed on user computing
device 704 and may be configured to transmit an indication of the
user actions to ESV computing device 102. For example, the user of
user computing device 704 may be an insurance adjuster that is
viewing the home to catalog the damage and provide an estimate of
the appropriate claim disbursement. Upon viewing the Enhanced
Situation Visualization 700 provided by ESV computing device 102,
the user may wish to save (e.g., copy to a document) or share
(e.g., send in an email or other message) the information
provided.
[0094] User computing device 704 may be configured to capture the
user actions by receiving a voice command, a touch command, a
button selection, or any other form of user input to user computing
device 704. For example, the user may initiate a "save" command,
and user computing device 704 may copy the content of the Enhanced
Situation Visualization to a document (e.g., a claims disbursement
form or other insurance-related tool) or may transmit an
instruction to ESV computing device 102 to do the same. A "save"
command may further generate a screen-capture of the displayed
content of the Enhanced Situation Visualization (e.g., the image of
the subject and any situation or reference information overlaid
thereupon), such that the user may view or reference the Enhanced
Situation Visualization at a later time.
[0095] FIG. 8 depicts a third exemplary view of post-disaster
Enhanced Situation Visualization 800 on user computing device 404.
In the exemplary embodiment, several icons 812, 814, 816, and 818
are displayed over the current view 300. Though there are four
icons illustrated, in other embodiments, there may be more, fewer,
or no icons displayed on the display area of user computing device
404. In the exemplary embodiment, two icons 812, 814 may generally
indicate a location of known elements of the property 202, based at
least in part on the reference information (e.g., one or more
historical images 204 of the property 202 and/or public works
information). Icon 812 may display the expected location of the
vehicle 208 shown in the historical image 204 (shown in FIG. 2),
calling the user's attention to its absence.
[0096] In some cases, the vehicle 208 may simply have been moved by
user or another person associated with the vehicle 208. In other
cases, the vehicle 208 may have been displaced due to the natural
disaster. Icon 814 may indicate the expected location of the fire
hydrant 210 shown in the historical image 204, calling attention to
the fact that the fire hydrant 210 may be covered by the debris 310
after the natural disaster.
[0097] In the exemplary embodiment, two icons 816, 818 may provide
additional situation information associated with a location of
beacon(s) 110 (shown in FIG. 1), relative to user computing device
404 ("beacon information"). These icons 816, 818 may guide the
recovery of items that may have been displaced as a result of the
natural disaster and/or the damage to the property 202. These icons
816, 818 may include an object or item tag, associating each icon
816, 818 with the item of interest that was associated with a
corresponding beacon 110 during the set-up or initialization phase,
as well as directional and/or distance information. For example,
icon 818 may display the following text: Jewelry Box|10 feet
West.
[0098] User computing device 404 may transmit an indication of user
action(s) to ESV computing device 102 after displaying the Enhanced
Situation Visualization 800. For example, user computing device 404
may transmit new device and/or location information indicating that
the user has moved in the direction of a beacon 110, such that ESV
computing device 102 may transmit an updated reference response
message with the new relative location of beacon 110. As another
example, the user may wish to focus on only one beacon 110 at a
time, and may request that one of the icons 816, 818 be removed
from the display. Such removal may be performed by user computing
device 404 or ESV computing device 102 (e.g., by sending an updated
reference response message without beacon information for the
beacon 110 corresponding to the removed icon 816 or 818).
[0099] ESV: Pre-Disaster Simulation
[0100] The natural risks inherent in a given property may be
difficult to visualize and understand--both by a consumer having
one or more insurance policies associated with properties and/or
items of interest, as well as by insurance associates (e.g., sales
associates, underwriters, insurance providers). Understanding the
risk to the property is paramount in the business of insurance, as
well as other property-related decision making. A greater
understanding of risks posed by natural conditions, hypothetical
events, and trends may increase confidence in both situations. As
used herein, a "natural event" includes any nature-related and/or
nature-generated event, including, for example, floods, storms,
lightning, wildfires, tornadoes, hurricanes, flash flooding, storm
surge, tsunamis, earthquakes, mudslides, landslides, snow, ice,
and/or blizzards.
[0101] As described above, one or more peril maps may be available
for a particular location, for example, a property. In at least
some cases, in-person surveys may traditionally have been necessary
to identify and characterize the various underwriting concerns
based upon real risks to the property. For example, an underwriter
may have had to travel to the location of the property with a
conventional peril map and use their judgement to attempt to
determine with the particular property had sufficient safeguards to
adequately temper those risks. For a property in a flood plain, the
underwriter may have had to measure or otherwise visually inspect
or estimate whether the property was far enough from a water source
or sufficiently raised from the ground to avoid damage to a main
floor of the property. In addition, the underwriter may have to
factor in various topographical or geographical features, typical
or maximum expected rainfall, and/or structural considerations.
Understandably, such visual inspection may be difficult and easily
inaccurate.
[0102] Accordingly, in one embodiment, reality augmentation system
100 (shown in FIG. 1) may be configured to provide a pre-disaster
Enhanced Situation Visualization (e.g., a disaster simulation)
including situation information associated with a property's risk
of peril. More specifically, reality augmentation system 100 may be
configured to provide a simulation of the effect(s) of various
natural phenomena on a subject property to a user interface of a
user computing device (e.g., user computing device 104, shown in
FIG. 1).
[0103] FIG. 9 illustrates one example of a screenshot of an
Enhanced Situation Visualization 900 ("ESV screenshot" 900)
displayed to a user on a user computing device 904. Although user
computing device 904 is illustrated as a tablet or smartphone, in
various other embodiments, user computing device 904 may be any
other user computing device or mobile device as described herein.
As described elsewhere herein, the user may access ESV computing
device 102 (shown in FIG. 1) to transmit and receive messages using
an app on user computing device 904, via a browser, or using any
other data transmission method. In the exemplary embodiment, the
subject of a reference request message (which may have been
previously transmitted by user computing device 904 to ESV
computing device 102) is a property 902.
[0104] Property 902, in the exemplary embodiment, may be located in
an area prone to heavy rainfall and/or flooding. Accordingly, the
reference request message may have been generated by an insurance
underwriter or a homeowner interested in the risk to property 902
for flood damage (or other water-related damage or loss). Moreover,
property 902 is located nearby a water source 906 (and accordingly
is depicted as casting a reflection of property 902 in water source
906). Using traditional methods, the insurance underwriter or
homeowner ("user") may have had to consult various two-dimensional
peril and/or topographical maps, as well as historical floodplain
data, to attempt to ascertain the level of risk to property 902 due
to a flood or other damage from water source 906.
[0105] In the exemplary embodiment, the user may generate the
reference request message using user computing device 904. The
reference request message may include various user-input variables.
In the exemplary embodiment, variables may include an amount of
rainfall (in inches), a duration of rainfall (in hours), and/or a
season (to set a groundwater baseline for the location including
property 902). Additional, less, or alternate variables may be
used, including wind speed, hail amount, and/or storm surge. The
user may adjust the variables using associated controls 910, 912,
and/or 914. Controls 910, 912, and 914 are illustrated as sliders
in the exemplary embodiment but may take any other form in
alternate embodiments, including, for example, text fields,
drop-down menus, arrow-key controls, or any other control.
[0106] The reference request message may also include an image of a
view of a property 902, which may further include water source 906.
Using a real-time (or relatively recent, e.g., taken within hours,
days, or weeks of the generation of the reference request message)
image of property 902 may be advantageous in that ESV computing
device 102 may incorporate any changes to topography or
specifications of property 902 in determining situation information
and generation ESV 900.
[0107] In some cases, the user may choose to capture an image of
property 902 and then transmit the reference request message
including the image at a later time. In such a case, the image may
not be of a "current view" in the sense that a user is still seeing
such a view, but may still be considered a "current view" in the
sense that the view is current (or relatively so) relative to the
situation about which the user is sending the reference request
message. Accordingly, the phrase "current view" should not be
construed to limit the embodiments described herein to only those
reference request messages sent immediately after capturing an
image.
[0108] As described above, in particular with respect to FIG. 5,
ESV computing device 102 may be configured to receive the reference
request message and determine a subject thereof.
[0109] In one embodiment, ESV computing device 102 may perform
object recognition on the received image and/or may use various
device information (in particular, location information such as GPS
coordinates) to determine a location of user computing device 904
and, therefore, a location of the subject of the image. Upon
determination that the subject of the reference request message is
property 902, ESV computing device 102 may retrieve reference
information associated with property 902 from reference database
108 (shown in FIG. 1), including, for example, various peril maps
(which may be internal or proprietary to an insurance provider
associated with insurance server 112, as shown in FIG. 1, and may
be provided to reference database 108 and/or to ESV computing
device by insurance server 112), topographical maps, risk analyses
for a location including property 902, and/or historical
weather/flood/rainfall data for the location.
[0110] Using the reference information and the user-input
variables, ESV computing device 102 may generate ESV screenshot 900
to indicate at least one effect on the property due to the proposed
natural event (i.e., a rainfall having the user-input variables
described above). An effect may include no damage, some damage,
total property loss, partial property loss, land damage, and/or
various other potential effects, which may be characterized using a
loss estimate, as will be described further herein. ESV screenshot
900 may include various situation information determined by ESV
computing device 102, such as a 100-year flood project water level
920 (which corresponds to a maximum anticipated water level under
the worst possible conditions) and/or a custom projected water
level 922 (which corresponds to the water level resulting from the
conditions expressed by the user-input variables).
[0111] In the exemplary embodiment, 100-year flood level 920 and
custom water level 922 are overlaid on a real-time view of property
902 as captured by a camera on user computing device 904. In
particular, 100-year flood level 920 and custom water level 922 are
depicted as lines or levels overlaid corresponding to the unique
topography associated with property 902, which may substantially
reduce or eliminate any "guesswork" or estimation on the part of
the user. Levels 920 and 922 may be transposed onto
three-dimensional and/or two-dimensional images and/or video
captures of property 902.
[0112] ESV screenshot 900 may further include a loss estimate 924,
which communicates to the user the potential for loss or damage to
the property by the 100-year and/or custom water level(s) 920, 922.
Loss estimate 924 may be generated by ESV computing device 102
using various internal underwriting information associated with
property 902, including property value of insured property 902
and/or structural specifications. For example, if a flood
simulation showed water levels that would damage a basement and a
detached structure associated with property 902, the value (or
percentage of total value) of those structures may be displayed to
the user, relative to the estimated severity and duration of the
disaster (i.e., water levels high enough to flood a basement would
not necessarily result in a total loss, but may result in a total
loss for a detached structure that is submerged or washed
away).
[0113] Incentives may be offered to those homeowners (or other
users associated with property 902) that actively respond to the
content shown in the ESV (such as that displayed in ESV screenshot
900), for example, by taking additional preventative or mitigating
measures. For example, an insurance provider may offer a discount,
credit, or service rebate to a homeowner that adds sandbags around
property 902 if ESV screenshot 900 were to indicate that property
902 is at risk of damage caused by a flood or other water-related
event. Additionally, ESV computing device 102 may be configured to
generate and/or display suggestions for such mitigating or
preventative measures. ESV computing device 102 may receive an
indication of user compliance with a risk-mitigating suggestion.
ESV computing device 102 may then recommend or facilitate an
adjustment in an insurance property associated with the property
902 (e.g., one or more of the incentives described above).
[0114] As another example, though not shown in FIG. 9, similar ESV
generation and display may be performed for a property at risk for
wildfire damage. The user may use their user computing device
(e.g., user computing device 904) to capture an image of a property
in a location having a moderate to high wildfire risk. The image
may further include any surrounding and/or nearby structures and/or
vegetation, which may contribute to the property's risk of wildfire
damage (e.g., a nearby barn, silo, forest, farm, brush, hedge,
etc.). Using a real-time (or relatively recent) image of the
property may be advantageous in that older pictures may be out of
date (in terms of vegetation and/or topography, for example) and
thereby cause inaccurate determination of situation information
associated with the property. There may be user-input variables
available for adjustment, including rainfall, average temperature,
maximum temperature, and/or maximum wind speed. Additional, less,
or alternate variables may be used. Upon receiving a reference
request message including the image and any available user-input
variables, ESV computing device 102 may determine that the subject
of the message is the property and may retrieve reference
information therefor. Reference information may include various
peril maps (which may be internal or proprietary to an insurance
provider associated with insurance server 112, and may be provided
to reference database 108 and/or to ESV computing device 102 by
insurance server 112), topographical maps, risk analyses for a
location including property 902, and/or historical
weather/wildfire/rainfall data for the location.
[0115] ESV computing device 102 may generate an ESV for the
property including wildfire approach limits, similar to the
overlaid water levels 920, 922 described and illustrated above,
which correspond to an expected approach path and/or limit (i.e.,
how close to the property a wildfire may be expected to reach).
Additionally, the ESV may identify whether nearby vegetation or
other objects pose an added risk to the property. Not only may a
setback area be more precisely and/or accurately identified to the
user, but incentives may be subsequently offered to those
homeowners (or other users associated with the property) that
actively respond to the ESV. For example, an insurance provider may
offer a discount, credit, or service rebate to a homeowner that
removes or relocates vegetation or other objects that may endanger
the property in the event of a wildfire. Additionally, ESV
computing device 102 may be configured to generate and/or display
suggestions for such mitigating or preventative measures.
[0116] FIG. 10 illustrates a screenshot of an exemplary
pre-disaster Enhanced Situation Visualization (ESV) 1000 ("ESV
screenshot" 1000) including a simulation. In the exemplary
embodiment, property 902, as shown in FIG. 9, is the subject of
another reference request message. Variables of rainfall, duration,
and season are still available to the user, along with
corresponding controls 910, 912, and 914, for manipulation on user
computing device 904 (shown in FIG. 9). However, in the exemplary
embodiment of FIG. 10, the user has adjusted all of controls 910,
912, and 914 to their respective maxima (i.e., a maximum rainfall
of 3 inches, a continuous duration of 9 hours, and a Late Spring
season, which may historically result in the highest groundwater
baseline).
[0117] ESV computing device 102 (shown in FIG. 1) may have received
the reference request message including an image of property 902
and water source 906, as shown in FIG. 9, as well as the user input
variables described above. After retrieving associated reference
information, ESV computing device 102 may determine that, in the
custom situation reference in the reference request message, a
water level 1020 may exceed a boundary surrounding property 902
such that property 902 may be washed away from its location. Using
such processes as augmented reality, image capture-overlay, and
animation capabilities, ESV computing device 102 may generate ESV
screenshot 1000 including an animated simulation of water levels
from water source 906 rising until property 902 is dislodged from
its location and is washed away, as shown in FIG. 10.
[0118] A more severe loss estimate 1024 may additionally be
displayed to the user, which may be generated as described above.
In an alternate embodiment in which damage may not have been
predicted to be so severe, an animated simulation and/or a
still-image ESV (e.g., that displayed in ESV screenshot 900) may
include discoloration or other overlays to show areas that may be
damaged under the conditions selected by the user. The user may
have an option (e.g., a radio button, toggle, or check box) to
choose between an animated ESV or a still-image ESV. The user may
further have an option to generate a screen-capture of the content
of the ESV, such that any content may be viewed and/or referenced
at a later time.
[0119] In a further enhancement of the systems and methods
described herein, particularly those described with respect to
FIGS. 9 and 10, a user may generate and transmit a reference
request message including a user-input risk tolerance threshold.
The user (e.g., a property owner or an insurance underwriter or
associate) may establish a risk tolerance threshold based upon a
level or severity of damage that may be acceptable or the maximum
allowed for coverage. For example, the user may establish a risk
tolerance threshold that flooding in a basement of the property is
an acceptable amount of risk, but that flooding or damage to a main
floor is unacceptable. ESV computing device 102 may use the risk
tolerance threshold to generate one or more ESVs (e.g.,
simulation(s) or overlays) that report the particular conditions
under which the risk tolerance threshold may be met (i.e., a
particular characteristic of the natural event that satisfies the
risk tolerance threshold). Accordingly, if a risk tolerance
threshold established by an insurance provider indicates that
flooding of any kind is unacceptable, and an ESV generated by ESV
computing device 102 indicates that only minor weather conditions
(e.g., average rainfall in any season) may produce a flood in the
property and thereby satisfy the risk tolerance threshold, the
insurance provider may choose not to offer flood insurance for the
property.
[0120] Additionally, ESV computing device 102 may use any situation
information, simulation information, ESV, overlays, etc., to update
and/or enhance exiting peril and/or topographical maps stored at
reference database 108. For example, if a generated ESV indicates a
higher level of flood or wildfire risk than indicated in a
particular peril map for a particular property, ESV computing
device 102 may associate new peril map values and/or information
for the property and/or location including the property, based upon
the generated ESV.
[0121] The present embodiments may include (1) the use of augmented
reality technology to produce disaster simulations in the context
of insurance; and/or (2) the mapping of peril maps to video capture
images on smartphones and smart glasses. Potential application
and/or uses of the present embodiments may include (i) enabling
agents and/or underwriters to better understand and communicate the
natural risk inherent in a particular property (e.g., by
visualizing for a homeowner their need for increased coverage
related to a particular natural hazard); (ii) enabling better
underwriting decision making (i.e., on the spot, instead of using
abstract peril maps only); (iii) enabling consumers and realtors to
make decisions about prospective property purchases, with respect
to natural hazard risks; and/or (iv) enabling consumers to assess
their own insurance risks, including their tolerance for risk and
possible preventative mitigation efforts that could be taken (for
example, as a free application on a smartphone that could simulate
the effects of different wildfire setbacks/defensible space,
suggesting the right distance to be maintained between the house
and surrounding brush areas).
[0122] Exemplary User Computing Device
[0123] FIG. 11 depicts an exemplary configuration of a user
computing device 1102. User computing device 1102 may include, but
is not limited to, user computing device 104 shown in FIG. 1, user
computing device 404 shown in FIG. 4, user computing device 704
shown in FIG. 7, and/or user computing device 904 shown in FIG.
9.
[0124] User computing device 1102 may include a processor 1105 for
executing instructions. In some embodiments, executable
instructions may be stored in a memory area 1110. Processor 1105
may include one or more processing units (e.g., in a multi-core
configuration). Memory area 1110 may be any device allowing
information such as executable instructions and/or other data to be
stored and retrieved. Memory area 1110 may include one or more
computer-readable media.
[0125] User computing device 1102 may also include at least one
media output component 1115 for presenting information to a user
1100. Media output component 1115 may be any component capable of
conveying information to user 1100. In some embodiments, media
output component 1115 may include an output adapter such as a video
adapter and/or an audio adapter. An output adapter may be
operatively coupled to processor 1105 and operatively coupleable to
an output device such as a display device (e.g., a liquid crystal
display (LCD), organic light emitting diode (OLED) display, cathode
ray tube (CRT), or "electronic ink" display) or an audio output
device (e.g., a speaker or headphones). In some embodiments, media
output component 1115 may be configured to present an interactive
user interface (e.g., a web browser or client application) to user
1100. The interactive user interface may include, for example, a
reality augmentation interface for requesting and viewing Enhanced
Situation Visualization.
[0126] In some embodiments, user computing device 1102 may include
an input device 1120 for receiving input from user 1100. Input
device 1120 may include, for example, a keyboard, a pointing
device, a mouse, a stylus, a touch sensitive panel (e.g., a touch
pad or a touch screen), a camera, a gyroscope, an accelerometer, a
position detector, and/or an audio input device. A single component
such as a touch screen may function as both an output device of
media output component 1115 and input device 1120.
[0127] User computing device 1102 may also include a communication
interface 1125, which is communicatively coupleable to a remote
device such as ESV computing device 102 (shown in FIG. 1).
Communication interface 1125 may include, for example, a wired or
wireless network adapter or a wireless data transceiver for use
with a mobile phone network (e.g., Global System for Mobile
communications (GSM), 3G, 4G or Bluetooth) or other mobile data
network (e.g., Worldwide Interoperability for Microwave Access
(WIMAX)).
[0128] Stored in memory area 1110 are, for example,
computer-readable instructions for providing a user interface to
user 1100 via media output component 1115 and, optionally,
receiving and processing input from input device 1120. A user
interface may include, among other possibilities, a web browser and
client application. Web browsers enable users 1100 to display and
interact with media and other information typically embedded on a
web page or a website from a web server associated with a merchant.
A client application allows users 1100 to interact with a server
application associated with, for example, a vendor or business.
[0129] Exemplary Server Computing Device
[0130] FIG. 12 depicts an exemplary configuration of a server
computing device 1202. Server computing device 1202 may be
representative of ESV computing device 102, database server 106,
and/or insurance server 112 (all shown in FIG. 1). Server computing
device 1202 may include a processor 1204 for executing
instructions. Instructions may be stored in a memory area 1206, for
example. Processor 1204 may include one or more processing units
(e.g., in a multi-core configuration).
[0131] Processor 1204 may be operatively coupled to a communication
interface 1208 such that server computing device 1202 may be
capable of communicating with a remote device such as user
computing device 1102 or another server computing device 1202. For
example, communication interface 1208 may receive requests from
user computing device 1102 via the Internet.
[0132] Processor 1204 may also be operatively coupled to a storage
device 1210. Storage device 1210 may be any computer-operated
hardware suitable for storing and/or retrieving data. In some
embodiments, storage device 1210 may be integrated in server
computing device 1202. For example, server computing device 1202
may include one or more hard disk drives as storage device 1210. In
other embodiments, storage device 1210 may be external to server
computing device 1202 and may be accessed by a plurality of server
computing devices 1202. For example, storage device 1210 may
include multiple storage units such as hard disks or solid state
disks in a redundant array of inexpensive disks (RAID)
configuration. Storage device 1210 may include a storage area
network (SAN) and/or a network attached storage (NAS) system.
[0133] In some embodiments, processor 1204 may be operatively
coupled to storage device 1210 via a storage interface 1212.
Storage interface 1212 may be any component capable of providing
processor 1204 with access to storage device 1210. Storage
interface 1212 may include, for example, an Advanced Technology
Attachment (ATA) adapter, a Serial ATA (SATA) adapter, a Small
Computer System Interface (SCSI) adapter, a RAID controller, a SAN
adapter, a network adapter, and/or any component providing
processor 1204 with access to storage device 1210.
[0134] Memory areas 1110 (shown in FIG. 11) and 1206 may include,
but are not limited to, random access memory (RAM) such as dynamic
RAM (DRAM) or static RAM (SRAM), read-only memory (ROM), erasable
programmable read-only memory (EPROM), electrically erasable
programmable read-only memory (EEPROM), and non-volatile RAM
(NVRAM). The above memory types are example only, and are thus not
limiting as to the types of memory usable for storage of a computer
program.
[0135] Exemplary Computer-Implemented Methods for ESV
[0136] FIG. 13 illustrates exemplary Enhanced Situation
Visualization 1300, for example, in a post-disaster situation. In
the exemplary embodiment, any or all of the steps depicted may be
performed by components of reality augmentation system 100 (shown
in FIG. 1). The Enhanced Situation Visualization 1300 may begin at
point 1302, which may represent, in general, a user's desire for
Enhanced Situation Visualization (e.g., the user wishes to receive
more information about their situation).
[0137] Group 1304 may generally represent the generation of a
reference request message. The user, at user computing device 104
(shown in FIG. 1), may capture 1306 an image of their situation,
representative of their current view. User computing device 104 may
substantially simultaneously capture 1308 device information, which
may include location information associated with the situation, the
current view, and/or a subject of the current view. The user may
also wish to generate 1310 a beacon activation request, should the
user need to locate associated items of interest. Point 1312 may
generally represent the transmission of the reference request
message to ESV computing device 102 (shown in FIG. 1) and/or the
receipt of the reference request message by ESV computing device
102.
[0138] Group 1314 may generally represent the generation of a
reference response message. ESV computing device 102 may determine
1316 a subject of the reference request message, for example, using
object recognition, device information, and any user input (e.g.,
the user may specify a subject of the image, or the image may
include SKU data or barcode data that identifies the subject of the
image). ESV computing device 102 may further retrieve 1318
historical images of the subject, and may retrieve 1320 insurance
policy information for any insurance policy associated with the
subject of the image. For example, if the subject of the reference
request message is a property, ESV computing device 102 may
retrieve a homeowner's insurance policy or renter's insurance
policy, as well as any additional policies, such as flood insurance
or auto insurance for vehicles associated with the owner or renter
of the property. ESV computing device 102 may also retrieve 1322
any public records associated with the subject (e.g., tax records,
MLS listings, etc.). Point 1324 may generally represent the
transmission of the reference response message and/or the receipt
of the reference response message by user computing device 104.
[0139] In one exemplary embodiment, the reference response message
may include instructions for user computing device 104 to display
the content of the message (i.e., the Enhanced Situation
Visualization) overlaid on, beside, or instead of the current view
on a display of user computing device 104. In another exemplary
embodiment, user computing device 104 may determine 1326 which
content to display and/or how to display the content without
instruction from ESV computing device 102. User computing device
104 may subsequently capture user actions 1328 associated with the
Enhanced Situation Visualization and may transmit an indication of
such user actions to ESV computing device 102, where necessary.
[0140] FIG. 14 illustrates an exemplary Enhanced Situation
Visualization for disaster simulation provided by the reality
augmentation system 100 shown in FIG. 1. In the exemplary
embodiment, any or all of the actions depicted may be performed by
components of reality augmentation system 100 (shown in FIG. 1).
The Enhanced Situation Visualization 1400 may begin at point 1402,
which may represent, in general, a user's desire for pre-disaster
Enhanced Situation Visualization (e.g., the user wishes to receive
more information about risk to their property).
[0141] Group 1404 may generally represent the generation of a
reference request message. The user, at user computing device 104
(shown in FIG. 1), may capture 1406 an image of their situation,
representative of their current view. For example, the image may be
of a property in a pre-disaster state, including topographical or
natural features associated with the property. User computing
device 104 may substantially simultaneously capture 1408 user-input
variables associated with the nature of the reference request. For
example, the user-input variables may be associated with a natural
event for which the property may be at risk. User computing device
104 may substantially simultaneously capture 1410 device
information, such location information associated with the
property. Point 1412 may generally represent the transmission of
the reference request message to ESV computing device 102 (shown in
FIG. 1) and/or the receipt of the reference request message by ESV
computing device 102.
[0142] Group 1414 may generally represent the generation of a
reference response message. ESV computing device 102 may determine
a subject of the reference request message, for example, by
performing 1416 object recognition, and/or using location
information and any user input (e.g., the user may specify the
subject of the image). ESV computing device 102 may further
retrieve and compare 1318 peril maps associated with property,
using the information captured in the image and the
topographical/natural risk assessments in the peril map(s) to
determine a level of risk of peril at the property. ESV computing
device 102 may further generate 1420 a loss estimate for the
property in the event that the natural disaster occurs (which may
be as low $0, if the property has a low risk associated with the
natural event, but may have no upper limit). Generating 1420 the
loss estimate may include retrieving insurance policies and/or
other underwriting information associated with the property. Point
1422 may generally represent the transmission of the reference
response message and/or the receipt of the reference response
message by user computing device 104.
[0143] In one exemplary embodiment, the reference response message
may include instructions for user computing device 104 to display
1424 the content of the message (i.e., the Enhanced Situation
Visualization, which may include a still-image or animated
simulation) overlaid on, beside, or instead of the current view on
a display of user computing device 104.
[0144] FIG. 15 illustrates a flowchart of an exemplary
computer-implemented method 1500 for providing Enhanced Situation
Visualization using reality augmentation system 100 shown in FIG.
1. In the exemplary embodiment, method 1500 may be performed by ESV
computing device 102 (shown in FIG. 1). Method 1500 may include
receiving 1502 a reference request message from a user computing
device (e.g., user computing device 104, shown in FIG. 1). The
reference request message may include an image representative of a
current view of a user of the user computing device. In accordance
with the various embodiments described herein, the image may
include a property in a current state. The reference request
message may further include at least one user-input variable, the
user-input variable associated with a characteristic of a natural
event. Method 1500 may also include determining 1504 a first
subject of the image. In accordance with the various embodiments
described herein, the subject may be the property.
[0145] Method 1500 may further include retrieving 1506 reference
information associated with the first subject of the image. The
reference information may be retrieved 1506 from a reference
database (e.g., reference database 108, shown in FIG. 1). In
accordance with the various embodiments described herein, the
reference information may include a peril map indicating a risk
level associated with the natural event for the property.
[0146] In addition, method 1500 may include determining 1508
situation information specific to a situation of the first subject.
The situation information may be determined 1508 based upon the
reference request message and the reference information. In
accordance with the various embodiments described herein, the
situation information may include the situation information
includes a loss estimate based upon the peril map, the at least one
user-input variable, and the natural event. Method 1500 may also
include generating a reference response message including the
situation information and the reference information.
[0147] Method 1500 may further include transmitting 1512 the
reference response message to the user computing device. The
reference response message may cause at least one of the situation
information and the reference information to be displayed on the
user computing device as overlaid upon the current view of the user
to provide an Enhanced Situation Visualization to the user. Method
1500 may include additional, less, or alternate actions, including
those discussed elsewhere herein.
[0148] Exemplary Computing Device for ESV
[0149] FIG. 16 depicts a diagram 1600 of components of one or more
exemplary computing devices 1610, for example ESV computing device
102 (shown in FIG. 1), that may be used in embodiments of the
described systems and methods. FIG. 18 further shows a
configuration of data in database 1620, which may be similar to
database 108 (also shown in FIG. 1). Database 1620 may include, for
example, reference information 1622 associated with one of more
subjects 1624, and other data as described elsewhere herein.
Database 1620 may be in communication with several separate
components within computing device 1610, which perform specific
tasks.
[0150] More specifically, computing device 1610 may include a
receiving component 1630, which may be configured to receive a
reference request message from a user computing device (e.g., user
computing device 104, shown in FIG. 1). The reference request
message may include an image representative of a current view of a
user of the user computing device. Computing device 1610 may also
include a determining component 1640, which may be configured to
determine a first subject 1624 of the image. Additionally,
determining component 1640 may be further configured to determine
situation information specific to a situation of the first subject
1624, based upon the reference request message and reference
information 1622. A retrieving component 1650 may be configured to
retrieve reference information 1622 associated with the determined
first subject 1624. Retrieving component 1650 may retrieve
reference information 1622 from database 1620 and/or any other
memory device.
[0151] Computing device 1610 may further include a generating
component 1660, which may be configured to generate a reference
response message including the situation information and the
reference information 1622. Computing device 1610 may still further
include a transmitting component 1670, which may configured to
transmit the reference response message to the user computing
device.
[0152] Exemplary Insurance-Related Functionality
[0153] As noted herein, the present embodiments may relate to
collecting, generating, providing, and/or displaying information
associated with an insurance-related event (such as damage to a
property due to a natural disaster, home fire, or other incident)
via a reality augmentation system 100, as shown in FIG. 1.
Information related to a current state of a subject, such as a
post-disaster damaged property or a pre-disaster property, may be
collected by a user computing device 104 (also shown in FIG. 1).
For example, user computing device 104 may capture an image of the
property in a post-disaster state. The image may be stored locally
on memory 1110 of user computing device 1104 (shown in FIG. 11)
and/or transmitted to insurance server 112, ESV computing device
102, (both shown in FIG. 1) and/or to an external database for
storage and/or analysis.
[0154] In particular, ESV computing device 102 may perform analysis
of the image to determine that the subject of the image is the
damaged property. The image may be analyzed using, for example,
object recognition (OR), optical character recognition (OCR), or
other computing techniques. ESV computing device 102 may further
generate situation information unique to the situation provided by
user computing device 104. More specifically, ESV computing device
102 may compare the current state of the property to a previous
(i.e., undamaged) state. Such situation information may facilitate,
or be used to, (i) determine, catalog, characterize, or quality an
extent, severity, and/or nature of the damage to the property; (ii)
prepare or populate a claims form associated with the damage to the
property; (iii) determine whether there is any additional potential
for damage or injury (e.g., whether there is a downed power line);
(iv) avoid unnecessary and/or inaccurate claims; and/or other
insurance-related purposes or activities.
[0155] In one aspect, ESV computing device 102 may be programmed to
identify any damaged areas of the property in the image to populate
and/or adjust a claims form associated with the property.
Additionally, ESV computing device 102 may be programmed to
identify any missing features, elements, and/or aspects of the
image such that the missing features, elements, and/or aspects may
be located (e.g., by debris removal or using beacon information) or
the loss may be added to an existing claims form or loss
report.
[0156] In accordance with one exemplary embodiment, an insurance
server (e.g., insurance server 112) may be provided. The insurance
server may include at least one processor and may be in
communication with a database. The at least one processor may be
configured to store a claims form, wherein the claims form includes
at least one fillable field associated with an insurance claims.
The at least one processor may also be configured to receive a
reference response message generated at an Enhanced Situation
Visualization computing device (e.g., ESV computing device 102).
The reference response message may identify a property in a current
state, and may include reference information associated with the
property and situation information associated with the current
state of the property. The situation information may include at
least one difference between the current state of the property and
a prior state of the property. The at least one processor may be
further configured to identify an insurance policy associated with
the property, and populate the at least one field of the claims
form with at least one of the reference information and the
situation information. The populated claims form may be used to
update and/or adjust the insurance policy associated with the
property.
[0157] The situation information may alternatively include a loss
estimate for the property. The insurance server may be configured
to generate the loss estimate and/or to provide various
underwriting and/or public information to facilitate the generation
of the loss estimate. For example, the insurance server may use (or
allow use of) an insurance policy associated with the property to
determine a value of the property. Accordingly, if a pre-disaster
reference request message is received that requests a risk
assessment for the property for flood damage, and associated peril
maps indicate that a basement of the property is at risk of being
flooded, the insurance server may use the value of the property to
determine a loss estimate in the event that the basement (e.g., a
particular percentage of the property) is flooded.
[0158] Additional Exemplary Computer-Implemented Methods
[0159] In one aspect, a computer-based method for updating an
insurance policy based upon an Enhanced Situation Visualization
(ESV) on a user interface of a user computing device may be
provided. The method may include storing, by an insurance server at
a database, a claims form, wherein the claims form includes at
least one fillable field associated with an insurance claim. The
method may further include receiving, by the insurance server, an
ESV from a user computing device, wherein the ESV identifies at
least one difference between a current state of a property and a
prior state of the property. In addition, the method may include
identifying, by the insurance server, an insurance policy
associated with the property, and populating, by the insurance
server, the at least one field of the claims form based upon the at
least one difference. The method may also include updating and/or
adjusting, by the insurance server at the database, the insurance
policy based upon the populated claims form.
[0160] In another aspect, another computer-based method for
updating an insurance policy based upon an Enhanced Situation
Visualization (ESV) on a user interface of a user computing device
may be provided. The method may include receiving, at an insurance
server, an ESV from an ESV computing device and/or a user computing
device. The ESV includes situation information, the situation
information including a loss estimate based upon an image of a
property, at least one user-input variable associated with a
characteristic of a natural event, and a peril map indicating a
risk level associated with the natural event for the property. The
method may further include generating, by the ESV computing device
and/or the insurance server, a risk-mitigating suggestion based
upon the situation information. The method may further include
receiving, by the ESV computing device and/or the insurance server,
an indication of user compliance with the risk-mitigating
suggestion, and/or adjusting, by the insurance server, an insurance
policy, premium, or discount associated with the property.
[0161] Exemplary Methods and Simulations
[0162] In one aspect, a computer-implemented method of natural
disaster simulation may be provided. The method may include (1)
property via a mobile device camera, and/or GPS coordinates from a
mobile device GPS unit; (2) determining an extent (such as a
predicted future or past extent) of a natural disaster at a
location of the insured property based upon computer analysis of
(i) the video or images, (ii) GPS coordinates, and/or (iii) weather
data via a local or remote processor, such as the mobile device
processor or a remote server associated with an insurance provider;
(3) determining a graphic (e.g., one or more icons) virtually
representing an extent of the natural disaster at the location of
the insured property via the local or remote processor; and/or (4)
superimposing the graphic virtually representing the extent of the
natural disaster on the video or images of the insured property to
visually depict whether the insured property is at risk, or has
incurred damage, due to the natural disaster and facilitate
mitigating or preventive measures that lessen or prevent damage to
the insured property caused by the natural disaster. The method may
include additional, less, or alternate actions, including those
discussed elsewhere herein, and/or implemented via one or more
local or remote processors, and/or computer-executable instructions
stored on non-transitory computer media or medium.
[0163] For instance, the mitigating or preventive measure may
involve removing personal belongings from the insured property
prior to the natural disaster causing damage. The mitigating or
preventive measure may involve adding sandbags around the insured
property if the graphic indicates that the insured property is at
risk of damage caused by the natural disaster, the natural disaster
being a flood or water-related. The mitigating or preventive
measure may involve cutting/removing shrubbery or trees around the
insured property if the graphic indicates that the insured property
is at risk of damage caused by the natural disaster, the natural
disaster being a wild fire or fire-related.
[0164] The method may include generating insurance-related
recommendations for display on the mobile device display if the
graphic indicates that the insured property is at risk of damage
caused by the natural disaster, the natural disaster being a flood
or wild fire. The method may include adjusting insurance premiums
or discount based upon the extent of the natural disaster
estimated, and/or an insurance customer taking or preforming the
insurance-related recommendations if the graphic indicates that the
insured property is at risk of damage caused by the natural
disaster, the natural disaster being a flood or wild fire.
[0165] Determining the extent (such as a predicted future or past
extent) of the natural disaster at the location of the insured
property may be based upon computer analysis of (i) the video or
images taken by the mobile device, (ii) GPS coordinates generated
by the mobile device GPS unit, (iii) weather data, including water
maps, rain amounts, wind speed and direction, and/or weather
forecasts and radar information, and/or (iv) topographical or other
map information. Additionally or alternatively, the graphic
depicting the extent of the natural disaster may be computed using
various variables, including amount of rainfall based upon the GPS
information, continuous duration of rain, and/or groundwater
amount.
[0166] In another aspect, a computer-implemented method of natural
disaster simulation may be provided. The method may include (1)
determining, via one or more processors (such as a mobile device or
remote server processor) an extent of a natural disaster, such as a
future or past extent of an ongoing natural disaster at a specific
location (such as a GPS location associated with the mobile device
and/or images/video acquired by the mobile device); (2) generating
a graphic, via the one or more processors, virtually representing
the extent of the natural disaster at the specific location; and/or
(3) displaying, via the one or more processors, the graphic on top
of still images or video of the specific location taken by the
mobile device, the still images or video including images of an
insured property, and the graphic depicting whether or not the
insured property is at risk of damage due to the natural disaster,
or has incurred damage caused by the natural disaster to facilitate
taking preventive or mitigating actions and/or providing
insurance-related recommendations to an insurance customer via
their mobile device.
[0167] The method may include additional, less, or alternate
actions, including those discussed elsewhere herein, and/or
implemented via one or more local or remote processors, and/or
computer-executable instructions stored on non-transitory computer
media or medium. For instance, the one or more processors may use
the extent of the natural disaster calculated to estimate damage to
an insured property, and/or may handle or adjust an insurance claim
covering the insured property using the estimated damage. In one
embodiment, the natural disaster may be a flood or wild fire,
and/or the damage to the insured property may be water or fire
damage.
[0168] Exemplary Systems for Disaster Simulation
[0169] In one aspect, a computer system or mobile device configured
for natural disaster simulation may be provided. The computer
system or mobile device may include a one or more local or remote
processors configured to: (1) capture video or images of an insured
property via a mobile device camera (and/or the one or more local
or remote processors), and GPS coordinates from a mobile device GPS
unit or nearby GPS units (such as a house smart controller); (2)
determine an extent (such as a predicted future or past extent) of
a natural disaster at a location of the insured property based upon
computer analysis by the one or more local or remote processors of
(i) the video or images, (ii) GPS coordinates, and/or (iii) weather
data, such as via the mobile device processor or a remote server
associated with an insurance provider; (3) determine or generate a
graphic (e.g., one or more icons) virtually representing an extent
of the natural disaster at the location of the insured property via
the one or more local or remote processors; and/or (4) superimpose
the graphic virtually representing the extent of the natural
disaster on the video or images of the insured property (via the
one or more local or remote processors) to visually depict whether
the insured property is at risk, or has incurred damage, due to the
natural disaster and facilitate mitigating or preventive measures
that lessen or prevent damage to the insured property caused by the
natural disaster.
[0170] The mitigating or preventive measure may involve removing
personal belongings from the insured property prior to the natural
disaster causing damage; adding sandbags around the insured
property if the graphic indicates that the insured property is at
risk of damage caused by the natural disaster, the natural disaster
being a flood or water-related; and/or cutting/removing shrubbery
or trees around the insured property if the graphic indicates that
the insured property is at risk of damage caused by the natural
disaster, the natural disaster being a wild fire or
fire-related.
[0171] The one or more local or remote processors may be configured
to generate insurance-related recommendations for display on the
mobile device (or another local or remote computing device) if the
graphic indicates that the insured property is at risk of damage
caused by the natural disaster, the natural disaster being a flood
or wild fire. Additionally or alternatively, the one or more local
or remote processors may be configured to adjust insurance premiums
or discount based upon the extent of the natural disaster
estimated, and/or generate and present one or more
insurance-related recommendations if the graphic indicates that the
insured property is at risk of damage caused by the natural
disaster, the natural disaster being a flood or wild fire.
Determining the extent (such as a predicted future or past extent)
of the natural disaster at the location of the insured property is
based upon computer analysis by the one or more local or remote
processors of (i) the video or images taken by the mobile device,
(ii) GPS coordinates generated by the mobile device GPS unit, (iii)
weather data, including water maps, rain amounts, wind speed and
direction, and/or weather forecasts and radar information, and/or
(iv) topographical or other map information. The graphic depicting
the extent of the natural disaster may be computed by the one or
more local or remote processors using various variables, including
amount of rainfall based upon the GPS information, continuous
duration of rain, and/or groundwater amount.
[0172] In another aspect, a computer system or mobile device for
presenting a natural disaster simulation on a display may be
provided. The computer system or mobile device may include one or
more local or remote processors configured to: (1) determine, via
the one or more processors (such as a mobile device or remote
server processor), an extent of a natural disaster, such as a
future or past extent of an ongoing natural disaster at a specific
location (such as a GPS location associated with the mobile device
and/or images/video acquired by the mobile device); (2) generate a
graphic, via the one or more processors, virtually representing the
extent of the natural disaster at the specific location; and/or (3)
display, via the one or more processors, the graphic on top of
still images or video of the specific location taken by the mobile
device, the still images or video including images of an insured
property, and the graphic depicting whether or not the insured
property is at risk of damage due to the natural disaster, or has
incurred damage caused by the natural disaster to facilitate taking
preventive or mitigating actions and/or providing insurance-related
recommendations to an insurance customer via their mobile
device.
[0173] The one or more processors use the extent of the natural
disaster calculated to estimate damage to an insured property,
and/or handle or adjust an insurance claim covering the insured
property using the estimated damage. The natural disaster may be a
flood or wild fire, and the damage to the insured property is water
or fire damage.
[0174] In another aspect, a computer system or mobile device
configured for natural disaster simulation may be provided. The
computer system or mobile deice may include one or more local or
remote processors configured to: (1) capture current or updated
video or images of an insured property via a mobile device camera
(and/or the one or more local or remote processors, or other smart
cameras), and GPS coordinates from a mobile device GPS unit or
nearby GPS units (such as a house smart controller); (2) access a
database of one or more historical digital or other images of the
insured property, such as one or more historical images stored
along with GPS coordinates and/or timestamp information; (3)
compare the current or updated video or images of the insured
property with the one or more historical digital or other images of
the insured property; (4) identify or determine an extent of damage
to the insured property caused by the natural disaster, or
estimated to be caused by the natural disaster; (5) identify or
determine a cost to repair the damage to the insured property
and/or an replacement cost for the insured property; and/or (6)
display on a mobile device or computing device display the
estimated extent of damage to the insured property, and/or the
estimated repair or replacement cost for an insured's review and/or
approval.
[0175] The one or more local or remote processors configured to:
prepare a virtual insurance claim (or pre-populate a virtual
insurance claim form) for the insured based upon the estimated
extent of damage to the insured property, and/or the estimated
repair or replacement cost. The one or more local or remote
processors may be configured to: adjust or modify an insurance
policy, premium, or discount based upon the estimated extent of
damage to the insured property, and/or the estimated repair or
replacement cost; generate insurance-related recommendations based
upon the estimated extent of damage to the insured property, and/or
the estimated repair or replacement cost; adjust or modify an
insurance policy, premium, or discount based upon an insured
following one or more insurance-related recommendations generated;
and/or generate a graphic based upon, and/or depicting, the
estimated extent of damage to the insured property, and/or
displaying the graphic on top of an image of the insured
property.
[0176] The computer systems and/or mobile devices discussed above
may include additional, less, or alternate functionality, including
that discussed elsewhere herein. The computer systems and/or mobile
devices may be implemented via computer-executable instructions
stored on non-transitory computer-readable media or medium.
[0177] Exemplary Computer-Implemented Methods
[0178] In one aspect, a computer-implemented method for natural
disaster simulation may be provided. The method may include (1)
capturing, via one or more local or remote processors (such as
mobile device or insurance provider remote server processors),
current, more recent, or updated video or images of an insured
property via a mobile device camera (and/or the one or more local
or remote processors, or other smart cameras), and GPS coordinates
from a mobile device GPS unit or nearby GPS units (such as a house
smart controller); (2) accessing, via the one or more local or
remote processors, a database of one or more historical digital or
other images of the insured property stored in a non-transitory
memory unit, such as one or more historical images stored along
with GPS coordinates and/or timestamp information; (3) comparing,
via the one or more local or remote processors, the current or
updated video or images of the insured property with the one or
more historical digital or other images of the insured property;
(4) identifying or determining, via the one or more local or remote
processors, an extent of damage to the insured property caused by
the natural disaster, or estimated to be caused by the natural
disaster; (5) identifying or determining, via the one or more local
or remote processors, a cost to repair the damage to the insured
property and/or an replacement cost for the insured property;
and/or (6) displaying or presenting on a mobile device or computing
device display, via the one or more local or remote processors, the
estimated extent of damage to the insured property, and/or the
estimated repair or replacement cost for an insured's review and/or
approval.
[0179] The method may include, via the one or more local or remote
processors: (i) preparing a virtual insurance claim (or
pre-populate a virtual insurance claim form) for the insured based
upon the estimated extent of damage to the insured property, and/or
the estimated repair or replacement cost; (ii) adjusting or
modifying an insurance policy, premium, or discount based upon the
estimated extent of damage to the insured property, and/or the
estimated repair or replacement cost; (iii) generating
insurance-related recommendations based upon the estimated extent
of damage to the insured property, and/or the estimated repair or
replacement cost; (iv) adjusting or modifying an insurance policy,
premium, or discount based upon an insured following one or more
insurance-related recommendations generated; and/or (v) generating
a graphic based upon, and/or depicting, the estimated extent of
damage to the insured property, and/or displaying the graphic on
top of an image of the insured property.
[0180] In another aspect, a computer-implemented method for natural
disaster simulation may be provided. The method may, via one or
more local or remote processors (such as mobile device or insurance
provider remote server processors), include (1) capturing current,
more recent, or updated video or images of an insured property via
a mobile device camera (and/or the one or more local or remote
processors, or other smart cameras), and GPS coordinates
(associated with the insured property) from a mobile device GPS
unit or nearby GPS units (such as a house smart controller); (2)
accessing a database of one or more historical digital or other
images of the insured property stored in a non-transitory memory
unit, such as one or more historical images stored along with GPS
coordinates and/or timestamp information; (3) comparing, via the
one or more local or remote processors, the current or updated
video or images of the insured property with the one or more
historical digital or other images of the insured property; (4)
identifying or determining, via the one or more local or remote
processors, an extent of damage to the insured property caused by
the natural disaster or other causes (such as age, weather, wear
and tear, usage, etc.), or estimated to be caused by the natural
disaster other causes (such as age, weather, wear and tear, usage,
etc.); (5) identifying or determining, via the one or more local or
remote processors, a cost to repair the damage to the insured
property for the insured property; and/or (6) displaying or
presenting on a mobile device or computing device display, via the
one or more local or remote processors, the estimated repair cost
for an insured's review and/or approval.
[0181] The method may include, via the one or more local or remote
processors: (1) preparing a virtual insurance claim (or
pre-populate a virtual insurance claim form) for the insured based
upon the estimated extent of damage to the insured property, and/or
the estimated repair cost; (2) adjusting or modifying an insurance
policy, premium, or discount based upon the estimated extent of
damage to the insured property, and/or the estimated repair cost;
(3) generating insurance-related recommendations based upon the
estimated extent of damage to the insured property, and/or the
estimated repair cost, such as recommendations related to cut back
brush in the case of a wild fire in the vicinity of the insured
property; (4) identifying a current type of roofing product, an
extent of damage to the current roofing product (e.g., shingle type
or manufacturer, or damaged square footage), an replacement type of
roofing product, and/or a cost of the replacement type of roofing
product; and/or (5) identifying, via the one or more local or
remote processors, a current type of house siding product, an
extent of damage to the current siding product (e.g., vinyl type or
manufacturer, or damaged square footage), an replacement type of
siding product, and/or a cost of the replacement type of siding
product.
[0182] The fore-going methods may include additional, less, or
alternate actions, including that discussed elsewhere herein. The
methods may be implemented via one or more local or remote
processors (such as mobile devices and/or insurance provider remote
servers), and/or via computer-executable instructions stored on
non-transitory computer-readable media or medium.
[0183] Exemplary Insurance Applications
[0184] The chaotic situation following a significant natural
disaster event (hurricane, wildfire, storm surge, etc.) may involve
many difficulties for property owners, local governments, insurance
companies and other interested parties. Displacement of objects,
piles of debris, inaccessible or damaged infrastructure services,
mixture of property belonging to multiple owners, and other factors
increase the time, expense and frustration of all affected parties
in an already traumatic situation. The lack of ready-at-hand
historical images may exacerbate these problems. Similar in effect,
though differing in cause, is the situation faced by those
experiencing memory loss and cognitive impairments. The ability of
those individuals to live unassisted in their own homes is
significantly reduced by the chaos of previously familiar objects
and tasks presenting themselves on a daily basis.
[0185] The present embodiments involve the use of historical image
data (video, still image, 3D mapping), object recognition
technology, location awareness technology, and personal digital
projection apparatuses (e.g. Google Glass, Meta Pro, Magic Leap,
GlassUp, Optinvent Ora-S AR, Vuzix Smart Glasses, etc.),
supplemented by augmented reality information in order to produce
visualizations that facilitate user action in chaotic or unfamiliar
situations. These technologies will be collectively referred to
below as "glasses," though the same results could be achieved
through a smartphone, tablet, or other computer that has embedded
camera and screen technology.
[0186] For claims adjusters, historical images of a pre-disaster
property (or set of properties) could be displayed to a user for
comparison purposes on their glasses. Historical images may be
drawn from existing public image data, per location (e.g., Google
Maps "street view", TomTom satellite-based 3D mapping) or from
company-owned image data related to the property. The historical
images may be supplemented with visual indicators to highlight
significant changes between the historical view and the current
view of a property. Where only a limited number of images are
available, the glasses may indicate the correct location from which
the image comparison may be made by the user. In areas with greater
image availability, the glasses could use location and directional
data to generate equivalent visualizations of a property,
regardless of viewing angle, by correlating location and
directional data from stored images. Customer contact information,
policy provisions and limits, and material depreciation (based upon
object recognition and/or standard calculations) may also be
displayed as an overlaying image on the glasses in a claims
scenario, making conversations and decisions more informed and
prompt. The same apparatus may be used to recognize, measure, and
evaluate roofing materials and damage to them.
[0187] The same apparatus could also show a virtual grid overlay of
pre-damage images of structures, highlighting the changes incurred,
and using object recognition and spatial mapping technologies to
establish the nature of the materials damaged, their approximate
size (square footage, etc.) and the approximate cost of repair or
replacement (by referring to established tables of such data--i.e.,
per square foot replacement cost for average quality vinyl siding).
These numbers may be captured and stored, being compiled together
as part of a claim disbursement estimate. The visual grid
(constructed from historical images of a property and/or
satellite-based 3D mapping) may be used to establish spatial
relationships between multiple badly-damaged structures (for
example, those that have collapsed into one another).
[0188] The same underlying apparatus and set of applications may be
used by field underwriters surveying a property to ensure proper
coverages and replacement costs. The creation dates of historical
images of a property could be compared to coverage effective dates,
to correlate property conditions to policy periods--as well as to
judge the changes in the property over time (i.e., if multiple
historical images are available). In this use, the process may be
further extended to include existing policy information (e.g.,
coverage amounts, deductibles, estimated replacement amounts), so
that any changes in the underlying property could be used as the
basis for answering specific business questions (e.g., changes in
eligibility based upon additions to a home, changes in coverage
needs and premium resulting from home improvements, changes in data
elements such as Roof Age when updates as detected). These
required/recommended policy changes may be communicated to the user
of the glasses, enabling them to immediately communicate those to
customers or other interested parties. The user of the glasses may
use the same apparatus to initiate those policy changes and/or
prepare a communication that invites the customer and/or their
agent to complete the work necessary to make those policy
changes.
[0189] Public record and non-image historical company data on a
property (on-hand statistics) may also be displayed on the glasses
and made available to field underwriting for more convenient
comparisons (with differences between recorded data and newly-seen
images being flagged or otherwise highlight--e.g., a previously
reported square footage that no longer matches the glasses-detected
square footage, because of a new addition to the property).
[0190] In combination with beacon/Internet of Things (IoT)
technology, such an apparatus may be used to facilitate recovery of
important objects obscured by debris, moved from their original
location by wind or water, or otherwise lost. The augmented reality
function of the apparatuses mentioned above may be used to visually
highlight the location of objects and provide information (name of
object, direction, distance). This information may be based upon a
correlation of beacon information with specific contents
pre-arranged as "key objects" in a household contents catalog
(e.g., car keys, fire-proof document boxes, etc.).
[0191] Private individuals could take advantage of the same
underlying technology and process to locate specific objects in or
around their property, outside of the context of a loss event. For
example, individuals suffering from impairments to memory and/or
cognition could be guided through their own home by an augmented
reality enhanced set of glasses that enables the location of lost
objects, and assistance with the use of important objects (e.g.,
prescription drugs). The glasses may visually highlight and provide
stock information about objects, based upon object recognition,
pre-set criteria and metadata related to important objects
(example: assistance in locating prescription medicine, and dosage
instructions once the object is located). The glasses may further
be employed to give process directions and reminders, in
conjunction with the augmented reality visuals (example: reminding
someone that it is time to take out the trash, and then, if
desired, giving directions for the sequential performance of the
steps needed for completing the task).
[0192] Private individuals may also be provided with simulations of
critical maintenance activities in their home or business. For
example, based upon object recognition, water valves, circuit
breakers, gas valves and other critical mechanical interfaces may
be identified, and when identified, simulations of their use could
be displayed on the smart glasses, showing the action a user could
take in the event of a mechanical failure or weather event. For
example, a simulation of locating and turning on a supplemental
generator may be provided via smart glasses when a power loss is
experienced.
[0193] The present embodiments may include (i) the combinative use
of personal digital projection apparatuses ("smart glasses") and
augmented reality technology in the context of insurance industry
operations; (ii) the correlation of historical visual (video, still
images, 3D mapping) data with current images, seen "live" through
smart glasses, smartphones, tablets, and other computer
technologies; (iii) the combination of technologies necessary for
enhanced object location services, in the context of insured
property (e.g., Claims) and independent living assistance; and/or
(iv) the use of beacon and/or Internet of Things (IoT) technology
in the context of insurance-specific processes, such as Claims
Adjustment and Underwriting Surveys.
[0194] The present embodiment may also (v) enable claims adjusters
to provide better, faster service in existing capacity, and
allowing them to fulfill new services on behalf of customers (i.e.,
assisting individual locate key objects that would be counted as
part of a total loss otherwise); (vi) reduce catastrophic claims
handling time service; (vii) reduce overall claims (i.e., from
total loss to partial loss, when key objects may be located and
recovered, or when just-in-time simulations and alerts are provided
to property owners in order to minimize the extent of a loss);
(viii) offer new loss recovery services (in catastrophic loss event
and theft situations), particularly for property such as
watercraft; (ix) offer a unique and valuable service that improves
the claims experience (e.g., for purposes of customer retention);
(x) allowing more accurate and expedited re-inspection of
properties, including allowing for analysis of differences at the
beginning of the process (i.e., when data are captured, rather than
after data are delivered into analysis systems after the fact);
(xi) reduce the time between re-inspections and coverage/premium
adjustments; (xii) enable customers to live longer in their own
homes without assistance (or minimizing such assistance); and/or
(xiii) offer a software application for smart glasses or smartphone
for sale to the public (i.e., to homeowners or realtors) that may
help them assess the condition of a home, its maintenance needs; as
well as to individuals (and caretakers) who need partial assistance
in order to maintain independent living.
[0195] Additional Insurance-Based Applications
[0196] The natural risks inherent in a given property may be
difficult to visualize and understand--both by insurance customer
and associates (whether in sales or underwriting). Understanding
risk is paramount in the business of insurance, as well as other
property-related decision making. A greater understanding of risks
posed by natural conditions, hypothetical events, and trends may
increase confidence in both situations.
[0197] Location-based systematic peril mapping and in-person (and
virtual) inspections (aka underwriting surveys) of real property
may help to give point-in-time view of the underwriting concerns
and risks of natural disasters occurring in a given location, with
relation to that property. For example, an in-person observation of
landslide activity may give an insurance carrier pause when it
comes to writing business for a particular property. Additionally,
certain peril mapping values may be avoided for whole sets of
properties in given locales (for example, those that may be likely
to experience hurricane and storm surge activity).
[0198] The present embodiments may involve the use of peril mapping
data, object recognition technology, location awareness technology,
and/or video/screen-enabled computing devices (whether "smart"
glasses, smartphones, or others), supplemented by augmented reality
information in order to produce simulations of the effects of
various natural phenomena on selected properties.
[0199] In one embodiment, existing two-dimensional peril maps may
be correlated to topographical data (e.g. elevation values) to
produce an augmented reality overlay of simulated water levels,
simulated wildfire paths, and other natural disasters. Object
recognition and location awareness may be used to correlate mapping
data to particular topography and insured objects (for example,
showing where a hypothetical flood could touch a property, or where
a wildfire under certain wind conditions may touch a property).
[0200] In the example of wildfire, object recognition may be used
to establish more accurate wildfire setback areas (as opposed to
peril map-based setbacks). Object recognition may also be used to
validate existing topographical maps (for example, following a
recent earth movement event or the growth of brush since the most
recent aerial imagery was captured), allowing for disaster
simulations based upon actual topography instead of the mapped
topography.
[0201] Simulations may be produced based upon factors such as a
user-selected disaster type, severity, and duration. The simulation
may be as simple as a visual overlay of a line representing a
projected water level, given the selected input, or as complex as
simulated property damage animations (discoloration to show water
damage, destruction of a detached structure in a fire, movement of
a building down a slope in an earthquake, etc.).
[0202] Existing insurance policy information may be used as the
basis for estimations of damage to a particular property. For
example, if a flood simulation showed water levels that would
damage a basement and a detached structure, the value (or
percentage of total value) of those structures could be displayed
to the user, relative to the estimated severity and duration of the
disaster (i.e. water levels high enough to flood a basement would
not necessarily result in a total loss, but they might result in a
total loss for a detached structure that is submerged or washed
away).
[0203] In a further enhancement, the input may be gathered in a
reverse direction, such that a risk tolerance (e.g., on the part of
the insurance customer or the insurance provider) may be
established first, and simulations could be run to determine the
level of severity, duration and extent that could occur and remain
under that risk tolerance threshold. For example, if a given
Customer decided that flooding of basements would be acceptable,
but flooding of 1st floors would not, simulations may be run to
determine the severity of weather events that could occur that
would cause the first, but not the second, type of damage.
[0204] The present embodiments may include (1) the use of augmented
reality technology to produce disaster simulations in the context
of insurance; and/or (2) the mapping of peril maps to video capture
images on smartphones and smart glasses. Potential application
and/or uses of the present embodiments may include (i) enabling
agents and/or underwriters to better understand and communicate the
natural risk inherent in a particular property (e.g., by
visualizing for a homeowner their need for increased coverage
related to a particular natural hazard); (ii) enabling better
underwriting decision making (i.e., on the spot, instead of using
abstract peril maps only); (iii) enabling consumers and realtors to
make decisions about prospective property purchases, with respect
to natural hazard risks; and/or (iv) enabling consumers to assess
their own insurance risks, including their tolerance for risk and
possible preventative mitigation efforts that could be taken (for
example, as a free application on a smartphone that could simulate
the effects of different wildfire setbacks/defensible space,
suggesting the right distance to be maintained between the house
and surrounding brush areas).
Additional Considerations
[0205] As will be appreciated based upon the foregoing
specification, the above-described embodiments of the disclosure
may be implemented using computer programming or engineering
techniques including computer software, firmware, hardware or any
combination or subset thereof. Any such resulting program, having
computer-readable code means, may be embodied or provided within
one or more computer-readable media, thereby making a computer
program product, i.e., an article of manufacture, according to the
discussed embodiments of the disclosure. The computer-readable
media may be, for example, but is not limited to, a fixed (hard)
drive, diskette, optical disk, magnetic tape, semiconductor memory
such as read-only memory (ROM), and/or any transmitting/receiving
medium such as the Internet or other communication network or link.
The article of manufacture containing the computer code may be made
and/or used by executing the code directly from one medium, by
copying the code from one medium to another medium, or by
transmitting the code over a network.
[0206] These computer programs (also known as programs, software,
software applications, "apps", or code) include machine
instructions for a programmable processor, and can be implemented
in a high-level procedural and/or object-oriented programming
language, and/or in assembly/machine language. As used herein, the
terms "machine-readable medium" "computer-readable medium" refers
to any computer program product, apparatus and/or device (e.g.,
magnetic discs, optical disks, memory, Programmable Logic Devices
(PLDs)) used to provide machine instructions and/or data to a
programmable processor, including a machine-readable medium that
receives machine instructions as a machine-readable signal. The
"machine-readable medium" and "computer-readable medium," however,
do not include transitory signals. The term "machine-readable
signal" refers to any signal used to provide machine instructions
and/or data to a programmable processor.
[0207] As used herein, a processor may include any programmable
system including systems using micro-controllers, reduced
instruction set circuits (RISC), application specific integrated
circuits (ASICs), logic circuits, and any other circuit or
processor capable of executing the functions described herein. The
above examples are example only, and are thus not intended to limit
in any way the definition and/or meaning of the term
"processor."
[0208] As used herein, the terms "software" and "firmware" are
interchangeable, and include any computer program stored in memory
for execution by a processor, including RAM memory, ROM memory,
EPROM memory, EEPROM memory, and non-volatile RAM (NVRAM) memory.
The above memory types are example only, and are thus not limiting
as to the types of memory usable for storage of a computer
program.
[0209] In one embodiment, a computer program is provided, and the
program is embodied on a computer readable medium. In an example
embodiment, the system is executed on a single computer system,
without requiring a connection to a sever computer. In a further
embodiment, the system is being run in a Windows.RTM. environment
(Windows is a registered trademark of Microsoft Corporation,
Redmond, Wash.). In yet another embodiment, the system is run on a
mainframe environment and a UNIX.RTM. server environment (UNIX is a
registered trademark of X/Open Company Limited located in Reading,
Berkshire, United Kingdom). The application is flexible and
designed to run in various different environments without
compromising any major functionality. In some embodiments, the
system includes multiple components distributed among a plurality
of computing devices. One or more components may be in the form of
computer-executable instructions embodied in a computer-readable
medium. The systems and processes are not limited to the specific
embodiments described herein. In addition, components of each
system and each process can be practiced independent and separate
from other components and processes described herein. Each
component and process can also be used in combination with other
assembly packages and processes.
[0210] As used herein, an element or step recited in the singular
and proceeded with the word "a" or "an" should be understood as not
excluding plural elements or steps, unless such exclusion is
explicitly recited. Furthermore, references to "example
embodiment," "exemplary embodiment," or "one embodiment" of the
present disclosure are not intended to be interpreted as excluding
the existence of additional embodiments that also incorporate the
recited features.
[0211] The patent claims at the end of this document are not
intended to be construed under 35 U.S.C. .sctn. 112(f) unless
traditional means-plus-function language is expressly recited, such
as "means for" or "step for" language being expressly recited in
the claim(s).
[0212] This written description uses examples to disclose the
disclosure, including the best mode, and also to enable any person
skilled in the art to practice the disclosure, including making and
using any devices or systems and performing any incorporated
methods. The patentable scope of the disclosure is defined by the
claims, and may include other examples that occur to those skilled
in the art. Such other examples are intended to be within the scope
of the claims if they have structural elements that do not differ
from the literal language of the claims, or if they include
equivalent structural elements with insubstantial differences from
the literal language of the claims.
* * * * *