U.S. patent application number 13/035321 was filed with the patent office on 2011-08-25 for firearm training systems and methods of using the same.
Invention is credited to David Alfred A. Lagettie, Paige Tedford Manard.
Application Number | 20110207089 13/035321 |
Document ID | / |
Family ID | 44476805 |
Filed Date | 2011-08-25 |
United States Patent
Application |
20110207089 |
Kind Code |
A1 |
Lagettie; David Alfred A. ;
et al. |
August 25, 2011 |
FIREARM TRAINING SYSTEMS AND METHODS OF USING THE SAME
Abstract
Firearm training systems and associated methods are provided. A
firearm training system may include a firearm, a scope, a tracking
system, and at least one controller. The scope may be mounted to
the firearm, and the scope may include a display. The tracking
system may be configured to collect information associated with a
position and an orientation of the firearm. The at least one
controller may be configured to (i) determine, based at least in
part upon the information collected by the tracking component, a
viewpoint of the firearm, (ii) determine, based at least in part
upon the viewpoint of the firearm, image data to be rendered by the
display, and (iii) direct output of the image data for receipt by
the display.
Inventors: |
Lagettie; David Alfred A.;
(Anna Bay, AU) ; Manard; Paige Tedford; (Richmond,
TX) |
Family ID: |
44476805 |
Appl. No.: |
13/035321 |
Filed: |
February 25, 2011 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
61308169 |
Feb 25, 2010 |
|
|
|
Current U.S.
Class: |
434/22 |
Current CPC
Class: |
F41G 3/2611 20130101;
F41G 3/2627 20130101 |
Class at
Publication: |
434/22 |
International
Class: |
F41G 3/26 20060101
F41G003/26 |
Claims
1. A firearm training system, comprising: a firearm; a scope
mounted to the firearm, the scope comprising a display; a tracking
system configured to collect information associated with a position
and an orientation of the firearm; and at least one controller
configured to (i) determine, based at least in part upon the
information collected by the tracking system, a viewpoint of the
firearm, (ii) determine, based at least in part upon the viewpoint
of the firearm, image data to be rendered by the display, and (iii)
direct output of the image data for receipt by the display.
2. The firearm training system of claim 1, wherein the at least one
controller is further configured to adjust, prior to determining
the image data to be rendered, the position of the firearm within a
virtual environment.
3. The firearm training system of claim 2, wherein the display
comprises a first display, and further comprising: a second display
external to the firearm, the second display configured to render a
representation of the virtual environment; wherein the at least one
controller is configured to adjust the position of the firearm to a
calibration point associated with the second display.
4. The firearm training system of claim 1, wherein the tracking
system comprises a laser tracking system comprising: a laser
emitter attached to the firearm; and at least one laser detector
configured to detect a laser emitted by the laser emitter.
5. The firearm training system of claim 1, wherein the tracking
system is configured to track motion along one, two, or three axes
with two to six degrees of freedom.
6. The firearm training system of claim 1, wherein the tracking
system comprises one of (i) an electromagnetic tracking system,
(ii) an optical tracking system, (iii) an infrared tracking system,
(iv) an ultrasonic tracking system, (v) an inertia tracking system,
(vi) a mechanical tracking system, or (vii) a global positioning
system tracking system.
7. The firearm training system of claim 1, wherein the display is
mounted at least partially within a housing of the scope.
8. The firearm training system of claim 1, wherein the display
comprises one of (i) a microdisplay, (ii) a liquid crystal display,
(iii) a light emitting diode display, or (iv) an organic light
emitting diode display.
9. The firearm training system of claim 1, further comprising: a
trigger assembly associated with the firearm, the trigger assembly
configured to detect a user actuation of a trigger, wherein the at
least one controller is further configured to (i) receive an
indication of the trigger actuation, (ii) calculate a trajectory of
a shot within a virtual environment, and (iii) direct output of
visual information associated with the shot for receipt by the
display.
10. The firearm training system of claim 1, further comprising: a
camera attached to the firearm and configured to read one or more
grid patterns associated with a surface external to the firearm,
wherein the at least one controller is further configured to (i)
receive grid pattern information from the camera and (ii) verify
the determined viewpoint of the firearm utilizing at least a
portion of the received grid pattern information.
11. A system for modifying a firearm for use in a simulation
environment, the system comprising: a display configured to be
attached to a scope associated with the firearm; a tracking
component configured to be attached to the firearm, the tracking
component further configured to collect information associated with
a position and an orientation of the firearm; and at least one
controller configured to (i) determine, based at least in part upon
the information collected by the tracking component, a viewpoint of
the firearm, (ii) determine, based at least in part upon the
viewpoint of the firearm, image data to be rendered by the display,
and (iii) direct output of the image data for receipt by the
display.
12. The system of claim 11, wherein the at least one controller is
further configured to adjust, prior to determining the image data
to be rendered, the position of the firearm within a virtual
environment.
13. The system of claim 1 I, wherein the tracking system comprises
a laser tracking system comprising: a laser emitter attached to the
firearm; and at least one laser detector configured to detect a
laser emitted by the laser emitter.
14. The system of claim 11, wherein the display comprises one of
(i) a microdisplay, (ii) a liquid crystal display, (iii) a light
emitting diode display, or (iv) an organic light emitting diode
display.
15. The system of claim 11, further comprising: a trigger detection
mechanism configured to be attached to the firearm, wherein the
trigger detection mechanism detects a user actuation of a trigger
associated with the firearm, and wherein the at least one
controller is further configured to (i) receive an indication of
the trigger actuation, (ii) calculate a trajectory of a shot within
a virtual environment, and (iii) direct output of visual
information associated with the shot for receipt by the
display.
16. A firearm scope, comprising: a housing configured to be mounted
onto a firearm; a display mounted at least partially within the
housing; and a video receiver in communication with the display and
configured to output video to be rendered by the display, wherein
the video receiver receives the video from at least one controller
configured to (i) determine a viewpoint of the firearm, (ii) adjust
a position of the firearm within a virtual environment, (iii)
determine the video to be rendered, and (iv) direct communication
of the video to the video receiver.
17. The firearm scope of claim 16, further comprising: a laser
tracking system comprising a laser emitter and a laser detector
configured to detect a laser emitted by the laser emitter, wherein
the at least one controller determines the viewpoint of the firearm
based at least in part upon tracking information received from the
laser tracking system.
18. A method for providing a firearm simulation, the method
comprising: tracking, via a tracking component attached to a
firearm, motion of the firearm; determining, by a controller
comprising one or more computers and based upon information
collected by the tracking component, a viewpoint of the firearm;
determining, by the controller based at least in part upon the
determined viewpoint, image data to be rendered by a display
associated with a scope mounted onto the firearm; outputting, by
the controller for receipt by the display, the determined image
data; and presenting the image data by the display.
19. The method of claim 18, further comprising: adjusting, by the
controller and prior to determining the image data to be rendered,
a position of the firearm within a virtual environment.
20. The method of claim 18, wherein tracking motion of the firearm
comprises tracking motion utilizing a laser tracking component.
Description
CROSS-REFERENCE TO RELATED APPLICATION
[0001] The present application claims the benefit of U.S.
Provisional Application Ser. No. 61/308,169, filed Feb. 25, 2010,
and entitled "Firearm Scope and System for Simulated Environments,"
the contents of which are incorporated by reference herein in their
entirety.
FIELD OF THE INVENTION
[0002] Aspects of the invention relate generally to firearm
training systems and, more particularly, to firearm training
systems that utilize a simulated virtual environment.
BACKGROUND OF THE INVENTION
[0003] Training for marksmen and snipers, including military and
law enforcement personnel, is often critical for the development of
targeting and accuracy skills. However, live training often
requires relatively expensive ammunition and the maintenance of
shooting range facilities. Additionally, live training typically
fails to provide a variety of targets, including moving targets. In
order to provide additional avenues for firearms training, various
shooting simulators have been developed. These conventional
shooting simulators typically present images via projection screens
or other image displays. Simulated projectiles or lasers are then
typically identified as shots when contact is made with the display
surface.
[0004] However, due to the utilization of two-dimensional surfaces
for shot recognition and capture, accuracy is typically compromised
in these conventional shooting simulators. These conventional
shooting simulators also fail to accurately represent ballistic
physics during the capture of shooting events. Additionally, these
conventional shooting simulators fail to effectively utilize scopes
to represent digital imagery within a virtual environment. For
example, if a user were to look through a scope at an image display
utilized in a conventional shooting simulator, the user will be
presented with a pixilated magnified view of the image display.
Accordingly, improved firearm training systems and associated
methods are desirable.
BRIEF DESCRIPTION OF THE INVENTION
[0005] Some or all of the above needs and/or problems may be
addressed by certain embodiments of the invention. Embodiments of
the invention may include firearm training systems and methods of
using the same. In one embodiment, a firearm training system may be
provided. The firearm training system may include a firearm, a
scope, a tracking system, and at least one controller. The scope
may be mounted to the firearm, and the scope may include a display.
The tracking system may be configured to collect information
associated with a position and an orientation of the firearm. The
at least one controller may be configured to (i) determine, based
at least in part upon the information collected by the tracking
component, a viewpoint of the firearm, (ii) determine, based at
least in part upon the viewpoint of the firearm, image data to be
rendered by the display, and (iii) direct output of the image data
for receipt by the display.
[0006] In accordance with another embodiment of the invention, a
system for modifying a firearm for use in a simulation environment
may be provided. The system may include a display, a tracking
component, and at least one controller. The display may be
configured to be attached to a scope associated with the firearm.
The tracking component may be configured to be attached to the
firearm, and the tracking component may be further configured to
collect information associated with a position and an orientation
of the firearm. The at least one controller may be configured to
(i) determine, based at least in part upon the information
collected by the tracking system, a viewpoint of the firearm, (ii)
determine, based at least in part upon the viewpoint of the
firearm, image data to be rendered by the display, and (iii) direct
output of the image data for receipt by the display.
[0007] In accordance with yet another embodiment of the invention,
a firearm scope may be provided. The firearm scope may include a
housing, a display, and a video receiver. The housing may be
configured to be mounted onto a firearm. The display may be mounted
at least partially within the housing. The video receiver may be in
communication with the display and configured to output video to be
rendered by the display. The video receiver may receive the video
from at least one controller configured to (i) determine a
viewpoint of the firearm, (ii) adjust a position of the firearm
within a virtual environment, (iii) determine the video to be
rendered, and (iv) direct communication of the video to the video
receiver.
[0008] In accordance with yet another embodiment of the invention,
a method for providing a firearm simulation may be provided. Motion
of a firearm may be tracked via a tracking component attached to
the firearm. Based upon information collected by the tracking
component, a viewpoint of the firearm may be determined by a
controller, such as a controller including one or more computers.
Based at least in part upon the determined viewpoint, the
controller may determine or identify image data to be rendered by a
display associated with a scope mounted onto the firearm. The
determined image data may be output by the controller for receipt
by the display, and the image data may be presented by the
display.
[0009] Additional systems, methods, apparatus, features, and
aspects may be realized through the techniques of various
embodiments of the invention. Other embodiments and aspects of the
invention are described in detail herein with reference to the
description and to the drawings and are considered a part of the
claimed invention.
BRIEF DESCRIPTION OF THE DRAWINGS
[0010] Reference will now be made to the accompanying drawings,
which are not necessarily drawn to scale, and wherein:
[0011] FIG. 1 illustrates a block diagram of an example firearm
training system that may be utilized in accordance with various
embodiments of the invention.
[0012] FIG. 2A is an illustration of an actual firearm that has
been modified to include attached simulation components.
[0013] FIG. 2B is an exploded illustration of the firearm shown in
FIG. 2A.
[0014] FIG. 3 illustrates a block diagram of an example scope that
may be mounted on a firearm in accordance with various embodiments
of the invention.
[0015] FIG. 4 is an illustration of an actual firearm onto which a
replica scope has been mounted to facilitate use of the firearm
within a simulation environment.
[0016] FIG. 5 is an illustration of a firearm into which simulation
components have been integrated.
[0017] FIG. 6 is a diagram of a simulation environment that may be
utilized in accordance with various embodiments of the
invention.
[0018] FIG. 7 is a flow diagram of an example method for presenting
image data via a display associated with a scope mounted onto a
firearm, according to an example embodiment of the invention.
[0019] FIG. 8 is a flow diagram of an example method for
determining image data to be presented via a display associated
with a scope mounted onto a firearm, according to an example
embodiment of the invention.
DETAILED DESCRIPTION
[0020] Embodiments of the invention now will be described more
fully hereinafter with reference to the accompanying drawings, in
which embodiments of the invention are shown. This invention may,
however, be embodied in many different forms and should not be
construed as limited to the embodiments set forth herein; rather,
these embodiments are provided so that this disclosure will be
thorough and complete, and will fully convey the scope of the
invention to those skilled in the art. Like numbers refer to like
elements throughout.
[0021] Embodiments of the invention may include firearm training
systems, apparatus, and methods for using the same. In certain
embodiments, a scope may be mounted to a firearm, and a display may
be associated with the scope. For example, the display may be
integrated into, incorporated into, or at least partially situated
within the scope. As another example, the display may be attached
to the scope. A wide variety of suitable displays may be utilized
as desired in various embodiments of the invention, such as a
microdisplay, a liquid crystal display ("LCD"), a light emitting
diode ("LED") display, an organic light emitting diode ("OLED")
display, or other suitable display. The display may facilitate the
presentation of image data, such as video, to a user of the
firearm. For example, in a virtual environment, one or more
displays external to the firearm (e.g., projection screens, etc.)
may be utilized to present a virtual environment to a user. When a
user looks through the scope, an enlarged representation of a
portion of the virtual environment may be presented to the user via
the display.
[0022] Additionally, at least one tracking system may be associated
with the firearm. The tracking system may be configured to track
motion, position, and/or orientation data associated with the
firearm. A wide variety of suitable tracking systems may be
utilized as desired in various embodiments of the invention,
including but not limited to, a laser tracking system, an
electromagnetic tracking system, a magnetic tracking system, an
optical tracking system, an infrared tracking system, an ultrasonic
tracking system, an inertia tracking system, a mechanical tracking
system, a Global Positioning System ("GPS") tracking system, or a
combination of tracking systems. As desired in various embodiments,
motion and/or movement of the firearm and/or scope may be tracked
along any number of axes and/or in a wide variety of different
directions. For example, movement may be tracked along three axes
(e.g., the x, y, and z axes) with anywhere from two to six degrees
of freedom.
[0023] Information collected by the tracking system may be utilized
to determine a viewpoint associated with the firearm. For example,
information collected by the tracking system may be provided to one
or more controllers configured to process the information and
calculate a viewpoint for the firearm and/or the firearm scope. In
this regard, image data to be presented via the display may be
determined. As desired in various embodiments of the invention, a
position of the firearm may be adjusted or corrected prior to
determining the image data to be presented. For example, an
external display (e.g., one or more projection screens, etc.) may
be configured to present a virtual environment from the perspective
of a single calibration point. Prior to determining a viewpoint of
the firearm, the position of the firearm may be corrected to the
calibration point. In this regard, the image data presented via the
display may correspond to that presented to the user via the
external display, regardless of the user's movement within a
simulation environment. This position correction may assist in
facilitating a simulation environment in which multiple firearms
are utilized.
[0024] A trigger mechanism or trigger actuation detection mechanism
may also be associated with the firearm. For example, a trigger
mechanism that facilitates the detection of a trigger pull event
may be incorporated into a firearm. As another example, a trigger
detection mechanism, such as a false or inert round of ammunition
that detects actuation of a tiring pin, may be inserted into a
firearm. Based upon the detection of a trigger actuation, a firing
event may be identified, and an indication of the firing event may
be communicated to one or more controllers. The controllers may
utilize a wide variety of information associated with the firearm,
scope, simulated ammunition, and/or simulated conditions (e.g.,
windage, elevation, etc.), as well as one or more ballistic models,
to determine a ballistic path for the shot within the virtual
environment. As a result of calculating a shot trajectory and/or a
ballistic path, the accuracy of the simulated shot may be enhanced.
Following the determination of information associated with a
simulated shot, the image data output for presentation via the
scope display may be updated.
[0025] System Overview
[0026] An example system 100 for facilitating firearms training
will now be described illustratively with respect to FIG. 1. The
system 100 may include, for example, a firearm 105, a scope 110
mounted onto the firearm 105, a display 115 associated with the
scope 110, one or more motion and/or position tracking systems 120,
and/or one or more controllers 125. In various embodiments, the
motion tracking systems 120 may be configured to track or monitor
motion, position, and/or orientation information associated with
the firearm. Information collected by the motion tracking systems
120 may be utilized to determine a viewpoint associated with the
firearm 105, and the determined viewpoint may be utilized to
determine and/or generate image data (e.g., video data) to be
rendered via the display 115.
[0027] With reference to FIG. 1, the firearm 105 may include any
suitable firearm, modified firearm, or replica firearm suitable for
use in a shooting or training simulation. For example, in certain
embodiments, the firearm 105 may be a replica firearm (e.g., a
replica or model rifle, pistol, shotgun, etc.) that has been
modified to include one or more simulation components (e.g., a
scope-mounted display, a scope-integrated display, one or more
motion tracking systems, etc.). In other embodiments, the firearm
105 may be an actual firearm (e.g., rifle, pistol, shotgun, etc.)
that has been modified to include or be otherwise associated with
one or more simulation components. As a result of utilizing an
actual firearm, a user may experience authentic trigger pressure,
weapon weight, and/or weapon adjustments within a simulation
environment. As desired, components may be incorporated into the
actual firearm 105 and/or removably attached to the firearm 105.
For example, simulation components and/or equipment may be
integrated into a firearm to form a simulation firearm. As another
example, a portion of the simulation components may be integrated
into a firearm while other simulation components are attached to
the firearm. As yet another example, simulation components and/or
equipment may be attached to the outside of an actual weapon. For
example, a user may attach simulation components to a personal
firearm in order to utilize the personal firearm in a simulation
environment. Indeed, a wide variety of firearm configurations may
be utilized in accordance with various embodiments of the
invention.
[0028] Additionally, in certain embodiments, a plurality of
firearms may be utilized in association with a training system 100.
Each of the firearms may include or be associated with respective
simulation components (e.g., a display, one or more motion tracking
systems, etc.). In the event that the system 100 includes one or
more display devices situated external to the firearms, such as the
one or more projection screens 130 illustrated in FIG. 1, the
respective users associated with the firearms may be presented the
same simulation or scenario by the external display devices.
However, the individual displays associated with the respective
firearm scopes may present the users with individual image
data.
[0029] The scope 110 may be any suitable scope device, modified
scope device, or replica scope device suitable for use in a
shooting or training simulation. For example, in certain
embodiments, the scope 110 may be a replica scope that has been
modified to include or otherwise be associated with a display 115
and/or other simulation components. In other embodiments, the scope
11.0 may be an actual scope (i.e., a telescopic sight) that has
been modified to include or otherwise be associated with a display
and/or other simulation components. According to an aspect of the
invention, a display 115 may be associated with the scope 110. In
certain embodiments, the display 115 may be at least partially
mounted within a housing of the scope 110. For example, a display
115 may be integrated into a scope 110 utilized in association with
a modified simulation firearm. In other embodiments, such as
embodiments in which a user utilizes a personal firearm, the
display 115 may be removably attached to the scope. In yet other
embodiments, a display 11.5 may be at least partially mounted
within the housing of a scope 110 that is attached to an actual
firearm when the firearm is utilized in conjunction with the
simulation system 100.
[0030] The display 115 may be any suitable display device
configured to present image data, such as video data, to a user of
the firearm 105. Examples of suitable display devices include, but
are not limited to, microdisplays, liquid crystal displays
("LCD's"), light emitting diode ("LED") displays, organic light
emitting diode ("OLED") displays, active matrix OLED ("AMOLED")
displays, pico front or rear projection displays, stereoscopic
displays, holographic displays, and/or cathode ray tube ("CRT")
displays. In certain embodiments, the display 115 may be configured
to present high-definition imagery. Additionally, in certain
embodiments, the display 115 may be at least partially mounted
within the housing of the scope 110. In other embodiments, the
display 115 may be removably attached to the scope 110. For
example, the display may be attached to either end of the scope
110.
[0031] In certain embodiments, a video receiver 135 may be provided
to route image and video signals to the display 115 for
presentation to a user. For example, a video receiver 135 may
receive image data from the one or more controllers 125, and the
video receiver 135 may process and/or format the received image
data. The processed and/or formatted image data may then be
provided to the display 115. The video receiver 135 may be either a
wired or wireless video receiver. For example, the video receiver
135 may be connected to the one or more controllers 125 via a video
graphics array ("VGA") cable, an Ethernet cable, a Universal Serial
Bus ("USB") cable, a mini-USB cable, a Firewire cable, a digital
video interface ("DVI") cable, a high definition multimedia
interface ("HDMI") cable, or other suitable cable or wired
connection. As another example, the video receiver 135 may be
connected to the one or more controllers 125 via a suitable
Bluetooth connection, wireless Ethernet connection, Wi-Fi
connection, or other wireless connection. In certain embodiments,
one or more suitable communications interfaces (e.g., wired
interfaces and/or wireless interfaces) 140 may facilitate
communications between the video receiver 135 and the one or more
controllers 125. Additionally, in certain embodiments, the video
receiver 135 may be incorporated and/or integrated into the firearm
105 or the scope 110. In other embodiments, the video receiver 135
may be removably attached to the firearm 105.
[0032] As desired, an audio video ("AV") receiver may be provided,
and the AV receiver may process both audio and video data. The
video data may be provided to the display 110, and the audio data
may be provided to one or more suitable audio presentation devices
(e.g., speakers, headphones, etc.). In certain embodiments, the AV
receiver may receive separate audio and video signals. For example,
separate audio and video signals may be received from the one or
more controllers 125. As another example, a video signal may be
received from the one or more controllers 125, and a separate audio
signal may be received from one or more other controllers
associated with the generation and presentation of a primary or
main simulation scenario, such as a scenario presented via the one
or more projection screens 130.
[0033] As desired, any number of suitable motion tracking systems
120 may be utilized in various embodiments of the invention. A
motion tracking system 120 may be configured to monitor and/or
track the motion, position, and/or orientation of the firearm 105
and/or the scope 110. In this regard, a viewpoint or perspective of
the firearm 105 and/or the scope 110 may be determined, and image
data may be determined and/or generated for presentation by the
display 115. Examples of suitable motion tracking systems 120 that
may be utilized include, but are not limited to, a laser tracking
system, an electromagnetic tracking system, an optical tracking
system, an infrared tracking system, an ultrasonic tracking system,
an inertia tracking system, a mechanical tracking system, and/or a
Global Positioning System ("GPS") tracking system. In certain
embodiments, a single motion tracking system may be utilized. In
other embodiments, a plurality or combination of motion tracking
systems may be utilized. In this regard, positional and/or
orientation information for the firearm 105 and/or the scope 110
may be verified and/or accuracy may be improved.
[0034] A laser tracking system may include a laser emitter that is
attached to or incorporated into the firearm 105 or the scope 110.
For example, a laser emitter may be inserted into the barrel of the
firearm 105. One or more laser detection devices, such as a laser
camera or other laser detector, may also be attached to the firearm
105 and/or at one or more other positions suitable for viewing an
area in which the laser emitter is pointed. For example, one or
more laser detection devices may be utilized to identify an emitted
laser that contacts and/or that is reflected off of the one or more
projection screens 130. Based upon the capture of the emitted
laser, positional and/or orientation data for the firearm 105 may
be determined or calculated.
[0035] An electromagnetic tracking system may include one or more
electromagnetic sources attached to the firearm 105 and one or more
sensors utilized to monitor a calibrated area for position and
motion of the electromagnetic sources. In this regard, positional
and/or orientation data for the firearm 105 may be determined or
calculated. In an optical tracking system, a video device may be
calibrated to one or more unique optical markers attached to the
firearm 105 and/or to one or more visual identifiers inherent to
the firearm 105. Motion of the firearm 105 may be monitored by the
video device, and information associated with the monitored motion
may be translated into positional and/or orientation data for the
firearm 105. With an infrared tracking system, a series of LEDs
and/or reflective markers may be attached to the firearm 105. The
LEDs may be utilized to emit infrared signals that are reflected by
the markers. One or more infrared cameras may be utilized to
monitor a field of view encompassing the firearm, such as a
simulation area. As the firearm 105 is moved, the infrared camera
may identify movement of the LEDs and/or reflective markers, and
information associated with the identified movement may be
translated into positional and/or orientation data for the firearm
105. In certain embodiments, movement of the LEDs and/or reflective
markers may be identified relative to one another in order to
provide a perspective for the firearm 105.
[0036] An ultrasonic tracking system may include one or more
ultrasonic transmitters that are attached to the firearm 105. One
or more audio sensors situated in predetermined positions may be
utilized to measure the time-of-flight and/or the phase of the
ultrasonic sound emitted by the ultrasonic transmitters. The
measurements data may then be processed in order to determine
distances between the sensors and the firearm, and the distances
may be utilized to determine positional and/or orientation data for
the firearm 105. With an inertia tracking system, one or more
inertia sensors including gyro meters and/or accelerometers may be
attached to the firearm 105. Motion or movement of the firearm 105
may be measured as inertia and translated into positional and/or
orientation data for the firearm 105. A mechanical tracking system
may include one or more mechanical devices and/or mechanical
sensors that are attached to the firearm 105 such that movement of
the firearm 105 directly affects the mechanical devices. Movement
of the gun may be translated into movement of portions of the
mechanical devices and may be translated into positional and/or
orientation data. With a GPS tracking system, one or more GPS
receivers may be attached to the firearm 105. A GPS receiver may
precisely measure the timing of signals output by orbital GPS
satellites. As the firearm 105 moves, the GAS receiver continually
calculates positional information in real-time and tracks movement
of the firearm 105. The movement data may then be translated into
positional and/or orientation information for the firearm 105.
[0037] According to an aspect of the invention, movement and/or
motion of the firearm 105 may be tracked along any number of axes
utilizing two to six degrees of freedom. For example, motion may be
tracked along an x, y, and/or z axis. Additionally, two, three,
four, five, or six degrees of freedom may be utilized to track or
monitor motion of the firearm 105. In this regard, motion may be
selectively tracked as the firearm is moved forward and backward
(e.g., closer to or further away from a projection screen), left or
right, and/or up or down. Additionally, yaw, pitch, and/or roll of
the firearm 105 may be tracked. The tracking of the firearm may be
utilized to determine a viewpoint and/or perspective of the firearm
in order to determine image data for presentation via the display
115.
[0038] As desired in certain embodiments, a wide variety of other
components may be incorporated into, attached to, and/or utilized
in conjunction with the firearm 105. For example, one or more
communications interfaces 140 may be incorporated into and/or
attached to the firearm 105 or the scope 110. The communications
interfaces may include wired and/or wireless communications devices
(e.g., transceivers, communications cards, etc.) that facilitate
communication with the controllers 125. In this regard, data and/or
information collected by simulation components attached to the
firearm 105 (e.g., the motion tracking systems 120, a camera, a
trigger recognition device, a keypad, etc.) may be communicated to
the controllers 125. Additionally, data output by the controllers
125 (e.g., image data, calibration data etc.) may be received and
distributed to the various simulation components.
[0039] In certain embodiments, a camera 145 may be integrated into
or incorporated into the firearm 105 or the scope 110. The camera
145 may be utilized to detect and/or monitor grid patterns
projected onto or output by an external display surface, such as
the one or more projection screens 130, The data captured by the
camera 145 may be processed in order to determine an orientation
and/or perspective of the firearm 105. In this regard, the motion
tracking of the firearm 105 may be verified. Additionally, a
determination may be made as to whether the firearm 105 is
positioned correctly.
[0040] Additionally, one or more trigger mechanisms 150 and/or
trigger detection devices may be incorporated into, inserted into,
and/or attached to the firearm 105. These mechanisms and/or devices
may facilitate the detection of a trigger pull or a trigger
actuation. Once a trigger actuation has been detected, an
indication of the detected actuation may be communicated to the
controllers 125, and a shot event may be registered. A wide variety
of trigger mechanisms 150 and/or trigger detection devices may be
utilized as desired in various embodiments of the invention. For
example, a firearm 105 may be modified to include a trigger
mechanism that includes one or more actuation sensors. As another
example, an inert or dummy round may be inserted into a firearm
105, and the dummy round may include one or more suitable
electrical components that detect actuation of the firing pin. An
inert round is explained in greater detail below with reference to
FIGS. 2A and 2B. The trigger mechanisms and/or trigger detection
devices described above are provided by way of example only, and
other suitable devices may be utilized as desired in various
embodiments of the invention.
[0041] A wide variety of other sensors may be incorporated into
and/or attached to the firearm 105 as desired in various
embodiments. For example, one or more sensors may be attached to
the scope 110 in order to detect user adjustment made to the scope,
such as windage adjustments, elevation adjustments, focus
adjustments, zoom adjustments, and/or mode adjustments (e.g., a
standard mode, a thermal imaging mode, a night vision mode, etc.).
As another example, one or more sensors may be utilized to monitor
the removal and/or insertion of magazines. As desired, information
collected by the sensors may be communicated to the controllers 125
and utilized during the presentation of audio and/or image data
during a simulation.
[0042] With continued reference to FIG. 1, any number of
controllers 125 may be provided. In certain embodiments, a
controller 125 may be situated or positioned external to the
firearm 105 and/or the scope 110. In other embodiments, a
controller 125 may be incorporated into and/or attached to the
firearm 105 and/or the scope 110. Each controller 125 may be a
suitable processor-driven device that facilitates the provision of
a firearms simulation. For example, a controller 125 may be a
suitable processor-driven device that receives motion information
associated with the firearm 105, determines a viewpoint of the
firearm 105, generates or determines image data for presentation
via the display 115, and/or outputs the image data for presentation
via the display 115. As such, the controller 125 may include any
number of computing devices, such as a personal computer, a digital
assistant, a personal digital assistant, a digital tablet, an
Internet appliance, an application-specific circuit, a
microcontroller, a minicomputer, or any other processor-based
device. The execution of suitable computer-implemented instructions
or computer-executable instructions by the controller 125 may form
a special purpose computer or other particular machine that is
operable to facilitate the provision of a firearms simulation.
[0043] With reference to FIG. 1, a controller 125 may include one
or more processors 152, one or more memory devices 154, one or more
network or communications interfaces 156, and/or one or more
input/output ("I/O") interfaces 158. The processors 152 may be
configured to execute any number of software applications and/or
computer-readable or computer-executable instructions. The memory
devices 154 may include any number of suitable memory devices, such
as caches, read-only memory devices, random access memory devices,
flash memory devices, magnetic storage devices, removable storage
devices (e.g., memory cards, memory sticks, etc.), etc. The memory
devices 154 may include internal memory devices and/or external
memory devices in communication with the controller 1.25. The
memory devices 154 may store data, executable instructions, and/or
various program modules utilized by the processors 152, such as
data files 160, an operating system ("OS") 162, a tracking module
164, a video module 166, and/or a ballistics module 168.
[0044] The data files 160 may include any suitable data that
facilitates the operation of the controller 125 and/or the
provision of a firearms simulation. For example, the data files 160
may include, but are not limited to, received motion and/or
movement data for one or more firearms 105, calibration information
for a simulation environment, information associated with types of
rounds or ammunition to be simulated, information associated with
desired scope settings, information associated with desired
simulation conditions (e.g., wind, etc.), one or more ballistic
models, stored image and/or audio data, and/or information utilized
to generate and/or format image and/or audio data for presentation
during a simulation.
[0045] The OS 162 may be a suitable software module that controls
the general operation of the controller 125 and/or the execution of
other software modules, for example, the tracking module 164, the
video module 166, and/or the ballistics module 168. The tracking
module 164 may be a suitable software module or application that
facilitates the determination of a viewpoint of the firearm 105.
The viewpoint of the firearm 105 may then be utilized to determine
video to be presented via the display 115. As desired, the
viewpoint of the firearm 105 may take a wide variety of position,
orientation, and/or perspective information into account, as well
as various settings associated with the scope 110 (e.g., a
magnification setting, an operational mode, etc.).
[0046] In operation, the tracking module 164 may receive
measurements and/or tracking data from the one or more motion
tracking systems 120, and the tracking module 164 may utilize at
least a portion of the received data to determine a viewpoint of
the firearm 105. For example, received measurements data may be
translated into positional and/or orientation information for the
firearm 105. The positional and/or orientation information may then
be utilized to determine a viewpoint for the firearm 105. For
example, an area of the one or more projection screens 130 at which
the firearm 105 is aimed or pointed may be determined. In certain
embodiments, various tilts and/or angles of the firearm (e.g.,
pitch, yaw, roll) may be taken into consideration when determining
the viewpoint. Additionally, in certain embodiments, a distance
between the firearm 105 and a relevant projection screen 130 may be
taken into consideration when determining the viewpoint.
[0047] In certain embodiments, the tracking module 164 may receive
tracking data from a plurality of motion tracking systems 120, and
the tracking module 164 may process the respective data to verify
the positional and/or orientation information for the firearm 105.
As desired, positional and/or orientation information may be
averaged or otherwise processed in the event that tracking data is
received from a plurality of sources. Additionally, in certain
embodiments, grid pattern information may be received from the
camera 145 and utilized to verify a viewpoint of the firearm
105.
[0048] In certain embodiments, such as embodiments in which a
plurality of firearms are utilized in conjunction with a
simulation, the tracking module 164 may correct or adjust
respective positions of the one or more firearms. In this regard,
image data that is determined or generated for display via the
firearm scopes may correspond to the perspective illustrated via
the external projection screens 130, which is the same for all
users within the simulation environment. In other words, prior to
determining a viewpoint of a firearm 105, the position of the
firearm 105 may be corrected to a calibration point for the
simulation environment. In this regard, the image data presented
via the display 115 may correspond to that presented to the user
via the projection screens 130, regardless of the user's movement
within a simulation environment. In other embodiments, such as
embodiments with a single user or embodiments in which each user
has a personal external display (e.g., a personal monitor, a
headset, etc.), a simulation scenario illustrated on the one or
more external displays may be modified as the user moves within a
simulation environment.
[0049] The video module 166 may be a suitable software module or
application configured to determine and/or generate image data for
presentation via the display 115. For example, the video module 166
may receive viewpoint information for the firearm 105, and the
video module 166 may utilize the viewpoint information to generate
image data for communication to the video receiver 135 and/or the
display 115. As desired, the video module 166 may additionally
utilize one or more settings associated with the scope (e.g., a
magnification, a display mode, etc.) and/or an indication of a shot
event to determine or generate image data for presentation. A wide
variety of image data may be generated by the video module 166,
including simulation scenario imagery, enhanced or enlarged imagery
corresponding to a portion or section of the imagery displayed on
the projection screens 130, imagery associated with a simulated
view through the scope 110 (e.g., a reticle, etc.), and/or imagery
associated with a simulated shot. In certain embodiments, the video
module 166 may also include an audio component configured to
determine or generate audio for presentation to a user of the
firearm 105.
[0050] The ballistics module 168 may be a suitable software module
or application configured to calculate or determine a simulated
trajectory of a shot within a virtual environment. Once a shot
event has been identified, information associated with the
viewpoint and/or orientation of the firearm 105 may be provided to
the ballistics module 168, and the ballistics module 168 may
calculate a trajectory for the shot. As desired, the ballistics
module 168 may utilize a wide variety of other information to
calculate a trajectory for the shot. For example, information
associated with the firearm 105 (e.g., caliber information, etc.),
information associated with a simulated ammunition, elevation
information and/or wind information may be provided to any number
of suitable ballistic models to calculate or determine a trajectory
for the simulated shot. Results of the simulated shot may then be
provided to the video module 166 to facilitate the generation of
image data associated with the simulated shot. As a result of
calculating shots by the ballistics module 168 within a virtual
environment, shot accuracy may be improved or enhanced. In
conventional training systems, such as systems that detect a shot
by detecting a laser projected onto a two-dimensional surface, the
size of the laser point or other shot may be bigger than the target
aimed at, thereby reducing accuracy and ballistic fidelity. By
simulating shots directly within a virtual environment, external
shot detection mechanisms may be bypassed, and overall accuracy may
be improved.
[0051] A wide variety of suitable operations may be performed by
the simulation components associated with a firearm 105 and/or by
the one or more controllers 125 to facilitate provision of a
training simulation. The various operations described above for a
controller 125 are provided by way of example only. Another example
of the operations that may be performed by the controllers 125 is
described in greater detail below with reference to FIG. 8.
Additionally, an example of the operations that may be performed by
the simulation components associated with the firearm 105 is
described in greater detail below with reference to FIG. 7.
[0052] With continued reference to the controller 125 of FIG. 1,
the one or more I/O interfaces 158 may facilitate communication
between the controller 125 and one or more input/output devices,
for example, one or more user interface devices, such as a display,
keypad, mouse, pointing device, control panel, touch screen
display, remote control, microphone, speaker, etc., that facilitate
user interaction with the controller 125. In this regard, user
commands may be received by the controller 125, and a wide variety
of data may be output for presentation to a user. The one or more
network interfaces 156 may facilitate connection of the controller
125 to one or more suitable wired and/or wireless networks 145, for
example, a local area network, a Wi-Fi enabled network, a
Bluetooth-enabled network, a radio frequency network, the Internet,
etc. Communications between various components of the system 100
may be facilitated via the networks 145.
[0053] With continued reference to FIG. 1, one or more projection
screens 130 and/or other external display devices may be utilized
in conjunction with the training system 100. For example, one or
more projection screens 130 may be utilized to present a simulation
scenario (e.g., a military training exercise, a law enforcement
training exercise, a hunting scenario, etc.) to any number of users
associated with respective firearms. A wide variety of projection
screen arrangements may be utilized in various embodiments of the
invention. For example, a plurality of screens may be placed next
to each other to enlarge a simulation viewing area. As another
example, the projections screens may positioned in a curve to
create up to a 360 degree simulation viewing area. Indeed,
projection screens may even be positioned in a half dome or full
dome configuration. As desired, one or more projectors 170 or other
image generation devices may be configured to output images and/or
video presented via the projection screens 130.
[0054] Additionally, in certain embodiments, one or more simulation
controllers 175 may be provided. A simulation controller 175 may be
a suitable processor-driven device configured to control the
presentation of a simulation via the one or more projection screens
130. As desired, a simulation controller 175 may include components
similar to those described above for the controller 125. For
example, a simulation controller 175 may include one or more
processors and/or memory devices, and simulation software may be
executed by the processors to facilitate the generation of image
data to be. displayed via the projection screens 130. In certain
embodiments, a simulation controller 175 may be separate from one
or more controllers 125 associated with individual firearms. In
other embodiments, the simulation controller 175 may be combined
with one or more other controllers 125.
[0055] Those of ordinary skill in the art will appreciate that the
system 100 shown in and described with respect to FIG. 1 is
provided by way of example only. Numerous other operating
environments, system architectures, and device configurations are
possible. Other system embodiments can include fewer or greater
numbers of components and may incorporate some or all of the
functionality described with respect to the system components shown
in FIG. 1.
[0056] As set forth above, a wide variety of suitable techniques
may be utilized as desired to modify firearms for use in a
simulation environment. FIGS. 2-5 illustrate a few example modified
firearms that may be utilized. FIG. 2A is an illustration of an
actual firearm 200 that has been modified to include attached
simulation components. The firearm 200 may be any suitable firearm
onto which a scope 205 has been mounted. With reference to FIG. 2A,
the scope 205 may be an actual scope mounted to the firearm 200. A
scope attachment 210 may be connected (e.g., removably connected)
to one end of the scope 205 to facilitate the presentation of image
data to a user. The scope attachment 210 may include a suitable
visual display unit configured to present image data. A visual
display unit may include, for example, a lens, a digital display
micro-screen or other suitable display, a control board, and/or a
communications interface.
[0057] With continued reference to FIG. 2A, one or more tracking
sensors 215 may be attached to the firearm 200 to facilitate motion
tracking of the firearm 200. Additionally, an inert round 220 may
be inserted into the firearm 200 to facilitate the detection of a
trigger pull event. FIG. 2B is an exploded illustration of the
firearm 200 shown in FIG. 2A. Also shown in FIG. 2B is the interior
225 of the inert round 220. The interior 225 may be a
spring-mounted metal bolt. A circuit 230, such as an approximately
five (5) volt circuit, may be connected to the inert round 220 and
grounded to the firearm 200. A trigger pull even may lead to the
inert round being contacted by the firing pin of the firearm 200,
thereby causing the circuit 230 to detect a shot being fired.
[0058] Although FIGS. 2A-2B illustrate an actual firearm, replica
or model firearms may be modified in a similar manner for use in a
simulation system.
[0059] FIG. 3 illustrates a block diagram of an example scope 300
that may be mounted on a firearm in accordance with various
embodiments of the invention. The scope 300 may include any number
of incorporated, integrated, and/or attached simulation components.
For example, the scope 300 may include a suitable display device
305, a video receiver 310, and/or one or more communications
interfaces 315. Each of these components may be similar to the
components described in greater detail above with reference to FIG.
1. These components may facilitate the receipt of image data output
by a controller, and the presentation of the image data to a
user.
[0060] As desired, the communications interfaces 315 associated
with the scope 300 may also be utilized to facilitate communication
of measurements data to one or more controllers. For example, the
communications interfaces 315 may be in communication with a
trigger mechanism 320, a motion tracking system 325, and/or a
camera 330. In this regard, shot indication data, motion data,
and/or grid pattern data may be communicated to one or more
controllers for processing. Although FIG. 3 illustrates a motion
tracking system 325 and camera 330 that have been integrated into
the scope 300, one or more of these components may be external to
the scope 300.
[0061] FIG. 4 is an illustration of an actual firearm 400 onto
which a replica scope 405 has been mounted to facilitate the use of
the firearm 400 within a simulation environment. The replica scope
405 may include components similar to those described above with
reference to FIG. 3. With additional reference to FIG. 4, in
certain embodiments, one or more tracking sensors 410 may be
attached to the firearm 400 to facilitate motion tracking.
Additionally, an inert round 415 may be inserted into the firearm
to facilitate detection of shot events.
[0062] FIG. 5 is an illustration of a firearm 500 into which
simulation components have been integrated. The firearm 500
illustrated in FIG. 5 may be a replica firearm or an actual firearm
that has been modified to include integrated simulation components.
For example, the firearm 500 may be modified to include a trigger
mechanism, a tracking system, and/or a communications interface.
Similarly, the scope may be modified to include a display device
and a video receiver.
[0063] Indeed, a wide variety of different firearm configurations
may be utilized in accordance with various embodiments of the
invention. These configurations should not be limited to those
described in FIGS. 2-5. Additionally, a wide variety of different
types of firearms and/or other weapons may be utilized in
conjunction with the embodiments of the invention. Embodiments of
the invention are applicable to a myriad of weapon types that use
an optical aiming system, such as rifles, handguns, missile
launchers, etc.
[0064] FIG. 6 is a diagram of a simulation environment 600 that may
be utilized in accordance with certain embodiments of the
invention. The simulation environment 600 may include any number of
displays 605, such as projection screens, that facilitate
presentation of simulation scenarios. Although three projection
screens are illustrated in FIG. 6, any number of suitable displays
may be utilized. For example, a relatively circular simulation
environment may be provided in which displays surround users in
three hundred and sixty degrees. Additionally, as desired, any
number of firearms 610a-n may be utilized in conjunction with the
simulation environment 600. Each of the firearms 610a-n may be
associated with a respective scope-integrated or scope-mounted
display. Thus, users may view a simulation scenario within the
simulation environment 600 via the external displays 605. Each user
may then use a scope-mounted display associated with his/her
firearm to view image data determined based upon the viewpoint of
the firearm.
[0065] As shown in FIG. 6, a perspective may be predetermined or
assumed for the external displays 605. For example, a calibration
point 615 may be situated at a relatively central point for viewing
the external displays 605. As desired, the firearms may be
calibrated for use with the simulation system 600 at the
calibration point 615.
[0066] During a simulation, a user may move a firearm, such as
firearm 610a, within the simulation environment 600. As the firearm
is moved, the firearm 610a may be pointed at different portions of
the external displays 605. Accordingly, the viewpoint of the
firearm 610a may change as the firearm 610a is moved. During the
generation or determination of image data to be presented via a
scope-mounted display associated with the firearm 610a, the
position of the firearm 610a within the simulation environment 600
may be autocorrected to the calibration point 615. For example, an
area or portion of the external displays 605 at which the firearm
610a is aimed may be determined. A viewpoint of that area or
portion from the perspective of the calibration point may then be
determined as the viewpoint of the firearm 610a. In other words,
regardless of the user's position within the simulation
environment, it may be assumed that the user is positioned at the
calibration point 615. Accordingly, the imagery illustrated via the
scope-mounted display may correlate or correspond to the imagery
presented to the user via the external displays 605.
[0067] Operational Overview
[0068] FIG. 7 is a flow diagram of an example method 700 for
presenting image data via a display associated with a scope mounted
onto a firearm, according to an example embodiment of the
invention. In certain embodiments, the method 700 may be performed
by a suitable firearm, such as the firearm 105 illustrated in FIG.
1. The method 700 may begin at block 705.
[0069] At block 705, communication may be initiated with a host
controller, such as a host computer. In this regard, motion
tracking information may be communicated to the host controller,
and the image data may be received from the host controller for
presentation via the display associated with the scope.
[0070] At block 710, interaction with one or more tracking systems
(e.g., a laser tracking system, an electromagnetic tracking system,
an optical tracking system, etc.) may be initiated. Each tracking
system may be configured to track motion and/or movement of the
firearm 105 within a simulation environment. In certain
embodiments, a tracking system may be configured to determine
position and/or orientation information for the firearm 105. For
example, as illustrated in optional block 715, position and/or
orientation information for the firearm 105 may be determined, and
the position and/or orientation information may be communicated to
the host controller at block 720.
[0071] In other embodiments; tracking information collected by a
tracking system may be provided or communicated to the host
controller at block 720, and the host controller may determine
position and/or orientation information for the firearm 105.
Regardless of the component or system that determines the position
and/or orientation information for the firearm 105, a viewpoint of
the firearm 105 or the firearm scope may be determined by the host
controller, and the host controller may generate image data in
accordance with the determined viewpoint. The image data may be
output by the host controller for receipt by the firearm 105, and
the firearm 105 may receive the image data at block 725. The
received image data, such as a received image feed, may then be
output by the display for presentation to the user at block 730.
Accordingly, when a user views the scope associated with the
firearm 105, the user may view the image data output by the host
controller.
[0072] At block 735, a determination may be made as to whether the
image is calibrated. In other words, a determination may be made as
to whether the firearm 105 has been properly calibrated within the
simulation system and/or the virtual environment. If it is
determined at block 735 that the image is not calibrated, then
operations may continue at block 740. At block 740, the firearm 105
and/or the position of the firearm 105 may be calibrated within the
virtual environment. For example, a user may be prompted to utilize
a suitable calibration routine (e.g., aiming the firearm 105 at the
four corners of an external display, etc.) to calibrate the firearm
105. Operations may then continue at block 725 described above.
[0073] If, however, it is determined at block 735 that the image is
calibrated, then operations may continue at block 745. At block
745, a determination may be made as to whether a trigger associated
with the firearm 105 has been pulled or actuated by a user. In
other words, a determination may be made as to whether a shot has
occurred. As described in greater detail above with reference to
FIG. 1, a wide variety of suitable trigger mechanisms and/or
trigger detection devices (e.g., an inert round, etc.) may be
utilized to detect a trigger actuation. If it is determined at
block 745 that the trigger has not been actuated or pulled, then
operations may continue at block 765 described in greater detail
below. If, however, it is determined at block 745 that the trigger
has been actuated or pulled, then operations may continue at block
750.
[0074] At block 750, a shot indication may be communicated from the
firearm 105 to the host controller. The shot indication may be
processed by the host controller to simulate a shot within a
virtual environment. For example, one or more ballistic models may
be utilized to determine a trajectory of a shot within the virtual
environment. Image data (e.g., an image feed) may then be updated
by the host controller to reflect the shot, and the updated image
data may be output by the host controller for receipt by the
firearm 105. The firearm 105 may receive the updated image data at
block 755 and present information associated with a virtual shot
via the display at block 760. Operations may then continue at block
765.
[0075] At block 765, a determination may be made as to whether a
position and/or orientation of the firearm 105 has changed. For
example, a determination may be made as to whether the firearm 105
has been moved or whether a tracking system has detected movement
of the firearm 105. If it is determined at block 765 that the
position and/or orientation of the firearm 105 has not changed,
then operations may continue at block 730, and image data may
continue to be output. If, however, it is determined at block 765
that the position and/or orientation of the firearm 105 has
changed, then operations may continue at block 715 (or block 720).
In this regard, a new viewpoint of the firearm 105 may be
determined, and the image data output by the display may be updated
to reflect the new viewpoint.
[0076] The method 700 may end following the completion of a
simulation scenario in which the firearm 105 is utilized.
[0077] FIG. 8 is a flow diagram of an example method 800 for
determining image data to be presented via a display associated
with a scope mounted onto a firearm, according to an example
embodiment of the invention. In certain embodiments, the method 800
may be performed by one or more suitable controllers associated
with a simulation system, such as the controllers 125 illustrated
in FIG. 1. The method 800 may begin at block 805.
[0078] At block 805, the controller 125 may initiate communication
with a firearm and/or a scope, such as the firearm 105 and/or the
scope 110 illustrated in FIG. 1. In this regard, the controller 125
may receive a wide variety of data and/or indications from the
firearm 105, such as indications associated with shot events or
trigger actuations and/or information associated with scope
settings. A communications channel may also be established for
communicating image data to the firearm 105 and/or the scope
110.
[0079] At block 810, one or more firearm and/or scope parameters
may be determined. In certain embodiments, the one or more
parameters may be determined or identified within suitable
ballistics simulation software, such as the ballistics module 168
illustrated in FIG. 1. A wide variety of different firearm and/or
scope parameters may be determined as desired in various
embodiments of the invention, such as a type associated with the
firearm 105 (e.g., a caliber, etc.), a ballistic profile associated
with the firearm 105, a type of ammunition to be simulated, a
ballistic profile associated with the ammunition, a zoom or
magnification for the scope 110, a mode of operation for the scope
110 (e.g., normal operation, an infrared mode, a night vision mode,
etc.), and/or windage and/or elevation adjustments for the scope
110, etc. In this regard, the trajectory of simulated shots may be
determined in a relatively accurate manner.
[0080] At block 815, the controller 125 may initiate communication
with one or more suitable tracking systems associated with the
firearm 105 and/or the scope 110, such as the motion tracking
system 120 illustrated in FIG. 1. In this regard, the controller
125 may receive motion, position, and/or orientation data
associated with the firearm 105. At block 820, position,
orientation, and/or viewpoint information for the firearm 105 may
be determined. For example, tracking data received from the motion
tracking systems 120 may be translated into position and/or
orientation data for the firearm 105. A viewpoint of the firearm
105 within the virtual environment may then be determined.
[0081] At block 825, which may be optional in certain embodiments
of the invention, a position of the firearm 105 within the virtual
environment may be corrected or adjusted. For example, the position
of the firearm 105 may be set to a predetermined position, such as
a predetermined calibration point. In this regard, a determined
viewpoint for the firearm 105 may correspond with the viewpoint
presented via one or more external (i.e., non-firearm-mounted)
displays.
[0082] At block 830, an image feed for display via a scope-mounted
display may be determined and/or generated based at least in part
upon the determined viewpoint for the firearm 105. The image feed
may then be output at block 835 for receipt by the firearm 105 and
presentation to the user via the scope-mounted display. At block
840, a determination may be made as to whether the image
communicated to the firearm 105 is calibrated. In other words, a
determination may be made as to whether the firearm 105 has been
properly calibrated within the simulation system and/or the virtual
environment. If it is determined at block 840 that the image is not
calibrated, then operations may continue at block 845. At block
845, the firearm 105 and/or the position of the firearm 105 may be
calibrated within the virtual environment. For example, a user may
be prompted to utilize a suitable calibration routine (e.g., aiming
the firearm 105 at the four corners of an external display, etc.)
to calibrate the firearm 105. Operations may then continue at block
830 described above.
[0083] If, however, it is determined at block 840 that the image is
calibrated, then operations may continue at block 850. At block
850, a determination may be made as to whether a shot event has
occurred. For example, a determination may be made as to whether a
shot indication has been received from the firearm 105. If it is
determined at block 850 that a shot event has not occurred, then
operations may continue at block 865 described in greater detail
below. lf, however, it is determined at block 850 that a shot event
has occurred, then operations may continue at block 855.
[0084] At block 855, the shot event may be registered within the
ballistic simulation software, and a trajectory of the shot may be
calculated or determined. As desired, a wide variety of ballistic
factors, such as weather conditions, elevation, ammunition type,
firearm type, etc., may be taken into consideration during the
calculation or determination of a shot trajectory. Once a shot
trajectory has been calculated, operations may continue at block
860, and the image feed may be adjusted or modified to correspond
with the registered shot. Image information associated with the
simulated shot may then be output for receipt via the firearm 105
and presentation via the scope-mounted display. Operations may then
continue at block 865.
[0085] At block 865, a determination may be made as to whether a
position and/or orientation of the firearm 105 has changed. For
example, a determination may be made as to whether motion or
movement information has been received from one or more tracking
systems. As another example, a determination may be made as to
whether updated position and/or orientation information has been
received. If it is determined at block 865 that the position and/or
orientation of the firearm 105 has not changed, then operations may
continue at block 835, and image data may continue to be output for
receipt by the firearm 105. If, however, it is determined at block
865 that the position and/or orientation of the firearm 105 has
changed, then operations may continue at block 820, and a new
viewpoint for the firearm 105 may be determined. In this regard,
the image data output for display by the scope may be updated to
reflect the new viewpoint.
[0086] The method 800 may end following the completion of a
simulation scenario associated with the controller 125.
[0087] The operations described and shown in the methods 700 and
800 of FIGS. 7-8 may be carried out or performed in any suitable
order as desired in various embodiments of the invention.
Additionally, in certain embodiments, at least a portion of the
operations may be carried out in parallel. Furthermore, in certain
embodiments, less than or more than the operations described in
FIGS. 7-8 may be performed.
[0088] Various block and/or flow diagrams of systems, methods,
apparatus, and/or computer program products according to example
embodiments of the invention are described above. It will be
understood that one or more blocks of the block diagrams and flow
diagrams, and combinations of blocks in the block diagrams and flow
diagrams, respectively, can be implemented by computer-executable
program instructions. Likewise, some blocks of the block diagrams
and flow diagrams may not necessarily need to be performed in the
order presented, or may not necessarily need to be performed at
all, according to some embodiments of the invention.
[0089] These computer-executable program instructions may be loaded
onto a special purpose computer or other particular machine, a
processor, or other programmable data processing apparatus to
produce a particular machine, such that the instructions that
execute on the computer, processor, or other programmable data
processing apparatus create means for implementing one or more
functions specified in the flow diagram block or blocks. These
computer program instructions may also be stored in a
computer-readable memory that can direct a computer or other
programmable data processing apparatus to function in a particular
manner, such that the instructions stored in the computer-readable
memory produce an article of manufacture including instruction
means that implement one or more functions specified in the flow
diagram block or blocks. As an example, embodiments of the
invention may provide for a computer program product, comprising a
computer-usable medium having a computer-readable program code or
program instructions embodied therein, said computer-readable
program code adapted to be executed to implement one or more
functions specified in the flow diagram block or blocks. The
computer program instructions may also be loaded onto a computer or
other programmable data processing apparatus to cause a series of
operational elements or steps to be performed on the computer or
other programmable apparatus to produce a computer-implemented
process such that the instructions that execute on the computer or
other programmable apparatus provide elements or steps for
implementing the functions specified in the flow diagram block or
blocks.
[0090] Accordingly, blocks of the block diagrams and flow diagrams
support combinations of means for performing the specified
functions, combinations of elements or steps for performing the
specified functions and program instruction means for performing
the specified functions. It will also be understood that each block
of the block diagrams and flow diagrams, and combinations of blocks
in the block diagrams and flow diagrams, can be implemented by
special purpose, hardware-based computer systems that perform the
specified functions, elements or steps, or combinations of special
purpose hardware and computer instructions.
[0091] Many modifications and other embodiments of the invention
set forth herein will be apparent having the benefit of the
teachings presented in the foregoing descriptions and the
associated drawings. Therefore, it is to be understood that the
invention is not to be limited to the specific embodiments
disclosed and that modifications and other embodiments are intended
to be included within the scope of the appended claims. Although
specific terms are employed herein, they are used in a generic and
descriptive sense only and not for purposes of limitation.
* * * * *