U.S. patent number RE47,648 [Application Number 15/041,784] was granted by the patent office on 2019-10-15 for integrated multi-sensor control system and method.
This patent grant is currently assigned to AGJUNCTION LLC. The grantee listed for this patent is AgJunction LLC. Invention is credited to Joshua M. Gattis, Malcolm B. Jones, Andreas F. Ramm, David R. Reeve, Aaron C. Stichter.
![](/patent/grant/RE047648/USRE047648-20191015-D00000.png)
![](/patent/grant/RE047648/USRE047648-20191015-D00001.png)
![](/patent/grant/RE047648/USRE047648-20191015-D00002.png)
![](/patent/grant/RE047648/USRE047648-20191015-D00003.png)
![](/patent/grant/RE047648/USRE047648-20191015-D00004.png)
![](/patent/grant/RE047648/USRE047648-20191015-D00005.png)
![](/patent/grant/RE047648/USRE047648-20191015-D00006.png)
![](/patent/grant/RE047648/USRE047648-20191015-D00007.png)
![](/patent/grant/RE047648/USRE047648-20191015-D00008.png)
![](/patent/grant/RE047648/USRE047648-20191015-D00009.png)
![](/patent/grant/RE047648/USRE047648-20191015-D00010.png)
View All Diagrams
United States Patent |
RE47,648 |
Reeve , et al. |
October 15, 2019 |
Integrated multi-sensor control system and method
Abstract
A GNSS integrated multi-sensor guidance system for a vehicle
assembly includes a suite of sensor units, including a global
navigation satellite system (GNSS) sensor unit comprising a
receiver and an antenna. An inertial measurement unit (IMU) outputs
vehicle dynamic information for combining with the output of the
GNSS unit. A controller with a processor receives the outputs of
the sensor suite and computes steering solutions, which are
utilized by vehicle actuators, including an automatic steering
control unit connected to the vehicle steering for guiding the
vehicle. The processor is programmed to define multiple
behavior-based automatons comprising self-operating entities in the
guidance system, which perform respective behaviors using data
output from one or more sensor units for achieving the behaviors. A
GNSS integrated multi-sensor vehicle guidance method is also
disclosed.
Inventors: |
Reeve; David R. (Chapel Hill,
AU), Jones; Malcolm B. (Bracken Ridge, AU),
Ramm; Andreas F. (Deception Bay, AU), Stichter; Aaron
C. (Apache Junction, AZ), Gattis; Joshua M. (Robinson,
KS) |
Applicant: |
Name |
City |
State |
Country |
Type |
AgJunction LLC |
Scottsdale |
AZ |
US |
|
|
Assignee: |
AGJUNCTION LLC (Hiawatha,
KS)
|
Family
ID: |
44647877 |
Appl.
No.: |
15/041,784 |
Filed: |
February 11, 2016 |
Related U.S. Patent Documents
|
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
Issue Date |
|
|
61243417 |
Sep 17, 2009 |
|
|
|
Reissue of: |
12884038 |
Sep 16, 2010 |
8649930 |
Feb 11, 2014 |
|
|
Current U.S.
Class: |
1/1 |
Current CPC
Class: |
G05D
1/0278 (20130101); B62D 1/28 (20130101); A01B
79/005 (20130101); G05D 1/027 (20130101); B62D
1/02 (20130101); G05D 1/0278 (20130101); A01B
69/008 (20130101); B62D 1/28 (20130101); G05D
1/027 (20130101); B62D 1/02 (20130101); A01B
69/008 (20130101); A01B 79/005 (20130101); G05D
2201/0201 (20130101); G05D 1/0246 (20130101); G05D
1/024 (20130101); G05D 1/0274 (20130101); G05D
1/0257 (20130101); G05D 1/0274 (20130101); G05D
2201/0201 (20130101); G05D 1/0257 (20130101) |
Current International
Class: |
B62D
1/02 (20060101); A01B 79/00 (20060101); B62D
1/28 (20060101); G05D 1/02 (20060101); A01B
69/04 (20060101) |
Field of
Search: |
;701/23,24,25,26,28,468,470,472,475 |
References Cited
[Referenced By]
U.S. Patent Documents
Foreign Patent Documents
|
|
|
|
|
|
|
07244150 |
|
Sep 1995 |
|
JP |
|
WO9836288 |
|
Aug 1998 |
|
WO |
|
WO0024239 |
|
May 2000 |
|
WO |
|
WO03019430 |
|
Mar 2003 |
|
WO |
|
WO2005/119386 |
|
Dec 2005 |
|
WO |
|
WO2009/066183 |
|
May 2009 |
|
WO |
|
WO2009126587 |
|
Oct 2009 |
|
WO |
|
WO2009/148638 |
|
Dec 2009 |
|
WO |
|
Other References
Information Disclosure Statement, Listing of Related Cases, Aug. 2,
2016, 1 page. cited by applicant .
Noh, Kwang-Mo, Self-tuning controller for farm tractor guidance,
Iowa State University Retrospective Theses and Dissertations, Paper
9874, (1990). cited by applicant .
Van Zuydam,. R.P., Centimeter-Precision Guidance of Agricultural
Implements in the Open Field by Means of Real Tim Kinematic DGPS,
ASA-CSSA-SSSA, pp. 1023-1034 (1999). cited by applicant .
Parkinson, Bradford W., et al., "Global Positioning System: Theory
and Applications, vol. II", Bradford W. Parkinson and James J.
Spiker, Jr., eds., Global Postioning System: Theory and
Applicaitons, vol. II, 1995, AIAA, Reston, VA, USA, pp. 3-50,
(1995), 3-50. cited by applicant .
"Orthman Manufacturing Co., www.orthman.com/htm;guidance.htm", 2004
regarding the "Tracer Quick-Hitch". cited by applicant .
Lin, Dai et al., "Real-time Attitude Determination fro
Microsatellite by Lamda Method Combined with Kalman Filtering", A
Collection fof the 22nd AIAA International Communications Satellite
Systems Conference and Exhibit Technical Paers vol. 1, Monetrey,
California American Institute of Aeronautics and Astronautics,
Inc., (May 2004),136-143. cited by applicant .
Xu, Jiangning et al., "An EHW Architecture for Real-Time GPS
Attitude Determination Based on Parallel Genetic Algorithm", The
Computer SocietyProceedings of the 2002 NASA/DOD Conference on
Evolvable Hardware (EH'02), (2002). cited by applicant .
Han, Shaowel et al., "Single-Epoch Ambiguity Resolution for
Real-Time GPS Attitude Determination with the Aid of
One-Dimensional Optical Fiber Gyro", GPS Solutions, vol. 3, No. 1,
pp. 5-12 (1999) John Wiley & Sons, Inc. cited by applicant
.
Park, Chansik et al., "Integer Ambiguity Resolution for GPS Based
Attitude Determination System", SICE 1998, Jul. 29-31, Chiba,
1115-1120. cited by applicant .
Last, J. D., et al., "Effect of skywave interference on coverage of
radiobeacon DGPS Stations", IEEE Proc.--Radar, Sonar Navig., vol.
144, No. 3, Jun. 1997, pp. 163-168. cited by applicant .
"International Search Report and Written Opinion",
PCT/US2004/015678, filed May 17, 2004, Jun. 21, 2005. cited by
applicant .
"ISO", 11783 Part 7 Draft Amendment 1 Annex, Paragraphs B.6 and
B.7.ISO 11783-7 2004 DAM1, ISO: Mar. 8, 2004. cited by applicant
.
Kaplan, E D., "Understanding GPS: Principles and Applications",
Artech House, MA, 1996. cited by applicant .
Irsigler, M et al., "PPL Tracking Performance in the Presence of
Oscillator Phase Noise", GPS Solutions, vol. 5, No. 4, pp. 45-57
(2002). cited by applicant .
Ward, Phillip W., "Performance Comparisons Between FLL, PLL and a
Novel FLL-Assisted-PLL Carrier Tracking Loop Under RF Interference
Conditions", 11th Int. Tech Meeting of the Satellite Division of
the U.S. Inst. of Navigation, Nashville, TN, Sep. 15-18, 783-795,
1998. cited by applicant .
Bevly, David M., "Comparison of INS v. Carrier-Phase DGPS for
Attitude Determination in the Control of Off-Road Vehicles", ION
55th Annual Meeting; Jun. 28-30, 1999; Cambridge, Massachusetts;
pp. 497-504. cited by applicant .
"International Search Report and Written Opinion", International
Searching Authortiy, PCT/US08/88070, Feb. 9, 2009. cited by
applicant .
Keicher, R. et al., "Automatic Guidance for Agricultural Vehicles
in Europe", Computers and Electronics in Agriculture, vol. 25,
(Jan. 2000),169-194. cited by applicant .
Takac, Frank et al., "SmartRTK: A Novel Method of Processing
Standardised RTCM Network TRK Information For High Precision
Positioning", Proceedings of ENC GNSS 2008, Toulouse, France,(Apr.
22, 2008). cited by applicant .
"International Search Report", PCT/US09/33567, (Feb. 9, 2009).
cited by applicant .
"International Search Report", PCT/US09/49776, (Aug. 11, 2009).
cited by applicant .
"International Search Report", PCT/AU/2008/000002, (Feb. 28, 2008).
cited by applicant .
"International Search Report and Written Opinion",
PCT/IB2008/003796 (Jul. 15, 2009). cited by applicant .
"International Search Report", PCT/US09/33693, (Mar. 30, 2009).
cited by applicant .
"International Search Report", PCT/US09/039686, (May 26, 2009).
cited by applicant .
"International Search Report", PCT/US09/34376, (Nov. 2, 2009).
cited by applicant .
"International Search Report / Written Opinion", PCT/US09/63594,
(Jan. 11, 2010). cited by applicant .
"International Search Report", PCT/US09/60668, (Dec. 9, 2009).
cited by applicant .
"International Search Report", PCT/US09/067693, (Jan. 26, 2010).
cited by applicant .
"International Search Report and Written Opinion", PCT/US10/21334,
(Mar. 12, 2010). cited by applicant .
Rho, Hyundho et al., "Dual-Frequency GPS Precise Point Positioning
with WADGPS Corrections", [retrieved on May 18, 2010]. Retrieved
from the Internet: ,URL:
http://gauss.gge.unb.ca/papers.pdf/iongnss2005.rho.wadgps.pdf,
(Jul. 12, 2006). cited by applicant .
"Eurocontrol, Pegasus Technical Notes on SBAS", report [online],
Dec. 7, 2004 [retrieved on May 18, 2010]. Retrieved from the
Internet: <URL:
http://vvww.icao.int/icao/en/ro/nacc/meetings/2004/gnss/documentation/Peg-
asus/tn.pdf>, (Dec. 7, 2004),p. 89 paras [0001]-[0004]. cited by
applicant .
"ARINC Engineering Services, Interface Specification IS-GPS-200,
Revision D", Online [retrieved on May 18, 2010]. Retrieved from the
Internet<URL:
http://www.navcen.uscg.gov/gps/geninfo/IS-GPS-200D.pdf>, (Dec.
7, 2004),p. 168 para [0001]. cited by applicant .
Schaer, et al., "Determination and Use of GPS Differential Code
Bias Values", Presentation [online]. Revtrieved May 18, 2010.
Retrieved from the
internet:.<http://nng.esoc.esa.de/ws2006/REPR2.pdf> (May 8,
2006). cited by applicant .
"International Search Report", PCT/US10/26509 (Apr. 20, 2010).
cited by applicant.
|
Primary Examiner: Sager; Mark
Attorney, Agent or Firm: Schwabe Williamson & Wyatt
Parent Case Text
CROSS REFERENCE TO RELATED APPLICATION
.[.This application relates to U.S. Provisional Patent Application
Ser. No. 61/243,417, filed Sep. 17, 2009, filed concurrently
herewith, which is incorporated herein by reference..].
.Iadd.The present application is a reissue application of U.S. Pat.
No. 8,649,930, issued Feb. 11, 2014, entitled: GNSS INTEGRATED
MULTI-SENSOR CONTROL SYSTEM AND METHOD; which claims benefit of
U.S. Provisional Patent Application No. 61/243,417, filed Sep. 17,
2009, the contents and disclosures of which are hereby incorporated
by reference in their entireties..Iaddend.
Claims
Having thus described the disclosed subject matter, what is claimed
as new and desired to be secured by Letters Patent is:
1. An integrated multi-sensor guidance system for a vehicle
assembly including a steering subsystem, which guidance system
includes: said vehicle assembly having a dynamic attitude
comprising a geo-reference location, vehicle assembly orientation
and vehicle assembly speed; a processor with multiple sensor inputs
and actuator outputs; a suite of sensor units each connected to a
respective sensor input; said sensor unit suite includes a GNSS
unit with an antenna and a receiver connected to said antenna,
.Iadd.a wheel angle sensor unit (WAS), and an inertial measurement
unit (IMU) sensor, .Iaddend.said GNSS unit providing output signals
corresponding to the GNSS-defined locations of said vehicle
assembly dynamic attitude to a respective processor input; a
guidance controller adapted for receiving signal input and
generating control output based on said signal input; a data
storage device including memory storage; a suite of actuator units
each connected to a respective actuator output; said guidance
controller being adapted for receiving and storing in said memory
storage device GNSS-based positioning signals; said processor being
adapted for computing a GNSS-based. guide pattern; said guidance
controller being adapted for providing output signals to a display
device for displaying vehicle motion relative to guide patterns and
contrasting displays of areas treated by said vehicle along
previously-traveled portions of said guide patterns; said guidance
controller being adapted for calibrating and storing in said memory
multiple vehicle profiles, each said profile including multiple,
independent vehicle-specific automatons; an accepting interface for
accepting requests from other automatons; a requesting interface
for making requests to another automaton; a knowledge input for
receiving a behavioral definition for affecting the behavior of the
automatons; a data input for receiving input data; a data output
for sending output data; said processor programmed to determine
.Iadd.different .Iaddend.variable confidence levels in real time
for each .Iadd.of .Iaddend.said .[.sensor.]. .Iadd.GNSS
.Iaddend.unit.Iadd., WAS, and IMU sensor .Iaddend.based on its
current relative performance; said processor programmed to utilize
said .[.sensor.]. .Iadd.GNSS .Iaddend.unit.Iadd., WAS, and IMU
sensor .Iaddend.outputs proportionally based on their respective
confidence levels in generating said control output signals; and
wherein said processor is programmed to define multiple
behavior-based automatons comprising self-operating entities in
said guidance system, said automatons performing respective
behaviors using data output from said .[.one or more sensor
units.]. .Iadd.GNSS unit, said WAS, and said IMU sensor
.Iaddend.for achieving said behaviors and wherein said .[.one or
more sensor units.]. .Iadd.GNSS unit, WAS, and IMU sensor
.Iaddend.provide .Iadd.at least some of .Iaddend.the same or
similar data.
2. The guidance system as claimed in claim 1, wherein .[.said
sensor unit suite includes an inertial measurement unit (IMU)
sensor providing output signals corresponding to an inertial aspect
of a dynamic attitude of said vehicle assembly to a respective
processor input.]. .Iadd.said guidance controller sends steering
control instructions instructing said automatons to perform
respective behaviors using data output from one or more GNSS unit,
WAS, and IMU sensor.Iaddend..
3. The guidance system as claimed in claim .[.2.].
.Iadd.1.Iaddend., wherein said guidance controller is adapted for
receiving inertial measurement signals .Iadd.from the IMU
.Iaddend.and integrating said inertial measurement signals with
said GNSS-based positioning signals.
4. The guidance system as claimed in claim 1, wherein said.
actuator unit suite includes a steering unit connected to said
steering subsystem and receiving said control output signals from
said processor.
5. The guidance system as claimed in claim 4, wherein said steering
subsystem includes: a steering controller including a steering
processor and connected to said guidance controller; said steering
controller receiving guidance signals as inputs from said guidance
controller and computing steering signals as outputs from said
steering controller; and said steering actuator receiving said
steering signals from said steering controller and steering said
vehicle in response thereto.
6. The guidance system as claimed in claim 1, wherein said sensor
suite includes sensor units chosen from among the group comprising:
a video camera unit oriented in the vehicle assembly direction of
travel; a radar unit; a laser unit; radio input; telemetry;
.[.material application exclusion areas input;.]. satellite image
inputs; .Iadd.and.Iaddend. .[.contour/elevation overlay inputs;.].
prescription mapping.[.; and a wheel angle sensor (WAS).]..
7. The guidance system as claimed in claim 1, wherein said actuator
suite includes actuator units chosen from among the group
comprising: an implement steering unit, an implement sectional
control unit, personal computer (PC) office software, material
application rate control, secondary vehicle control, mapping, crop
yield, and mapping skips and overlaps.
8. A method of vehicle control and guidance, which method comprises
the steps: providing a vehicle assembly including a steering
subsystem and dynamic attitude comprising a geo-reference location,
vehicle assembly orientation, and vehicle assembly speed; providing
a guidance system including a processor with multiple sensor inputs
and actuator outputs, a suite of sensor units connected to a
respective sensor input, a suite of actuator units connected to a
respective actuator output, and a data storage device including
memory storage; providing a guidance controller; inputting signal
input data to said guidance controller; generating control output
signals with said guidance controller based on said signal input;
receiving and storing in said memory storage device GNSS-based
positioning signals with said guidance controller; computing a
GNSS-based guide pattern with said processor; providing output
signals with said guidance controller to a display device for
displaying vehicle motion relative to guide patterns and
contrasting displays of areas treated by said vehicle along
previously-traveled portions of said guide patterns; calibrating
and storing in said memory multiple vehicle profiles with said
guidance controller, each said profile including multiple,
independent vehicle-specific automatons; wherein said sensor unit
suite includes an inertial measurement unit (IMU) sensor providing
output signals corresponding to an inertial aspect of a dynamic
attitude of said vehicle assembly to a respective processor input;
generating inertial measurement signals with said IMU sensor;
receiving the inertial measurement signals with said guidance
controller; integrating said inertial measurement signals with said
GNSS-based positioning signals; defining multiple behavior-based
automatons comprising self-operating entities in said guidance
system; instructing said automatons to perform respective behaviors
using data output from one or more sensor units for achieving said
behaviors wherein one or inure sensor units provide the same or
similar data; providing each automaton with an accepting interface
for accepting requests from other automatons; providing each
automaton with a requesting interface for making requests to
another automaton; providing each automaton with a knowledge input
for receiving a behavioral definition for affecting the behavior of
the automatons; providing each automaton with a data input for
receiving input data; providing each automaton with a data output
for sending data; requesting instructions by each automaton from
each other automaton; and accepting instructions by each automaton
provided from each other automaton.
9. A method of vehicle control and guidance as claimed by claim 8,
including the steps: determining variable confidence levels with
the processor in real time for each said sensor unit based on
current relative performance; and utilizing said sensor unit
outputs proportionally based on the respective confidence levels in
generating said control output signals.
10. A method of vehicle control and guidance as claimed by claim 8,
including the steps: providing a steering unit connected to said
steering system; and receiving said control output signals at said
steering unit as steering control instructions.
11. A method of vehicle control and guidance as claimed by claim
10, including the steps: providing a steering processor connected
to said guidance controller; receiving guidance signals at said
steering controller as inputs from said guidance controller;
computing steering signals as outputs from said steering
controller; receiving said steering signals with said steering
actuator; and steering said vehicle assembly in response to said
steering signals.
12. An integrated multi-sensor guidance system for a vehicle
.[.assembly including a steering subsystem, which guidance system
includes.]..Iadd., comprising.Iaddend.: .[.said vehicle assembly
having a dynamic attitude comprising a geo-reference location,
vehicle assembly orientation and vehicle assembly speed;.]. a
processor with multiple sensor inputs and actuator outputs; a suite
of sensor units each connected to a respective sensor input;
.[.said sensor unit suite including a GNSS unit with an antenna and
a receiver connected to said antenna, said GNSS unit providing
output signals corresponding to the GNSS-defined locations of said
vehicle assembly dynamic attitude to a respective processor input;
said sensor unit suite including an inertial measurement unit (IMU)
sensor providing output signals corresponding to an inertial aspect
of a dynamic altitude of said vehicle assembly to a respective
processor input; said guidance controller being adapted for
receiving said inertial measurement signals and integrating said
inertial measurement signals with said GNSS-based positioning
signals;.]. said processor being programmed to determine variable
confidence levels in real time for each said sensor unit based on
its current relative performance; said processor being programmed
to utilize .[.said.]. sensor unit outputs proportionally based on
their respective confidence levels in generating .[.said.].
steering signals; .[.a suite of actuator units each connected to a
respective actuator output; said actuator unit suite including a
steering unit connected to said steering subsystem and receiving
said steering signals from said processor;.]. said processor being
programmed to define multiple behavior-based automatons comprising
self-operating entities in said guidance system, said automatons
performing respective behaviors using data output from one or more
said sensor units for achieving said behaviors and wherein said one
or more sensor units provide the same or similar data; each said
automaton having: an accepting interface for accepting requests
from other automatons; a requesting interface for making requests
to another automatons; a knowledge input for receiving a behavioral
definition for affecting the behavior of the automatons; a data
input for receiving input data; and a data output for sending out
the data; .[.said guidance controller being adapted for receiving
and storing.]. .Iadd.the same or another processor further
configured to: receive and store .Iaddend.in .[.said.]. .Iadd.a
.Iaddend.memory storage device .Iadd.global navigation satellite
system (.Iaddend.GNSS.Iadd.).Iaddend.-based positioning signals
.Iadd.and inertial measurement signals from an inertial measurement
unit (IMU) sensor;.Iaddend. .[.said processor being adapted for
computing.]. .Iadd.compute .Iaddend.a GNSS-based guide pattern;
.[.said guidance controller being adapted for providing.].
.Iadd.provide .Iaddend.output signals to a display device .[.for
displaying.]. .Iadd.to display .Iaddend.vehicle motion .Iadd.for
the vehicle .Iaddend.relative to .[.guide patterns.]. .Iadd.the
guide pattern .Iaddend.and .[.contrasting.]. .Iadd.to contrast
.Iaddend.displays of areas treated by .[.said.]. .Iadd.the
.Iaddend.vehicle along previously-traveled portions of .[.said.].
.Iadd.the .Iaddend.guide .[.patterns.]. .Iadd.pattern.Iaddend.;
.[.and.]. .[.said guidance controller being adapted for calibrating
and storing.]. .Iadd.calibrate and store .Iaddend.in .[.said.].
.Iadd.the .Iaddend.memory .Iadd.storage device .Iaddend.multiple
vehicle profiles, each said profile including multiple, independent
vehicle-specific automatons.
.Iadd.13. A guidance system for operating multiple vehicles, the
guidance system including a guidance controller to: receive and
store in a memory storage device data from sensor units located on
a first vehicle including velocity, acceleration and GNSS-based
position data derived from a wheel angle sensor (WAS), inertial
measurement unit (IMU), and global navigation satellite system
(GNSS); determine variable confidence levels in real time the said
sensor units based on its current relative performance; define
multiple behavior-based automatons comprising self-operating
entities in said guidance system, said automatons performing
respective behaviors using data output from said one or more sensor
units for achieving said behaviors and wherein said one or more
sensor units provide the same or similar data; each said automaton
having: an accepting interface for accepting requests from other
automatons; a requesting interface for making requests to another
automatons; a knowledge input for receiving a behavioral definition
for affecting the behavior of the automatons; a data input for
receiving input data; and a data output for sending out the data;
calibrate and store in the memory storage device multiple vehicle
profiles, each said profile including multiple, independent
vehicle-specific automatons; compute a guide path for the first
vehicle from at least the GNSS-based position data; send the
GNSS-based position data to a display device to display movement of
the first vehicle relative to the guide path; provide output
signals to the display device to display vehicle motion for the
first vehicle relative to the guide path and to contrast displays
of areas treated by the first vehicle along previously-traveled
portions of the guide path; store in the memory storage device a
vehicle profile for a second vehicle; send guidance instructions to
the second vehicle to use data output from one or more of the
sensor units that provide the same position data as derived from
the GNSS, WAS, and IMU to control movement of the second vehicle in
cooperation with movement of the first vehicle along the guide
path; and update the guidance instructions to the second vehicle
based on computed differences in the position data of the first
vehicle..Iaddend.
.Iadd.14. An integrated multi-sensor guidance system comprising a
guidance controller to: receive signals from different sensor units
including a global navigation satellite system (GNSS)-based sensor
unit, a wheel angle sensor (WAS), and an inertial measurement unit
(IMU) sensor providing positions for a vehicle; define multiple
behavior-based automatons comprising self-operating entities in
said guidance system performing respective behaviors using data
output from the sensor units and wherein said sensor units provide
at least some same or similar data, each said automaton having: an
accepting interface for accepting requests from other automatons; a
requesting interface for making requests to another automatons; a
knowledge input for receiving a behavioral definition for affecting
the behavior of the automatons; a data input for receiving input
data; and a data output for sending out the data; calibrate and
store in a memory storage device multiple vehicle profiles, each
said profile including multiple, independent vehicle-specific
automatons; compute a GNSS-based guide path for the vehicle;
determine variable confidence levels in real-time for the
GNSS-based sensor unit, the WAS, and the IMU sensor based on
current relative performance of the sensor units; use output
signals from a first one of the sensor units with a highest one of
the confidence levels to provide output signals to determine
movements of the vehicle relative to the guide path; use output
signals from the first one of the sensor units with the highest one
of the confidence levels to calibrate a second one of the sensor
units with a lower one of the confidence levels; and provide output
signals to a display device to display vehicle motion for the
vehicle relative to the guide path and to contrast displays of
areas treated by the vehicle along previously-traveled portions of
the guide path..Iaddend.
Description
BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention relates generally to a versatile integrated
multi-sensor apparatus which combines positional data from a
variety of sensor types including a GNSS system. The various sensor
data is ranked according to its confidence level, and using that
data as a means to automatically create a planned path and steer a
vehicle along that planned path. Elements of the present invention
allow the system to be easily interchangeable among a multitude of
vehicles and to communicate with other vehicles to allow for
autonomous cooperative vehicle behavior building and task
delegation.
2. Description of the Related Art
Global navigation satellite system (GNSS) guidance and control are
widely used for vehicle and personal navigation and a variety of
other uses involving precision location in geodesic reference
systems. GNSS, which includes the Global Positioning System (GPS)
and other satellite-based positioning systems, has progressed to
sub-centimeter accuracy with known correction techniques, including
a number of commercial satellite based augmentation systems
(SBASs).
For even more accurate information, higher frequency signals with
shorter wavelengths are required. It is known in the art that by
using GNSS satellites' carrier phase transmissions, and possibly
carrier phase signal components from base reference stations or
satellite based augmentation systems (SBAS), including the Wide
Area Augmentation System (WAAS) (U.S.), and similar systems such as
EGNOS (European Union) and MSAS (Japan), a position may readily be
determined to within millimeters. When accomplished with two
antennas at a fixed spacing, an angular rotation may be computed
using the position differences. In an exemplary embodiment, two
antennas placed in the horizontal plane may be employed to compute
a heading (rotation about a vertical axis) from a position
displacement. Heading information, combined with position, either
differentially corrected (DGPS) or carrier phase corrected
real-time kinematic (RTK), provides the feedback information
desired for a proper control of the vehicle direction.
Another benefit achieved by incorporating a GNSS-based heading
sensor is the elimination or reduction of drift and biases
resultant from a gyro-only or other inertial sensor approach. Yet
another advantage is that heading may be computed while movable
equipment is stopped or moving slowly, which is not possible in a
single-antenna, GNSS-based approach that requires a velocity vector
to derive a heading. Yet another advantage of incorporating a
GNSS-based heading sensor is independence from a host vehicle's
sensors or additional external sensors. Thus, such a system is
readily maintained as equipment-independent and may be moved from
one vehicle to another with minimal effort. Yet another exemplary
embodiment of the sensor employs global navigation satellite system
(GNSS) sensors and measurements to provide accurate, reliable
positioning information. GNSS sensors include, but are not limited
to, GPS, Global Navigation System (GLONAS), Wide Area Augmentation
System (WAAS) and the like, as well as combinations including at
least one of the foregoing.
An example of a GNSS is the Global Positioning System (GPS)
established by the United States government, which employs a
constellation of 24 or more satellites in well-defined orbits at an
altitude of approximately 26,500 km. These satellites continually
transmit microwave L-band radio signals in two frequency bands,
centered at 1575.42 MHz and 1227.6 MHz, denoted as L1 and L2
respectively. These signals include timing patterns relative to the
satellite's onboard precision clock (which is kept synchronized by
a ground station) as well as a navigation message giving the
precise orbital positions of the satellites, an ionosphere model
and other useful information. GPS receivers process the radio
signals, computing ranges to the GPS satellites, and by
triangulating these ranges, the GPS receiver determines its
position and its internal clock error.
In standalone GPS systems that determine a receiver's antenna
position coordinates without reference to a nearby reference
receiver, the process of position determination is subject to
errors from a number of sources. These include errors in the GPS
satellite's clock reference, the location of the orbiting
satellite, ionosphere induced propagation delay errors, and
troposphere refraction errors. The overall positional signal is
weakened with each satellite target lost. These targets may be lost
due to obstructions such as trees, hills, or merely because the
satellite has orbited out of view.
To overcome these positioning errors of standalone GPS systems,
many positioning applications have made use of data from multiple
GPS receivers. Typically, in such applications, a reference or base
receiver, located at a reference site having known coordinates,
receives the GPS satellite signals simultaneously with the receipt
of signals by a remote or rover receiver. Depending on the
separation distance between the two GPS receivers, many of the
errors mentioned above will affect the satellite signals equally
for the two receivers. By taking the difference between signals
received both at the reference site and the remote location, these
errors are effectively eliminated. This facilitates an accurate
determination of the remote receiver's coordinates relative to the
reference receiver's coordinates. Additional sensors may also be
used to support weak GNSS positional data, such as an inertial
measurement unit which may include a gyroscope. Such additional
sensors are, however, prone to lose calibration and then need to be
corrected.
Differential global navigation satellite system (DGNSS) guidance
utilizes a localized base receiver of known location in combination
with a rover receiver on a moving vehicle for obtaining accurate
vehicle positions from GNSS data. Differential positioning, using
base and rover receivers, provides more accurate positioning
information than standalone systems because the satellite ranging
signal transmission errors tend to effect the base and rover
receivers equally and therefore can be cancelled out in computing
position solutions. In other words, the base-rover position signal
"differential" accurately places the rover receiver "relative" to
the base receiver. Because the "absolute" geo-reference location of
the fixed-position base receiver is precisely known, the absolute
position of the rover receiver can be computed using the base
receiver known, absolute position and the position of the rover
receiver relative thereto.
Differential GPS is well known and exhibits many forms. GPS
applications have been improved and enhanced by employing a broader
array of satellites such as GNSS and WAAS. For example, see
commonly assigned U.S. Pat. No. 6,469,663 to Whitehead et al.
titled Method and System for GPS and WAAS Carrier Phase
Measurements for Relative Positioning, dated Oct. 22, 2002, the
disclosures of which are incorporated by reference herein in their
entirely. Additionally, multiple receiver DGPS has been enhanced by
utilizing a single receiver to perform differential corrections.
For example, see commonly assigned U.S. Pat. No. 6,397,147 to
Whitehead titled Relative GPS Positioning Using A Single GPS
Receiver With Internally Generated Differential Correction Terms,
dated May 28, 2002 the disclosures of which are incorporated by
reference herein in their entireties.
It is not uncommon to utilize a GNSS system in combination with an
automatic-steering module linked to a vehicle's steering manifold
through a steering controller unit. The guidance unit receives
positional information from the GNSS unit and compares it with a
pre-planned path or map. Because the GNSS positional information
allows the guidance unit to know exactly where the vehicle is
located along a path, it can use this information to automatically
guide and steer the vehicle along this path.
A steering controller is required to accept instructions from the
guidance unit and actually perform the steering controls on the
vehicle. This device connects to the vehicle steering manifold
and/or hydraulic steering valves. Signals from the guidance unit
are delivered to the steering controller, which then commands
hydraulic valves to open or close depending on the desired
results.
Automatic steering systems using GNSS data tend to lose accuracy.
If the system calibration is off the steering controller may tend
to over-correct, resulting in erratic turns. Additionally, loss of
the GNSS signal could affect the automatic steering function.
SUMMARY OF THE INVENTION
Disclosed herein is a method for providing accurate and precise
vehicle positioning guidance and control with automatic steering
capabilities. The present invention utilizes a series of separate
sensors which may serve as temporary reliable guidance devices when
GNSS signals are weak, and are recalibrated when GNSS signals are
strong. This reliable positioning information gathering allows
multiple vehicles to operate in cooperation with each other using
autonomous task delegation and control. A versatile system is
described that facilitates a number of precise steering tasks for a
variety of functions using proportional hydraulic control and
state-of-the-art GNSS positional systems.
BRIEF DESCRIPTION OF THE DRAWINGS
In the accompanying drawings, which illustrate the principles of
the present invention and an exemplary embodiment thereof.
FIG. 1 is an isometric view of a tractor demonstrating the
preferred embodiment.
FIG. 1A is an isometric view of a tractor demonstrating the three
axes of orientation (X, Y and Z) and three possible directions of
rotation (pitch, roll, and yaw).
FIG. 2 is an alternative line diagram demonstrating the
relationship between devices in an embodiment of the invention.
FIG. 3 is an alternative line diagram demonstrating the
relationship between devices in an embodiment of the invention.
FIG. 4 is an alternative line diagram demonstrating the
relationship between devices in an embodiment of the invention.
FIG. 5 is an alternative line diagram demonstrating the
relationship between devices in an embodiment of the invention.
FIG. 6 is an alternative line diagram demonstrating the
relationship between devices in an embodiment of the invention.
FIG. 7 is an alternative line diagram demonstrating the
relationship between devices in an embodiment of the invention.
FIG. 8 is an alternative line diagram demonstrating the
relationship between devices in an embodiment of the invention.
FIG. 9 is an alternative line diagram demonstrating the
relationship between devices in an embodiment of the invention.
FIG. 10 is a line diagram demonstrating the step-by-step method by
which the sensor suite determines confidence levels of various
sensors.
FIGS. 11A-G demonstrate various path-finding, path-creating, and
object avoidance possibilities available when a tractor is equipped
with the present invention.
DETAILED DESCRIPTION OF THE PREFERRED ASPECTS
I. Introduction, Environment, and Preferred Embodiment
Generally, a preferred embodiment of the present invention consists
of components which allow a farming vehicle, with or without an
attached farming implement, to automatically guide itself around a
field and perform a plurality of functions, leading to precision
farming. Said vehicle may be in constant communication with other
vehicles similarly equipped for the same or different tasks. The
vehicles within such a network are capable of making decisions
amongst themselves about where to go and what to do to best perform
assigned tasks based on the global position of each vehicle
relative to each other and the location of said tasks.
The preferred embodiment or the integrated multi-sensor guidance
system (guidance system) 2, as shown in FIG. 1, includes a vehicle
4, which may be equipped with a farming implement 6, a sensor suite
7, a guidance unit 10 capable of versatile path guidance, a
steering controller 12 providing proportional hydraulic control, a
hydraulic steering manifold 14, a wheel angle/speed sensor (WAS)
16, and an implement steering manifold 50. Additionally, the
guidance system 2 includes a base station antenna 18 to communicate
with a differential receiver 20, a GNSS receiver 22 connected to a
plurality of antennas 24 located on the vehicle 4, and an inertial
measurement unit (IMU) 26 for providing additional information to
said guidance unit 10. Also included is a user interface 28 within
the cab of the vehicle 4 allowing the driver of that vehicle to
manually input commands into the guidance system 2 and override
automatic control.
The preferred embodiment of the present invention has at least four
particular applications. First, there is a command center approach
that can be applied, where the guidance system is a one-time
capital investment that can be moved and used with each piece of
farming equipment, regardless of the season or the task being
performed. Second, a highly accurate yet economical automatic
steering application is available. Such an application can allow
for high accuracy work to be performed 24 hours a day, 7 days a
week with limited stress on human drivers. The third particular
application of the present invention deals with sectional control
of implements; that is the guidance unit can selectively shut off
portions of the working implement where overlap would otherwise
occur. Finally, site-specific farming using variable rate control
can be applied. Depending on the site and the crop being grown, the
system can fluctuate how much work the implement does, whether that
be spraying, seeding, or tilling.
FIGS. 11A-G demonstrate the versatility of the automatic steering
capabilities of the present invention on straight, contour, and
circle pivot driving paths. These figures demonstrate a vehicle
being self-driven around a series of different path-types 64 which
are automatically generated by the guidance system 10 of the
preferred embodiment due to information both manually input and
gathered by the sensor suite 7. These figures demonstrate how the
guidance system 2 will recognize field borders 62 and obstacles 66.
When the planned path 64 encounters the obstacles 66, the system
will either automatically create an alternative path 72 or return
manual control to the vehicle driver. When the planned path 64
encounters the field border 62, the system will automatically shut
off all implement controls and either perform an automatic turn in
the headlands 68 or return manual vehicle control to the vehicle
operator.
II. Sensor Suite
The sensor suite 7 is comprised of a plurality of sensors,
including at least a GNSS system 8, a wheel angle sensor (WAS) 16
and an inertial measurement unit (IMU) 26. Additional sensors may
include a video camera unit oriented in the vehicle towards the
direction of travel. For example, the video camera unit can be
oriented towards a landmark on the horizon, which can provide an
aiming point or point of reference corresponding to a predetermined
geo-reference location. Other sensors in the sensor suite 7 can
include a radar unit for ranging and direction finding, e.g., to a
particular radar target. A laser unit, radio input, telemetry, and
other sensor units capable of aiding in precision position and
trajectory mapping can also be utilized. This suite of sensors
gathers position and heading data and relay this information to the
guidance unit 10 discussed in detail in section III.
In the preferred embodiment of this invention, the GNSS system 8
will be assigned the highest confidence level as a default, and is
thus a primary and important element to this guidance system 2.
Global navigation satellite systems (GNSS) are broadly defined to
include GPS (U.S.), Galileo (proposed), GLONASS (Russia),
Beidou/Compass (China, proposed), IRNSS (India, proposed), QZSS
(Japan, proposed) and other current and future positioning
technology using signals from satellites, with or without
augmentation from terrestrial sources. Inertial navigation systems
(INS) include gyroscopic (gyro) sensors, accelerometers and similar
technologies for providing output corresponding to the inertia of
moving components in all axes, i.e. through six degrees of freedom
(positive and negative directions along transverse X, longitudinal
Y and vertical Z axes). Yaw, pitch and roll refer to moving
component rotation about the Z, X and Y axes respectively. Said
terminology will include the words specifically mentioned,
derivatives thereof and words of similar meaning.
Disclosed herein in an exemplary embodiment is a sensor system for
vehicle guidance. The sensor system can utilize a plurality of GNSS
code or carrier phase differenced antennas to derive attitude
information, herein referred to as a GNSS attitude system.
Moreover, the GNSS attitude system may optionally be combined with
one or more rate gyro(s) used to measure turn, roll or pitch rates
and to further calibrate bias and scale factor errors within these
gyros. In an exemplary embodiment, the rate gyros and GNSS
receiver/antenna are integrated together within the same unit, to
provide multiple mechanisms to characterize a vehicle's motion and
position to make a robust vehicle steering control mechanism.
The preferred embodiment of the present invention includes a
vehicle 4, an implement 6, and a sensor suite 7. The sensor suite
is comprised of a plurality of sensors, containing at least a GNSS
system 8, a WAS 16, and an IMU 26. Said GNSS system 8 is further
comprised of a receiver 22, a differential receiver 20, a base
station antenna 18, and a plurality of antennas 24 located on said
vehicle 4 and implement 6. The GNSS system provides position
information to the guidance unit 10. This information can be used
or creating a path 64 around a field 60, establishing alternatives
72 to said path when obstacles 66 are encountered.
The sensor suite 7 will integrate all connected sensors with the
ultimate result being robust tight wheel control; that is, wheel
and vehicle control at a very precise level. This sensor
integration implements a confidence level or reliance level
checklist by which certain sensors are given higher-priority when
position information is used unless those sensors are reporting
weak or no signal. Higher priority sensor systems are used to
recalibrate lower priority systems while said higher priority
systems remain at their default signal levels. This ensures that
when the higher priority systems lose signal, the lower priority
systems are timely calibrated to compensate for the higher priority
system for the short time period of reduced signal.
III. Guidance Unit 10
A guidance unit 10, otherwise known as an electronic control unit
(ECU), can be put to several different uses on an agricultural
vehicle. One common use is to provide heading data based on a
pre-planned or calculated path 64. The guidance unit might have the
path manually input into the unit, or it might be capable of
receiving GNSS positional data and information regarding a
particular piece of land and calculate a path based off of this
information. The guidance unit 10 can display information to the
vehicle's driver through a user interface (UI) 28 and allow the
driver to manually steer the vehicle along the displayed path. A
more precise application of such a guidance unit 10 is to introduce
automatic steering to a farming vehicle 4. The vehicle 4 will then
guide itself along said calculated or pre-planned path 64 with
greater precision than manual steering could provide.
The guidance unit 10 can be put to additional uses as well,
including automated implement control and advanced mapping and data
management. The automated implement control comprises sectional
implement control, including application rate control and variable
rate control. The advanced mapping and data management, as
mentioned above, includes the system's ability to take known
landscape information from the GNSS system and store that
information for processing during jobs. This leads to real-time map
creation as the vehicle self-guides the piece of land to be
worked.
The preferred embodiment of the present invention includes the
sensor suite 7 mentioned above which is connected to the guidance
unit 10. The guidance unit 10 interprets positional data received
from the sensor suite 7 and puts it to use in several ways. The
guidance unit 10 is further divided into at least a logic portion
30 and a guidance portion 32. The guidance unit receives data from
the sensor suite 7, determines what to do with the data in the
logic portion 30, including computing a path 64 or selectively
controlling the implement, and then transmits that data through the
guidance portion 32 to the steering controller 12 and the implement
steering controller 50.
As demonstrated in FIG. 10, a confidence loop 100 is employed by
the guidance unit 10 against multiple sensors in the sensor suite 7
to determine which sensor systems should be relied on when
determining position and heading information. The confidence loop
100 is comprised of several steps. The start step 102 is initiated
when the guidance system 2 is booted up. This can either be
directly connected to the start-up of the vehicle 4 to which the
system is attached, or completely independent of that vehicle. The
system is then initialized at 104 and a default process 106 is
begun. In this default process, the sensor systems are placed in a
default reliance list whereby a particular sensor is given a higher
confidence than other sensors and that high-confidence sensor is
used to calibrate all other sensors. This highest confidence sensor
is also used for initial position and heading information gathering
and to instruct the guidance unit 10. In a preferred embodiment of
the invention, the GNSS system 8 could delimit as the initial
highest confidence system, followed by the IMU 26 and the WAS
16.
Once the default process 106 is begun, the loop 100 begins a sensor
signal check 108. During this step, each sensor's signal is checked
internally to determine whether it is communicating properly with
the rest of the guidance system 2 and whether incoming signals are
present. For example, the GNSS system 8 will be checked several
times per second to determine the strength of the satellite signal
being received by the antennas 24 and receiver 22. These sensor
signal levels are then compared 110 with the default signal levels
that are expected. If these detected signals are equal to or exceed
the strength of the expected signal, a "yes" command is entered and
the sensor signal check begins again.
If, however, the detected signal is lower than the expected default
signal, a "no" command is reported and the loop 100 enters a
confidence level reduction step 112 whereby the particular sensor's
confidence level is reduced according to the strength of the
detected signal. A confidence level comparison step 114 is then
performed, comparing the updated confidence levels of all sensors
in the sensor suite 7. If the result of the sensor-reliance
reordering step 116 is a change in reliance levels, a "yes" command
is returned and the reliance priority list is reordered at 118.
This occurs when the confidence level of a particular sensor drops
so low due to a weak or loss of signal that its information is no
longer reliable. That sensor drops down in the reliance list and
the new most reliable sensor is used to produce position and
heading information until a sensor signal check 108 results in the
original sensor regaining its signal and thus priority level. If
the result of the sensor-reliance reordering step 116 is "no," then
the reliance list is not reordered and the confidence loop 100
returns to the sensor signal checking step 104.
This process of steps ensures that only the most reliable sensors
are used to determine current vehicle position and heading and to
recalibrate less reliable sensors. The listed steps are an example
of such a confidence loop 100 and are not intended to be the only
means to achieve the desired results. Additional or fewer steps may
be used to return an appropriate confidence or reliance level
list.
As an example of this process, the guidance unit 10 is connected to
the steering controller 12 and the WAS 16. The guidance unit can
relay correction information from the GNSS positioning system 8 to
the WAS for calibration purposes. The WAS 16 is initially
calibrated with a zero-heading and receives information from the
steering controller 12 regarding turn data, and in turn relays
actual data back to the steering controller and the guidance unit.
The guidance unit knows exact position and heading information
because of data received from the GNSS system 8 and other sensors
high on the reliability list. By comparing the highly reliable GNSS
information with the less reliable WAS information, the guidance
unit can tell whether the WAS is correct or not. If it is
determined that the WAS information is incorrect, the guidance unit
can recalibrate the WAS and create a new zero-heading. In the
alternative, if the confidence loop 100 were to determine that the
GNSS system 8 had a weak signal at a particular point, the guidance
unit 10 could rely on data from the IMU 26 and/or WAS 16 until the
GNSS signal returns. These additional sensors are better suited for
short-term accurate guidance, but quickly degrade and must be
recalibrated.
IV. Steering Controller 12
The steering controller 12 is the third major component of the
guidance system 2. The steering controller is designed to accept
guidance inputs and transform those inputs into outputs that result
in actual motion and steering of the vehicle 4.
The steering controller 12 portion of the guidance system 2 is
designed to transmit and receive steering information from all
associated parts and to provide the means for actually controlling
the direction of the vehicle 4 based upon position and guidance
data gathered by the sensor suite 7 and interpreted by the guidance
unit 10. The steering controller is directly connected to the
guidance unit 10, the WAS 16, the hydraulic steering manifold 14,
and the implement controller 50. The steering controller 12 is the
primary step for transforming data from the guidance system into
actual movement of the vehicle itself.
Although the WAS 16 is part of the sensor suite 7 as discussed
above, there is a direct connection between the WAS 16 and the
steering controller 12. This results in a "wheel loop" whereby the
steering controller 12 transmits steering commands to the hydraulic
steering manifold 14 which proceeds to turn the wheels of the
vehicle 4 in a direction. The angle of the turn is reported back to
the steering controller, which may order further steering
corrections depending on the pre-planned path 64. This angle can
also reported to the guidance unit 10 where it is compared with
other sensors in the confidence loop 100. Assuming another sensor,
such as the GNSS system 8, is currently at the top of the reliance
list, the WAS may be recalibrated if it turns out that the applied
turning angle was incorrect when applied to the calculated path
64.
V. Automaton Control
The process of controlling several machines as automatons in a
smart and accurate system, such as the one presented herein, is
accomplished with the combination of the above-described units into
a single, autonomous system allowing one system to control the
positioning, guidance, and workload of a fleet of agricultural
vehicles.
VI. Alternative Examples of a Guidance System 2
The above sections discuss the preferred embodiment of the
invention, comprising generally a sensor suite 7, a guidance unit
10 and a steering controller 12. Several alternative methods of
forming the guidance system 2 exist. A primary example is using the
GNSS system 8 to completely replace the sensor suite 7, and moving
the IMU 26 to the guidance unit 10. Other examples of said guidance
system 2 follow.
As shown in FIG. 3, the IMU 26 and an optional "smart" antenna 34
may be directly connected to the guidance unit 10 providing direct
information used to compare position and heading information with
data received from the sensor suite 7. The smart antenna 34 is a
combination receiver 22 and antenna 24. The user interface (UI) 28
is connected directly to the GNSS system 8. Additionally, the WAS
16 is connected separately and entirely to the steering controller
12.
FIG. 4 provides another detailed breakdown of the guidance unit 10
and its relationship with a GNSS system 8 and the steering
controller 12. The IMU 26 is composed of a plurality of Kalman
filters 36 which relay information regarding the various degrees of
pitch, roll, and heading of the vehicle. The guidance portion 32 is
further composed of a cross-track PID 38, a state dependent Ricatti
equation (SDRE) 40 and a heading proportional integral derivative
(PID) component 42. The steering controller 12 is again in direct
communication with the WAS 16.
FIG. 5 demonstrates another alternative example of the relationship
between the guidance unit 10 and the other elements of the guidance
system 2. In this example the guidance unit 10 is independently
connected to several unique elements, including a GNSS system 8, an
implement control system 50, a variable rate transfer controller
52, personal computer office software 44, and a steering controller
12. The steering controller is separately connected to steering
sensors 46 and the steering interface 48. The steering sensors may
in turn contain WAS 16 or other sensor types. An important aspect
demonstrated in this figure is the relationship between the
guidance unit 10 and cooperative PC office software 44. This
relationship is a key element because it allows the guidance unit
10 to be updated, controlled, and calibrated through a connection
with a standard office PC. This allows the end-user to create
paths, identify field boundaries, and update equipment software
while using familiar PC technology instead of new,
single-application user interfaces associated solely with the
guidance unit.
FIG. 6 demonstrates another alternative example of the present
invention. In this example, the guidance unit 10 contains a
majority of the standard system elements, including the GNSS system
8, a UI 28, a variable rate controller 52, guidance 32 and steering
46 sensors, an accessory input 54, a mapping device 56, and a
section controller 58 containing input/output connections. The
accessory input device 54 allows the guidance controller to connect
to external devices such as a weather radio, wireless service,
video sensors, and monitoring devices. A wireless receiver 22 is
connected to the GNSS 8 portion of the guidance unit 10 externally.
A steering controller 12 is also connected to the guidance 32 and
steering 46 sensors externally. The steering controller has
additional connections to control the vehicle steering manifold
14.
FIG. 8 demonstrates another alternative example of the present
invention. As is typical, the guidance unit 10 is the central
element. Two steering controllers 12 are connected to the guidance
unit, providing options to the guidance unit. A smart antenna 34,
UI 28, and GNSS 8 system are connected to the guidance unit 10
along a first connection, and a second GNSS 8 system, along with a
virtual terminal (VT) and/or an original equipment manufacturer
(OEM) terminal. These two connections again provide options upon
which the guidance unit 10 can make decisions to base path-making
and steering choices on.
FIG. 9 demonstrates another example of the present invention
generally comprising a kit for installation in an existing vehicle,
such as a tractor with a hydraulic steering system. This example is
again divided into three main components; the GNSS system 8, the
guidance unit 10, and the steering controller 12. The GNSS system 8
includes an internal receiver 22 and at least one external antenna
24, along with various input/output connections. A CAN connection
links the GNSS system 8 to the guidance unit 10. The guidance unit
includes an internal IMU and an optional connection to an external
smart antenna 34. The guidance unit 10 connects to the steering
controller 12 through another CAN connection. The steering
controller 12 is connected to and controls the vehicle steering
valve/manifold 14. An analog connection links the WAS to the
steering controller 12. Additionally, several switches are
connected to the steering controller that will cancel auto-steer
and return the vehicle to manual steering, stopping the vehicle
immediately unless the driver is ready to continue. These switches
include, but are not limited to, a steering wheel switch that
detects when the operator's hand touches the steering wheel, a
magnetic shutoff switch that is attached to the operator's seat and
can determine if and when the operator stands to leave the seat,
and a manual shut-off switch.
.Iadd.An integrated multi-sensor guidance system for a vehicle
assembly including a steering subsystem, which guidance system
includes: said vehicle assembly having a dynamic attitude
comprising a geo-reference location, vehicle assembly orientation
and vehicle assembly speed; a processor with multiple sensor inputs
and actuator outputs; a suite of sensor units each connected to a
respective sensor input..Iaddend.
.Iadd.Said sensor unit suite includes a GNSS unit with an antenna
and a receiver connected to said antenna, said GNSS unit providing
output signals corresponding to the GNSS-defined locations of said
vehicle assembly dynamic attitude to a respective processor
input..Iaddend.
.Iadd.A guidance controller is adapted for receiving signal input
and generating control output based on said signal input; a data
storage device including memory storage; and a suite of actuator
units each connected to a respective actuator output..Iaddend.
.Iadd.Said sensor unit suite includes an inertial measurement unit
(IMU) sensor providing output signals corresponding to an inertial
aspect of a dynamic attitude of said vehicle assembly to a
respective processor input..Iaddend.
.Iadd.Said guidance controller is adapted for receiving inertial
measurement signals and integrating said inertial measurement
signals with said GNSS-based positioning signals. Said processor is
programmed to determine variable confidence levels in real time for
each said sensor unit based on its current relative performance;
and said processor is programmed to utilize said sensor unit
outputs proportionally based on their respective confidence levels
in generating said control output signals..Iaddend.
.Iadd.Said processor is programmed to define multiple
behavior-based automatons comprising self-operating entities in
said guidance system, said automatons performing respective
behaviors using data output from one or more sensor units for
achieving said behaviors and wherein one or more sensor units
provide the same or similar data..Iaddend.
.Iadd.Each said automaton has an accepting interface for accepting
requests from other automatons; a requesting interface for making
requests to another automaton; a knowledge input for receiving a
behavioral definition for affecting the behavior of the automatons;
and a data input for receiving input data; and a data output for
sending out the data..Iaddend.
.Iadd.Said actuator unit suite includes a steering unit connected
to said steering subsystem and receiving said control output
signals from said processor. Said steering subsystem includes: a
steering controller including a steering processor and connected to
said guidance controller. Said steering controller receives
guidance signals as inputs from said guidance controller and
computing steering signals as outputs from said steering
controller; and said steering actuator receives said steering
signals from said steering controller and steering said vehicle in
response thereto..Iaddend.
.Iadd.Said sensor suite includes sensor units chosen from among the
group comprising: a video camera unit oriented in the vehicle
assembly direction of travel; a radar unit; a laser unit; radio
input; telemetry; material application exclusion areas input;
satellite image inputs; contour/elevation overlay inputs;
prescription mapping; and a wheel angle sensor (WAS)..Iaddend.
.Iadd.Said actuator suite includes actuator units chosen from among
the group comprising: an implement steering unit, an implement
sectional control unit, personal computer (PC) office software,
material application rate control, secondary vehicle control,
mapping, crop yield, and mapping skips and overlaps..Iaddend.
.Iadd.Said guidance controller is adapted for receiving and storing
in said memory storage device GNSS-based positioning signals. Said
processor is adapted for computing a GNSS-based guide pattern. Said
guidance controller is adapted for providing output signals to a
display device for displaying vehicle motion relative to guide
patterns and contrasting displays of areas treated by said vehicle
along previously-traveled portions of said guide patterns. Said
guidance controller is adapted for calibrating and storing in said
memory multiple vehicle profiles, each said profile including
multiple, independent vehicle-specific automatons..Iaddend.
.Iadd.A method of vehicle control and guidance, comprises the
steps: providing a vehicle assembly including a steering subsystem
and dynamic attitude comprising a geo-reference location, vehicle
assembly orientation, and vehicle assembly speed; providing a
guidance system including processor with multiple sensor inputs and
actuator outputs, a suite of sensor units connected to a respective
sensor input, a suite of actuator units connected to a respective
actuator output, and a data storage device including memory
storage; providing a guidance controller; inputting signal input
data to said guidance controller; and generating control output
signals with said guidance controller based on said signal
input..Iaddend.
.Iadd.The sensor unit suite includes an inertial measurement unit
(IMU) sensor providing output signals corresponding to an inertial
aspect of a dynamic attitude of said vehicle assembly to a
respective processor input. The method of vehicle control and
guidance also includes generating inertial measurement signals with
said IMU sensor; receiving the inertial measurement signals with
said guidance controller; and integrating said inertial measurement
signals with said GNSS-based positioning signals..Iaddend.
.Iadd.The method of vehicle control and guidance also includes
determining variable confidence levels with the processor in real
time for each said sensor unit based on current relative
performance; and utilizing said sensor unit outputs proportionally
based on the respective confidence levels in generating said
control output signals..Iaddend.
.Iadd.The method of vehicle control and guidance also includes
defining multiple behavior-based automatons comprising
self-operating entities in said guidance system; and instructing
said automatons to perform respective behaviors using data output
from one or more sensor units for achieving said behaviors wherein
one or more sensor units provide the same or similar
data..Iaddend.
.Iadd.The method of vehicle control and guidance also includes
providing each automaton with an accepting interface for accepting
requests from other automatons; providing each automaton with a
requesting interface for making requests to another automaton;
providing each automaton with a knowledge input for receiving a
behavioral definition for affecting the behavior of the automatons;
providing each automaton with a data input for receiving input
data; and providing each automaton with a data output for sending
data..Iaddend.
.Iadd.The method of vehicle control and guidance also includes
providing a steering unit connected to said steering subsystem; and
receiving said control output signals at said steering unit as
steering control instructions. The method of vehicle control and
guidance also includes providing a steering processor connected to
said guidance controller; receiving guidance signals at said
steering controller as inputs from said guidance controller;
computing steering signals as outputs from said steering
controller; receiving said steering signals with said steering
actuator; and steering said vehicle assembly in response to said
steering signals..Iaddend.
.Iadd.The method of vehicle control and guidance also includes
determining variable confidence levels with the processor in real
time for each said sensor unit based on current relative
performance; and utilizing said sensor unit outputs proportionally
based on the respective confidence levels in generating said
control output signals..Iaddend.
.Iadd.An integrated multi-sensor guidance system for a vehicle
assembly including a steering subsystem, includes: said vehicle
assembly having a dynamic attitude comprising a geo-reference
location, vehicle assembly orientation and vehicle assembly speed;
a processor with multiple sensor inputs and actuator outputs; a
suite of sensor units each connected to a respective sensor
input..Iaddend.
.Iadd.Said sensor unit suite includes a GNSS unit with an antenna
and a receiver connected to said antenna, said GNSS unit provides
output signals corresponding to the GNSS-defined locations of said
vehicle assembly dynamic attitude to a respective processor input;
said sensor unit suite includes an inertial measurement unit (IMU)
sensor providing output signals corresponding to an inertial aspect
of a dynamic attitude of said vehicle assembly to a respective
processor input..Iaddend.
.Iadd.Said guidance controller is adapted for receiving inertial
measurement signals and integrating said inertial measurement
signals with said GNSS-based positioning signals; said processor is
programmed to determine variable confidence levels in real time for
each said sensor unit based on its current relative
performance..Iaddend.
.Iadd.Said processor is programmed to utilize said sensor unit
outputs proportionally based on their respective confidence levels
in generating said steering signals; a suite of actuator units are
each connected to a respective actuator output; said actuator unit
suite includes a steering unit connected to said steering subsystem
and receiving said steering signals from said
processor..Iaddend.
.Iadd.Said processor is programmed to define multiple
behavior-based automatons comprising self-operating entities in
said guidance system, said automatons performing respective
behaviors using data output from one or more sensor units for
achieving said behaviors and wherein one or more sensor units
provide the same or similar data..Iaddend.
.Iadd.Each said automaton has an accepting interface for accepting
requests from other automatons; a requesting interface for making
requests to another automatons; a knowledge input for receiving a
behavioral definition for affecting the behavior of the automatons;
a data input for receiving input data; and a data output for
sending out the data..Iaddend.
It will be appreciated that the components of the system 2 can be
used for various other applications. Moreover, the subsystems,
units and components of the system 2 can be combined in various
configurations within the scope of the present invention. For
example, the various units could be combined or subdivided as
appropriate for particular applications. The system 2 is scalable
as necessary for applications of various complexities. It is to be
understood that while certain aspects of the disclosed subject
matter have been shown and described, the disclosed subject matter
is not limited thereto and encompasses various other embodiments
and aspects.
* * * * *
References