U.S. patent application number 14/920413 was filed with the patent office on 2016-02-11 for method of controlling a component of a vehicle with a user device.
This patent application is currently assigned to GM GLOBAL TECHNOLOGY OPERATIONS LLC. The applicant listed for this patent is GM GLOBAL TECHNOLOGY OPERATIONS LLC. Invention is credited to Micah R. Jones, Roy J. Mathieu, Joseph F. Szczerba.
Application Number | 20160041562 14/920413 |
Document ID | / |
Family ID | 51792397 |
Filed Date | 2016-02-11 |
United States Patent
Application |
20160041562 |
Kind Code |
A1 |
Mathieu; Roy J. ; et
al. |
February 11, 2016 |
METHOD OF CONTROLLING A COMPONENT OF A VEHICLE WITH A USER
DEVICE
Abstract
A method of controlling a component of a vehicle with a user
device includes receiving a data signal from the user device,
within a vehicle controller. Graphics, corresponding to the data
signal received from the user device, are displayed on a display
screen of the vehicle. A signal is received from the user device
that indicates the user device has received an input into a user
interface of the user device. A determination is made, in the
controller, that the signal received from the user device
corresponds to controlling the component of the vehicle.
Inventors: |
Mathieu; Roy J.; (Rochester
Hills, MI) ; Szczerba; Joseph F.; (Grand Blanc,
MI) ; Jones; Micah R.; (Ferndale, MI) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
GM GLOBAL TECHNOLOGY OPERATIONS LLC |
Detroit |
MI |
US |
|
|
Assignee: |
GM GLOBAL TECHNOLOGY OPERATIONS
LLC
Detroit
MI
|
Family ID: |
51792397 |
Appl. No.: |
14/920413 |
Filed: |
October 22, 2015 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
PCT/US2014/035396 |
Apr 25, 2014 |
|
|
|
14920413 |
|
|
|
|
61816089 |
Apr 25, 2013 |
|
|
|
Current U.S.
Class: |
701/2 |
Current CPC
Class: |
B60R 2300/8093 20130101;
B60Y 2302/03 20130101; B60K 2370/52 20190501; B60K 2370/177
20190501; G02B 6/005 20130101; B60K 2370/334 20190501; B60C 9/00
20130101; B60K 2370/1529 20190501; B60K 2370/25 20190501; B60K
35/00 20130101; G02B 2027/0118 20130101; B60R 2300/8006 20130101;
G02B 2027/0194 20130101; B60W 50/14 20130101; A61B 5/6893 20130101;
B60K 2370/197 20190501; G07C 5/08 20130101; G06K 9/00845 20130101;
G02B 27/0961 20130101; B60Q 9/00 20130101; A61B 5/0533 20130101;
G02B 27/0101 20130101; A61B 5/0205 20130101; B60K 2370/21 20190501;
G05D 3/00 20130101; A61B 3/113 20130101; A61B 5/02405 20130101;
B60W 2050/146 20130101; B60K 2370/785 20190501; G02B 27/0927
20130101; G02B 2027/0112 20130101; G02B 2027/0141 20130101; B60K
28/066 20130101; G08C 17/02 20130101; G07C 5/02 20130101 |
International
Class: |
G05D 3/00 20060101
G05D003/00; G08C 17/02 20060101 G08C017/02 |
Claims
1. A method of controlling a component of a vehicle with a user
device, the method comprising: receiving a data signal from the
user device, wherein the data signal is received within a vehicle
controller; receiving a signal from the user device indicating the
user device has received an input into a user interface of the user
device, wherein the signal is received within the vehicle
controller; determining, in the controller, that the signal
received from the user device corresponds to actuating the
component of the vehicle. transmitting a signal from the vehicle
controller to the component; receiving, by the component, the
signal from the vehicle controller; and actuating the component as
a function of the transmitted signal.
2. A method, as set forth in claim 1, wherein the signal received
from the user device is a function of a sensor reading of at least
one sensor within the user device.
3. A method, as set forth in claim 2, further comprising
determining a vehicle operating condition; wherein transmitting a
signal from the vehicle controller to the component is a function
of the vehicle operating condition and the sensor reading of the at
least one sensor.
4. A method, as set forth in claim 2, wherein transmitting a signal
from the vehicle controller to the component is proportional to the
sensor reading of the at least one sensor within the user
device.
5. A method, as set forth in claim 1, further comprising:
displaying graphics corresponding to the data signal received from
the user device on a display screen of the vehicle; and altering
the display of graphics on the display screen as a function of the
signal received from the user device.
6. A method, as set forth in claim 1, further comprising:
operatively connecting the user device to an interface system of
the vehicle; receiving a signal, in a vehicle controller, from the
user device; and authenticating the user device such that at least
one application is enabled on the user device.
7. A method, as set forth in claim 1, actuating a haptic actuator
within the user device to provide a haptic response through the
user device to indicate the signal received from the user device
corresponds to actuating the component of the vehicle.
8. A method, as set forth in claim 1, wherein actuating the
component is further defined as changing an orientation of the
component as a function of the transmitted signal.
9. A vehicle configured to communicate with a user device, the
vehicle comprising: a component; and an interface system including
a vehicle controller; wherein the vehicle controller is configured
to be in selective communication with a user device, the vehicle
controller is operable for: receiving a data signal from the user
device, wherein the data signal is received within the vehicle
controller; receiving a signal from the user device indicating the
user device has received an input into a user interface of the user
device; determining, in the controller, that the signal received
from the user device, corresponds to actuating the component of the
vehicle; transmitting a signal from the vehicle controller to the
component; and receiving, by the component, the signal from the
vehicle controller; and actuating the component as a function of
the transmitted signal.
10. A vehicle, as set forth in claim 9, wherein the signal received
from the user device is a function of a sensor reading of at least
one sensor within the user device.
11. A vehicle, as set forth in claim 10, wherein the vehicle
controller is further operable for determining a vehicle operating
condition; and wherein transmitting a signal from the vehicle
controller to the component is a function of the vehicle operating
condition and the sensor reading of the at least one sensor.
12. A vehicle, as set forth in claim 10, wherein transmitting a
signal from the vehicle controller to the component is proportional
to the sensor reading of the at least one sensor within the user
device.
13. A vehicle, as set forth in claim 9, wherein the interface
system further includes a display screen and the vehicle controller
is further operable for: displaying graphics corresponding to the
data signal received from the user device on the display screen of
the interface system; and altering the display of graphics on the
display screen as a function of the signal received from the user
device.
14. A vehicle, as set forth in claim 9, wherein the vehicle
controller is further operable for: operatively connecting the user
device to the interface system of the vehicle. receiving a signal,
in a vehicle controller, from the user device; and authenticating
the user device such that at least one application is enabled on
the user device.
15. A vehicle, as set forth in claim 9, wherein the vehicle
controller is further operable for providing haptic feedback
through the user device to indicate the signal received from the
user device corresponds to actuating the component of the
vehicle.
16. A vehicle, as set forth in claim 9, wherein actuating the
component is further defined as changing an orientation of the
component as a function of the transmitted signal.
17. An interface system for controlling a component of a vehicle
with a user device, the interface system comprising: a vehicle
controller configured to be in operative communication with the
component and the user device, the vehicle controller operable for:
receiving a data signal from the user device, wherein the data
signal is received within the vehicle controller; receiving a
signal from the user device indicating the user device has received
an input into a user interface of the user device; determining, in
the controller, that the signal received from the user device,
corresponds to actuating the component of the vehicle; transmitting
a signal from the vehicle controller to the component to actuate
the component.
18. A vehicle, as set forth in claim 9, wherein the interface
system further includes a display screen and the vehicle controller
is further operable for: displaying graphics corresponding to the
data signal received from the user device on the display screen of
the interface system; and altering the display of graphics on the
display screen as a function of the signal received from the user
device.
19. A vehicle, as set forth in claim 9, wherein the vehicle
controller is further operable for: operatively connecting the user
device to the interface system of the vehicle. receiving a signal,
in a vehicle controller, from the user device; and authenticating
the user device such that at least one application is enabled on
the user device.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application is a continuation of International Patent
Application No. PCT/US2014/035396, filed on Apr. 25, 2014, which
claims the benefit of U.S. Provisional Application No. 61/816,089,
filed Apr. 25, 2013, which are hereby incorporated by reference in
their entirety.
TECHNICAL FIELD
[0002] The present disclosure is related to a system and method of
controlling a component of a vehicle with a user device.
BACKGROUND
[0003] Vehicles, such as cars, typically include displays or
indicators to provide information to the vehicle user. Such
displays or indicators may, for example, provide information
regarding mileage, fuel consumption, and vehicle speed. The vehicle
user usually has to shift his eye gaze away from the road scene and
on to an in-vehicle display in order to visually process the
information presented by these displays or indicators. In order to
interact with the displayed information, the user has to utilize
input controls that are built into the vehicle.
SUMMARY
[0004] One aspect of the disclosure provides a method of
controlling a component of a vehicle with a user device includes
receiving a data signal from the user device, within a vehicle
controller. Graphics, corresponding to the data signal received
from the user device, are displayed on a display screen of the
vehicle. A signal is received from the user device that indicates
the user device has received an input into a user interface of the
user device. A determination is made, in the controller, that the
signal received from the user device corresponds to controlling the
component of the vehicle.
[0005] Another aspect of the disclosure provides a vehicle
including a component and an interface system. The interface system
includes a display screen and a vehicle controller. The vehicle
controller is configured to be in selective communication with a
user device. The vehicle controller is operable for receiving a
data signal from the user device. The data signal is received
within the vehicle controller. Graphics are displayed that
correspond to the data signal received from the user device on the
display screen of the interface system. A signal is received from
the user device that indicates the user device has received an
input into a user interface of the user device. A determination is
made in the controller that the signal received from the user
device corresponds to actuating the component of the vehicle.
[0006] Yet another aspect of the disclosure provides an interface
system for controlling a component of a vehicle with a user device.
The interface system includes a display screen and a vehicle
controller. The vehicle controller is configured to be in operative
communication with the component and the user device. The vehicle
controller operable for receiving, by the interface system, a data
signal from the user device. The data signal corresponds to the
execution of at least one software application. Graphics are
displayed that correspond to the data signal received from the user
device on the display screen of the interface system. A signal is
received from the user device indicating the user device has
received an input into a user interface of the user device. A
determination is made, in the controller, that the signal received
from the user device, corresponds to actuating the component of the
vehicle
[0007] The above features and advantages and other features and
advantages of the present teachings are readily apparent from the
following detailed description of the best modes for carrying out
the present teachings when taken in connection with the
accompanying drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
[0008] FIG. 1 is a schematic illustrative side view of a
vehicle.
[0009] FIG. 2 is a schematic diagrammatic view of an interior of a
vehicle having an interface system and a user device.
[0010] FIG. 3 is a schematic view of the vehicle, including a
component and the interface system, illustrating the vehicle in
communication with a user device.
[0011] FIG. 4 is a schematic flow chart diagram of a method of
alerting the user of the vehicle as to a scene, external to the
vehicle, requiring the user's attention.
DETAILED DESCRIPTION
[0012] Those having ordinary skill in the art will recognize that
terms such as "above," "below," "upward," "downward," "top,"
"bottom," etc., are used descriptively for the figures, and do not
represent limitations on the scope of the invention, as defined by
the appended claims. Furthermore, the invention may be described
herein in terms of functional and/or logical block components
and/or various processing steps. It should be realized that such
block components may be comprised of any number of hardware,
software, and/or firmware components configured to perform the
specified functions.
[0013] Referring now to the drawings, wherein the like numerals
indicate corresponding parts throughout the several views, FIG. 1
schematically illustrates a vehicle 10 including a body 12. The
vehicle 10 may be a land vehicle, such as a car, or any other type
of vehicle such as an airplane, farm equipment, construction
equipment, a boat, etc.
[0014] With reference to FIG. 3, the vehicle 10 includes an
interface system 14 and at least one component 16. The interface
system 14 is configured to allow the operative connection between a
user device 18 and a component 16 of the vehicle 10. The interface
system 14 may include a display screen 21 and the vehicle
controller 20.
[0015] The component 16, which is resident within the vehicle 10,
may include, but should not be limited to a head unit, a heads-up
display (HUD), instrument cluster, center display, speakers, a
video screen, air blowers, speedometer, seat motors, door locks,
window motors, window defrost actuators, power doors, include, or
be included in, a head unit, an infotainment system, a navigation
system, an on-board telephone system, a heating, ventilation, and
air conditioning (HVAC) system, and other like devices within the
vehicle 10. The video screen and/or display screen 21 may be
located anywhere within the interior of the vehicle 10, including,
but not limited to, extending from a headliner, rear seat
entertainment displays, side window displays, center front
displays, and any other area capable of receiving a display.
[0016] The component 16 is configured to operatively interact with
the interface system 14. More specifically, the vehicle 10
component 16 may be operatively connected to the vehicle controller
20. The vehicle 10 component 16 may be operatively interconnected
to the vehicle controller 20 (arrow 42) using, a wireless
communication medium, for example, Bluetooth, Wi-Fi, etc., or may
be wired communication medium, for example, a universal serial bus
(USB) or other hardwire cable via hardwiring. It would be
understood that the elements of the component 16 may include, but
should not be limited to, a vehicle interface, while such things as
the operating system 34, the applications, and the like, are
resident within the user device 18.
[0017] The user device 18 may be a portable device that is carried
by the user 24 of the interface system 14, i.e., a user 24 of the
vehicle 10. The component 16 may be a smart phone, a tablet, a
computer, a netbook, an e-reader, a personal digital assistant
(PDA), a gaming device, a video player, a wristwatch, and other
like devices with a at least one sensor and capable of running a
plurality of software applications, either preloaded or downloaded
by the user 24, which may be stored on and executed by the user
device 18 to interact with one or more of the components 16, via
the interface system 14. Examples of the plurality of software
functions may include, but should not be limited to, providing
music, DVD, video, phone, navigation, weather, e-mail, climate
control, seat motor actuation, window motor actuation, window
defrost actuation, and other like applications.
[0018] The user device 18 may include a device memory 26, a
transmitter 28, at least one sensor 29, and a user interface 30,
also referred to as a human machine interface (HMI). The sensors 29
may be one or more of an accelerometer, a touch sensor, a pressure
sensor, a camera, a proximity sensor, a physiological sensor, and
other like sensors. The sensors 29 may be operatively connected to,
or otherwise in operative communication, with the user interface
30. The touch sensor may be configured to sense gestures, while
operatively contacting the touch sensor. Therefore, the touch
sensor may be able to discern gestures corresponding to on/off
and/or directional gestures, including, but not limited to, left
right, up, down, etc. The pressure sensor may be configured to
sense on/off, pressure range, and pressure distribution into the
user interface 30. The temperature sensor may be configured to
sense a temperature in an area of the sensor that may be configured
to enact a change temperature in an area of the vehicle 10. The
camera may be configured to recognize an object, features of an
object, cause a state change, recognize motion to thereby adjust a
feature, and the like. The accelerometer may be configured to sense
a rate of acceleration to thereby bring about, i.e., via the
vehicle controller 20, a corresponding movement of a component in
the vehicle and/or change an orientation of the display screen
within the vehicle 10. The proximity sensor may be configured to
adjust a level within the vehicle. By way of a non-limiting
example, the proximity sensor may be configured to sense light
intensity to bring about, i.e., via the vehicle controller 20, a
dimming of the display screen 21, adjusting a light intensity
within the vehicle 10, and the like. The physiological sensor may
be configured to monitor physiological state thresholds so as to
bring about, i.e., via the engine controller 20, a change in
vehicle operating conditions. By way of a non-limiting example, the
physiological sensor may determine that the user 24 is perspiring
and would, in turn, transmit a signal to the vehicle controller 20
to turn on air conditioning and play classical music through
speakers within the vehicle. It should be appreciated that the user
device may include other sensors 29 and the software of the vehicle
controller 20 may be configured to recognize signals from any type
of sensor 29.
[0019] The user interface 30 may include audio input and/or output,
a keypad, touch screen, a display screen 21, a switch, and/or other
like interfaces. In use, the user 24 may activate the user device
18 by actuating one or more of the user interfaces 30. By way of a
non-limiting example, the user 24 may turn on the air conditioning
within the vehicle 10 by touching the touch pad on the user device
18. As such, the software and sensors 29 configured for controlling
such components 16 is resident within the user device 18 and not
resident within the vehicle 10.
[0020] Further, the vehicle controller 20 may be configured to
receive a signal (arrow 36) in response to input by the user 24
into the user interface 30. Such input into the user interface 30
may be sensed by at least one of the sensors 29 and subsequently be
transmitted to the vehicle controller 20. By way of a non-limiting
example, accelerometers within the user device 18 may be configured
to sense movement of the user device 18. More specifically, the
accelerometers may sense an orientation of the user device 18 as
the user device 18 is tilted and turned. Signals, corresponding to
the orientation of the user device may be transmitted to the
vehicle controller 20. In response, the vehicle controller 20 may
transmit a signal (arrow 42) to a corresponding component 16 to
cause the component 16 to move to mimic or otherwise correspond to
the movement and orientation of the user device 18. The component
may be an outside mirror, a rearview mirror, a display on the
display screen 21, position of a vehicle seat, and the like.
[0021] In another non-limiting example, the pressure sensors within
the user device 18 may be configured to sense the pressure being
applied to the user interface 30. As such, the signal transmitted
from the user device 18 to the vehicle controller 20 may be
proportionate to the amount of pressure applied to the user
interface 30. Therefore, if a large pressure is applied, the signal
transmitted to the vehicle controller may instruct the vehicle
controller to send a corresponding signal to the HVAC blower to
increase a fan speed. Conversely, application of a lesser pressure
to the user interface 30 may cause the signal transmitted to the
vehicle controller to set the fan speed of the HVAC blower to be
proportionately less than the fan speed with the higher pressure.
It should be appreciated that the pressure is not limited to being
able to control fan speed of the HVAC blower, but may also be used
to control temperatures of the HVAC system, seat temperatures, seat
positions, lumbar support, lighting levels, speaker levels,
etc.
[0022] In yet another non-limiting example, the camera and/or the
proximity sensor may be configured to sense a level of light
surrounding the user device 18. As such, the user device 18 may
transmit a signal to the vehicle controller 20 to adjust light
levels within the vehicle 10 and/or exterior to the vehicle 10. It
should be appreciated that the camera and/or proximity sensors may
be used to control other features of the vehicle 20, as well.
[0023] In another example, the display screen 21 may display a menu
of selectable software applications. The user 24 may physically
gesture with their hand, whereby a camera, resident within the user
device 18, recognizes the gesture. A selectable software
application, displayed on the display screen 21 within the
interface system 14, may be highlighted in response to the gesture.
Such a highlighting of the software application on the display
screen 21 may be sufficient to signal the user device 18 to execute
the selected software application. Alternatively, another input
into the user interface 30 may be required to execute the
highlighted software application to run on the user device 18, such
as, a voice command, toggling a switch, touching the touch screen,
and the like.
[0024] The software of the vehicle controller 20 may also be
configured to modify a sensitivity of input gain into the user
interface 30 of the user device 18 as a function of vehicle 10
operating conditions, user 24 preference, driving context, and the
like. By way of a non-limiting example, the graphical display of a
visual menu list, displayed on the display screen 21, may respond
to a gesture at different vehicle velocities and/or vehicle
accelerations. Therefore, the vehicle controller 20 may be
configured to interpret the input signal from the user device 18
such that response from the vehicle controller 20 to the component
16 corresponds to the operating state of the vehicle.
[0025] Once the software application is executed, the vehicle
controller 20 may transmit a signal to the component 16, e.g., via
the transmitter 28, to operate. By way of a non-limiting example,
if the selected software application was music, the vehicle
controller 20 may transmit signals to speakers within the vehicle
10 to play the music.
[0026] The user interface 30 of the user device 18 may also be
configured to provide a haptic feedback to inputs into the user
interface 30. More specifically, haptic response provides tactile
feedback, which takes advantage of the sense of touch by applying
forces, vibrations, or motions to the user 24. By way of example,
the user 24 may move a finger across the touchpad to change an
operating mode on the HVAC controls simulated on the display screen
21 of the interface system 14. As a different HVAC operating mode
is highlighted and/or selected on the display screen 21, the user
interface 30 may vibrate, signaling to the user 24 that the
different operating mode was selected. It should be appreciated
that the haptic feedback may be provided for any desired software
application. The user interface 30 may also be equipped with
tactile sensors that may measure forces exerted by the user 24 on
the user interface 30.
[0027] The user device 18 may include an operating system 34, which
may provide functionality such as authenticating the user device 18
to the component 16 through a handshaking process or other
authenticating process and enabling one or more applications. The
operating system 34 and/or user device 18 may include memory
configured of sufficient size and type to store data and other
information and to store and/or execute the plurality of
applications.
[0028] The vehicle controller 20 may be configured to interact with
the user device 18 through a first communication link (arrow 36).
The first communication link may be a wireless communication
medium, for example, Bluetooth, Wi-Fi, etc., or may be wired
communication medium, for example, a universal serial bus (USB) or
other hardwire cable. A protocol may be used over the communication
link to project graphics, sound, operating instructions, and the
like from the user device 18 to the vehicle 10 component 16. The
user device 18 may also utilize a direct hardware video and/or
audio out signals to project the contents of the user interface 30
of the user device 18 onto the display screen 21 included in the
interface system 14. The display screen 21 may be, for example, a
video display screen 21 configured to display video content, an
electronic visual display configured to display images, and other
like devices for displaying content.
[0029] The user device 18 may further include a communications
interface 38 to selectively communicate with other devices, via a
second communication link (arrow 40), which may include telephones,
portable devices, and one or more off-board (e.g., off vehicle)
servers or systems. The second communication link may be a wireless
communication link in communication with a telecommunications
network or the internet.
[0030] An example of an off-board system may include one or more
service providers, which may be configured as a server located
off-board the vehicle 10, e.g., at a location remote from the
vehicle 10. The off-board server may be a vehicle integrated
service providers, such as the OnStar.RTM. service system, which
may be selectively linked to the vehicle 10 component 16 via the
user device 18. The server may include an operating system 34,
which may provide functionality such as authenticating a device in
communication with the server that may be, for example, the user
device 18 or the component 16, through a handshaking process or
other authenticating process, and enabling one or more
applications. The operating system 34 and/or server may include
memory that is configured of sufficient size and type to store data
and information and store and execute the plurality of
applications. The plurality of applications may include, for
example, phone, voicemail, text messaging, email, navigation, web
browser, message analysis including information feature extraction,
message transcription including voice-to-text transcription using,
for example, automatic speech recognition (ASR), and text-to-speech
(TTS) conversion. The server further includes a communications
interface 38 which may be used to enable interaction between the
user device 18 and/or the vehicle 10 component 16 which may include
sending and receiving data and information including a message
and/or an information feature through the communications link, or
providing other services, such as navigation instructions,
telephone text, email and/or other messaging services.
[0031] One or more servers may be selectively linked to the user
device 18 to, in turn, operate the component 16 of the vehicle 10,
through the vehicle controller 20. For example, a first server 46
may be selectively linked to the user device 18, where the first
server 46 is configured as a service provider or back-end server to
process information features and provide services related thereto
to the vehicle 10. In one example the first server 46 may be
configured as a back-end such as the OnStar.RTM. system. A second
server 48 may be selectively linked to the user device 18 and
configured to receive a message from the vehicle 10 component 16,
and the integration application, and to extract the information
feature(s) from the message and/or transcribe or convert the
message and/or information feature(s).
[0032] The vehicle controller 20 is programmed to provide
communication between the user device 18 and the component 16 via
execution of instructions embodying a method 100, an example of
which is described below with reference to FIG. 3.
[0033] The vehicle controller 20 of FIG. 1 may be embodied as one
or more computer devices having a processor (P) 22 and tangible,
non-transitory memory (M) 25 on which is recorded instructions for
executing the method 100. The memory 25 may include magnetic or
optical memory, electrically-erasable programmable read only memory
(EEPROM), and the like. Additional transitory memory may be
included as needed, e.g., random access memory (RAM), memory for
internal signal buffers, etc. Other hardware of the vehicle
controller 20 may include a high-speed clock, analog-to-digital
(A/D) circuitry, digital-to-analog (D/A) circuitry, and any
required input/output (I/O) circuitry and devices, as well as
signal conditioning and buffer electronics. Individual control
algorithms resident in the vehicle controller 20 or readily
accessible by the vehicle controller 20 may be stored in memory 25
and/or other suitable memory, and automatically executed via the
processor 22 to provide the required control functionality.
[0034] Referring to FIG. 4, an example embodiment of the method 100
begins with step 102, where the vehicle controller 20 receives a
signal (arrow 36) from the user device 18 requesting to connect the
user device 18 to the interface system 14. The signal may
automatically result from the user device 18 being docked in a
docking station within the vehicle 10. Alternatively, the signal
may be the result of the vehicle 10 user 24 affirmatively
activating the user device 18, i.e., actuating the user interface
30. Once the vehicle controller 20 receives the signal, the method
proceeds to step 104.
[0035] At step 104, the vehicle controller 20 may authenticate the
user device 18 to the component 16 through a handshaking process or
other authenticating process and enabling one or more applications.
Once the vehicle controller 20 authenticates the user device 18,
the method proceeds to step 106.
[0036] At step 106, the user device 18 is connected to the
interface system 14. As already described, the connection between
the user device 18 and the interface system 14 may be employed by a
wireless communication medium or a wired communication medium. Once
the user device 18 is connected to the interface system 14, the
method proceeds to step 108.
[0037] At step 108, the controller may receive a data signal (arrow
36) from the user device 18. The data signal may correspond to
graphics, video, etc., to be displayed on the display screen 21 of
the interface system 14. The data signal may correspond to graphics
that may be a menu of an application, which was activated and is
running on the user device 18. Once the controller receives the
data signal, the method proceeds to step 110.
[0038] At step 110, the display screen 21 of the interface system
14 may display the graphics corresponding to the data signal
received from the user device 18. By way of example, the display
screen 21 may display a menu including a plurality of choices for
music channels available for selection. The method then proceeds to
step 112.
[0039] At step 112, controller receives a signal from the user
device 18 that the user device 18 has received an input (arrow 44)
into the user interface 30. Such an input into the user interface
30 may include, but should not be limited to, an audio input,
keypad actuation, touch screen actuation, switch actuation,
activation of accelerometers within the user device 18, and the
like. The method proceeds to step 114.
[0040] At step 114, the display screen 21 alters the display as a
function of the signal received from the user device 18. In
continuing with the exemplary menu for the music channels, provided
above, the display screen 21 may highlight a selected music channel
on the menu in response to receiving the signal from the user
device 18. The method then proceeds to step 116.
[0041] At step 116, the controller determines whether the signal
received from the user device 18 corresponds to activating one or
more components 16 of the vehicle 10. If the vehicle controller 20
determines that the signal does not correspond to activating one or
more components 16, the method returns to step 114. However, if the
vehicle controller 20 determines the signal does correspond to
activating one or more components 16, the method proceeds to step
118.
[0042] At step 118, the vehicle controller 20 transmits a signal to
the corresponding component 16 of the vehicle 10. The method then
proceeds to step 120.
[0043] At step 120, the component 16 receives the transmitted
signal.
[0044] It should be appreciated that the method is not limited to
using the applications and components 16 described herein. Other
applications may be executed and/or other components 16 may be
actuated, so long as the applications remain resident within, and
are executed by, the user device 18.
[0045] While the best modes for carrying out the many aspects of
the present teachings have been described in detail, those familiar
with the art to which these teachings relate will recognize various
alternative aspects for practicing the present teachings that are
within the scope of the appended claims.
* * * * *