U.S. patent application number 16/144789 was filed with the patent office on 2019-01-31 for 3d sound positioning with distributed sensors.
This patent application is currently assigned to BRAGI GmbH. The applicant listed for this patent is BRAGI GmbH. Invention is credited to Peter Vincent Boesen, Darko Dragicevic.
Application Number | 20190029571 16/144789 |
Document ID | / |
Family ID | 62020711 |
Filed Date | 2019-01-31 |
United States Patent
Application |
20190029571 |
Kind Code |
A1 |
Boesen; Peter Vincent ; et
al. |
January 31, 2019 |
3D Sound positioning with distributed sensors
Abstract
A method of providing audiometric feedback from a network of
distributed body sensors using one or more earpieces includes
receiving signals from the network of distributed body sensors at
the one or more wireless earpieces, processing the signals received
at the one or more wireless earpieces to determine a location of
individual body sensors within the network of distributed body
sensors relative the one or more earpieces, and producing
audiometric feedback at the one or more wireless earpieces at least
partially based on the locations of the individual body sensors
relative to the one or more earpieces.
Inventors: |
Boesen; Peter Vincent;
(Munchen, DE) ; Dragicevic; Darko; (Munchen,
DE) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
BRAGI GmbH |
Munchen |
|
DE |
|
|
Assignee: |
BRAGI GmbH
Munchen
DE
|
Family ID: |
62020711 |
Appl. No.: |
16/144789 |
Filed: |
September 27, 2018 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
15799623 |
Oct 31, 2017 |
10117604 |
|
|
16144789 |
|
|
|
|
62416587 |
Nov 2, 2016 |
|
|
|
Current U.S.
Class: |
1/1 |
Current CPC
Class: |
A61B 5/12 20130101; H04R
2201/107 20130101; G06F 3/017 20130101; A61B 5/7415 20130101; H04R
1/1091 20130101; A61B 2560/0242 20130101; G06F 3/165 20130101; A61B
5/6817 20130101; H04R 1/1016 20130101; H04R 1/105 20130101; H04R
2420/07 20130101; H04R 1/1066 20130101; A61B 5/486 20130101; H04R
1/1041 20130101 |
International
Class: |
A61B 5/12 20060101
A61B005/12; H04R 1/10 20060101 H04R001/10; G06F 3/01 20060101
G06F003/01 |
Claims
1. A method of locating a distributed body sensor with a network of
distributed body sensors using one or more earpieces comprising:
receiving a sound signal from the individual body sensor within the
network of distributed body sensors at a first wireless earpiece;
receiving the sound signal from the individual body sensor within
the network of distributed body sensors at a second wireless
earpiece; and processing the sound signal at the first and the
second wireless earpieces by a processor of the first and the
second wireless earpieces to determine a location of the individual
body sensor within the network of distributed body sensors relative
to the first and the second earpieces.
2. The method of claim 1, further comprising the step of producing
audiometric feedback at the first or the second wireless earpieces
based on the location of the individual body sensor relative to the
first or the second earpieces.
3. The method of claim 2 wherein the audiometric feedback is based
on changes in the location of the individual body sensor relative
to the first or the second earpieces due to movement.
4. The method of claim 3 wherein the audiometric feedback comprises
an audiometric rhythmic signal synchronized to the movement.
5. The method of claim 4 further comprising the step of encoding
sensor information from the individual body sensor within the sound
signal and decoding the sensor information from the individual body
sensor at the the first or the second wireless earpieces.
6. The method of claim 1 wherein the signal is communicated over
skin of a user of the distributed body sensor and the first or the
second earpieces.
7. The method of claim 1 wherein the distributed body sensor is
associated with an article of clothing.
8. The method of claim 1 wherein at least one of the distributed
body sensors is associated with a band.
9. The method of claim 1 wherein the audiometric feedback is
represented by three-dimensional sound.
10. The method of claim 9 wherein the audiometric feedback is
positioned at a location to be perceived as emanating from a
position of the distributed body sensor.
11. The method of claim 1 wherein the signals are wireless
signals.
12. A method of locating a body sensor from a network of
distributed body sensors using one or more earpieces comprising:
receiving at least a first signal from the body sensor within the
network of distributed body sensors at a first wireless earpiece;
receiving at least a second signal from the body sensor within the
network of distributed body sensors at a second wireless earpiece;
receiving at least a third signal from either the first or second
wireless earpiece at the second or first wireless earpiece; wherein
the first, the second and third signal are received by the first
and the second wireless earpieces at a different time or angle of
incidence; processing at least the first signal, the second signal
and the third signal received at the first or the second wireless
earpieces to determine location of the body sensor within the
network of distributed body.
13. The method of claim 12, further comprising the step of
producing audiometric feedback at the first or the second wireless
earpieces at least partially based on the location of at least one
of the body sensor relative to the first or the second wireless
earpiece.
14. The method of claim 13 wherein the first signal is received at
the first earpiece and has a first arrival time and a first angle
of incidence and the second signal is received at the second
earpiece and has a second arrival time and a second angle of
incidence.
15. The method of claim 14 wherein a processor disposed of within
the first earpiece or the second earpiece determines the location
of at least one of the individual body sensors using the first
arrival time and the first angle of incidence and the second
arrival time and the second angle of incidence.
16. The method of claim 12 wherein the first or the second wireless
earpieces is configured to triangulate the location of at least one
of the individual body sensors.
17. The method of claim 12 wherein the first or the second wireless
earpiece further comprise at least one wireless transceiver
configured to communicate with an external electronic device.
18. The method of claim 13 wherein the audiometric feedback is
positioned at a location to be perceived as emanating from a
position of the body sensor.
19. The method of claim 12 wherein the first or the second wireless
earpieces further comprises a plurality of microphones operatively
connected to the wireless earpieces and a speaker is operatively
connected to the first or the second wireless earpieces.
20. The method of claim 19 wherein the location of at least one of
the individual body sensors is determined by sound intensity
differential.
Description
PRIORITY STATEMENT
[0001] This application is a continuation of U.S. Non-Provisional
patent application Ser. No. 15/799,623, filed on Oct. 31, 2017 and
claims priority to U.S. Provisional Patent Application No.
62/416,587, filed on Nov. 2, 2016, both of which are Titled 3D
Sound Positioning with Distributed Sensors all of which are hereby
incorporated by reference in their entireties.
FIELD OF THE INVENTION
[0002] The present invention relates to wearable devices. More
particularly, but not exclusively, the present invention relates to
earpieces.
BACKGROUND
[0003] What is needed are methods and systems for wearable devices.
Methods and systems which allow for increased functionality for
earpieces when used in combination with other wearable devices.
SUMMARY
[0004] Therefore, it is a primary object, feature, or advantage of
the present invention to improve over the state of the art.
[0005] It is a further object, feature, or advantage of the present
invention to provide audiometric feedback in a three-dimensional
manner from body sensor signals.
[0006] It is a still further object, feature, or advantage of the
present invention to locate a body sensor using a multitude of
different methods.
[0007] According to one aspect, a method of providing audiometric
feedback from a network of distributed body sensors using one or
more earpieces is provided. The method includes receiving signals
from the network of distributed body sensors at the one or more
wireless earpieces, processing the signals received at the one or
more wireless earpieces to determine a location of individual body
sensors within the network of distributed body sensors relative the
one or more earpieces, and producing audiometric feedback at the
one or more wireless earpieces at least partially based on the
locations of the individual body sensors relative to the one or
more earpieces. The signals may be sound signals. The audiometric
feedback may be based on changes in the location of the individual
body sensors relative to the one or more earpieces due to
movement.
[0008] The audio metric feedback may include an audiometric
rhythmic signal synchronized to the movement. The method may
further include encoding sensor information from the individual
body sensors within the sound signals and decoding the sensor
information from the individual body sensors at the one or more
wireless earpieces. Alternatively, the signals may be communicated
over skin of a user of the distributed body sensors and the one or
more earpieces.
[0009] One or more of the distributed body sensors may be
associated with an article of clothing. One or more of the
distributed body sensors may be associated with a band.
[0010] In one implementation, a system includes an earpiece having
an earpiece housing, a processor disposed within the earpiece
housing, a microphone operatively connected to the earpiece housing
and the processor, wherein the microphone is configured to receive
voice commands, a wireless transceiver operatively connected to the
earpiece housing and the processor, wherein the wireless
transceiver is configured to receive a signal from at least one
body sensor, and a speaker operatively connected to the earpiece
housing and the processor, wherein the speaker is configured to
provide audiometric feedback from one or more body sensors in a
three-dimensional manner.
[0011] One or more of the follow features may be included. The
earpiece may comprise a set of earpieces. The set of earpieces may
comprise a left earpiece and a right earpiece. The earpiece housing
may be composed of soundproof materials. The earpiece housing may
be configured to substantially enclose an ear canal. The earpiece
housing may be further configured to substantially fit within the
ear canal. The earpiece housing may also have a sheath attached to
a side proximate to the ear canal. The wireless transceiver of the
left earpiece and the wireless transceiver of the right earpiece
may be configured to triangulate a position of one or more body
sensors from the signal. A gestural control interface may be
operatively connected to the earpiece housing and the processor.
The gestural control interface may be further configured to
reconfigure the processor, the microphone, or the wireless
transceiver in response to a gesture. A sensor may be operatively
connected to the earpiece housing and the processor, wherein the
sensor may be configured to determine the position of one or more
body sensors operatively connected to the user. The sensor may be
configured to sense one or more motions of the user. A bone
conduction microphone may be positioned along a temporal bone and
may be configured to receive body sounds from the user. The
processor may be configured to provide noise cancellation sound at
the speaker, wherein the noise cancellation sound may be configured
to substantially neutralize one or more body sounds via destructive
interference. The speaker may be configured to short out if the
decibel level of the audiometric feedback exceeds a certain
level.
[0012] In another implementation, a method of determining a
position of a body sensor using a set of earpieces having a left
earpiece and a right earpiece includes receiving an audio signal
from the body sensor at a microphone disposed within the left
earpiece and a microphone disposed within the right earpiece,
wherein the signal at the left earpiece has a first arrival time
and a first angle of incidence and the signal at the right earpiece
has a second arrival time and a second angle of incidence,
transmitting the first arrival time and the first angle of
incidence to the wireless transceiver of the right earpiece via the
wireless transceiver of the left earpiece, and triangulating the
position of the body sensor from the first arrival time, second
arrival time, first angle of incidence and the second angle of
incidence using a processor operatively connected to the right
earpiece.
[0013] One or more of the following features may be included. At
least one signal may encode data related to the user. The data may
be position or motion data. Audiometric feedback may be created
from the position or motion data. Audiometric feedback may also be
provided using a speaker operatively connected to the right
earpiece, and the audiometric feedback may be provided in a
three-dimensional manner.
[0014] In another implementation, a method of determining a
position of a body sensor using a plurality of microphones
operatively connected to an earpiece includes transmitting a sound
signal using a speaker operatively connected to the earpiece,
receiving a reflected sound signal at the plurality of microphones,
and determining the position of the body sensor from a sound
intensity differential between the reception of the reflected sound
at a first microphone and the reception of the reflected sound at a
second microphone and a time delay between the reception of the
reflected sound at the first microphone and the reception of the
reflected sound at the second microphone.
[0015] One or more of the following features may be included. The
sound signal may be transmitted through the user. The sound signal
may be modulated. The reflected sound signal may be correlated with
the sound signal using pulse compression.
[0016] In another implementation, a method of providing audiometric
feedback from a body sensor using an earpiece includes receiving a
signal from the body sensor, processing information encoded in the
signal using a processor disposed within the earpiece, and
producing audiometric feedback derived from the information encoded
in the signal in a three-dimensional manner.
[0017] One or more of the following features may be included. Body
sounds may be received via a bone conduction microphone operatively
connected to the earpiece. Noise cancellation sound may be provided
using a processor operatively connected to the earpiece, wherein
the noise cancellation sound is configured to substantially
neutralize the body sounds via destructive interference. The
information encoded in the signal may comprise kinematic
information. The audiometric feedback may be synchronized to the
kinematic information. Audiometric feedback derived from the
kinematic information may lead the user.
[0018] One or more of these and/or other objects, features, or
advantages of the present invention will become apparent from the
specification and following claims. No single embodiment needs
provide every object, feature, or advantage. Different embodiments
may have different objects, features, or advantages. Therefore, the
present invention is not to be limited to or by an object, feature,
or advantage stated herein.
BRIEF DESCRIPTION OF THE DRAWINGS
[0019] FIG. 1 illustrates a block diagram of one embodiment of the
system.
[0020] FIG. 2 illustrates a block diagram of a second embodiment of
the system.
[0021] FIG. 3 illustrates a left earpiece and a right earpiece.
[0022] FIG. 4 illustrates a flowchart of a method of providing
audiometric feedback from a body sensor using an earpiece.
[0023] FIG. 5 illustrates a network of distributed sensors in
operative communication with a set of wireless earpieces.
DETAILED DESCRIPTION
[0024] A network of distributed sensors which may be wearable
devices (such as shoes, clothing, and socks) is provided. Sensors
may be placed on the various wearable devices at various positions
within the body. The position of the sensors on the body is
determined. Audiometric feedback may be provided. One example is to
produce an audiometric rhythmic signal which can synchronize to a
person's movement or lead a person's movements.
[0025] FIG. 1 illustrates one embodiment of a system 10 which
includes an earpiece 12 having an earpiece housing 14, a processor
16 disposed within the earpiece housing 14, a microphone 18
operatively connected to the processor 16, a wireless transceiver
20 disposed within the earpiece housing 14 and operatively
connected to the processor 16, and a speaker 22 disposed within the
earpiece housing 14 and operatively connected to the processor 16.
The earpiece housing 14 may be composed of soundproof materials or
any material resistant to shear and strain and may also have a
sheath attached to improve comfort, sound transmission, or reduce
the likelihood of skin or ear allergies. In addition, the earpiece
housing 14 may also substantially encompass the outer opening of
the ear canal of a user to substantially reduce or eliminate
external sounds.
[0026] The processor 16 may be configured to process the signals
received from the body sensor, wherein the signals may encode for
sound, instructions, or information which may be stored within a
data storage device, transmitted to another electronic device, or
provided to the user or a third party via the speaker 22. The
processor 16 may also process inputs received from the microphone
18 or another external electronic device and store the inputs in a
data storage device or communicate them to the user or a third
party via the speaker 22.
[0027] The microphone 18, may be positioned to receive certain
ambient sounds. For example, where body worn devices within a
network of body worn sensors communicate with the earpiece by
emitting sounds, the microphone 18 may be used to detect these
sounds. Such sounds may be emitted at a frequency higher than human
hearing such as over 20 kHz and the microphone may be configured to
detect sounds of such frequencies.
[0028] The wireless transceiver 20 may be configured to receive one
or more signals from a wearable device or another electronic
device. In one embodiment the wireless transceiver 20 may provide
for wireless radio signals such as Bluetooth, BLE, or other types
of radio signals. The signals from each source may be received in
any order and may be received simultaneously. For example, the
wireless transceiver 20 may receive one signal from a body sensor
related to kinetic information regarding one of the user's feet,
receive another signal from an external electronic device which may
encode information related to a topic of interest of the user, and
receive yet another signal from someone trying to call the user.
The signals received by the wireless transceiver 20 may travel
through the air, may be sent as sound signals, or may be
communicated through the user's skin.
[0029] The speaker 22, in addition to being operatively connected
to the processor 16, may be used to provide audiometric feedback in
a three-dimensional manner by reproducing audio generated by the
wireless earpiece. For example, the audio may provide motion
feedback either from one or more body sensors and/or provide
feedback created by the processor 16 from data received from one or
more body sensors the user's feet are not properly aligned or the
user's gait is not optimal. The audiometric feedback may be
provided in such a manner the user interprets the feet-related
feedback as emanating from the ground, or the gait feedback as
emanating from the user's upper legs. This may be performed by
generating three-dimensional sound and setting sound source
locations appropriately to provide the desired effect on the user's
perception of the sound. The speaker 22 may also be used to produce
music, sounds, instructions, or information originating from a
signal of a wearable device, a signal from another electronic
device, the microphone 18 or a data storage device. For example,
the speaker 22 may produce a song, a sound, a podcast, a newscast,
a forecast, or anything else of interest to the user selected from
a menu by the user or third party, which may be prompted by a voice
command or a gesture. The speaker 22 may also communicate a
notification of the choice the user or a third party selected so
the user or third party may change the selection if an error was
made.
[0030] FIG. 2 illustrates another embodiment of the system 10
comprising an earpiece 12 having an earpiece housing 14, a
processor 16 disposed within the earpiece housing 14, at least one
microphone 18 operatively connected to the processor 16, a wireless
transceiver 20 operatively connected to the processor 16, a speaker
22 operatively connected to the processor 16, a sensor 24 which may
comprise a bone conduction microphone 32 and/or an inertial sensor
34, a gesture control interface 26 with at least one emitter 42 and
at least one detector 44 operatively connected to the processor 16,
a transceiver 28 disposed within the earpiece housing 14, a data
storage device 30 operatively connected to the processor 16, and at
least one LED 36, and a battery 40 disposed within the earpiece
housing 14 and operatively connected to various components.
[0031] The earpiece housing 14 may be composed of one or more metal
or plastic materials substantially resistant to straining and
shearing stresses and may also have a sheath attached to improve
comfort, sound transmission, or reduce the likelihood of skin or
ear allergies. In addition, the earpiece housing 14 may also
substantially encompass an outer opening of a user's ear canal to
substantially reduce or eliminate external sounds to further
improve audio transparency and may be also be configured to
substantially fit with the user's ear canal to facilitate audio
transmission. The earpiece housing may be an ear bud style
configuration.
[0032] A processor 16 may be disposed within the earpiece housing
14 and may be configured to process one or more signals from one or
more body sensors, one or more signals from another wearable
device, process voice commands from a user or a third party,
process one or more signals from the wireless transceiver 20,
process one or more signals from the transceiver 28, process one or
more signals originating from the data storage device 30, process
one or more signals from the bone conduction microphone 32, or
process one or more signals from the inertial sensor 34, wherein
the signals may encode for music, newscasts, podcasts, commentary,
instructions, information related to sensor readings, or other
forms of digital media and/or information. The processor 16 may
also, in addition to processing the signals, produce the signals
from the microphone 18, the wireless transceiver 20, the
transceiver 28, the data storage device 30, the bone conduction
microphone 32, or the inertial sensor 34 at the speaker 22. The
processor 16 may also be reconfigured by the user or a third-party
using gestures read by a gestural control interface 26, a voice
command received by one or more microphones 18, or a signal
received by the wireless transceiver 20 or the transceiver 28.
[0033] One or more microphones 18 may be operatively connected to
the earpiece housing 14 and the processor 16 and may be configured
to receive one or more voice commands which may be used to cease,
commence, change, or modify one or more functions of the earpiece
12. For example, a voice command to cease receiving ambient sounds
may be provided by the user or a third party saying, "Cease
reception of outside sounds," or a voice command to play the fifth
song in a playlist may be provided by the user or a third party
saying, "Play song five in playlist," or "Skip to song five." Other
commands may be used to cease, commence, change or modify other
functions of the earpiece 12. In addition, one or more microphones
18 may also be configured to receive ambient sounds from one or
more outside sources, which may originate from the user, a third
party, a machine, an animal, another earpiece, another electronic
device, or even nature itself. The ambient sounds received by the
microphone 18 may include a word, a combination of words, a sound,
a combination of sounds, or any combination. The sounds may be of
any frequency and need not necessarily be audible to the user and
may be used to reconfigure one or more components of the earpiece
12.
[0034] A wireless transceiver 20 may be disposed within the
earpiece housing 14 and operatively connected to the processor 16
and may be configured to, in addition to receiving one or more
signals from a body sensor, receive one or more signals from and
transmit one or more signals to at least one wearable device,
and/or receive one or more signals from one or more external
electronic devices. All the signals may be transmitted to the
processor 16 for further processing. The external electronic
devices the wireless transceiver 20 may be configured to receive
signals from include Bluetooth devices, mobile devices, desktops,
laptops, tablets, modems, routers, communications towers, cameras,
watches, third-party earpieces, earpieces, or other electronic
devices capable of transmitting or receiving wireless signals. The
signals received by the wireless transceiver 20 may encode for
sound, instructions, or information. The wireless transceiver 20
may receive or transmit more than one signal simultaneously.
[0035] A speaker 22 may be operatively connected to the processor
16 and, in addition to being configured to provide audiometric
feedback in a three-dimensional manner, may be configured to
produce one or more ambient and/or non-ambient sounds from one or
more microphones 16 or produce audio output derived from one or
more signals from the wireless transceiver 20, the transceiver 28,
or the data storage device 30. The produced sounds may consist of
musical sounds, non-musical sounds, commentary, instructions,
miscellaneous information, or anything else of interest or
importance to the user.
[0036] One or more sensors 24 may be operatively connected to the
earpiece housing 14 and the processor 16 and may be configured to
sense at least one user motion and may also be configured to read
sounds or motions not ascertainable by other components of the
earpiece 12. For example, a bone conduction microphone 32 may be
configured to receive body sounds from the temporal bone of the
user's skull and transmit the body sounds to the processor 16,
which may then create a noise cancellation sound configured to
substantially neutralize each unique body sound the processor 16
receives using destructive interference techniques. An inertial
sensor 34 may also be employed to ascertain the movement of the
user. For example, the inertial sensor 34 may sense a running speed
of the user or an arm speed of a third party which may be
communicated to the processor 16, which may be used in providing
audiometric feedback related to an athletic or personal goal of the
user. Each sensor 24 may be positioned at any location on the
earpiece housing 14 conducive to receiving information and need not
necessarily be in direct contact with either the user or the
external environment.
[0037] A gesture control interface 26 having at least one emitter
42 and a detector 44 may be operatively connected to the earpiece
housing 14 and the processor 16 and may be configured to allow the
user or a third party to control one or more functions of the
earpiece 12. For example, a menu may be prompted through the use of
a gesture with the gestural control interface 26, which may allow
the user or a third party to listen to a song either stored within
the data storage device 30 or received through the wireless
transceiver 20, listen to a playlist, newscast, podcast, or a
weather report received through the wireless transceiver 20 or
stored within the data storage device 30, obtain information on the
user's current surroundings, or anything else of interest to the
user or a third party, and the aforementioned list is
non-exclusive. The selections may be chosen using one or more
additional gestures or using one or more voice commands from the
user and/or a third party. The types of gestures used with the
gesture control interface 26 to control the earpiece 12 include,
without limitation, touching, tapping, swiping, use of an
instrument, or any combination of the gestures. Touching gestures
used to control the earpiece 12 may be of any duration and may
include the touching of areas not part of the gesture control
interface 26. Tapping gestures used to control the earpiece 12 may
include one or more taps and need not be brief. Swiping gestures
used to control the earpiece 12 may include a single swipe, a swipe
changing direction at least once, a swipe with a time delay, a
plurality of swipes, or any combination.
[0038] A transceiver 28 may be disposed within the earpiece housing
14 and may be configured to receive signals from and to transmit
signals to a second earpiece of the user if the user is using more
than one earpiece. The transceiver 28 may receive or transmit more
than one signal simultaneously. The transceiver 28 may be of any
number of types including a near field magnetic induction (NFMI)
transceiver.
[0039] One or more data storage devices 30 may be operatively
connected to the earpiece housing 14 and the processor 16 and may
be configured to store data or information related to one or more
signals received from a wearable device, body sensor, external
electronic device, or a combination. One or more data storage
devices 30 may also have one or more programs preinstalled which
may be (1) used by the processor 16 in processing one or more
signals, (2) used by the processor 16 in executing one or more
commands to be carried out by one or more components of the
earpiece 12, (3) accessible via a gesture or voice command, or (4)
transmitted to an external electronic device.
[0040] One or more LEDs 36 may be operatively connected to the
processor 16 and may be configured to emit light to convey
information to a user concerning the earpiece 12. The LEDs 36 may
be in any area on the earpiece 12 suitable for viewing by the user
or a third party and may consist of as few as one diode which may
be provided in combination with a light guide. In addition, the
LEDs 36 may be discernable by a human eye or an electronic device
and need not have a minimum luminescence.
[0041] FIG. 3 illustrates a pair of earpieces 12 which includes a
left earpiece 12A and a right earpiece 12B. The left earpiece 12A
has a left earpiece housing 14A. The right earpiece 12B has a right
earpiece housing 14B. The left earpiece 12A and the right earpiece
12B may be configured to substantially encompass an outer opening
of a user's ear canal to substantially prevent external sounds from
reaching the user's ear canal and/or fit within the user's ear
canal to minimize the distance between the speakers and a user's
tympanic membranes. The earpiece housings 14A and 14B may be
composed of metallic materials, plastic materials, or any material
with substantial shear and strain resistance and may also be
configured to be soundproof to improve audio transparency. A
microphone 18A is shown on the left earpiece 12A and a microphone
18B is shown on the right earpiece 12B. The microphones 18A and 18B
may be located anywhere on the left earpiece 12A and the right
earpiece 12B respectively and each microphone may be configured to
receive one or more voice commands. Speakers 22A and 22B may be
configured to communicate audiometric feedback 46A and 46B.
[0042] Various methods may be used for determining a position of a
body sensor using a set of earpieces having a left earpiece and a
right earpiece. For example, a signal from the body sensor may be
received by both a wireless transceiver disposed within the left
earpiece and a wireless transceiver disposed within the right
earpiece. The signal may arrive at each wireless transceiver at
different times and at different angles of incidence. The body
sensors transmitting the signal may either be directly worn by a
user or may be part of another wearable device such as a watch or
item of clothing. The signal transmitted by the body sensor may
encode positional data related to the user (for example, it may
indicate it is worn on a left side of the body or right side of the
body), kinetic or motion data related to the user, data or
information related to persons, animals, or objects adjacent to the
user, terrain data, health data related to the user, weather data,
data or information related to a task or goal of the user or a
third party, or anything else a body sensor is capable of sensing.
In addition, signals from more than one body sensor may be received
by the wireless transceivers. For example, a body sensor located on
the user's foot may sense and transmit a signal with information
concerning the user's speed or gait, another body sensor located on
the user's chest may sense the user's heart rate or objects related
to the user's goal, such as a building along a jogging route also
acts as a checkpoint for the user's progress, and another body
sensor located on the user may relay terrain information of the
user's immediate vicinity. The signals transmitted by the body
sensors may also be transmitted galvanically through the user's
skin and may be received by the wireless transceivers at any point
in time and may be received in any order. Where communicated
through the user's skin, electrical contact areas may be present on
both the wearable devices as well as one or more earpieces. In
addition, the signals may be transmitted through generating audio
transmissions which are outside of the range of a person's normal
hearing but may be captured with microphones within the earpieces.
In addition, signals from two or more body sensors may reach the
wireless transceivers at different times and rates.
[0043] In one embodiment the arrival time and the angle of
incidence of the signal received by the left earpiece may be
transmitted to the right earpiece via wireless transceiver 20.
Alternatively, the arrival time and the angle of incidence of the
signal received by the right earpiece may be transmitted to the
left earpiece via wireless transceiver, or both earpieces may
transmit their signal data to the other earpiece. The transmission
of the arrival time or the angle of incidence to the other
processor may be performed continuously or discretely, and the
information may be transmitted in any order. A processor disposed
within one of the earpieces uses the two arrival times and the two
angles of incidence to determine the position of a body sensor
using triangulation algorithms, direction finding techniques or
otherwise. Other algebraic or geometric techniques may also be
employed to determine the position of the body sensor. The
processor may receive more than one reading from a body sensor or
receive multiple readings from multiple body sensors operatively
connected to the user; in fact, the position of a body sensor may
be a position function when the user is in motion, with time as the
independent variable and position as the dependent variable.
[0044] FIG. 4 illustrates a flowchart of a method of providing
audiometric feedback from a body sensor using an earpiece 300.
First, in step 302, a signal is received from the body sensor at
the earpiece. More than one signal or signals from multiple body
sensors may be received, and the signals may be received
continuously or intermittently. The signals may encode data related
to user position, user motion, user health, tasks, goals, persons,
animals, objects, terrain, weather or anything else a body sensor
may be capable of sensing. For example, a signal may encode data
related to the user's body temperature and heart rate or may encode
lower body motion data for further analysis.
[0045] In step 304, a processor operatively connected to the
earpiece processes data or information encoded in the signal
received from the body sensor. The processor may be any number of
different processors or combinations of processors. More than one
signal or one body sensor may be involved, and the processor may
process the signals it receives in any order. The processor may
create audiometric feedback derived from the data or information
encoded in the signals it receives, wherein the audiometric
feedback may be related to running speed, jogging speed, user gait,
body temperature, heart rate, blood sugar, electrolytes, body
moisture, current temperature, current obstacles, checkpoints,
goals, tasks, or anything else a body sensor may sense or the user
may desire. The processor may also process signals from a data
storage device operatively connected to the processor or an
external electronic device, wherein the signals may encode for
music, sound, instructions, information or other media or
informational-related topics, and may incorporate the signals in
the audiometric feedback. For example, the processor may integrate
instructions to take a rest stop stored in the data storage device
with instructions on how to reach the rest stop into the
audiometric feedback if both the current temperature and the body
temperature exceed a specified value. In addition, the processor
may incorporate certain sounds or music cues into the audiometric
feedback if certain milestones have been met or certain checkpoints
have been reached. The processor may also create noise cancellation
sounds for any bodily-related sounds distracting the user. The
bodily-related sounds may originate from any part of the user's
body and may also include any clothing or jewelry the user is
wearing or any objects the user may be carrying.
[0046] In step 306, the processor produces the audiometric feedback
derived from the encoded information at a speaker operatively
connected to the earpiece. The audiometric feedback may be provided
in a three-dimensional manner. For example, if the user is
listening to an instructional audio detailing how to properly swing
a golf club and the user's feet placement is incorrect or otherwise
suboptimal, the user may hear, "spread your feet six more inches"
as if the feedback originated from the user's feet. In addition,
the audiometric feedback may be produced in such a manner as to
synchronize with the user's movements or lead the user. For
example, if the user is listening to an instructional dancing
audio, three-dimensional audiometric feedback may be provided
synchronizing with the dance movements. In addition, if the user is
listening to a workout routine which is configured to operate with
one or more body sensors present on the user's body, sound cues may
be provided to lead the user as to proper body movement or proper
body mechanics to optimize the workout. Additional feedback, music,
sound, instructions, information, or other media-related output
unrelated to a body sensor may also be produced at the speaker, and
the speaker may short out if the sound level at any point exceeds a
certain sound intensity threshold.
[0047] FIG. 5 illustrates a pictorial representation of a network
of body sensors associated with body worn or wearable devices 204
which may include smart socks 200, articles of clothing, jewelry,
watches with watch bands 202, fitness bands, or any other number of
different body worn items. As shown in FIG. 5, the set of wireless
earpieces 10 may communicate with each individual body sensor
within the network of body sensors. The communication may be
one-way or bi-directional depending upon the embodiment. This
communication may take place in various ways.
[0048] In one embodiment, audio may be generated at the body worn
or wearable devices which is detected at the wireless earpieces 20.
The audio may be outside a range of normal human hearing such as
above 20 kHz. Alternatively, this communication may take place such
as by communicating data through the skin such as by method
disclosed in U.S. Pat. No. 6,754,472 to Williams et al., hereby
incorporated by reference in its entirety. Where communication
takes place through the skin each of the wearable devices may have
contact areas for electrically connecting with the skin. In other
embodiments, this communication may take place through radio
communications. Where this communication takes place through radio
communications, it is preferred both a left earpiece and a right
earpiece be used with each of the left earpiece and the right
earpiece capable of communication with the wearable devices. Signal
strength and other signal characteristics may be used to identify
relative location of the wearable devices to the wireless
earpieces. For example, a sock with a stronger signal strength with
a left earpiece than a right earpiece may be indicative the sock is
being worn on a left foot instead of a right foot. Of course,
direction finding, triangulation, and/or other techniques may be
applied to locate body worn devices relative to one or more
wireless earpieces.
[0049] The audiometric feedback produced may be in a
three-dimensional manner. Thus, different perceived sound sources
210 may be placed at different locations within a three-dimensional
sound space as shown. Thus, for example, audiometric feedback may
be perceived as being reproduced from locations. For example,
audiometric feedback may be perceived as being reproduced from a
location of the sensor for which the audiometric feedback is
provided.
[0050] Therefore, various method, system and apparatus have been
shown and described. Numerous options, variations, and alternatives
are contemplated. The present invention is not to be limited to the
specific embodiments shown and described herein.
* * * * *