U.S. patent application number 16/567116 was filed with the patent office on 2021-03-11 for systems and methods for providing and coordinating vagus nerve stimulation with audio therapy.
This patent application is currently assigned to Bose Corporation. The applicant listed for this patent is Bose Corporation. Invention is credited to Kathleen Kremer, Harsh Anilkant Mankodi, John Trotter.
Application Number | 20210069488 16/567116 |
Document ID | / |
Family ID | 1000004362183 |
Filed Date | 2021-03-11 |
![](/patent/app/20210069488/US20210069488A1-20210311-D00000.png)
![](/patent/app/20210069488/US20210069488A1-20210311-D00001.png)
![](/patent/app/20210069488/US20210069488A1-20210311-D00002.png)
![](/patent/app/20210069488/US20210069488A1-20210311-D00003.png)
United States Patent
Application |
20210069488 |
Kind Code |
A1 |
Trotter; John ; et
al. |
March 11, 2021 |
SYSTEMS AND METHODS FOR PROVIDING AND COORDINATING VAGUS NERVE
STIMULATION WITH AUDIO THERAPY
Abstract
A method of providing vagus nerve stimulation to a user is
provided. The method includes the step of prompting, via a first
audio cue emitted by an earpiece worn by a user, the user to inhale
at a first point in time. The method may further include the step
of prompting, via a second audio cue emitted by the earpiece, the
user to exhale at a second point in time. The method may further
include the step of stimulating a vagus nerve of the user with a
first electrical signal generated by the earpiece at the second
point in time. The first electrical signal may be applied to a
concha of an ear of the user. The first electrical signal may be
applied to an ear canal of the user. The first electrical signal
may be applied to a tragus of the user.
Inventors: |
Trotter; John; (Sudbury,
MA) ; Kremer; Kathleen; (Southborough, MA) ;
Mankodi; Harsh Anilkant; (Brighton, MA) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
Bose Corporation |
Framingham |
MA |
US |
|
|
Assignee: |
Bose Corporation
Framingham
MA
|
Family ID: |
1000004362183 |
Appl. No.: |
16/567116 |
Filed: |
September 11, 2019 |
Current U.S.
Class: |
1/1 |
Current CPC
Class: |
A61N 1/0472 20130101;
A61N 1/36038 20170801; A61N 1/36031 20170801; A61N 1/36017
20130101 |
International
Class: |
A61N 1/04 20060101
A61N001/04; A61N 1/36 20060101 A61N001/36 |
Claims
1. A method of providing vagus nerve stimulation to a user,
comprising the steps of: prompting, via a first audio cue emitted
by an earpiece worn by a user, the user to inhale at a first point
in time; prompting, via a second audio cue emitted by the earpiece,
the user to exhale at a second point in time; and stimulating a
vagus nerve of the user with a first electrical signal generated by
the earpiece at the second point in time.
2. The method of claim 1, further comprising the steps of:
prompting, via a third audio cue emitted by the earpiece, the user
to inhale at a third point in time; prompting, via a fourth audio
cue emitted by the earpiece, the user to exhale at a fourth point
in time; and stimulating the vagus nerve of the user with a second
electrical signal generated by the earpiece at the fourth point in
time.
3. The method of claim 2, wherein a time period between the third
and fourth points in time is greater than a time period between the
first and second points in time.
4. The method of claim 1, wherein the first audio cue is an
inhalation recording and the second audio cue is an exhalation
recording.
5. The method of claim 1, further comprising the step of measuring,
via a sensor, a biometric property of the user.
6. The method of claim 1, wherein the second point in time is
adjusted according to the biometric property.
7. The method of claim 2, further comprising the steps of:
prompting, via a fifth audio cue emitted by the earpiece, the user
to inhale at a fifth point in time; prompting, via a sixth audio
cue emitted by the earpiece, the user to exhale at a sixth point in
time; and stimulating the vagus nerve of the user with a third
electrical signal generated by the earpiece at the sixth point in
time.
8. The method of claim 7, wherein a time period between the fifth
and sixth points in time is greater than the time period between
the third and fourth points in time, and wherein a time period
between the third and fourth points in time is greater than a time
period between the first and second points in time.
9. The method of claim 1, further comprising the step of inputting,
by the user via a user interface, one or more personal
characteristics.
10. The method of claim 9, further comprising the step of adjusting
a time period between the first point in time and the second point
in time according to the one or more personal characteristics.
11. A system for providing vagus nerve stimulation to a user,
comprising: an earpiece comprising an audio transducer and a first
electrode, the earpiece configured to: form a first electrical
contact between the electrode and an ear canal, tragus, or concha
of a first ear of a user; emit a first audio cue from the audio
transducer at a first point in time, wherein the first audio cue
prompts the user to inhale; emit a second audio cue from the audio
transducer at a second point in time, wherein the second audio cue
prompts the user to exhale; and deliver a vagus nerve stimulation
signal to the user via the first electrode at the second point in
time; and a second electrode configured to form a second electrical
contact with the user.
12. The system of claim 11, wherein the earpiece further comprises
the second electrode.
13. The system of claim 11, further comprising a sensor configured
to measure a biometric property of the user.
14. The system of claim 11, further comprising a user interface
configured for the user to input one or more personal
characteristics.
15. A method of providing vagus nerve stimulation to a user,
comprising the steps of: delivering, via an earpiece worn by a
user, audio content to a user; and stimulating a vagus nerve of the
user with one or more electrical signals generated by the earpiece
in coordination with the delivery of the audio content.
16. The method of claim 15, wherein the audio content comprises
guided breathing instructions prompting the user to adjust their
respiration rate.
17. The method of claim 16, wherein the guided breathing
instructions comprise audio cues to prompt the user to adjust their
respiration rate, and wherein the step of stimulating the vagus
nerve of the user further comprises stimulating the vagus nerve of
the user in coordination with the audio cues.
18. The method of claim 15, further comprising the step of
measuring, via a sensor, a biometric property of the user.
19. The method of claim 18, further comprising the step of
adjusting the audio content based on the biometric property.
20. The method of claim 19, wherein the audio content comprises
audio cues to prompt the user to adjust their respiration rate, and
wherein playback timing of the audio cues is adjusted based on the
biometric property.
Description
BACKGROUND
[0001] This disclosure generally relates to methods and systems for
providing and coordinating vagus nerve stimulation (VNS) with audio
therapy.
SUMMARY
[0002] This disclosure is directed to methods and systems for
providing and coordinating vagus nerve stimulation (VNS) with audio
therapy.
[0003] In one aspect, a method of providing VNS to a user is
provided. The method includes the step of prompting, via a first
audio cue emitted by an earpiece worn by a user, the user to inhale
at a first point in time. The method may farther include the step
of prompting, via a second audio cue emitted by the earpiece, the
user to exhale at a second point in time. The method may further
include the step of stimulating a vagus nerve of the user with a
first electrical signal generated by the earpiece at the second
point in time. The first electrical signal may be applied to a
concha of an ear of the user. The first electrical signal may be
applied to an ear canal of the user. The first electrical signal
may be applied to a tragus of the user.
[0004] According to an example, the method may further include the
step of prompting, via a third audio cue emitted by the earpiece,
the user to inhale at a third point in time. The method may further
include the step of prompting, via a fourth audio cue emitted by
the earpiece, the user to exhale at a fourth point in time. The
method may further include the step of stimulating the vagus nerve
of the user with a second electrical signal generated by the
earpiece at the fourth point in time. A time period between the
third and fourth points in time may be greater than a time period
between the first and second points in time.
[0005] According to an example, the first and second audio cues may
be verbal commands. The first and second audio cues may be audio
tones. The first audio cue may be an inhalation recording. The
second audio cue may be an exhalation recording.
[0006] According to an example, the method may further include the
step of measuring, via a sensor, a biometric property of the user.
The second point in time may be adjusted according to the biometric
property. The biometric property may be heart rate variability. The
biometric property may be respiration rate.
[0007] According to an example, the method may further include the
step of prompting, via a fifth audio cue emitted by the earpiece,
the user to inhale at a fifth point in time. The method may further
include prompting, via a sixth audio cue emitted by the earpiece,
the user to exhale at a sixth point in time. The method may further
include stimulating the vagus nerve of the user with a third
electrical signal generated by the earpiece at the sixth point in
time. A time period between the fifth and sixth points in time may
be greater than the time period between the third and fourth points
in time. A time period between the third and fourth points in time
may be greater than a time period between the first and second
points in time.
[0008] According to an example, the method may further include the
step of inputting, by the user via a user interface, one or more
personal characteristics. The method may further include the step
of adjusting a time period between the first point in time and the
second point in time according to the one or more personal
characteristics.
[0009] In another aspect, a system for providing VNS to a user is
provided. The system may include an earpiece. The earpiece may
include an audio transducer. The earpiece may include a first
electrode. The earpiece may be configured to form a first
electrical contact between the electrode and an ear canal, tragus,
or concha of a first ear of a user. The earpiece may be further
configured to emit a first audio cue from the audio transducer at a
first point in time. The first audio cue may prompt the user to
inhale. The earpiece may be further configured to emit a second
audio cue from the audio transducer at a second point in time. The
second audio cue may prompt the user to exhale. The earpiece may be
configured to deliver a VNS signal to the user via the first
electrode at the second point in time. The system may further
include a second electrode. The second electrode may be configured
to form a second electrical contact with the user.
[0010] According to an example, the earpiece may further include
the second electrode. Alternatively, the second electrode may be
incorporated into a second earpiece or other device.
[0011] According to an example, the second electrode may be
configured to form the second electrical contract with the ear
canal, tragus, or concha of the first ear of the user. In an
alternative example, the second electrode may form the second
electrical contact with an ear canal, tragus, or concha of a second
ear of the user. In another alternative example, the second
electrode may form the second electrical contact on a back portion
of an earlobe of the first ear.
[0012] According to an example, the system may further include a
sensor configured to measure a biometric property of the user.
[0013] According to an example, the system may further include a
user interface configured for the user to input one or more
personal characteristics.
[0014] In another aspect, an earpiece for providing VNS to a user
is provided. The earpiece may include an audio transducer. The
audio transducer may be configured to emit a first audio cue from
the audio transducer at a first point in time. The first audio cue
may prompt the user to inhale. The audio transducer may be
configured to emit a second audio cue from the audio transducer at
a second point in time. The second audio cue may prompt the user to
exhale. The earpiece may further include a first electrode. The
first electrode may be configured to form a first electrical
contact with an ear canal, tragus, or concha of a user. The first
electrode may be further configured to deliver a vagus nerve
stimulation signal to the user via the first electrode at the
second point in time.
[0015] According to an example, the earpiece may further include a
second electrode configured to form a second electrical contact
with the user.
[0016] According to an example, the earpiece may further include a
sensor configured to measure a biometric property of the user.
[0017] In another aspect, a method of providing VNS to a user is
provided. The method may include the steps of delivering, via an
earpiece worn by a user, audio content to a user. The method may
include stimulating a vagus nerve of the user with one or more
electrical signals generated by the earpiece in coordination with
the delivery of the audio content.
[0018] According to an example, the audio content may include
guided breathing instructions prompting the user to adjust their
respiration rate. The guided breathing instructions may include
audio cues to prompt the user to adjust their respiration rate. The
step of stimulating the vagus nerve of the user may further include
stimulating the vagus nerve of the user in coordination with the
audio cues.
[0019] According to an example, the method may further include the
step of measuring, via a sensor, a biometric property of the user.
The method may further include the step of adjusting the guided
breathing instructions based on the biometric property. The guided
breathing instructions may include audio cues to prompt the user to
adjust their respiration rate. Playback timing of the audio cues
may be adjusted based on the biometric property.
[0020] In another aspect, a system for providing VNS to a user is
provided. The system may include an earpiece. The earpiece may
include an audio transducer. The earpiece may include a first
electrode. The earpiece may be configured to form a first
electrical contact between the electrode and an ear canal, tragus,
or concha of a first ear of a user. The earpiece may be further
configured to deliver audio content to a user. The earpiece may be
further configured to stimulate a vagus nerve of the user with one
or more electrical signals generated by the earpiece in
coordination with the delivery of the audio content. The system may
further include a second electrode configured to form a second
electrical contact with the user.
[0021] According to an example, the audio content may include
guided breathing instructions prompting the user to adjust their
respiration rate. The guided breathing instructions may include
audio cues to prompt the user to adjust their respiration rate. The
earpiece may be further configured to stimulate the vagus nerve of
the user in coordination with the audio cues.
[0022] According to an example, the system may further include a
sensor. The sensor may be configured to measure a biometric
property of the user.
[0023] According to an example, the earpiece may be further
configured to adjust the guided breathing instructions based on the
biometric property. The guided breathing instructions may include
audio cues to prompt the user to adjust their respiration rate.
Playback timing of the audio cues may be adjusted based on the
biometric property.
[0024] In another aspect, an earpiece for providing VNS to a user
is provided. The earpiece may include an audio transducer. The
audio transducer may be configured to deliver audio content to a
user. The earpiece may further include a first electrode. The first
electrode may be configured to form a first electrical contact with
an ear canal, tragus, or concha of a user. The first electrode may
be further configured to stimulate a vagus nerve of the user with
one or more electrical signals generated by the earpiece in
coordination with the delivery of the audio content.
[0025] According an example, the audio content may include guided
breathing instructions prompting the user to adjust their
respiration rate. The guided breathing instructions may include
audio cues to prompt the user to adjust their respiration rate. The
first electrode may be further configured to stimulate the vagus
nerve of the user in coordination with the audio cues.
[0026] According to an example, the earpiece may be further
configured to adjust the guided breathing instructions based on a
biometric property of the user measured by a sensor. The guided
breathing instructions may further include audio cues to prompt the
user to adjust their respiration rate. Playback timing of the audio
cues may be adjusted based on the biometric property.
[0027] Other features and advantages will be apparent from the
description and the claims.
BRIEF DESCRIPTION OF THE DRAWINGS
[0028] In the drawings, like reference characters generally refer
to the same parts throughout the different views. Also, the
drawings are not necessarily to scale, emphasis instead generally
being placed upon illustrating the principles of the various
examples.
[0029] FIG. 1 is a flowchart of an example of a method of providing
VNS to a user.
[0030] FIG. 2 is an example earpiece.
[0031] FIG. 3 is a flowchart of a further example of a method of
providing VNS to a user.
DETAILED DESCRIPTION
[0032] This disclosure is directed to methods and systems for
providing and coordinating vagus nerve stimulation (VNS) with audio
therapy. VNS is a medical treatment that involves delivering
electrical impulses to the vagus nerve of a user. The electrical
impulses may be delivered via an ear canal, tragus, and/or concha
of the user. VNS has been used to treat physiological disorders
such as, for example, depression, insomnia, and tinnitus, among
others. This disclosure relates generally to providing VNS in
conjunction with an audio therapy regimen to help relieve insonmia
or anxiety. The audio therapy may be, but is not limited to, guided
breathing or binaural beats. In one example, the user follows a
guided breathing regimen based on a soundtrack played through a set
of earpieces. The soundtrack may include audio cues for the user to
follow to guide their respiration rate. The earpieces may also
carry electrodes for delivering VNS treatment in or around the ears
of the user. As the user follows the guided breathing audio cure,
VNS may be applied and controlled such that treatment is delivered
based upon the cadence of the audio cues and/or the user's
corresponding respiration rate.
[0033] The examples and implementations disclosed or otherwise
envisioned herein can be utilized with any suitable earpiece.
Examples of suitable earpieces include Bose.RTM. Sleepbuds.TM.
(manufactured by Bose Corporation of Framingham, Mass.), ear tips,
earbuds, in-ear headphones, over-the-ear headphones, noise-blocking
earplugs, and hearing aids. However, the disclosure is not limited
to these devices, and thus the disclosure and embodiments disclosed
herein can encompass any earpiece configured to be placed at least
partially within human ears.
[0034] In one aspect, and with reference to FIG. 1, a method 10 of
providing VNS to a user is provided. The method 10 includes the
step of prompting 12, via a first audio cue emitted by an earpiece
worn by a user, the user to inhale at a first point in time. The
method 10 may further include the step of prompting 14, via a
second audio cue emitted by the earpiece, the user to exhale at a
second point in time. The method 16 may further include the step of
stimulating a vagus nerve of the user with a first electrical
signal generated by the earpiece at the second point in time. By
following this method 10, the VNS signal may be applied
simultaneously (at the second point in time) to the exhalation of
the user for optimum therapeutic impact. In other examples, the
method 10 may be altered to delay stimulation the vagus nerve until
sometime after exhalation.
[0035] The first electrical signal (and any other electrical signal
applied using method 10) may be applied to an ear canal of the
user. The ear canal is considered to be the pathway running from
the outer ear to the middle ear. The first electrical signal may
also be applied to a concha of an ear of the user. The concha is
considered to be the hollow area of the ear next to the ear canal.
The first electrical signal may further be applied to a tragus of
the user. The tragus is considered to be the small pointed eminence
of the external ear, positioned in front of the concha, and
projecting backwards over the meatus. The first electrical signal
may also be applied to other areas of the user through which the
vagus nerve may be stimulated by an electrical signal.
[0036] According to an example, the first and second audio cues may
be verbal commands. The first and second audio cues may be audio
tones. The first audio cue may be an inhalation recording. The
second audio cue may be an exhalation recording. The first and
second audio cue may also be any other sound to aid the user's
guided breathing. The first and second audio cue may be the same or
different, depending on the application. The first and second audio
cue may be pre-recorded audio tracks, or they may be audio tones
generated by the internal circuitry of the earpiece.
[0037] According to an example, the method 10 may further include
the step of prompting 18, via a third audio cue emitted by the
earpiece, the user to inhale at a third point in time. The method
10 may further include the step of prompting 20, via a fourth audio
cue emitted by the earpiece, the user to exhale at a fourth point
in time. The method 10 may further include the step of stimulating
22 the vagus nerve of the user with a second electrical signal
generated by the earpiece at the fourth point in time. A time
period between the third and fourth points in time may be greater
than a time period between the first and second points in time. The
increase in time periods between inhalation and exhalation may be
used for sleep therapy guided breathing. The increase in time
periods may be controlled by an algorithm processing biometric
information of the user.
[0038] Similarly, the method 10 may further include the step of
prompting 24, via a fifth audio cue emitted by the earpiece, the
user to inhale at a fifth point in time. The method 10 may further
include prompting 26, via a sixth audio cue emitted by the
earpiece, the user to exhale at a sixth point in time. The method
10 may further include stimulating 22 the vagus nerve of the user
with a third electrical signal generated by the earpiece at the
sixth point in time. A time period between the fifth and sixth
points in time may be greater than the time period between the
third and fourth points in time. A time period between the third
and fourth points in time is greater than a time period between the
first and second points in time. As the user's respiration rate
decreases over the course of the guided breathing regimen. VNS may
no longer be applied at every exhalation. Further, as the user gets
closer to the optimum relaxation or sleep state, VNS may stop
entirely.
[0039] According to an example, the method 10 may further include
the step of measuring 36, via a sensor, a biometric property of the
user. The second point in time may be adjusted according to the
biometric property. The biometric property may be any property
which would impact the rate of optimum application of VNS. The
biometric property may be heart rate variability. The biometric
property may be respiration rate. For example, as the respiration
rate of the user decreases over time, the rate guided breathing
audio cues and corresponding VNS may similarly decrease. In
addition to the second point in time, any other point in time of
the audio therapy regimen may be adjusted according to the
biometric property.
[0040] According to an example, the method 10 may further include
the step of inputting 32, by the user via a user interface, one or
more personal characteristics. These personal characteristics may
include weight, age, or any other characteristics which would
impact the rate of optimum application of VNS. The method may
further include the step of adjusting a time period between the
first point in time and the second point in time according to the
one or more personal characteristics. In addition to the second
point in time, any other point in time of the audio therapy regimen
may be adjusted according to the biometric property.
[0041] In another aspect, and with reference to FIG. 2, a system 50
for providing VNS to a user is provided. The system may include an
earpiece 100. The earpiece 100 may include an audio transducer 170.
An example audio transducer 170 is shown disposed within hollow
passage 104.
[0042] The earpiece 100 may include a first electrode 120. The
earpiece 100 may be configured to form a first electrical contact
between the electrode 120 and an ear canal, tragus, or concha of a
first ear of a user.
[0043] The earpiece 100 may be further configured to emit a first
audio cue from the audio transducer 170 at a first point in time.
The first audio cue may prompt the user to inhale. The earpiece 100
may be further configured to emit a second audio cue from the audio
transducer 170 at a second point in time. The second audio cue may
prompt the user to exhale. The first and second audio cues may be
verbal commands. The first and second audio cues may be audio
tones. The first audio cue may be an inhalation recording. The
second audio cue may be an exhalation recording. The first and
second audio cue may also be any other sound to aid the user's
guided breathing. The first and second audio cue may be the same or
different, depending on the application. The first and second audio
cue may be pre-recorded audio tracks, or they may be audio tones
generated by the internal circuitry of the earpiece.
[0044] The earpiece 100 may be configured to deliver a VNS signal
to the user via the first electrode 120 at the second point in
time. The system 100 may further include a second electrode 122.
The second electrode 122 is used to complete a circuit with the
first electrode and the stimulated portion of the user's ear. The
second electrode 122 may be configured to form a second electrical
contact with the user. According to an example, and as shown in
FIG. 2 the earpiece 122 may include the both the first electrode
120 and the second electrode 122.
[0045] According to an example, the second electrode 122 may be
configured to form the second electrical contract with the ear
canal, tragus, or concha of the first ear of the user. As shown in
FIG. 2, both the first electrode 120 and the second electrode 122
are positioned to make electrical contact with the same ear canal
of the user when the earpiece 100 is properly worn. A third
electrode 124 may also be incorporated. This third electrode 124
may be configured to form an electrical contact with a different
portion of the ear than the second 122 electrode, such as the
concha or tragus. It should be understood that to properly deliver
a VNS signal to the user, only two electrodes may be active at
once.
[0046] The second electrode 122 may form the second electrical
contact with an ear canal, tragus, and/or concha of a second ear of
the user. In this example, the second electrode may be incorporated
into a second earpiece. In another alternative example, the second
electrode 122 may form the second electrical contact on a back
portion of an earlobe of the first ear. In this example, the second
electrode 122 may be incorporated into the ear cup of a headset.
Alternatively, the electrode 122 may be a discrete electrical
contact which the user attaches to the back of their earlobe.
[0047] According to an example, the system 50 may further include a
sensor 180 configured to measure a biometric property of the user.
The sensor may be embedded into earpiece 100 as shown in FIG. 2.
The sensor 180 may be a discrete device electrically connected to
earpiece 100. The electrical connection may be wired or wireless.
The second point in time may be adjusted according to the biometric
property. The biometric property may be any property which would
impact the rate of optimum application of VNS. The biometric
property may be heart rate variability. The biometric property may
be respiration rate. For example, as the respiration rate of the
user decreases over time, the rate guided breathing audio cues and
corresponding VNS may similarly decrease. In addition to the second
point in time, any other point in time of the audio therapy regimen
may be adjusted according to the biometric property.
[0048] According to an example, the system 50 may further include a
user interface 190 configured for the user to input one or more
personal characteristics, as shown in FIG. 2 (not to scale). The
user interface 190 may be a discrete device, such as a smartphone,
electrically connected to earpiece 100. The user interface 190 may
include one or more buttons, one or more knobs, a processor, a
memory, and/or a display screen. The electrical connection may be
wired or wireless. These personal characteristics may include
weight, age, or any other characteristics which would impact the
rate of optimum application of VNS. The time period between the
first point in time and the second point in time may be adjusted
according to the one or more personal characteristics. In addition
to the second point in time, any other point in time of the audio
therapy regimen may be adjusted according to the biometric
property.
[0049] In another aspect, and with reference to FIG. 2, an earpiece
100 for providing VNS to a user is provided. The earpiece may
include an audio transducer 170 disposed within hollow passage 104.
The audio transducer 170 may be configured to emit a first audio
cue from the audio transducer at a first point in time. The first
audio cue may prompt the user to inhale. The audio transducer 170
may be configured to emit a second audio cue from the audio
transducer 170 at a second point in time. The second audio cue may
prompt the user to exhale. The earpiece 100 may further include a
first electrode 120. The first electrode 120 may be configured to
form a first electrical contact with an ear canal, tragus, or
concha of a user. The first electrode 120 may be further configured
to deliver a vagus nerve stimulation signal at the second point in
time.
[0050] According to an example, the earpiece 100 may further
include a second electrode 122 configured to form a second
electrical contact with the user. According to another example, the
earpiece 100 may further include a sensor 180 configured to measure
a biometric property of the user.
[0051] Turning now to FIG. 2 in more detail, FIG. 2 shows an
example earpiece 100 including an ear tip that is configured to fit
at least partially into a person's ear canal and a retaining
structure configured to hold the ear tip in place when worn. The
following should be viewed in light of FIG. 2. Earpiece 100 can
include body 102, and hollow passage 104, and retaining legs 106.
Although FIG. 2 shows retaining legs 106 as one embodiment of a
retaining structure, this disclosure is not limited to such a
configuration. Any type of retaining structure is contemplated.
Alternatively, the retaining structure can be omitted altogether.
The retaining legs 106 are optional.
[0052] Body 102 includes first end 108 and second end 110 opposite
the first end 108. Second end 110 is proximate to the retaining
legs 106. Body 104 further includes inner wall 107 extending
between the first end 108 and the second end 110. Inner wall 107
defines and surrounds hollow passage 104 which can be configured to
conduct sound waves. Body 102 also includes outer wall 112
connected to the inner wall 107 at the first end 108. Outer wall
112 tapers away from the inner wall 107 toward the second end 110.
In example aspects, outer wall 112 is frustoconical in shape. As
shown in FIG. 2, outer wall 112 tapers toward the second end 110
but does not necessarily reach the second end 110. In alternate
embodiments not shown, outer wall 112 could extend to the second
end 110 or beyond second end 110. Body 102 can be made of any
suitable soft, flexible materials, including, for example,
silicone, polyurethane, polynorbornene (e.g., Norsorex.RTM.
material available from D-NOV GmbH of Vienna, Austria),
thermoplastic elastomer (TPE), and/or fluoroelastomer.
[0053] Earpiece 100 includes electrically conductive elements 120,
122, and 124, which function as electrodes when positioned to
contact skin within the ear, which generally is a stable
environment offering good electrical and/or mechanical contact
between skin and electrodes. Although FIG. 2 shows one particular
configuration of an ear tip with electrically conductive elements
120 and 122 arranged on opposite sides of an outer surface of outer
wall 112 (configured to fit at least partially into a person's ear
canal), and electrically conductive element 124 arranged on body
102, this disclosure is not limited to such a configuration, and
any number of electrically conductive elements can be placed in any
configuration on earpiece 100 as long as the electrically
conductive elements are arranged to contact skin and provide VNS
within the ear when the earpiece is worn by a user. It should be
appreciated that only two electrically conductive elements may be
used to function as electrodes.
[0054] Examples of electrically conductive elements include a metal
pad, metal button, metal foils (e.g., gold, silver), metal-salt
hybrids (e.g., silver/silver-chloride), polymeric composites (e.g.,
rubbers compounded with conductive fillers such as carbon black,
carbon nanotubes, graphene, silver, glass-coated silver),
intrinsically conductive polymers (e.g.,
poly(3,4-ethylenedioxythiophene) polystyrene sulfonate or
PEDOT:PSS), and/or conductive fabric (e.g., fabrics with conductive
yarns, fabrics coated with conductive materials).
[0055] Electrically conductive elements 120, 122, and/or 124 can be
used to deliver an electrical signal to the ear, for example, for
use in VNS. Electrically conductive elements 120, 122, and/or 124
can also function as physiological electrodes for detecting
bioelectrical signals of an individual, including, for example,
electroencephalogram (EEG), electrooculogram (EOG),
electrocardiography (ECG), and electromyogram (EMG) signals, and
may also be used to detect pulse rate, respiration rate, body
temperature, sweat levels, and glucose, among other health
parameters. Accordingly, electrically conductive elements 120, 122,
and 124 may be used in place of or in conjunction with sensor
180.
[0056] Earpiece 100 may include an embedded integrated circuit, for
example, inside the body 102. In examples, earpiece 100 is
communicably coupled with an integrated circuit that is separate
and remotely located, for example, in a computer or a mobile
device. The integrated circuit can include a data processor, a
memory, and a communication processor. Commands to be executed by
the processor can be obtained via the communication processor. The
communication processor facilitates wired or wireless communication
for earpiece 100 and can be facilitated via one or more antennas,
for example. The communication processor can facilitate
communication with one or more networks or other devices, for
example, by using wireless methods that are known, including but
not limited to Wi-Fi. Bluetooth, 3G, 4G, LTE, and/or ZigBee, among
others. Earpiece 100 can further include an embedded power source
(e.g., a battery) required to carry out various functionalities
involving the integrated circuit and the one or more electronic
components described herein.
[0057] In another aspect, and with reference to FIG. 3, a method 40
of providing VNS to a user is provided. The method 40 may include
the steps of delivering 42, via an earpiece worn by a user, audio
content to a user. The method 40 may include stimulating 44 a vagus
nerve of the user with one or more electrical signals generated by
the earpiece in coordination with the delivery of the audio
content. Coordination between the one or more electrical signals
and the delivery of the audio content may include any embodiments
in which the vagus nerve is stimulated in accordance with the audio
content. For example, if the audio content includes a series of
audio tones, the vagus nerve may be stimulated simultaneously with
the playback of each tone.
[0058] According to an example, the audio content may include
guided breathing instructions prompting the user to adjust their
respiration rate. The guided breathing instructions include audio
cues to prompt the user to adjust their respiration rate. The step
of stimulating 44 the vagus nerve of the user may further include
stimulating the vagus nerve of the user in coordination with the
audio cues. In another example, the audio content may be a series
of binaural beats. The audio content may include any sounds,
recordings, or soundtracks useful for audio therapy.
[0059] According to an example, the method may further include the
step of measuring 46, via a sensor, a biometric property of the
user. The method may further include the step of adjusting 48 the
audio content based on the biometric property. The adjusted audio
content may include guided breathing instructions. The adjusted
audio content may include audio cues to prompt the user to adjust
their respiration rate. Playback timing of the audio cues may be
adjusted based on the biometric property.
[0060] According to an example, the earpiece may be further
configured to adjust the guided breathing instructions based on a
biometric property of the user measured by a sensor. The guided
breathing instructions may further include audio cues to prompt the
user to adjust their respiration rate. Playback timing of the audio
cues may be adjusted based on the biometric property.
[0061] In another aspect, and with reference with FIG. 2, a system
50 for providing VNS to a user is provided. The system 50 may
include an earpiece 100. The earpiece may include an audio
transducer 170. The earpiece 100 may include a first electrode 120.
The earpiece may be configured to form a first electrical contact
between the electrode 120 and an ear canal, tragus, or concha of a
first ear of a user. The earpiece 100 may be further configured to
deliver audio content to a user. The earpiece 100 may be further
configured to stimulate a vagus nerve of the user with one or more
electrical signals generated by the earpiece 100 in coordination
with the delivery of the audio content. The system 50 may further
include a second electrode 122 configured to form a second
electrical contact with the user.
[0062] According to an example, the audio content may include
guided breathing instructions prompting the user to adjust their
respiration rate. The guided breathing instructions may include
audio cues to prompt the user to adjust their respiration rate. The
earpiece 100 may be further configured to stimulate the vagus nerve
of the user in coordination with the audio cues.
[0063] According to an example, the system may further include a
sensor 180. The sensor 180 may be configured to measure a biometric
property of the user.
[0064] According to an example, the earpiece 100 may be further
configured to adjust the audio content based on the biometric
property. The adjusted audio content may be guided breathing
instructions. The adjusted audio content may include audio cues to
prompt the user to adjust their respiration rate. Playback timing
of the audio cues may be adjusted based on the biometric
property.
[0065] In another aspect, and with reference with FIG. 2, an
earpiece 100 for providing VNS to a user is provided. The earpiece
100 may include an audio transducer 170. The audio transducer 170
may be configured to deliver audio content to a user. The earpiece
100 may further include a first electrode 120. The first electrode
120 may be configured to form a first electrical contact with an
ear canal, tragus, or concha of a user. The first electrode 120 may
be further configured to stimulate a vagus nerve of the user with
one or more electrical signals generated by the earpiece 100 in
coordination with the delivery of the audio content.
[0066] According to an example, the audio content may include
guided breathing instructions prompting the user to adjust their
respiration rate. The guided breathing instructions may include
audio cues to prompt the user to adjust their respiration rate. The
first electrode 120 may be further configured to stimulate the
vagus nerve of the user in coordination with the audio cues.
[0067] According to an example, the earpiece 100 may be further
configured to adjust the guided breathing instructions based on a
biometric property of the user measured by a sensor. The guided
breathing instructions may further include audio cues to prompt the
user to adjust their respiration rate. Playback timing of the audio
cues may be adjusted based on the biometric property.
[0068] The above-described examples of the described subject matter
can be implemented in any of numerous ways. For example, some
aspects may be implemented using hardware, software or a
combination thereof. When any aspect is implemented at least in
part in software, the software code can be executed on any suitable
processor or collection of processors, whether provided in a single
device or computer or distributed among multiple
devices/computers.
[0069] The present disclosure may be implemented as a system, a
method, and/or a computer program product at any possible technical
detail level of integration. The computer program product may
include a computer readable storage medium (or media) having
computer readable program instructions thereon for causing a
processor to carry out aspects of the present disclosure.
[0070] The computer readable storage medium can be a tangible
device that can retain and store instructions for use by an
instruction execution device. The computer readable storage medium
may be, for example, but is not limited to, an electronic storage
device, a magnetic storage device, an optical storage device, an
electromagnetic storage device, a semiconductor storage device, or
any suitable combination of the foregoing. A non-exhaustive list of
more specific examples of the computer readable storage medium
includes the following: a portable computer diskette, a hard disk,
a random access memory (RAM), a read-only memory (ROM), an erasable
programmable read-only memory (EPROM or Flash memory), a static
random access memory (SRAM), a portable compact disc read-only
memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a
floppy disk, a mechanically encoded device such as punch-cards or
raised structures in a groove having instructions recorded thereon,
and any suitable combination of the foregoing. A computer readable
storage medium, as used herein, is not to be construed as being
transitory signals per se, such as radio waves or other freely
propagating electromagnetic waves, electromagnetic waves
propagating through a waveguide or other transmission media (e.g.,
light pulses passing through a fiber-optic cable), or electrical
signals transmitted through a wire.
[0071] Computer readable program instructions described herein can
be downloaded to respective computing/processing devices from a
computer readable storage medium or to an external computer or
external storage device via a network, for example, the Internet, a
local area network, a wide area network and/or a wireless network.
The network may comprise copper transmission cables, optical
transmission fibers, wireless transmission, routers, firewalls,
switches, gateway computers and/or edge servers. A network adapter
card or network interface in each computing/processing device
receives computer readable program instructions from the network
and forwards the computer readable program instructions for storage
in a computer readable storage medium within the respective
computing/processing device.
[0072] Computer readable program instructions for carrying out
operations of the present disclosure may be assembler instructions,
instruction-set-architecture (ISA) instructions, machine
instructions, machine dependent instructions, microcode, firmware
instructions, state-setting data, configuration data for integrated
circuitry, or either source code or object code written in any
combination of one or more programming languages, including an
object oriented programming language such as Smalltalk, C++, or the
like, and procedural programming languages, such as the "C"
programming language or similar programming languages. The computer
readable program instructions may execute entirely on the user's
computer, partly on the user's computer, as a stand-alone software
package, partly on the user's computer and partly on a remote
computer or entirely on the remote computer or server. In the
latter scenario, the remote computer may be connected to the user's
computer through any type of network, including a local area
network (LAN) or a wide area network (WAN), or the connection may
be made to an external computer (for example, through the Internet
using an Internet Service Provider). In some examples, electronic
circuitry including, for example, programmable logic circuitry,
field-programmable gate arrays (FPGA), or programmable logic arrays
(PLA) may execute the computer readable program instructions by
utilizing state information of the computer readable program
instructions to personalize the electronic circuitry, in order to
perform aspects of the present disclosure.
[0073] Aspects of the present disclosure are described herein with
reference to flowchart illustrations and/or block diagrams of
methods, apparatus (systems), and computer program products
according to examples of the disclosure. It will be understood that
each block of the flowchart illustrations and/or block diagrams,
and combinations of blocks in the flowchart illustrations and/or
block diagrams, can be implemented by computer readable program
instructions.
[0074] The computer readable program instructions may be provided
to a processor of a general purpose computer, special purpose
computer, or other programmable data processing apparatus to
produce a machine, such that the instructions, which execute via
the processor of the computer or other programmable data processing
apparatus, create means for implementing the functions/acts
specified in the flowchart and/or block diagram block or blocks.
These computer readable program instructions may also be stored in
a computer readable storage medium that can direct a computer, a
programmable data processing apparatus, and/or other devices to
function in a particular manner, such that the computer readable
storage medium having instructions stored therein comprises an
article of manufacture including instructions which implement
aspects of the function/act specified in the flowchart and/or block
diagram or blocks.
[0075] The computer readable program instructions may also be
loaded onto a computer, other programmable data processing
apparatus, or other device to cause a series of operational steps
to be performed on the computer, other programmable apparatus or
other device to produce a computer implemented process, such that
the instructions which execute on the computer, other programmable
apparatus, or other device implement the functions/acts specified
in the flowchart and/or block diagram block or blocks.
[0076] The flowchart and block diagrams in the Figures illustrate
the architecture, functionality, and operation of possible
implementations of systems, methods, and computer program products
according to various examples of the present disclosure. In this
regard, each block in the flowchart or block diagrams may represent
a module, segment, or portion of instructions, which comprises one
or more executable instructions for implementing the specified
logical function(s). In some alternative implementations, the
functions noted in the blocks may occur out of the order noted in
the Figures. For example, two blocks shown in succession may, in
fact, be executed substantially concurrently, or the blocks may
sometimes be executed in the reverse order, depending upon the
functionality involved. It will also be noted that each block of
the block diagrams and/or flowchart illustration, and combinations
of blocks in the block diagrams and/or flowchart illustration, can
be implemented by special purpose hardware-based systems that
perform the specified functions or acts or carry out combinations
of special purpose hardware and computer instructions.
[0077] While several inventive examples have been described and
illustrated herein, those of ordinary skill in the art will readily
envision a variety of other means and/or structures for performing
the function and/or obtaining the results and/or one or more of the
advantages described herein, and each of such variations and/or
modifications is deemed to be within the scope of the inventive
examples described herein. More generally, those skilled in the art
will readily appreciate that all parameters, dimensions, materials,
and configurations described herein are meant to be exemplary and
that the actual parameters, dimensions, materials, and/or
configurations will depend upon the specific application or
applications for which the inventive teachings is/are used. Those
skilled in the art will recognize, or be able to ascertain using no
more than routine experimentation, many equivalents to the specific
inventive examples described herein. It is, therefore, to be
understood that the foregoing examples are presented by way of
example only and that, within the scope of the appended claims and
equivalents thereto, inventive examples may be practiced otherwise
than as specifically described and claimed. Inventive examples of
the present disclosure are directed to each individual feature,
system, article, material, and/or method described herein. In
addition, any combination of two or more such features, systems,
articles, materials, and/or methods, if such features, systems,
articles, materials, and/or methods are not mutually inconsistent,
is included within the inventive scope of the present
disclosure.
[0078] Other implementations are within the scope of the following
claims and other claims to which the applicant may be entitled.
* * * * *