U.S. patent number 9,965,059 [Application Number 15/096,240] was granted by the patent office on 2018-05-08 for methods, systems and devices for physical contact activated display and navigation.
This patent grant is currently assigned to Fitbit, Inc.. The grantee listed for this patent is Fitbit, Inc.. Invention is credited to Christine Brumback, Dave Knight, Jayson Messenger, Nicholas Myers, James Park, Timothy Roberts, Shelten Yuen.
United States Patent |
9,965,059 |
Myers , et al. |
May 8, 2018 |
Methods, systems and devices for physical contact activated display
and navigation
Abstract
Methods, systems and devices are provided for providing user
interface navigation of screen display metrics of a device. In one
example, a device is configured for capture of activity data for a
user. The device includes a housing and a screen disposed on the
housing to display a plurality of metrics which include metrics
that characterize the activity captured over time. The device
further includes a sensor disposed in the housing to capture
physical contact upon the housing. A processor is included to
process the physical contact to determine if the physical contact
qualifies as an input. The processor enables the screen from an off
state when the physical contact qualifies as the input. The screen
is configured to display one or more of the plurality of metrics in
accordance with a scroll order, and a first metric displayed in
response to the physical contact that qualifies as the input.
Inventors: |
Myers; Nicholas (Oakland,
CA), Brumback; Christine (San Francisco, CA), Roberts;
Timothy (San Francisco, CA), Park; James (Berkeley,
CA), Knight; Dave (San Francisco, CA), Yuen; Shelten
(Berkeley, CA), Messenger; Jayson (San Francisco, CA) |
Applicant: |
Name |
City |
State |
Country |
Type |
Fitbit, Inc. |
San Francisco |
CA |
US |
|
|
Assignee: |
Fitbit, Inc. (San Francisco,
CA)
|
Family
ID: |
50974078 |
Appl.
No.: |
15/096,240 |
Filed: |
April 11, 2016 |
Prior Publication Data
|
|
|
|
Document
Identifier |
Publication Date |
|
US 20160224130 A1 |
Aug 4, 2016 |
|
Related U.S. Patent Documents
|
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
Issue Date |
|
|
14192282 |
Feb 27, 2014 |
9310909 |
|
|
|
14050270 |
Oct 9, 2013 |
|
|
|
|
13959714 |
Aug 5, 2013 |
8762101 |
|
|
|
13693334 |
Dec 4, 2012 |
8548770 |
|
|
|
13667229 |
Nov 2, 2012 |
8437980 |
|
|
|
13469027 |
May 10, 2012 |
8311769 |
|
|
|
13246843 |
Sep 27, 2011 |
8180591 |
|
|
|
13156304 |
Jun 8, 2011 |
9167991 |
|
|
|
13759485 |
Feb 5, 2013 |
8543351 |
|
|
|
13667229 |
Nov 2, 2012 |
8437980 |
|
|
|
13469027 |
May 10, 2012 |
8311769 |
|
|
|
13246843 |
Sep 27, 2011 |
8180591 |
|
|
|
13156304 |
Jun 8, 2011 |
9167991 |
|
|
|
13913726 |
Jun 10, 2013 |
8670953 |
|
|
|
61885959 |
Oct 2, 2013 |
|
|
|
|
61388595 |
Sep 30, 2010 |
|
|
|
|
61390811 |
Oct 7, 2010 |
|
|
|
|
Current U.S.
Class: |
1/1 |
Current CPC
Class: |
A61B
5/681 (20130101); G06F 1/3265 (20130101); G06F
1/3212 (20130101); G16H 40/63 (20180101); G06F
3/0485 (20130101); A61B 5/0022 (20130101); H04B
7/26 (20130101); A61B 5/7475 (20130101); G06F
3/0488 (20130101); A61B 5/7455 (20130101); G06F
3/0482 (20130101); G06F 19/00 (20130101); A63B
24/0062 (20130101); A61B 5/7425 (20130101); G06F
1/163 (20130101); A61B 5/7405 (20130101); A61B
5/7264 (20130101); A61B 5/7435 (20130101); G16H
20/40 (20180101); A61B 5/1118 (20130101); G06F
3/041 (20130101); A61B 5/746 (20130101); G06F
3/038 (20130101); A61B 2562/0219 (20130101); A61B
5/4812 (20130101); A61B 5/222 (20130101); G16H
20/30 (20180101); A61B 5/021 (20130101); A61B
5/1112 (20130101); A61B 5/1123 (20130101); A61B
5/6838 (20130101); G01C 22/006 (20130101); Y02D
10/00 (20180101); A61B 5/02055 (20130101); A61B
5/024 (20130101); A61B 5/4809 (20130101); G16H
40/67 (20180101); A61B 5/4815 (20130101); A61B
2560/0475 (20130101); A61B 2560/0242 (20130101); G06F
2200/1636 (20130101); A61B 2560/0214 (20130101) |
Current International
Class: |
G06F
3/038 (20130101); H04B 7/26 (20060101); G06F
1/16 (20060101); G06F 3/0482 (20130101); G06F
3/0485 (20130101); G06F 19/00 (20180101); G06F
3/041 (20060101); A63B 24/00 (20060101); A61B
5/00 (20060101); G01C 22/00 (20060101); A61B
5/22 (20060101); A61B 5/11 (20060101); A61B
5/0205 (20060101); A61B 5/021 (20060101); A61B
5/024 (20060101) |
References Cited
[Referenced By]
U.S. Patent Documents
Foreign Patent Documents
|
|
|
|
|
|
|
100998921 |
|
Jul 2007 |
|
CN |
|
101380252 |
|
Mar 2009 |
|
CN |
|
102598086 |
|
Jul 2012 |
|
CN |
|
11-347021 |
|
Dec 1999 |
|
JP |
|
2178588 |
|
Jan 2002 |
|
RU |
|
WO 02/011019 |
|
Feb 2002 |
|
WO |
|
WO 06/055125 |
|
May 2006 |
|
WO |
|
WO 06/090197 |
|
Aug 2006 |
|
WO |
|
WO 08/038141 |
|
Apr 2008 |
|
WO |
|
WO 09/042965 |
|
Apr 2009 |
|
WO |
|
WO 12/170586 |
|
Dec 2012 |
|
WO |
|
WO 12/170924 |
|
Dec 2012 |
|
WO |
|
WO 12/171032 |
|
Dec 2012 |
|
WO |
|
WO 15/127067 |
|
Aug 2015 |
|
WO |
|
WO 16/003269 |
|
Jan 2016 |
|
WO |
|
Other References
Chandrasekar et al., Aug. 28-Sep. 1, 2012, Plug-and-Play,
Single-Chip Photoplethysmography, 34th Annual International
Conference of the IEEE EMBS, San Diego, California USA, 4 pages.
cited by applicant .
Fang et al., Dec. 2005, Design of a Wireless Assisted Pedestrian
Dead Reckoning System--The NavMote Experience, IEEE Transactions on
Instrumentation and Measurement, 54(6):2342-2358. cited by
applicant .
Godfrey et al., 2008, Direct Measurement of Human Movement by
Accelerometry, Medical Engineering & Physics, 30:1364-1386.
cited by applicant .
Godha et al., May 2008, Foot Mounted Inertia System for Pedestrian
Naviation, Measurement Science and Technology, 19(7):1-9. cited by
applicant .
Intersema App., Using MS5534 for altimeters and barometers, Note
AN501, Jan. 2006. cited by applicant .
Ladetto et al., Sep. 2000, On Foot Navigation: When GPS alone is
not Enough, Journal of Navigation, 53(2):279-285. cited by
applicant .
Lammel et al., Sep. 2009, Indoor Navigation with MEMS Sensors,
Proceedings of the Eurosensors XIII conference, 1(1):532-535. cited
by applicant .
Lester et al., 2005, A Hybrid Discriminative/Generative Approach
for Modeling Human Activities, Proc. of the Int'l Joint Conf.
Artificial Intelligence, pp. 766-772. cited by applicant .
Lester et al., 2009, Validated caloric expenditure estimation using
a single body-worn sensor, Proc. of the Int'l Conf. on Ubiquitous
Computing, pp. 225-234. cited by applicant .
Ohtaki et al., Aug. 2005, Automatic classification of ambulatory
movements and evaluation of energy consumptions utilizing
accelerometers and barometer, Microsystem Technologies,
11(8-10:)1034-1040. cited by applicant .
Parkka et al., Jan. 2006, Activity Classification Using Realistic
Data From Wearable Sensors, IEEE Transactions on Information
Technology in Biomedicine, 10(1):119-128. cited by applicant .
Perrin et al., 2000, Improvement of Walking Speed Prediction by
Accelerometry and Altimetry, Validated by Satellite Positioning,
Medical & Biological Engineering & Computing, 38:164-168.
cited by applicant .
Retscher, 2006, An Intelligent Multi-Sensor system for Pedestrian
Navigation, Journal of Global Positioning Systems, 5(1):110-118.
cited by applicant .
SCP 1000-D01/D11 Pressure Sensor as Barometer and Altimeter, VTI
Technologies Application, Jun. 2006, Note 33. cited by applicant
.
Stirling et al., 2005, Evaluation of a New Method of Heading
Estimation of Pedestrian Dead Reckoning Using Shoe Mounted Sensors,
Journal of Navigation, 58:31-45. cited by applicant .
Suunto LUMI User Guide, Jun. and Sep. 1997. cited by applicant
.
Tanigawa et al., Mar. 2008, Drift-free dynamic height sensor using
MEMS IMU aided by MEMS pressure sensor, Workshop on Positioning,
Navigation and Communication, pp. 191-196. cited by applicant .
International Search Report dated Aug. 15, 2008, in related
application PCT/IB07/03617. cited by applicant .
Sagawa et al., Aug.-Sep. 1998, Classification of Human Moving
Patterns Using Air Pressure and Acceleration, Proceedings of the
24.sup.th Annual Conference of the IEEE Industrial Electronics
Society, 2:1214-1219. cited by applicant .
Sagawa et al., Oct. 2000, Non-restricted measurement of walking
distance, IEEE Int'l Conf. on Systems, Man, and Cybernetics,
3:1847-1852. cited by applicant .
Specification of the Bluetooth.RTM. System, Core Package, version
4.1, Dec. 2013, vols. 0 & 1, 282 pages. cited by
applicant.
|
Primary Examiner: Chang; Kent
Assistant Examiner: Au; Scott
Attorney, Agent or Firm: Knobbe, Martens, Olson & Bear,
LLP
Parent Case Text
CLAIM OF PRIORITY
This application is a continuation of U.S. patent application Ser.
No. 14/192,282, entitled "Methods, Systems and Devices for Physical
Contact Activated Display and Navigation," filed on Feb. 27, 2014,
which claims priority from U.S. patent application Ser. No.
14/050,270, entitled "Methods, Systems and Devices for Physical
Contact Activated Display and Navigation", filed on Oct. 10, 2013,
which claims priority to U.S. Provisional Application No.
61/885,959, entitled "Methods, Systems and Devices for Physical
Contact Activated Display and Navigation", filed on Oct. 2, 2013,
all of which are incorporated herein by reference.
U.S. patent application Ser. No. 14/192,282, entitled "Methods,
Systems and Devices for Physical Contact Activated Display and
Navigation," filed on Feb. 27, 2014 is a continuation-in-part of
U.S. patent application Ser. No. 13/959,714 (now issued as U.S.
Pat. No. 8,762,101, issued Jun. 24, 2014), filed on Aug. 5, 2013,
titled "Methods and Systems for Identification of Event Data Having
Combined Activity and Location Information of Portable Monitoring
Devices", which is a continuation-in-part of U.S. patent
application Ser. No. 13/693,334 (now U.S. Pat. No. 8,548,770,
issued Oct. 1, 2013), filed on Dec. 4, 2012, titled "Portable
Monitoring Devices and Methods for Operating Same", which is a
divisional of U.S. patent application Ser. No. 13/667,229 (now U.S.
Pat. No. 8,437,980, issued on May 7, 2013), filed on Nov. 2, 2012,
titled "Portable Monitoring Devices and Methods for Operating
Same", which is a divisional of U.S. patent application Ser. No.
13/469,027 (now U.S. Pat. No. 8,311,769, issued Nov. 13, 2012)
filed May 10, 2012, titled "Portable Monitoring Devices and Methods
for Operating Same", which is a divisional of U.S. patent
application Ser. No. 13/246,843 (now U.S. Pat. No. 8,180,591,
issued on May 15, 2012) filed on Sep. 27, 2011, which is a
divisional of U.S. patent application Ser. No. 13/156,304 (now U.S.
Pat. No. 9,167,991, issued on Oct. 27, 2015) filed on Jun. 8, 2011,
titled "Portable Monitoring Devices and Methods for Operating
Same", which claims the benefit of and priority to, under 35 U.S.C.
119 .sctn. (e) to U.S. Provisional Patent Application No.
61/388,595, filed on Sep. 30, 2010, and titled "Portable Monitoring
Devices and Methods for Operating Same", and to U.S. Provisional
Patent Application No. 61/390,811, filed on Oct. 7, 2010, and
titled "Portable Monitoring Devices and Methods for Operating
Same", all of which are hereby incorporated by reference in their
entirety.
U.S. patent application Ser. No. 13/959,714, (now issued as U.S.
Pat. No. 8,762,101) filed Aug. 5, 2013, titled "Methods and Systems
for Identification of Event Data Having Combined Activity and
Location Information of Portable Monitoring Devices", is a
continuation-in-part of U.S. patent application Ser. No.
13/759,485, (now issued as U.S. Pat. No. 8,543,351, issued on Sep.
24, 2013), filed on Feb. 5, 2013, titled "Portable Monitoring
Devices and Methods for Operating Same", which is a divisional of
U.S. patent application Ser. No. 13/667,229 (now U.S. Pat. No.
8,437,980, issued on May 7, 2013) filed on Nov. 2, 2012, titled
"Portable Monitoring Devices and Methods for Operating Same", which
is a divisional of U.S. patent application Ser. No. 13/469,027 (now
U.S. Pat. No. 8,311,769, issued on Nov. 13, 2012), filed on May 10,
2012, titled "Portable Monitoring Devices and Methods for Operating
Same", which is a divisional of U.S. patent application Ser. No.
13/246,843 (now U.S. Pat. No. 8,180,591, issued on May 15, 2012)
filed on Sep. 27, 2011, which is a divisional of U.S. patent
application Ser. No. 13/156,304 (now U.S. Pat. No. 9,167,991,
issued on Oct. 27, 2015), filed on Jun. 8, 2011, titled "Portable
Monitoring Devices and Methods for Operating Same", which claims
the benefit of and priority under 35 U.S.C. 119 .sctn. (e), to U.S.
Provisional Patent Application No. 61/388,595, filed on Sep. 30,
2010, and titled "Portable Monitoring Devices and Methods for
Operating Same" and to U.S. Provisional Patent Application No.
61/390,811, filed on Oct. 7, 2010, and titled "Portable Monitoring
Devices and Methods for Operating Same", all of which are hereby
incorporated by reference in their entirety.
U.S. patent application Ser. No. 14/192,282, entitled "Methods,
Systems and Devices for Physical Contact Activated Display and
Navigation," filed on Feb. 27, 2014 is a continuation-in-part of
U.S. patent application Ser. No. 13/913,726 (now U.S. Pat. No.
8,670,953, issued on Mar. 11, 2014), filed on Jun. 10, 2013, titled
"Portable Monitoring Devices and Methods for Operating Same", all
of which are incorporated by reference in its entirety.
Claims
What is claimed is:
1. A method, comprising, receiving motion data corresponding to
first physical contact, the motion data generated by a sensor of an
activity tracking device, the first physical contact being received
on a touch sensitive screen mounted on a housing of the activity
tracking device, the touch sensitive screen of the activity
tracking device configured to display a plurality of metrics, the
plurality of metrics being arranged in a pre-defined scroll order,
wherein the physical contact received on the touch sensitive screen
comprises one or more taps by a bare finger of a user of the
activity tracking device, wherein the activity tracking device is
wearable by the user and programmable by the user on a computing
device associated with the user to display a pre-defined first
metric from among the plurality of metrics on the touch sensitive
screen; determining, based on the motion data, that the first
physical contact qualifies as a double tap input for the activity
tracking device; in response to determining that the first physical
contact qualifies as the double tap input, causing the touch
sensitive screen of the activity tracking device to be activated,
the touch sensitive screen being configured to display, upon
activation, the first metric of the plurality of metrics;
determining, based on additional motion data generated by the
sensor in response to second physical contact received on the touch
sensitive screen subsequent to the first physical contact, whether
the second physical contact qualifies as a single tap input that is
different from the double tap input; in response to determining
that the second physical contact qualifies as the single tap input,
causing the first metric displayed on the touch sensitive screen to
be replaced with a next one of the plurality of metrics in the
scroll order that is different from the first metric to be
displayed on the touch sensitive screen such that only one metric
of the plurality of metrics in the scroll order is displayed at a
time; and for each subsequent physical contact that qualifies as
the single tap input, causing another next one of the plurality of
metrics in the scroll order to be displayed on the touch sensitive
screen, the method being executed by one or more processors of the
activity tracking device.
2. The method of claim 1, wherein the plurality of metrics
comprises one or more of step count, floors climbed, stairs
climbed, distance traveled, active minutes, or calories burned.
3. The method of claim 1, wherein the scroll order is user
configurable and defines which respective ones of the plurality of
metrics follow the first metric.
4. The method of claim 3, wherein the activity tracking device is
associated to a user account, the user account provides access to
configure the scroll order.
5. The method of claim 4, wherein the computing device is usable to
access a website for the user account, the computing device is
provided with one or more graphical user interfaces to configure
the scroll order.
6. The method of claim 5, wherein the computing device is a
portable computing device, the activity tracking device being
configured to receive the scroll order from the portable computing
device via a wireless connection.
7. The method of claim 4, further comprising enabling wireless
communication with a portable computing device paired with the
activity tracking device, the portable computing device provided
with access to a website for access to the user account associated
with the activity tracking device, the scroll order being stored in
association with the user account, wherein the activity tracking
device is configured to receive the scroll order obtained from the
website by the portable computing device.
8. The method of claim 4, wherein the computing device is usable to
access a website for the user account, the computing device
provided with one or more graphical user interface to configure the
scroll order of the plurality of metrics and to remove or add
metrics to the plurality of metrics.
9. The method of claim 1, wherein the first metric is a home screen
metric specified by the user of the activity tracking device, the
home screen metric being initially displayed on the touch sensitive
screen upon determining that the first physical contact qualifies
as the double tap input.
10. The method of claim 1, further comprising, causing the touch
sensitive screen to be deactivated, subsequent to the second
physical contact, when a period of time passes without further
detected user input; and causing the touch sensitive screen to be
activated, after the period of time passes, and to display the
first metric in response to determining that third physical contact
is qualified as the double tap input.
11. The method of claim 1, further comprising, transitioning the
touch sensitive screen to an off state in an absence of user input
to the activity tracking device.
12. The method of claim 1, wherein the first metric comprises a
time metric and one or more remaining metrics of the plurality of
metrics each comprises an activity metric indicative of activity
data of the user captured over time by the activity tracking
device.
13. An activity tracking device configured for capture of activity
data for a user, comprising, a housing; a screen disposed on the
housing, the screen configured to display a plurality of metrics
which include a time of day metric and activity metrics that
characterize the activity data captured over time, the plurality of
metrics being arranged in a pre-defined scroll order, wherein the
activity tracking device is wearable by the user and programmable
by the user on a computing device to display a pre-defined first
metric from among the plurality of metrics on the screen; a sensor
disposed in the housing, the sensor configured to generate motion
data in response to physical contact upon the housing; and one or
more processors configured to: receive first motion data
corresponding to first physical contact, the first motion data
generated by the sensor in response to the first physical contact
received upon the housing of the activity tracking device,
determine, based on the first motion data, that the first physical
contact qualifies as a double tap input, in response to determining
that the first physical contact qualifies as the double tap input,
cause the screen to be activated and to display, upon activation,
the first metric of the plurality of metrics, and determine, based
on second motion data generated by the sensor in response to second
physical contact received upon the housing subsequent to the first
physical contact, whether the second physical contact qualifies as
a single tap input that is different from the double tap input, and
in response to determining that the second physical contact
qualifies as the single tap input, cause the first metric displayed
on the screen to be replaced with a next one of the plurality of
metrics in the scroll order that is different from the first metric
to be displayed on the screen such that only one metric of the
plurality of metrics in the scroll order is displayed at a time;
and for each subsequent physical contact that qualifies as the
single tap input, cause another next one of the plurality of
metrics in the scroll order to be displayed on the screen.
14. The activity tracking device of claim 13, wherein the housing
is part of a wearable wrist attachable structure, or an attachable
structure that can be carried or worn by the user of the activity
tracking device.
15. The activity tracking device of claim 13, wherein the physical
contact produces motion that is captured by the sensor, and the
physical contact is received from a finger or hand.
16. The activity tracking device of claim 13, further comprising
one or both of an altimeter and an accelerometer.
17. The activity tracking device of claim 13, wherein the housing
further includes a memory for storing the activity data, captured
data, and also includes wireless communication logic, and the
wireless communication logic includes one of WiFi processing logic,
or Bluetooth (BT) processing logic, or radio processing logic.
18. The activity tracking device of claim 17, wherein the wireless
communication logic is configured to pair with a portable computing
device or a computer, and the portable computing device or the
computer is configured for communication over the Internet with a
server, the server having processing instructions for configuring
the scroll order of the plurality of metrics, or removing metrics
from, or adding metrics to, the plurality of metrics.
19. The activity tracking device of claim 13, wherein the one or
more processors are further configured to examine motion profiles
captured by the sensor to determine whether the physical contact
qualifies as the double tap input or the single tap input.
20. The activity tracking device of claim 13, wherein the plurality
of metrics comprises one or more of step count, floors climbed,
stairs climbed, distance traveled, active minutes, or calories
burned.
21. The activity tracking device of claim 13, wherein the scroll
order is user configurable and defines which respective ones of the
plurality of metrics follow the first metric.
Description
FIELD
The present disclosure relates to systems and methods for capturing
activity data over a period of time and methods and systems for
navigating metric data to a display.
BACKGROUND
In recent years, the need for health and fitness has grown
tremendously. The growth has occurred due to a better understanding
of the benefits of good fitness to overall health and wellness.
Unfortunately, although today's modern culture has brought about
many new technologies, such as the Internet, connected devices and
computers, people have become less active. Additionally, many
office jobs require people to sit in front of computer screens for
long periods of time, which further reduces a person's activity
levels. Furthermore, much of today's entertainment options involve
viewing multimedia content, computer social networking, and other
types of computer involved interfacing. Although such computer
activity can be very productive as well as entertaining, such
activity tends to reduce a person's overall physical activity.
To provide users concerned with health and fitness a way of
measuring or accounting for their activity or lack thereof, fitness
trackers are often used. Fitness trackers are used to measure
activity, such as walking, motion, running, sleeping, being
inactive, bicycling, exercising on an elliptical trainer, and the
like. Usually, the data collected by such fitness trackers can be
transferred and viewed on a computing device. However, such data is
often provided as a basic accumulation of activity data with
complicated or confusing interfaces.
It is in this context that embodiments described herein arise.
SUMMARY
Embodiments described in the present disclosure provide systems,
apparatus, computer readable media, and methods for analyzing
tracked activity data and providing navigation screens and
interfaces on a device used by a user. The activity tracking device
includes sensor(s) for detecting when physical contact occurs onto
the activity tracking device and logic for providing a display
action to the screen of the activity tracking device. The physical
contact, in one embodiment, can be qualified as an input when the
physical contact has a particular characteristic or pattern that is
predefined. The characteristic can be, when the contact is the
result of one or more taps, e.g., physical contact to the activity
tracking device by a finger or hand of the user, or object held by
a user and used to impart the contact.
In one embodiment, a method is provided. The method includes
detecting a physical contact by a sensor of a device that is
configured to display a plurality of metrics on a screen of the
device and examining the physical contact to determine if the
physical contact qualifies as an input for the device. The method
acts to maintain the screen of the device in an off state for
physical contact that does not qualify as the input and actives the
screen of the device to display a first metric when the examining
determines that the physical contact qualifies as the input. The
method is executed by a processor.
In another embodiment, a device configured for capture of activity
data for a user is provided. The device includes a housing and a
screen disposed on the housing to display a plurality of metrics
which include metrics that characterize the activity captured over
time. The device further includes a sensor disposed in the housing
to capture physical contact upon the housing. A processor is
included to process the physical contact to determine if the
physical contact qualifies as an input. The processor enables the
screen from an off state when the physical contact qualifies as the
input. The screen is configured to display one or more of the
plurality of metrics in accordance with a scroll order, and a first
metric of the plurality of metrics is displayed in accordance with
user configuration identifying that the first metric is to be
displayed in response to the physical contact that qualifies as the
input, as determined by the processor.
In another embodiment, computer readable medium for storing program
instructions executable by a processor is provided. The computer
readable medium includes (a) program instructions for detecting a
physical contact by a sensor of a device that is configured to
display a plurality of metrics on a screen of the device; (b)
program instructions for examining the physical contact to
determine if the physical contact qualifies as an input for the
device; (c) program instructions for maintaining the screen of the
device in an off state for physical contact that does not qualify
as the input; (d) program instructions for activating the screen of
the device to display a first metric when the examining determines
that the physical contact qualifies as the input, (e) program
instructions for detecting user input to transition from the first
metric to a next metric in a scroll order or to an off state when
no user input is detected. If user input is received within a
predetermined time after the off state, a process turns the screen
on and displays the last displayed metric, and if the user input is
received after the predetermined time after the off state, a
process turns the screen on and displays the first metric. The
plurality of metrics include a time of day metric and metrics
representing activity data captured by the device when associated
with a user, the activity data being of the user.
In some embodiments, the input is associated with qualified
physical contact, and the user input is associated with one of
qualified physical contact, or non-touch proximity input, or voice
input, or button press input.
In still another embodiment, an activity tracking device is
configured for capturing data or activity data for a user. The
device includes a housing configured as a wearable wrist attachable
structure or a structure that can accompany the user to capture the
activity data. The device includes a screen disposed on the housing
to display a plurality of metrics which include metrics that
characterize the activity captured over time. The device has an
accelerometer sensor disposed in the housing to capture physical
contact upon the housing. The device further includes a processor
to process the physical contact to determine if the physical
contact qualifies as an input. The processor enables the screen
from an off state when the physical contact qualifies as the input,
and the screen is configured to display one or more of the
plurality of metrics in accordance with a scroll order. A first
metric of the plurality of metrics is displayed in accordance with
user configuration identifying that the first metric is to be
displayed in response to the physical contact that qualifies as the
input, as determined by the processor. The physical contact
captured by the accelerometer sensor is from sensing or detecting
one or more taps having a predefined tap profile upon the housing
by a finger, or hand or object. The user configuration identifying
the first metric enables setting of a shortcut to any metric in the
scroll order to be the first metric.
Other aspects will become apparent from the following detailed
description, taken in conjunction with the accompanying drawings,
illustrating by way of example the principles of embodiments
described in the present disclosure.
BRIEF DESCRIPTION OF THE DRAWINGS
Various embodiments described in the present disclosure may best be
understood by reference to the following description taken in
conjunction with the accompanying drawings.
FIG. 1A shows a block diagram of an activity tracking device, in
accordance with one embodiment of the present invention.
FIG. 1B illustrates an example of an activity tracking device, in
accordance with one embodiment of the present invention.
FIG. 1C illustrates another example of an activity tracking device,
in accordance with one embodiment of the present invention.
FIG. 2A illustrates an example of activity tracking device
including example components utilized for tracking activity and
motion of the device, and associated interfaces to a display
screen, in accordance with one embodiment of the present
invention.
FIG. 2B illustrates an example of activity tracking device in
communication with a remote device, in accordance with one
embodiment of the present invention.
FIG. 3A illustrates a diagram of input interfaces for accessing
screens that display metrics associated with the activity tracking
device, in accordance with one embodiment of the present
invention.
FIG. 3B illustrates an example where the activity tracking device
has been programmed or customize by a user to select a specific
metric to display on the display screen upon receiving input, in
accordance with one embodiment of the present invention.
FIG. 3C illustrates another example where the activity tracking
device 100 has been programmed or customize by a user to select a
different metric to be the first metric, in accordance with one
embodiment of the present invention.
FIG. 3D illustrates a flowchart defining example interactivity and
display functionality of the activity tracking device, in
accordance with one embodiment of the present invention.
FIG. 3E illustrates an example of a table showing example scroll
orders and inputs, in accordance with one embodiment of the present
invention.
FIG. 3F illustrates an example of the scroll order that extends
from metric 1 to metric 4, in accordance with one embodiment of the
present invention.
FIG. 3G illustrates an example where metric 3 is designated as the
first metric, which will be the metric data displayed initially
upon receiving the double tap input.
FIG. 4A-1 illustrates an example in accordance with an alternative
embodiment, associated with navigating the screens of an activity
tracking device showing ready mode.
FIG. 4A-2 illustrates an example in accordance with an alternative
embodiment, associated with navigating the screens of an activity
tracking device showing a first metric.
FIG. 4A-3 illustrates an example in accordance with an alternative
embodiment, associated with navigating the screens of an activity
tracking device showing a next metric.
FIG. 4A-4 illustrates an example in accordance with an alternative
embodiment, associated with navigating the screens of an activity
tracking device showing a next metric.
FIG. 4A-5 illustrates an example in accordance with an alternative
embodiment, associated with navigating the screens of an activity
tracking device showing a next metric.
FIG. 4A-6 illustrates an example in accordance with an alternative
embodiment, associated with navigating the screen of an activity
tracking device showing a next metric.
FIG. 4A-7 illustrates an example in accordance with an alternative
embodiment, associated with navigating the screens of an activity
tracking device showing no physical contact.
FIG. 4A-8 illustrates an example in accordance with an alternative
embodiment, associated with navigating the screens of an activity
tracking device showing a ready mode.
FIG. 4B illustrates an example of a system that includes the
device, a remote device, and the server interfacing over the
Internet, in accordance with one embodiment of the present
invention.
FIG. 5 illustrates a table providing a scroll order and the various
inputs that can be defined to traverse the scroll order, in one
embodiment.
FIG. 6 illustrates an example of a physical contact represented by
tap(s), which act to activate the display in a first metric, in
accordance with one embodiment of the present invention.
FIG. 7 illustrates another example where a single tap activates the
display in the first metric, in accordance with one embodiment of
the present invention.
FIG. 8 illustrates an example of a double tap physical contact
acting to initiate the display with the first metric, in accordance
with one embodiment of the present invention.
FIG. 9 illustrates an example where various types of activities of
users 900A-900I can be captured or collected by activity tracking
devices, in accordance with various embodiments of the present
invention.
DETAILED DESCRIPTION
Embodiments described in the present disclosure provide systems,
apparatus, computer readable media, and methods for analyzing
tracked activity data and providing navigation screens and
interfaces. Some embodiments are directed to providing navigation
interfaces for an activity tracking device. The activity tracking
device includes sensors for detecting when physical contact occurs
onto the activity tracking device and logic for providing a display
action to the screen of the activity tracking device. The physical
contact, in one embodiment, can be qualified as an input when the
physical contact has a particular characteristic that is
predefined. The characteristic can be, when the contact is the
result of one or more taps, e.g., physical contact to the activity
tracking device by a finger or hand of the user, or object held by
a user and used to impart the contact.
In other embodiments, the input can be non-physical, such as
proximity sensing input. The proximity sensing input can be
processed by an infrared proximity sensor, a thermal sensor, etc.
The input can also be by way of a button, voice input, gaze
detected input, input processed in response to motion or motion
profiles, etc.
It should be noted that there are many inventions described and
illustrated herein. The present inventions are neither limited to
any single aspect nor embodiment thereof, nor to any combinations
and/or permutations of such aspects and/or embodiments. Moreover,
each of the aspects of the present inventions, and/or embodiments
thereof, may be employed alone or in combination with one or more
of the other aspects of the present inventions and/or embodiments
thereof. For the sake of brevity, many of those permutations and
combinations will not be discussed separately herein.
Further, in the course of describing and illustrating the present
inventions, various circuitry, architectures, structures,
components, functions and/or elements, as well as combinations
and/or permutations thereof, are set forth. It should be understood
that circuitry, architectures, structures, components, functions
and/or elements other than those specifically described and
illustrated, are contemplated and are within the scope of the
present inventions, as well as combinations and/or permutations
thereof.
FIG. 1A shows a block diagram of an activity tracking device 100,
in accordance with one embodiment of the present invention. The
activity tracking device 100 is contained in a housing, which may
be worn or held by a user. The housing may be in the form of a
wristband, a clip on device, a wearable device, or may be held by
the user either in the user's hand or in a pocket or attached to
the user's body. The activity tracking device 100 includes device
components 102, which may be in the form of logic, storage, and
glue logic, one or more processors, microelectronics, and
interfacing circuitry. In one example, the components 102 will
include a processor 106, memory 108, a wireless transceiver 110, a
user interface 114, biometric sensors 116, and environmental
sensors 118.
The environmental sensors 118 may be in the form of motion
detecting sensors. In some embodiments, a motion sensor can be one
or more of an accelerometer, or a gyroscope, or a rotary encoder,
or a calorie measurement sensor, or a heat measurement sensor, or a
moisture measurement sensor, or a displacement sensor, or an
ultrasonic sensor, or a pedometer, or an altimeter, or a linear
motion sensor, or an angular motion sensor, or a multi-axis motion
sensor, or a combination thereof. The biometric sensors 116 can be
defined to measure physiological characteristics of the user that
is using the activity tracking device 100. The user interface 114
provides a way for communicating with the activity tracking device
100, in response to user interaction 104. The user interaction 104
can be in the form of physical contact (e.g., without limitation,
tapping, sliding, rubbing, multiple taps, gestures, etc.).
In some embodiments, the user interface 114 is configured to
receive user interaction 104 that is in the form of noncontact
input. The noncontact input can be by way of proximity sensors,
button presses, touch sensitive screen inputs, graphical user
interface inputs, voice inputs, sound inputs, etc. The activity
tracking device 100 can communicate with a client and/or server 112
using the wireless transceiver 110. The wireless transceiver 110
will allow the activity tracking device 100 to communicate using a
wireless connection, which is enabled by wireless communication
logic. The wireless communication logic can be in the form of a
circuit having radio communication capabilities. The radio
communication capabilities can be in the form of a Wi-Fi
connection, a Bluetooth connection, a low-energy Bluetooth
connection, or any other form of wireless tethering or near field
communication. In still other embodiments, the activity tracking
device 100 can communicate with other computing devices using a
wired connection (not shown). As mentioned, the environmental
sensors 118 can detect motion of the activity tracking device
100.
The motion can be activity of the user, such as walking, running,
stair climbing, etc. The motion can also be in the form of physical
contact received on any surface of the activity tracking device
110, so long as the environmental sensors 118 can detect such
motion from the physical contact. As will be explained in more
detail below, the physical contact may be in the form of a tap or
multiple taps by a finger upon the housing of the activity tracking
device 100.
FIG. 1B illustrates an example of an activity tracking device 100
having a housing 130 in the form of a wearable wrist attachable
device. The sensors of the activity tracking device 100 can, as
mentioned above, detect motion such as physical contact that is
applied and received on a surface 120 of the housing 130. In the
example shown, the physical contact 124 is in the form of a tap or
multiple taps on the surface 120. Device components 102 are, in one
embodiment, contained within the housing 130. The location at which
the device components 102 are integrated into the housing 130 can
vary. For example, the device components 102 can be integrated
throughout various locations around the housing 130, and not
limited to the central portion of the wrist attachable device. In
some embodiments, the device components 102 can be integrated into
or with a smart watch device.
In other embodiments, the device components 102 are positioned
substantially in a central position of the wrist attachable device,
such as under or proximate to a location where a display screen 122
is located. In the illustrated example, the housing 130 also
includes a button 126. The button 126 can be pressed to activate
the display screen 122, navigate to various metrics displayed on
the screen 122, or turn off the screen 122.
FIG. 1C illustrates another example of an activity tracking device
100, in accordance with one embodiment of the present invention.
The form factor of the activity tracking device 100 is shown as a
clickable device that includes a screen 122, a button 126, and
device components 102 integrated within the housing 130'. The
housing 130' can include a clip that allows for attachment to
clothing or articles of the user, or to simply place the device
within a pocket or holder of the user. Accordingly, the physical
contact 124 shown with respect to FIG. 1B can also be implemented
upon the surface 120 of activity tracking device 100 of FIG. 1C. It
should be understood, therefore, that the form factor of the
activity tracking device 100 can take on various configurations and
should not be limited to the example configurations provided
herein.
FIG. 2A illustrates an example of activity tracking device 100 of
FIG. 1A, showing some additional example components utilized for
tracking activity and motion of the device, and associated
interfaces to display screen 122. In this example, the finger of a
user can be used to tap and provide physical contact 124 onto any
surface 120 of activity tracking device 100. The physical contact,
when sensed by sensors 156 of the activity tracking device 100,
will cause a response by the activity tracking device 100, and
therefore provide some metric on the display screen 122. In one
embodiment, examples of a display screen 122 can include, but are
not limited to, liquid crystal display (LCD) screens, light
emitting diode (LED) screens, organic light emitting diode (OLED)
screens, plasma display screens, etc.
As shown in FIG. 2A, the activity tracking device 100 includes
logic 158. Logic 158 may include activity tracking logic 140,
physical contact logic 142, display interface logic 144, alarm
management logic 146, wireless communication logic 148, processor
106, and sensors 156. Additionally, storage (e.g. memory) 108, and
a battery 154 can be integrated within the activity tracking device
100. The activity tracking logic 140 can include logic that is
configured to process motion data produced by sensors 156, so as to
quantify the motion and produce identifiable metrics associated
with the motion.
Some motions will produce and quantify various types of metrics,
such as step count, stairs climbed, distance traveled, very active
minutes, calories burned, etc. The physical contact logic 142 can
include logic that calculates or determines when particular
physical contact can qualify as an input. To qualify as an input,
the physical contact detected by sensors 156 should have a
particular pattern that is identifiable as input. For example, the
input may be predefined to be a double tap input, and the physical
contact logic 142 can analyze the motion to determine if a double
tap indeed occurred in response to analyzing the sensor data
produced by sensors 156.
In other embodiments, the physical contact logic can be programmed
to determine when particular physical contacts occurred, the time
in between the physical contacts, and whether the one or more
physical contacts will qualify within predefined motion profiles
that would indicate that an input is desired. If physical contact
occurs that is not within some predefined profile or pattern, the
physical contact logic will not indicate or qualify that physical
contact as an input.
The display interface logic 144 is configured to interface with the
processor and the physical contact logic to determine when specific
metric data will be displayed on the display screen 122 of the
activity tracking device 100. The display interface logic 144 can
act to turn on the screen, display metric information, display
characters or alphanumeric information, display graphical user
interface graphics, or combinations thereof. Alarm management logic
146 can function to provide a user interface and settings for
managing and receiving input from a user to set an alarm. The alarm
management logic can interface with a timekeeping module (e.g.,
clock, calendar, time zone, etc.), and can trigger the activation
of an alarm. The alarm can be in the form of an audible alarm or a
non-audible alarm.
A non-audible alarm can provide such alarm by way of a vibration.
The vibration can be produced by a motor integrated in the activity
tracking device 100. The vibration can be defined to include
various vibration patterns, intensities, and custom set patterns.
The vibration produced by the motor or motors of the activity
tracking device 100 can be managed by the alarm management logic
146 in conjunction with processing by the processor 106. The
wireless communication logic 148 is configured for communication of
the activity tracking device with another computing device by way
of a wireless signal. The wireless signal can be in the form of a
radio signal. As noted above, the radio signal can be in the form
of a Wi-Fi signal, a Bluetooth signal, a low energy Bluetooth
signal, or combinations thereof. The wireless communication logic
can interface with the processor 106, storage 108 and battery 154
of device 100, for transferring activity data, which may be in the
form of motion data or processed motion data, stored in the storage
108 to the computing device.
In one embodiment, processor 106 functions in conjunction with the
various logic components 140, 142, 144, 146, and 148. The processor
106 can, in one embodiment, provide the functionality of any one or
all of the logic components. In other embodiments, multiple chips
can be used to separate the processing performed by any one of the
logic components and the processor 106. Sensors 156 can communicate
via a bus with the processor 106 and/or the logic components. The
storage 108 is also in communication with the bus for providing
storage of the motion data processed or tracked by the activity
tracking device 100. Battery 154 is provided for providing power to
the activity tracking device 100.
FIG. 2B illustrates an example of activity tracking device 100 in
communication with a remote device 200. Remote device 200 is a
computing device that is capable of communicating wirelessly with
activity tracking device 100 and with the Internet 160. Remote
device 200 can support installation and execution of applications.
Such applications can include an activity tracking application 202.
Activity tracking application 202 can be downloaded from a server.
The server can be a specialized server or a server that provides
applications to devices, such as an application store. Once the
activity tracking application 202 is installed in the remote device
200, the remote device 200 can communicate or be set to communicate
with activity tracking device 100 (Device A). The remote device 200
can be a smartphone, a handheld computer, a tablet computer, a
laptop computer, a desktop computer, or any other computing device
capable of wirelessly interfacing with Device A and the
Internet.
In one embodiment, remote device 200 communicates with activity
tracking device 100 over a Bluetooth connection. In one embodiment,
the Bluetooth connection is a low energy Bluetooth connection
(e.g., Bluetooth LE, BLE, or Bluetooth Smart). Low energy Bluetooth
is configured for providing low power consumption relative to
standard Bluetooth circuitry. Low energy Bluetooth uses, in one
embodiment, a 2.4 GHz radio frequency, which allows for dual mode
devices to share a single radio antenna. In one embodiment, low
energy Bluetooth connections can function at distances up to 50
meters, with over the air data rates ranging between 1-3 megabits
(Mb) per second. In one embodiment, a proximity distance for
communication can be defined by the particular wireless link, and
is not tied to any specific standard. It should be understood that
the proximity distance limitation will change in accordance with
changes to existing standards and in view of future standards
and/or circuitry and capabilities.
Remote device 200 can also communicate with the Internet 160 using
an Internet connection. The Internet connection of the remote
device 200 can include cellular connections, wireless connections
such as Wi-Fi, and combinations thereof (such as connections to
switches between different types of connection links). The remote
device, as mentioned above, can be a smartphone or tablet computer,
or any other type of computing device having access to the Internet
and with capabilities for communicating with the activity tracking
device 100.
A server 220 is also provided, which is interfaced with the
Internet 160. The server 220 can include a number of applications
that service the activity tracking device 100, and the associated
users of the activity tracking device 100 by way of user accounts.
For example, the server 220 can include an activity management
application 224. The activity management application 224 can
include logic for providing access to various devices 100, which
are associated with user accounts managed by server 220. Server 220
can include storage 226 that includes various user profiles
associated with the various user accounts. The user account 228a
for user A and the user account 228n for user N are shown to
include various information.
The information can include, without limitation, data associated
with a display scroll order 230, user data, etc. As will be
described in greater detail below, the display scroll order 230
includes information regarding a user's preferences, settings, and
configurations which are settable by the user or set by default at
the server 220 when accessing a respective user account. The
storage 226 will include any number of user profiles, depending on
the number of registered users having user accounts for their
respective activity tracking devices. It should also be noted that
a single user account can have various or multiple devices
associated therewith, and the multiple devices can be individually
customized, managed and accessed by a user. In one embodiment, the
server 220 provides access to a user to view the user data 232
associated with activity tracking device.
The data viewable by the user includes the tracked motion data,
which is processed to identify a plurality of metrics associated
with the motion data. The metrics are shown in various graphical
user interfaces of a website enabled by the server 220. The website
can include various pages with graphical user interfaces for
rendering and displaying the various metrics for view by the user
associated with the user account. In one embodiment, the website
can also include interfaces that allow for data entry and
configuration by the user.
The configurations can include defining which metrics will be
displayed on the activity tracking device 100. In addition, the
configurations can include identification of which metrics will be
a first metric to be displayed on the activity tracking device. The
first metric to be displayed by the activity tracking device can be
in response to a user input at the activity tracked device 100. As
noted above, the user input can be by way of physical contact. The
physical contact is qualified by the processor and/or logic of the
activity tracking device 100 to determine if the physical contact
should be treated as an input. The input can trigger or cause the
display screen of the activity tracking device 100 to be turned on
to display a specific metric, that is selected by the user as the
first metric to display. In another embodiment, the first metric
displayed in response to the input can be predefined by the system
as a default.
The configuration provided by the user by way of the server 220 and
the activity management application 224 can also be provided by way
of the activity tracking application 202 of the computing device
200. For example, the activity tracking application 202 can include
a plurality of screens that also display metrics associated with
the captured motion data of the activity tracking device 100. The
activity tracking application 202 can also allow for user input and
configuration at various graphical user interface screens to set
and define which input will produce display of the first metric. In
other embodiments, in addition to identifying the first metric to
be displayed in response to the input, which may be physical
contact, the configuration can allow an ordering of which metrics
will be displayed in a specific scroll order.
In another embodiment, the scroll order of the metrics is
predefined. In some embodiments, the input provided by the user by
way of the physical contact can be pre-assigned to a specific
metric in the scroll order. For example, the scroll order can
remain the same, while the input can allow the screen to jump to a
specific entry in the scroll order. Jumping to a specific entry can
be viewed as a shortcut to a specific entry that is desired to be
seen first by the user upon providing physical contact or input to
the device 100.
FIG. 3A illustrates a diagram of input interfaces for accessing
screens that display metrics associated with the activity tracking
device 100, in accordance with one embodiment of the present
invention. As shown, the display screen 122 is in the off state
before access is requested of the device 100. In one embodiment,
the device 100 will shift to and off state when input is not
received for a predetermined period of time, which allows the
device 102 to save energy (i.e., battery power).
In mode 1, the display screen 122 is off 300 and upon a button
press (e.g., of button 126 in FIG. 1B), the display screen 122
changes in transition 302a in direction 303a, to illustrate the
time of day metric. However, if the battery is determined to be
critical (e.g., the battery is too low to activate the device or
operate it properly for a period of time), the display screen 122
will illustrate a graphic of a battery 304. The graphical battery
304 will indicate to the user that the device must be charged. If
it is determined that the battery is low, a low battery indicator
306 will be shown in the display 122. The user can continue to use
the device, but the critical battery notification may be
re-displayed. After showing the low battery indicator 306, the
display will advance to the clock metrics 310 after a predetermined
period of time (e.g., 1 second). If it is determined that the timer
function has been running, then the transition from the low battery
indicator 306 would be automatic to the timer metric 308.
Therefore, if the timer function is not running, the system will
transition to the clock metrics 310, in this embodiment.
In mode 2, if it is determined that a button of the activity
tracking device 100 had been pressed and held for a predetermined
period of time (e.g., 1 second), then the transition would be
directly to a timer metric 308. The timer metric 308 operates a
stopwatch function, which first shows a graphic of a stopwatch and
then automatically transitions to the time kept by the stopwatch
function. If the user desires to transition from the timer metric
308 to one of the other metrics in the scroll order shown in FIG.
3A, the user can press the button 126 of the activity tracking
device 100. For example, the downward facing arrow indicates that a
button press will transition the display screen 122 to the next
metric and so on. It should be understood that the downward facing
arrow can also be activated by other than a button press, such as
by physical contact (e.g., one or more taps onto the surface of the
activity tracking device 100). As noted above, other input
functions can also be provided, such as proximity sensing,
touchscreens, voice activation, gesture detection, etc.
In mode 3, if it is determined that a double tap was detected on
the surface of the activity tracking device 100 by a sensor, the
display screen 122 will go from being on 302, to displaying a
predetermined first metric. In this example, the predetermined
first metric is a main goal 330 of the user, and is shown to be a
step count metric 312. As shown, the display screen 122 will
transition 302b in a direction 303a, which exposes an icon or
graphic associated with the main goal 330, which are steps. The
steps are shown as feet icon. The display screen 122, in one
embodiment also transitions from the feet icon to the numerical
value of the steps taken by the user utilizing the activity
tracking device 100. If the user wishes to transition and view the
other metrics in the scroll order, such as distance metric,
calories burned metric, floors metric, very active minutes metric,
alarm metric, the user can transition by pressing buttons 126 on
the device 100. Again, transitioning downward (or through a list in
any direction) is shown by the downward facing arrows, which are
activated in response to a button press (or other types of inputs).
These example transitions allow for display of other screens/data
concerning metrics 308, 310, 312, 314, 316, 318, 320, 322, etc. It
should be understood that additional metrics can be added to the
scroll order, certain metrics can be deleted from the scroll order,
the scroll order can be re-arranged, and these customizations can
be made in response to user configurations or system configurations
or default configurations.
After a predetermined period of time that no input is received by
the activity tracking device 100, the display screen 122 will
transition to and off state 300. The transition, in one embodiment
allows for the display screen 122 to transition off 302c in
direction 303b. Therefore, the display screen 122 will move to the
off state 300 where battery consumption is reduced. In one
embodiment, the transition to the off state 300 will occur after
about 6 seconds. It should be understood that this predefined
period of time can be modified for the specific configuration and
should not be limited to the specific example. As will be described
below with reference to FIG. 3D, one embodiment will allow for
custom transitions between an off state to specific metrics based
on the amount of time determined to have passed when no user input
has been received.
FIG. 3B illustrates an example where the activity tracking device
100 has been programmed or customize by a user to select a specific
metric to display on the display screen 122 upon receiving input.
In this example, the configuration has been made to the
functionality of mode 3. In mode 3, the user has configured a
double tap (physical contact) detected to activate or jump to a
specific metric selected by the user. In this example, the user has
selected that a double tap input will cause the display screen 122
to transition from off state 302 to the clock metric 310. This will
jump 332a, in the scroll order, to the clock metric 310. The clock
metric 310 will be considered the first metric, of a plurality of
metrics to be displayed on the display screen 122 of the activity
tracking device, in response to a double tap. Once the device jumps
to display the clock metric 310, the user can then scroll through
the remaining metrics by selecting or pressing a button 126 of the
activity tracking device 100. As noted above, transition to the
other metrics can also be enabled by additional physical contact or
non-physical contact inputs.
FIG. 3C illustrates another example where the activity tracking
device 100 has been programmed or customize by a user to select a
different metric to be the first metric. The first metric is a
metric that will be displayed on the display screen 122 upon
receiving the input, in accordance with mode 3. In this example,
mode 3 is a transition from the off state 300 to display the first
metric in response to a double tap. Although specific mention of a
double tap is provided herein, it should be understood that any
number of tap(s) can be provided to qualify as the physical contact
sensed by a sensor of the device 100. Once the double tap is
received, the display screen of the activity tracking device 100
will jump 332b to calories burned metric 316. Calories burned
metric 316 is a metric that the user has configured to be the first
metric that will be displayed upon receiving a double tap when the
display was in an off state. By way of this illustration, it should
be understood that any one of the metrics can be configured to be
the first metric that is shown on the display screen 122 upon
receiving the double tap. A user can then transition to other
metrics in the scroll order or list of metrics by providing
additional input. The downward facing arrow between the metrics
shows the transition through the various metrics. Upon displaying
each of the metrics, in one embodiment, the display screen 122 will
display a graphic icon indicative of the type of metric that will
be displayed. The right facing arrow indicates that the transition
between the graphic icons to the numerical metrics will be
automatically transitioned without additional user input.
FIG. 3D illustrates a flowchart defining example interactivity and
display functionality of the activity tracking device 100, in
accordance with one embodiment of the present invention. In one
embodiment, the screen is in an off state in operation 390. In
operation 391, physical contact or input is detected by the
activity tracking device 100. The physical contact or input, as
defined above can include taps, touches, proximity, button pushes,
etc. Once the input is detected in operation 391, the display
screen of the device will be activated to show the first metric. As
described above, the first metric can be user configured so as to
allow selection of which metric will be displayed first upon
receiving the input. In operation 393, it is determined if
additional input is received to scroll or move to the next
metric.
If additional input is received in operation 394, the display
screen will move in scroll to the next metric. This will continue
as the user is allowed to select the next metric in the list or
scroll order. In one embodiment, the scroll order can wrap around
and continue to display metrics. If no input is received for a
period of time, the display screen will turn off in operation 395.
In operation 396, it is determined if input is received within a
predetermined amount of time after the screen was turned off. For
example, if input, such as a button press is received within 3
seconds of the screen turning off, the screen will turn back on and
display the last metric that had been displayed.
For instance, in FIG. 3A, if the last metric being viewed was
calories burned 316, and no input was received, the screen will
turn off. If within 3 seconds of the screen being off a button
press is detected, the calories burned metric 316 will be displayed
once again in operation 397. The method will then proceed back to
operation 394 where it is determined if additional input is
received to transition to view a next metric 393. If no input is
received after a predetermined amount of time since the screen was
off in operation 396, the method will transition to operation 398.
In operation 398 it is determined that input occurred after the
predetermined amount of time that the screen was off, the method
will return back to activate the screen and show the first metric.
The input received in operation 398 may be a button press or a
physical contact, such as input received through mode 1 or mode 3.
If after the predetermined amount of time that the screen was off
no input was received, the screen will transition back or remain in
the off state 390.
FIG. 3E illustrates an example of a table 350 showing example
scroll order 352 and input 354. In this example, the scroll order
will include screen off 300, clock metric 310, distance metric 314,
calories burned metric 316, floors metric 318, very active minutes
metric 320, alarm metric 322. In this example, when no physical
contact qualifying as an input is received, the screen will remain
off. This of course also includes receiving no button presses so
that the screen will remain off. In this example, a double tap (DT)
physical contact 124 will activate the display screen 122 with the
first metric, which is identified by the user to be the clock
metric 310. The user can then transition by various button presses
126a to the various metrics in the scroll order after receiving
each of the button presses. As noted above, instead of button
presses other input can also be used to transition between the
metrics.
FIG. 3F illustrates an example of the scroll order 380 that extends
from metric 1 to metric 4. In this example, the first metric has
been identified to be metric 1. Additionally, a double tap input is
predefined to access metric 1 as the initially displayed metric on
the activity tracking device. FIG. 3G illustrates an example where
metric 3 is designated as the first metric, which will be the
metric data displayed initially upon receiving the double tap input
124. The custom designation of the first metric to any one of the
metrics in the scroll order functions as a jump 332 to the specific
metric identified by the user. The specific metric identified as
the first metric can also be predefined by the system or set as a
default. As used herein, a jump should be interpreted to also
include a shortcut to a specific metric in a list.
FIGS. 4A-1 to 4A-8 illustrate examples in accordance with an
alternative embodiment, associated with navigating the screens of
activity tracking device 100. In this example, that device 100 has
a screen 122 that is in the off state. In FIG. 4A-1, the screen is
in an off state, which may indicate that the device is in a ready
mode and ready to receive input. In FIG. 4A-2, the device receives
physical contact 124, which may be a double tap. The physical
contact and 24 will activate the display screen 122 and display a
first metric. In this example the first metric is the time of day.
In FIG. 4A-3, additional physical contact is received by device 100
which causes the display screen 122 to show a next metric, which in
this example is step count. In FIG. 4A-4, the device 100 has
received another double tap 124 which then causes the display
screen 122 to show a next metric. This next metric is a step count.
In FIG. 4A-5, the device receives another physical contact 124
which causes the device to illustrate a next metric. The next
metric is calories burned. In FIG. 4A-6, the device receives
another physical contact 124 which causes the display screen 122 to
show a next metric. The next metric is a distance. In FIG. 4A-7, it
is determined that no physical contact has occurred for a
predetermined period of time 402. This will cause the device 100,
as shown in FIG. 4A-8, to transition to turn off the screen, and
back to the ready mode.
FIG. 4B illustrates an example of a system that includes the device
100, a remote device 200, and the server 220 interfacing over the
Internet 160. The interfacing allows for the remote device 200 to
set the specific configuration of the scroll order settings of
device 100, in accordance with a user configuration. In this
example, using the remote device 200, the user navigates to a
screen provided by the application 202 of the remote device. The
application 202 is an activity tracking application that can
communicate with activity monsoon application 224 managed by the
server 220. As discussed above with reference to FIG. 2B, activity
tracking application 202 can include a plurality of user interfaces
and screens to allow user input and access to the activity data or
configure settings.
The settings being configured in this example includes settings
associated with the scroll order and the definition of the first
metric. In this example, the activity tracking application 202 will
allow the user to login to his or her user accounts and access and
identify device a, which is the activity tracking device 100. Using
screens and menus provided by the activity tracking application,
the user is able to identify the display scroll order settings in
the screen 200a of the device 100. In this example, the user has
decided to set the home screen metric (e.g. first metric), as time
or clock. Optionally, the user may then select the scroll order of
the various screens or GUIs (e.g., screens 200a-200h) to be
traversed. In this example, the user has selected step count metric
to follow, then stairs metric, then calories burned metric, and
then distance metric. The user may also be prompted or can elect to
edit, remove or add additional metrics to the scroll order setting.
Once the user approves of the settings, the user can save the
settings to the user profile. Saving the settings to the user
profile can act to update the settings to the user display scroll
order configuration 230 in the user account (user A). This
configuration setting is synchronized with the server 220 and then
transferred to the device 100 by way of a wireless connection over
a predefined proximity distance 404. As noted above, in one
embodiment, communication between the device 100 and the remote
device 200 (computing device) is by way of a wireless link. The
wireless link may be, for example Bluetooth radio communication,
and in one embodiment, low-energy Bluetooth radio
communication.
FIG. 5 illustrates a table providing a scroll order 352 and the
various inputs 354 that can be defined to traverse the scroll
order, in one embodiment. In this embodiment, the screen is off 300
and ready for use. When the screen is off, the device is on and
operating, but energy to power the screen does not consume battery
power. In this state, no physical contact has been received during
a period of time. When a double tap is received, the configuration
will allow the screen to transition to a first metric, which is the
time of day. As noted above, the first metric can be defined to be
any one of the metrics in the scroll order. Additionally, the
scroll order can remain in the same order yet the first metric can
be identified as any one of the metrics in the scroll order, and
access to the specific metric can be by way of a jump or shortcut.
In another embodiment, the scroll order configuration can be
reordered. In this example, transitioning from metric to next
metric is by way of double tap (DT) physical contact. However as
noted above, the transitioning between the metrics can be by way of
any number of inputs, such as physical and nonphysical interfaces,
voice communication, proximity communication, button presses,
etc.
FIG. 6 illustrates an example of a physical contact 124 represented
by taps, which act to activate the display 602 in a first metric.
The next input can be an additional tap or another suitable input
604, to transition to the next metric 606.
FIG. 7 illustrates another example where a single tap 124 activates
the display 602 in the first metric. In this example, the first
metric 1.0 is time. If within a period of time a double tap is
received 610, a sub-data metric 1.1 can be displayed. Sub-data
metric can be an additional metric that further describes or
provides additional information related to the first metric. For
example, sub-data metric 1.1 can be the date. Another double tap
612 can provide sub metric 1.2 which can be a calendar. Another
double tap 614 can provide another sub metric 1.3, which can be
sunrise metrics. A single tap 616 can then provide the next metric
2.0, which will define the metric data 617 associated with that
next metric. Accordingly, it should be understood that a specific
metric can have additional information (e.g., multiple levels of
information or formats) that relate to that same metric, and that
additional information can be provided by additional screen
displays of metric data that relate to one of the metrics.
FIG. 8 illustrates an example of a double tap physical contact 124
acting to initiate the display 602 with the first metric 1.0. A
single tap 620 can then expose or present sub-data 1.1, single tap
622 can present sub-data 1.2, single tap 624 can present sub-data
1.3. Similar to FIG. 7, the provides additional data/information
regarding a specific metric. However, in FIG. 8, the transitions
between the sub-data screens are by way of a single tap, as opposed
to a double tap. Thus a double tap 626 can then expose a next
metric 627, shown as metric 2.0. Metric 2.0 can also have sub-data
(not shown). Another double tap 626 can then expose the next metric
630, shown as metric 3.0. Each of the various metrics organized in
a scroll order can have sub-data associated therewith, if desired
for the specific configuration. These examples are provided to show
that physical contact input can be configured and quantified to
define a specific input for traversing a specific display screen,
or selecting specific metrics to show in specific screen
displays.
In some embodiments, a device is provided. The device is defined in
a form of a wearable wrist attachable structure. In one embodiment,
the device has a housing that is at least partially constructed or
formed from a plastic material. In one embodiment, the housing of
the device includes an altimeter. The defines can further include a
transiently visible display, or a dead-front display, a touch
screen display, a monochrome display, a digital display, a color
display, or combination thereof. In yet another embodiment, the
device can include one or more accelerometers. In one specific
example, the device can include a 3-axis accelerometer. On still
another embodiment, a 3-axis accelerometer can be replaced with or
replicated by use of separate accelerometers (e.g., 3
accelerometers) positioned orthogonally to each other.
FIG. 9 illustrates an example where various types of activities of
users 900A-900I can be captured by activity tracking devices 100,
in accordance with one embodiment of the present invention. As
shown, the various types of activities can generate different types
of data that can be captured by the activity tracking device 100.
The data, which can be represented as motion data (or processed
motion data) can be transferred 920 to a network 176 for processing
and saving by a server, as described above. In one embodiment, the
activity tracking device 100 can communicate to a device using a
wireless connection, and the device is capable of communicating and
synchronizing the captured data with an application running on the
server. In one embodiment, an application running on a local
device, such as a smart phone or tablet or smart watch can capture
or receive data from the activity tracking device 100 and represent
the tract motion data in a number of metrics.
In one embodiment, the device collects one or more types of
physiological and/or environmental data from embedded sensors
and/or external devices and communicates or relays such metric
information to other devices, including devices capable of serving
as Internet-accessible data sources, thus permitting the collected
data to be viewed, for example, using a web browser or
network-based application. For example, while the user is wearing
an activity tracking device, the device may calculate and store the
user's step count using one or more sensors. The device then
transmits data representative of the user's step count to an
account on a web service, computer, mobile phone, or health station
where the data may be stored, processed, and visualized by the
user. Indeed, the device may measure or calculate a plurality of
other physiological metrics in addition to, or in place of, the
user's step count.
Some physiological metrics include, but are not limited to, energy
expenditure (for example, calorie burn), floors climbed and/or
descended, heart rate, heart rate variability, heart rate recovery,
location and/or heading (for example, through GPS), elevation,
ambulatory speed and/or distance traveled, swimming lap count,
bicycle distance and/or speed, blood pressure, blood glucose, skin
conduction, skin and/or body temperature, electromyography,
electroencephalography, weight, body fat, caloric intake,
nutritional intake from food, medication intake, sleep periods
(i.e., clock time), sleep phases, sleep quality and/or duration, pH
levels, hydration levels, and respiration rate. The device may also
measure or calculate metrics related to the environment around the
user such as barometric pressure, weather conditions (for example,
temperature, humidity, pollen count, air quality, rain/snow
conditions, wind speed), light exposure (for example, ambient
light, UV light exposure, time and/or duration spent in darkness),
noise exposure, radiation exposure, and magnetic field.
Still further, other metrics can include, without limitation,
calories burned by a user, weight gained by a user, weight lost by
a user, stairs ascended, e.g., climbed, etc., by a user, stairs
descended by a user, steps taken by a user during walking or
running, a number of rotations of a bicycle pedal rotated by a
user, sedentary activity data, driving a vehicle, a number of golf
swings taken by a user, a number of forehands of a sport played by
a user, a number of backhands of a sport played by a user, or a
combination thereof. In some embodiments, sedentary activity data
is referred to herein as inactive activity data or as passive
activity data. In some embodiments, when a user is not sedentary
and is not sleeping, the user is active. In some embodiments, a
user may stand on a monitoring device that determines a
physiological parameter of the user. For example, a user stands on
a scale that measures a weight, a body fat percentage, a biomass
index, or a combination thereof, of the user.
Furthermore, the device or the system collating the data streams
may calculate metrics derived from this data. For example, the
device or system may calculate the user's stress and/or relaxation
levels through a combination of heart rate variability, skin
conduction, noise pollution, and sleep quality. In another example,
the device or system may determine the efficacy of a medical
intervention (for example, medication) through the combination of
medication intake, sleep and/or activity data. In yet another
example, the device or system may determine the efficacy of an
allergy medication through the combination of pollen data,
medication intake, sleep and/or activity data. These examples are
provided for illustration only and are not intended to be limiting
or exhaustive.
This information can be associated to the users account, which can
be managed by an activity management application on the server. The
activity management application can provide access to the users
account and data saved thereon. The activity manager application
running on the server can be in the form of a web application. The
web application can provide access to a number of websites screens
and pages that illustrate information regarding the metrics in
various formats. This information can be viewed by the user, and
synchronized with a computing device of the user, such as a smart
phone.
In one embodiment, the data captured by the activity tracking
device 100 is received by the computing device, and the data is
synchronized with the activity measured application on the server.
In this example, data viewable on the computing device (e.g. smart
phone) using an activity tracking application (app) can be
synchronized with the data present on the server, and associated
with the user's account. In this way, information entered into the
activity tracking application on the computing device can be
synchronized with application illustrated in the various screens of
the activity management application provided by the server on the
website.
The user can therefore access the data associated with the user
account using any device having access to the Internet. Data
received by the network 176 can then be synchronized with the
user's various devices, and analytics on the server can provide
data analysis to provide recommendations for additional activity,
and or improvements in physical health. The process therefore
continues where data is captured, analyzed, synchronized, and
recommendations are produced. In some embodiments, the captured
data can be itemized and partitioned based on the type of activity
being performed, and such information can be provided to the user
on the website via graphical user interfaces, or by way of the
application executed on the users smart phone (by way of graphical
user interfaces).
In an embodiment, the sensor or sensors of a device 100 can
determine or capture data to determine an amount of movement of the
monitoring device over a period of time. The sensors can include,
for example, an accelerometer, a magnetometer, a gyroscope, or
combinations thereof. Broadly speaking, these sensors are inertial
sensors, which capture some movement data, in response to the
device 100 being moved. The amount of movement (e.g., motion
sensed) may occur when the user is performing an activity of
climbing stairs over the time period, walking, running, etc. The
monitoring device may be worn on a wrist, carried by a user, worn
on clothing (using a clip, or placed in a pocket), attached to a
leg or foot, attached to the user's chest, waist, or integrated in
an article of clothing such as a shirt, hat, pants, blouse,
glasses, and the like. These examples are not limiting to all the
possible ways the sensors of the device can be associated with a
user or thing being monitored.
In other embodiments, a biological sensor can determine any number
of physiological characteristics of a user. As another example, the
biological sensor may determine heart rate, a hydration level, body
fat, bone density, fingerprint data, sweat rate, and/or a bio
impedance of the user. Examples of the biological sensors include,
without limitation, a biometric sensor, a physiological parameter
sensor, a pedometer, or a combination thereof.
In some embodiments, data associated with the user's activity can
be monitored by the applications on the server and the users
device, and activity associated with the user's friends,
acquaintances, or social network peers can also be shared, based on
the user's authorization. This provides for the ability for friends
to compete regarding their fitness, achieve goals, receive badges
for achieving goals, get reminders for achieving such goals,
rewards or discounts for achieving certain goals, etc.
As noted, an activity tracking device 100 can communicate with a
computing device (e.g., a smartphone, a tablet computer, a desktop
computer, or computer device having wireless communication access
and/or access to the Internet). The computing device, in turn, can
communicate over a network, such as the Internet or an Intranet to
provide data synchronization. The network may be a wide area
network, a local area network, or a combination thereof. The
network may be coupled to one or more servers, one or more virtual
machines, or a combination thereof. A server, a virtual machine, a
controller of a monitoring device, or a controller of a computing
device is sometimes referred to herein as a computing resource.
Examples of a controller include a processor and a memory
device.
In one embodiment, the processor may be a general purpose
processor. In another embodiment, the processor can be a customized
processor configured to run specific algorithms or operations. Such
processors can include digital signal processors (DSPs), which are
designed to execute or interact with specific chips, signals,
wires, and perform certain algorithms, processes, state diagrams,
feedback, detection, execution, or the like. In some embodiments, a
processor can include or be interfaced with an application specific
integrated circuit (ASIC), a programmable logic device (PLD), a
central processing unit (CPU), or a combination thereof, etc.
In some embodiments, one or more chips, modules, devices, or logic
can be defined to execute instructions or logic, which collectively
can be viewed or characterized to be a processor. Therefore, it
should be understood that a processor does not necessarily have to
be one single chip or module, but can be defined from a collection
of electronic or connecting components, logic, firmware, code, and
combinations thereof.
Examples of a memory device include a random access memory (RAM)
and a read-only memory (ROM). A memory device may be a Flash
memory, a redundant array of disks (RAID), a hard disk, or a
combination thereof.
Embodiments described in the present disclosure may be practiced
with various computer system configurations including hand-held
devices, microprocessor systems, microprocessor-based or
programmable consumer electronics, minicomputers, mainframe
computers and the like. Several embodiments described in the
present disclosure can also be practiced in distributed computing
environments where tasks are performed by remote processing devices
that are linked through a wire-based or wireless network.
With the above embodiments in mind, it should be understood that a
number of embodiments described in the present disclosure can
employ various computer-implemented operations involving data
stored in computer systems. These operations are those requiring
physical manipulation of physical quantities. Any of the operations
described herein that form part of various embodiments described in
the present disclosure are useful machine operations. Several
embodiments described in the present disclosure also relate to a
device or an apparatus for performing these operations. The
apparatus can be specially constructed for a purpose, or the
apparatus can be a computer selectively activated or configured by
a computer program stored in the computer. In particular, various
machines can be used with computer programs written in accordance
with the teachings herein, or it may be more convenient to
construct a more specialized apparatus to perform the required
operations.
Various embodiments described in the present disclosure can also be
embodied as computer-readable code on a non-transitory
computer-readable medium. The computer-readable medium is any data
storage device that can store data, which can thereafter be read by
a computer system. Examples of the computer-readable medium include
hard drives, network attached storage (NAS), ROM, RAM, compact
disc-ROMs (CD-ROMs), CD-recordables (CD-Rs), CD-rewritables (RWs),
magnetic tapes and other optical and non-optical data storage
devices. The computer-readable medium can include computer-readable
tangible medium distributed over a network-coupled computer system
so that the computer-readable code is stored and executed in a
distributed fashion.
Although the method operations were described in a specific order,
it should be understood that other housekeeping operations may be
performed in between operations, or operations may be performed in
an order other than that shown, or operations may be adjusted so
that they occur at slightly different times, or may be distributed
in a system which allows the occurrence of the processing
operations at various intervals associated with the processing.
Although the foregoing embodiments have been described in some
detail for purposes of clarity of understanding, it will be
apparent that certain changes and modifications can be practiced
within the scope of the appended claims. Accordingly, the present
embodiments are to be considered as illustrative and not
restrictive, and the various embodiments described in the present
disclosure are not to be limited to the details given herein, but
may be modified within the scope and equivalents of the appended
claims.
* * * * *