U.S. patent application number 13/741016 was filed with the patent office on 2013-05-16 for graphical user interface for presenting location information.
This patent application is currently assigned to APPLE INC.. The applicant listed for this patent is Apple Inc.. Invention is credited to Christopher Blumenberg, Gregory N. Christie, Mike Matas.
Application Number | 20130124091 13/741016 |
Document ID | / |
Family ID | 40521910 |
Filed Date | 2013-05-16 |
United States Patent
Application |
20130124091 |
Kind Code |
A1 |
Matas; Mike ; et
al. |
May 16, 2013 |
Graphical User Interface For Presenting Location Information
Abstract
Methods, systems, and apparatus, including computer program
products, for presenting location information. A first geographic
area in which a device is currently located is estimated using a
first positioning system. A second geographic area in which the
device is currently located is estimated using a second positioning
system. A first map view including an indication of the first
geographic area on the first map view is presented if the first
geographic area is contained within the second geographic area. A
second map view including an indication of the second geographic
area is presented if the first geographic area is not contained
within the second geographic area.
Inventors: |
Matas; Mike; (San Francisco,
CA) ; Christie; Gregory N.; (San Jose, CA) ;
Blumenberg; Christopher; (San Francisco, CA) |
|
Applicant: |
Name |
City |
State |
Country |
Type |
Apple Inc.; |
Cupertino |
CA |
US |
|
|
Assignee: |
APPLE INC.
Cupertino
CA
|
Family ID: |
40521910 |
Appl. No.: |
13/741016 |
Filed: |
January 14, 2013 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
11969901 |
Jan 6, 2008 |
8355862 |
|
|
13741016 |
|
|
|
|
Current U.S.
Class: |
701/533 ;
715/810; 715/859 |
Current CPC
Class: |
H04M 1/72544 20130101;
H04M 2250/12 20130101; G01S 19/48 20130101; G01C 21/00 20130101;
H04M 2250/22 20130101; G01C 21/3614 20130101; H04M 2250/10
20130101; G01C 21/3664 20130101; G06F 3/04886 20130101; H04M
1/72572 20130101; G01S 5/0263 20130101; G01C 21/367 20130101 |
Class at
Publication: |
701/533 ;
715/859; 715/810 |
International
Class: |
G01C 21/00 20060101
G01C021/00 |
Claims
1. A method comprising: providing for display a map view; receiving
a user input requesting an arbitrary position marker; and providing
for display the map view with the arbitrary position marker placed
at an arbitrary position on the map view, the arbitrary position
marker comprising: a flag providing information on the arbitrary
position on the map view; and a user selectable object, wherein the
method is performed by one or more computing devices.
2. The method of claim 1, comprising: receiving user input moving
the arbitrary position marker; and positioning the arbitrary
position marker at a new position based on the user input.
3. The method of claim 1, comprising: generating route information
for a route, wherein an endpoint of the route is a position on the
map view indicated by the arbitrary position marker; and presenting
the route information.
4. The method of claim 1, wherein: the arbitrary position is a
street, and the flag is a street flag that displays a name of the
street, the street flag including the user selectable object.
5. The method of claim 1, wherein the flag comprises a user
selectable arrow that, when selected, causes a user selectable menu
of options to be displayed.
6. The method of claim 5, wherein the user selectable menu of
options is a dropped pin menu operable to cause the position marker
to be saved as a bookmark in a bookmark list.
7. The method of claim 1, wherein the flag is configured to present
instructions or information about the arbitrary position.
8. A system, comprising: a portable device; a non-transitory
program carrier storing computer instructions operable to cause the
portable device to perform operations comprising: providing for
display a map view; receiving a user input requesting an arbitrary
position marker; and providing for display the map view with the
arbitrary position marker placed at an arbitrary position on the
map view, the arbitrary position marker comprising: a flag
providing information on the arbitrary position on the map view;
and a user selectable object.
9. The system of claim 8, the operations comprising: receiving user
input moving the arbitrary position marker; and positioning the
arbitrary position marker at a new position based on the user
input.
10. The system of claim 8, the operations comprising: generating
route information for a route, wherein an endpoint of the route is
a position on the map view indicated by the arbitrary position
marker; and presenting the route information.
11. The system of claim 8, wherein: the arbitrary position is a
street, and the flag is a street flag that displays a name of the
street, the street flag including the user selectable object.
12. The system of claim 8, wherein the flag comprises a user
selectable arrow that, when selected, causes a user selectable menu
of options to be displayed.
13. The system of claim 12, wherein the user selectable menu of
options is a dropped pin menu operable to cause the position marker
to be saved as a bookmark in a bookmark list.
14. The system of claim 8, wherein the flag is configured to
present instructions or information about the arbitrary
position.
15. A non-transitory program carrier storing computer instructions
operable to cause a portable device to perform operations
comprising: providing for display a map view; receiving a user
input requesting an arbitrary position marker; and providing for
display the map view with the arbitrary position marker placed at
an arbitrary position on the map view, the arbitrary position
marker comprising: a flag providing information on the arbitrary
position on the map view; and a user selectable object.
16. The non-transitory program carrier of claim 15, the operations
comprising: receiving user input moving the arbitrary position
marker; and positioning the arbitrary position marker at a new
position based on the user input.
17. The non-transitory program carrier of claim 15, the operations
comprising: generating route information for a route, wherein an
endpoint of the route is a position on the map view indicated by
the arbitrary position marker; and presenting the route
information.
18. The non-transitory program carrier of claim 15, wherein: the
arbitrary position is a street, and the flag is a street flag that
displays a name of the street, the street flag including the user
selectable object.
19. The non-transitory program carrier of claim 15, wherein the
flag comprises a user selectable arrow that, when selected, causes
a user selectable menu of options to be displayed.
20. The non-transitory program carrier of claim 19, wherein the
user selectable menu of options is a dropped pin menu operable to
cause the position marker to be saved as a bookmark in a bookmark
list.
21. The non-transitory program carrier of claim 15, wherein the
flag is configured to present instructions or information about the
arbitrary position.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application is a divisional of and claims priority to
U.S. application Ser. No. 11/969,901, entitled "Graphical User
Interface for Presenting Location Information," filed Jan. 6, 2008,
the entire contents of which is incorporated herein by
reference.
BACKGROUND
[0002] This specification is related generally to graphical user
interfaces.
[0003] The role of traditional printed maps is being supplanted by
modern devices capable of rendering dynamic map displays. Devices
that include mapping or navigation applications provide information
regarding an area selected by a user by recalling map data from
local memory or networked services.
[0004] When coupled with any of a number of positioning
technologies, a mapping device can display a current position on a
map as well as deliver navigation instructions based on the current
position to route a user to a desired destination. Positioning
technologies include satellite positioning systems, such as the
Global Positioning System (GPS), cell tower triangulation and
network-based positioning systems (e.g., Wi-Fi positioning system).
Positioning technologies provide varying degrees of accuracy or
precision. For example, GPS is generally accurate to a few meters,
while other positioning technologies are generally less
accurate.
SUMMARY
[0005] In general, one aspect of the subject matter described in
this specification can be embodied in methods that include the
actions of estimating a first geographic area in which a device is
currently located using a first positioning system, estimating a
second geographic area in which the device is currently located
using a second positioning system, presenting a first map view
including an indication of the first geographic area on the first
map view if the first geographic area is contained within the
second geographic area, and presenting a second map view including
an indication of the second geographic area if the first geographic
area is not contained within the second geographic area. Other
embodiments of this aspect include corresponding systems,
apparatus, computer program products, and computer readable
media.
[0006] In general, another aspect of the subject matter described
in this specification can be embodied in methods that include the
actions of presenting a map view, receiving user input requesting
an arbitrary position marker, and presenting the map view with the
arbitrary position marker placed at an arbitrary position on the
map view. Other embodiments of this aspect include corresponding
systems, apparatus, computer program products, and computer
readable media.
[0007] In general, another aspect of the subject matter described
in this specification can be embodied in methods that include the
actions of presenting a map view on a first virtual graphical
layer, receiving a request to present other content, presenting a
flipped-page graphical effect with respect to the map view, where
the flipped-page effect simulates a partial peeling of the first
virtual graphical layer to reveal a second virtual graphical layer,
and presenting the other content on the second virtual graphical
layer. Other embodiments of this aspect include corresponding
systems, apparatus, computer program products, and computer
readable media.
[0008] In general, another aspect of the subject matter described
in this specification can be embodied in methods that include the
actions of presenting a map view, and presenting an indication of a
region on the map view, where the region represents a geographic
area in which a device is located. Other embodiments of this aspect
include corresponding systems, apparatus, computer program
products, and computer readable media.
[0009] In general, another aspect of the subject matter described
in this specification can be embodied in graphical user interfaces
that include a map view and an indication of a region on the map
view, where the region represents a geographic area in which a
device is located, and where the indication includes a circle and
one or more crosshairs. Other embodiments of this aspect include
corresponding methods, systems, apparatus, computer program
products, and computer readable media.
[0010] Particular embodiments of the subject matter described in
this specification can be implemented to realize one or more of the
following advantages. The area of uncertainty associated with a
current position determined using a positioning technology can be
indicated on a map. Geographic areas in which a device is currently
located can be determined or estimated using one or more
positioning technologies and the most accurate or precise
geographic area can be indicated on a map displayed on a mobile
device, for example.
[0011] The details of one or more embodiments of the subject matter
described in this specification are set forth in the accompanying
drawings and the description below. Other features, aspects, and
advantages of the subject matter will become apparent from the
description, the drawings, and the claims.
BRIEF DESCRIPTION OF THE DRAWINGS
[0012] FIG. 1 is a block diagram of an example mobile device.
[0013] FIG. 2 is a block diagram of an example network operating
environment for the mobile device of FIG. 1.
[0014] FIG. 3 is a block diagram of an example implementation of
the mobile device of FIG. 1.
[0015] FIG. 4 illustrates an example implementation of a navigation
system.
[0016] FIG. 5A illustrates an example of a map with a geographic
area displayed on the mobile device of FIG. 1.
[0017] FIG. 5B illustrates another example of the map from FIG.
5A.
[0018] FIG. 6 illustrates a zoomed out view of the map from FIG.
5A.
[0019] FIG. 7 illustrates another view of the map from FIG. 5A.
[0020] FIG. 8 illustrates another view of map with a larger
geographic area than shown in FIG. 6.
[0021] FIG. 9 illustrates an example of a user interface for
getting directions.
[0022] FIG. 10 illustrates an example of route information
displayed on a map.
[0023] FIG. 11 illustrates an example menu in a mapping
application.
[0024] FIG. 12 shows an example of a map with a position marker
dropped onto an arbitrary position on the map.
[0025] FIG. 13 illustrates a dropped position marker menu.
[0026] FIG. 14 illustrates another example menu in a mapping
application.
[0027] FIG. 15 illustrates another example of a user interface for
getting directions.
[0028] FIG. 16 illustrates a dropped pin in the center of a
geographic area.
[0029] Like reference numbers and designations in the various
drawings indicate like elements.
DETAILED DESCRIPTION
Example Mobile Device
[0030] FIG. 1 is a block diagram of an example mobile device 100.
The mobile device 100 can be, for example, a handheld computer, a
personal digital assistant, a cellular telephone, a network
appliance, a camera, a smart phone, an enhanced general packet
radio service (EGPRS) mobile phone, a network base station, a media
player, a navigation device, an email device, a game console, or a
combination of any two or more of these data processing devices or
other data processing devices.
Mobile Device Overview
[0031] In some implementations, the mobile device 100 includes a
touch-sensitive display 102. The touch-sensitive display 102 can
implement liquid crystal display (LCD) technology, light emitting
polymer display (LPD) technology, or some other display technology.
The touch sensitive display 102 can be sensitive to haptic and/or
tactile contact with a user.
[0032] In some implementations, the touch-sensitive display 102 can
comprise a multi-touch-sensitive display 102. A
multi-touch-sensitive display 102 can, for example, process
multiple simultaneous touch points, including processing data
related to the pressure, degree, and/or position of each touch
point. Such processing facilitates gestures and interactions with
multiple fingers, chording, and other interactions. Other
touch-sensitive display technologies can also be used, e.g., a
display in which contact is made using a stylus or other pointing
device. Some examples of multi-touch-sensitive display technology
are described in U.S. Pat. Nos. 6,323,846, 6,570,557, 6,677,932,
and 6,888,536, each of which is incorporated by reference herein in
its entirety.
[0033] In some implementations, the mobile device 100 can display
one or more graphical user interfaces on the touch-sensitive
display 102 for providing the user access to various system objects
and for conveying information to the user. In some implementations,
the graphical user interface can include one or more display
objects 104, 106. In the example shown, the display objects 104,
106, are graphic representations of system objects. Some examples
of system objects include device functions, applications, windows,
files, alerts, events, or other identifiable system objects.
Example Mobile Device Functionality
[0034] In some implementations, the mobile device 100 can implement
multiple device functionalities, such as a telephony device, an
e-mail device, a network data communication device, a Wi-Fi base
station device (not shown), and a media processing device. In some
implementations, particular display objects 104 can be displayed in
a menu bar 118. In some implementations, device functionalities can
be accessed from a top-level graphical user interface, such as the
graphical user interface illustrated in FIG. 1. Touching one of the
objects 104 can, for example, invoke corresponding
functionality.
[0035] In some implementations, the mobile device 100 can implement
network distribution functionality. For example, the functionality
can enable the user to take the mobile device 100 and provide
access to its associated network while traveling. In particular,
the mobile device 100 can extend Internet access (e.g., Wi-Fi) to
other wireless devices in the vicinity. For example, mobile device
100 can be configured as a base station for one or more devices. As
such, mobile device 100 can grant or deny network access to other
wireless devices.
[0036] In some implementations, upon invocation of device
functionality, the graphical user interface of the mobile device
100 changes, or is augmented or replaced with another user
interface or user interface elements, to facilitate user access to
particular functions associated with the corresponding device
functionality. For example, in response to a user touching a phone
object, the graphical user interface of the touch-sensitive display
102 may present display objects related to various phone functions;
likewise, touching of an email object may cause the graphical user
interface to present display objects related to various e-mail
functions; touching a Web object may cause the graphical user
interface to present display objects related to various Web-surfing
functions; and touching a media player object may cause the
graphical user interface to present display objects related to
various media processing functions.
[0037] In some implementations, the top-level graphical user
interface environment or state of FIG. 1 can be restored by
pressing a button 120 located near the bottom of the mobile device
100. In some implementations, each corresponding device
functionality may have corresponding "home" display objects
displayed on the touch-sensitive display 102, and the graphical
user interface environment of FIG. 1 can be restored by pressing
the "home" display object.
[0038] In some implementations, the top-level graphical user
interface can include additional display objects 106, such as a
short messaging service (SMS) object, a calendar object, a photos
object, a camera object, a calculator object, a stocks object, a
weather object, a maps object 144, a notes object, a clock object,
an address book object, and a settings object. Touching the maps
object 144 can, for example, invoke a mapping and location-based
services environment and supporting functionality; likewise, a
selection of any of the display objects 106 can invoke a
corresponding object environment and functionality.
[0039] Additional and/or different display objects can also be
displayed in the graphical user interface of FIG. 1. For example,
if the device 100 is functioning as a base station for other
devices, one or more "connection" objects may appear in the
graphical user interface to indicate the connection. In some
implementations, the display objects 106 can be configured by a
user, e.g., a user may specify which display objects 106 are
displayed, and/or may download additional applications or other
software that provides other functionalities and corresponding
display objects.
[0040] In some implementations, the mobile device 100 can include
one or more input/output (I/O) devices and/or sensor devices. For
example, a speaker 160 and a microphone 162 can be included to
facilitate voice-enabled functionalities, such as phone and voice
mail functions. In some implementations, an up/down button 184 for
volume control of the speaker 160 and the microphone 162 can be
included. The mobile device 100 can also include an on/off button
182 for a ring indicator of incoming phone calls. In some
implementations, a loud speaker 164 can be included to facilitate
hands-free voice functionalities, such as speaker phone functions.
An audio jack 166 can also be included for use of headphones and/or
a microphone.
[0041] In some implementations, a proximity sensor 168 can be
included to facilitate the detection of the user positioning the
mobile device 100 proximate to the user's ear and, in response, to
disengage the touch-sensitive display 102 to prevent accidental
function invocations. In some implementations, the touch-sensitive
display 102 can be turned off to conserve additional power when the
mobile device 100 is proximate to the user's ear.
[0042] Other sensors can also be used. For example, in some
implementations, an ambient light sensor 170 can be utilized to
facilitate adjusting the brightness of the touch-sensitive display
102. In some implementations, an accelerometer 172 can be utilized
to detect movement of the mobile device 100, as indicated by the
directional arrow 174. Accordingly, display objects and/or media
can be presented according to a detected orientation, e.g.,
portrait or landscape. In some implementations, the mobile device
100 may include circuitry and sensors for supporting a location
determining capability, such as that provided by the Global
Positioning System (GPS) or other positioning systems (e.g.,
systems using Wi-Fi access points, television signals, cellular
grids, Uniform Resource Locators (URLs)). In some implementations,
a positioning system (e.g., a GPS receiver) can be integrated into
the mobile device 100 or provided as a separate device that can be
coupled to the mobile device 100 through an interface (e.g., port
device 190) to provide access to location-based services.
[0043] In some implementations, a port device 190, e.g., a
Universal Serial Bus (USB) port, or a docking port, or some other
wired port connection, can be included. The port device 190 can,
for example, be utilized to establish a wired connection to other
computing devices, such as other communication devices 100, network
access devices, a personal computer, a printer, a display screen,
or other processing devices capable of receiving and/or
transmitting data. In some implementations, the port device 190
allows the mobile device 100 to synchronize with a host device
using one or more protocols, such as, for example, the TCP/IP,
HTTP, UDP and any other known protocol.
[0044] The mobile device 100 can also include a camera lens and
sensor 180. In some implementations, the camera lens and sensor 180
can be located on the back surface of the mobile device 100. The
camera can capture still images and/or video.
[0045] The mobile device 100 can also include one or more wireless
communication subsystems, such as an 802.11b/g communication device
186, and/or a Bluetooth.TM. communication device 188. Other
communication protocols can also be supported, including other
802.x communication protocols (e.g., WiMax, Wi-Fi, 3G), code
division multiple access (CDMA), global system for mobile
communications (GSM), Enhanced Data GSM Environment (EDGE),
etc.
Network Operating Environment
[0046] FIG. 2 is a block diagram of an example network operating
environment 200. In FIG. 2, mobile devices 202a and 202b each can
represent mobile device 100. Mobile devices 202a and 202b can, for
example, communicate over one or more wired and/or wireless
networks 210 in data communication. For example, a wireless network
212, e.g., a cellular network, can communicate with a wide area
network (WAN) 214, such as the Internet, by use of a gateway 216.
Likewise, an access device 218, such as an 802.11g wireless access
device, can provide communication access to the wide area network
214. In some implementations, both voice and data communications
can be established over the wireless network 212 and the access
device 218. For example, the mobile device 202a can place and
receive phone calls (e.g., using VoIP protocols), send and receive
e-mail messages (e.g., using POP3 protocol), and retrieve
electronic documents and/or streams, such as web pages,
photographs, and videos, over the wireless network 212, gateway
216, and wide area network 214 (e.g., using TCP/IP or UDP
protocols). Likewise, in some implementations, the mobile device
202b can place and receive phone calls, send and receive e-mail
messages, and retrieve electronic documents over the access device
218 and the wide area network 214. In some implementations, the
mobile device 202a or 202b can be physically connected to the
access device 218 using one or more cables and the access device
218 can be a personal computer. In this configuration, the mobile
device 202a or 202b can be referred to as a "tethered" device.
[0047] The mobile devices 202a and 202b can also establish
communications by other means. For example, the wireless device
202a can communicate with other wireless devices, e.g., other
mobile devices 202a or 202b, cell phones, etc., over the wireless
network 212. Likewise, the mobile devices 202a and 202b can
establish peer-to-peer communications 220, e.g., a personal area
network, by use of one or more communication subsystems, such as
the Bluetooth.TM. communication devices 188 shown in FIG. 1. Other
communication protocols and topologies can also be implemented.
[0048] The mobile device 202a or 202b can, for example, communicate
with one or more services 230, 240, 250, 260, and 270 over the one
or more wired and/or wireless networks 210. For example, one or
more navigation services 230 can provide navigation information,
e.g., map information, location information, route information, and
other information, to the mobile device 202a or 202b. A user of the
mobile device 202b can invoke a map functionality, e.g., by
pressing the maps object 144 on the top-level graphical user
interface shown in FIG. 1, and can request and receive a map for a
particular location, request and receive route directions, or
request and receive listings of businesses in the vicinity of a
particular location, for example.
[0049] A messaging service 240 can, for example, provide e-mail
and/or other messaging services. A media service 250 can, for
example, provide access to media files, such as song files, audio
books, movie files, video clips, and other media data. In some
implementations, separate audio and video services (not shown) can
provide access to the respective types of media files. A syncing
service 260 can, for example, perform syncing services (e.g., sync
files). An activation service 270 can, for example, perform an
activation process for activating the mobile device 202a or 202b.
Other services can also be provided, including a software update
service that automatically determines whether software updates
exist for software on the mobile device 202a or 202b, then
downloads the software updates to the mobile device 202a or 202b
where the software updates can be manually or automatically
unpacked and/or installed.
[0050] The mobile device 202a or 202b can also access other data
and content over the one or more wired and/or wireless networks
210. For example, content publishers, such as news sites, RSS
feeds, web sites, blogs, social networking sites, developer
networks, etc., can be accessed by the mobile device 202a or 202b.
Such access can be provided by invocation of a web browsing
function or application (e.g., a browser) in response to a user
touching, for example, a Web object.
Example Mobile Device Architecture
[0051] FIG. 3 is a block diagram 300 of an example implementation
of the mobile device 100 of FIG. 1. The mobile device 100 can
include a memory interface 302, one or more data processors, image
processors and/or central processing units 304, and a peripherals
interface 306. The memory interface 302, the one or more processors
304 and/or the peripherals interface 306 can be separate components
or can be integrated in one or more integrated circuits. The
various components in the mobile device 100 can be coupled by one
or more communication buses or signal lines.
[0052] Sensors, devices, and subsystems can be coupled to the
peripherals interface 306 to facilitate multiple functionalities.
For example, a motion sensor 310, a light sensor 312, and a
proximity sensor 314 can be coupled to the peripherals interface
306 to facilitate the orientation, lighting, and proximity
functions described with respect to FIG. 1. Other sensors 316 can
also be connected to the peripherals interface 306, such as a
positioning system (e.g., GPS receiver), a temperature sensor, a
biometric sensor, or other sensing device, to facilitate related
functionalities.
[0053] A camera subsystem 320 and an optical sensor 322, e.g., a
charged coupled device (CCD) or a complementary metal-oxide
semiconductor (CMOS) optical sensor, can be utilized to facilitate
camera functions, such as recording photographs and video
clips.
[0054] Communication functions can be facilitated through one or
more wireless communication subsystems 324, which can include radio
frequency receivers and transmitters and/or optical (e.g.,
infrared) receivers and transmitters. The specific design and
implementation of the communication subsystem 324 can depend on the
communication network(s) over which the mobile device 100 is
intended to operate. For example, a mobile device 100 may include
communication subsystems 324 designed to operate over a GSM
network, a GPRS network, an EDGE network, a Wi-Fi or WiMax network,
and a Bluetooth.TM. network. In particular, the wireless
communication subsystems 324 may include hosting protocols such
that the device 100 may be configured as a base station for other
wireless devices.
[0055] An audio subsystem 326 can be coupled to a speaker 328 and a
microphone 330 to facilitate voice-enabled functions, such as voice
recognition, voice replication, digital recording, and telephony
functions.
[0056] The I/O subsystem 340 can include a touch screen controller
342 and/or other input controller(s) 344. The touch-screen
controller 342 can be coupled to a touch screen 346. The touch
screen 346 and touch screen controller 342 can, for example, detect
contact and movement or break thereof using any of a plurality of
touch sensitivity technologies, including but not limited to
capacitive, resistive, infrared, and surface acoustic wave
technologies, as well as other proximity sensor arrays or other
elements for determining one or more points of contact with the
touch screen 346.
[0057] The other input controller(s) 344 can be coupled to other
input/control devices 348, such as one or more buttons, rocker
switches, thumb-wheel, infrared port, USB port, and/or a pointer
device such as a stylus. The one or more buttons (not shown) can
include an up/down button for volume control of the speaker 328
and/or the microphone 330.
[0058] In one implementation, a pressing of the button for a first
duration may disengage a lock of the touch screen 346; and a
pressing of the button for a second duration that is longer than
the first duration may turn power to the mobile device 100 on or
off. The user may be able to customize a functionality of one or
more of the buttons. The touch screen 346 can, for example, also be
used to implement virtual or soft buttons and/or a keyboard.
[0059] In some implementations, the mobile device 100 can present
recorded audio and/or video files, such as MP3, AAC, and MPEG
files. In some implementations, the mobile device 100 can include
the functionality of an MP3 player, such as an iPod.TM.. The mobile
device 100 may, therefore, include a 36-pin connector that is
compatible with the iPod. Other input/output and control devices
can also be used.
[0060] The memory interface 302 can be coupled to memory 350. The
memory 350 can include high-speed random access memory and/or
non-volatile memory, such as one or more magnetic disk storage
devices, one or more optical storage devices, and/or flash memory
(e.g., NAND, NOR). The memory 350 can store an operating system
352, such as Darwin, RTXC, LINUX, UNIX, OS X, WINDOWS, or an
embedded operating system such as VxWorks. The operating system 352
may include instructions for handling basic system services and for
performing hardware dependent tasks. In some implementations, the
operating system 352 can be a kernel (e.g., UNIX kernel).
[0061] The memory 350 may also store communication instructions 354
to facilitate communicating with one or more additional devices,
one or more computers and/or one or more servers. The memory 350
may include graphical user interface instructions 356 to facilitate
graphic user interface processing; sensor processing instructions
358 to facilitate sensor-related processing and functions; phone
instructions 360 to facilitate phone-related processes and
functions; electronic messaging instructions 362 to facilitate
electronic-messaging related processes and functions; web browsing
instructions 364 to facilitate web browsing-related processes and
functions; media processing instructions 366 to facilitate media
processing-related processes and functions; GPS/Navigation
instructions 368 to facilitate GPS and navigation-related processes
and instructions; camera instructions 370 to facilitate
camera-related processes and functions; and/or other software
instructions 372 to facilitate other processes and functions, e.g.,
security processes and functions. The memory 350 may also store
other software instructions (not shown), such as web video
instructions to facilitate web video-related processes and
functions; and/or web shopping instructions to facilitate web
shopping-related processes and functions. In some implementations,
the media processing instructions 366 are divided into audio
processing instructions and video processing instructions to
facilitate audio processing-related processes and functions and
video processing-related processes and functions, respectively. An
activation record and International Mobile Equipment Identity
(IMEI) 374 or similar hardware identifier can also be stored in
memory 350.
[0062] Each of the above identified instructions and applications
can correspond to a set of instructions for performing one or more
functions described above. These instructions need not be
implemented as separate software programs, procedures, or modules.
The memory 350 can include additional instructions or fewer
instructions. Furthermore, various functions of the mobile device
100 may be implemented in hardware and/or in software, including in
one or more signal processing and/or application specific
integrated circuits.
Example Navigation System
[0063] FIG. 4 illustrates an example implementation of a navigation
system 400. In FIG. 4, mobile device 402 can represent mobile
device 100. The mobile device 402 can, for example, communicate to
one or more network access points 404 (e.g., Wi-Fi base station
devices) or one or more cell towers 406. In some implementations,
the access points 404 can be any combination of 802.11b/g wireless
routers, 802.11n wireless routers, and some other Wi-Fi devices
that implement any suitable Wi-Fi or other wireless networking
technology or protocol. Using the communication with the access
points 404 or the cell towers 406, a location-based service 408
(Location-Based Service A) or a location-based service 410
(Location-Based Service B) can estimate geographic areas where the
mobile device 402 is currently located. The actual location of the
mobile device 402 can be anywhere within the estimated geographic
area. An estimated geographic area is not necessarily circular but
can be indicated as a circular area on a map display for
convenience.
[0064] The mobile device 402 can, for example, receive a
communication 412A from an access point 404 (e.g., a Wi-Fi access
point). The communication 412A can include information about the
access point 404, e.g., the Internet Protocol (IP) address and/or
the Media Access Control (MAC) address of the access point 404. The
communication 412A can include other information, such as the
latitude and longitude of the access point 404. The information
received in communication 412A can be sent to the location-based
service 408 in communication 412B. The location-based service 408
can, for example, with a degree of uncertainty or error, estimate a
first geographic area in which the mobile device 402 is currently
located using the information sent in communication 412B. In some
implementations, the location-based service 408 is a system or
service that estimates, with some degree of uncertainty or error,
the position of a device using a database of access points mapped
to geographic locations. The accuracy or precision (or the degree
of uncertainty or error) of the estimated position can, for
example, be based on the range of the technology, the accuracy of
the range, or some other metric. Accuracy or precision of an
estimated position can be affected by one or more factors
including, for example, inherent properties or limitations of the
technology or system, and a level of deployment of the given
technology or system (e.g., number of access points or cell towers
in the vicinity of the device).
[0065] In some implementations, the accuracy or precision of the
estimated position is stated in units of distance (e.g., "the
estimated position is accurate up to 50 meters"). That is, the
actual position of the mobile device 402 can be within the accuracy
distance from the estimated position. For example, the first
geographic area can be a circle centered at the latitude and
longitude of the estimated position with a radius equal to the
stated accuracy or precision (e.g. 38 meters if the accuracy of the
estimated position is up to 38 meters). The first geographic area
can alternatively be represented on a map display as a square,
rectangle, oval, diamond, triangle, or some other shaped enclosed
region.
[0066] In some other implementations, unique signatures of multiple
access points (e.g. five or more) can be compared to a local cache
on the mobile device 402 or a central reference database at
location-based service 408 via network communication (e.g.
communication 412B can be sent to the location-based service 408).
The location-based service 408 can use the unique signatures to
estimate the latitude and longitude of the center of the first
geographic circle with an m meter radius (e.g., about 20
meters).
[0067] In some implementations, location-based service 408 includes
positioning services and reference database information provided by
SKYHOOK WIRELESS of Boston, Mass.
[0068] The mobile device 402 can receive a communication 414A from
cell tower 406. The cell communication 414A can include, for
example, information identifying the cell tower 406. In some
implementations, the cell communication 414A can also include the
latitude and longitude of the cell tower 406. The identifying
information and/or the latitude and longitude of the cell tower 406
can be sent to the location-based service 410 in communication
414B. The location-based service 410 can estimate a position of the
mobile device 402 using the information included in communication
414B and estimate an accuracy of the estimate position. Thus, for
example, the location-based service 410 can estimate a second
geographic area in which the mobile device 402 is currently
located. In some implementations, the second geographic area is
represented on a map as a circular region centered at the estimated
position and with a radius equal to the accuracy of the estimated
position. In some other implementations, the second geographic area
can represented on a map by a square or rectangular shaped enclosed
region, to name a few additional examples.
[0069] In some implementations, the position and the geographic
area of the mobile device 402 can be estimated using a "cell of
origin" positioning technology. In some other implementations, the
second geographic area can be determined by cell tower
triangulation.
[0070] The first and second geographic areas can be sent to the
mobile device 402 by communications 416 and 418, respectively. The
mobile device 402 can present, on the touch-sensitive display 102
for example, a map view including an indication of one of the
geographic areas if that one is entirely contained within the other
geographic area (including the case where the boundaries of the
geographic areas touch but otherwise no part of the one of the
geographic area is outside of the other geographic area). For
example, a map view that includes an indication of the first
geographic area can be presented if the first geographic area is
completely contained within the second geographic area. The mobile
device 402 can present a map view that includes an indication of
the second geographic area if the first geographic area is not
completely contained within the second geographic area. For
example, if the technology, system, or service used to estimate the
first geographic area has a greater level of accuracy or precision
detail than the technology, system, or service used to estimate the
second geographic area, and the first geographic area is determined
to be more accurate or precise based on the first geographic area
being entirely within the second geographic area, the mobile device
402 can use the first geographic area. The level of detail of the
technology can, for example, be determined by a priority table
stored in the mobile device 402.
[0071] In another example, the first geographic area can be
estimated using the access point 404, and the second geographic
area can be estimated using the cell tower 406. In this example,
the access point 404 was moved from its original location and the
latitude and longitude coordinates of the access point 404 as they
are known to location-based service 408 are incorrect and the
mobile device 402 is not actually located within the first
geographic area. The cell tower 406 provided correct location-based
data and the mobile device 402 is located within the second
geographic area. In this situation, the first and second geographic
areas do not overlap, so the mobile device 402 would display a map
view indicating the second geographic area. The mobile device 402
can be used to determine the accuracy of the first and second
geographic areas, or an external server or servers can be used.
[0072] The location-based service 408 and location-based service
410 can run on the same device or on separate devices. For example,
the location-based services 408 and 410 can run on servers
communicating with the mobile device 100 through a network (e.g.,
WAN 214). The servers can be separate servers or the same server.
The location-based services 408 and 410 can alternatively run on
the mobile device 402.
[0073] The mobile device 402 can, for example, connect to
additional devices or services (not shown) for location-based
services, instead of, or in addition to the access point 404 and
the cell tower 406. Such devices or services could include a
Bluetooth.TM. device, GPS, radio or TV towers, or cellular grids,
to name a few examples. For example, the mobile device 402 can
connect to peer devices with the Bluetooth.TM. communication device
188 (FIG. 1) and receive location-based information from other
mobile devices and/or Bluetooth.TM. enabled devices. In some
implementations, the mobile device 402 can determine or estimate
its position and/or geographic area using other technologies (e.g.,
GPS). In some implementations, the geographic area determined or
estimated using any of these other technologies can be used (e.g.,
displayed) in lieu of the geographic area estimated using
location-based services 408 or 410 (.e.g., Wi-Fi or cellular
positioning techniques) if the geographic area determined or
estimated using the other technologies is contained entirely within
the geographic area estimated using location-based services 408 or
410 and if the other technologies are more accurate or precise
according to the priority table stored in the mobile device
402.
Example Map View with Indication of Current Geographic Area
[0074] FIG. 5A illustrates an example of a map 502 with a
geographic area displayed on mobile device 100. In some
implementations, the mobile device 100 can display the map 502 on
the touch sensitive display 102 of mobile device 100. The map 502
can be displayed when a user selects the maps object 144 to view
mapping and location based services. In some implementations,
objects, such as the maps object 144 (FIG. 1), can be selected by
voice activation. A search bar 504 and a bookmarks list object 506
can be displayed at the top of the map 502. Below the bottom of the
map one or more display objects can be displayed, for example a
search object 508, a directions object 510, a map view object 512,
and a current location object 514.
[0075] The search bar 504 can be used to find an address or other
location on the map. For example, a user can enter their home
address in the search bar 504, and the region containing the
address would be displayed on the map 502. The bookmarks list
object 506 can, for example, bring up a Bookmarks list which
contains addresses that are frequently visited, such as a user's
home address. The Bookmarks list can also, for example, contain
special bookmarks such as the current location, e.g. the current
location of the mobile device 100.
[0076] The search object 508 can be used to display the search bar
504 and other map related search menus. The directions object 510
can, for example, bring up a menu interface that allows the user to
enter a start and end location and then displays information for a
route from the start location to the end location, e.g. directions
and travel time. The map view object 512 can bring up a menu that
will allow the user to select display options for the map 502. An
example of a menu that can be displayed when map view 512 is
selected is further described below in reference to FIGS. 11 and
14. The map 502 could be changed from black and white to color, the
background of the map could be changed, or the user could change
the brightness of the map, to name a few examples.
[0077] The current location object 514 can allow the user to see a
geographic area 516 on the map 502 indicating where the device 100
is currently located. The special current location bookmark can be
placed in the Bookmarks list when the current location object 514
is selected. If the special current location bookmark was
previously set in the Bookmarks list, the old bookmark information
can, for example, be replaced with the new current location
information. In some implementations, the special current location
bookmark is tied to the centroid of the geographic area 516. That
is, the special current location bookmark includes the address for
the centroid of the geographic area 516. The geographic area 516
can be based on location data determined or estimated using
location-based services 408 or 410, or some other location data,
such as the data previously described in reference to FIG. 4. The
geographic area 516 can, for example, be depicted by a circle,
rectangle, square, or other enclosed region with crosshairs, or
some other distinctive element to differentiate the geographic area
516 from the map 502.
[0078] In some implementations, the geographic area 516 indicates a
region in which the mobile device 100 is determined or estimated to
be located, and the geographic area may not necessarily be centered
on the actual current position of the mobile device 100. In this
example, the mobile device 100 may be located off-center within the
geographic area. In another example, the geographic area 516 can be
centered on an estimated current position of the mobile device
100.
[0079] The mobile device 100 can, for example, center the map view
on the geographic area 516 when the current location object 514 is
tapped or otherwise selected. In some implementations, the zoom
level of the map can be adjusted based on the accuracy or precision
of the location data or the technology, system, or service that
provided the location data. For example, the map can be zoomed in
for higher accuracy GPS location data and zoomed out for lower
accuracy cell tower or Wi-Fi location data. In another
implementation, the zoom level can be based on the velocity of the
mobile device 100, e.g. the map can be zoomed out at higher
velocities and zoomed in when the mobile device 100 is not moving.
A combination of accuracy or precision and velocity can also be
used. If all methods for retrieving location-based data fail, e.g.
both location-based service A 408 and location-based service B 410
fail and there are no other systems or services available for
determining or estimating the current position of the mobile device
100, an error can be displayed to the user and no geographic area
is displayed on the map 502. The error can, for example, contain a
message to the user informing them of the failure and the possible
reason for the failure.
[0080] The current location object 514 can be selected, for
example, to activate the estimation and displaying of the
geographic area 516 on the map 502, to get directions to or from
the estimated current location (i.e., the centroid of the
geographic area 516), to send the estimated current location of the
mobile device 100 to a friend (e.g. so that the friend can go to
the same location), or to create a bookmark for the estimated
current location, to name a few examples.
[0081] FIG. 5B illustrates another example of the map 502 with a
geographic area 518 on the mobile device 100. In this example the
geographic area 518 is shown as a circle with ticks on the
perimeter of the circle. Other distinctive features could also be
used to differentiate the geographic area 518 from the remainder of
the map 502, such as a square, a triangle, or a shaded region. The
perimeter of the geographic area 518 can, for example, have
crosshairs, tick marks, shading, or some other feature to make it
stand out from the map 502. In another example, the geographic area
518 can be shaded and/or highlighted, or the area of the map 502
outside of the geographic area 518 can be shaded and/or
highlighted.
[0082] The circle outlining the geographic area 518 can, for
example, have animation, or be motionless. The circle can, for
example, change color (e.g. user defined colors or pre-programmed
colors on the mobile device 100), change distinctive features (e.g.
switch between crosshairs and ticks), or have moving distinctive
features (e.g. the ticks change position). The circle could, for
example, have white space overlaid on top of the map directly
around the perimeter, or some combination of features. In some
implementations, the contrast of the circle's color oscillates or
pulses to differentiate the circle from the map 502.
[0083] FIG. 6 illustrates a zoomed out view 602 of the map 502 from
FIG. 5A. In this example the user is able to view a large area of
the map 602 around a geographic area 604. The geographic area 604
can, for example, be the geographic area 516 rescaled to match the
scale of the zoomed out view 602. The user can, for example, find
restaurants, entertainment, or some other destination located in or
in the vicinity of the geographic area 604. This view of the map
602 can be presented when the mobile device 100 is moving, e.g. the
user of the mobile device 100 is a passenger in a moving vehicle.
In some implementations, the geographic area 604 is determined or
estimated when the current location object 514 is selected and does
not change with the movement of the mobile device 100. In some
other implementations, the geographic area 604 can move on the map
602 with the movement of the mobile device 100.
[0084] In some implementations, the geographic area 604 can be
based on data from the most recent query to the Wi-Fi base station
device 404. The geographic area 604 could also, for example, be
based on historical data of where the mobile device 100 recently
was located. For example, if the geographic area 604 was last
updated a short time before the newest activation, the previous
location can be used to identify the new location. Continuing the
example, if the previous current location showed that the user is
in California, but the new location 30 minutes later would show a
current location in Florida, an error can be displayed on the
mobile device 100.
[0085] FIG. 7 illustrates another view of the map 502 from FIG. 5A.
In this example, a geographic area 702 encloses the entire view of
the map 502. In some implementations, this view of the map 502 and
geographic area 702 can be shown, for example, when the mobile
device is not moving or not moving very quickly, e.g. the user with
mobile device 100 is walking. The geographic area 702 can be
displayed to contain the entire view of the map 502 when, at a
given zoom level of the map 502, the accuracy or precision of the
estimated position of the device 100 spans the entirety of the
displayed area of the map 502, i.e., the geographic area 702 covers
the entirety of the displayed portion of the map 502 at the given
zoom level. For example, if the map is zoomed to display a 50-meter
region around the estimated current position and the accuracy of
the estimated current position is 100 meters, then the geographic
area 702 can be displayed to enclose the displayed region of the
map.
[0086] FIG. 8 illustrates another view of map 602 with a larger
geographic area 802 than shown in FIG. 6. The geographic area 604
could, for example, be determined using Wi-Fi, such as through
access point 404 described in reference to FIG. 4, and the
geographic area 802 could be determined using cell phone towers,
such as the cell tower 406 described in reference to FIG. 4.
[0087] For example, the first geographic area from FIG. 4 can be
determined using access points 404 and the second geographic area
from FIG. 4 can be determined by cell tower 406. If the first
geographic area is determined to be wholly contained within the
second geographic area, the first geographic area can be used as
geographic area 604 and be presented on the mobile device 100. If
the first geographic area is determined to be at most partially
contained within the second geographic area, the second geographic
area could be used as geographic area 802 and be presented on the
mobile device 100.
[0088] Some examples of the first geographic area being at most
partially contained within the second geographic area include the
access point 404 being moved, or the access point 404 and the cell
tower 406 having a partial overlap of geographic areas (e.g. the
first geographic area is on the edge and partially outside of the
second geographic area). Other examples of the cause of geographic
areas 604 and 802 being different would be for other positioning
technologies, systems or services that have varying levels of
accuracy or precision, such as a GPS and a radio tower.
[0089] In some implementations, the scaling of the map 502 can be
different based on the geographic area selected, the
technology/system/service used to determine or estimate the
geographic area, or some other requirement. For example, if the
first geographic area determined using access points 404, as
described above, is selected to be presented as the geographic area
516 from FIG. 5A, the scaling can be larger than if the second
geographic area, determined using cell tower 406, was selected to
be presented as geographic area 802. The scaling of the map 602
can, for example, stay the same when the current location object
514 is selected. Continuing the example, if the default zoom level
determined to be used on the mobile device 100 is close to the
currently used zoom level, the zoom level can remain unchanged,
e.g. if the default level is a street zoom and the current zoom
level in on a city view, no change is made. If the current zoom
level is a state view, and the default level is a street view, the
map 602 can, for example, be changed to a street view.
[0090] The line width of the circle around the geographic area 802
can be changed based on the zoom level, or the line width can stay
the same. For example, the line width can get larger or smaller
when zooming in closer to the map. The change to the line width
and/or to lengths of the crosshairs, tick marks, or the like on the
circle can be based on an exponential algorithm, a linear
algorithm, or some other algorithm.
[0091] FIG. 9 illustrates an example of an interface for getting
directions on the mobile device 100. Selection of the directions
object 510 can, for example, bring up the directions interface. The
directions interface can allow the user of mobile device 100 to get
directions to and from any location. In some implementations, the
current location bookmark in the Bookmarks list can automatically
be inserted a start location 902. A start location 902 can also be
manually entered using a keyboard 908. An end location 904 can be
selected from the Bookmarks list with an end bookmarks list object
906 or can be manually entered using the keyboard 908. If the
current location is not used as the start location 902, it can be
used as the end location 904 (e.g., by selecting the current
location bookmark from the Bookmarks list).
[0092] The start location 902 and end location 904 can be swapped
with a switch object 910. For example, the switch object can be
used to make the end location 904 the new start location and the
start location 902 the new end location.
[0093] In some implementations, the user can, for example, remove
the current location bookmark from the start location 902 if a
different start location is desired with a remove start location
object 912. Alternatively, the user can manually enter a new
location as the start location 902 using the keyboard 908. In other
implementations, the user can select the current location bookmark
or another location bookmark from the Bookmarks list as the start
location. The remove start location object 912 can be replaced with
a start bookmarks list object (not shown), similar to the end
bookmarks list object 906, if the start location 902 is empty. Once
the start location 902 and end location 904 have been entered the
route object 914 can be selected and directions and optionally
other information for a route from the start location 902 to the
end location 904 can be displayed on the map, e.g. the map 502.
[0094] FIG. 10 illustrates an example of directions for a route
1002 displayed on a map 1001. The user, for example, can use the
directions interface of FIG. 9 to get directions from the center or
centroid of the geographic area 1003 in city A, to a destination
address (e.g. an address entered on the keyboard 908, a selected
bookmarked address). A start pin 1004 or other position marker can
be placed in the center or centroid of the geographic area 1003 to
represent an estimated current location of the mobile device 100;
the actual location of the mobile device 100 can be anywhere within
the geographic area 1003. An end pin 1006 can be placed at the
destination address.
[0095] The route of the directions 1002 between the start pin 1004
and the end pin 1006 can be highlight or depicted in some other way
to clearly show the route. In some implementations, the start pin
1004 has a street flag 1008 that displays the name of the street on
which the start pin 1004 is placed. The street flag 1008 can have
an object 1010 that is user selectable. Selection of the 1010 can
display an interface to allow the user to change the display of the
route 1002, to show turn by turn directions, or to select other
options, to name a few examples.
Example Map View with Dropped Pin
[0096] FIG. 11 illustrates a menu presented when the user selects
the map view object 512. When the map view object 512 is selected,
the map 502 can be displayed with a page curl or flipped-page
graphical effect 1102, to allow presentation of a menu 1103 on a
bottom portion of the touch-sensitive display 102. In some
implementations, this simulates a peeling of a layer with the map
502 to reveal the menu 1103 in a layer below. The flipped-page
graphical effect 1102 can, for example, also allow presentation of
other content such as the Bookmarks list, an address book, the
directions interface, a phonebook, a calendar, a calculator, or
other information. In some other implementations, the flipped-page
graphical effect 1102 can allow additional content to be displayed
toward the top, left, right, or middle of the touch-sensitive
display 102. The flipped-page graphical effect 1102 can, for
example, be a half page curl, or a whole page curl, to name a few
examples.
[0097] The menu 1103 presented with the selection of the map view
object 512 can contain user selectable objects including a drop pin
object 1104 and a show traffic object 1106. Map views for the map
502 can be changed with a map object 1108, a satellite object 1110,
and a hybrid object 1112. A list object 1114 can also be
displayed.
[0098] The show traffic object 1106 can display traffic related
information on the map 502. For example, roads with heavy traffic
can be highlighted in red on the map 502. The map object 1108 can
select a basic background (e.g. just streets, highways, etc are
displayed) for the map 502. The satellite object 1110 can select a
satellite view of the map 502 to be presented. The satellite view
can contain terrain, such as trees, hills, mountains, buildings,
etc. The hybrid object 1112 can have a combination of the basic
background and the satellite view presented on the map 502. For
example, terrain information is displayed with road information
overlaid on top.
[0099] The list object 1114 can, for example, be used to display a
list of one or more businesses, landmarks, addresses, or other
locations located in the area displayed on the map 502 or in the
vicinity of the current geographic area 516. In some
implementations, the list of businesses can be a scroll list, or
the list can be overlaid on top of map 502, e.g. the names and
locations of businesses can be displayed on the map 502. In some
implementations, the businesses, etc. in the list can be organized
by name or by street. In some implementations, the businesses,
landmarks, etc. are ordered by a location-based service from which
the information regarding the businesses, landmarks, etc. are
received. The location-based service can sort the list based on
various criteria (e.g., popularity, relevance to a search query
entered by the user of the device, etc.). Selection of the list
object 1114 can change the zoom level of map 502 so that too much
information is not displayed on the map 502 (e.g., if the user is
in a densely populated area the map can be zoomed in more than if
the user is in a rural area).
[0100] In some implementations, when the list object 1114 is
selected while a directions mode is active, e.g., the map is
showing direction for a route, the list that is displayed is a list
of the turn-by-turn directions for the route.
[0101] The drop pin object 1104 can, for example, allow the user to
drop a pin or other position marker on the map 502. When the user
drops a pin on a map (e.g., map 502), a special drop pin bookmark
is placed in the Bookmarks list. In some implementations, the map
can have one custom drop pin at any time. In some other
implementations, the map can have more than one dropped pin at any
time. The user can drop a pin on the map to mark and to keep track
of an arbitrary location. For example, the user received a current
location of a friend and wants to place a pin on the map where
their friend is located. The user can use the dropped pin to get
directions from the current location of the mobile device 100 to
the location of the dropped pin, e.g. their friend's location. In
the case of multiple dropped pins, the user can use the dropped
pins to get direction from one arbitrary location to another
arbitrary location.
[0102] FIG. 12 shows an example of a dropped pin 1202 on a map
1201. The dropped pin 1202 can, for example, have a flag 1204 that
provides information to the user of the mobile device 100. In some
implementations, the flag 1204 can display information on the
location where the dropped pin 1202 is placed, such as the city and
state or street name, for example. The flag 1204 can include an
arrow 1206 that is user selectable. In some implementations, when
the user selects the arrow 1206, other content (e.g., additional
information, a user-selectable menu of options, etc.) is displayed.
In some implementations, the dropped pin 1202 can be moved by
dragging it on the touch-sensitive display 102 to another location.
The dropped pin 1202 can be any arbitrary location, e.g., a
location that the user will need to get to in the future, the
location of a friend needing directions, or some other point of
reference.
[0103] FIG. 13 illustrates a dropped pin menu 1302 displayed on the
mobile device 100. The dropped pin menu 1302 can, for example, be
accessed by double tapping on the dropped pin 1202 in FIG. 12, or
by selecting the arrow 1206. The dropped pin menu 1302 can, for
example, be associated with a dropped pin, e.g. the dropped pin
1202. The dropped pin menu 1302 can contain an add to bookmarks
object 1304, a directions to here object 1306, a directions from
here object 1308, and a remove pin object 1310.
[0104] In some implementations, the add to bookmarks object 1304
can create a bookmark in the Bookmarks list for the current
location of the dropped pin 1202. When the bookmark is added to the
Bookmarks list, the special dropped pin bookmark can, for example,
be removed from the Bookmarks list, or the special dropped pin
bookmark can remain in the Bookmarks list.
[0105] The directions to here object 1306 can bring up the
directions interface for getting directions to the current location
of the dropped pin 1202. For example, if the dropped pin 1202 is
placed at a location that the user needs to get back to at a later
time, the user can double tap on the dropped pin 1202 when they
need directions. Continuing the example, if the user has the
estimated current location (e.g., the centroid of an estimated
geographic area) selected, selection of the directions to here
object 1306 can bring up the directions interface with the
estimated current location filled out as the start location, and
the end location as the dropped pin 1202. This can allow the user
of the mobile device 100 to easily get back to another location
they were at recently, or to get to a friend's location, to name a
few examples.
[0106] The directions from here object 1308 can bring up the
directions interface and allow the user to select or enter an end
destination. The directions from here object 1308 can, for example,
be chosen when the user is giving directions to a friend located at
the dropped pin 1202. The directions from here object 1308 can be
selected, continuing the example, and a bookmark (e.g. a home
bookmark with the home address of the user of the mobile device
100) or a manually entered location can be chosen as the end
location. The directions to here object 1308 and the directions
from here object 1308 can also be used to provide route information
to or from the dropped pin 1202 for a doctor's appointment or a
sporting event, to name a few additional examples.
[0107] The remove pin object 1310 can allow the user to remove the
currently selected dropped pin 1202 from the map 502. In some
implementations, the dropped pin 1202 is unique (e.g. there is only
one dropped pin on the map at any time), and in other
implementations, there can be more than one dropped pin on the map
at any time. The user can select the remove pin object 1310 if the
location of the dropped pin 1202 is no longer needed, e.g. the
location of the dropped pin has been added to the Bookmarks list
with the add to bookmarks object 1304.
[0108] FIG. 14 illustrates the menu presented by the map view
object 512 when there is a dropped pin on the map. In this
implementation, there can only be one dropped pin on the map at any
time for keeping track of a point of interest, or a destination
that will only be traveled to once, to list a few examples. The
user can, for example, save the current dropped pin 1202 to the
Bookmarks list to keep track of multiple points of interest.
Continuing the example, the drop pin object 1104 is replaced with a
replace pin object 1402. When the replace pin object 1402 is
selected, for example, the currently dropped pin (e.g. dropped pin
1202) is removed from the map and a new pin is dropped onto the
map. The old dropped pin (e.g. dropped pin 1202) may not currently
be visible on the map.
[0109] In some other implementations, there can be more than one
dropped pin. In this example, both the drop pin object 1104 and the
replace pin object 1402 can be visible in the menu presented by
selection of the map view object 512. The user could, for example,
name the dropped pins to keep track of them on the map 502.
[0110] FIG. 15 illustrates an example of the directions interface
after selecting the directions from here object 1308. In this case,
when the directions interface is displayed, the location of the
dropped pin 1202 is filled in for a start location 1502. The user
can select the end location 904 with the end bookmark list object
906 or the keyboard 908. The switch object 910 can be selected to
swap the start location 1502 and the end location 904. If, for
example, the user changes plans, the start location 1502 can be
changed from the dropped pin 1202 to another location by selecting
the remove start location object 912 and entering a new start
location with the keyboard 908 or by selecting a location from the
Bookmarks list. When both the start location 1502 and the end
location 904 have been entered, the route object 914 can be
selected to display directions on a map, e.g. the map 1201.
[0111] The directions interface can also be entered when the user
selects the directions to here object 1306. In this case, when the
directions interface is displayed, the location of the dropped pin
1202 is filled in as the end location 904.
[0112] FIG. 16 illustrates a dropped pin 1602 placed at the
centroid of the geographic area 802 from FIG. 8. When a geographic
area (e.g., geographic area 802) of the mobile device 100 is
displayed on a map (e.g., the map 602) and the dropped pin 1602 is
placed on the map, the dropped pin 1602 is placed at the centroid
of the geographic area. The user can drag the dropped pin 1602 to
another location within the geographic area 802 or on the map 602
if desired by selecting the dropped pin, e.g. by touching the
touch-sensitive display 102 at the location of the dropped pin
1602. In some implementations, the dropped pin 1602 can have a flag
1604 that can present information about the location of the dropped
pin 1602 or instructions to the user. The dropped pin 1602 can, for
example, be saved as a permanent bookmark in the Bookmarks list via
the dropped pin menu 1302 accessed with an arrow 1606 in the flag
1604.
[0113] The disclosed and other embodiments and the functional
operations described in this specification can be implemented in
digital electronic circuitry, or in computer software, firmware, or
hardware, including the structures disclosed in this specification
and their structural equivalents, or in combinations of one or more
of them. The disclosed and other embodiments can be implemented as
one or more computer program products, i.e., one or more modules of
computer program instructions encoded on a computer-readable medium
for execution by, or to control the operation of, data processing
apparatus. The computer-readable medium can be a machine-readable
storage device, a machine-readable storage substrate, a memory
device, a composition of matter effecting a machine-readable
propagated signal, or a combination of one or more them. The term
"data processing apparatus" encompasses all apparatus, devices, and
machines for processing data, including by way of example a
programmable processor, a computer, or multiple processors or
computers. The apparatus can include, in addition to hardware, code
that creates an execution environment for the computer program in
question, e.g., code that constitutes processor firmware, a
protocol stack, a database management system, an operating system,
or a combination of one or more of them. A propagated signal is an
artificially generated signal, e.g., a machine-generated
electrical, optical, or electromagnetic signal, that is generated
to encode information for transmission to suitable receiver
apparatus.
[0114] A computer program (also known as a program, software,
software application, script, or code) can be written in any form
of programming language, including compiled or interpreted
languages, and it can be deployed in any form, including as a
stand-alone program or as a module, component, subroutine, or other
unit suitable for use in a computing environment. A computer
program does not necessarily correspond to a file in a file system.
A program can be stored in a portion of a file that holds other
programs or data (e.g., one or more scripts stored in a markup
language document), in a single file dedicated to the program in
question, or in multiple coordinated files (e.g., files that store
one or more modules, sub-programs, or portions of code). A computer
program can be deployed to be executed on one computer or on
multiple computers that are located at one site or distributed
across multiple sites and interconnected by a communication
network.
[0115] The processes and logic flows described in this
specification can be performed by one or more programmable
processors executing one or more computer programs to perform
functions by operating on input data and generating output. The
processes and logic flows can also be performed by, and apparatus
can also be implemented as, special purpose logic circuitry, e.g.,
an FPGA (field programmable gate array) or an ASIC
(application-specific integrated circuit).
[0116] Processors suitable for the execution of a computer program
include, by way of example, both general and special purpose
microprocessors, and any one or more processors of any kind of
digital computer. Generally, a processor will receive instructions
and data from a read-only memory or a random access memory or both.
The essential elements of a computer are a processor for performing
instructions and one or more memory devices for storing
instructions and data. Generally, a computer will also include, or
be operatively coupled to receive data from or transfer data to, or
both, one or more mass storage devices for storing data, e.g.,
magnetic, magneto-optical disks, or optical disks. However, a
computer need not have such devices. Computer-readable media
suitable for storing computer program instructions and data include
all forms of non-volatile memory, media and memory devices,
including by way of example semiconductor memory devices, e.g.,
EPROM, EEPROM, and flash memory devices; magnetic disks, e.g.,
internal hard disks or removable disks; magneto-optical disks; and
CD-ROM and DVD-ROM disks. The processor and the memory can be
supplemented by, or incorporated in, special purpose logic
circuitry.
[0117] To provide for interaction with a user, the disclosed
embodiments can be implemented on a computer having a display
device, e.g., a CRT (cathode ray tube) or LCD (liquid crystal
display) monitor, for displaying information to the user and a
keyboard and a pointing device, e.g., a mouse or a trackball, by
which the user can provide input to the computer. Other kinds of
devices can be used to provide for interaction with a user as well;
for example, feedback provided to the user can be any form of
sensory feedback, e.g., visual feedback, auditory feedback, or
tactile feedback; and input from the user can be received in any
form, including acoustic, speech, or tactile input.
[0118] The disclosed embodiments can be implemented in a computing
system that includes a back-end component, e.g., as a data server,
or that includes a middleware component, e.g., an application
server, or that includes a front-end component, e.g., a client
computer having a graphical user interface or a Web browser through
which a user can interact with an implementation of what is
disclosed here, or any combination of one or more such back-end,
middleware, or front-end components. The components of the system
can be interconnected by any form or medium of digital data
communication, e.g., a communication network. Examples of
communication networks include a local area network ("LAN") and a
wide area network ("WAN"), e.g., the Internet.
[0119] The computing system can include clients and servers. A
client and server are generally remote from each other and
typically interact through a communication network. The
relationship of client and server arises by virtue of computer
programs running on the respective computers and having a
client-server relationship to each other.
[0120] While this specification contains many specifics, these
should not be construed as limitations on the scope of what being
claims or of what may be claimed, but rather as descriptions of
features specific to particular embodiments. Certain features that
are described in this specification in the context of separate
embodiments can also be implemented in combination in a single
embodiment. Conversely, various features that are described in the
context of a single embodiment can also be implemented in multiple
embodiments separately or in any suitable subcombination. Moreover,
although features may be described above as acting in certain
combinations and even initially claimed as such, one or more
features from a claimed combination can in some cases be excised
from the combination, and the claimed combination may be directed
to a subcombination or variation of a subcombination.
[0121] Similarly, while operations are depicted in the drawings in
a particular order, this should not be understand as requiring that
such operations be performed in the particular order shown or in
sequential order, or that all illustrated operations be performed,
to achieve desirable results. In certain circumstances,
multitasking and parallel processing may be advantageous. Moreover,
the separation of various system components in the embodiments
described above should not be understood as requiring such
separation in all embodiments, and it should be understood that the
described program components and systems can generally be
integrated together in a single software product or packaged into
multiple software products.
[0122] Particular embodiments of the subject matter described in
this specification have been described. Other embodiments are
within the scope of the following claims. For example, the actions
recited in the claims can be performed in a different order and
still achieve desirable results. As one example, the processes
depicted in the accompanying figures do not necessarily require the
particular order shown, or sequential order, to achieve desirable
results. In certain implementations, multitasking and parallel
processing may be advantageous.
* * * * *