U.S. patent application number 13/025539 was filed with the patent office on 2011-06-09 for peripheral with a display.
This patent application is currently assigned to SONY ERICSSON MOBILE COMMUNICATIONS AB. Invention is credited to Eral Denis FOXENLAND.
Application Number | 20110134016 13/025539 |
Document ID | / |
Family ID | 38961077 |
Filed Date | 2011-06-09 |
United States Patent
Application |
20110134016 |
Kind Code |
A1 |
FOXENLAND; Eral Denis |
June 9, 2011 |
PERIPHERAL WITH A DISPLAY
Abstract
A headset may include a communication interface to receive one
or more images from a device. The headset may further include a
pair of earpieces, each earpiece including a speaker and a display
to show at least one of the one or more images from the device.
Inventors: |
FOXENLAND; Eral Denis;
(Malmo, SE) |
Assignee: |
SONY ERICSSON MOBILE COMMUNICATIONS
AB
Lund
SE
|
Family ID: |
38961077 |
Appl. No.: |
13/025539 |
Filed: |
February 11, 2011 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
11682371 |
Mar 6, 2007 |
7911410 |
|
|
13025539 |
|
|
|
|
Current U.S.
Class: |
345/8 |
Current CPC
Class: |
H04R 5/033 20130101;
H04R 25/554 20130101; H04R 25/552 20130101; G02B 27/017 20130101;
H04R 2420/07 20130101; G02B 2027/014 20130101 |
Class at
Publication: |
345/8 |
International
Class: |
G09G 5/00 20060101
G09G005/00 |
Claims
1-20. (canceled)
21. A system comprising: a peripheral device comprising one or more
displays to: selectively display, based on audible information
received from a user of a communication device, a first image, of a
plurality of images, on the one or more displays when a
communication application is being executed on the communication
device, where the communication device is communicatively linked to
the peripheral device, and selectively display, based on an audio
signal transmitted from the communication device and output to a
user through the peripheral device, a second image, of the
plurality of images, on the one or more displays when the
communication application is executing on the communication device,
where the second image and the first image comprise different
images.
22. The system of claim 21, where the peripheral device further
comprises: a microphone to receive the audible information from the
user.
23. The system of claim 21, where the one or more displays
comprises a plurality of displays, and the peripheral device
further comprises: a communication interface to communicatively
link the peripheral device and the communication device, and a
processor to cause an alternate display of the first image, of the
plurality of images, via a first display of the plurality of
displays, and the second image, via a second display of the
plurality of displays, based on display information that is
received from the communication device at the communication
interface.
24. The system of claim 21, where the second image includes a video
associated with the audio signal.
25. The system of claim 21, where the peripheral device further
comprises: a memory to store each of the plurality of images,
including the first image and the second image, that are received
from the communication device.
26. The system of claim 21, where the peripheral device is further
to: selectively display a third image, of the plurality of images,
via the one or more displays, when the communication application is
not currently being executed at the communication device.
27. The system of claim 21, where the first image comprises an
image corresponding to a particular person associated with a
telephone number of an incoming call received at the communication
device, and where the peripheral device is further to: display the
image corresponding to the particular person via the one or more
displays before the incoming call is accepted by the user of the
communication device.
28. A device comprising: a plurality of displays for displaying one
or more images of a plurality of images; and a processor to:
generate display information, identifying a first display, of the
plurality of displays, to display a first image of the plurality of
images, and identifying a second display, of the plurality of
displays, to display a second image, based on received audible
information, that is transmitted to a communication device
associated with the device, and an audio signal, that is
transmitted from the communication device, respectively, where the
first image comprises a different image than the second image.
29. The device of claim 28, further comprising: a microphone to
receive the audible information from the user.
30. The device of claim 28, further comprising: a communication
interface to communicatively link the device and the communication
device, and where the processor is further to: cause an alternate
display of the first image via the first display and the second
image via the second display, based on display information that is
received from the communication device at the communication
interface.
31. The device of claim 28, further comprising: a speaker to
transmit the audio signal, and where the second image includes a
video associated with the audio signal.
32. The device of claim 28, where the first image and the second
image are received from the communication device, and the device
further comprises: a memory to store each of the plurality of
images, including the first image and the second image received
from the communication device.
33. The device of claim 28, where the processor is further to:
selectively display, when a communication application is not
currently being executed on the communication device, a third
image, of the plurality of images, via one of the plurality of
displays.
34. The device of claim 28, where the first image comprises an
image corresponding to a particular person associated with a
telephone number of an incoming call received at the communication
device, and where the processor is further to: display the image
corresponding to the particular person via one or more of the
plurality of displays before the incoming call is accepted by the
user of the communication device.
35. A method comprising: retrieving, by a main device and based on
a communication activity of a peripheral device, a plurality of
images and corresponding display information, from a memory, where
the display information relates each of a plurality of images to a
particular display, of a plurality of displays, of the peripheral
device; transmitting, by the main device and via one or more
communication links established between the main device and the
peripheral device, the plurality of images and corresponding
display information to the peripheral device; detecting, by the
main device, audible information received via a microphone of the
peripheral device; transmitting, by the main device and via the one
or more communication links, an audio signal to the peripheral
device; displaying, based on the display information, one of the
plurality of images on at least one of the plurality of displays of
the peripheral device in response to detecting the audible
information; and displaying, based on the display information,
another one of the plurality of images on the at least one of the
plurality of displays of the peripheral device in response to
transmitting the audio signal.
36. The method of claim 35, further comprising: displaying, based
on a display theme and the communication activity, one or more
images, of the plurality of images, on one or more of the plurality
of displays of the peripheral device.
37. The method of claim 35, further comprising: creating an
association between the communication activity, the one or more
images of the plurality of images, and the one or more of the
plurality of displays of the peripheral device; and storing the
association in a memory of the main device.
38. The method of claim 37, further comprising: identifying the
execution of a communication application corresponding to the
communication activity; and retrieving the plurality of images and
the corresponding display information in response to the
identification of the execution of the communication
application.
39. The method of claim 35, where the one or more communication
links comprise a plurality of information links and each of the
plurality of information links is associated with one of the
plurality of displays of the peripheral device, the method further
comprising: transmitting the one of the plurality of images via one
of the plurality of communication links; and transmitting the other
one of the plurality of images via another one of the plurality of
communication links.
40. The method of claim 35, where the one or more communication
links comprise a plurality of information links and each of the
plurality of information links is associated with one of the
plurality of displays of the peripheral device, the method further
comprising: transmitting the one of the plurality of images via one
of the plurality of communication links; and transmitting the
display information via another one of the plurality of information
links.
Description
TECHNICAL FIELD OF THE INVENTION
[0001] Implementations described herein are related to displays and
in particular, pertain to methods and devices for displaying images
on peripherals that are located in a user's personal space.
DESCRIPTION OF RELATED ART
[0002] Peripherals, such as headphones, earphones, headsets, and
earpieces, allow a user to conveniently interact with communication
and multimedia devices. For example, when a headset is attached to
a cellular phone, the headset may provide a pair of speakers and a
microphone for a user to exchange information with a caller without
using the hands.
SUMMARY
[0003] According to one aspect, a headset may comprise a
communication interface to receive one or more images from a
device, and a pair of earpieces, each earpiece including a speaker
and a display to show at least one of the one or more images from
the device.
[0004] Additionally, the communicating interface may include a
wireless communication interface.
[0005] Additionally, the headset may further comprise a processor
to choose a display among the two displays included in the
earpieces based on display information that is received at the
communication interface and is associated with a first one of the
one or more images, and to provide the first one of the one or more
images to the chosen display.
[0006] Additionally, the one or more images may include a
video.
[0007] Additionally, the headset may further comprise a memory to
store images that are received from the device.
[0008] Additionally, the one or more images may include an image
associated with a telephone number of an incoming call at the
device.
[0009] Additionally, the one or more images may include an image
associated with an application that is currently being executed at
the device.
[0010] According to another aspect, a device may comprise a
communication interface that establishes a communication link with
a headset that includes a pair of earpieces, each earpiece
including a display. Additionally, the device may further comprise
a processor to retrieve an image, obtain display information
associated with the image, identify one or more displays among the
displays included in the earpieces based on the display
information, and send the retrieved image to the identified one or
more displays on the headset through the communication link.
[0011] Additionally, the device may further comprise a database
that includes pieces of information, each piece of information
relating an image to one or more displays.
[0012] Additionally, the device may include a mobile device, an
audio player, or a video game console.
[0013] According to yet another aspect, an earpiece may comprise a
communication interface to receive an image from a device, and a
display to display the image, the display being integrated into a
surface of the earpiece and being not viewable by a person wearing
the earpiece.
[0014] Additionally, the communication interface may include a
wireless communication interface.
[0015] Additionally, the image may include an image that is
associated with a telephone number of an incoming call at the
device.
[0016] Additionally, the image may include an image associated with
a state of an application that is currently being executed at the
device.
[0017] Additionally, the earpiece may further comprise at least one
of a microphone, a speaker, or a display to display images to the
person wearing the earpiece.
[0018] Additionally, a method may comprise retrieving images and
pieces of display information from a database based on an identity
of an active application, each piece of display information
relating one or more destination displays to an image.
Additionally, the method may further comprise establishing, in a
personal space, wireless communication links from a device to one
or more personal peripherals that are to receive the images, the
one or more personal peripherals including the one or more
destination displays. Additionally, the method may further comprise
sending each of the images to the one or more destination displays
through the wireless communication links.
[0019] Additionally, the method may further comprise causing
personal peripherals to show one or more of the images at the one
or more destination displays in accordance with a display theme and
the identity of the active application.
[0020] Additionally, the method may further comprise creating an
association between an application, images, and displays, and
storing the association in the database.
[0021] According to yet another aspect, a system may include a
headset and a device. Additionally, the headset may comprise a
communication interface to receive one or more images, and a pair
of earpieces, each earpiece including a speaker and a display to
show at least one of the one or more images. Additionally, the
device may comprise a communication interface that establishes a
communication link with the headset. Additionally, the device may
further comprise a processor to retrieve an image, obtain display
information associated with the image, identify one or more
displays among the displays included in the earpieces based on the
display information, and send the retrieved image to the identified
one or more displays on the headset through the communication
link.
[0022] According to a further aspect, a headset may comprise means
for receiving an image from a device through a wireless
communication link within a personal space of an individual, and
means for showing the received image on a display that is
integrated into an earpiece of the headset, the image being outside
a visual range of the individual when the individual is wearing the
headset.
BRIEF DESCRIPTION OF THE DRAWINGS
[0023] The accompanying drawings, which are incorporated in and
constitute a part of this specification, illustrate one or more
embodiments described herein and, together with the description,
explain the embodiments. In the drawings,
[0024] FIG. 1 depicts an exemplary network in which systems and
methods described herein may be implemented;
[0025] FIG. 2 is an exemplary diagram of the main device of FIG.
1;
[0026] FIG. 3 is an exemplary diagram of the personal peripheral of
FIG. 1;
[0027] FIG. 4 is an exemplary block diagram of the main device of
FIG. 1;
[0028] FIG. 5 is an exemplary block diagram of the personal
peripheral of FIG. 1;
[0029] FIG. 6 is an exemplary functional block diagram of exemplary
components that may be included in the main device of FIG. 4;
[0030] FIG. 7 is an exemplary functional block diagram of exemplary
components that may be included in the personal peripheral of FIG.
5;
[0031] FIG. 8 shows an exemplary process for sending one or more
images to the personal peripheral of FIG. 1;
[0032] FIG. 9 is an exemplary process for displaying one or more
images at the personal peripheral of FIG. 1;
[0033] FIG. 10 illustrates a local network in which different
implementations of personal peripherals communicate with the main
device of FIG. 1; and
[0034] FIG. 11A shows a cellular phone and a person using a headset
with displays;
[0035] FIG. 11B shows an image that may be displayed when Erika is
speaking; and
[0036] FIG. 11C shows an image that may be displayed when Erika is
listening.
DETAILED DESCRIPTION OF EMBODIMENTS
[0037] The following detailed description refers to the
accompanying drawings. The same reference numbers in different
drawings may identify the same or similar elements.
[0038] The term "personal space," as used herein, may refer to the
region surrounding a person (e.g., a bubble of 1-2 meter radius
from the center of the person's body) that the person may consider
the person's own domain or territory. A communication link may be
in a personal space, if two devices that communicate through the
communication link are in the same personal space.
[0039] The term "personal peripheral," as used herein, may describe
a device that is separate from and used in conjunction with a main
device that stays within the personal space of a user for periods
of time (e.g., headset, watch, handbag, etc.).
[0040] The term "image," as used herein, may refer to a digital
representation of visual information (e.g., a picture, a video, a
photograph, animations, etc).
[0041] In the following implementations, a personal peripheral may
include a display. By using a main device that is linked to the
personal peripheral through a cable or a wireless communication
channel, different images may be shown on the display of the
personal peripheral. For example, a user may speak with a caller
through a headset that is wirelessly linked to the user's cellular
phone. When the user speaks, a display that is attached to the
headset may show the image that indicates that the user is
currently engaged in a conversation.
Exemplary Main Device and Peripheral
[0042] FIG. 1 depicts an exemplary network in which systems and
methods described herein may be implemented. As shown, network 100
may include a main device 102, a personal peripheral 104, and a
wireless network 106. In practice, network 100 may include fewer or
additional main devices, personal peripherals, and/or wireless
networks.
[0043] Main device 102 may include any of the following devices
that have the ability to communicate with personal peripherals: an
audio player; a video player; a telephone, such as a radio
telephone; a video game console; a personal communications system
(PCS) terminal that may combine cellular radiotelephone with data
processing, facsimile, and/or data communications capabilities; a
mobile telephone; an electronic notepad; a laptop; a personal
computer (PC); a personal digital assistant (PDA) that can include
a radiotelephone, pager, Internet/intranet access, web browser,
organizer, calendar, and/or GPS receiver; or any device with
sufficient computing power and memory to support functions
described herein.
[0044] Personal peripheral 104 may include devices that operate as
peripherals for main device 102 within a user's personal space, and
may communicate with main device 102 either wirelessly or through a
cable. In the description to follow, personal peripheral 104 will
be described in terms of a headset. A "headset," as the term is
used herein, is to be broadly interpreted to include one or more
of: a headphone; an earpiece; an earphone; a microphone; etc.
[0045] Wireless network 106 may include a local network that
interconnects wireless devices (e.g., wireless personal
peripherals) within their immediate proximity. These wireless
devices may be located nearby and may frequently or occasionally
come within range and go out of range of each other. In FIG. 1,
personal peripheral 104 may communicate with main device 102
through wireless network 106.
[0046] FIG. 2 is an exemplary diagram of main device 102 of FIG. 1.
As shown, main device 102 may include a speaker 202, a display 204,
control buttons 206, a keypad 208, a microphone 210, a camera 212,
and housing 214. Speaker 202 may provide audible information to a
user of main device 102. Display 204 may provide visual information
to the user. For example, display 204 may provide information
regarding incoming or outgoing calls, media, games, phone books,
the current time, etc. In one implementation, display 204 may
provide the user with a graphical user interface for inputting
various parameters associated with communication and image
processing. Control buttons 206 may permit the user to interact
with main device 102 to cause main device 102 to perform one or
more operations. Keypad 208 may include a standard telephone
keypad. Microphone 210 may receive audible information from the
user. Camera 212 may enable a user to capture and store images
(e.g., pictures, video clips). Housing 214 may provide a casing for
components of main device 102 and may protect the components from
outside elements.
[0047] FIG. 3 is an exemplary diagram of personal peripheral 104 of
FIG. 1. As shown, personal peripheral 104 may include an
interconnect 302, a microphone 304, and earpieces 306. Interconnect
302 may deliver signals from main device 102 to personal peripheral
104 and from personal peripheral 104 to main device 102. Microphone
304 may receive audible information from a user and deliver the
information to personal peripheral 104.
[0048] As shown in FIG. 3, earpieces 306 may include a display 308
and a speaker 310. Display 308 may include a display that is
integrated into earpieces 306 or that can be attached to or adapted
to fit the surface of earpieces 306. Display 308 may be thin, bent,
or bendable, and may provide a sufficient color and/or screen
resolution for displaying images such as icons, logos, pictures
(e.g., pictures of one's loved ones and pets, a real size picture
of gems, jewelry, etc.), animations, etc. Speaker 310 may provide
audible information to a user of personal peripheral 104.
[0049] FIG. 4 is an exemplary block diagram of main device 102 of
FIG. 1. As illustrated, main device 102 may include memory 402, a
processing unit 404, a network interface 406, I/O devices 408, and
bus 410. In other implementations, main device 102 may include
fewer, additional, or different components than those illustrated
in FIG. 4.
[0050] Memory 402 may include static memory, such as read only
memory (ROM), and/or dynamic memory, such as random access memory
(RAM), or onboard cache, for storing data and machine-readable
instructions. Processing unit 404 may include one or more
processors, microprocessors, and/or processing logic capable of
controlling main device 102. Network interface 406 may include any
transceiver-like mechanism that enables main device 102 to
communicate with other devices and/or systems, such as personal
peripheral 104. I/O devices 408 may include a display (e.g.,
display 204), keyboard, keypad (e.g., keypad 208), button, mouse,
speaker, microphone (e.g., microphone 210), Digital Video Disk
(DVD) writer, DVD reader, USB lines, and/or another type of device
for converting physical events or phenomena to and/or from digital
signals that pertain to main device 102. Bus 410 may provide an
interface through which components of main device 102 can
communicate with one another.
[0051] FIG. 5 is an exemplary block diagram of personal peripheral
104 of FIG. 1. As illustrated, personal peripheral 104 may include
memory 502, a processing unit 504, a network interface 506, a
display 508, a speaker 510, a microphone 512, and a bus 514. In
other implementations, personal peripheral 104 may include fewer,
additional, or different components than those illustrated in FIG.
5.
[0052] Memory 502 may include static memory, such as ROM, and/or
dynamic memory, such as RAM, or onboard cache, for storing data and
machine-readable instructions. Processing unit 504 may include one
or more processors, microprocessors, and/or processing logic
capable of controlling personal peripheral 104.
[0053] Network interface 506 may include any transceiver-like
mechanism that enables personal peripheral 104 to communicate with
other devices and/or systems. For example, network interface 506
may include mechanisms for communicating with main device 102
through a cable (e.g., interconnect). Additionally or
alternatively, network interface 506 may include a wireless
interface (e.g., a Bluetooth radio interface or an Ultra Wide Band
(UWB) radio interface) for permitting wireless communication with
main device 102.
[0054] Display 508 may include a device that can display signals
that are generated or received by personal peripheral 104 as images
on a screen. In one example, display 508 may provide information
regarding incoming or outgoing calls on main device 102. Examples
of display 408 include a liquid crystal display (LCD), organic
light-emitting diode (OLED) display, bistable display, a touch
screen, a bendable or flexible display (e.g., electronic paper),
and any display that is capable of being integrated into, adapted
to fit a surface of, or attached to personal peripheral 104.
[0055] Speaker 510 may provide audible information to the user of
personal peripheral 104. An example of speaker 510 may include
speaker 310 (FIG. 3). Microphone 512 may receive audible
information from the user and relay the audible information to
personal peripheral 104 to be processed for sending to main device
102. An example of microphone 512 may include microphone 304.
[0056] Bus 514 may provide an interface through which components of
personal peripheral 104 can communicate with one another.
[0057] FIG. 6 is an exemplary functional block diagram of exemplary
components that may be included in main device 102 of FIG. 1. As
illustrated, main device 102 may include applications 602, user
interface 604, display logic 606, database 608, and support logic
610. In practice, main device 102 may include additional, fewer, or
different components than those illustrated in FIG. 6.
[0058] Applications 602 may include hardware and/or software for
supporting various functionalities of main device 102, such as text
messaging, email, Multimedia Messaging, wireless (e.g., Bluetooth)
communications, infrared communications, web access, file uploading
and downloading, image transfer, etc. User interface 604 may
include hardware and/or software for allowing a user to interact
with applications 602, display logic 606, database 608, and/or
support logic 610.
[0059] Display logic 606 may include hardware and/or software for
causing one or more images to be displayed on personal peripheral
104. In one implementation, display logic 606 may allow a user to
select, through user interface 504, an image to be shown on display
308 on each of earpieces 302. The user may select the same image
for both displays on earpieces 302, or a different image for each
display 308. Display logic 606 may send the selected image or
images to personal peripheral 104.
[0060] In another implementation, display logic 606 may monitor a
user's activity at main device 102 and send one or more images
relating to the activity to personal peripheral 104. If a user
begins an activity, display logic 606 may send to personal
peripheral 104 one or more images that have been selected by the
user to represent the activity. Display logic 606 may accomplish
the preceding by identifying an application that is active (e.g., a
phone call handling application), fetching one or more images that
are associated with the application from database 608, retrieving
display information that specifies one or more displays for each of
the images, and causing each image to be displayed on the
corresponding displays. For example, when main device 102 receives
a call, display logic 606 may identify the phone call handling
application as the active application and may retrieve two images
based on the identification. During the call, display logic 606 may
cause one image to be shown on display 308 (FIG. 3) when a user is
talking, and may cause the other image to be shown when the other
party is talking.
[0061] If no application is in active (i.e., main device 102 is
idle), display logic 606 may still retrieve one or more images and
display information associated with the idle condition of main
device 102, and cause the images to be displayed on personal
peripheral 104. In one implementation, the display information may
include logical addresses of the displays and may designate a
destination display for each image. Because there may be more than
one display on a personal peripheral and more than one personal
peripheral that is in communication with main device 102, the
logical addresses may be used to send an image to one or more of
these displays.
[0062] In yet another implementation, display logic 606 may send
images and/or display information for each image to be loaded, and
allow a user to later select one or more of the loaded images to be
displayed at personal peripheral 104. For example, display logic
606 may send images (e.g., a color, a pattern, etc.) that match a
user's clothes to personal peripheral 104. Once the images have
been loaded, by using display logic 606 of main device 102, the
user may cause one or more of the images to be shown on display 308
(FIG. 3) of personal peripheral 104.
[0063] In yet another implementation, display logic 606 may provide
the ability to set themes. For example, a user may employ display
logic 606 to store a theme that includes one or more images (e.g.,
images with color patterns). When the user activates the theme,
display logic 606 may cause one or more displays to show the images
for the theme.
[0064] Returning to FIG. 6, database 608 may act as an information
repository for applications 602, user interface 604, display logic
606, and/or support logic 610. For example, applications 602 may
store and/or retrieve images and sounds to and from database 608.
In addition, an association between one or more images, displays,
and/or an application may be created and stored in database
608.
[0065] Database 608 may store images, information about callers
(e.g., phone numbers, addresses, etc.), information about displays
(e.g., an address of display 308), and/or information that
specifies one or more images and displays for an application on
main device 102 (i.e., an association between an image, a display,
and an application).
[0066] For example, database 608 may include a video, a logical
address of display 308, and an association that specifies the video
and display 308 for an audio application (e.g., Motion Pictures
Experts Group layer 3 (MP3) player application). If main device 102
runs the audio application, display logic 606 may identify the
audio application, and retrieve the association from database 608
based on the identity of the application. Through the use of the
association, display logic 606 may dispatch the correct images to
the correct displays, to indicate that the user may be listening to
music, sound, speech, etc.
[0067] In another example, database 608 may include a picture of a
person, a pet or anything that may be associated with a particular
caller. If a call is from the particular caller, display logic 606
may retrieve the picture from database 608 and show the picture at
display 308 (FIG. 3), based on an identifier (e.g., phone number)
associated with the incoming call. Such a picture may allow the
user to grab and view one of earpieces 306 to determine who is
calling and decide whether to take the call.
[0068] Additional examples of images that can be stored in database
608 may include: a logo, which may be sold with main device 102 as
the default image for branding purposes; an image of a user's skin
or of a user's ear that changes and morphs to produce a
chameleon-like effect; videos of animations that pulse with music;
images that make fashion statements; an image of an equalizer or a
static music icon to be displayed when the user is listening to
music or a radio station; etc.
[0069] Returning to FIG. 6, support logic 610 may include hardware
and/or software for performing various support functions for
applications 602, user interface 604, display logic 606, and
database 608. For example, support logic 610 may provide a TCI/IP
stack to support communication applications. Support logic 610 may
also provide interfaces between the components illustrated in FIG.
6 (e.g., applications 602) and the components in FIG. 4 (e.g.,
network interface 406 and I/O devices 408).
[0070] As described above, main device 102 may cause images to be
displayed on personal peripheral 104. FIG. 7 illustrates an
exemplary functional block diagram of exemplary components that may
be included in personal peripheral 104 of FIG. 1.
[0071] As illustrated, personal peripheral 104 may include image
logic 702, audio logic 704, microphone logic 706, image database
708, and/or support logic 710. In different implementations,
personal peripheral 104 may include additional, fewer or different
components than those illustrated in FIG. 7.
[0072] Image logic 702 may include hardware and/or software for
generating image signals from data received from main device 102
and for causing images to be shown on display 308 (FIG. 3). Audio
logic 704 may include hardware and/or software for generating audio
signals from data received from main device 102 and for causing
sounds to be generated by speaker 310. Microphone logic 706 may
include hardware and/or software for processing signals from
microphone 304 and causing sound signals to be sent to main device
102. Image database 708 may include images and/or display
information that have been received from main device 102. When
requested by main device 102, image logic 702 may select and
retrieve one or more images from image database 610. The images may
be displayed at display 308.
[0073] Support logic 710 may provide a support for receiving,
processing and/or transmitting signals to and/or from main device
102. For example, support logic 710 may provide a Transmission
Control Protocol (TCP)/Internet Protocol (IP) stack, buffers for
received data, etc. In addition, support logic 710 may provide a
support for interfacing image logic 702, audio logic 704, and/or
microphone logic 706 to hardware components, such as microphone
304, display 308, and/or speaker 310.
Exemplary Process for Displaying Images at a Personal
Peripheral
[0074] FIG. 8 is an exemplary process 800 for sending one or more
images to personal peripheral 104 of FIG. 1.
[0075] At block 802, one or more images that are to be displayed at
personal peripheral 104 may be selected. Depending on
implementation, the selection may be made by a user and/or by
display logic 606. In one implementation, a user may select one or
more images by using display logic 606. In another implementation,
display logic 606 may automatically make the selection by
monitoring a user's activity and selecting one or more images that
are associated with the activity. If main device 102 is in an idle
state (i.e., a user is not actively using main device 102), default
images may be selected. In yet another implementation, a user may
make the selection by activating a theme through display logic 606.
When the user activates the theme, the selection may be made by
display logic 606.
[0076] At block 804, image information may be fetched from database
608 based on the selection. In one implementation, the image
information may include images and display information. In another
implementation, the image information may specify which images
among those stored at personal peripheral 104 are to be
displayed.
[0077] At block 806, the fetched image information may be sent to
personal peripheral 104. If a communication channel (e.g., wireless
channel) is not established with personal peripheral 104, the
communication channel may be established before sending the image
information.
[0078] FIG. 9 is an exemplary process for displaying one or more
images at personal peripheral 104 of FIG. 1. At block 902, the
image information may be received from main device 102.
[0079] At block 904, one or more displays may be identified. In one
implementation, the identification may be made based on display
information that is included in the image information. In another
implementation, the display identification may be made based on
display information that is retrieved from image database 708. In
yet another implementation, image information for each display may
arrive on a separate communication channel or a sub-channel. In
such an implementation, it may not be necessary to identify the
display based on the display information.
[0080] At block 906, one or more images may be obtained. In one
implementation, the images may be obtained from the received image
information, if the received image information includes the images
themselves. In another implementation, the images may be obtained
from image database 708 by performing a database lookup using the
received image information. At block 908, the obtained images may
be shown at the identified displays.
Alternative Implementations
[0081] Many changes to main device 102 and personal peripheral 104,
along with associated processes, may be implemented. FIG. 10
illustrates a local network in which different implementations of
personal peripherals communicate with main device 102 of FIG. 1. As
shown, network 1000 may include personal peripherals 1002-1008 and
104, main device 102, and wireless network 106. Personal
peripherals 1002-1008 may receive and display images in accordance
with image information that is sent from main device 102. Main
device 102, personal peripheral 104, and wireless network 106 have
been described with respect to FIG. 1, and may operate in the
manner already described.
[0082] In FIG. 10, each of personal peripherals 1002-1008 may take
the form of a button (e.g., a button on a coat, or a jacket). In
other implementations, personal peripherals 1002-1008 may take
other forms, such as a headband, a wristband, earmuffs, a purse, a
belt, and/or a personal object (i.e., an object in a personal
space) with a body and a surface on which it is possible to
integrate, attach, or fit a display and circuitry to interact with
main device 102. By the use of main device 102, different images
may be shown on the displays (not shown) of personal peripherals
1002-1008 in ways described above.
Example
[0083] The following example illustrates processes involved in
sending an image to and displaying the image at a personal
peripheral, with reference to FIGS. 11A-11C. The example is
consistent with the exemplary processes described above with
references to FIGS. 8 and 9.
[0084] Assume that Erika has a personal peripheral in the form of a
headset 1102 and that Erika uses headset 1102 in conjunction with a
main device, in the form of a cellular phone 1104, as illustrated
in FIG. 11A.
[0085] When cellular phone 1104 receives a call from Erika's
friend, cellular phone 1104 may identify a phone call handling
application in cellular phone 1104 as the application that is
active, and may retrieve images that are associated with the phone
call handling application.
[0086] During the call, when Erika speaks, headset 1102 or cellular
phone 1104 may detect an audio signal from the microphone on
headset 1102, and one of the images 1108 may be displayed on
display 1106. FIG. 11B shows image 1108 that may be displayed on
display 1106 when Erika speaks. If headset 1102 or cellular phone
1104 detects a transmission of an audio signal from cellular phone
1104 to headset 1102, headset 1102 or cellular phone 1104 may cause
the other image 1110 to be displayed on display 1106. FIG. 11C
shows image 1110 that may be displayed on display 1106 when Erika
listens to the audio signal.
[0087] By displaying images 1108 and 1110 while Erika interacts
with her friend, headset 1102 may let others in her office know
whether Erika is actively engaged in a phone conversation or not,
and may provide them with opportunities to speak to Erika without
interfering with her call.
CONCLUSION
[0088] The foregoing description of embodiments provides
illustration, but is not intended to be exhaustive or to limit the
embodiments to the precise form disclosed. Modifications and
variations are possible in light of the above teachings or may be
acquired from practice of the teachings.
[0089] For example, while series of blocks have been described with
regard to processes illustrated in FIGS. 8 and 9, the order of the
blocks may be modified in other implementations. For example, block
904 may be performed before 906 in some implementations. In
addition, non-dependent blocks may represent acts that can be
performed in parallel with a number of other blocks.
[0090] It will be apparent that aspects described herein may be
implemented in many different forms of software, firmware, and
hardware in the implementations illustrated in the figures. The
actual software code or specialized control hardware used to
implement aspects should not be construed as limiting. Thus, the
operation and behavior of the aspects were described without
reference to the specific software code--it being understood that
software and control hardware can be designed to implement the
aspects based on the description herein.
[0091] No element, act, or instruction used in the present
application should be construed as critical or essential to the
implementations described herein unless explicitly described as
such. Also, as used herein, the article "a" is intended to include
one or more items. Where only one item is intended, the term "one"
or similar language is used. Further, the phrase "based on" is
intended to mean "based, at least in part, on" unless explicitly
stated otherwise.
[0092] It should be emphasized that the term "comprises/comprising"
when used in this specification is taken to specify the presence of
stated features, integers, steps or components but does not
preclude the presence or addition of one or more other features,
integers, steps, components, or groups thereof.
[0093] Further, certain portions of the invention have been
described as "logic" that performs one or more functions. This
logic may include hardware, such as a processor, a microprocessor,
an application specific integrated circuit, or a field programmable
gate array, software, or a combination of hardware and
software.
* * * * *