U.S. patent application number 11/146907 was filed with the patent office on 2006-10-19 for figurine using wireless communication to harness external computing power.
This patent application is currently assigned to Speechgear, Inc.. Invention is credited to Robert D. Palmquist.
Application Number | 20060234602 11/146907 |
Document ID | / |
Family ID | 35510282 |
Filed Date | 2006-10-19 |
United States Patent
Application |
20060234602 |
Kind Code |
A1 |
Palmquist; Robert D. |
October 19, 2006 |
Figurine using wireless communication to harness external computing
power
Abstract
The invention is directed toward a figurine that utilizes
wireless communication to harness computing power of an external
computer. The figurine may capture visual or audible input and
wirelessly transfer the input to the external computer, either
directly or via a network. The external computer processes the
input, generates output, and transfers the output to the figurine.
The output can then be presented to a child as though the figurine
processed and generated the output directly.
Inventors: |
Palmquist; Robert D.;
(Faribault, MN) |
Correspondence
Address: |
SHUMAKER & SIEFFERT, P. A.
8425 SEASONS PARKWAY
SUITE 105
ST. PAUL
MN
55125
US
|
Assignee: |
Speechgear, Inc.
Northfield
MN
|
Family ID: |
35510282 |
Appl. No.: |
11/146907 |
Filed: |
June 7, 2005 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
60578101 |
Jun 8, 2004 |
|
|
|
Current U.S.
Class: |
446/297 ;
446/175 |
Current CPC
Class: |
A63H 3/003 20130101;
A63H 2200/00 20130101; A63H 3/28 20130101; A63H 30/04 20130101 |
Class at
Publication: |
446/297 ;
446/175 |
International
Class: |
A63H 3/28 20060101
A63H003/28 |
Goverment Interests
STATEMENT REGARDING FEDERALLY SPONSORED RESEARCH OR DEVELOPMENT
[0002] This invention was made with Government support under
contract number N00014-02-C-0122 awarded by the Office of Naval
Research. The government has certain right in the invention.
Claims
1. A system comprising: a figurine that captures input and
wirelessly communicates the input; and a computer that receives the
input from the figurine, generates output based on the input, and
wirelessly communicates the output to the figurine, wherein the
figurine presents the output to a user.
2. The system of claim 1, wherein: the figurine captures speech
data from a user and wirelessly communicates the speech data; and
the computer receives the speech data from the figurine, generates
an audible response to the speech data, and wirelessly communicates
the response to the figurine, wherein the figurine outputs the
audible response to the user.
3. The system of claim 2, wherein the computer receives additional
data via a network, and wherein the computer generates output based
on the additional data.
4. The system of claim 1, wherein: the figurine captures voice data
from a user and wirelessly communicates the voice data; and the
computer receives the voice data from the figurine, identifies a
person associated with the voice data, and generates an audible
response to the voice data identifying the person.
5. The system of claim 1, wherein: the captured input comprises
speech data from a user; and the computer receives the speech data
from the figurine, generates a translation of the speech data, and
wirelessly communicates the translation to the figurine, wherein
the figurine outputs the translation to the user.
6. The system of claim 1, wherein: the captured input comprises
image data including one or more words or phrases; and the computer
receives the image data from the figurine, generates a translation
of the words or phrases, and wirelessly communicates the
translation to the figurine, wherein the figurine outputs the
translation to the user.
7. A system of claim 1, wherein: the captured input comprises image
data including one or more words or phrases; and the computer
receives the image data from the figurine, generates audio data
corresponding to the words or phrases, and wirelessly communicates
the audio data to the figurine, wherein the figurine outputs the
audio data to the user.
8. The system of claim 1, further comprising a wireless hub,
wherein the figurine wirelessly communicates with the computer via
the wireless hub.
9. The system of claim 8, further comprising the Internet, wherein
the figurine wirelessly communicates with the computer over the
Internet via the wireless hub.
10. The system of claim 1, wherein: the figurine captures image
data including an identifiable face or object; and the computer
receives the image data from the figurine, determines an
identification that identifies the identifiable face or object, and
wirelessly communicates the identification to the figurine.
11. The system of claim 1, wherein: the figurine captures
additional data from a compatible object; and the computer
generates output based on the additional data.
12. The system of claim 1, wherein the output is first output and
wherein the computer is programmed to proactively cause the
figurine to output second output to a user.
13. An interactive toy figurine comprising: an data capture device
to capture audio or video data; and a wireless transmitter/receiver
to wirelessly transfer data captured by the data capture device and
receive output associated with the data captured by the data
capture device.
14. The interactive toy figurine of claim 13, wherein the data
capture device comprises an image capture device, the figurine
further comprising a display to display at least one of an image
captured by the image capture device and the output.
15. The interactive toy figurine of claim 13, further comprising a
speaker to output audio data associated with the data captured by
the data capture device.
16. The interactive toy figurine of claim 13, wherein the data
captured by the data capture device include one or more written
words or phrases, and the audio data comprise an audible recitation
of the words or phrases.
17. The interactive toy figurine of claim 13, wherein the data
captured by the data capture device include one or more words or
phrases in a first language, and the audio data comprise a
translation of the words or phrases in a second language.
18. A method comprising: capturing data from a user at a figurine;
wirelessly communicating the data to an external computer;
receiving from the external computer a response to the data; and
outputting the response to the user from the figurine.
19. The method of claim 18, wherein the data comprise speech data
captured from the user, wherein outputting the response comprises
outputting an audible response to the speech data.
20. The method of claim 19, wherein receiving the response
comprises receiving a translation of the speech data and wherein
outputting the response comprises outputting the translation to the
user from the figurine.
21. The method of claim 18, wherein the data comprise image data
including one or more words or phrases, wherein receiving the
response comprises receiving audio data corresponding to the words
or phrases, and wherein outputting the response comprises
outputting the audio data from the figurine.
22. The method of claim 18, wherein the data comprise image data
including one or more words or phrases in a first language, wherein
receiving the response comprises receiving a translation of the
words or phrases in a second language, and wherein outputting the
response comprises outputting the translation from the
figurine.
23. The method of claim 22, wherein outputting the translation
includes driving a speaker to generate an audible recitation of the
translation.
24. The method of claim 22, wherein outputting the translation
includes driving a display to generate a visual translation.
25. A system comprising: a computer; and a figurine communicatively
coupled to the computer, wherein the figurine provides input to the
computer, receives output from the computer and outputs the output
to a user, and wherein functionality of the figurine is expandable
via upgrades to the computer.
26. A system comprising: a figurine that captures input and
wirelessly communicates the input; and a parents unit that receives
the input and generates an alarm based on the input.
27. The system of claim 26, further comprising a computer that
receives the input from the figurine, forwards the input to the
parents unit and causes the parents unit to generate the alarm.
28. The system of claim 26, wherein the input comprises breathing
information associated with a child and the parents unit generates
the alarm if the child stops breathing.
29. A system comprising: a computer; and a figurine communicatively
coupled to the computer; one or more system compatible objects,
wherein the figurine interacts with the one or more objects by
harnessing computing power of the computer, wherein the system
compatible objects include indicia identifiable by the figurine to
ensure that software on the computer can ensure interaction between
the figurine and the object.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This Application claims the benefit of U.S. Provisional
Patent Application No. 60/578,101, filed Jun. 8, 2004, the entire
content of which is incorporated herein by reference.
TECHNICAL FIELD
[0003] The invention relates to figurines such as stuffed animals,
teddy bears, dolls, toy robots, action figures, and the like, and
more particularly, to figurines that include electronics.
BACKGROUND
[0004] In this disclosure, the term "figurine" refers to a doll, a
teddy bear, a stuffed animal, a toy robot, a toy statue, an action
figure, and the like. Figurines are commonly used by children to
pass the time and facilitate imaginative thought. In recent times,
more advanced computerized figurines have been developed. These
more advanced figurines, for example, may incorporate electronics
that allow the figurine to interact with the child.
SUMMARY
[0005] In general, the invention is directed to a system including
a figurine that utilizes wireless communication to harness
computing power of an external computer. In particular,
applications that require intensive processing power can be
seemingly executed by the figurine, with the intensive processing
actually being performed by the external computer. The figurine may
capture input and wirelessly transfer the input to an external
computer, which processes the input. The external computer returns
output to the figurine, which presents the output to a child.
[0006] Speech recognition applications, speech interpretation
applications, image processing applications, voice recognition
applications, and language translation applications are some
examples of applications that typically require intensive
processing power and large amounts of memory. The invention
contemplates a figurine that utilizes wireless communication to
harness computing power of an external computer in order to
facilitate the presentation of speech recognition applications,
speech interpretation applications, image processing applications,
voice recognition applications, and language translation
applications through the figurine. By performing the intensive
processing external to the figurine, the internal electronics of
the figurine can be greatly simplified. In particular, the need for
intensive processing power and a large amount of memory in the
figurine can be avoided. Accordingly, the need to protect powerful
processors and memory from misuse by a child handing the figurine
can also be avoided. In addition, battery life in the figurine may
be extended by using the techniques described herein.
[0007] In one embodiment, the invention provides a system
comprising a figurine that captures input from a user and
wirelessly communicates the input. The input can be image data, for
example, or audio data such as speech data. The system also
includes a computer that receives the speech data from the
figurine, generates a response to the speech data, and wirelessly
communicates the response to the figurine. The figurine then
outputs the response to the user.
[0008] In another embodiment, the invention provides a system
comprising a figurine that captures speech data from a user and
wirelessly communicates the speech data. The system also includes a
computer that receives the speech data from the figurine, generates
a translation of the speech data, and wirelessly communicates the
translation to the figurine. The figurine outputs the translation
to the user.
[0009] In another embodiment, the invention provides a system
comprising a figurine that captures image data from a user and
wirelessly communicates the image data, wherein the image data
includes one or more words or phrases. The system also includes a
computer that receives the image data from the figurine, generates
a translation of the words or phrases, and wirelessly communicates
the translation to the figurine. The figurine outputs the
translation to the user.
[0010] In another embodiment, the invention provides a system
comprising a figurine that captures image data from a user and
wirelessly communicates the image data, wherein the image data
includes one or more words or phrases. The system also includes a
computer that receives the image data from the figurine, generates
audio data corresponding to the words or phrases, and wirelessly
communicates the audio data to the figurine. The figurine outputs
the audio data to the user.
[0011] In another embodiment, the invention provides an interactive
toy figurine comprising a data capture device and a wireless
transmitter/receiver to wirelessly transfer data captured by the
data capture device and receive output associated with the data
captured by the data capture device. For example, the data capture
device may be an image capture device to capture image data, such
as a camera deployed in one or both of the eyes of the toy
figurine, or elsewhere.
[0012] In another embodiment, a method comprises capturing speech
data from a user at a figurine, and wirelessly communicating the
speech data to an external computer. The method also comprises
receiving from the external computer a response to the speech data,
and outputting the response to the user from the figurine.
[0013] In another embodiment, a method comprises capturing speech
data from a user at a figurine, and wirelessly communicating the
speech data to an external computer. The method also comprises
receiving from the external computer a translation of the speech
data, and outputting the translation to the user from the
figurine.
[0014] In another embodiment, a method comprises capturing image
data with a figurine, and wirelessly communicating the image data
to an external computer. The image data includes one or more words
or phrases. The method also comprises receiving from the external
computer a translation of the words or phrases, and outputting the
translation from the figurine.
[0015] In another embodiment, a method comprises capturing image
data with a figurine, and wirelessly communicating the image data
to an external computer. The image data includes one or more words
or phrases. The method also comprises receiving from the external
computer audio data corresponding to the words or phrases, and
outputting the audio data from the figurine.
[0016] In another embodiment, a system comprises a figurine that
captures input and wirelessly communicates the input. The system
also includes a computer that receives the input from the figurine,
generates output based on the input, and wirelessly communicates
the output to the figurine. The figurine presents the output to a
user.
[0017] In another embodiment, a system comprises a figurine
communicatively coupled to a computer, which is in-turn
communicatively coupled to a server via a network. The figurine
provides input to the computer and receives output from the
computer. The computer can receive software updates from the server
such that functionality of the figurine can be changed or expanded
via computer software upgrades. Of course, upgrades may also be
loaded on the computer via a conventional disk or other storage
medium, in which case, communication with the server would not be
necessary.
[0018] In another embodiment, a system comprises a figurine
communicatively coupled to a computer. In addition, the system
includes one or more system compatible objects that the figurine
can interact with, harnessing the power of the computer. The
compatible objects may include indicia identifiable by the
figurine, which can ensure that the software on the computer can
provide useful interaction between the figurine and the object.
[0019] In another embodiment, a system comprises a figurine, a
computer, and a parent unit. The parent unit may comprise a
software module on the computer, or a separate hardware device. In
any case, the parent unit allows parents to exert parental control
over the functionality of the figurine by interacting with software
modules on the computer that control operation and interactive
features of the figurine. The parent unit may also function as a
baby monitor, e.g., a smart baby monitor that can generate an alarm
if a baby in proximity to the figurine ceases to breath, or has
other detectable problems.
[0020] The details of one or more embodiments of the invention are
set forth in the accompanying drawings and the description below.
Other features, objects, and advantages of the invention will be
apparent from the description and drawings, and from the
claims.
BRIEF DESCRIPTION OF DRAWINGS
[0021] FIG. 1 is a conceptual diagram illustrating a figurine
wirelessly communicating with a computer.
[0022] FIGS. 2 and 3 are block diagrams of a figurine wirelessly
communicating with a computer.
[0023] FIGS. 4-6 are flow diagrams according to embodiments of the
invention, illustrating application of the invention to translation
of spoken or written messages.
[0024] FIG. 7 is a conceptual diagram illustrating a figurine
wirelessly communicating with a computer via a wireless hub.
[0025] FIG. 8 is a conceptual diagram illustrating a figurine
wirelessly communicating with a computer thorough a network.
[0026] FIG. 9 is a conceptual diagram illustrating a figurine
wirelessly communicating with a computer and a compatible
object.
[0027] FIG. 10 is a conceptual diagram illustrating a figurine
wirelessly communicating with a computer, with a parents' unit.
[0028] FIG. 11 is a conceptual diagram illustrating a system in
which a server communicates with clients that wirelessly
communicate with figurines.
DETAILED DESCRIPTION
[0029] The invention is directed to a system including a figurine
that utilizes wireless communication to harness computing power of
an external computer. In particular, certain applications that
require intensive processing power can be seemingly executed by the
figurine, with the intensive processing actually being performed
external to the figurine in another computer. The figurine may
capture audio data, video data or both, and wirelessly transfer the
captured data to the external computer, either directly or via a
network. Audio data includes, but is not limited to, speech data,
voice data and music data. The external computer receives the data
as input, processes the data, generates output based on the input
data, and transfers the output to the figurine. The output can then
be presented to a child as though the figurine processed and
generated the output directly.
[0030] Speech recognition applications, voice recognition
applications, speech interpretation applications, and language
translation applications are some examples of applications that may
require intensive processing power and a large amount of memory.
The invention contemplates a figurine that utilizes wireless
communication to harness computing power of an external computer in
order to facilitate the presentation of speech recognition
applications, voice recognition applications, speech interpretation
applications, and language translation applications through the
figurine.
[0031] FIG. 1 is a diagram illustrating a system 10 according to an
embodiment of the invention. System 10 generally includes a
figurine 12 such as a doll, teddy bear, stuffed animal, toy robot,
toy statue, action figure or the like. System 10 also includes an
external computer 14 such as a personal computer (PC), Macintosh,
workstation, laptop, notebook, palm computer, or any other computer
external to figurine 12. Figurine 12 and external computer 14
communicate either directly or indirectly via one or more wireless
communication links 16. In some cases, external computer 14 may be
networked to one or more wireless hubs or other devices that
facilitate wireless communication.
[0032] Figurine 12 harnesses the computing power of external
computer 14 in order to facilitate execution of processor-intensive
and/or memory-intensive applications. A child can interact with
figurine 12. Accordingly, figurine 12 can facilitate learning and
provide instruction and guidance to the child. As figurine 12
harnesses the computing power of external computer 14 in order to
facilitate execution of these applications, the computing power and
memory needed in figurine 12 can be significantly reduced.
Accordingly, the need to protect processors and/or memory from
misuse, by a child handing figurine 12, can also be reduced. Also,
the power used by figurine 12 can be reduced, prolonging battery
life within figurine 12.
[0033] In one example, figurine 12 may present a speech recognition
application to the child, e.g., a program that teaches the child
the meanings of one or more words or phrases. In that case, the
child may speak to figurine 12, which captures the speech and
wirelessly communicates the captured speech to external computer
14. External computer 14 parses the speech and generates one or
more meanings, which are communicated back to figurine 12. Figurine
12 can then output the meanings to the child in any number of
ways.
[0034] By way of illustration, a child can utter the word "travel"
to figurine 12, which captures the utterance and wirelessly
communicates the captured speech to external computer 14. External
computer 14 parses the captured speech and generates one or more
definitions, which are communicated back to figurine 12. Figurine
12 can output the definition by, for example, responding "the word
`travel` means to go on a trip."
[0035] In another example, figurine 12 may be capable of holding
intelligent conversation with the child by harnessing the computing
power of external computer. In that case, the child may speak to
figurine 12, which captures the speech and wirelessly communicates
the captured speech to external computer 14. External computer 14
parses the speech and generates one or more responses, which are
communicated back to figurine 12. Figurine 12 can then output the
responses to the child. Thus, if the child asks a question to
figurine 12, figurine 12 may respond with an intelligent answer.
Software executing on external computer 14 may adapt over time to
the questions posed by the child, and may also be upgradeable.
Upgrades to software on external computer 14, for example, may
cause figurine 12 to appear to grow intellectually with the
child.
[0036] By way of illustration, a child can utter the words "I love
you" to figurine 12, which captures the utterance and wirelessly
communicates the captured speech to external computer 14. External
computer 14 parses the child's message and generates one or more
responses, which are communicated back to figurine 12. Figurine 12
can output the response by, for example, responding "I love you,
too." Or the child can utter the question "What is a triangle?" to
figurine 12, which can respond "A triangle is a shape that has
three sides."
[0037] In another example, figurine 12 may help the child with
reading. For example, one or more image capture devices, such as
digital cameras, may be located in figurine 12, such as in one or
more eyes 18 of figurine 12. The child may present a page of a book
to figurine 12 by directing the eyes 18 of figurine 12 toward the
page and pressing a button (not shown). In that case, a captured
image of the page can be wirelessly communicated to external
computer 14. External computer 14 parses the page and identifies
the words printed on the page. External computer 14 then
communicates back to figurine, so that figurine 12 can output the
words written on the page. In this sense, figurine 12 may appear to
be reading to the child. In that case, a display (not shown) may
also be incorporated into figurine 12 to present the child with the
captured words being read by figurine 12. Therefore, figurine 12
may aid in the learning process of the child by helping to teach
the child to read.
[0038] In another example, figurine 12 may facilitate translation
of words spoken by the child. For example, the child may speak to
figurine 12, which captures the speech and wirelessly communicates
the captured speech to external computer 14. External computer 14
parses the speech and identifies a translation of the words or
phrases spoken by the child. External computer 14 then communicates
back to figurine, so that figurine 12 can output the translations
to the child. In this example, figurine 12 serves as an
interpreter.
[0039] For purposes of illustration, a child can utter the words
"Thank you" to figurine 12, which captures the utterance and
wirelessly communicates the captured speech to external computer
14. External computer 14 parses the utterance and identifies a
translation of the phrase. External computer 14 then communicates
the translation back to figurine, and figurine 12 can output the
translations by, for example, responding with "`Gracias` means
`thank you` in Spanish.".
[0040] In another example, figurine 12 may facilitate translation
of written words. For example, one or more image capture devices,
such as digital cameras, may be located in the eyes 18 of figurine
12. The child may present words or phrases to figurine 12 by
directing the eyes 18 of figurine 12 toward the words or phrases.
The child may press a button (not shown) on figurine 12 to capture
the words or phrases being "viewed" by the figurine. In that case,
a captured image of the words or phrases can be wirelessly
communicated to external computer 14. External computer 14 parses
the words or phrases, and identifies a translation of the words or
phrases. External computer 14 then communicates back to figurine,
so that figurine 12 can output the translation. In that case, a
display (not shown) may also be incorporated into figurine 12 to
present the child with the captured words being translated. The
display, for example, may be located anywhere on figurine, but is
preferably located on the back of figurine so that the words can be
viewed by the child as eyes 18 of figurine 12 are directed away
from the child towards a page to be read.
[0041] In a further example, figurine 12 harnesses the computing
power of external computer 14 to perform image processing unrelated
to words and phrases. One or more image capture devices, such as
digital cameras, may be located in the eyes 18 of figurine 12, and
can capture image data to be processed by external computer 14.
Image processing can include recognition of faces, objects, colors,
numbers, places, activities, and the like. When external computer
14 runs face recognition software, for example, figurine 12 can
seem to recognize the person or persons interacting with figurine
12. Figurine 12 can use the recognition in its interaction by, for
instance, calling a child by name. When external computer 14 runs
object recognition software, figurine 12 can seem to recognize
objects and attributers of objects such as shape, type or quantity.
In an exemplary application, figurine 12 can teach a child to
recognize shapes, count objects, become familiar with colors, and
the like.
[0042] In yet another example, figurine 12 harnesses the computing
power of external computer 14 to perform voice recognition in order
to identify the speaker. Figurine 12 can use the voice recognition
in its interaction by, for instance, calling a child by name. When
external computer 14 runs voice recognition software, figurine 12
can seem to recognize the speaker. Typically, voice recognition
applications would be used along with speech recognition
applications.
[0043] Voice recognition applications refer to applications that
identify who is talking and may allow for programmed figurine
interaction only with those persons associated with a recognized
voice. Speech recognition applications refer to applications that
recognize what is being said and may be generally used with any
voice. In some cases, the invention may utilize both speech and
voice applications together in order to determine what is being
said and who is speaking. This can improve interaction with
figurine 12 such that figurine 12 may only respond to the child for
which it is programmed to respond. With voice recognition, a child
could say "sing me a song" or "tell me a story" and the figurine
may select a song or store from a library and respond to the
recognized voice, as directed. Responses from the figurine to
others, however, may be limited or prohibited if the requesting
voice is not recognized.
[0044] Interaction between a user and figurine 12 can be proactive
as well as reactive. In other words, external computer 14 can cause
figurine 12 to take action that is not in response to action by a
user. For example, figurine 12 may serve as an alarm clock, telling
a child that it is time to get out of bed. Figurine 12 may also
proactively remind a user of the day's appointments, birthdays of
friends or relatives, and the like. Thus, first output may be
provided, which is responsive to input to the figurine, and
computer 12 can be programmed to proactively cause figurine 12 to
output second output to a user, e.g., an alarm or reminder.
[0045] FIG. 2 is a functional block diagram of system 10 including
a figurine 12 and an external computer 14. Again, figurine 12 uses
wireless communication in order to harnesses the processing
horsepower of external computer 14. In this manner, complex
applications can be performed by computer 14, yet presented by
figurine 12 to a user.
[0046] Figurine 12 includes one or more input devices 22 to capture
input from a user, e.g., a child. Figurine 12 also includes one or
more output devices 23 present output to the user. Input device 22
may comprise, for example, a sound-detecting transducer such as a
microphone, or an image capture device, such as a digital camera. A
button or other actuator may be disposed on figurine 12 to turn on
the microphone or to cause the digital camera to take a picture.
Output device 23 may comprise a sound-generating transducer such as
a speaker, or possibly a display screen.
[0047] Sounds or images detected by input device 22 may be
processed locally by local central processing unit (CPU) 24 in
order to facilitate communication of the data to external computer
14. For example, local CPU 24 may package the captured input for
transmission to external computer 14. Local CPU 24 may also control
transmitter/receiver 26 to cause transmission of data indicative of
the sounds or images detected by input device 22. Local CPU 24, for
example, may comprise a relatively simple controller implemented in
application specific integrated circuit (ASIC). If images are
captured by figurine 12, local CPU 24 may compress the image file
to simplify wireless transfer of the image file. In any case,
transmitter/receiver 26 transfers data collected by figurine 12 so
that the data can be processed external to figurine 12.
[0048] The wireless communication between transmitter/receiver 26
of figurine 12 and transmitter/receiver 27 of external computer 14
may conform to any of a wide variety of wireless communication
protocols. Examples include, but are not limited to a wireless
networking standard such as one of the IEEE 802.11 standards, a
standard according to the Bluetooth Special Interest Group, or the
like. The IEEE 802.11 standards include, for example, the original
802.11 standard having data transfer rates of 1-2 Megabits per
second (Mbps) in a 2.4-2.483 Gigahertz (GHz) frequency band, as
well as the IEEE 802.11b standard (sometimes referred to as 802.11
wireless fidelity or 802.11 Wi-Fi) that utilizes binary phase shift
keying (BPSK) for 1.0 MBPS transmission, and quadrature phase shift
keying (QPSK) for 2.0, 5.5 and 11.0 Mbps transmission, the IEEE
802.11g standard that utilizes orthogonal frequency division
multiplexing (OFDM) in the 2.4 GHz frequency band to provide data
transmission at rates up to 54 Mbps, and the IEEE 802.11a standard
that utilizes OFDM in a 5 GHz frequency band to provide data
transmission at rates up to 54 Mbps. These and other wireless
networks have been developed. Additional extensions to the IEEE
802.11 standard, as well as other wireless standards will likely
emerge in the future. The invention is not limited to the type of
wireless communication techniques used, and could also be
implemented with wireless protocols similar to those used for cell
phone communication or direct two-way pagers, or any other wireless
protocol, either known or later developed.
[0049] Transmitter/receiver 27 of external computer 14 receives
data sent by transmitter/receiver 26 of figurine 12. Remote CPU 28
performs extensive processing on the data to generate output. For
example, remote CPU 28 may comprise a general purpose
microprocessor that executes software to generate the output. The
output is then transmitted back to figurine 12. Output device 23 of
figurine 12 can then present the output to the user.
[0050] By performing the intensive processing in external computer
14, the local electronics of figurine 12 can be greatly simplified.
In particular, the need for intensive processing power and a large
amount of memory in figurine 12 can be avoided. Accordingly, the
need to protect powerful processors and memory from misuse by a
child handing the figurine can also be avoided. Battery power in
figurine 12 can also be extended by performing processing tasks
externally in computer 14. Moreover, software upgrades may be
easily implemented for execution by remote CPU 28 without requiring
upgrade of the components of figurine 12.
[0051] The processing tasks performed in remote CPU 28 of external
computer 14, generally depend on the given application being
presented to the user by figurine 12. In one example, figurine 12
may present a speech recognition application to the child, e.g., a
program that teaches the child the meanings of one or more words or
phrases. In that case, the child may speak to figurine 12, and
input device 22 can capture the speech. Local CPU 24 packages the
speech and causes transmitter/receiver 26 to wirelessly
communicates the captured speech to external computer 14. Remote
CPU 28 parses the speech and generates one or more meanings, which
are communicated back to figurine 12 by transmitter/receiver 27.
Output device 23 of figurine 12 can then output the meanings to the
child.
[0052] In another example, figurine 12 may be capable of holding
intelligent conversation with the child by harnessing the remote
CPU 28 of external computer 14. In that case, the child may speak
to figurine 12, and input device 22 can capture the speech. Local
CPU 24 packages the speech and causes transmitter/receiver 26 to
wirelessly communicate the captured speech to external computer 14.
Remote CPU 28 parses the speech and generates one or more
responses, which are communicated back to figurine 12 by
transmitter/receiver 27. Output device 23 of figurine 12 can then
output the responses to the child. Thus, if the child asks a
question to figurine 12, figurine 12 may respond with an
intelligent answer.
[0053] In another example, figurine 12 may help the child with
reading. In that case, input device 22 in the form of an image
capture device may capture images of a page. Local CPU 24 packages
the image and causes transmitter/receiver 26 to wirelessly
communicate the captured image to external computer 14. Local CPU
may also compress the image prior to transmission. Once external
computer 14 has received the captured image, remote CPU 28 parses
the image and generates one or more meanings, which are
communicated back to figurine 12 by transmitter/receiver 27. For
example, remote CPU 28 may perform character recognition on the
image in order to identify characters, and may then decipher the
meaning of the identified characters using one or more dictionaries
stored in memory and accessible by remote CPU 28. Output device 23
of figurine 12 can then output the meanings to the child. In this
sense, figurine 12 may appear to be reading to the child. Output
device 23 may include speakers for verbal output and possibly a
display to present the child with the captured words being read by
figurine 12.
[0054] In another example, figurine 12 may facilitate translation
of words spoken by the child. In that case, the child may speak to
figurine 12, and input device 22 can capture the speech. Local CPU
24 packages the speech and causes transmitter/receiver 26 to
wirelessly communicate the captured speech to external computer 14.
Remote CPU 28 parses the speech and identifies a translation of the
word or phrases spoken by the child, which are communicated back to
figurine 12 by transmitter/receiver 27. Output device 23 of
figurine 12 can then output the translation to the child. In this
example, figurine 12 serves as an interpreter.
[0055] In a further example, figurine 12 harnesses the computing
power of external computer 14 to perform other types of image
processing. One or more image capture devices can capture image
data to be processed by external computer 14. Image processing can
include recognition of faces, objects, colors, numbers, places,
activities, and the like. When external computer 14 runs face
recognition software, for example, figurine 12 can seem to
recognize the person or persons interacting with figurine 12.
Figurine 12 can use the recognition in its interaction by, for
instance, calling a child by name. When external computer 14 runs
object recognition software, figurine 12 can seem to recognize
objects and attributers of objects such as shape, type or quantity.
In an exemplary application, figurine 12 can teach a child to
recognize shapes, count objects, become familiar with colors, and
the like.
[0056] Interaction between a user and figurine 12 can be proactive
as well as reactive. In other words, external computer 14 can cause
figurine 12 to take action that is not in response to action by a
user. For example, figurine 12 may serve as an alarm clock, telling
a child that it is time to get out of bed. Figurine 12 may also
proactively remind a user of the day's appointments, birthdays of
friends or relatives, and the like. For elderly applications,
figurine 12 may remind a user to take medication. Any such alarms
or reminders may be standard audio tones, music, or possibly
programmed or recorded audio of a familiar voice, making figurine
12 speak with a pleasant tone to the user when providing reminders.
For example, a parent's voice may be recorded such that figurine 12
speaks with such recordings. Voice emulation software may also be
used by computer 14 so that figurine 12 speaks new words or phrases
in a voice that emulates that of the parents.
[0057] FIG. 3 is a more detailed block diagram of system 30
illustrating application of the invention to one of the example
applications described above, in particular, translation of written
words. System 30 may correspond to system 10 (FIGS. 1 and 2). In
this case, image capture device 33 may capture images of a page.
For example, image capture device 33 may comprise a digital camera
located in the eyes of figurine 32 so that when a child directs the
eyes of figurine 32 toward a page and presses an actuator, the
image of the page is captured. The actuator, for example, may be
disposed on the back of figurine 32 so that when the eyes of
figurine 32 are directed toward a page, the actuator is easily
accessible. The image capture device and actuator, however, may be
deployed in other locations on figurine 32.
[0058] Local CPU 34 packages the image and causes
transmitter/receiver 36 to wirelessly communicate the captured
image to external computer 31. Remote CPU 38 parses the image and
generates a translation, which is communicated back to figurine 32
by transmitter/receiver 37. Remote CPU 28 may invoke software
modules 37, 39 to specifically perform optical character
recognition 39 and translation 40. Once the image has been
translated and the translation has been communicated back to
figurine 32, output device 35 of figurine 32 can then output the
translation.
[0059] Different optical character recognition modules and
translator modules may also be invoked for different languages.
Optical character recognition module 39, for example, may recognize
English, and translator module 40 may translate from English to
Spanish. Any other languages, however, could also be supported. In
another embodiment, optical character recognition may be performed
locally at figurine, with the more processor-intensive translation
being performed by external computer 31.
[0060] Other exemplary applications can be supported in a manner
similar to that depicted in FIG. 3. In the context of face
recognition, for example, image capture device 33 may capture one
or more images of a face. Local CPU 34 packages the image and
causes transmitter/receiver 36 to wirelessly communicate the
captured image to external computer 31. Remote CPU 38 executes face
recognition software modules to identify the face in the image.
Once the face has been identified remote CPU 38 may then
incorporate that identity into the output of figurine 32 by, for
example, referring to the user by name. Voice recognition could
also be used to cause figurine 32 to refer to the user by name.
[0061] In addition, remote CPU 38 can execute shape recognition
software modules, color recognition software modules, object
recognition software modules, quantification software modules.
Remote CPU 38 can use such software modules to help a user
recognize shapes, count objects, become familiar with colors, and
the like. Although a user perceives all action occurring through
figurine 32, processor-intensive image processing is actually being
performed by remotely by external computer 31. An object
recognition module, for example, may be designed to recognize
currency (such as coins) and allow the figurine to teach a child
how to accurately count change.
[0062] The various modules and components described herein may be
implemented in hardware, software, firmware, or any combination.
The invention is not limited to any particular hardware or software
implementation. If implemented in software, the modules may be
stored on a computer readable medium such as memory or a
non-volatile storage medium. Indeed, one advantage of the
techniques described herein is that the need for large amounts of
memory in a figurine can be avoided. Instead, the memory needed to
execute very memory-intensive applications is included in an
external computer. Further, the exemplary applications described
above are not exclusive of one another. External computer 31 can
execute any combination of translation, voice and speech
recognition, image processing or other types of software modules,
and the invention is not limited to systems that perform a single
application to the exclusion of other applications. On the
contrary, an advantage of the invention is its versatility. The
invention can be adapted to implement one or more applications as
desired by the user.
[0063] FIGS. 4-6 are flow diagrams according to some embodiments of
the invention, illustrating application of the invention to one of
the example applications described above, in particular,
translation of spoken or written messages. In the technique shown
in FIG. 4, figurine 12 performs speech capture (41), and then
transmits speech data to external computer 14 (42). External
computer 14 receives the speech data (43) and may perform speech
recognition (44) to identify the spoken words or phrases. External
computer 14 performs translation with respect to the identified
spoken words or phrases (45) and transmits the translation back to
figurine 12 (46). Figurine 12 receives the translation (47) and
outputs the translation to the user. Figurine 12 may drive an
output device such as a display screen, thereby providing a written
output. A user may find it more desirable, however, to have
figurine 12 drive a speaker in figurine 12, thereby providing an
audible output, such as a synthesized speech recitation of the
translation. In this manner, figurine 12 acts as a translator of
spoken words or phrases, invoking an external computer 14 to reduce
the local processing at figurine 12. Also, as mentioned above, in
addition to performing speech recognition, computer 14 may also
perform voice recognition so that figurine 12 only responds to
recognized voices, or responds differently, e.g., by identifying
different persons, in response to computer 14 recognizing different
voices.
[0064] In the technique shown in FIG. 5, figurine 12 captures an
image (51), and then transmits the image to external computer 14
(52). External computer 14 receives the image (53) and decodes the
image (54), e.g., by performing optical character recognition.
External computer then translates the characters to generate a
translation (55). External computer 12 then transmits the
translation back to figurine 12 (56). Figurine 12 receives the
translation (57) and outputs the translation to the user (58). In
this manner, figurine 12 acts as a translator of written words or
phrases, invoking an external computer 14 to reduce the local
processing at figurine 12. The translation may be output in audio,
video, or both.
[0065] In the technique shown in FIG. 6, figurine 12 captures an
image that includes written words or phrases (61), and then
transmits the images to external computer 14 (62). External
computer 14 receives the image (63) and decodes the image (64)
e.g., by performing optical character recognition to identify
written words or phrases. External computer 14 then generates an
audio signal (65) as a function of the identified words or phrases.
External computer 12 then transmits the audio signal back to
figurine 12 (66). Figurine 12 receives the audio signal (67) and
outputs the translation to the user. In this manner, figurine 12
appears to be reading the written words or phrases, by invoking an
external computer 14 to reduce the local processing at figurine
12.
[0066] In the various embodiments described herein, a figurine is
described that utilizes wireless communication to harness computing
power of an external computer. However, the figurine need not be
directly coupled to the external computer in order to take
advantage of the computing resources of the external device. FIG. 7
is a block diagram of a system 70, similar to system 10. In system
70, however, figurine 72 wirelessly communicates with external
computer 74 via a wireless hub 75. In particular, wireless hub 75
communicates wirelessly with figurine 72, and is coupled to
external computer 74.
[0067] FIG. 8 illustrates another system 80, similar to system 10,
in which figurine 82 wirelessly communicates to take advantage of
computing power of external computer 84. In system 80, figurine 82
wirelessly communicates with external computer 84 via a wireless
hub 85 that couples to external computer 84 via network 86. In
particular, wireless hub 85 communicates wirelessly with figurine
82, and is coupled to external computer 84 via network 86. Network
86 may comprise a small local area network (LAN), a wide area
network, or even a global network such as the Internet.
Communication between hub 85 and external computer 84 may be, but
need not be, wireless. Importantly, the wireless capabilities of
figurine 82 allow for communication with external computer 84,
thereby allowing figurine 82 to make use of the processing
capabilities of external computer 84.
[0068] When a figurine is configured to communicate with a global
network such as the Internet, such as is depicted in FIG. 8 or 11,
the figurine can serve as an input-output device for interaction
with the network and other stations or servers coupled to network
82. In a typical application, figurine 82 reports information
obtained from one or more network servers (not shown). For example,
a child could ask figurine 82, "What is the weather forecast for
today?" The request is relayed to external computer 84, which
accesses a server via network 86 that can provide the local
forecast. Upon retrieving the local forecast, external computer 84
supplies that information to figurine 82, which answers the child's
question.
[0069] The invention supports wireless communication in other
configurations as well. The figurine need not be directly coupled
to the external computer in order to take advantage of the
computing resources of the external device. Figurine 82 may
communicate wirelessly with any intermediate device, such as
another figurine, a wireless access point, or a computer that does
not serve as external computer 84.
[0070] FIG. 9 is a diagram illustrating a system 90 according to an
additional embodiment of the invention. System 90 includes a
figurine 92 and an external computer 94, which communicate either
directly or indirectly via one or more wireless communication
links. In addition, system 90 includes a compatible object 95,
embodied in FIG. 9 as a book. Compatible object 95 can be an object
of any type, but in a typical implementation, compatible object 95
is an accessory for figurine 92.
[0071] Compatible object 95 includes a wireless identifier, by
which a detector in figurine 92 can detect the presence of
compatible object 95. An example of such a wireless identifier is a
radio frequency identification (RFID) tag 96. RFID tag 96 may be
hidden in compatible object 95 and not readily observable to a
user. An RFID tag reader 98 in figurine 92 detects and reads RFID
tag 96. Bar codes or other indicia might also be used, in which
case reader 98 would facilitate the reading of such indicia.
[0072] RFID tag 96 is a wireless electronic device that
communicates with RFID tag reader 98. RFID tag 96 may include an
integrated circuit (not shown) and a coil (not shown). The coil may
act as a source of power, as a receiving antenna, and a
transmitting antenna. The coil may be coupled to capacitor to store
power when interrogated in order to drive the integrated circuit.
The integrated circuit may include wireless communications
components and memory. RFID tag reader 98 may include an antenna
and a transceiver.
[0073] RFID tag reader 98 may "interrogate" RFID tag 96 by
directing an electromagnetic (i.e., radio) signal to RFID tag 96.
RFID tag 96 may include, but need not include, an independent power
source. In a typical embodiment, RFID tag 96, receives power from
the interrogating signal from RFID tag reader 98. Upon power-on,
RFID tag 96 may perform certain operations, which may include
transmitting data stored in the memory of the RFID tag 96 to RFID
tag reader 98. The transmitted data may include an identification
of compatible object 95. In this manner, the manufacturer of
figurine 92 may exert better control over those objects that will
be used to interact with figurine, and can help ensure that a child
will not become frustrated, e.g., if figurine 92 were used with an
incompatible book or object. When RFID tag reader 98 identifies
RFID tag 96, external computer 94 becomes aware of compatible
object 95 proximate to figurine 92. External computer 94 can use
the identity of compatible object 95 to communicate with a user
more effectively. For example, exemplary compatible object 95 in
FIG. 9 is a book. When external computer 94 learns the identity of
the book, external computer 94 can generate output appropriate for
that book. Figurine 92 may, for example, direct the attention of a
child to illustrations shown in the book, and explain how the
illustrations pertain to the story. Figurine 92 can also describe
how the story relates to other books, such as other books dealing
with the same characters, or figurine 92 can explain background
information about the story or its author.
[0074] FIG. 10 is a diagram illustrating a system 100 according to
another embodiment of the invention. System 100 includes a figurine
102 and an external computer 104, which communicate either directly
or indirectly via one or more wireless communication links. In
addition, system 100 includes a parents' unit 106. Parents' unit
106 may comprise any device, including, but not limited to, a
television, a computer, a telephone, a speaker, a video monitor and
the like. Parents' unit 106 may communicate with external computer
104 in any fashion, such as by an electrical connection, by an
optical link, or by radio frequency.
[0075] System 100 is configured to serve as a child monitoring
system. A parent can deploy figurine 102 proximate to a child so
that figurine 102 can capture video information or audio
information or both about that child. Figurine 102 transmits the
captured information to external computer 104. External computer
104 in turn sends information to parents' unit 106. In this
example, the parent may also communicate in real time to the child
through figurine 102, e.g., by speaking into a microphone of
parents' unit 106. Parents' unit 106 may be a separate unit, or may
be implemented as a software module that executes directly on
external computer 104.
[0076] In one application, external computer 104 simply relays
captured information to parents' unit 106. For example, captured
audio and video data showing the child's location, condition and
activity may be relayed to parents' unit 106. In another
application, external computer 104 can also process the captured
audio and video data and provide useful information to parents'
unit 106. For example, external computer 104 can process audio data
captured via figurine 102 and determine whether the child is
crying, sleeping, breathing abnormally, and the like. External
computer 104 can also process video data captured via figurine 102
and determine whether the child is awake or has gotten out of bed
or the like.
[0077] FIG. 11 is a diagram illustrating a system 110 according to
another embodiment of the invention. System 110 is a server-client
system tin which a server 112 supplies one or more functionalities
to one or more client figurine-computer systems 114, 116. Server
112 manages a database 113 that stores software that can provide
figurines with one or more functionalities. Client
figurine-computer systems 114, 116 download one or more
functionalities from server 112 via a network 118. Network 118 may
comprise any network, including a global network such as the
Internet.
[0078] Examples of functionalities include, but are not limited to,
the functionalities described herein. The owner of client
figurine-computer system 114, for example, may desire that her
child's figurine 122 should be capable of helping teach her child
about numbers, letters, basic shapes and basic colors. In addition,
her child's figurine 122 should be capable of reciting stories
suitable for a child four years of age. Accordingly, the owner of
client figurine-computer system 114 downloads software for such
functionalities from server 112 via network 118. The software is
stored locally at external computer 120. By contrast, the owner of
client figurine-computer system 116 may desire that his child's
figurine 124 should be capable of reading a book, helping teach his
child speak and write in English and Spanish, and playing games
appropriate for a child six years of age. Accordingly, the owner of
client figurine-computer system 116 downloads software for such
functionalities from server 112 via network 118.
[0079] With system 110, each parent can customize his or her
child's figurine for the child's age, needs or desires. As the
child develops, the parent can obtain more advanced functionality.
Further, as new functionalities are developed and added to database
113, the parents can download the new functionalities. As a result,
the figurines seem to "grow" with the children, and can be enabled
to perform new or more sophisticated functions. Because the new and
more advanced functionality can be executed in external computer
120, the need to upgrade figurine 122 may be avoided, which can be
important to a child that has become emotionally attached to
figurine 122.
[0080] Although the figures may depict a single figurine with a
single external computer, the invention encompasses embodiments in
which a single external computer interacts with two or more
figurines. A parent with two children can, for example, give a
different figurine to each child, and each figurine can communicate
wirelessly with the same or different external computers. Each
child will perceive that each figurine operates independently of
the other. Furthermore, each figurine may be separately empowered
with functionality appropriate for each child.
[0081] The invention may offer one or more advantages. A child's
toy can be very versatile, capable of a wide range of
functionality. Furthermore, the functionality can be customized to
the child, and can change as the child develops. Moreover, the
invention supports an interesting and adaptable system that can
help a child learn a wide range of subjects, making interaction
with the figurine not only fun, but educational as well.
[0082] Several embodiments of the invention have been described.
Various modifications may be made without departing from the scope
of the invention. These and other embodiments are within the scope
of the following claims.
* * * * *