U.S. patent application number 11/140483 was filed with the patent office on 2006-11-30 for interactive animated characters.
Invention is credited to Jeffrey M. Ford, Peter Hall, Richard J. Maddocks, Eduardo J. Rodriguez.
Application Number | 20060270312 11/140483 |
Document ID | / |
Family ID | 37464067 |
Filed Date | 2006-11-30 |
United States Patent
Application |
20060270312 |
Kind Code |
A1 |
Maddocks; Richard J. ; et
al. |
November 30, 2006 |
Interactive animated characters
Abstract
A very compact interactive animated character is provided that
provides highly life-like and intelligent seeming interaction with
the user thereof. The animated character can take the form of a
small animal-like creature having a variety of moving body parts
including a smile/emotion assembly which are coordinated to exhibit
life-like emotional states by precisely controlling and
synchronizing their movements in response to external sensed
conditions. The animated character also includes sound generating
circuitry to generate speech sounds as well as sounds associated
with various emotional states, which are coordinated with a lip
sync assembly simulating speech mouth movement. The drive system
utilizes first and second reversible motors which are able to power
and precisely coordinate the lip sync assembly producing speech
mouth movement, with the movable parts and the smile/emotion
assembly to produce life-like interactions and emotional
expressions.
Inventors: |
Maddocks; Richard J.;
(Barrington, RI) ; Rodriguez; Eduardo J.;
(Tiverton, RI) ; Ford; Jeffrey M.; (Warwick,
RI) ; Hall; Peter; (Norfolk, GB) |
Correspondence
Address: |
PERRY HOFFMAN & ASSOCIATES P.C.
PO BOX 1649
DEERFIELD
IL
60015
US
|
Family ID: |
37464067 |
Appl. No.: |
11/140483 |
Filed: |
May 27, 2005 |
Current U.S.
Class: |
446/337 |
Current CPC
Class: |
A63H 3/28 20130101; A63H
13/005 20130101; A63H 3/40 20130101 |
Class at
Publication: |
446/337 |
International
Class: |
A63H 3/36 20060101
A63H003/36 |
Claims
1. An electrically controlled animating apparatus for simulating
life-like movements, the apparatus comprising: a front facial area,
a body and a plurality of movable body parts thereof; sound
generating circuitry for generating speech including multisyllabic
words; a plurality of sensors for detecting external inputs; a
controller responsive to the plurality of sensors; said controller
being operable to control the sound generating circuitry; a mouth
assembly on the front facial area comprising flexible molded
material having upper and lower mouth portions and having first and
second opposing corners thereof; a first mouth mechanism operable
with the mouth assembly for controlling the first and
second-corners of the mouth assembly to define smile/emotion states
of the mouth assembly responsive to external input from sensed
conditions of the plurality of sensors; a second mouth mechanism
operable with the mouth assembly for controlling the upper and
lower mouth portions to provide lip synchronism response to the
multisyllabic words generated with the sound generating circuitry;
a drive system that powers movement of the first and second mouth
mechanisms independently to simulate life-like responses to sensed
conditions; a first control shaft for the first mouth mechanism
driven for rotation by the drive system; a second control shaft for
the second mouth mechanism driven by the drive system; at least one
of the first or second control shafts having a predetermined range
of rotation for causing movement of at least one of the plurality
of movable body parts in addition to the first or second
mechanisms.
2. The animating apparatus as recited in claim 1 wherein said first
control shaft causes movement of at least one of the plurality of
movable body parts in addition to the first and second corners of
the mouth assembly.
3. The animating apparatus as recited in claim 2 wherein said drive
system further comprises a first reversible motor which powers the
rotation of said first control shaft for movement of the corners of
the mouth assembly, and a second reversible motor which powers the
rotation of said second control shaft for movement of the upper and
lower mouth portions of the mouth assembly.
4. The animating apparatus as recited in claim 3 further comprising
a mechanical coupling between said second motor and said second
control shaft for transmitting rotary output power from the second
motor to the second shaft for rotation thereof.
5. The animating apparatus as recited in claim 4 wherein said at
least one of the plurality of movable body parts includes a foot
portion, and rotation of said second control shaft in a direction
causes movement of said foot portion and rotation of said second
control shaft in an opposite direction causes movement of said
second mouth mechanism for controlling upper and lower mouth
portions of said mouth assembly.
6. The animating apparatus as recited in claim 5 wherein said
mechanical coupling comprises one or more of a shuttle gear or a
clutch mechanism between said second motor and said second control
shaft for transmitting rotary output power from the second motor to
the second shaft for rotation of said second control shaft in the
direction causing movement of said foot portion and rotation of
said second control shaft in the opposite direction causing
movement of said second mouth mechanism.
7. The animating apparatus as recited in claim 1 wherein said
controller provides the animating apparatus with a plurality of
states comprising animating apparatus modes that include excited,
sleeping, and waking modes.
8. The animating apparatus as recited in claim 2 wherein said at
least one of a plurality of body parts includes one or more of the
following: an ear assembly; an eye assembly; an eye lid assembly; a
plume assembly; a brow assembly; and a chest assembly.
9. The animating apparatus as recited in claim 2 wherein said first
and second corners of the mouth assembly pivot to a neutral state
an up/smile state and a down/frown state.
10. The animating apparatus as recited in claim 9 wherein said
first control shaft has a neutral position in the predetermined
range of shaft rotation with the corners of the mouth assembly in a
neutral state.
11. An animating system comprising: means for generating speech
including multisyllabic words; means for detecting external sensor
inputs; means for controlling the means for generating speech
responsive to the means for detecting external sensor inputs; said
means for controlling being operable to control a mouth assembly
comprising flexible molded material having upper and lower mouth
portions and having first and second opposing corners thereof; said
means for controlling being operable to control a first mouth
mechanism operable with the mouth assembly for controlling the
first and second corners of the mouth assembly to define
smile/emotion states of the mouth assembly responsive to external
input from sensed conditions; said means for controlling being
operable to control a second mouth mechanism operable with the
mouth assembly for controlling the upper and lower mouth portions
to provide lip synchronism response to the multisyllabic words
generated with the sound generating; and means for driving movement
of the first and second mouth mechanisms independently to simulate
life-like responses to sensed conditions including a first control
shaft for the first mouth mechanism driven for rotation by the
drive system, and a second control shaft for the second mouth
mechanism driven by the drive system with at least one of the first
or second control shafts having a predetermined range of rotation
for causing movement of at least one of the plurality of movable
body parts in addition to the first or second mechanisms.
12. The animating system as recited in claim 11 wherein said first
control shaft causes movement of at least one of the plurality of
movable body parts in addition to the first and second corners of
the mouth assembly.
13. The animating system as recited in claim 12 wherein said drive
system further comprises a first reversible motor which powers the
rotation of said first control shaft for movement of the corners of
the mouth assembly, and a second reversible motor which powers the
rotation of said second control shaft for movement of the upper and
lower mouth portions of the mouth assembly.
14. The animating system as recited in claim 13 further comprising
means for coupling between said second motor and said second
control shaft for transmitting rotary output power from the second
motor to the second shaft for rotation thereof.
15. The animating system as recited in claim 14 wherein said at
least one of the plurality of movable body parts includes a foot
portion, and rotation of said second control shaft in a direction
causes movement of said foot portion and rotation of said second
control shaft in an opposite direction causes movement of said
second mouth mechanism for controlling upper and lower mouth
portions of said mouth assembly.
16. An animating method comprising: generating speech including
multisyllabic words; detecting external sensor inputs; controlling
a mouth assembly comprising flexible molded material having upper
and lower mouth portions and having first and second opposing
corners thereof for generating speech responsive thereto and
detecting external sensor inputs; controlling a first mouth
mechanism operable with the mouth assembly for controlling the
first and second corners of the mouth assembly to define
smile/emotion states of the mouth assembly responsive to external
input from sensed conditions; controlling a second mouth mechanism
operable with the mouth assembly for controlling the upper and
lower mouth portions to provide lip synchronism response to the
multisyllabic words generated with the sound generating; and
driving movement of the first and second mouth mechanisms
independently to simulate life-like responses to sensed conditions
including a first control shaft for the first mouth mechanism
driven for rotation by the drive system, and a second control shaft
for the second mouth mechanism driven by the drive system with at
least one of the first or second control shafts having a
predetermined range of rotation for causing movement of at least
one of the plurality of movable body parts in addition to the first
or second mechanisms.
17. The animating method as recited in claim 16 wherein said first
control shaft causes movement of at least one of the plurality of
movable body parts in addition to the first and second corners of
the mouth assembly.
18. The animating method as recited in claim 16 wherein said
controlling provides a plurality of states comprising modes that
include excited, sleeping, and waking modes.
19. The animating method as recited in claim 18 wherein at least
one of a plurality of body parts includes one or more of the
following: an ear assembly; an eye assembly; an eye lid assembly; a
plume assembly; a brow assembly; and a chest assembly.
20. The animating method as recited in claim 17 wherein said first
and second corners of the mouth assembly pivot to a neutral state
an up/smile state and a down/frown state, and said first control
shaft has a neutral position in the predetermined range of shaft
rotation with the corners of the mouth assembly in a neutral
state.
21. An electrically controlled animating apparatus for simulating
life-like movements, the apparatus comprising: a front facial area
of a character head and a plurality of movable parts thereof; a
mouth assembly on the front facial area comprising flexible molded
material having upper and lower lips and having first and second
opposing corners where the lips meet; a first mouth mechanism
operable with the mouth assembly for animating the first and second
corners of the mouth assembly; a second mouth mechanism operable
with the mouth assembly for animating the upper and lower mouth
portions for animating the lips between opened and closed positions
independent of the first mouth mechanism; a drive system that
powers movement of the first and second mouth mechanisms
independently to simulate life-like responses and for causing
movement of at least one of the plurality of movable parts in
addition to the first or second mouth mechanisms.
22. The animating apparatus as recited in claim 21 wherein said
first and second corners of the first mouth mechanism are moveable
between neutral, up/smile state and down/frown states.
23. The animating apparatus as recited in claim 21 wherein said
drive system further comprises a first reversible motor which
powers the movement of the corners of the mouth assembly, and a
second reversible motor which powers the movement of the upper and
lower mouth portions of the mouth assembly independently.
24. The animating apparatus as recited in claim 21 wherein the
drive system facilitates the animating apparatus with a plurality
of states comprising animating apparatus modes that include
excited, sleeping, and waking modes.
25. The animating apparatus as recited in claim 24 wherein the
plurality of modes includes movement of one or more of the
following moveable parts: an ear assembly; an eye assembly; an eye
lid assembly; a plume assembly; a brow assembly; and a chest
assembly.
26. An electrically controlled animating apparatus for simulating
life-like movements in a front facial area of a character head, the
apparatus comprising: a molded flexible elastomeric mouth including
upper and lower lips having center portions and opposite side
corners where the lips meet, the mouth including attachment points
at the center portions and corners; a first animation mechanism for
animating the mouth for simulated talking movement by moving the
attachment points at the center portions; and a second animating
mechanism independent from the first animating mechanism for moving
the attachment points at the corners to simulate mouth movement
corresponding to facial expression; wherein the first and second
animating mechanisms act to distort the elastomeric mouth to
achieve simulated talking movement and facial expression as the
animating mechanisms are operated.
27. The animating apparatus as recited in claim 26 comprising a
drive system that powers movement of the first animation mechanism
independent of the movement of the second animation mechanism for
providing simulated talking movement or facial expression, or a
combination thereof as the animating mechanisms are operated.
28. The animating apparatus as recited in claim 26, said
elastomeric mouth comprising a neutral undeformed position, wherein
the second animating mechanism moves the attachment points at the
corners to deform the elastomeric mouth from the neutral undeformed
position.
29. The animating apparatus as recited in claim 28 wherein said
opposite side corners where the lips meet are moveable between
neutral, up/smile state and down/frown states with the second
animating mechanism.
Description
FIELD OF THE INVENTION
[0001] The present invention relates to interactive toys and, more
particularly, to animated characters that can perform movements
with body parts thereof in a precisely controlled and coordinated
manner in response to external sensed conditions.
BACKGROUND OF THE INVENTION
[0002] One major challenge with toys in general is keeping a child
interested in playing with the toy for more than a short period of
time. To this end, toy dolls and animals have been developed that
can talk and/or have moving body parts. The goal with these devices
is to provide a plaything that appears to interact with the child
when they play with the toy.
[0003] Whereas prior art interactive toys have several moving
parts, the life-like action attributed to these moving parts is due
to the random nature of their movements with respect to each other
as the individual parts tend to move in a predictable cyclic
action; in other words, there is no control over the motion of a
specific part individually on command in prior toys, and highly
controlled coordination of one part with the movement of other
parts is generally not done. Emotion has proven difficult to
capture with conventional mechanical actuators, and thus it would
be desirable to provide better coordinated constituent assemblies
to exhibit life-like emotional states by precisely controlling and
synchronizing their movements in response to external sensed
conditions. Moreover, coordination with sound generating circuitry
would be desirable to generate speech sounds as well as sounds
associated with various emotional states that are coordinated with
a lip synchronization simulating speech and mouth movements. Thus,
there is a need for an animated character that provides for more
precisely controlled and coordinated movements between its various
moving parts and allows for individual parts to be moved in a more
realistic manner.
SUMMARY OF THE INVENTION
[0004] In accordance with the present invention, a very compact
animated character is provided that provides highly life-like and
intelligent seeming interaction with the user thereof. The animated
character can take the form of a small animal-like creature having
a variety of moving body parts including a smile/emotion assembly
which are coordinated to exhibit life-like emotional states by
precisely controlling and synchronizing their movements in response
to external sensed conditions. The animated character also includes
sound generating circuitry to generate speech sounds as well as
sounds associated with various emotional states, which are
coordinated with a lip sync assembly simulating speech mouth
movement. The drive system utilizes first and second reversible
motors which are able to power and precisely coordinate the lip
sync assembly producing speech mouth movement, with the movable
parts and the smile/emotion assembly to produce life-like
interactions and emotional expressions.
[0005] More particularly, the drive system that powers the movement
of the character body parts, e.g. eye, brow, mouth, ear, plume,
chest, and foot assemblies, includes a mouth assembly on the front
facial area including a flexible molded material having upper and
lower mouth portions and having first and second opposing corners
thereof. The animated character has a lip sync assembly to simulate
speech mouth movement that is independently controlled and
coordinated with the movable parts resulting in coordinated speech
mouth movement with the desired life-like emotional states. A first
mouth mechanism is operable with the mouth assembly for controlling
the first and second corners of the mouth assembly to define
smile/emotion states of the mouth assembly responsive to external
input from sensed conditions of the plurality of sensors. A second
mouth mechanism is further operable with the mouth assembly for
controlling the upper and lower mouth portions to provide lip
synchronism response to the multisyllabic words generated with the
sound generating circuitry.
[0006] The drive system that powers movement of the first and
second mouth mechanisms thus uses each of these mechanisms
independently to simulate life-like responses to sensed conditions,
at least one of which may also be used for causing movement of
another of movable body parts in addition to the first or second
mechanisms. The cams have surfaces that are programmed for very
precise and controlled movements of the body parts in particular
ranges of shaft movements such that generally every point on a
particular cam surface has meaning to the controller in terms of
what type of movement the body part is undergoing and where it
needs to be for its subsequent movement, or for when the body part
is to remain stationary. In this manner, the controller can
coordinate movements of the body parts to provide the animated
character with different states such as sleeping, waking or excited
states. Further, the controller is provided with sound generating
circuitry for generating words that complement the different states
such as snoring in the sleeping state or various exclamations in
the excited state. In addition, the programmed surfaces of the cam
mechanisms are preferably provided on the walls of slots with the
cam mechanisms including followers that ride in the slots.
[0007] The animated character herein is also capable of playing
games with the user in a highly interactive and intelligent seeming
manner. These and other advantages are realized with the described
interactive plaything. The invention advantages may be best
understood from the following detailed description taken in
conjunction with the accompanying flow charts of Appendix A and the
drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
[0008] FIG. 1 is a perspective view of the animated character body
showing the various components thereof;
[0009] FIGS. 2A, 2B and 2C illustrate independently driven Lip Sync
and Smile/emotion actuator linkages to simulate life-like responses
in accordance with an embodiment of the invention;
[0010] FIGS. 3A and 3B illustrate controls activation of reversible
motors to enable the animated character to exhibit speech and
expression of emotional states simultaneously;
[0011] FIGS. 4A and 4B illustrate mechanical cam and gear
positioning and control including the left and right ear cam
members respectively;
[0012] FIG. 5 illustrates the animated character from a front
elevational view to show emotional states of the mouth
assembly;
[0013] FIG. 6 illustrates actuation of the upper and lower mouth
portions of the mouth assembly and including a tongue member;
[0014] FIG. 7 illustrates positioning of plume, brow and eye
assemblies, and the pinion gear of the ear assembly;
[0015] FIGS. 8 and 9 illustrate the gear train transmission portion
of the drive system to show the optical counting assembly;
[0016] FIG. 10 illustrates the positioning of the movable body
parts expressing various emotional states as the first control
shaft rotates in a clockwise direction; and
[0017] FIG. 11 shows a schematic block diagram of the embedded
processor circuitry which includes information processing and voice
speech synthesis functionality in accordance with the present
invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
[0018] An animated character 100 as shown in FIGS. 1-4, has a
number of movable parts, generally designated 102, which are
coordinated to exhibit life-like emotional states by precisely
controlling and synchronizing their movements in response to
external sensed conditions. The animated character 100 also has a
lip sync assembly 140 to simulate speech mouth movement that is
independently controlled and coordinated with the movable parts 102
resulting in seemingly intelligent and highly life-like interaction
with the user. The animated character's lip sync assembly 140 and
movable body parts 102 are controlled and coordinated in response
to predetermined sensory inputs detected by various sensors,
generally designated 104 (where in figures?).
[0019] The sensors 104 signal a controller or processor circuitry
400, described hereinafter, which controls a drive systems 106 for
coordinating speech mouth movement with the desired life-like
emotional states. The drive system 106, utilizes two, low power
reversible electric motors, 108 & 110, which are able to power
and precisely control the lip sync assembly 140 and the movable
parts 102 to produce life-like interactions and emotional
expressions. Further, the control processor circuitry 400 includes
sound generating circuitry to generate speech sounds as well as
sounds associated with various emotional states, such as a laugh,
gasp, sigh, growl & snore, etc. coordinated with speech mouth
movement driven by the lip sync assembly 140 and movement of the
various body parts 102 of the animated character 100. A prior form
of the device was available from the Assignee herein under the name
"Furby" .TM., for which prior issued U.S. Pat. Nos. 6,544,098,
6,537,128, 6,514,117, 6,497,607, 6,149,490 for "Interactive toy" to
Hampton et al. of Applicant's Assignee are hereby incorporated by
reference in their entirety.
[0020] The animated character 100, as seen in FIG. 1, includes a
main body 112 which has a relatively small and compact form and
which contains, in an interior thereof 114, the drive system 106
and all the circuitry and various linkages for generating speech
and sounds, and for activating the lip sync assembly 140 and moving
body parts 102 in a coordinated manner for seemingly intelligent
and life-like interaction with the user. The sensors 104
effectively detect predetermined external conditions and signal the
control processor circuitry 400 of the animated character 100 which
then controls activation of reversible motors 108 & 110, as
seen in FIGS. 3A &3B, to enable the animated character 100 to
exhibit speech and expression of emotional states simultaneously.
The animated character 100 responds to being held, petted, and
tickled. The child can pet the animated character's tummy, rub its
back, rock it, and play with it, e.g., via sensory input buttons.
Whenever the child does these things, the animated character will
speak and make sounds using the speech synthesizer.
[0021] Many of the movable body parts 102 of the animated character
100 herein are provided in a front facial area 117 toward the upper
end 116 of the animated character body 112. The facial area
contains eye and eye lid assemblies 123 and 124 respectively and
mouth assembly 126, with a brow assembly 122 adjacent thereto, as
seen in FIG. 1. The movable body parts 102 of the animated
character 100 herein also include an ear assembly 118, including
left and right movable ear devices 132 and 133 respectively, set
apart from one another and adjacent the brow assembly 122, and a
plume assembly 120 also adjacent the brow assembly 122 and disposed
generally between movable ear devices 132 and 133. The animated
character 100 also includes a movable chest assembly 128 toward a
lower end 115 of the body 112, and a movable foot assembly 130
adjacent thereto.
[0022] A face frame 119 is mounted to the body 112 in an upper
opening and includes a pair of upper eye openings and a lower mouth
opening centered therebelow (openings not shown). An eye assembly
123 is provided including a pair of semi-spherical eyeballs 248 and
250 sized to fit in the eye openings of the frame 119 and pivotally
attached thereto via pivot eye shafts 252 and 254 respectively.
Thus, the pivot shafts 252 and 254 are spaced forwardly and
vertically higher than a first control shaft 136 (discussed further
below) and extends perpendicular thereto.
[0023] Movement of each body part 102 is driven either by motor
108, as seen in FIG. 3A, via a cam operating mechanism, generally
designated 134, or by motor 110, as seen in FIGS. 2C and 3B, via a
shuttle gear 142. In an embodiment, the cam operating mechanism
134, driven by motor 108, rotates a first control shaft 136 fixed
to the interior 114 of the main body 112, and attached thereto are
cam mechanisms, generally designated 138. The first shaft 136 is
rotated when the motor 108 is activated via gear train transmission
278 by meshing of worm gear 280 with peripheral teeth 282a of a
gear cam member 282 fixed on and for rotation with the first
control shaft 136, as seen in FIG. 4B.
[0024] More specifically, cam mechanisms 138 are associated with
the ear assembly 118, the plume assembly 120, the brow assembly
122, the eye lid assembly 124, the eye assembly 125, the
smile/emotion assembly 126, and the chest assembly 128,
coordinating the movement of the aforementioned assemblies for the
expression of various life-like emotional states. Simultaneously,
motor 110, rotates the shuttle gear 142 associated with a second
control shaft 139 in one direction driving movement of the lip sync
assembly 140 simulating speech mouth movement, or rotates the
shuttle gear 142 and associated second control shaft 139 in an
opposite direction driving movement of the foot assembly 130. The
dual motor system of the present embodiment, separates the lip sync
assembly 140 operations from operations of the movable body parts
102 associated with the first control shaft 136 and cam operating
mechanisms 138. Thus any speech generated by animated character 100
can be combined and coordinated with any of the exhibited life-like
emotional states, such that the animated character 100 is able to
coordinate the same speech with any exhibited emotional state, and
likewise, coordinate different speech with the same emotional
state.
[0025] Cam mechanisms 138 each include a disc-shaped cam member and
a follower or actuator linkage thereof. More specifically, and
referencing FIG. 4B with respect to the ear assembly 118, an ear
cam mechanism 144 is provided including left and right ear cam
members 146 and 150 respectively. Left ear cam member 146 (shared
with left smile/emotion cam member as set forth herein) is fixed on
shaft 136, rotates about axis 136a and has an arcuate slot 154
formed on one side thereof, defined by slot wall 156 and including
programmed cam surfaces 156a which engages a left ear cam follower
148. More specifically, a follower pin 158 projecting from left ear
cam follower 148 rides in slot 154 against cam surfaces 156a as
shaft 136 is rotated by motor 108. Left ear cam follower 148 has a
window 160 through which shaft 136 extends and projects
perpendicular to axis 136a thereof. At an upper end of the follower
148 are teeth 162 for engagement with a pinion gear 164 (shown in
FIG. 7) which also engages left movable ear device 132, providing
pivoting movement of left movable ear device 132 upon rotation of
shaft 136 which generates vertical up and down movement of the
follower member 148. Right ear cam member 150 and follower 152 are
fixed on shaft 136 at an end opposite left ear cam member 146 and
follower 148 and engage and pivot right movable ear device 133 in
the same manner though similar mechanisms as described herein for
pivoting left movable ear device 132.
[0026] Left and right movable ear devices 132 and 133 respectively,
are able to bend forward and back in addition to their pivoting
action upon rotation of the control shaft 136. As shown in FIGS. 1
and 3B) left movable ear device 132 includes a flexible strip 166
that has a first surface 168 and second surface 170 and at least
one plate 172 connected to each first and second surface 168 and
170 respectively. A first elongated device 174, made of a flexible
material, intersects at least one plate 172 on the first surface
168 of device 132 and connects to a first lever on an ear actuator.
Upon rotation of the shaft 136, the ear actuator is rotated in a
first direction about a drive axis fixed relative to the animated
character body and connected to the first elongated device 174
engaging at least one plate 172 on the first surface 168 pulling
the elongated device 174 toward the ear actuator thus bending
forward the left movable ear device 132. A second elongated device
176, also made of flexible material, intersects at least one plate
172 on the second surface 170 of device 132 and connects to a
second lever on the ear actuator. When the ear actuator is rotated
in a direction opposite the first direction upon rotation of shaft
136, second elongated device 176 engages at least one plate 172 on
the second surface 170 and is pulled toward the ear actuator thus
bending back the left movable ear device 132.
[0027] Right movable ear device 133 also bends forward and back
upon rotation of control shaft 136 in the same manner through
similar mechanisms as described herein for bending left movable ear
device 132. A prior form of a movable appendage was available from
the Assignee herein for which prior issued U.S. Pat. No. 6,773,327
for "Apparatus for actuating a toy" to Felice et al. issued Aug.
10, 2004 to Applicant's Assignee is hereby incorporated by
reference in its entirety with reference to apparatus for an
appendage including a moveable device within a toy appendage that
is attached to a body of a toy and an actuator connected to the
moveable device. The actuator is configured to rotate the moveable
device about a drive axis that is fixed relative to the body of the
toy. The actuator is configured to rotate at least a first portion
of the moveable device relative to at least a second portion of the
moveable device about a device axis that is fixed relative to the
moveable device.
[0028] The mouth assembly 126 includes a first mouth mechanism
herein the smile/emotion assembly 127 and a second mouth mechanism
herein the lip sync assembly 140 which are operate with the mouth
assembly 126 to independently drive two different types of mouth
movement. The mouth assembly 126 has a mouth member 196 comprised
of substantially identical upper and lower mouth portions 204 and
206, covered with a flexible molded material, in the form of upper
and lower halves of a beak in the present embodiment, as seen in
FIGS. 2A and 2B. Thus the molded flexible elastomeric mouth
includes upper and lower lips having center portions and opposite
side corners where the lips meet. The mouth has attachment points
at the center portions and corners. The first animation mechanism
provides for animating the mouth for simulated talking movement by
moving the attachment points at the center portions. The second
animating mechanism, independent from the first animating
mechanism, provides for moving the attachment points at the corners
to simulate mouth movement corresponding to facial expression.
Accordingly, the first and second animating mechanisms act to
distort the elastomeric mouth to achieve simulated talking movement
and facial expression as the animating mechanisms are operated and
for providing simulated talking movement or facial expression, or a
combination thereof as the animating mechanisms are operated. As
discussed herein, the elastomeric mouth has a neutral undeformed
position, wherein the second animating mechanism moves the
attachment points at the corners to deform the elastomeric mouth
from the neutral undeformed position. The smile/emotion assembly
127 attaches to left and right corners 198 and 200 respectfully, of
the mouth member 196 to achieve a first type of mouth movement. The
smile/emotion assembly 127 moves corners 198 and 200 of the mouth
member 196 in a vertical up direction toward the upper end 116 of
the main body 112 and a vertical down direction toward the lower
end 115 of the main body, such that the mouth member 196 appears to
smile, or frown, or neither, as seen in FIG. 5.
[0029] A smile/emotion cam mechanism 178 drives the vertical up and
down movement of the smile/emotion assembly 127 and includes a left
cam member 146 (shared with left ear cam member as set forth
herein) and a right cam member 179 as shown in FIG. 4B. Left
smile/emotion cam member 146 has an arcuate slot formed on a side
opposite arcuate slot 154, including programmed cam surfaces which
engage a pin 186 of a left smile/emotion follower 184 as shaft 136
is rotated by motor 108. Smile/emotion follower 184 is L-shaped and
includes a first arm 188 with a window 190 through which shaft 136
extends, and a second arm 192 which projects forward substantially
perpendicular to axis 136a thereof.
[0030] FIGS. 2A and 2B illustrate independently driven Lip Sync and
Smile/emotion actuator linkages to simulate life-like responses. As
shown, a distal end 193 of projected second arm 192 connects to a
linkage 194 attached to the left corner 198 of the mouth member 196
moving the left corner 198 in a vertical up and down fashion upon
rotation of shaft 136. Likewise, the right corner 200 of mouth
member 196 is moved vertically up and down in synchronization with
the left corner 198 in the same manner and through similar
mechanisms as described for the left corner 198 of mouth member
196. Thus, the mouth member 196 is able to change expressions
between a smile, frown, and neutral states upon rotation of first
control shaft 136.
[0031] The second mouth mechanism herein the lip sync assembly 140
attaches to upper and lower mouth portions 204 and 206
respectively, of mouth member 196 to achieve a second type of mouth
movement. In particular, FIG. 2C illustrates motor 110 driving
movement of the foot assembly 130 through activation of the shuttle
gear 142 and rotation of the second control shaft 139 in a
direction opposite the rotation which drives the lip sync assembly
140 as well as discussed further below. Motor 110 activates shuttle
gear 142 and associated second control shaft 139 driving the lip
sync assembly 140 to open and close the mouth member 196, such that
the upper and lower mouth portions 204 and 206 respectively, move
away from and toward each other in order to simulate speech mouth
movement which is coordinated with speech sounds, by the control
processor circuitry 400. The mouth portions 204 and 206 include
upper and lower pairs of oppositely facing hook-shaped coupling
portions 216 to allow an associated actuator 218, to cause opening
and closing movement of the mouth portions 204 and 206 upon
rotation of second control shaft 139, as seen in FIG. 6 and
discussed further below. Thus, to achieve the entire range of mouth
movement of the mouth member 196 both motors 108 and 110 activate
at the same time, simultaneously achieving two types of mouth
movement. As a result, the mouth member 192, coordinated by the
control processor circuitry 400 according to predetermined
conditions, seemingly appears to speak with coordinated speech
mouth movement synchronized with the appropriate mouth expression
such as a smile, frown or neutral state, providing seemingly
intelligent interaction with the user.
[0032] The upper and lower mouth portions 204 and 206 are pivotally
mounted on shaft 208 by rear semi-circular boss portions thereof
spaced on either side of the mouth portions 204 and 206 so as to
provide space for a tongue member 210 therebetween, as seen in FIG.
6. The tongue member 210 includes an intermediate annular bearing
portion 212 through which the pivot shaft 208 extends and has a
rearwardly extending switch actuator portion 214 so that depressing
the tongue 210 pivots the portion 214 for actuating the tongue
sensor assembly. Referring to FIG. 6, the tongue sensor assembly is
illustrated. As previously discussed, the tongue member 210 that
has an actuator portion 214 that projects rearwardly and pivots.
The switch actuator portion 214 extends further in the rearward
direction than the forward tongue portion 210 and is designed so
that normally the switch actuator portion 214 is in its lower
position and the tongue portion 210 is disposed for activation to
simulate feeding the animated character 100.
[0033] The eye lid assembly 124, as shown in FIG. 7 includes upper
and lower lid portions 220 and 222 respectively, pivotally mounted
on shaft 224 interconnecting the pair of eyelids. Assembly 124 has
associated eye lid cam mechanism 226 which includes a cam member
228 fixed on shaft 136 having an arcuate slot 230 defined by slot
walls 232. An eyelid cam follower 234 includes a pin 236 projecting
therefrom for engagement with cam surfaces 232a on slot walls 232.
The cam follower 234 has a window 238 through which shaft 136
extends and projects forwardly from the shaft 136 substantially
perpendicular to axis 136a thereof. Toward the forward end of the
cam follower 234 are a pair of vertically spaced flexible arcuate
arm portions 240 having small pairs of pivot pin portions 242
extending oppositely and laterally from forked distal ends therof
spaced forwardly of the shaft 136 and extending parallel thereto,
for engagement with upper and lower lid portions 220 and 222.
[0034] Accordingly, rotation of shaft 136 rotates cam member 228
with pin 236 riding in slot 230 thereof to cause the follower 234
to translate in a fore and aft direction while engaged with upper
and lower lid portions 220 and 222. The follower shifting forwardly
causes upper and lower lid portions 220 and 222 to move away from
one another and seemingly close the eyes of animated character 100,
and the follower 234 shifting rearwardly causes the lid portions
220 and 222 to move toward each other seemingly opening the eyes of
animated character 100.
[0035] The eye assembly 123, as also seen in FIG. 7, has left and
right eye balls 248 and 250 respectively, pivotally mounted on left
and right eye shafts 252 and 254 respectively, and associated with
barrel cam 146 mounted on first control shaft 136. Shafts 252 and
254 fit into follower 255 which has shaft 257 that engages barrel
cam 146, for back and forth sideways movement of left and right eye
balls 248 and 250 respectively, as first control shaft 136 is
rotated.
[0036] Further expressive features of the animated character 100
which are driven for movement by rotation of the first control
shaft 136 include the plume assembly 120 and the brow assembly 122.
The plume assembly 120 and the brow assembly 122 are pivotally
attached to a brow bracket 278 fixed to the upper end 116 of the
body 112. The plume assembly 120 as seen in FIGS. 1 and 7 has a
plume member 256 which pivots about a shaft 258 and is associated
with plume cam mechanism 260 including plume cam member 262. Cam
member 262 is fixed on shaft 136 and includes an arcuate slot 264
defined by slot walls 266. Plume follower 268 includes a pin 270
projecting therefrom for engagement with slot surfaces 264a on slot
walls 264. Plume follower 268 has a window 272 through which shaft
136 extends and projects upwardly from the shaft 136 substantially
perpendicular to the axis 136a thereof. At an upper end of follower
268 is a offset bent hook portion 274 having a shaft 276 for
engagement with an end of plume member 256, such that rotation of
first control shaft 136 translates into a vertical up and down
movement of follower 268 causing attached plume member 256 to pivot
in a back and forth fashion.
[0037] The brow assembly 122 as seen in FIGS. 1 and 7 has an
eyebrow member 280 pivotally attached to brow bracket 278 and is
associated with brow cam mechanism 283 including brow cam member
284. Cam member 284 is fixed on shaft 136 and includes an arcuate
slot 286 defined by slot walls 288. Brow follower 290 includes a
pin 292 projecting therefrom for engagement with slot surfaces 288a
on slot walls 288. Brow follower 290 has a window 294 through which
shaft 136 extends and projects upwardly from the shaft 136
substantially perpendicular to the axis 136a thereof. At an upper
end of follower 290 is a hook portion 296 having a shaft 298 for
engagement with eyebrow member 280, such that rotation of first
shaft 136 translates into a vertical up and down movement of
follower 290 causing attached eyebrow member 280 to pivot in a
vertical up and down fashion.
[0038] A chest assembly 128, as seen in FIG. 1, includes an
apertured disc actuator 320 having an upper arm portion 322
attached to a speaker/speaker grill 324 fixed to a chest portion
328 movably fixed to a bracket (not shown) which in turn is rigidly
mounted to the animated character body 112. The disc actuator 320
is made of a plastic material and the arm portion 322 thereof
spaces the disc actuator 320 forwardly of the speaker grill 324.
The speaker grill 324 is preferably of a plastic material with arm
portions thereof spaced forwardly allowing the disc actuator 320 to
be flexibly and resiliently shifted or pushed back and forth. The
chest assembly 128 is associated with cam mechanism 330 including
cam member 332 fixed to first control shaft 136, and follower 334
having a window 336 an end of the follower through which shaft 136
extends. Follower 334 further includes a hook portion 338 at an end
opposite the window 336 which couples to the chest portion 328,
such that the chest portion 328 moves in and out with respect to
the character body 112 as the first control shaft 136 is
rotated.
[0039] The animated character 100 also includes a foot assembly 130
including a pair of feet 300, as seen in FIG. 2C, that are movable
relative to the animated character body 112 which allows the
animated character to rock back and forth and, if done
repetitively, gives the appearance that the animated character 100
is dancing. Motor 110 drives movement of the foot-assembly 130
through activation of the shuttle gear 142 and rotation of the
second control shaft 139 in a direction opposite the rotation which
drives the lip sync assembly 140. The foot assembly 130 also
includes a battery compartment (not shown). As the splined
connection between the shaft and pinion portions allows for
relative motion such as when a child grabs the feet 300 during
movement thereof. It is possible for a particular shaft to become
out of alignment, however the splined connection will allow
subsequent movement of feet 300 in alignment with each other absent
a braking force applied thereto.
[0040] The control processor circuitry 400 is able to precisely
control and determine the position of the first control shaft 136
when the motor 108 is activated; however, it is also desirable to
avoid the expense and moving parts of utilizing a closed loop servo
mechanism for providing the necessary feedback. The drive system
106 of an embodiment herein instead includes an optical counting
assembly 302 which counts intervals of the rotation of a slotted
gear wheel 304 in gear train transmission of the drive system 106,
as seen in FIGS. 8 and 9. The gear wheel 304 is mounted at the
lower end of a common vertical shaft 306 having worm gear 280
formed at its upper end, and is driven for rotation by the upper
portion 308a of intermediate compound gear 308 which, in turn, is
driven for rotation by gear 108a on the output shaft of the motor
108 which drives the larger lower portion 308b of compound gear 308
for rotation.
[0041] For programming of the cam surfaces, modeling of the
animated character's different states is based on puppeteering
actions to achieve positions of body parts for generating animated
character movements. From the neutral position as a starting point,
the cam is designed to actuate the leaf spring switch to zero out
the count for the motor on a regular basis. In this manner, the
position of the shaft will not become out of synchronization, the
count of the processor thus being zeroed to provide for recurrent
and regular calibration of the position of the shaft. From the
neutral position, rotation/direction is determined to cause certain
coordinated movements of various body parts with precise movements
thereof. In this regard, the cams are provided with cam surfaces
that have active regions and inactive regions so that in the active
regions, the part associated with the particular cam is undergoing
movement, and in the inactive region the part is stationary.
[0042] As shown in FIG. 9, an integral IR transmitter element 312
facilitates optical servo control. The optical servo control
circuitry employing the slotted wheel for generating an infrared
light source is used. The phototransistor is used as an infrared
photo detector for generating a light pulse count signal. The IR
receiver element 314 is rigidly mounted to the frame of the
box-shaped housing portion 314 thereof integrally formed with frame
wall. In this manner, the optical counting assembly 302 herein is
improved over prior feedback mechanisms that require moving parts
or impart frictional resistance to motor operation, as the assembly
302 utilizes elements 312 and 314 that are fixed in the body
interior using the optical servo control circuitry for controlling
the operation of the motor. By incrementally counting slots 310 in
the wheel 304 as the wheel 304 is rotated when the motor 108 is
activated as the slots 310 pass between an IR transmitter 312 and
an IR receiver 314 on either side of the gear wheel 304, the
control processor circuitry 400 can receive accurate information
regarding the position of the control shaft 136 for precisely
controlling the movements of the body parts 102. Preferably six
slots 310 are equally spaced at 60 degree intervals about the wheel
304. In addition, an initialization switch assembly is provided
mechanically affixed to a frame for the cam operating mechanism 134
to zero out the count in the control circuitry 400 on a regular
basis when the switch assembly is actuated. In this regard, the
optical counter assembly counts intervals of the revolutions of an
apertured gear wheel with the use of standard types of IR
transmitters and receivers on either side thereof that are small
components fixed in housings rigidly mounted inside the animated
character 100.
[0043] Contacts of a leaf spring switch are mounted between the
disc 320 and the speaker grill 324 and affixed thereto. Thus,
depressing the disc 320 as by pushing or rubbing on the hide of the
character thereover causes engagement of the contact strips which
signals the processor circuitry 400. Actuating a front sensor
assembly can simulate tickling of the animated character 100 in its
belly region.
[0044] As previously stated, cam surfaces of the cam mechanisms 138
herein are provided with precise predetermined shapes which are
coordinated with the programming of the processor circuitry 400 so
that at every point of the cam surfaces, the processor circuitry
400 can be used to determine the position of the moving body parts
102 associated therewith. In this manner, the animated character
100 can be provided with a number of different expressions to
simulate different predetermined physical and emotional states. For
instance, changes in emotional expressions of animated character
100 upon rotation of first control shaft 136 are provided as shown
in FIG. 10, and are described herein with reference to life-like
expressions.
[0045] A neutral position is provided at a zero degree position of
the control shaft 136 wherein the eyes lids 220 and 222 are open,
the ear devices 132 and 133 are up at a forty-five degree angle,
the chest is in, the plume 256 is down, and the mouth corners 198
and 200 and brow are in neutral positions neither up nor down. A
happy expression is provided at a thirty-six slot count clockwise
rotation of the control shaft 136 wherein the eye lids 220 and 222
are open the ear devices 132 and 133 are pivoted up to a
twenty-five degree angle, the chest is in, the plume 256 is up, the
mouth corners 198 and 200 are up in a smile, and the brow is up. A
surprised expression is provided at a seventy-two slot count
clockwise rotation of the control shaft 136 wherein the eyes lids
220 and 222 are wide open, the ear devices 132 and 133 remain up at
a twenty-five degree angle, the chest is in, the plume 256 is up,
the mouth corners 198 and 200 are in a neutral position neither up
nor down, and the brow remains up. A sad expression is provided at
a one hundred eight slot count rotation of the control shaft 136
wherein the eyes lids 220 and 222 lower to open, the ear devices
132 and 133 are down at a ninety degree angle, the chest is in, the
plume 256 is down, the mouth corners 198 and 200 are down in a
frown, and the brow remains up. An angry expression is provided at
a one hundred forty-four slot count rotation of the control shaft
136 wherein the eyelids 220 and 222 are narrow, the ear devices 132
and 133 are down at a ninety degree angle, the chest is about
thirty percent out, the plume 256 is up, the mouth corners 198 and
200 are down in a frown, and the brow is down. A sleep expression
is provided at a one hundred eighty slot count rotation of the
control shaft 136 wherein the eye lids 220 and 222 are wide open,
the ear devices 132 and 133 are up at a forty-five degree angle,
the chest is about fifty percent out (i.e., the chest is fully out
at a one hundred sixty-eight rotation of control shaft 136) the
plume 256 is down, the mouth corners 198 and 200 and brow are in
neutral positions neither up nor down. Total slot count for 1
revolution of cam system is 206 counts (0 to 205).
[0046] The embedded microprocessor circuit for the animated
character 100 is identified in FIG. 11 as the processor circuitry
400. The schematic block diagram of FIG. 11 shows the embedded
processor circuitry in accordance with the present embodiment in
which an information processor 402 CMOS integrated circuit
providing the RISC processor and read only memory (ROM). In the
present described embodiment, the information processor 402 is
provided by Sensory, Inc. (Santa Clara, Calif.) as an RSC-4.times.
Speech Recognition and Synthesis Microcontroller. The information
processor 402 provides various functional controls facilitated with
on board static random access memory (SRAM), a timer/counter, input
and output ports (I/O) as well as an audio current mode Pulse Width
Modulator (PWM). The RSC-4.times. facilitates speech processing
with advanced audio features based on an 8-bit microcontroller. The
RSC-4.times. integrates speech-optimized digital and analog
processing blocks into a single chip solution capable of accurate
speech recognition as well as high quality, low data-rate
compressed speech. The RSC-4.times. information processor 402 also
provides on-chip integration of a microphone preamplifier, twin-DMA
units, vector accelerator, hardware multiplier, 3 timers, and 4.8
Kbytes of RAM with multiple ROM options.
[0047] The circuitry employs wireless transmission 404. The
input/output (I/O) port of the information processor 402 is
capacitively coupled to the data lines from the port of the
information processor 402. Capacitive coupling methods are employed
to initiate simple wireless communication between two bodies by
placing them within a few inches apart. For example, communication
is facilitated through the use of two small plates (406, 408) about
0.75 square inches in size and mounted side by side about 1/2 inch
apart horizontally. A receiver amplifier 410 is provided as a
receiver module preamplifier, herein Waitrony Module No. WPI-T2100
used for amplification of the capacitively coupled electrical
carrier signals. Accordingly, an emitter plate 406 is used as a
transmitter, with the other plate 408 used as a receiver. When
located near a matching pair of plates, communication can be
established by initiating a capacitive coupling between the aligned
plates. Transmit and receive protocol is assigned on the fly (i.e.
who talks first). Such capacitive coupling techniques know in the
art include, e.g., expired U.S. Pat. No. 4,242,666 to Reschovsky et
al. for Range selectable contactless data acquisition system for
rotating machinery, issued Dec. 30, 1980 which discloses a
multichannel data acquisition system uses radio telemetry for data
transfer by providing a capacitive coupling link between rotating
and stationary members with a pulse-code modulated signal
containing the measured information for transmission through the
capacitive coupling link.
[0048] As described, the wireless transmission 404 provides
circuitry under control of the speech processing incorporated with
the information processor 402 which serves to receive, transmit and
process speech and other information. The wireless receive circuit
block 408 is coupled to the information processor 402 for receiving
wireless signals from the transmit circuitry 404 of another
animated character device as described herein.
[0049] The information processor 402 is provided for speech and
wireless communications capabilities. The RSC-4.times. speech
recognition and synthesis are supported with its Sensory Speech.TM.
7, providing advanced algorithms having substantial on-chip speech
recognition algorithms accuracy for speaker-dependent recognition
and as well as for speaker-independent recognition. Additionally
audible speech synthesis is also provided.
[0050] The described information processor 402 of FIG. 11 also
facilitates speech synthesis processing that using integrated
microprocessor control facilitating music and sound effects as well
as speech and system control functions. Accordingly, the mouth
mechanism is operable with the mouth assembly for controlling the
upper and lower mouth portions to provide lip synchronism with a
Lip Sync switch 422 responsive to the multisyllabic words generated
with the sound generating circuitry. A motor calibration switch is
provided as switch 424.
[0051] The sound detection and voice recognition are provided via
microphone (Mic In 1 and Mic In 2) inputs to allow the information
processor 402 to receive audible information as sensory inputs from
the child which is interacting with the animated character 100.
Optical control circuitry 412 is used with the motor control
circuitry 414 as discussed herein to provide an electronic motor
control interface for controlling the position and direction of the
electric motors. An H-bridge circuit for operating the motor in
either forward or reverse directions. A power control block 416 is
used to voltage regulate the battery power to the processor CPU,
nonvolatile memory (EEPROM) and other functional components of the
processor circuitry 400.
[0052] Various other sensory inputs 418 provide a plurality of
sensory inputs coupled to the information processor 402 allowing
the animated character 100 to be responsive to its environment and
sensory signals from the child. A tilt/invert sensor 420 is
provided to facilitate single pull double throw switching with a
captured conductive metal ball allowing the unswitched CPU voltage
to be provided at either of two input ports indicating tilt and
inversion of the plaything respectively. The sensory inputs 418 of
the described embodiment are provided as push button switches,
although pressure transducers and the like may also be provided for
sensory input. The sensory inputs 418 are provided as a momentary
push button controlled, e.g., a mouth sensor of the tongue of the
plaything is acquired with the audio ADC provided as a
switch-select allowing the processor 402 to receive, e.g. the feed
input with other I/O inputs. Additional momentary switches are
provided for the front and back sensors of the plaything
respectively as push button sensory inputs 418.
[0053] The motor interface provided between the information
processor 402 and the motor control block 414 controls the actuator
linkages with the information processor 402. As described, the
plurality of sensory inputs, e.g., switch sensory inputs 418, and
the audio (Mic In 1 and Mic In 2), and wireless wireless
transmission 404, are coupled to the information processor 402 for
receiving corresponding sensory signals. Computer programs
referenced below in connection with the program flow diagram for
operating the embedded processor design embodiment of Appendix A
facilitates processing of the sensory signals for a plurality of
operational modes provided by the computer program with respect to
the actuator linkage operation and corresponding sensory signal
processing for controlling the at least one actuator linkage to
generate voice interaction with the child with the plurality of
movable members corresponding to each of the operational modes of
the plaything which provides interactive artificial intelligence
for the animated character 100. As discussed, the animated
character includes a doll-plush toy or the like having movable body
parts with one or more of the body parts of the doll being
controlled by the plurality of movable members for interacting with
the child in a life-like manner.
[0054] The of the software program flow diagrams of Appendix_A, see
programs P.1-P.48 below; provide for the operating of the embedded
processor circuitry described above. The program flow diagram uses
the embedded processor circuitry 400 for initialization,
diagnostics, and calibration routines are executed prior to the
normal run mode of the processor circuitry 400. As provided in
connection with the general random table (P.48), pseudo random
values are introduced. For example, when push button sensory inputs
418 are activated, inputs e.g., 0, 1, 2 are provided instead of
always adding 1 when an input is triggered once. This adds a random
increment when sensory inputs are being triggered. The mean value
of the pseudo output may be set to unity (1) to have randomness
factor of trigger for fluency calculation. There are three
conditions for fluency counting to increase: (a) Each time that the
animated character HEARS and UNDERSTANDS a word, that word's
Fluency increases (+5 each time for "each" word). The fluency
parameter will be updated in each VR response; (b) each time that
the animated character SPEAKS a word, increase that word's Fluency
by +1--the fluency parameter will be updated in each Phrase in
ENGLISH response; or (c) every hour that Furby has some interaction
with the player, ALL words increase by +1, e.g, a hourly check of
any key/VR within the previous hour. The flow charts also address
the now fluency mechanism references to the fluency increasing (OOV
response and `I don't understand` response). Various artificial
intelligence (AI) and sensor training functions are provided in
which training between the random and sequential behavior
modification of the animated character, allowing the child to
provide reinforcement of desirable activities and responses. In
connection with the AI functions, appropriate responses are
provided in response to particular activities or conditions, e.g.,
bored, hungry, sick, sleep. Such predefined conditions have
programmed responses which are undertaken by the animated character
at appropriate times in its operative states. Accordingly,
summarizing the wide range of life-like functions and activities
the compact and cost-effective toy 100 herein can perform to
entertain and provide intelligent seeming interaction with a child,
the following is a description of the various abilities the
preferred animated character 100 has and some of the specifics in
terms of how these functions can be implemented, in subroutines or
programs P.1-P.48 (Appendix A) as follows. [0055] P.1 Game Play
Flowchart [0056] P.2 Power Up Sequence [0057] P.3 Game Play Loop
[0058] P.4 Idle Mode [0059] P.5 OOV Response [0060] P.6
Acknowledgement Response [0061] P.7 Time Out Response [0062] P.8
Mischief Mode [0063] P.9 Initiate Response [0064] P.10 Main Input
Mode [0065] P.11 Main Input Mode (Furbish) [0066] P.12 Story Mode
[0067] P.13 Song Mode [0068] P.14 Joke Mode Entry [0069] P.15 Joke
Player Response [0070] P.16 Joke Correct Respond Sequence [0071]
P.17 End Joke Sequence [0072] P.18 Hungry Mode [0073] P.19 Play
Mode Entry [0074] P.20 Furby Select Action (Play Mode) [0075] P.21
Game Over Handle (Play Mode) [0076] P.22 Player Tilted Furby (Play
Mode) [0077] P.23 Player Too Slow (Play Mode) [0078] P.24 Dance
Mode [0079] P.25 Love Mode [0080] P.26 "How are you" mode [0081]
P.27 Sleep Response [0082] P.28 Sleep Mode [0083] P.29 Deep Sleep
Mode [0084] P.30 Sensor Check [0085] P.31 Tilt Mode [0086] P.32
Petting Mode [0087] P.33 Tickle Mode [0088] P.34 Feeding Mode
[0089] P.35 Question Mode [0090] P.36 Yes No Response [0091] P.37
Don't Understand Handle [0092] P.38 Furby to Furby Mode Entry
[0093] P.39 Furby to Furby VR Check [0094] P.40 Furby to Furby VR
Check pt 1/2 [0095] P.41 Furby to Furby VR Check pt 3/4 [0096] P.42
Furby to Furby VR Check pt 5 [0097] P.43 Play Phrase Handling
[0098] P.44 Fluency Handle During VR Check [0099] P.45 Hourly
Fluency Handle [0100] P.46 Try Me Mode (Part 1) [0101] P.47 Try Me
Mode (Part 2) [0102] P.48 General Random Table
[0103] While there have been illustrated and described particular
embodiments of the present invention, it will be appreciated that
numerous changes and modifications will occur to those skilled in
the art, and it is intended in the appended claims to cover all
those changes and modifications which fall within the true spirit
and scope of the present invention.
* * * * *