U.S. patent application number 13/236516 was filed with the patent office on 2012-01-12 for method of changing processor-generated individual characterizations presented on multiple interacting processor-controlled objects.
This patent application is currently assigned to SMALTI TECHNOLOGY LIMITED. Invention is credited to Martin OWEN.
Application Number | 20120007870 13/236516 |
Document ID | / |
Family ID | 34566497 |
Filed Date | 2012-01-12 |
United States Patent
Application |
20120007870 |
Kind Code |
A1 |
OWEN; Martin |
January 12, 2012 |
METHOD OF CHANGING PROCESSOR-GENERATED INDIVIDUAL CHARACTERIZATIONS
PRESENTED ON MULTIPLE INTERACTING PROCESSOR-CONTROLLED OBJECTS
Abstract
Processor-controlled objects, such as inter-communicating
processor-controlled blocks, are adapted to present changeable
individual characterizations to a user. A user manipulating the
objects can cause, over time, a designated object to inherit
characterizations and properties from other interacting objects to
permit scalability in a set of such objects. The communication of
individual characterization between interacting objects allows
generation of sensory responses (in a response generator of a
specific object or otherwise in a response generator associated
with at least one other similar objects) based on proximity,
relative position and the individual characterization presented on
and by those interacting objects at the time of interaction. In
this way, a set of objects has vastly extended interactive
capabilities since each object is capable of dynamically taking on
different characterizations arising from a meaningful combination
of properties from different conjoined objects.
Inventors: |
OWEN; Martin; (Bristol,
GB) |
Assignee: |
SMALTI TECHNOLOGY LIMITED
HOLYWELL
GB
|
Family ID: |
34566497 |
Appl. No.: |
13/236516 |
Filed: |
September 19, 2011 |
Related U.S. Patent Documents
|
|
|
|
|
|
Application
Number |
Filing Date |
Patent Number |
|
|
11142955 |
Jun 2, 2005 |
8057233 |
|
|
13236516 |
|
|
|
|
Current U.S.
Class: |
345/473 ;
345/156 |
Current CPC
Class: |
A63F 9/28 20130101; G09B
1/36 20130101; G09B 5/06 20130101; A63F 3/0423 20130101; G09B 1/34
20130101 |
Class at
Publication: |
345/473 ;
345/156 |
International
Class: |
G06T 13/00 20110101
G06T013/00; G09G 5/00 20060101 G09G005/00 |
Foreign Application Data
Date |
Code |
Application Number |
Mar 24, 2005 |
GB |
0506159.3 |
Claims
1. A method of controllably presenting changeable,
processor-generated individual characterizations presented on each
of a plurality of objects that selectively interact, the method
comprising: generating and displaying under powered
processor-control first visual display material on a first movable
object, the first visual display material having a first changeable
individual characterization having a first property; sensing
proximity and relative position of second visual display material
generated and displayed under powered processor-control on a second
movable object separate to the first movable object, the second
object brought into processor-resolvable interacting proximity with
the first moveable object by manipulation of one or more of the
first movable object and the second movable object, the second
visual display material having a second changeable individual
characterization independent to the first individual
characterization, the second changeable individual characterization
having a second property; in response to processor-resolved
interaction between said first and second changeable individual
characterizations arising from sensed proximity and relative
position of the first and second objects, generating a
user-perceivable sensory response from a response generator,
wherein the user-perceivable sensory response is dependent upon
said sensed relative positions of the first visual display material
to the second visual display material and is indicative of a
contextual relationship that arises between said first property of
the first changeable individual characterizations and the second
property of the second changeable individual characterizations; and
under processor-control, selectively and autonomously changing with
time the individual characterizations on at least one of the first
and second objects such that the first property and the second
property change to allow different processor-resolvable
interactions to take place between the first and second objects,
which different processor-resolvable interactions give rise to new
user-perceivable sensory responses.
2. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
1, wherein the change in property and individual characterization
presented by one of the plurality of objects is dependent upon the
processor-resolved interaction and contextual relationship between
the first changeable individual characterization presented on the
first object and the second changeable individual characterization
presented on the second object.
3. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
1, wherein the change to the first property and/or the second
property includes an auditory change.
4. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
1, wherein the change to the first property and/or the second
property includes a visual change.
5. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
1, wherein the user-perceivable sensory response is a phonetic
sound produced by combined interaction of said first and second
changeable individual characterizations and their respective first
and second meanings.
6. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
1, wherein generating the user-perceivable sensory response further
generates: an audible representation of the first changeable
individual characterization; and an audible representation of the
second changeable individual characterization.
7. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
1, the method further comprising: under processor-control,
selectively changing with time said changeable individual
characterizations displayed on said plurality of objects based
upon: i) nearby detection and relative position of objects; and ii)
individual characterizations presented by interacting objects at
the time of their detection and interaction.
8. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
1, wherein the user-perceivable sensory response is generated for
only meaningful non-random interactions between said first and
second changeable individual characterizations.
9. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
1, wherein the user-perceivable sensory response generated from the
response generator includes an audible response, a visual response
or a combination of an audible and visual response.
10. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
8, wherein the user-perceivable sensory response includes a related
visual animation.
11. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
1, wherein generating the user-perceivable sensory response
includes generating at least one of: a phoneme; a letter, word,
phrase or sentence; speech relating to mathematical properties of a
visually presented number; and an audio musical response
corresponding to a musical symbol.
12. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
1, further comprising: arranging said first and second objects in a
line so that said first visual display material and said second
visual display material reads in a meaningful manner along said
line; and locating a third object adjacent to one side of said line
of objects, the third object having a third changeable individual
characterization presented as third visual display material, the
step of adjacently locating causing the third visual display
material presented on said third object to change to take on the
combination of said first visual display material and said second
visual display material that reads in the meaningful manner along
said line.
13. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
1, further comprising: locating a third object proximate to the
first and second objects, the step of locating causing positioning
of the third object in one of a multiplicity of different positions
below the first and second objects, the third object having a third
changeable individual characterization; determining relative
positions between the first, second and third objects; in the event
that said relative position between the first and second objects
results in production of a non-random meaningful, interaction
between at least the first and second individual characterizations
that produces one of a phoneme, a word a mathematical property or a
musical response, effecting a processor-controlled change to the
third changeable individual characterization presented on the third
object by generating a new changeable individual characterization
that is presented on the third object and which new changeable
individual characterization reflects: i) the meaningful interaction
taking place between the first and second individual
characterizations; and ii) the relative position between the first,
second and third objects.
14. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
12, wherein each object is arranged to generate and display a
changeable individual characterization in the form of: at least one
letter; a word; a number; a mathematical symbol; or a musical
symbol.
15. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
1, further comprising: selectively programming each of the
plurality of objects to present over time a multiplicity of
properties under processor-control.
16. The method of controllably presenting changeable,
processor-generated individual characterizations according to claim
1, further comprising: based on relative position between the first
and second movable objects, having the first movable object inherit
properties associated with at least the second changeable
individual characterization of the second movable object.
17. A method of controllably presenting changeable,
processor-generated individual characterizations presented on each
of a plurality of objects that selectively interact, the method
comprising: generating and presenting under powered
processor-control first visual display material on a
processor-controlled object, the first visual display material
having a first changeable individual characterization with a first
property; sensing proximity and relative position of second visual
display material generated and presented under processor-control on
a second object movable independently of said processor-controlled
object, the second object brought into processor-resolvable
interacting proximity with said processor-controlled object by
manipulation of one or more of said processor-controlled object and
the second object, the second visual display material having a
processor-controlled second changeable individual characterization
independent of the first individual characterization, the second
changeable individual characterization having a second property;
sensing proximity and relative position of third visual display
material generated and presented under processor-control on a third
object movable independently of said processor-controlled object,
the third object brought into processor-resolvable interacting
proximity with said processor-controlled object by its relative
manipulation with respect to the second object, the third visual
display material having a processor-controlled third changeable
individual characterization independent of the first and second
individual characterizations, the third changeable individual
characterization having a third property, the second and third
properties of the second and third objects interacting with one
another when located side-by-side in a line to produce a
combination; under processor control changing the first changeable
individual characterization with the first property to a fourth
changeable individual characterization with a fourth property such
that the fourth changeable individual characterization is presented
on said processor-controlled object, the fourth property different
to the first property, the fourth individual characterization being
the combination inherited from the second and third individual
characterizations provided that the in-line side-by-side
combination has a contextual relevance that is not random; and
determining new processor-resolvable interactions involving said
processor-controlled object now having the fourth changeable
individual characterization and generating and outputting audible
and/or visual user-perceivable sensory responses from a response
generator to reflect contextually relevant processor-resolvable
interactions that reflect and/or involve the fourth changeable
individual characterization.
18. The method of controllably presenting changeable,
processor-generated individual characterizations of claim 17,
further comprising: assembling a first line of objects containing
only the processor-controlled object and assembling a second line
of objects contains the second and third objects but not the
processor-controlled object; determining that the first and second
lines are interacting; and determining whether the second and third
changeable individual characterizations combine to provide a
contextually relevant combination that is not random, subject to
the second and third changeable individual characterizations
combining to provide a contextually relevant combination that is
not random, causing the first changeable individual
characterization to change to the fourth changeable individual
characterization by inheriting the contextually relevant
combination based on an edge justification determined by the
processor-controlled object in the first line relative to the
contextually relevant combination in the second line.
19. The method of controllably presenting changeable,
processor-generated individual characterizations of claim 17,
wherein the contextually relevant combination inherited as the
fourth changeable individual characterization includes at least the
second and third changeable individual characterizations.
20. The method of controllably presenting changeable,
processor-generated individual characterizations of claim 17,
further comprising: assembling a first line of objects containing
only the processor-controlled object and assembling a second line
of objects containing at least the second and third objects but not
the processor-controlled object; determining that the first and
second lines are interacting; and determining whether at least the
second and third changeable individual characterizations combine to
provide a contextually relevant combination that is not random,
subject to the second and third changeable individual
characterizations combining to provide a contextually relevant
combination that is not random, causing the first changeable
individual characterization to the change to the fourth changeable
individual characterization by inheriting the contextually relevant
combination based on an edge justification determined by the
processor-controlled object in the first line relative to the
contextually relevant combination in the second line.
21. A method of controllably presenting changeable,
processor-generated individual characterizations presented on each
of a plurality of objects that selectively interact, the method
comprising: generating and presenting under powered
processor-control first visual display material on a
processor-controlled object, the first visual display material
having a first changeable individual characterization with a first
property; sensing proximity and relative position of second visual
display material generated and presented under processor-control on
a second object movable independently of said processor-controlled
object, the second object brought into processor-resolvable
interacting proximity with said processor-controlled object by
manipulation of one or more of said processor-controlled object and
the second object, the second visual display material having a
processor-controlled second changeable individual characterization
independent of the first individual characterization, the second
changeable individual characterization having a second property;
sensing proximity and relative position of third visual display
material generated and presented under processor-control on a third
object movable independently of said processor-controlled object,
the third object brought into processor-resolvable interacting
proximity with said processor-controlled object by its relative
manipulation with respect to the second object, the third visual
display material having a processor-controlled third changeable
individual characterization independent of the first and second
individual characterizations, the third changeable individual
characterization having a third property, the second and third
properties of the second and third objects interacting with one
another when located side-by-side in a line to produce a
combination; sensing proximity and relative position of fourth
visual display material generated and presented under
processor-control on a fourth object movable independently of said
processor-controlled object, the fourth object brought into
processor-resolvable interacting proximity with said
processor-controlled object by its relative manipulation with
respect to the second and third object, the fourth visual display
material having a processor-controlled fourth changeable individual
characterization independent of the first, second and third
individual characterizations, the fourth changeable individual
characterization having a fourth property, wherein the second,
third and fourth properties of the second, third and fourth objects
interacting with one another when located side-by-side in a line to
produce a combination; under processor control changing the first
changeable individual characterization with the first property to a
fifth changeable individual characterization with a fifth property
such that the fifth changeable individual characterization is
presented on said processor-controlled object, the fifth property
different to the first property, the fifth individual
characterization being the combination inherited from the second,
third and fourth individual characterizations provided that the
in-line side-by-side combination has a contextual relevance that is
not random; and determining new processor-resolvable interactions
involving said processor-controlled object now having the fifth
changeable individual characterization and generating and
outputting audible and/or visual user-perceivable sensory responses
from a response generator to reflect contextually relevant
processor-resolvable interactions that reflect and/or involve the
fifth changeable individual characterization.
22. The method of controllably presenting changeable,
processor-generated individual characterizations of claim 21,
further comprising: assembling a first line of objects containing
only the processor-controlled object and assembling a second line
of objects containing the second, third and fourth objects but not
the processor-controlled object; determining whether the first and
second lines are interacting; and determining whether the second,
third and fourth changeable individual characterizations combine to
provide a contextually relevant combination that is not random,
subject to the second, third and fourth changeable individual
characterizations combining to provide a contextually relevant
combination that is not random, causing the first changeable
individual characterization to the change to the fifth changeable
individual characterization by inheriting the contextually relevant
combination based on an edge justification determined by the
processor-controlled object in the first line relative to the
contextually relevant combination in the second line.
23. The method of controllably presenting changeable,
processor-generated individual characterizations of claim 21,
wherein the contextually relevant combination inherited as the
fifth changeable individual characterization includes at least the
second, third and fourth changeable individual
characterizations.
24. The method of controllably presenting changeable,
processor-generated individual characterizations of claim 21,
further comprising: inheriting and displaying a word produced from
multiple similar interacting objects positioned in the line above
said processor-controlled object.
25. The method of controllably presenting changeable,
processor-generated individual characterizations of claim 17,
further comprising: inheriting phonetic properties produced from at
least two similar interacting objects positioned in a line above
said processor-controlled object and provided that interacting
properties of the at two similar objects combine to provide a
non-random phonetically meaningful combination.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application is a continuation of U.S. patent
application Ser. No. 11/142,955, filed on Jun. 2, 2005, entitled
"MANIPULABLE INTERACTIVE DEVICES," which is incorporated herein by
reference in its entirety for all purposes.
FIELD OF THE INVENTION
[0002] This invention generally relates to processor controlled
objects, such as inter-communicating processor-controlled blocks.
More particularly, the present invention is directed to a
controllably presenting changeable, processor-generated individual
characterizations presented on each of a plurality of objects that
selectively interact.
BACKGROUND OF THE INVENTION
[0003] This invention relates to a manually manipulable device,
especially a device which is adapted to interact with a similar
device or devices according to their relative locations so as to
produce a sensory response for a user, such a device being
especially suitable for educational or entertainment purposes.
SUMMARY OF THE INVENTION
[0004] The present invention generally comprises a method of using
processor controlled objects including controllably presenting
changeable, processor-generated individual characterizations
presented on each of a plurality of processor-controlled objects to
present changeable individual characterizations to a user. A user
manipulating the objects can cause, over time, designated objects
to inherit characterizations and properties from other interacting
objects. Interactions further generate sensory responses in the
response generator of a specific object or otherwise in a response
generator associated with another similar object based on
proximity, relative position and the individual characterization
presented on and by those interacting objects at the time of
interaction. A set of objects has extended interactive capabilities
since each object is capable of dynamically taking on different
characterizations arising from a meaningful combination of
properties from different conjoined objects.
[0005] In one embodiment of a method of controllably presenting
changeable, processor-generated individual characterizations
presented on each of a plurality of objects that selectively
interact includes generating and displaying under powered
processor-control first visual display material on a first movable
object. The first visual display material has a first changeable
individual characterization having a first property. The method
includes sensing proximity and relative position of second visual
display material generated and displayed under powered
processor-control on a second movable object separate to the first
movable object. The second object is brought into
processor-resolvable interacting proximity with the first moveable
object by manipulation of one or more of the first movable object
and the second movable object. The second visual display material
has a second changeable individual characterization independent to
the first individual characterization and the second changeable
individual characterization has a second property. In response to
processor-resolved interaction between the first and second
changeable individual characterizations arising from sensed
proximity and relative position of the first and second objects,
the method includes generating a user-perceivable sensory response
from a response generator, wherein the user-perceivable sensory
response is dependent upon the sensed relative positions of the
first visual display material to the second visual display material
and is indicative of a contextual relationship that arises between
said first property of the first changeable individual
characterizations and the second property of the second changeable
individual characterizations. The method also includes under
processor-control, selectively and autonomously changing with time
the individual characterizations on at least one of the first and
second objects such that the first property and the second property
change to allow different processor-resolvable interactions to take
place between the first and second objects, which different
processor-resolvable interactions give rise to new user-perceivable
sensory responses.
[0006] According to a first aspect, the invention consists of a
manually manipulable device adapted to present an individual
characterisation to a user comprising a processor, a power source,
a communications unit, a response generator, and a proximity sensor
adapted to sense the close proximity of a similar device, such that
a user can manipulate the device and generate a sensory response in
said response generator or a response generator of a similar
device, in accordance with the proximity of one or more similar
devices.
[0007] According to still another aspect, the invention consists of
a set of two or more manually manipulable devices, each adapted to
present an individual characterisation to a user and to be
locatable relative to other such devices in multiple different
arrangements, wherein each device comprises a processor, a power
source, a response generator, and a communications unit, such that
the devices generate a sensory response through said response
generators in accordance with the arrangement of the devices
selected by a user.
[0008] The characterisation may comprise visual display material or
audio output material, and will vary depending on the particular
application or purpose of the device or devices. For example,
visual display material may comprise a letter or group of letters
(e.g. phoneme) or word or words, and the sensory response may
comprise speech corresponding to a word or phrase or sentence spelt
out by the letters or words. In another application, visual display
material may comprise a number or mathematical symbol, and the
sensory response may comprise speech relating to mathematical
properties of the numbers on the devices. In yet another
application, visual display material may comprise a musical symbol
and the sensory response may be an audio musical response. In an
example in which the characterisation comprises audio output
material, this may comprise the audio equivalent of any of the
examples of visual display material given above.
[0009] In some implementations of the invention, the sensory
response comprises an audio response which may be generated by one
or more devices. Thus, each device incorporates an audio generator
to provide an audio response. However, in other examples of the
invention, the sensory response may instead, or in addition,
comprise a visual response, which may be generated by one or more
devices.
[0010] In a preferred embodiment of the invention, each device
incorporates a visual display unit which displays visual display
material and/or is able to generate a visual sensory response,
which may be a static or animated visual display. Preferably, each
device is programmable to allow the visual display material and the
sensory response to be programmed to suit different applications,
for example, to accommodate letters or words or numbers or musical
symbols as described above, or any other visual display material,
and to generate corresponding audio or visual responses.
[0011] Therefore, a device according to the invention is preferably
a fully programmable, multifunctional device which can be adapted
for use as a learning aid in relation to language, mathematics or
music or other subjects. Such a device can be readily adapted to be
used in the manner of known multi-component, educational apparatus
such as Cuisinaire rods (used to teach arithmetic), dominoes and
jigsaws, each component (rod, domino or jigsaw piece) being
embodied in the form of a device according to the invention, which
is then able to respond visually or audibly to enhance the
experience of the user of the apparatus.
[0012] The communications unit incorporated in the device is
adapted to communicate with similar devices with which it is used
to co-ordinate the sensory response appropriate to an array of
multiple devices. Each device communicates relevant information
about itself corresponding to its characterisation and may be a
simple identity code. The sensory response is made evident through
one or more of the devices, and could include a separate response
generator.
[0013] Communication of a sensory response to any device preferably
occurs via the communications unit.
[0014] Preferably, the communications unit is a wireless device,
that may be implemented using mobile telephone technology or the
like.
[0015] Each device is preferably provided with a proximity sensor,
or multiple proximity sensors, adapted to sense the proximity of a
similar device in any one of multiple adjacent positions, for
example, adjacent to each of multiple edges of the device. Each
device is preferably further adapted to identify an adjacent device
and to communicate information of both the identity and position of
an adjacent device to other devices or to the central control unit
via said communication unit so that an appropriate response can be
generated.
[0016] The proximity sensor may comprise a magnetic or an
electrical device, and may require physical contact between
adjacent devices to be operational.
[0017] Preferably, a manually manipulable device according to the
invention is constructed with a robust outer casing suitable for
handling by a child aged 3 or older.
[0018] Preferably, a manually manipulable device according to the
invention has registration features, such as protrusions and
indents, in its outer surface that allow the device to be placed in
registration with other such devices. Preferably, the registration
features provide a visual guide during the registration process.
The registration features may interlock adjacently located manually
manipulable devices according to the invention. In one embodiment,
a manually manipulable device according to the invention is
arranged to provide an indication when registration with another
such device is achieved. The indication may be audible or visible
in nature.
[0019] Adjacent contacting edges of devices may be adapted to fit
together or interlock only when correctly orientated so that both
display said visual display material the same way up (i.e. top to
bottom). A rectangularly shaped device may be adapted to be
orientated with a similar device adjacent to each of its four side
edges, and the proximity sensor is then adapted to sense each
adjacent device.
[0020] In an alternative embodiment of the invention, the devices
are used in conjunction with a board, tray or base on which they
are placed and which is capable of identifying the location and
identity of each device and communicating this to a central control
unit or one or more of the devices so that they can generate the
sensory response. The board itself may consist of a screen which is
able to generate a display appropriate for the particular
application and/or to generate the sensory response. In this
alternative embodiment of the invention, the individual devices may
not need to incorporate the proximity sensor because of the
location sensing ability of the board.
[0021] In the above alternative embodiment, the board may be
adapted so that it can recharge individual devices when placed in
contact with it. Furthermore, this recharging feature may be
provided in a board not having the device location capability.
[0022] A device according to the invention may also incorporate a
camera that allows an image to be captured, this image being used
as said visual display material on a visual display unit also
incorporated in the device, or the image can be used in a visual
sensory response.
[0023] A device according to the invention may also incorporate a
microphone to allow sound to be captured and used in an audio
sensory response.
[0024] A device according to the invention may also incorporate
data input means in the form of a handwriting recognition device to
input words, letters, symbols or numbers for use in
characterisation of the device or programming a sensory response to
be produced by the device.
[0025] It will be appreciated that an audio sensory response, such
as incorporated in any of the embodiments described above, may take
the form of a directional or stereo/audio response by arranging
that two or more devices are controlled simultaneously or
sequentially to generate appropriate sounds.
[0026] Programming of each device may be achieved by any of a
number of different methods including connection to memory media
such as smart cards or memory sticks; via a personal computer or
hand-held computing device; or via said communications unit. In one
example, each device may make use of the communications unit to
receive information from a television broadcast so that the device
is adapted for use in conjunction with a television programme being
broadcast.
[0027] A device according to the invention is preferably further
adapted so that it incorporates a user sensor sensitive to touch
and/or movement so that it can trigger a characterisation output
when handled by a user. The characterisation output may comprise a
visual or audio output or both.
[0028] Specific technologies that can be used in embodiments of the
invention include networked distributed intelligent small computers
known as Specks or Motes; micro-electromechanical-systems MEMs,
especially for audio components and sensors; and ZigBee radio or
similar communications technology.
[0029] A manually manipulable device according to the invention is,
from one aspect, a computing unit and as such can be designed to be
a thin client in a client-server relationship with some other
entity.
[0030] In one embodiment, a manually manipulable device according
to the invention comprises a 32 Bit RISC (or better) CPU, memory, a
graphics processor, an audio processor, a communications processor,
internal data storage, a rechargeable power source and a
touch-sensitive audio-visual display unit. The CPU is preferably
capable of processing 200 Million Instructions Per Second (MIPS) or
better. The CPU can preferably address 16 Mb (or better) of Random
Access Memory. The graphics processor and visual display will
preferably be capable of rendering screen resolutions of
160.times.160 pixels (or better) in 8 bit colour (or better). Other
versions will be able to process full motion video at 12.5 frames
per second (or better) with 16 bit colour (or better) synchronised
to audio. Other versions will have live video or still image
capture via a built-in camera. The audio processor will preferably
be capable of playback of 4 bit, 4 kHz mono audio (or better) and
polyphonic tones. Enhanced versions will feature audio recording
capability. The internal storage may be provided by Secure Digital
(SD) cards, MultiMedia Cards (MMC) or a hard disc arrangement. The
communications processor will preferably include support for
industry standard wireless protocols including Bluetooth and in
future will support other emergent protocols including IEEE
802.15.4 and other near field communication protocols. It is
presently preferred that a manually manipulable device according to
the invention will have a real time operating system (RTOS).
[0031] Video apparatus could for example involve the use of screens
5 cm.times.5 cm, but 8 cm.times.8 cm might also be acceptable. The
screens could for example comprise thin film transistor TFT screens
with an active matrix 2.5'' (4:3), a resolution 880.times.228 RGB
delta, pixel size 56.5.times.164 HM, fully integrated single-phase
analogue display drivers, signal input voltage 3V, driver frequency
3 MHz, driver power consumption 15 MW.
[0032] The power source is preferably a rechargeable battery and
might comprise a photovoltaic generator.
[0033] The user sensor may also sense manipulation of the device by
a user indicative of a positioning movement of the device requiring
an assessment of its proximity relative to similar devices and the
need to generate a sensory response corresponding to one of said
arrangements of devices.
[0034] According to a further feature of the invention, each of
said manually manipulable devices incorporates a visual display
unit to display visual display material, and two or more of said
devices are adapted to be arranged in a row so that said visual
display material "reads" in a meaningful manner along said row. A
similar device is locatable adjacent to one side of said row of
devices, and thereby triggers a change in the visual display
material on said similar device so that it matches that of said row
of devices. For example, said similar device can be located below
said row of devices to acquire a combination of characters from the
row above it. This device displaying said combination of characters
can then be re-used in a further row of devices to create a new
combination of characters.
[0035] Each device may have an ON/OFF switch to allow it to be
reset to a start up condition, for example, displaying initial
pre-programmed visual display material.
BRIEF DESCRIPTION OF THE DRAWINGS
[0036] By way of example only, certain embodiments of the invention
will now be described with reference to the accompanying drawings,
in which:
[0037] FIG. 1 illustrates the external physical constitution of an
interactive block;
[0038] FIG. 2 illustrates, conceptually, the internal constitution
of the interactive block of FIG. 1,
[0039] FIG. 3 illustrates how blocks of the kind illustrated in
FIG. 1 can be connected in registration with one another;
[0040] FIG. 4 illustrates how blocks of the kind shown in FIG. 1
can be used in a learning activity;
[0041] FIG. 5 illustrates schematically an interactive block;
and
[0042] FIG. 6 illustrates schematically a tray or board which can
interact with blocks of the kind shown in FIG. 5.
DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
[0043] One embodiment of the invention consists of a set of blocks,
say 12 blocks, each being rectangular in shape and adapted to be
positioned edge-to-edge with other blocks on either side (referred
to as left-hand and right-hand edge) in the manner of a row, and
top edge and bottom edge in the manner of a column. Each block
incorporates a display screen over most of its front or upper
surface, which forms part of an electronic visual display unit
capable of displaying visual display material according to display
data derived from a database. In this embodiment, the visual
display material consists of a lower case letter of the alphabet
which is displayed on the screen when the block is first activated.
Each block may incorporate a switch that allows it to be activated
to deactivated, and operation of the switch initiates a start-up
condition in which a pre-programmed letter is displayed.
Programming of the blocks may be such that different combinations
in a row can spell out fifteen to twenty different words
appropriate for teaching a young child to read.
[0044] Each block incorporates a means of displaying its
orientation as far as top and bottom is concerned, which may invoke
the shape of the block or an indicator displayed in the display
screen.
[0045] Each block further incorporates a proximity sensor or sensor
adapted to allow it so sense the proximity of another block aligned
edge-to-edge with it, preferably involving contact between said
adjacent edges, either at the left hand edge or right-hand edge or
top edge or bottom edge. The proximity sensor, or other ID sensor
means independent of it, is adapted to sense the identity of each
adjacent block.
[0046] Each block further incorporates a touch and/or movement
sensor.
[0047] Each block further incorporates a wireless communications
unit through which it can communicate with another block to
transmit information relating to its own identity and visual
display material and the identity and location of adjacent blocks
and to receive information causing the visual display unit to
change the visual display unit material.
[0048] Each block preferably further incorporates an audio
generator which is adapted to produce an audio response in
accordance with internal programming information received via the
wireless communications unit.
[0049] It will be appreciated that a block with the communications,
visual display and audio generator capability described above can
be readily implemented using mobile telephone technology. Proximity
sensors, ID sensors and touch and movement sensors can also be
readily implemented using known technology. It will be appreciated
that each block has its own power supply and incorporates a
processor or processors which provide the required
functionality.
[0050] A set of blocks is adapted to be sufficient in itself to
provide the functionality described below with the processors
operating in accordance with pre-programmed instructions and the
inputs from the sensors of each so as to produce visual and audio
responses in the blocks.
[0051] The constitution of an example one of the blocks is shown in
FIGS. 1, 2 and 5. The internal construction of a block is shown
conceptually in FIG. 2 and in block diagram form in FIG. 5. FIG. 3
illustrates how blocks of this kind can be placed in registration
with one another both vertically and horizontally.
[0052] FIG. 5 illustrates the main components of a block. It will,
of course, be apparent to the skilled person that this is a high
level diagram illustrating only key components of the block. As
shown in FIG. 5, a block 500 comprises a processor 510, a memory
512, an RF transceiver 514, a screen 516, a speaker 518, a magnetic
switch 520, a touch sensor 522, a movement sensor 524, a docking
port 526 and a battery 528. The RF transceiver 514 enables the
block 500 to communicate wirelessly with other, at least similar,
blocks in the vicinity. The screen 516 and the speaker 518 allow
visual and audio information to be presented to a user of the block
500. The magnetic switch 520 is activated by the proximity of
another, at least similar, block. The touch sensor 522 is provided
at the exterior of the block 500 to detect a user touching at least
that area of the block 500. The movement switch 524 detects
movement of the block 500 by a user. The docking port 526 is for
receiving a memory card to load software/data into the block 500.
The block 500 also includes a battery 528 that provides power to
allow the various devices within the block to operate. The
processor 510 processes, with the aid memory 512, information
received from the RF transceiver 514, the switch 520, the touch
sensor 522, the movement sensor 524 and the docking port 526 to
cause, as appropriate, the RF transceiver 514 to communicate with
other blocks and/or cause the screen 516 and/or the speaker 518 to
present information to a user of the block 500.
[0053] FIG. 6 shows a tray 600 for use with blocks, e.g. 610 to 616
of the kind described above with respect to FIGS. 1, 2 and 5. The
board 600 comprises a detector 618 for determining the location and
identity of blocks placed on the board. The board 600 also includes
a charger for recharging the batteries of blocks that are placed on
the board. The board also includes a screen 622 and is configured
to present information to a user via the screen in response to
interactions of the user with blocks on the board.
[0054] Examples of how the set of blocks can be used as alphabet
blocks will now be now described.
[0055] Sam is four and a half. She's just started in her reception
year at school where she's learning to read and write. Her parents
are keen to help her learn at home and buy her a set of blocks with
some preloaded reading software appropriate for her age.
[0056] Sam opens the box and takes out the blocks. Her parents are
standing over, curious about how they work.
[0057] Each is displaying a different lower case letter.
[0058] She goes to pick one up and the unit sounds the letter it is
displaying. For example, `/c/`. Moving each of the blocks she
realises they all do the same.
[0059] In an alternative option, if there's 15 seconds inactivity,
one block could say `Try spelling a word, how about cat`.
Alternatively pressing on a block could say, `c sounds like /c/.
/c/ is for cat. Move the blocks together to spell cat?`
[0060] Sam puts two of the blocks next to each other. Starting with
the one on the left, the blocks read in turn the letters they are
displaying. For example `/d/, /o/`. They then read the combined
sound. For this example the blocks would say `do`.
[0061] When she puts three `random` letters together ('/c/, /f/,
/g/'), they make no sound.
[0062] She plays around with some different combinations until a
word is spelt. For example, `/c/, /a/, /t/. You've spelt cat. Well
done.` At this point a cat leaps onto the screen, runs around and
miaows.
[0063] In an alternative option, the blocks prompt the child what
to do next? For example, Now you can copy the word you've made onto
its own block, by placing one below. Or you can try and spell
another word.'
[0064] When Sam puts another block below the word she has spelt,
the word jumps down onto that single block. It's says `cat` when
she presses it.
[0065] The three blocks that originally spelt the word are now free
to be used for another word.
[0066] As described above, each block is individually responsive to
touch or movement and reacts audibly and visually depending upon
what it displays.
[0067] If each block is responsive to both truck and movement
separately, then each can have a secondary response, such as giving
an example of use.
[0068] If a letter is displayed, e.g. "c", the block sounds the
letter as in it is said in the alphabet and phonetically. For
example, `C. C sounds like /c/ for cat`. An animation may play on
the screen relating to the letter and the example given. A
secondary response might suggest what the user can do next? For
example, `Can you spell Cat?`
[0069] If a word is displayed e.g. "cat", the block sounds the
phonetic letters for the word. For example, `/c/, /a/, /t/ spells
cat`. An animation relating to the word plays on the screen. A
secondary response might suggest the spelling of another word from
the available letters if this is possible.
[0070] If a phonetic sound is displayed e.g. "ch", the block sounds
the combined phonetic sound `/ch/ as in lunch`. The screen displays
an animation of some food being eaten.
[0071] When blocks are placed next to each other they react
depending what is on each. This could be a phonetic sound e.g.
`/ch/`, a word e.g. `cat` or random letters e.g. `/k/, /r/,
/f/`.
[0072] If the user places individual blocks alongside each other
then they respond according to the combination of letters they
display.
[0073] If a phonetic sound is created "ch", the blocks sound the
combined sound, `/ch/`. They could also give a short example of use
`/ch/ as in lunch, yum, yum, yum`.
[0074] If a word is created "cat", the blocks sound the individual
letters followed by the word. For example, `/c/, /a/, /t/, spells
cat. Well done, you've spelt cat`. The displays play a short
animation. In this example a picture of a cat running between the
two blocks. This happens whenever one of the joined blocks are
pressed.
[0075] If a new word is created (plural or completely new) by
adding a letter or letters to a current word of phonetic sound, the
response might be, for example, `/c/, /a/, /r/, /t/, spells cart.
Are you coming for a ride?` or `/c, /a/, /t/, /s/ spells cats. Here
they come!`. The displays animate according to the word spelt if
the word has an associated animation in database. So in the above
examples, a horse and cart could drive on and off the screens, or
several cats could start playing around.
[0076] If a random set of letters are placed next to each other.
For example `/d/, /f/, /r/, /g/`, no sound is generated and no
animation is displayed.
[0077] Animation and sound will only be available for some of the
words that can be created using the blocks, as stated in a related
response database held in one or each block or a central control
unit.
[0078] If a user places one block adjacent the top edge of another,
the lower block inherits the property of the upper block. Placing
multiple blocks above or below will also cause a reaction between
the blocks. For example, if the user places one block above
another, and the top block shows `/b/` and the lower block shows
`/b/`, the lower block will also become a `/b/`.
[0079] A user can place a word spelt out over several blocks onto
one block by placing a block below. This could also be used to join
a `/c/` and an `/h/` on a single `/ch/` block.
[0080] If a user has spelt a word or phonetic sound using three
individual blocks, for example, `/c/`, `/a/` and `/t/` spelling
`cat`, the user can then place a fourth block anywhere under the
three letter blocks and the word "cat" moves onto a single block.
However, if a user tries to copy two random letters onto a single
block it will not work. For example `/g/` and `/f/` cannot be
joined on a single `/gf/` block.
[0081] Likewise if the user has two word blocks that don't make a
third word, they cannot be copied onto a single block. For example
`cat` and `sat` cannot be joined to make a `catsat` block.
[0082] If a user has the word cat on a single block and wants to
split it into three separate letters, they need to place three
blocks below the word block. The three letters each go into their
own block in right to left order below.
[0083] An example of use of a set of alphabet blocks operating
according to the above principles is illustrated in FIG. 4, in a
number of steps 1-6.
[0084] 1. Blocks are taken out of the box and arranged on the
floor.
[0085] 2. User puts `/c/` and `/h/` together, the blocks sound
`/ch/`. They put `/g/` underneath and copy `/ch/` onto it. Trying
to copy `/t/`, `/m/` onto `/g/` doesn't work.
[0086] 3. `/a/` and `/t/` are joined to make `at` and copied onto a
single block.
[0087] 4. `/m/` is put in front of `at` to make `mat`. The
individual `/a/` and `/t/` blocks are still joined to the top of
`at`, but have no direct effect to the `/m/` as they are not
directly above but to one side. `/u/` is put below the `/m/` of
`mat` and `mat` is copied onto the single block, which is then
removed (not illustrated).
[0088] 5. A `/s/` block is put in front of the `/a/` and `/t/`
blocks to spell `sat`. As the `/m/` of `mat` is now below the `/s/`
block the word `sat` is copied onto it. `sat` is also copied onto
the `at` block. The two `sat` blocks don't interact with each other
as a new word or sound hasn't been created. Likewise when a `r`
block is placed below either of the `sat` blocks nothing is copied
down.
[0089] 6. Using the blocks here is a chain of various words that
can be created following the principles described in the functional
specifications.
[0090] The invention is applicable to diverse areas, which include
but are not limited to, play, entertainment, adornment and
decoration, environment, industry and learning (of, for example,
languages, mathematics and musical skills/knowledge).
[0091] Play applications may include a variety of playful games
using the blocks and, optionally, a tray of the type mentioned in
the introduction. These include new games as well as enhancements
of typical existing board and card games with additional features
(by virtue of the fact the pieces (blocks) can change their image
and emit sounds) and the board (interactive base) can also change
its image. Further, new forms of toy such as farmyards and zoos can
be created and become elements of animated stories.
[0092] In relation to adornment and decoration, in the educational
context, IA blocks can be worn as badges that can enable students
to role play their various functions (letters, sounds, numbers) and
interact with other badge-wearing children to form words, tunes and
equations. Beyond this, IA blocks have implicit emotive, aesthetic,
interactive, and descriptive capabilities. Blocks in combination
can be used to trigger social and artistic interactions between
people or create more complex installations.
[0093] In environment and industrial settings variations of the
devices can enable audio and visual data/systems alone or in
combination (e.g. for health and safety measurement and
control).
* * * * *