US7116316B2 - Audible and visual effects as a result of adaptive tablet scanning - Google Patents

Audible and visual effects as a result of adaptive tablet scanning Download PDF

Info

Publication number
US7116316B2
US7116316B2 US10/094,583 US9458302A US7116316B2 US 7116316 B2 US7116316 B2 US 7116316B2 US 9458302 A US9458302 A US 9458302A US 7116316 B2 US7116316 B2 US 7116316B2
Authority
US
United States
Prior art keywords
writing
graphic image
tablet
audible
area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US10/094,583
Other versions
US20030169237A1 (en
Inventor
Lenka M. Jelinek
Timothy L. Brooke
Frank T. Brown
Herman D. D'Hooge
Wendy A. March
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Intel Corp
Original Assignee
Intel Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Intel Corp filed Critical Intel Corp
Priority to US10/094,583 priority Critical patent/US7116316B2/en
Assigned to INTEL CORPORATION reassignment INTEL CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BROOKE, TIMOTHY L., BROWN, FRANK T., JELINEK, LENKA M., MARCH, WENDY A., D'HOOGE, HERMAN D.
Publication of US20030169237A1 publication Critical patent/US20030169237A1/en
Application granted granted Critical
Publication of US7116316B2 publication Critical patent/US7116316B2/en
Adjusted expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/042Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
    • G06F3/0425Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means using a single imaging device like a video camera for tracking the absolute position of a single or a plurality of objects with respect to an imaged reference surface, e.g. video camera imaging a display or a projection screen, a table or a wall surface, on which a computer generated image is displayed or projected
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/042Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
    • G06F3/0421Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means by interrupting or reflecting a light beam, e.g. optical touch-screen
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B11/00Teaching hand-writing, shorthand, drawing, or painting
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B17/00Teaching reading
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B5/00Electrically-operated educational appliances
    • G09B5/06Electrically-operated educational appliances with both visual and audible presentation of the material to be studied

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Educational Administration (AREA)
  • Educational Technology (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Electrically Operated Instructional Devices (AREA)
  • Position Input By Displaying (AREA)

Abstract

One aspect of the invention involves a writing tablet comprising a housing having a writing surface and internal logic placed within the housing and situated below the writing surface. The internal logic includes imaging sensor, a light source and a controller. The imaging sensor is adapted to capture a location and orientation of each line segment forming a graphic image. The light source is adapted to produce a light beam adjusted for display on the writing surface. The controller is adapted to control performance of an event upon the light beam crossing a line segment of the graphic image.

Description

FIELD
The invention generally relates to the field of writing tablets. In particular, one embodiment of the invention relates to a writing tablet that performs an event upon detecting a graphic image written on the screen of the writing tablet.
GENERAL BACKGROUND
Over the past decade, both parents and institutions are relying more heavily on educational toys as supplemental learning tools. One popular education toy is a writing tablet that is communicatively coupled to a computer. Normally, a stylus or other non-ink producing writing instrument accompanies the tablet.
As one end of the stylus comes into contact with a writing surface of the tablet, its location is registered by an image sensing mechanism situated within the tablet. In response to the user gliding the stylus across the writing surface of the tablet, outlining a desired graphic image (e.g., a handwritten alphanumeric character, a doodle, an artistic rendering, etc.), the computer generates a corresponding image for display on its monitor screen.
One problem with conventional writing tablets is that they fail to provide the user any visual or audible feedback for learning enhancement.
BRIEF DESCRIPTION OF THE DRAWINGS
The features and advantages of the invention will become apparent from the following detailed description of the invention in which:
FIG. 1 is a first exemplary embodiment of a writing tablet.
FIG. 2 is a first exemplary embodiment of internal logic situated within the writing tablet of FIG. 1.
FIG. 3 is an exemplary embodiment of a scanning process conducted by the writing tablet of FIGS. 1 and 2.
FIG. 4 is a second exemplary embodiment of a writing tablet.
FIG. 5 is an exemplary embodiment of internal logic situated within the writing tablet of FIG. 4.
FIG. 6 is an exemplary flowchart featuring the operations of the writing tablet set forth in FIG. 1 or FIG. 4.
DETAILED DESCRIPTION
In general, one embodiment of the invention relates to a writing tablet that performs an event upon detecting a graphic image written on the screen of the writing tablet during an image scanning operation. For clarity, the term “writing” and related tenses used herein involve the act of handwriting and/or drawing.
Certain details are set forth below in order to provide a thorough understanding of the invention, albeit the invention may be practiced through many embodiments other that those illustrated. Well-known circuits and operations are not set forth in detail in order to avoid unnecessarily obscuring the invention.
In the following description, certain terminology is used to describe certain features of the invention. For example, a “computing device” includes logic, namely hardware, firmware, software module(s) or any combination thereof that performs a desired function. In one embodiment, the computing device is a computer such as a desktop computer, laptop computer, hand-held (e.g., personal digital assistant), mainframe, workstation, or server. Other examples of computing devices include, but are not limited or restricted to other communication equipment such as an alphanumeric pager, a printer, a facsimile machine, a set-top box or a wireless telephone for example.
A “physical-ink writing instrument” is any device that dispenses writing solution during use. Examples of writing solution includes ink, dry ink powder, chalk, crayon, lead and the like. Examples of a physical-ink writing instrument includes a marker, ink-filled pen, pencil, crayon, etc. A “virtual-ink writing instrument” is any device that, during use, does not dispense writing solution. Examples of a virtual-ink writing instrument includes a stylus.
A “software module” is a series of instructions that, when executed, performs a certain function. Examples of a software module include an operating system, an application, an applet, a program or even a routine. One or more software modules may be stored in a machine-readable medium, which includes but is not limited to an electronic circuit, a semiconductor memory device, a read only memory (ROM), a flash memory, a type of erasable programmable ROM (EPROM or EEPROM), a floppy diskette, a compact disk, an optical disk, a hard disk, or the like.
In addition, a “graphic image” includes one or more lines segments (e.g., a single point or multiple points connected together as a marking) that form (i) handwritten alphanumeric characters or symbols or (ii) images featuring one or more geometrical shaped objects or artistic renderings. The alphanumeric characters may be in any case (upper/lower) or style (cursive or printed) and in accordance with any character type (e.g., Roman, Kanji, Arabic, Chinese, etc.).
Referring to FIG. 1, a first exemplary embodiment of a writing tablet 100 is shown. The tablet 100 comprises a housing 110 made of a rigid material such as hardened plastic. The housing 110 protects internal logic 200 employed within a cavity formed by the housing 110. By substantially encasing or perhaps completely encapsulating the internal logic 200, the housing 110 protects the internal logic 200 from damage and contaminants.
One surface 120 of the housing 110 is adapted with a connector port 130 to be accessible through the housing 110. The connector port 130 enables communication with a computing device 170 via a link 140. In general, the link 140 is one or more physical or virtual information-carrying mediums that establishes a communication pathway. The link 140 may be adapted as an electrical wire, electrical or optical cable, wireless signaling technology, or another means of communication.
Referring still to FIG. 1, the connector port 130 may be configured as a Universal Serial Bus (USB) port that supports “plug and play” operations. Alternatively, the connector port 130 may be a serial port or a parallel port for direct communications with the computing device 170. The connector port 130 may also be a network adapter (e.g., Ethernet adapter) that enables communication with a network. Of course, multiple connector ports may be provided to support different types of adapters.
As shown, another surface 125 of the housing 110 features a writing area 150. The writing area 150 is a region made of either a semi-opaque material having a translucent or transparent quality (e.g., plastic, glass, etc.) or a liquid crystal display (LCD) or plasma screen. The writing area 150 may be sized to accommodate an overlay of standard letter size paper (8.5″×11″), although other sized screens may be implemented as an alternative.
In one embodiment, a writing instrument 190 is used by the user to produce a graphic image 160 within the writing area 150. A representation of the graphic image 160 is transferred to the computing device 170, which is in communications with the tablet 100 and controls the display of the representation of the graphic image 160 on its monitor screen 180.
As shown in both FIGS. 1 and 2, since the writing area 150 is semi-opaque, which means that light may pass through it, an imaging sensor 210 may be placed internally within the housing 110. As part of the internal logic 200, the imaging sensor 210 captures the graphic image 160 from below the writing area 150 of the tablet 100.
In one embodiment, the imaging sensor 210 is an optical sensor that optically captures the entire graphic image 160. A graphical location and orientation of each line segment forming the graphic image 160, perhaps in accordance with Cartesian coordinates at which a selected point on the writing surface 150 is an origin, is transferred to memory of the computing device 170. As an optional feature represented by dotted lines, the location and orientation information associated with the graphic image 160 may be temporarily stored in memory 220 before transfer to the computing device 170. Being part of the internal logic 200, the memory 220 may be volatile or nonvolatile memory.
In another embodiment, the imaging sensor 210 involves magnetic sensing, where magnetic displacement caused by the writing instrument 190 enables the imaging sensor 210 to locate and identify the writing instrument 190. As a result, specifics associated with the graphic image 160 placed on the writing area 130 may be identified as well. For example, the imaging sensor 210 may be adapted to only detect writing instruments that are in contact with the writing area 150. Also, the writing solution forming the graphic images may have magnetic characteristics that are detected when deposited on the writing area 150.
In yet another embodiment, the imaging sensor 210 may involve radio frequency (RF) or pressure sensing to determine the location of graphic images written on the writing area 150. Herein, for RF sensing, the imaging sensor 210 would receive wireless signals from one or more writing instruments in accordance with a unicast or multicast transmissions. The transmission range for the wireless signals may be adjusted so that the imaging sensor 210 can only detect the wireless signals from writing instruments that are in contact with or in close proximity to the writing area 150. Normally, this range may be varied by adjusting the level of effective isotropic radiated power (referred to as the “power level”) utilized by the imaging sensor 210. However, it is contemplated that the power level may be configured to be at a constant level.
Since the imaging sensor 210 is used to capture a graphic image 160 placed on the writing area 150, the graphic image 160 may include line segments produced by physical-ink or virtual-ink writing instruments or perhaps images printed, painted or carved on any object placed on and facing the writing area 150. For example, where the imaging sensor 210 is an optical sensor, an image from a page of a children's book placed on the writing area 150 may be replicated and displayed on the monitor 180 of the computing device 170.
The internal logic 200 further includes a controller 225, light source 230, and light beam adjustment (LBA) logic 235. The controller 225 controls the light source 230, which produces a light beam 240 that is reflected and adjusted by the light beam adjustment logic 235. The controller 225 controls the operations of the light beam adjustment logic 235. Examples of the controller 225 include at least one of a general microprocessor, a digital signal processor, a microcontroller, an application specific integrated circuit (ASIC), a state machine, or other types of data processing logic.
As shown in FIGS. 2 and 3, the light beam 240 defines what regions of the writing area 150 that the computing device 170 is scanning to detect graphic images placed on the writing area 150. The light beam 240 illuminated on the writing area 150 may be adjusted by the light beam adjustment logic 235 so as to have an longitudinal (as shown), lateral, diagonal or circular orientation. Of course, it is contemplated that the light beam 240 may be oriented in any geometric shape. The color of the light beam 240 is a design choice.
Referring to both FIGS. 1–3, it is contemplated that prior to generation of the light beam 240, the computing device 170 is already apprised of the location of any graphic images as described above. The light beam 240 is used to identify to the user what regions of the writing area 150 are being analyzed.
According to one embodiment of the invention, as the light beam 240 encounters a line segment of a graphic image, the tablet 100 reacts by triggering an event such as an audible effect, a visual effect or a combination. In general, an “audible effect” involves the generation and play back of audio, which may be used to reinforce the user's learning of letters, numbers and common shapes. A “visual effect” is generation and display of an image or perhaps changing a color of the graphic image displayed or its background.
For instance, with respect to one embodiment, the audible effect triggered by the tablet 100 may involve the transmission of control signals to the computing device 170, which causes the computing device 170 to play back a series of musical notes upon detecting a graphic image 160. Another event may involve the computing device 170 playing back an audible rendition of a detected alphanumeric character. For example, upon scanning the writing surface and encountering the letter “S”, the tablet 100 signals the computing device 170 to play a sound that verbally states the letter “S”.
Of course, it is contemplated that the tablet 100 may be adapted to perform an event itself upon detecting a graphic image 160 in lieu of the computing device 170. For instance, the tablet 100 may play back musical notes over a speaker as described in FIG. 5, render an audible sound of a detected character or other audible or visual effects.
Referring now to FIG. 4, a second exemplary embodiment of a writing tablet 400 is shown. The tablet 400 comprises the housing 410 that protects internal logic 500 employed within a cavity formed by the housing 410. One surface 425 of the housing 410 features a writing area 450 being a region made of either a semi-opaque material or an LCD or plasma display screen and a plurality of apertures 455 proximate to a speaker contained within the housing 410.
Similar to FIG. 1, the writing instrument (not shown) is used by the user to produce the graphic image 460 over the writing area 450. Since the writing area 450 is semi-opaque, so that light may pass through it, an imaging sensor 510 may be placed internally within the housing 410. As part of the internal logic 500 and shown in FIG. 5, the imaging sensor 510 captures the graphic image 460 from below the writing area 450 of the tablet 400.
Referring now to FIG. 5, a second exemplary embodiment of the internal logic 500 situated within the writing tablet 400 of FIG. 4 is shown. Herein, the internal logic 500 includes the imaging sensor 510, a controller 520, memory 530, a light source 540, light beam adjustment (LBA) logic 550 and a speaker 560.
As shown, the imaging sensor 510 is placed internally within the housing 410 and captures the graphic image 460 from below the writing area 450 of the tablet 400. The imaging sensor 510 may operate as an optical sensor, a magnetic sensor, a RF sensor or a pressure sensor as described above.
The controller 520 controls the operation of the imaging sensor 510 and the generation of the light beam over the writing area 450 during a scanning process. The functionality of the controller 520 is programmed by software modules placed within the internal memory 530 and perhaps memory within the controller 520 itself. Examples of the controller 520 include at least one of a general microprocessor, digital signal processor, microcontroller, ASIC, state machine, and the like.
The controller 520 further controls the light beam adjustment logic 550. For example, the controller 520 may include mirrors and galvanometers to adjust mirror positioning to reflect a light beam from the light source 540 (e.g., a laser, a light emitting diode, etc.). The reflected light beam is used to indicate the region of the writing area 450 that is being scanned for graphic images. However, the controller 520 is already apprised of the location of any graphic images based on the transmission of locations of the writing instrument being used on the writing area 450. The light beam may be oriented as a longitudinal line, a lateral line, diagonal line, circular object or in accordance with any geometric shape.
Upon the light beam encountering a line segment of a graphic image, the tablet 400 reacts by triggering an event. For example, upon detecting a graphic image, the tablet 400 plays back a series of musical notes via speaker 560. Upon detecting an alphanumeric character, the tablet 400 plays back an audible rendition of the alphanumeric character over speaker 560 as well.
Referring to FIG. 6, an exemplary flowchart featuring the operations of the writing tablet set forth in FIG. 1 is shown. Herein, the imaging sensor detects placement of a graphic image on a writing surface of the tablet (block 600). In one embodiment, the tablet notifies the computing device of the specific placement (e.g., graphical coordinates) of the graphic image (block 610). In another embodiment, the information associated with the specific placement of the graphic image is stored within internal memory (block 620). This detection and notification/storage process continues until the tablet enters into a scanning process state (block 630).
During the scanning process, a light beam is generated and directed along various regions of the writing surface (block 640). The light beam may be configured in any shape or orientation. In one embodiment, once the light beam intersect a portion of the graphic image, namely graphical coordinates of a portion of the light beam and image intersect, an event is triggered (blocks 650670). However, in another embodiment, an event may be triggered upon recognition of the image. For instance, it is contemplated that an event may be triggered only upon recognition of a particular image (e.g., alphanumeric character) so that any other image types (e.g., doodles by the user) are ignored.
In one embodiment, the triggered event is audio playback that identifies the graphic image as a particular alphanumeric character. Such identification may be accomplished when (1) the tablet is placed in a mode of operation associated with the generation of numbers or letters and (2) the tablet or computing device has already analyzed and determined the written character. One or more events may be triggered during the scanning process.
While certain exemplary embodiments have been described and shown in the accompanying drawings, it is to be understood that such embodiments are merely illustrative of and not restrictive on the broad invention, and that this invention not be limited to the specific constructions and arrangements shown and described. For example, it may be possible to implement the invention or some of its features in hardware, firmware, software or a combination thereof.

Claims (12)

1. A writing tablet comprising:
a housing including a wilting area; and
logic contained within the housing and situated below the writing area, the logic including
an imaging sensor to detect a graphic image on the writing area,
a light source to produce on a back surface of the writing area facing an interior of the housing a light beam that is adjusted for display on the writing area,
a speaker,
an internal memory, and
a controller coupled to at least the internal memory, the light source, the imaging sensor and the speaker, the controller to control generation of an audible rendition of an alphanumeric character for output from the speaker in response to both the light beam crossing a line segment of the graphic image previously captured by the imaging sensor and the controller identifying the graphic image as the alphanumeric character.
2. The writing tablet of claim 1, wherein the light source is a light emitting diode.
3. The writing tablet of claim 1, wherein the controller is a proccssor.
4. The writing tablet of claim 1, wherein the internal memory includes at least one software module executable by the controller to generate signals recognized as audible sounds by the speaker.
5. The writing tablet of claim 4, wherein the internal memory includes information related to a location and orientation of the graphic image placed on the writing area.
6. The writing tablet of claim 5, wherein the information includes Cartesian coordinates at which a selected point on the writing area is an origin.
7. The writing tablet of claim 1, wherein the controller is a microcontroller.
8. A method comprising:
detecting placement of a graphic image on a writing area of a writing tablet;
generating light beam on a back surface of and visible through the writing area to represent an area of the writing tablet being analyzed and adjusted to move over the graphic image; and
triggering at least one of an audible effect and a visual effect in response to the light beam intersecting a line segment of the graphic image previously drawn on and currently displayed within the writing area and recognition that the graphic image as a particular alphanumeric character, the audible effect being an audible rendition of the alphanumeric character.
9. The method of claim 8, wherein the triggering of the at least one audible effect and visual effect includes play back of the audible rendition of the alphanumeric character by the writing tablet.
10. The method of claim 8, wherein the triggering of the at least one audible effect and visual effect includes transmitting a signal to a computing device to control the computing device to playback the audible rendition of the alphanumeric character.
11. The method of claim 8, wherein detecting placement of the graphic image includes activating an imaging sensor to detect a graphical location at which a writing instrument comes in contact with writing area and storing the location within internal memory of the writing tablet.
12. The method of claim 8, wherein detecting placement of the graphic image includes activating an imaging sensor to detect a graphical location at which a writing instrument comes in contact with the writing area and transferring information representative of the graphical location to memory within a computing device coupled to the writing tablet.
US10/094,583 2002-03-07 2002-03-07 Audible and visual effects as a result of adaptive tablet scanning Expired - Fee Related US7116316B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US10/094,583 US7116316B2 (en) 2002-03-07 2002-03-07 Audible and visual effects as a result of adaptive tablet scanning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US10/094,583 US7116316B2 (en) 2002-03-07 2002-03-07 Audible and visual effects as a result of adaptive tablet scanning

Publications (2)

Publication Number Publication Date
US20030169237A1 US20030169237A1 (en) 2003-09-11
US7116316B2 true US7116316B2 (en) 2006-10-03

Family

ID=27788145

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/094,583 Expired - Fee Related US7116316B2 (en) 2002-03-07 2002-03-07 Audible and visual effects as a result of adaptive tablet scanning

Country Status (1)

Country Link
US (1) US7116316B2 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030117378A1 (en) * 2001-12-21 2003-06-26 International Business Machines Corporation Device and system for retrieving and displaying handwritten annotations
US20050266386A1 (en) * 2004-05-28 2005-12-01 Leapfrog Enterprises, Inc. Print media apparatus including stroke recognition
US20100245291A1 (en) * 2009-03-26 2010-09-30 Fuji Xerox Co., Ltd. Writing apparatus
US20110313771A1 (en) * 2005-11-01 2011-12-22 Leapfrog Enterprises, Inc. Method and device for audibly instructing a user to interact with a function
US8196041B2 (en) 2003-06-26 2012-06-05 International Business Machines Corporation Method and system for processing information relating to active regions of a page of physical document
US9335839B1 (en) 2014-04-08 2016-05-10 Clive Lynch Graphic artistic tablet computer
US20180297392A1 (en) * 2015-10-15 2018-10-18 Sanford, L.P. Generating mechanically rendered objects from digital input

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4596787B2 (en) * 2003-04-25 2010-12-15 豊田合成株式会社 Fuel tank
TWI447616B (en) * 2009-06-05 2014-08-01 Hon Hai Prec Ind Co Ltd Written panel
CN109147468A (en) * 2018-09-14 2019-01-04 邹玉平 A kind of writing system and calligraphy practice method
CN113243666A (en) * 2021-05-21 2021-08-13 中国石油大学胜利学院 But intelligent digital media art design is with remote control platform

Citations (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4464118A (en) 1980-06-19 1984-08-07 Texas Instruments Incorporated Didactic device to improve penmanship and drawing skills
US4633436A (en) 1983-12-16 1986-12-30 International Business Machines Corp. Real-time rub-out erase for an electronic handwriting facility
US4690644A (en) 1984-12-21 1987-09-01 Flanders Robert G Teaching apparatus, particularly for teaching shorthand
US4793810A (en) 1986-11-19 1988-12-27 Data Entry Systems, Inc. Interactive instructional apparatus and method
US4804328A (en) * 1986-06-26 1989-02-14 Barrabee Kent P Interactive audio-visual teaching method and device
US4911536A (en) * 1986-05-08 1990-03-27 Ditzik Richard J Interactive graphic comunications terminal
US5007085A (en) 1988-10-28 1991-04-09 International Business Machines Corporation Remotely sensed personal stylus
US5100329A (en) 1990-06-22 1992-03-31 Deesen Kenneth C Computer assisted coaching method
US5416610A (en) * 1992-10-27 1995-05-16 Cordata, Inc. Integral computer scanning system
US5501601A (en) * 1993-06-15 1996-03-26 Stuff Co., Ltd. Educational drawing toy with sound-generating function
US5730602A (en) * 1995-04-28 1998-03-24 Penmanship, Inc. Computerized method and apparatus for teaching handwriting
US5874722A (en) 1994-07-19 1999-02-23 Spectra-Physics Scanning Systems, Inc. Compact scanner module mountable to pointing instrument
US5911533A (en) 1997-11-24 1999-06-15 Eastman Kodak Company Microfluidic writing pen
US5999509A (en) * 1997-03-19 1999-12-07 Pioneer Electronic Corporation Optical pickup device with two independent light beams and an integrated prism for providing return light with astigmatism
US6088025A (en) * 1995-08-24 2000-07-11 Matsushita Electric Industrial Co., Ltd. Terminal device with built-in image sensor
US6215901B1 (en) 1997-03-07 2001-04-10 Mark H. Schwartz Pen based computer handwriting instruction
US6373492B1 (en) 1995-12-26 2002-04-16 Imax Corporation Computer-assisted animation construction system and method and user interface
US20020160342A1 (en) 2001-04-26 2002-10-31 Felix Castro Teaching method and device
US6491225B1 (en) * 1989-10-30 2002-12-10 Symbol Technologies, Inc. Electro-optical reader with electronic stylus
US6515654B1 (en) 2000-10-13 2003-02-04 Taiwan Regular Electronics Touch-type pointing device with wireless input capability
US6572014B1 (en) * 1997-04-16 2003-06-03 Francis Lambert Method and apparatus for non-intrusive biometric capture
US6928462B2 (en) * 2001-03-06 2005-08-09 Hewlett-Packard Development Company, L.P. System and method for distributed processing of non-processable elements of a document to be rendered on a client

Patent Citations (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4464118A (en) 1980-06-19 1984-08-07 Texas Instruments Incorporated Didactic device to improve penmanship and drawing skills
US4633436A (en) 1983-12-16 1986-12-30 International Business Machines Corp. Real-time rub-out erase for an electronic handwriting facility
US4690644A (en) 1984-12-21 1987-09-01 Flanders Robert G Teaching apparatus, particularly for teaching shorthand
US4911536A (en) * 1986-05-08 1990-03-27 Ditzik Richard J Interactive graphic comunications terminal
US4804328A (en) * 1986-06-26 1989-02-14 Barrabee Kent P Interactive audio-visual teaching method and device
US4793810A (en) 1986-11-19 1988-12-27 Data Entry Systems, Inc. Interactive instructional apparatus and method
US5007085A (en) 1988-10-28 1991-04-09 International Business Machines Corporation Remotely sensed personal stylus
US6491225B1 (en) * 1989-10-30 2002-12-10 Symbol Technologies, Inc. Electro-optical reader with electronic stylus
US5100329A (en) 1990-06-22 1992-03-31 Deesen Kenneth C Computer assisted coaching method
US5416610A (en) * 1992-10-27 1995-05-16 Cordata, Inc. Integral computer scanning system
US5501601A (en) * 1993-06-15 1996-03-26 Stuff Co., Ltd. Educational drawing toy with sound-generating function
US5874722A (en) 1994-07-19 1999-02-23 Spectra-Physics Scanning Systems, Inc. Compact scanner module mountable to pointing instrument
US5730602A (en) * 1995-04-28 1998-03-24 Penmanship, Inc. Computerized method and apparatus for teaching handwriting
US6088025A (en) * 1995-08-24 2000-07-11 Matsushita Electric Industrial Co., Ltd. Terminal device with built-in image sensor
US6373492B1 (en) 1995-12-26 2002-04-16 Imax Corporation Computer-assisted animation construction system and method and user interface
US6215901B1 (en) 1997-03-07 2001-04-10 Mark H. Schwartz Pen based computer handwriting instruction
US5999509A (en) * 1997-03-19 1999-12-07 Pioneer Electronic Corporation Optical pickup device with two independent light beams and an integrated prism for providing return light with astigmatism
US6572014B1 (en) * 1997-04-16 2003-06-03 Francis Lambert Method and apparatus for non-intrusive biometric capture
US5911533A (en) 1997-11-24 1999-06-15 Eastman Kodak Company Microfluidic writing pen
US6515654B1 (en) 2000-10-13 2003-02-04 Taiwan Regular Electronics Touch-type pointing device with wireless input capability
US6928462B2 (en) * 2001-03-06 2005-08-09 Hewlett-Packard Development Company, L.P. System and method for distributed processing of non-processable elements of a document to be rendered on a client
US20020160342A1 (en) 2001-04-26 2002-10-31 Felix Castro Teaching method and device

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030117378A1 (en) * 2001-12-21 2003-06-26 International Business Machines Corporation Device and system for retrieving and displaying handwritten annotations
US10664153B2 (en) 2001-12-21 2020-05-26 International Business Machines Corporation Device and system for retrieving and displaying handwritten annotations
US8196041B2 (en) 2003-06-26 2012-06-05 International Business Machines Corporation Method and system for processing information relating to active regions of a page of physical document
US20050266386A1 (en) * 2004-05-28 2005-12-01 Leapfrog Enterprises, Inc. Print media apparatus including stroke recognition
US20110313771A1 (en) * 2005-11-01 2011-12-22 Leapfrog Enterprises, Inc. Method and device for audibly instructing a user to interact with a function
US20100245291A1 (en) * 2009-03-26 2010-09-30 Fuji Xerox Co., Ltd. Writing apparatus
US8988361B2 (en) * 2009-03-26 2015-03-24 Fuji Xerox Co., Ltd. Writing apparatus
US9335839B1 (en) 2014-04-08 2016-05-10 Clive Lynch Graphic artistic tablet computer
US20180297392A1 (en) * 2015-10-15 2018-10-18 Sanford, L.P. Generating mechanically rendered objects from digital input

Also Published As

Publication number Publication date
US20030169237A1 (en) 2003-09-11

Similar Documents

Publication Publication Date Title
RU2536667C2 (en) Handwritten input/output system, handwritten input sheet, information input system and sheet facilitating information input
US20060077184A1 (en) Methods and devices for retrieving and using information stored as a pattern on a surface
RU2673275C2 (en) Method of reproducing information, a method of information input/output, a playback device information, a portable information input/output device and a electronic toy where a point raster is used
US6567078B2 (en) Handwriting communication system and handwriting input device used therein
US8773398B2 (en) Data input system
US7116316B2 (en) Audible and visual effects as a result of adaptive tablet scanning
US20130093733A1 (en) Handwriting input board and information processing system using handwriting input board
KR20070047198A (en) A mehod and device for associating a user writing with a user-writable element
JP2003508831A (en) System and apparatus for electronic recording of handwritten information
US20090248960A1 (en) Methods and systems for creating and using virtual flash cards
US8139048B2 (en) Method of raising resolution in locating on a micro dotmap
KR20010102224A (en) Apparatus and system for reproduction of handwritten input
GB2388239A (en) Hand-writing practising system
CN110853424A (en) Voice learning method, device and system with visual recognition
KR102446679B1 (en) Electronic board of improved hand-writing recognition on infrared touch panel and operating method thereof
US7671269B1 (en) Methods and systems for graphical actuation of a velocity and directionally sensitive sound generation application
CN110134261A (en) It is a kind of to restore the electronic pen really write
CN210573686U (en) Erasing device for white board covered with dot matrix codes
RU2349956C2 (en) Method of information playback, method of information input/output, device for information playback, portable device of information input/output and electronic toy in with dot raster
US20050134927A1 (en) Data management system and method
WO2012002915A1 (en) Computer integrated presentation device
JP2007272710A (en) Handwriting input system
CN211181137U (en) Multifunctional language learning terminal with visual recognition and handwriting board
CN210573714U (en) Erasing device of electronic whiteboard
CN210348429U (en) Blackboard eraser structure of electronic whiteboard

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTEL CORPORATION, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JELINEK, LENKA M.;BROOKE, TIMOTHY L.;BROWN, FRANK T.;AND OTHERS;REEL/FRAME:012686/0915;SIGNING DATES FROM 20020128 TO 20020220

CC Certificate of correction
FPAY Fee payment

Year of fee payment: 4

FPAY Fee payment

Year of fee payment: 8

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.)

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20181003