AU9088191A - Talking video games - Google Patents
Talking video gamesInfo
- Publication number
- AU9088191A AU9088191A AU90881/91A AU9088191A AU9088191A AU 9088191 A AU9088191 A AU 9088191A AU 90881/91 A AU90881/91 A AU 90881/91A AU 9088191 A AU9088191 A AU 9088191A AU 9088191 A AU9088191 A AU 9088191A
- Authority
- AU
- Australia
- Prior art keywords
- voice
- talking
- character
- voice sounds
- animated
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000001755 vocal effect Effects 0.000 claims description 36
- 238000000034 method Methods 0.000 claims description 35
- 230000014509 gene expression Effects 0.000 claims description 30
- 230000009471 action Effects 0.000 claims description 25
- 230000003111 delayed effect Effects 0.000 claims 1
- 230000003993 interaction Effects 0.000 claims 1
- 230000004044 response Effects 0.000 description 22
- 230000008921 facial expression Effects 0.000 description 3
- 239000004973 liquid crystal related substance Substances 0.000 description 3
- 241000792914 Valeriana Species 0.000 description 2
- 210000001260 vocal cord Anatomy 0.000 description 2
- 241001465754 Metazoa Species 0.000 description 1
- 230000002860 competitive effect Effects 0.000 description 1
- 230000002079 cooperative effect Effects 0.000 description 1
- 230000000875 corresponding effect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 230000002996 emotional effect Effects 0.000 description 1
- 230000001815 facial effect Effects 0.000 description 1
- 230000009474 immediate action Effects 0.000 description 1
- 230000002045 lasting effect Effects 0.000 description 1
- 239000012528 membrane Substances 0.000 description 1
- 230000036961 partial effect Effects 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 238000003756 stirring Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B7/00—Electrically-operated teaching apparatus or devices working with questions and answers
- G09B7/02—Electrically-operated teaching apparatus or devices working with questions and answers of the type wherein the student is expected to construct an answer to the question which is presented or wherein the machine gives an answer to the question presented by a student
- G09B7/04—Electrically-operated teaching apparatus or devices working with questions and answers of the type wherein the student is expected to construct an answer to the question which is presented or wherein the machine gives an answer to the question presented by a student characterised by modifying the teaching programme in response to a wrong answer, e.g. repeating the question, supplying a further explanation
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/40—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
- A63F13/42—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
- A63F13/424—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle involving acoustic input signals, e.g. by using the results of pitch or rhythm extraction or voice recognition
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/45—Controlling the progress of the video game
- A63F13/47—Controlling the progress of the video game involving branching, e.g. choosing one of several possible scenarios at a given point in time
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/80—Special adaptations for executing a specific game genre or game mode
- A63F13/822—Strategy games; Role-playing games
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/80—Special adaptations for executing a specific game genre or game mode
- A63F13/843—Special adaptations for executing a specific game genre or game mode involving concurrently two or more players on the same game device, e.g. requiring the use of a plurality of controllers or of a specific view of game data for each player
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/90—Constructional details or arrangements of video game devices not provided for in groups A63F13/20 or A63F13/25, e.g. housing, wiring, connections or cabinets
- A63F13/92—Video game devices specially adapted to be hand-held while playing
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B17/00—Guiding record carriers not specifically of filamentary or web form, or of supports therefor
- G11B17/005—Programmed access to indexed parts of tracks of operating discs, by guiding the disc
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B19/00—Driving, starting, stopping record carriers not specifically of filamentary or web form, or of supports therefor; Control thereof; Control of operating function ; Driving both disc and head
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/102—Programmed access in sequence to addressed parts of tracks of operating record carriers
- G11B27/105—Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/60—Methods for processing data by generating or executing the game program
- A63F2300/63—Methods for processing data by generating or executing the game program for controlling the execution of the game in time
- A63F2300/632—Methods for processing data by generating or executing the game program for controlling the execution of the game in time by branching, e.g. choosing one of several possible story developments at a given point in time
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/80—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game specially adapted for executing a specific type of game
- A63F2300/807—Role playing or strategy games
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/80—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game specially adapted for executing a specific type of game
- A63F2300/8088—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game specially adapted for executing a specific type of game involving concurrently several players in a non-networked game, e.g. on the same game console
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
- G11B2220/21—Disc-shaped record carriers characterised in that the disc is of read-only, rewritable, or recordable type
- G11B2220/213—Read-only discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
- G11B2220/25—Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
- G11B2220/2537—Optical discs
- G11B2220/2545—CDs
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Educational Technology (AREA)
- Educational Administration (AREA)
- Business, Economics & Management (AREA)
- General Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Human Computer Interaction (AREA)
- Processing Or Creating Images (AREA)
- Telephone Function (AREA)
- Closed-Circuit Television Systems (AREA)
Description
TALKING VIDEO GAMES
Technical Field
This invention relates to video games, animated cartoons, and picture/sound synchronization.
Background Art
We are all born with a desire to talk and to be talked to. Listening to other people talk and thereby sharing their emotional experiences is also a desire we are born with and this desire has been partly satisfied by motion picture film and television in which voice sounds are now essential. Until recently, voice sounds were seldom used in video games or were used in an optional manner. People have a desire to participate in voice conversations with other people, but this desire is not satisfied by current video games.
Adding voice sounds to conventional video games is not enough to simulate a face to face voice conversation. A talking video game called Thayer's Quest was attempted in 1984 and was played from an analog laser-readable video disc. One of the reasons for the commercial failure of Thayer's Quest was that each spoken sentence was programmed to accompany only one sequence of video frames. Since the video was not compressed, the maximum amount of play time was limited to about half an hour which was further reduced to a fraction of that by the branching story. Hence, only a few minutes of voice sounds were actually heard during the game. Whenever a human player saw a certain video character, the character usually spoke the same words. This greatly reduced the entertainment value of the game. Another consequence of programming the audio and video to be inseparable, was that branching scenes were not distinguished from branching dialog. Talking video games will be much more entertaining if each scene has a rich variety of possible dialog sequences.
Talking video games are disclosed in my U.S. Patents Nos. 4,305,131; 4,333,152; 4,445,187 and 4,569,026. These patents show only one talking animated character and one human player, although multiple characters and players were suggested in the texts. These patents also show push-button controllers that display alternative words to say to animated characters.
It is common practice in the video game art for stories to branch. It is also common practice for digital data of animated characters to be stored separately from background scenery and to generate each frame of an animated picture from both the background data and the character data to minimize the number of stored images. It is well known for human players to input choices using any of a variety of input devices such as push buttons, rotatable knobs, pressure sensitive membrane, prox.imity sensitive pads or screen overlay, light pen, light sensitive gun, joy stick, mouse, track ball, moving a cursor or crosshairs or scrolling through highlighted options, speech recognition, etc.
In the prior art, each choice by the human can be immediately followed by a synthesized voice or digitized voice recording that speaks the words selected by the human player, so the human will quickly adjust to the fact that the spoken words he hears for his side of the dialog are initiated by his fingers rather than his vocal cords.
The prior art also includes methods for generative video images of moving lips and facial expressions on a talking head or other animated character. See for example, U.S. Patent No. 4,884,972 issued December 5, 1989 to Elon Gasper who contemplates use in video games.
Drawing Figures
FIG. 1 illustrates an animated cartoon talking game with two human players who take turns selecting words to say to one or two animated characters on a TV screen who then respond to the words.
FIG. 2 illustrates an animated talking game with one human player and two animated cartoon characters on a TV screen who talk to the human and to each other.
FIG. 3 is an animated cartoon illustrating a kissing scene in which one character's unspoken thoughts are sounded.
FIG. 4 is an animated cartoon illustrating a ledge-hanging scene in which a talking cartoon character requires immediate action by the human player. FIG. 5 is an animated sequence illustrating branching dialog in which two cartoon characters respond to sentences selected by a human player by talking to the human player and to each other.
FIG. 6 is an animated sequence illustrating branching dialog in which one animated character responds to a sentence selected by a human player and refers to a second character (not shown) .
FIG. 7 is an animated sequence showing hands simulating the hands of a human player performing an action.
FIG. 8 is a flow diagram illustrating scene branching and dialog branching within each scene.
FIG. 9 is a pictorial view of a video game system having a CD-ROM disc drive connected to a television or video monitor with auxilary devices for displaying prompting words. Description of Preferred Embodiments
Characters in role-playing video games are of two kinds: player-controlled characters and non-player characters. A player-controlled character is a human player's animated counterpart and does what the human player chooses to have him do. Non-player characters are not directly controlled by a human player, but can be indirectly influenced by a human player, either by responding to an action selected by the human player or by responding to what a player-controlled character does or says. Non-player characters should be programmed to say
any different alternative things depending on what a player or player-controlled character has just "said". A scene may repeat several times with a non-player character saying something different each time and the alternative responses for each human player being different each time a scene is repeated.
Hand-held controllers are preferable to displaying menus on the video screen, because in two-person games the menu displayed to one player may be different from the other player's menu. Part of the fun comes from a player not knowing what choices the other player has. Also, video screen menus disrupt the illusion that a player is talking with a real video character. Hence, the preferred embodiment of my invention makes use of hand-held menu display controllers with a push-button or equivalent for each menu item. The invention also makes use of echo voices that repeat the selected words so that other human players will know what is being said.
Referring to FIG. 1, in a typical embodiment of this invention a video game system displays on a video screen
11 an animated picture sequence to two human game players. Human players 10 and 12 take turns selecting a phrase or sentence to "say" to a character or characters on a video screen who then talk back responsively. In this example it is player 10's turn. Player 12's hand-held controller is blank, indicating that she cannot say anything to a character at this moment. The video screen shows two animated characters traveling in a boat or other vehicle. The characters are non-player characters that talk directly to human players in this example. Each human player holds a hand-held controller with three push buttons next to a liquid-crystal display 13, shown enlarged in FIG. 1 for clarity.
As the boat scene begins, the video game system displays two or three or more alternative responses on display 13 or other display device. While player 10 is deciding which button to press, a linking picture sequence
(not shown) continues to show the same boat scene with one character looking back from the boat as if looking at player 10 standing on the dock. Player 10 selects one of the displayed responses (in this example "WHERE .ARE YOU GOING?") which is then echoed as voice sounds or is displayed as sub-titles on the video screen so that player 12 will know what player 10 has just said to an animated character. This character 18 answers responsively with voice sounds: "TO THE ISLAND." after which a second character 17 responds with "WE'LL BE THERE .AN HOUR." in response to what character 18 just said.
In the preferred embodiment, the game system generates a voice to echo the words selected by player 10 so that player 12 hears what player 10 has "said" to the animated character. When it is player 12's turn, player 12's hand-held controller will then display alternative words (not shown in FIG. 1) that she can "say" in response to what animated character 17 or 18 has said. Thus a three-way or four-way dialog can be s.imulated. Before each game begins, display 13 may show identifying information for each player-controlled character so that each human player may select which character he or she wants to play. Alternatively, each player-controlled character may be shown on video screen 11 and the question asked "Who wants to play this character?" for each character shown. The game system then records which hand-held controller responds to the question, so later that character's words will be shown on display 13 only for the player who is playing that character (in this example player 10). A human player may also choose to play a player- controlled character that need not always be shown on the video screen, but who may be off-screen (for example, standing on the dock in FIG. 1), and can carry on a dialog with on-screen characters. Referring to FIG. 2, in another embodiment of the invention, a video game system displays on a video screen 11 an animated picture sequence having two or more
animated talking characters 17 and 18. In this example, character 17 is a player-controlled character that human player 12 controls. Player 12 plays the role of character 17 and can talk to character 18 and other characters in the game through character 17. Player 12 holds a hand-held controller with three push buttons next to a liquid-crystal display 13, shown enlarged in FIG. 2 for clarity. The game system displays three alternative responses on display 13. Player 12 selects one of the displayed responses (in this example "KISS .ME AGAIN") with push button 14. The echo voice sounds 15 for character 17 then repeat the words selected from display 13. The words are directed at character 18 whose voice sounds 16 then respond to character 17. The two animated characters may respond in reverse sequence, that is, the non-player character 18 may say his line first so that the player-controlled character 17 can respond as ordered by human player 12. For example, after player 12 selects "KISS ME AGAIN", character 18 may say "WHAT DO YOU WANT TO DO?" a prerequisite sentence that is not one of the displayed alternatives, then character 17 can respond with "KISS ME AGAIN" which is responsive to what character 18 has just said.
Echo voices or sub-titles may also be used to express unspoken thoughts or the thoughts of non-speaking beings such as babies or animals or inanimate objects. Cloud balloon 19 in FIG. 2 represents an unspoken thought of character 18 which is sounded (with mouth not moving) in response to spoken sentence 15 of character 17. Voice sounds for unspoken thoughts may be electronically altered to indicate to players that a voice is not a normal spoken voice. For example, unspoken thoughts can be given a hollow or tremulous sound or a whispering sound by electronically or digitally editing voice sound data before converting to audio.
Referring to FIG. 3, when characters 17 and 18 perform a competitive or cooperative activity such as
kissing, one of the characters may speak (with moving mouth) or think unspoken thoughts (sounded with unmoving mouth) as in cloud balloon 29, responsive to the action being shown or to what was said or done in the prior sequence shown in FIG. 2 or in response to a phrase selected from display 13.
Referring to FIG. 4, in another embodiment of this invention, a video game system generates an animated picture sequence showing a character hanging by his fingers from a ledge on the outside of a building. His friends on the roof have thrown him a rope which is not long enough to reach the FIG. 4 character. This ledge-hanging situation is followed by the next scene shown in FIG. 5. Referring to FIG. 5, picture sequences 21, 25, and 28 are parts of one roof scene in which two talking characters discuss how to rescue the character in FIG. 4. One or two human players participate in the conversation by "saying" words or phrases or sentences to the animated characters who then answer responsively and ask questions or make remarks that lead to the next input by a player. The alternatives shown on display 22 are suggested solutions to the problem posed in sentence 20. When the human player 10 presses button 14 next to "CALL EMERGENCY", one of the characters responds by asking question 23 to the other character who responds with question 24 directed at the human player. Question 24 is also accompanied by alternative actions 26. When a player presses the button next to "SLIDE DOWN THE ROPE", a character comments on this choice of action with sentence 27 in sequence 28. Thus a simulated verbal dialog can continue through several exchanges of words within the same scene.
Referring to FIG. 6, in another embodiment of this invention a video game system generates animated picture sequence 31 and 32. The picture sequence has one scene showing a burning car in the background and a frightened
woman 17 in the foreground. During part 31 of the sequence the woman's voice 15 says "PLEASE HELP ME1 MY MOTHER IS TRAPPED IN THE CAR1" The game then displays two or more alternative responses. Human player 10 selects one of the displayed responses (such as "PRY THE DOOR OPEN") and presses the corresponding push button 14 or equivalent. While the player is deciding which button to press, a linking picture sequence (not shown) continues to show the same scene with the woman looking anxiously at the player. When the player selects a response, part 32 of the animated sequence continues showing the same burning car scene with the woman's voice 16 answering responsively: "I TRIED TO, BUT I CAN'T GET IT OPEN" Selecting a response by pushing a button 14 can result in a change to a different scene, but in this FIG. 6 example the scene remains the same and only the dialog changes. Each of the three alternative responses on display 13 will result in a different answer by animated character 17. The animated sequences 31 and 32 are generated by the video game system by overlaying three moving pictures: (1) the background sequence showing the burning car with flames continually flickering, (2) a sprite or mosaic of sprites showing the woman character 17 moving in a natural manner against the background and (3) mouth and other facial sprites selected (by table lookup) by the game system approximately lip-synchronized with voice sounds 15 and 16.
Since the player's response time is variable, sequence 31 or 32 with a closed mouth sprite should cycle continually until a button 14 is pressed or until a time limit is reached, at which time a prompting picture sequence with words such as "Hurry1 Hurry!" may be sounded and displayed with moving mouth sprites. The burning car can be any object such as a damaged bus, airplane, boat, or building that provides a danger situation.
Referring to FIG. 7, the game system displays on the video screen an animated sequence showing one or more
handε 36 simulating the hands of a human player performing an action. In this example, animated sequence 33 shows a hand holding pipe 35 or a fire extinguisher or other such tool while the human player whose real hand 10 is illustrated pressing button 14 next to display 13 selects one of the alternative actions shown on display 13. Note that display 13 in FIG. 7 shows alternative actions which would result in scene changes, but display 13 in FIG. 6 shows alternative words or phrases to say to animated character 17 in a common scene. In FIG. 7, pressing button 14 to select "PRY THE DOOR OPEN" results in picture sequence 34 showing hands 36 prying open the car door with pipe 35.
The flowchart shown in FIG. 8 illustrates the distinction between branching dialog and branching scenes. For example, in scene 64 branch point 60 is displayed to the player showing two alternative verbal responses 61 or 62. Either verbal response results in a different answer from an animated character, but not a scene change. However, the alternatives at branch point 63 will result in a scene change, either to scene 65 or to scene 66. Branch point 62 may result in either a scene change to scene 67 or a dialog branch depending on a player's choice. To allow each background scene to be used with different animated characters who can move around against the background scene, the digital animation data for the background scene should be stored separately from the digital animation data for each character. Similarly, to allow each character to say many different sentences without a scene change, the digitized voice data should be independent of the animation data. In the preferred embodiment, animated character video, voice sound sequences and guidance word sequences are generated independently from separately stored digital data. Dialog data that is not used in one scene may be used later in a different scene with the same or different characters.
The voice data may consist of sequences of codes or compressed digital recordings of words, phrases, word segments or phonemes in several distinctive voices so that each character can speak thousands of preprogrammed words or sentences. Similarly, the digital data for each animated character's body may be stored separately from sprite data for moving lips, facial expressions, and gestures, so that each character and its distinctive voice can be lip-synchronized with different mouth movements depending on the dialog. The digital data for each animated character may also combine body images, lips, facial expressions, hand gestures and voice sounds.
Referring to FIG. 9, video game system 42 is connected by cables to a television 11 or video monitor and to one or more hand-held control units 44 and 47 or portable game system 46, each having three push buttons 14 next to a liquid-crystal display 13. Infra-red or radio signals may be used instead of cables. System 42 includes a disk reader that reads digital data from a CD-ROM disk 43, or write-once disk or card or other medium containing digital data from which system 42 generates animated picture sequences, compressed audio for voice and other sounds, synchronizing data, and words to display on units 44, 46 and 47. Portable game system 46 with appropriate ROM program cartridge may substitute for units 44 and 47. Cable 45 connects game system 42 to controllers 44, 46 or 47 and transmits alternative guidance words or other verbal expressions for display on displays 13. Cable 45 also transmits push-button 14 responses to game system 42.
If portable game system 46 has only one or two available push-buttons, selection from among responses shown on display 13 may be done as shown in FIG. 7 by pressing push-button 14 multiple times to position a pointer 49, cursor, or other such indicator next to (or on) a selected sentence.
When a human player presses a button 14 (FIG. 2, 5, 6, or 7) , the game system may generate a voice sound speaking the selected sentence as a substitute for the player's side of the dialog. The animated character,then "responds." as if the generated voice sounds had been spoken by the human player. Because the player selects the words which are actually sounded, he will quickly adjust to the fact that the spoken words he hears for his side of the dialog are initiated by his fingers rather than his vocal cords. This echo voice is important for games with multiple human players so that each player will hear what each of the other players has "said" to on-screen characters. Pushing a button 14 selects both a simulated verbal response to the previous words spoken by an animated character and also selects a new dialog sequence that corresponds to the simulated verbal response shown on display 13. The selected dialog sequence includes the face and voice of the animated character speaking words which are responsive to the player's selected verbal response.
Alternatively, sub-titles such as illustrated on video screen 11 in FIG. 1 may be used instead of echo voices and be displayed on a video screen or on a hand-held controller as a substitute for the player's side of the dialog. Digitized live action picture sequences are equivalent to animated cartoon picture sequences for this invention. The term "verbal expression" means any word, words, phrase, sentence, question, expletive, curse, keyword, combination of keywords, symbol, or any meaningful human voice sound such as "huh?" or "hmmm" or laughter or scream. The word kissing is used herein to mean simulated touching of one animated character's mouth to another animated character's mouth or other body part. No video game yet satisfies the basic human desire for people to speak and respond to us as individuals.
Although the telephone provides a partial satisfaction of our desire to talk with other people, it is necessarily
limited to living people who are willing to talk with us. Historical and imaginary people cannot talk with us and famous living people do not want to. Hence, there is is a strong but unfulfilled human desire waiting to be satisfied by new technology. Talking animated characters will change the nature of video games as dramatically as talking pictures changed silent film. Talking video games will let human players talk with images of famous people (living or historical or imaginary) and with animal-like characters, and participate in simulated adventures and dramas and conversations with interesting characters and groups of characters who will talk to each player responsively. Talking video games that stir human emotions like dramatic films will have lasting appeal, because they will satisfy a basic human desire, the desire to talk with other people.
Although 1 have described the preferred embodiments of my invention with a degree of particularity, it is understood that the present disclosure has been made only by way of example and that equivalent steps and components may be substituted and design details changed without departing from the spirit and scope of my invention.
Claims (33)
1. A video game method of simulating voice conversations between at least two talking animated characters and one or more human players of the video game, comprising the steps of: generating animated pictures showing first and second talking characters having voices that are noticeably different; generating first voice sounds in the voice of one of said talking characters; displaying on a control device held by a first human player a first plurality of alternative verbal expressions responding to said first voice sounds; receiving from said control device an indication of which verbal expression is selected by said human player from said plurality thereof; generating second voice sounds responding to said selected verbal expression in the voice of said first talking character; generating third voice sounds in the voice of said second talking character responding to said second voice sounds; and displaying on a control device held by a second human player a second plurality of alternative verbal expressions responding to said third voice sounds.
2. The method of claim 1 wherein said first and second human players are the same person holding the same control device.
3. The method of claim 1 wherein said talking animated characters are animated cartoons.
4. The method of claim 1 wherein the animated pictures of said talking animated characters are generated from digitized images of living people.
5. The method of claim 1 wherein at least some of said voice sounds are altered to indicate that the altered voice sounds represent unspoken thoughts of , one of said characters.
6. The method of claim 1 wherein the step of generating said first voice sounds is delayed and performed between said receiving step and the generating of said second voice sounds.
7. The method of claim 1 wherein said selected verbal expression describes a selected action and said second voice sounds talk about the feasability of the selected action.
8. The method of claim 1 wherein said generated animated pictures show one of said talking characters in a dangerous situation and said first plurality of alternative verbal expressions describe alternative actions to rescue the endangered character from the dangerous situation.
9. The method of claim 1 wherein said second voice sounds say the words of said selected verbal expression.
10. The method of claim 9 wherein said second voice sounds say words in a voice noticeably different than the voices of said first and second talking characters.
11. The method of cla.im 1 wherein said generated animated pictures show one of said characters hugging or kissing the other character in accordance with any of said voice sounds.
12. The method of claim 1 wherein said animated pictures show one of said talking characters with a closed unmoving mouth during at least some of the character's voice sounds to indicate that the voice sounds represent unspoken thoughts of the talking character.
13. A video game method of simulating voice conversations between a talking animated character and at least two human players of the video game, comprising the steps of: generating animated pictures showing a talking an.imated character; displaying a first plurality of alternative verbal expressions on a first control device held by a first human player; receiving from said first control device an indication of which first verbal expression is selected by said first human player from said first plurality of verbal expressions; electronically communicating said first selected verbal expression to a second human player; generating voice sounds responding to said first selected verbal expression in the voice of said talking animated character; displaying on a second control device held by said second human player a second plurality of alternative verbal expressions responding to said voice sounds. receiving from said second control device an indication of which second verbal expression is selected by said second human player from said second plurality of verbal expressions; and electronically communicating said second selected verbal expression to said first human player.
14. The method of claim 13 wherein said first control device indicates which of the human players should respond to the plurality of verbal expressions.
15. The method of claim 13 wherein said first selected verbal expression is communicated to said second human player by displaying the first selected verbal expression.
16. The method of claim 13 wherein said first selected verbal expression is communicated to said second human player by generating voice sounds expressing the first selected verbal expression.
17. The method of claim 13 wherein said talking animated character is an animated cartoon.
18. The method of claim 13 wherein the animated pictures of said talking animated character are generated from digitized images of living people.
19. The method of claim 13 wherein said voice sounds say the words of said selected verbal expression.
20. The method of claim 13 wherein at least some of said voice sounds are noticably different than the voice of said talking character to represent unspoken thoughts of the talking character.
21. The method of claim 13 wherein said animated pictures show said talking character with a closed unmoving mouth during at least some of said voice sounds to indicate that said voice sounds represent unspoken thoughts of said talking character.
22. A video game method of simulating interaction between at least two talking animated characters and one or more human players of the video game, comprising the steps of: generating animated pictures showing first and second talking characters having voices that are noticeably different and encountering a situation requiring a choice of one action from a plurality of alternative actions; displaying on a control device held by a first human player words describing each action in said plurality of alternative actions; receiving from said control device an indication of which selected action in said plurality thereof is chosen by said first human player; generating first voice sounds in the voice of said first talking character repeating the displayed words describing said selected action; generating second voice sounds in the voice of said second talking character responding to said first voice sounds; and displaying on a control device held by a second human player words describing a second plurality of alternative actions responding to said second voice sounds.
23. The method of claim 22 wherein one of said characters is shown in a dangerous situation and said voice sounds describe methods of helping the endangered character.
24. The method of claim 23 wherein some of said voice sounds ask for guidance in helping said endangered character.
25. The method of claim 22 wherein said animated pictures show one of said talking characters or a portion thereof performing said selected action in accordance with said voice sounds.
26. The method of claim 22 wherein some of said voice sounds comment on said selected action;
27. The method of claim 22 wherein some of said voice sounds refer to a character affected by said selected action;
28. The method of claim 22 wherein said generated animated pictures show one of said characters hugging or kissing the other character in accordance with said first or second voice sounds.
29. A video game method of simulating voice conversations between at least two talking animated characters and one or more human players of the video game, comprising the steps of: generating animated pictures showing first and second talking characters having voices that are noticeably different; generating first voice sounds in the voice of one of said talking characters; displaying to a first human player a first plurality of alternative verbal expressions responding to said first voice sounds; receiving from said first human player an indication of which verbal expression is selected by the player from said plurality thereof; generating second voice sounds responding to said selected verbal expression in the voice of said first talking character; generating third voice sounds in the voice of said second talking character responding to said second voice sounds; and displaying to a second human player a second plurality of alternative verbal expressions responding to said third voice sounds.
30. The method of claim 29 wherein said first and second human players are the same person.
31. The method of claim 29 wherein said second voice sounds say the words of said selected verbal expression.
32. The method of claim 29 wherein said selected verbal expression describes a selected action and one of said characters performs the selected action.
33. The method of claim 32 wherein said generated animated pictures show one of said characters hugging or kissing the other character in accordance said selected action.
Applications Claiming Priority (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US61484390A | 1990-11-14 | 1990-11-14 | |
| US614843 | 1990-11-14 | ||
| PCT/US1991/007266 WO1992008531A1 (en) | 1990-11-14 | 1991-10-07 | Talking video games |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| AU9088191A true AU9088191A (en) | 1992-06-11 |
| AU652209B2 AU652209B2 (en) | 1994-08-18 |
Family
ID=24462942
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| AU90881/91A Ceased AU652209B2 (en) | 1990-11-14 | 1991-10-07 | Talking video games |
Country Status (6)
| Country | Link |
|---|---|
| US (1) | US5358259A (en) |
| EP (1) | EP0557444A1 (en) |
| JP (1) | JP2784409B2 (en) |
| AU (1) | AU652209B2 (en) |
| CA (1) | CA2095820C (en) |
| WO (1) | WO1992008531A1 (en) |
Cited By (12)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6001015A (en) * | 1995-10-09 | 1999-12-14 | Nintendo Co., Ltd. | Operation controlling device and video processing system used therewith |
| US6007428A (en) * | 1995-10-09 | 1999-12-28 | Nintendo Co., Ltd. | Operation controlling device and video processing system used therewith |
| US6022274A (en) * | 1995-11-22 | 2000-02-08 | Nintendo Co., Ltd. | Video game system using memory module |
| US6102803A (en) | 1995-05-10 | 2000-08-15 | Nintendo Co., Ltd. | Operating device with analog joystick |
| US6139433A (en) | 1995-11-22 | 2000-10-31 | Nintendo Co., Ltd. | Video game system and method with enhanced three-dimensional character and background control due to environmental conditions |
| US6139434A (en) | 1996-09-24 | 2000-10-31 | Nintendo Co., Ltd. | Three-dimensional image processing apparatus with enhanced automatic and user point of view control |
| US6155926A (en) * | 1995-11-22 | 2000-12-05 | Nintendo Co., Ltd. | Video game system and method with enhanced three-dimensional character and background control |
| US6241611B1 (en) | 1995-05-10 | 2001-06-05 | Nintendo Co., Ltd. | Function expansion device and operating device using the function expansion device |
| US6267673B1 (en) | 1996-09-20 | 2001-07-31 | Nintendo Co., Ltd. | Video game system with state of next world dependent upon manner of entry from previous world via a portal |
| US6421056B1 (en) | 1995-10-09 | 2002-07-16 | Nintendo Co., Ltd. | Three-dimensional image processing apparatus |
| US6679776B1 (en) | 1997-07-17 | 2004-01-20 | Nintendo Co., Ltd. | Video game system |
| US7126584B1 (en) | 1995-10-09 | 2006-10-24 | Nintendo Co., Ltd. | Operating device and image processing system using same |
Families Citing this family (122)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US5393073A (en) * | 1990-11-14 | 1995-02-28 | Best; Robert M. | Talking video games |
| JPH05111579A (en) * | 1991-09-09 | 1993-05-07 | Robert M Best | Method for playing video game |
| US6906700B1 (en) | 1992-03-05 | 2005-06-14 | Anascape | 3D controller with vibration |
| US6222525B1 (en) | 1992-03-05 | 2001-04-24 | Brad A. Armstrong | Image controllers with sheet connected sensors |
| US8821276B2 (en) | 1992-05-22 | 2014-09-02 | Bassilic Technologies Llc | Image integration, mapping and linking system and methodology |
| US5553864A (en) | 1992-05-22 | 1996-09-10 | Sitrick; David H. | User image integration into audiovisual presentation system and methodology |
| DE4222110C2 (en) * | 1992-07-06 | 1998-08-27 | Nsm Ag | Program-controlled entertainment and play equipment |
| US5465384A (en) * | 1992-11-25 | 1995-11-07 | Actifilm, Inc. | Automatic polling and display interactive entertainment system |
| US5603507A (en) * | 1994-04-22 | 1997-02-18 | Hasbro, Inc. | Method of input selection in an electronic game system |
| US5999172A (en) * | 1994-06-22 | 1999-12-07 | Roach; Richard Gregory | Multimedia techniques |
| US5692212A (en) * | 1994-06-22 | 1997-11-25 | Roach; Richard Gregory | Interactive multimedia movies and techniques |
| US5682469A (en) * | 1994-07-08 | 1997-10-28 | Microsoft Corporation | Software platform having a real world interface with animated characters |
| US5604855A (en) * | 1994-09-28 | 1997-02-18 | Crawford; Christopher C. | Computer story generation system and method using network of re-usable substories |
| JP3839501B2 (en) * | 1994-10-13 | 2006-11-01 | 株式会社スクウェア・エニックス | VIDEO GAME DEVICE, CONTROL METHOD AND CONTROL DEVICE THEREOF, AND MEMORY CARTRIDGE FOR VIDEO GAME |
| US6232965B1 (en) * | 1994-11-30 | 2001-05-15 | California Institute Of Technology | Method and apparatus for synthesizing realistic animations of a human speaking using a computer |
| US6010405A (en) * | 1994-12-30 | 2000-01-04 | Sega Enterprises, Ltd. | Videogame system for creating simulated comic book game |
| TW266277B (en) * | 1994-12-31 | 1995-12-21 | Sega Of America Inc | Videogame system and methods for enhanced processing and display of graphical character elements |
| US5607356A (en) * | 1995-05-10 | 1997-03-04 | Atari Corporation | Interactive game film |
| JP3274040B2 (en) * | 1995-05-26 | 2002-04-15 | 株式会社タイトー | A game console that can change the accent of the audio output of the screen display sentence including the name |
| JP3274041B2 (en) * | 1995-05-26 | 2002-04-15 | 株式会社タイトー | Game consoles that can change the tone of audio output |
| US5890123A (en) * | 1995-06-05 | 1999-03-30 | Lucent Technologies, Inc. | System and method for voice controlled video screen display |
| US5603658A (en) * | 1995-06-27 | 1997-02-18 | Cohen; Justin R. | Computer play toy for infants and very young children |
| US5556339A (en) * | 1995-06-27 | 1996-09-17 | Cohen; Justin R. | Computer picture toy for infants and very young children |
| GB9517806D0 (en) * | 1995-08-31 | 1995-11-01 | Philips Electronics Uk Ltd | Information handling for interactive apparatus |
| GB9517807D0 (en) * | 1995-08-31 | 1995-11-01 | Philips Electronics Uk Ltd | Interactive entertainment attribute setting |
| GB9517808D0 (en) * | 1995-08-31 | 1995-11-01 | Philips Electronics Uk Ltd | Interactive entertainment personalisation |
| GB9517788D0 (en) * | 1995-08-31 | 1995-11-01 | Philips Electronics Uk Ltd | Interactive entertainment apparatus |
| GB9517789D0 (en) * | 1995-08-31 | 1995-11-01 | Philips Electronics Uk Ltd | Interactive entertainment content control |
| JP3524247B2 (en) * | 1995-10-09 | 2004-05-10 | 任天堂株式会社 | Game machine and game machine system using the same |
| TW355774B (en) * | 1995-10-12 | 1999-04-11 | Konami Co Ltd | TV game device, its loading support method and control method |
| CN1109960C (en) | 1995-11-10 | 2003-05-28 | 任天堂株式会社 | Joystick apparatus |
| US5893720A (en) * | 1996-03-25 | 1999-04-13 | Cohen; Hannah R. | Development language system for infants |
| EP0834110A1 (en) * | 1996-04-03 | 1998-04-08 | Koninklijke Philips Electronics N.V. | A hand-held control unit for controlling a display screen-oriented computer game, and a display screen-oriented computer game having one or more such control units |
| JP3739800B2 (en) * | 1996-07-01 | 2006-01-25 | 株式会社セガ | Video game apparatus and medium storing video game program |
| US8674932B2 (en) | 1996-07-05 | 2014-03-18 | Anascape, Ltd. | Image controller |
| US6601103B1 (en) * | 1996-08-22 | 2003-07-29 | Intel Corporation | Method and apparatus for providing personalized supplemental programming |
| JP3012560B2 (en) * | 1997-06-25 | 2000-02-21 | 日本電気ソフトウェア株式会社 | Computer-based electronic dialogue method, computer-to-computer electronic dialogue device, and computer-readable recording medium recording computer-based electronic dialogue program |
| JPH1176609A (en) * | 1997-09-04 | 1999-03-23 | Konami Co Ltd | Method for diagnosing affinity with game character, video game device and record medium |
| JPH1176608A (en) * | 1997-09-04 | 1999-03-23 | Konami Co Ltd | Game development control method reflecting character of player, video game device and record medium |
| USD397370S (en) | 1997-09-19 | 1998-08-25 | Hohek Microelectronics, Inc. | Electronic game display panel |
| DE69838734T2 (en) * | 1997-11-25 | 2008-10-30 | Kabushiki Kaisha Sega Doing Business As Sega Corp. | IMAGING DEVICE |
| JP3125006B2 (en) * | 1998-04-07 | 2001-01-15 | コナミ株式会社 | Character image display control method and apparatus, recording medium |
| JP3936749B2 (en) * | 1998-04-16 | 2007-06-27 | クリエイター・リミテッド | Interactive toys |
| JPH11300039A (en) * | 1998-04-24 | 1999-11-02 | Namco Ltd | Game device and information storage medium |
| JP3178426B2 (en) * | 1998-07-29 | 2001-06-18 | 日本電気株式会社 | Natural language dialogue system and natural language dialogue program recording medium |
| JP2000116940A (en) * | 1998-10-15 | 2000-04-25 | Seta Corp | Bidirectional communication-type game system |
| US20030190940A1 (en) * | 1998-11-05 | 2003-10-09 | Meryl Greenwald Gordon | Multiplayer electronic games |
| US7198490B1 (en) * | 1998-11-25 | 2007-04-03 | The Johns Hopkins University | Apparatus and method for training using a human interaction simulator |
| JP3729323B2 (en) * | 1999-04-15 | 2005-12-21 | 株式会社ソニー・コンピュータエンタテインメント | Entertainment system, entertainment apparatus, recording medium, and image processing method |
| US6500070B1 (en) | 1999-05-28 | 2002-12-31 | Nintendo Co., Ltd. | Combined game system of portable and video game machines |
| US6296487B1 (en) * | 1999-06-14 | 2001-10-02 | Ernest L. Lotecka | Method and system for facilitating communicating and behavior skills training |
| JP2001009157A (en) * | 1999-06-30 | 2001-01-16 | Konami Co Ltd | Control method for video game, video game device and medium recording program of video game allowing reading by computer |
| US6371854B1 (en) | 1999-08-20 | 2002-04-16 | Ninetendo Co., Ltd. | Combined game system |
| TW581701B (en) * | 2000-01-14 | 2004-04-01 | Sony Computer Entertainment Inc | Recording medium, method of using a computer and computer for executing role-playing games |
| US7210104B2 (en) * | 2000-02-16 | 2007-04-24 | Sega Corporation | Information display method and information display system for finding another user in a plurality of users based upon registered information |
| US20010039203A1 (en) * | 2000-02-23 | 2001-11-08 | Brown Geoffrey Parker | Behavior modeling in a gaming environment with contextual accuracy |
| US6299535B1 (en) * | 2000-04-28 | 2001-10-09 | Square Co., Ltd. | Method of processing interactive game, program product and game system for the same |
| US6951516B1 (en) * | 2001-08-21 | 2005-10-04 | Nintendo Co., Ltd. | Method and apparatus for multi-user communications using discrete video game platforms |
| US6544040B1 (en) * | 2000-06-27 | 2003-04-08 | Cynthia P. Brelis | Method, apparatus and article for presenting a narrative, including user selectable levels of detail |
| US7330806B2 (en) * | 2000-08-08 | 2008-02-12 | Reohr Iii John | Virtualized network |
| US7827488B2 (en) | 2000-11-27 | 2010-11-02 | Sitrick David H | Image tracking and substitution system and methodology for audio-visual presentations |
| JP2002282543A (en) | 2000-12-28 | 2002-10-02 | Sony Computer Entertainment Inc | Object voice processing program, computer-readable recording medium with object voice processing program recorded thereon, program execution device, and object voice processing method |
| US20020084986A1 (en) * | 2001-01-04 | 2002-07-04 | Armstrong Brad A. | Computer mouse with specialized button(s) |
| JP4691268B2 (en) | 2001-05-02 | 2011-06-01 | 任天堂株式会社 | Game system and game program |
| US6966837B1 (en) * | 2001-05-10 | 2005-11-22 | Best Robert M | Linked portable and video game systems |
| US20020193157A1 (en) * | 2001-06-18 | 2002-12-19 | Canon Kabushiki Kaisha | Computer device for implementing a trading card game and control method therefor, program executed by computer device, controller, system, and game cards |
| US20030049014A1 (en) * | 2001-09-07 | 2003-03-13 | Tri-Vision Electronics Inc. | Method and apparatus for playing digital media and digital media for use therein |
| US20040068406A1 (en) * | 2001-09-27 | 2004-04-08 | Hidetsugu Maekawa | Dialogue apparatus, dialogue parent apparatus, dialogue child apparatus, dialogue control method, and dialogue control program |
| EP1434627A4 (en) * | 2001-10-10 | 2005-02-02 | Sony Comp Emtertainment Us | Providing game information via characters in a game environment |
| US7198568B2 (en) * | 2002-05-01 | 2007-04-03 | Nintendo Co., Ltd. | Game machine and game program for changing the movement of one character based on the movement of another character |
| US7640164B2 (en) * | 2002-07-04 | 2009-12-29 | Denso Corporation | System for performing interactive dialog |
| US6982716B2 (en) * | 2002-07-11 | 2006-01-03 | Kulas Charles J | User interface for interactive video productions |
| US7401295B2 (en) * | 2002-08-15 | 2008-07-15 | Simulearn, Inc. | Computer-based learning system |
| US7908324B2 (en) | 2002-10-02 | 2011-03-15 | Disney Enterprises, Inc. | Multi-user interactive communication network environment |
| US7309283B2 (en) * | 2002-11-13 | 2007-12-18 | Keith G. Nemitz | Interactive narrative operated by introducing encounter events |
| US20050095569A1 (en) * | 2003-10-29 | 2005-05-05 | Patricia Franklin | Integrated multi-tiered simulation, mentoring and collaboration E-learning platform and its software |
| US20050177412A1 (en) * | 2004-02-09 | 2005-08-11 | Hire The Winners, Llc | Automobile sales simulator |
| ITRM20040201A1 (en) * | 2004-04-23 | 2004-07-23 | Link Italia S R L | MEMORY SUPPORT, IN PARTICULAR A DVD, THAT STORES INTERACTIVE MOVIES. |
| CA2505234A1 (en) * | 2004-04-30 | 2005-10-30 | Esel International Co., Ltd. | Wireless communication systems |
| US7713127B1 (en) * | 2004-06-14 | 2010-05-11 | Ted De La Torre-Bueno | Simulating dialog in electronic games |
| US20060040748A1 (en) * | 2004-08-19 | 2006-02-23 | Mark Barthold | Branching storyline game |
| CN101119782B (en) | 2004-09-21 | 2010-11-17 | 时间游戏Ip公司 | System for multiplayer gaming |
| US20070060392A1 (en) * | 2005-08-11 | 2007-03-15 | Sullivan Paul J | Game controller |
| US20070111169A1 (en) * | 2005-08-15 | 2007-05-17 | University Of Southern California | Interactive Story Development System with Automated Goal Prioritization |
| JP3916648B1 (en) * | 2006-02-17 | 2007-05-16 | 株式会社コナミデジタルエンタテインメント | GAME SERVER DEVICE, GAME SERVICE METHOD, AND PROGRAM |
| US8082499B2 (en) * | 2006-03-21 | 2011-12-20 | Electronic Arts, Inc. | Graphical interface for interactive dialog |
| JP5048271B2 (en) * | 2006-05-02 | 2012-10-17 | 任天堂株式会社 | GAME PROGRAM AND GAME DEVICE |
| WO2007133716A2 (en) * | 2006-05-11 | 2007-11-22 | Cerebode, Inc. | Multimodal communication and command control systems and related methods |
| JP4987399B2 (en) * | 2006-09-12 | 2012-07-25 | 任天堂株式会社 | GAME PROGRAM AND GAME DEVICE |
| US8771071B2 (en) * | 2006-11-22 | 2014-07-08 | Sony Computer Entertainment America Llc | System and method of rendering controller information |
| US8571463B2 (en) * | 2007-01-30 | 2013-10-29 | Breakthrough Performancetech, Llc | Systems and methods for computerized interactive skill training |
| US8702432B2 (en) | 2007-03-28 | 2014-04-22 | Breakthrough Performancetech, Llc | Systems and methods for computerized interactive training |
| US20090049470A1 (en) * | 2007-08-13 | 2009-02-19 | Gal Peer | Method and device for interactive operation of television |
| US9126116B2 (en) * | 2007-09-05 | 2015-09-08 | Sony Computer Entertainment America Llc | Ranking of user-generated game play advice |
| US9108108B2 (en) * | 2007-09-05 | 2015-08-18 | Sony Computer Entertainment America Llc | Real-time, contextual display of ranked, user-generated game play advice |
| CA2732268C (en) | 2008-07-28 | 2020-07-21 | Breakthrough Performancetech, Llc | Systems and methods for computerized interactive skill training |
| US20100255894A1 (en) * | 2009-04-01 | 2010-10-07 | Chira Kidakarn | Method for combining multiple actions in single video game |
| US8340965B2 (en) * | 2009-09-02 | 2012-12-25 | Microsoft Corporation | Rich context modeling for text-to-speech engines |
| US20110106536A1 (en) * | 2009-10-29 | 2011-05-05 | Rovi Technologies Corporation | Systems and methods for simulating dialog between a user and media equipment device |
| US20110271194A1 (en) * | 2010-04-29 | 2011-11-03 | Google Inc. | Voice ad interactions as ad conversions |
| JP6046638B2 (en) | 2011-02-01 | 2016-12-21 | タイムプレイ インク. | Interactive experience system and method, and controller therefor |
| US8594993B2 (en) | 2011-04-04 | 2013-11-26 | Microsoft Corporation | Frame mapping approach for cross-lingual voice transformation |
| US8874444B2 (en) * | 2012-02-28 | 2014-10-28 | Disney Enterprises, Inc. | Simulated conversation by pre-recorded audio navigator |
| US9604132B1 (en) * | 2012-05-07 | 2017-03-28 | CP Studios LLC | Video gaming platform and user interface |
| US9833707B2 (en) | 2012-10-29 | 2017-12-05 | Sony Interactive Entertainment Inc. | Ambient light control and calibration via a console |
| US9433863B2 (en) * | 2012-12-31 | 2016-09-06 | Echostar Technologies L.L.C. | Video processing using concurrent outcome simulation threads |
| US8977113B1 (en) * | 2013-10-25 | 2015-03-10 | Joseph Rumteen | Mobile device video decision tree |
| US10142708B2 (en) | 2014-07-31 | 2018-11-27 | Podop, Inc. | Method, apparatus and article for delivering media content via a user-selectable narrative presentation |
| US9690468B2 (en) * | 2014-08-01 | 2017-06-27 | Content Maker, Inc. | Interactive media presentations using a plurality of selectable audio elements |
| US10279257B2 (en) | 2015-01-14 | 2019-05-07 | Podop, Inc. | Data mining, influencing viewer selections, and user interfaces |
| JP2017162268A (en) * | 2016-03-10 | 2017-09-14 | 国立大学法人大阪大学 | Dialog system and control program |
| US10878799B2 (en) * | 2016-08-29 | 2020-12-29 | Sony Corporation | Information presenting apparatus and information presenting method |
| US10272349B2 (en) * | 2016-09-07 | 2019-04-30 | Isaac Davenport | Dialog simulation |
| US10111035B2 (en) | 2016-10-03 | 2018-10-23 | Isaac Davenport | Real-time proximity tracking using received signal strength indication |
| US10561942B2 (en) | 2017-05-15 | 2020-02-18 | Sony Interactive Entertainment America Llc | Metronome for competitive gaming headset |
| US10128914B1 (en) | 2017-09-06 | 2018-11-13 | Sony Interactive Entertainment LLC | Smart tags with multiple interactions |
| JP6909682B2 (en) * | 2017-09-12 | 2021-07-28 | 株式会社バンダイナムコエンターテインメント | Programs, image generators, and image generators |
| US11036984B1 (en) * | 2018-06-08 | 2021-06-15 | Facebook, Inc. | Interactive instructions |
| CN109453526B (en) * | 2018-10-26 | 2023-07-21 | 努比亚技术有限公司 | Sound processing method, terminal and computer readable storage medium |
| CN109646952A (en) * | 2018-12-14 | 2019-04-19 | 北京智明星通科技股份有限公司 | A kind of interactive game interface display method, device and terminal |
| US20230201715A1 (en) * | 2021-12-23 | 2023-06-29 | TCL Research America Inc. | Interactive video player |
| GB2622405A (en) * | 2022-09-15 | 2024-03-20 | Sony Interactive Entertainment Inc | Systems and methods for controlling dialogue complexity in video games |
Family Cites Families (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US4333152A (en) * | 1979-02-05 | 1982-06-01 | Best Robert M | TV Movies that talk back |
| US4445187A (en) * | 1979-02-05 | 1984-04-24 | Best Robert M | Video games with voice dialog |
| US4305131A (en) * | 1979-02-05 | 1981-12-08 | Best Robert M | Dialog between TV movies and human viewers |
| EP0016314A1 (en) * | 1979-02-05 | 1980-10-01 | Best, Robert MacAndrew | Method and apparatus for voice dialogue between a video picture and a human |
| US4569026A (en) * | 1979-02-05 | 1986-02-04 | Best Robert M | TV Movies that talk back |
| US5006987A (en) * | 1986-03-25 | 1991-04-09 | Harless William G | Audiovisual system for simulation of an interaction between persons through output of stored dramatic scenes in response to user vocal input |
| US4884972A (en) * | 1986-11-26 | 1989-12-05 | Bright Star Technology, Inc. | Speech synchronized animation |
| US4846693A (en) * | 1987-01-08 | 1989-07-11 | Smith Engineering | Video based instructional and entertainment system using animated figure |
| US4847699A (en) * | 1987-07-16 | 1989-07-11 | Actv, Inc. | Method for providing an interactive full motion synched compatible audio/visual television display |
-
1991
- 1991-10-07 CA CA002095820A patent/CA2095820C/en not_active Expired - Fee Related
- 1991-10-07 EP EP92901085A patent/EP0557444A1/en not_active Withdrawn
- 1991-10-07 AU AU90881/91A patent/AU652209B2/en not_active Ceased
- 1991-10-07 WO PCT/US1991/007266 patent/WO1992008531A1/en not_active Ceased
- 1991-11-05 JP JP3315172A patent/JP2784409B2/en not_active Expired - Fee Related
-
1993
- 1993-10-21 US US08/140,266 patent/US5358259A/en not_active Expired - Fee Related
Cited By (25)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6241611B1 (en) | 1995-05-10 | 2001-06-05 | Nintendo Co., Ltd. | Function expansion device and operating device using the function expansion device |
| US6489946B1 (en) | 1995-05-10 | 2002-12-03 | Nintendo Co., Ltd. | Operating device with analog joystick |
| US6461242B2 (en) | 1995-05-10 | 2002-10-08 | Nintendo Co., Ltd. | Operating device for an image processing apparatus |
| US6102803A (en) | 1995-05-10 | 2000-08-15 | Nintendo Co., Ltd. | Operating device with analog joystick |
| US6186896B1 (en) | 1995-05-10 | 2001-02-13 | Nintendo Co., Ltd. | Operating device with analog joystick |
| US7102618B2 (en) | 1995-10-09 | 2006-09-05 | Nintendo Co., Ltd. | User controlled graphics object movement based on a amount of joystick angular rotation and point of view angle |
| US6200253B1 (en) | 1995-10-09 | 2001-03-13 | Nintendo Co., Ltd. | Controller pack |
| US6001015A (en) * | 1995-10-09 | 1999-12-14 | Nintendo Co., Ltd. | Operation controlling device and video processing system used therewith |
| US6007428A (en) * | 1995-10-09 | 1999-12-28 | Nintendo Co., Ltd. | Operation controlling device and video processing system used therewith |
| US6421056B1 (en) | 1995-10-09 | 2002-07-16 | Nintendo Co., Ltd. | Three-dimensional image processing apparatus |
| US6917356B1 (en) | 1995-10-09 | 2005-07-12 | Nintendo Co. Ltd. | User controlled graphics object movement based on amount of joystick angular rotation and point of view angle |
| US6325718B1 (en) | 1995-10-09 | 2001-12-04 | Nintendo Co., Ltd. | Operation controlling device and video processing system used therewith |
| US6676520B2 (en) | 1995-10-09 | 2004-01-13 | Nintendo Co., Ltd. | Video game system providing physical sensation |
| US7126584B1 (en) | 1995-10-09 | 2006-10-24 | Nintendo Co., Ltd. | Operating device and image processing system using same |
| US6331146B1 (en) | 1995-11-22 | 2001-12-18 | Nintendo Co., Ltd. | Video game system and method with enhanced three-dimensional character and background control |
| US6454652B2 (en) | 1995-11-22 | 2002-09-24 | Nintendo Co., Ltd. | Video game system and method with enhanced three-dimensional character and background control due to environmental conditions |
| US6383079B1 (en) | 1995-11-22 | 2002-05-07 | Nintendo Co., Ltd. | High performance/low cost video game system with multi-functional peripheral processing subsystem |
| US6155926A (en) * | 1995-11-22 | 2000-12-05 | Nintendo Co., Ltd. | Video game system and method with enhanced three-dimensional character and background control |
| US6139433A (en) | 1995-11-22 | 2000-10-31 | Nintendo Co., Ltd. | Video game system and method with enhanced three-dimensional character and background control due to environmental conditions |
| US6022274A (en) * | 1995-11-22 | 2000-02-08 | Nintendo Co., Ltd. | Video game system using memory module |
| US6267673B1 (en) | 1996-09-20 | 2001-07-31 | Nintendo Co., Ltd. | Video game system with state of next world dependent upon manner of entry from previous world via a portal |
| US6491585B1 (en) | 1996-09-24 | 2002-12-10 | Nintendo Co., Ltd. | Three-dimensional image processing apparatus with enhanced automatic and user point of view control |
| US6139434A (en) | 1996-09-24 | 2000-10-31 | Nintendo Co., Ltd. | Three-dimensional image processing apparatus with enhanced automatic and user point of view control |
| US6679776B1 (en) | 1997-07-17 | 2004-01-20 | Nintendo Co., Ltd. | Video game system |
| US7070507B2 (en) | 1997-07-17 | 2006-07-04 | Nintendo Co., Ltd. | Video game system |
Also Published As
| Publication number | Publication date |
|---|---|
| US5358259A (en) | 1994-10-25 |
| WO1992008531A1 (en) | 1992-05-29 |
| JP2784409B2 (en) | 1998-08-06 |
| CA2095820C (en) | 2004-04-13 |
| EP0557444A1 (en) | 1993-09-01 |
| AU652209B2 (en) | 1994-08-18 |
| CA2095820A1 (en) | 1992-05-15 |
| JPH04266781A (en) | 1992-09-22 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| AU652209B2 (en) | Talking video games | |
| US5393073A (en) | Talking video games | |
| US5393072A (en) | Talking video games with vocal conflict | |
| US5393071A (en) | Talking video games with cooperative action | |
| US5393070A (en) | Talking video games with parallel montage | |
| US8082499B2 (en) | Graphical interface for interactive dialog | |
| US4569026A (en) | TV Movies that talk back | |
| US4333152A (en) | TV Movies that talk back | |
| US4445187A (en) | Video games with voice dialog | |
| US4846693A (en) | Video based instructional and entertainment system using animated figure | |
| Goodwin | Audience diversity, participation and interpretation | |
| JPH06507084A (en) | Video games with audio player interaction using real-time video synchronization | |
| Lancaster | When spectators become performers: contemporary performance-entertainments meet the needs of an" unsettled" audience | |
| Ben-Zvi | Samuel Beckett's media plays | |
| EP1172132A2 (en) | Entertainment system, recording medium | |
| WO1993014844A1 (en) | Talking video games with cooperative action | |
| US5930757A (en) | Interactive two-way conversational apparatus with voice recognition | |
| Emunah | Drama therapy in action. | |
| WO1993004748A1 (en) | Video game with interactive audiovisual dialogue | |
| Kang et al. | One-Man Movie: A System to Assist Actor Recording in a Virtual Studio | |
| JPH05293252A (en) | Dialog video game method utilizing sound contrast | |
| WO1996003190A1 (en) | Interactive system with programmable toys | |
| CA2454004A1 (en) | Linked video game system | |
| Boles | “Only Write the Good Parts”: Playwright Lucas Hnath in Conversation with Jay Malarcher | |
| Mattsson | Dialect representation: Language varieties in The Witcher 3: The Wild Hunt |