US20100062860A1 - Operation of a set of devices - Google Patents
Operation of a set of devices Download PDFInfo
- Publication number
- US20100062860A1 US20100062860A1 US12/621,644 US62164409A US2010062860A1 US 20100062860 A1 US20100062860 A1 US 20100062860A1 US 62164409 A US62164409 A US 62164409A US 2010062860 A1 US2010062860 A1 US 2010062860A1
- Authority
- US
- United States
- Prior art keywords
- real
- world
- generating
- signal
- markup language
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
- G06Q50/10—Services
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/4104—Peripherals receiving signals from specially adapted client devices
- H04N21/4131—Peripherals receiving signals from specially adapted client devices home appliance, e.g. lighting, air conditioning system, metering devices
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/40—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
- A63F13/42—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/30—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by output arrangements for receiving control signals generated by the game device
- A63F2300/302—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by output arrangements for receiving control signals generated by the game device specially adapted for receiving control signals not targeted to a display device or game input means, e.g. vibrating driver's seat, scent dispenser
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/60—Methods for processing data by generating or executing the game program
- A63F2300/6045—Methods for processing data by generating or executing the game program for mapping control signals received from the input arrangement into game commands
Definitions
- This invention relates to a method of and apparatus for operating a set of devices.
- this invention relates to receiving a signal produced by a computer program, the signal comprising a game world model, analysing the signal to produce a real-world description in the form of an instruction set of a markup-language, and operating the devices according to the description.
- a method of operating a set of devices comprising receiving a signal, analysing said signal to produce a real-world description in the form of an instruction set of a markup language, and operating said devices according to said description.
- apparatus for operating a set of devices comprising receiving means for receiving a signal and analysing means for analysing said signal to produce a real-world description in the form of an instruction set of a markup language.
- the signal may be a broadcast signal or may be a signal produced by a computer program.
- the signal may comprise at least part of a game world model.
- the signal may comprise a video component.
- Preferably the description is distributed amongst said devices.
- the analysing can take place in real time, or can be carried out before it is needed, storing the real-world description locally and recalling that description on demand.
- the real-world description is preferably limited according to the functionality of the devices present in the set of devices.
- FIG. 1 is a schematic view of a method of operating a set of devices
- FIG. 2 is a perspective view of a real-world representation system
- FIG. 3 is a flowchart of a method of controlling a set of devices
- FIG. 4 is a flowchart of an alternative method of controlling a set of devices.
- the method of FIG. 1 illustrates how a user 2 of a computer game has their experience of the game augmented by the generation of a real-world description in the form of an instruction set of a markup language.
- This instruction set is used to operate a set of devices that make up a real-world representation system.
- the markup language is used, in general, to describe the real world elements associated with the virtual world of the computer game that fall outside the part of the visual field available to the user.
- the user 2 views a display 3 of a PC that displays a video component of the computer game.
- the display 3 can be thought of as showing a rendered part 4 of a game world model 6 .
- the part 4 is typically the portion of the game world model 6 that relates to the particular location that the user is currently experiencing. As the user 2 “moves” around the game world, the part that is rendered changes accordingly.
- a translator 7 receives a signal produced by a computer program, which comprises at least part of the game world model 6 , and analyses the signal to produce a real-world description 8 in the form of an instruction set of a markup language. In this example, the analysing takes place in real time, the signal that the translator 7 receives being the entire model 6 .
- the translator 7 analyses the signal to identify the environment that is presently being rendered on the display 3 , and generates an appropriate real-world description, which is passed to a browser 9 .
- the real-world description can be updated rapidly, as the game progresses.
- the browser 9 is either present on a local server, or is distributed amongst the set of devices.
- the translator 7 could be a plug in unit for the apparatus running the computer program.
- the unit would comprise receiving means for receiving the signal and analysing means for analysing the signal to produce the real-world description. Additionally distributing means could be included for distributing the description to the devices in the set. Alternatively the translator 7 could be implemented in software.
- the real-world description is provided by a markup language that communicates a description of physical environments and the objects within them, their relationship to the user, each other and to the physical space.
- the instruction set of the markup language is interpreted by a device or devices to render the experience.
- Each device that is enabled contains a component that interprets the instruction set to the best of its capability.
- the language contains a wide range of states that can be rendered by the devices in a real-world representation system. Such states relate to:
- the browser 9 is arranged to interpret the description, which is then rendered by the devices in a real-world representation system.
- FIG. 2 shows a real-world representation system comprising a set of devices including a display device 10 , audio speakers 12 , a lighting device 14 , a heating device 16 , walls 18 and the floor 20 . These devices together contribute to make up the ambient environment, each device being arranged to provide one or more real-world parameters. For example, the lighting device 14 contributes colour tones as well as to the luminance level.
- the devices may be electronic or they may be purely mechanical.
- the devices are interconnected by either a wireless network or a wired network such as a powerline carrier network.
- the walls 18 and floor 20 are provided with display functionality. This can be achieved either by the walls 18 and floor 20 being composed of fabric display material, or a centred ceiling projector can illuminate the visible portions of the walls 18 and the floor 20 . It is also feasible to back light the walls 18 if the environment allows it.
- the system of FIG. 2 is shown as being a room with a couch 22 for a user, the room being part of a private dwelling. However the room could be, for example, part of an entertainment complex or could form part of a public space.
- the devices making up the real-world representation system are arranged to receive a real-world description in the form of an instruction set of a markup language, the devices being operated according to said description.
- each markup language enabled device operates in principally the same manner.
- the lighting device 14 has receiving means for receiving the real-world description in the form of an instruction set of a markup language, the receiving means including part of the distributed browser 9 that interprets the instructions of the instruction set.
- the portion of the browser 9 in the receiving means communicates with adjusting means that is arranged to adjust one or more parameters of the lighting device 14 .
- the part of the browser 9 in the receiving means interprets this into specific instructions relating to the colour tones and luminance level for the adjusting means to adjust their levels accordingly.
- the likely colour tone would be a pleasant green and the light level would be low but warm.
- the browser 9 interprets instructions of a general type to generate specific parameter adjustments.
- the description is received by all of the other parts of the distributed browser 9 in each of the other devices in the real-world representation system.
- Each device interprets the instructions and adjusts its parameter configuration accordingly.
- the greater number of enabled devices that are present in a particular environment the more realistic the end result will be.
- the operability of the walls 18 and floor 20 as display devices will determine how immersed in the real-world representation the user feels.
- a local server which can be a dedicated device or could be contained within the capability of a device that nominally has a different purpose.
- a browser 9 or operating system is present on the local server and interprets the instructions of the real-world description and generates specific parameter adjustments for communicating to the relevant device. In this way devices currently available can be used without the requirement that they be modified or purpose built for use in the real-world representation system.
- the game world model 6 By using the game world model 6 to generate a markup language representation of the user's location, it is possible to extend the experience of the virtual location into the real world. This could also include audio, for example, a user could hear a monster creep up behind them. If a rocket just missed a user and hit the wall behind them, the lights at this location will glow yellow and red, and synchronise with the audio stream.
- An alternative way of operating the set of devices is to generate the real-world description before it is needed, storing the real-world description locally and recalling the description on demand.
- the real-world description, whether generated in real-time or before it is needed can be limited according to the functionality of the devices present in the set of devices. This saves on processing time and resources.
- the computer game that the user is playing need not have a video component.
- the user may be playing an electronic board game and the translator 7 will use the game world model 6 associated with the board game to generate a description in markup language. This description is rendered by the browser 9 to augment the user's experience of the electronic board game.
- the method of operating a set of devices can also be utilised when the signal received by the translator 7 comprises a broadcast signal.
- the creation of the description from a broadcast signal allows the user to be immersed in an environment that is linked to a broadcast that he is receiving, for example a broadcast television signal.
- the user may be watching a computer generated animation, that includes in the broadcast stream a signal comprising the model of the content.
- This signal can be analysed by the translator 7 to produce a real-world description that is used to operate the devices making up the real-world representation system.
- the user can operate a user interface to set limits on the operation of the devices in the system. For example, if the user wishes to set the volume of those devices with an audio capability to a specific range or set an upper limit on the volume then they can do so via the user interface. This prevents experiences becoming unpleasant for the user.
- the level of the light and the rate of any change in light intensity are also things that the user can control. All of the parameters of the system can be user defined.
- the user interface can be portable.
Abstract
A method of operating a set of devices comprises receiving a signal, for example at least part of a game world model from a computer program. The signal is analyzed to produce a real-world description in the form of an instruction set of a markup language and the set of devices is operated according to the description.
Description
- This invention relates to a method of and apparatus for operating a set of devices. In particular this invention relates to receiving a signal produced by a computer program, the signal comprising a game world model, analysing the signal to produce a real-world description in the form of an instruction set of a markup-language, and operating the devices according to the description.
- When playing a computer game, the user's experience of the game consists, in most cases, of the viewing of a simple display device while listening to the associated audio. Since the advent of computer games, it has been desired to augment this user experience. A number of ways of achieving this have been proposed, including head mounted displays, surround screen installations and game peripherals such as rumble pads. The object of these functional improvements has been to increase the user's immersion in the game environment, or game world model. However a number of problems are apparent with the known improvements, including the high cost of any extra devices, the fact that most solutions are proprietary and therefore not transferable between platforms, and that generally each solution adds only a small additional functionality to the game world.
- According to a first aspect of the invention, there is provided a method of operating a set of devices, comprising receiving a signal, analysing said signal to produce a real-world description in the form of an instruction set of a markup language, and operating said devices according to said description.
- According to a second aspect of the invention, there is provided apparatus for operating a set of devices, comprising receiving means for receiving a signal and analysing means for analysing said signal to produce a real-world description in the form of an instruction set of a markup language.
- The signal may be a broadcast signal or may be a signal produced by a computer program. The signal may comprise at least part of a game world model. The signal may comprise a video component. Preferably the description is distributed amongst said devices. The analysing can take place in real time, or can be carried out before it is needed, storing the real-world description locally and recalling that description on demand. The real-world description is preferably limited according to the functionality of the devices present in the set of devices.
- Owing to the invention, it is possible to provide an enhanced experience of, for example, a computer game. By analysing the content of the signal making up the game and producing from it a real-world description in the form of an instruction set of a markup language, the description can be used to operate a set of devices that will augment the experience.
- Embodiments of the invention will now be described, by way of example only, with reference to the accompanying drawings, in which:
-
FIG. 1 is a schematic view of a method of operating a set of devices, -
FIG. 2 is a perspective view of a real-world representation system, -
FIG. 3 is a flowchart of a method of controlling a set of devices, and -
FIG. 4 is a flowchart of an alternative method of controlling a set of devices. - The method of
FIG. 1 illustrates how auser 2 of a computer game has their experience of the game augmented by the generation of a real-world description in the form of an instruction set of a markup language. This instruction set is used to operate a set of devices that make up a real-world representation system. The markup language is used, in general, to describe the real world elements associated with the virtual world of the computer game that fall outside the part of the visual field available to the user. - The
user 2 views adisplay 3 of a PC that displays a video component of the computer game. Thedisplay 3 can be thought of as showing a rendered part 4 of a game world model 6. The part 4 is typically the portion of the game world model 6 that relates to the particular location that the user is currently experiencing. As theuser 2 “moves” around the game world, the part that is rendered changes accordingly. - A
translator 7 receives a signal produced by a computer program, which comprises at least part of the game world model 6, and analyses the signal to produce a real-world description 8 in the form of an instruction set of a markup language. In this example, the analysing takes place in real time, the signal that thetranslator 7 receives being the entire model 6. Thetranslator 7 analyses the signal to identify the environment that is presently being rendered on thedisplay 3, and generates an appropriate real-world description, which is passed to abrowser 9. The real-world description can be updated rapidly, as the game progresses. Thebrowser 9 is either present on a local server, or is distributed amongst the set of devices. - The
translator 7 could be a plug in unit for the apparatus running the computer program. The unit would comprise receiving means for receiving the signal and analysing means for analysing the signal to produce the real-world description. Additionally distributing means could be included for distributing the description to the devices in the set. Alternatively thetranslator 7 could be implemented in software. - The real-world description is provided by a markup language that communicates a description of physical environments and the objects within them, their relationship to the user, each other and to the physical space. Within a location that is enabled to produce a real-world experience, the instruction set of the markup language is interpreted by a device or devices to render the experience. Each device that is enabled contains a component that interprets the instruction set to the best of its capability.
- The language contains a wide range of states that can be rendered by the devices in a real-world representation system. Such states relate to:
-
- Image display—specific images, streamed video
- Audio—music, sound effects, voice
- Mood—emotional, ambient, animated
- Light—levels (relative/absolute), moods, colours, position, focus
- Location—absolute, fantasy, generic type.
- Any information that relates to the game world experience can be expressed in this markup language. The
browser 9 is arranged to interpret the description, which is then rendered by the devices in a real-world representation system. -
FIG. 2 shows a real-world representation system comprising a set of devices including adisplay device 10,audio speakers 12, alighting device 14, aheating device 16,walls 18 and thefloor 20. These devices together contribute to make up the ambient environment, each device being arranged to provide one or more real-world parameters. For example, thelighting device 14 contributes colour tones as well as to the luminance level. The devices may be electronic or they may be purely mechanical. The devices are interconnected by either a wireless network or a wired network such as a powerline carrier network. - In addition to the
display device 10, thewalls 18 andfloor 20 are provided with display functionality. This can be achieved either by thewalls 18 andfloor 20 being composed of fabric display material, or a centred ceiling projector can illuminate the visible portions of thewalls 18 and thefloor 20. It is also feasible to back light thewalls 18 if the environment allows it. - The system of
FIG. 2 is shown as being a room with acouch 22 for a user, the room being part of a private dwelling. However the room could be, for example, part of an entertainment complex or could form part of a public space. The devices making up the real-world representation system are arranged to receive a real-world description in the form of an instruction set of a markup language, the devices being operated according to said description. - In a first embodiment (outlined in the flowchart of
FIG. 3 ) this description is distributed amongst the devices. Each markup language enabled device operates in principally the same manner. For example, thelighting device 14 has receiving means for receiving the real-world description in the form of an instruction set of a markup language, the receiving means including part of thedistributed browser 9 that interprets the instructions of the instruction set. The portion of thebrowser 9 in the receiving means communicates with adjusting means that is arranged to adjust one or more parameters of thelighting device 14. For example, if the real-world description reads <FOREST>, <SUMMER>, <EVENING> then the part of thebrowser 9 in the receiving means interprets this into specific instructions relating to the colour tones and luminance level for the adjusting means to adjust their levels accordingly. In this example the likely colour tone would be a pleasant green and the light level would be low but warm. Thebrowser 9 interprets instructions of a general type to generate specific parameter adjustments. - The description is received by all of the other parts of the distributed
browser 9 in each of the other devices in the real-world representation system. Each device interprets the instructions and adjusts its parameter configuration accordingly. The greater number of enabled devices that are present in a particular environment the more realistic the end result will be. In particular the operability of thewalls 18 andfloor 20 as display devices will determine how immersed in the real-world representation the user feels. - In a second embodiment (outlined in the flowchart of
FIG. 4 ) the description is read at a local server, which can be a dedicated device or could be contained within the capability of a device that nominally has a different purpose. In this embodiment abrowser 9 or operating system is present on the local server and interprets the instructions of the real-world description and generates specific parameter adjustments for communicating to the relevant device. In this way devices currently available can be used without the requirement that they be modified or purpose built for use in the real-world representation system. - By using the game world model 6 to generate a markup language representation of the user's location, it is possible to extend the experience of the virtual location into the real world. This could also include audio, for example, a user could hear a monster creep up behind them. If a rocket just missed a user and hit the wall behind them, the lights at this location will glow yellow and red, and synchronise with the audio stream.
- An alternative way of operating the set of devices is to generate the real-world description before it is needed, storing the real-world description locally and recalling the description on demand.
- The real-world description, whether generated in real-time or before it is needed can be limited according to the functionality of the devices present in the set of devices. This saves on processing time and resources.
- The computer game that the user is playing need not have a video component. For example, the user may be playing an electronic board game and the
translator 7 will use the game world model 6 associated with the board game to generate a description in markup language. This description is rendered by thebrowser 9 to augment the user's experience of the electronic board game. - The method of operating a set of devices can also be utilised when the signal received by the
translator 7 comprises a broadcast signal. The creation of the description from a broadcast signal allows the user to be immersed in an environment that is linked to a broadcast that he is receiving, for example a broadcast television signal. For example, the user may be watching a computer generated animation, that includes in the broadcast stream a signal comprising the model of the content. This signal can be analysed by thetranslator 7 to produce a real-world description that is used to operate the devices making up the real-world representation system. - The user can operate a user interface to set limits on the operation of the devices in the system. For example, if the user wishes to set the volume of those devices with an audio capability to a specific range or set an upper limit on the volume then they can do so via the user interface. This prevents experiences becoming unpleasant for the user. The level of the light and the rate of any change in light intensity are also things that the user can control. All of the parameters of the system can be user defined. The user interface can be portable.
Claims (20)
1-13. (canceled)
14. A method of operating a real device capable of generating ambient stimuli external of a virtual game world generated from a game world model, the device having at least one adjustable parameter associated therewith, the method comprising:
A) deriving, from the game world model, a real-world description in the form of markup language instructions;
B) interpreting the markup language instructions and generating at least one device specific parameter adjustment therefrom;
C) modifying the at least one parameter of the device according to the specific parameter adjustments; and
D) generating ambient stimuli external of the virtual game world with the device according to the modified parameter.
15. The method of claim 14 wherein B) comprises:
B1) interpreting the markup language instructions and generating the at least one device specific parameter adjustment therefrom at the real device.
16. The method of claim 14 wherein B) comprises:
B1) interpreting the markup language instructions and generating the at least one device specific parameter adjustment therefrom remote from the real device; and
B2) distributing the interpreted instructions and the generated device specific parameter adjustment to the real device.
17. The method of claim 14 wherein A) comprises:
A1) receiving a signal; and
A2) deriving the real-world description from the received signal.
18. The method according to claim 17 , wherein the signal comprises a broadcast signal.
19. The method according to claim 17 , wherein the signal is produced by a computer program.
20. The method according to claim 17 , wherein the signal comprises at least part of a game world model.
21. The method according to claim 17 , wherein the signal comprises at least part of a video component.
22. The method according to claim 14 , wherein A) takes place in real time.
23. The method according of claim 14 wherein A) comprises:
A1) storing said real-world description in local memory.
24. The method according of claim 23 wherein A) further comprises:
A2) recalling said locally stored real-world description on demand.
25. A system capable of generating ambient stimuli external of a virtual game world comprising:
a memory for storing a game world model;
a computer game process coupled to the memory for generating a computer game representation of the game world model;
a mechanism for displaying a video representation of a portion of the game world model as generated by the computer game process, said portion representing a particular location of the game world model being experienced by a user;
a translator responsive to a signal produced by the computer game process and capable of analyzing the signal and deriving a real world description of the location in the form of a markup language instructions;
a browser portion, responsive to the at least one markup language instruction, for interpreting the instruction and for generating device specific parameter adjustments therefrom;
at least one device responsive to the browser portion and capable of generating ambient stimuli and having at least one adjustable parameter related thereto.
26. The system of claim 25 further comprising:
a plurality devices responsive to the browser portion and capable of generating ambient stimuli and having at least one adjustable parameter related thereto.
27. The system of claim 26 wherein each the plurality devices further comprise:
a browser portion, responsive to the at least one markup language instruction, for interpreting the instruction and for generating device specific parameter adjustments therefrom.
28. The system of claim 25 wherein the analysing of the signal takes place in real time.
29. The system of claim 25 wherein said real-world description is stored within said memory.
30. A real device comprising a you:
means capable of generating ambient stimuli and having at least one adjustable parameter related thereto,
means for receiving a real-world description in the form of at least one instruction of a markup language;
a browser portion, responsive to the at least one markup language instruction, for interpreting the instruction and for generating a specific parameter adjustment therefrom; and
means for modifying one or more of parameter configurations according to the generated parameter adjustment.
31. The device of claim 30 wherein the means capable of generating ambient stimuli has plural adjustable parameters related thereto.
32. The device of claim 30 in combination with other of said devices operatively coupled over a network.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/621,644 US20100062860A1 (en) | 2001-05-11 | 2009-11-19 | Operation of a set of devices |
Applications Claiming Priority (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GBGB0111431.3A GB0111431D0 (en) | 2001-05-11 | 2001-05-11 | A real-world representation system and language |
GB0111431.3 | 2001-05-11 | ||
GB0128524.6 | 2001-11-29 | ||
GBGB0128524.6A GB0128524D0 (en) | 2001-11-29 | 2001-11-29 | Operation of a set of devices |
US10/143,642 US7740531B2 (en) | 2001-05-11 | 2002-05-10 | Operation of a set of devices |
US12/621,644 US20100062860A1 (en) | 2001-05-11 | 2009-11-19 | Operation of a set of devices |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/143,642 Continuation US7740531B2 (en) | 2001-05-11 | 2002-05-10 | Operation of a set of devices |
Publications (1)
Publication Number | Publication Date |
---|---|
US20100062860A1 true US20100062860A1 (en) | 2010-03-11 |
Family
ID=26246053
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/143,642 Active 2028-01-17 US7740531B2 (en) | 2001-05-11 | 2002-05-10 | Operation of a set of devices |
US12/621,644 Abandoned US20100062860A1 (en) | 2001-05-11 | 2009-11-19 | Operation of a set of devices |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/143,642 Active 2028-01-17 US7740531B2 (en) | 2001-05-11 | 2002-05-10 | Operation of a set of devices |
Country Status (7)
Country | Link |
---|---|
US (2) | US7740531B2 (en) |
EP (1) | EP1401547B1 (en) |
JP (1) | JP4601255B2 (en) |
KR (1) | KR20030021250A (en) |
CN (1) | CN1232322C (en) |
AT (1) | ATE542577T1 (en) |
WO (1) | WO2002092182A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090280896A1 (en) * | 2006-06-19 | 2009-11-12 | Ambx Uk Limited | Game enhancer |
US20090322955A1 (en) * | 2006-06-13 | 2009-12-31 | Takuya Iwanami | Data transmitting device, data transmitting method, audio-visual environment control device, audio-visual environment control system, and audio-visual environment control method |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7689422B2 (en) * | 2002-12-24 | 2010-03-30 | Ambx Uk Limited | Method and system to mark an audio signal with metadata |
GB0305762D0 (en) * | 2003-03-13 | 2003-04-16 | Koninkl Philips Electronics Nv | Asset channels |
GB0328953D0 (en) * | 2003-12-12 | 2004-01-14 | Koninkl Philips Electronics Nv | Assets and effects |
TWI417788B (en) * | 2005-09-01 | 2013-12-01 | Koninkl Philips Electronics Nv | A data processing system and a method of operating a rendering platform |
EP1962980A1 (en) * | 2005-12-16 | 2008-09-03 | Koninklijke Philips Electronics N.V. | Shadow generation apparatus and method |
JP2009540909A (en) * | 2006-06-19 | 2009-11-26 | エーエムビーエックス ユーケー リミテッド | Game accelerator |
JP2009542297A (en) * | 2006-07-07 | 2009-12-03 | エーエムビーエックス ユーケー リミテッド | Ambient environmental effects |
JP5341755B2 (en) * | 2006-07-17 | 2013-11-13 | コーニンクレッカ フィリップス エヌ ヴェ | Determining environmental parameter sets |
WO2008047281A2 (en) * | 2006-10-18 | 2008-04-24 | Ambx Uk Limited | Method and system for detecting effect of lighting device |
JP2011501981A (en) * | 2007-09-07 | 2011-01-20 | エーエムビーエックス ユーケー リミテッド | How to generate effect scripts corresponding to game play events |
US20090079690A1 (en) * | 2007-09-21 | 2009-03-26 | Sony Computer Entertainment America Inc. | Method and apparatus for enhancing entertainment software through haptic insertion |
KR20140062892A (en) * | 2012-11-15 | 2014-05-26 | 삼성전자주식회사 | Wearable device, display device and system for providing exercise service and methods thereof |
Citations (40)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5004225A (en) * | 1989-11-01 | 1991-04-02 | Yuri Krukovsky | Simulated observation flight apparatus |
US5291556A (en) * | 1989-10-28 | 1994-03-01 | Hewlett-Packard Company | Audio system for a computer display |
US5695400A (en) * | 1996-01-30 | 1997-12-09 | Boxer Jam Productions | Method of managing multi-player game playing over a network |
US5771041A (en) * | 1994-06-03 | 1998-06-23 | Apple Computer, Inc. | System for producing directional sound in computer based virtual environment |
US5795228A (en) * | 1996-07-03 | 1998-08-18 | Ridefilm Corporation | Interactive computer-based entertainment system |
US5855483A (en) * | 1994-11-21 | 1999-01-05 | Compaq Computer Corp. | Interactive play with a computer |
US5887118A (en) * | 1997-02-12 | 1999-03-23 | Motorola, Inc. | Olfactory card |
US5945986A (en) * | 1997-05-19 | 1999-08-31 | University Of Illinois At Urbana-Champaign | Silent application state driven sound authoring system and method |
US5956038A (en) * | 1995-07-12 | 1999-09-21 | Sony Corporation | Three-dimensional virtual reality space sharing method and system, an information recording medium and method, an information transmission medium and method, an information processing method, a client terminal, and a shared server terminal |
US5964660A (en) * | 1997-06-18 | 1999-10-12 | Vr-1, Inc. | Network multiplayer game |
US5982445A (en) * | 1996-10-21 | 1999-11-09 | General Instrument Corporation | Hypertext markup language protocol for television display and control |
US5982903A (en) * | 1995-09-26 | 1999-11-09 | Nippon Telegraph And Telephone Corporation | Method for construction of transfer function table for virtual sound localization, memory with the transfer function table recorded therein, and acoustic signal editing scheme using the transfer function table |
US5991693A (en) * | 1996-02-23 | 1999-11-23 | Mindcraft Technologies, Inc. | Wireless I/O apparatus and method of computer-assisted instruction |
US6007338A (en) * | 1997-11-17 | 1999-12-28 | Disney Enterprises, Inc. | Roller coaster simulator |
US6050822A (en) * | 1997-10-01 | 2000-04-18 | The United States Of America As Represented By The Secretary Of The Army | Electromagnetic locomotion platform for translation and total immersion of humans into virtual environments |
US6056640A (en) * | 1996-11-11 | 2000-05-02 | Schaaij; Johan Michiel | Computerized gaming apparatus |
US6106399A (en) * | 1997-06-16 | 2000-08-22 | Vr-1, Inc. | Internet audio multi-user roleplaying game |
US6154553A (en) * | 1993-12-14 | 2000-11-28 | Taylor Group Of Companies, Inc. | Sound bubble structures for sound reproducing arrays |
US6159100A (en) * | 1998-04-23 | 2000-12-12 | Smith; Michael D. | Virtual reality game |
US6167353A (en) * | 1996-07-03 | 2000-12-26 | Interval Research Corporation | Computer method and apparatus for interacting with a physical system |
US6166496A (en) * | 1997-08-26 | 2000-12-26 | Color Kinetics Incorporated | Lighting entertainment system |
US6243707B1 (en) * | 1997-06-25 | 2001-06-05 | Samsung Electronics Co. Ltd. | Method and apparatus for creating home network macros |
US6354940B1 (en) * | 1999-04-30 | 2002-03-12 | Square Co., Ltd. | Game apparatus, game fight processing method, and computer readable storage medium storage program therefor |
US6379253B1 (en) * | 1999-06-08 | 2002-04-30 | Konami Co., Ltd. | Game device and computer-readable storage medium |
US20020087988A1 (en) * | 2000-12-28 | 2002-07-04 | International Business Machines Corporation | Interactive TV contextual content surfing using inter-channel hyperlinking: systems, methods & program products |
US20020110788A1 (en) * | 2001-02-09 | 2002-08-15 | Henry Michael L. | Acoustical enhancement of audio sound effects playback in video games |
US6498955B1 (en) * | 1999-03-19 | 2002-12-24 | Accenture Llp | Member preference control of an environment |
US20030018755A1 (en) * | 2001-03-30 | 2003-01-23 | Masterson Robert J. | Online system that facilitates configuration and administration of residential electronic devices |
US6553121B1 (en) * | 1995-09-08 | 2003-04-22 | Fujitsu Limited | Three-dimensional acoustic processor which uses linear predictive coefficients |
US20030076281A1 (en) * | 1997-08-26 | 2003-04-24 | Frederick Marshall Morgan | Diffuse illumination systems and methods |
US20040024656A1 (en) * | 2000-06-02 | 2004-02-05 | Coleman Kevin B. | Interactive product selector with inferential logic engine |
US6826725B1 (en) * | 1999-12-16 | 2004-11-30 | Microsoft Corporation | Techniques for invoking system commands from within a mark-up language document |
US6993722B1 (en) * | 1999-02-08 | 2006-01-31 | Cirrus Logic, Inc. | User interface system methods and computer program products for multi-function consumer entertainment appliances |
US7100118B1 (en) * | 1999-03-05 | 2006-08-29 | Amulet Technologies, Llc | Graphical user interface engine for embedded systems |
US7111297B1 (en) * | 2000-05-02 | 2006-09-19 | Microsoft Corporation | Methods and architectures for resource management |
US7118482B2 (en) * | 2000-05-29 | 2006-10-10 | Nintendo Co., Ltd. | Game system using game cards and game machine |
US7140033B1 (en) * | 2000-06-21 | 2006-11-21 | Bellsouth Intellectual Property Corporation | Methods and systems for controlling consumer electronics external devices via data delivered to a device |
US7147561B2 (en) * | 2003-08-20 | 2006-12-12 | Nintendo Co., Ltd. | Game system having a plurality of game machines for competing for card data |
US7505921B1 (en) * | 2000-03-03 | 2009-03-17 | Finali Corporation | System and method for optimizing a product configuration |
US8266539B2 (en) * | 2000-04-28 | 2012-09-11 | Intel Corporation | Enabling hypertext elements to work with software applications |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
PT97283A (en) * | 1991-04-08 | 1992-06-30 | Alcino Gomes Dos Santos | RECESSER OF CENIC ENVIRONMENTS |
KR960004813B1 (en) * | 1992-10-06 | 1996-04-13 | 엘지전자주식회사 | Scent occurring television receiver |
US6161126A (en) * | 1995-12-13 | 2000-12-12 | Immersion Corporation | Implementing force feedback over the World Wide Web and other computer networks |
US6018764A (en) * | 1996-12-10 | 2000-01-25 | General Instrument Corporation | Mapping uniform resource locators to broadcast addresses in a television signal |
DE29922084U1 (en) * | 1999-12-09 | 2000-03-23 | Stroetmann Christian | Plug-in module for remote regulation and control |
-
2002
- 2002-05-02 CN CNB028016092A patent/CN1232322C/en not_active Expired - Fee Related
- 2002-05-02 AT AT02724568T patent/ATE542577T1/en active
- 2002-05-02 JP JP2002589098A patent/JP4601255B2/en not_active Expired - Fee Related
- 2002-05-02 WO PCT/IB2002/001535 patent/WO2002092182A1/en active Application Filing
- 2002-05-02 KR KR10-2003-7000389A patent/KR20030021250A/en not_active Application Discontinuation
- 2002-05-02 EP EP02724568A patent/EP1401547B1/en not_active Expired - Lifetime
- 2002-05-10 US US10/143,642 patent/US7740531B2/en active Active
-
2009
- 2009-11-19 US US12/621,644 patent/US20100062860A1/en not_active Abandoned
Patent Citations (41)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5291556A (en) * | 1989-10-28 | 1994-03-01 | Hewlett-Packard Company | Audio system for a computer display |
US5004225A (en) * | 1989-11-01 | 1991-04-02 | Yuri Krukovsky | Simulated observation flight apparatus |
US6154553A (en) * | 1993-12-14 | 2000-11-28 | Taylor Group Of Companies, Inc. | Sound bubble structures for sound reproducing arrays |
US5771041A (en) * | 1994-06-03 | 1998-06-23 | Apple Computer, Inc. | System for producing directional sound in computer based virtual environment |
US5855483A (en) * | 1994-11-21 | 1999-01-05 | Compaq Computer Corp. | Interactive play with a computer |
US5956038A (en) * | 1995-07-12 | 1999-09-21 | Sony Corporation | Three-dimensional virtual reality space sharing method and system, an information recording medium and method, an information transmission medium and method, an information processing method, a client terminal, and a shared server terminal |
US6553121B1 (en) * | 1995-09-08 | 2003-04-22 | Fujitsu Limited | Three-dimensional acoustic processor which uses linear predictive coefficients |
US5982903A (en) * | 1995-09-26 | 1999-11-09 | Nippon Telegraph And Telephone Corporation | Method for construction of transfer function table for virtual sound localization, memory with the transfer function table recorded therein, and acoustic signal editing scheme using the transfer function table |
US5695400A (en) * | 1996-01-30 | 1997-12-09 | Boxer Jam Productions | Method of managing multi-player game playing over a network |
US5991693A (en) * | 1996-02-23 | 1999-11-23 | Mindcraft Technologies, Inc. | Wireless I/O apparatus and method of computer-assisted instruction |
US5795228A (en) * | 1996-07-03 | 1998-08-18 | Ridefilm Corporation | Interactive computer-based entertainment system |
US6167353A (en) * | 1996-07-03 | 2000-12-26 | Interval Research Corporation | Computer method and apparatus for interacting with a physical system |
US5982445A (en) * | 1996-10-21 | 1999-11-09 | General Instrument Corporation | Hypertext markup language protocol for television display and control |
US6056640A (en) * | 1996-11-11 | 2000-05-02 | Schaaij; Johan Michiel | Computerized gaming apparatus |
US5887118A (en) * | 1997-02-12 | 1999-03-23 | Motorola, Inc. | Olfactory card |
US5945986A (en) * | 1997-05-19 | 1999-08-31 | University Of Illinois At Urbana-Champaign | Silent application state driven sound authoring system and method |
US6106399A (en) * | 1997-06-16 | 2000-08-22 | Vr-1, Inc. | Internet audio multi-user roleplaying game |
US5964660A (en) * | 1997-06-18 | 1999-10-12 | Vr-1, Inc. | Network multiplayer game |
US6288716B1 (en) * | 1997-06-25 | 2001-09-11 | Samsung Electronics, Co., Ltd | Browser based command and control home network |
US6243707B1 (en) * | 1997-06-25 | 2001-06-05 | Samsung Electronics Co. Ltd. | Method and apparatus for creating home network macros |
US6166496A (en) * | 1997-08-26 | 2000-12-26 | Color Kinetics Incorporated | Lighting entertainment system |
US20030076281A1 (en) * | 1997-08-26 | 2003-04-24 | Frederick Marshall Morgan | Diffuse illumination systems and methods |
US6050822A (en) * | 1997-10-01 | 2000-04-18 | The United States Of America As Represented By The Secretary Of The Army | Electromagnetic locomotion platform for translation and total immersion of humans into virtual environments |
US6007338A (en) * | 1997-11-17 | 1999-12-28 | Disney Enterprises, Inc. | Roller coaster simulator |
US6159100A (en) * | 1998-04-23 | 2000-12-12 | Smith; Michael D. | Virtual reality game |
US6993722B1 (en) * | 1999-02-08 | 2006-01-31 | Cirrus Logic, Inc. | User interface system methods and computer program products for multi-function consumer entertainment appliances |
US7100118B1 (en) * | 1999-03-05 | 2006-08-29 | Amulet Technologies, Llc | Graphical user interface engine for embedded systems |
US6498955B1 (en) * | 1999-03-19 | 2002-12-24 | Accenture Llp | Member preference control of an environment |
US6354940B1 (en) * | 1999-04-30 | 2002-03-12 | Square Co., Ltd. | Game apparatus, game fight processing method, and computer readable storage medium storage program therefor |
US6379253B1 (en) * | 1999-06-08 | 2002-04-30 | Konami Co., Ltd. | Game device and computer-readable storage medium |
US6826725B1 (en) * | 1999-12-16 | 2004-11-30 | Microsoft Corporation | Techniques for invoking system commands from within a mark-up language document |
US7505921B1 (en) * | 2000-03-03 | 2009-03-17 | Finali Corporation | System and method for optimizing a product configuration |
US8266539B2 (en) * | 2000-04-28 | 2012-09-11 | Intel Corporation | Enabling hypertext elements to work with software applications |
US7111297B1 (en) * | 2000-05-02 | 2006-09-19 | Microsoft Corporation | Methods and architectures for resource management |
US7118482B2 (en) * | 2000-05-29 | 2006-10-10 | Nintendo Co., Ltd. | Game system using game cards and game machine |
US20040024656A1 (en) * | 2000-06-02 | 2004-02-05 | Coleman Kevin B. | Interactive product selector with inferential logic engine |
US7140033B1 (en) * | 2000-06-21 | 2006-11-21 | Bellsouth Intellectual Property Corporation | Methods and systems for controlling consumer electronics external devices via data delivered to a device |
US20020087988A1 (en) * | 2000-12-28 | 2002-07-04 | International Business Machines Corporation | Interactive TV contextual content surfing using inter-channel hyperlinking: systems, methods & program products |
US20020110788A1 (en) * | 2001-02-09 | 2002-08-15 | Henry Michael L. | Acoustical enhancement of audio sound effects playback in video games |
US20030018755A1 (en) * | 2001-03-30 | 2003-01-23 | Masterson Robert J. | Online system that facilitates configuration and administration of residential electronic devices |
US7147561B2 (en) * | 2003-08-20 | 2006-12-12 | Nintendo Co., Ltd. | Game system having a plurality of game machines for competing for card data |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090322955A1 (en) * | 2006-06-13 | 2009-12-31 | Takuya Iwanami | Data transmitting device, data transmitting method, audio-visual environment control device, audio-visual environment control system, and audio-visual environment control method |
US20090280896A1 (en) * | 2006-06-19 | 2009-11-12 | Ambx Uk Limited | Game enhancer |
US8376844B2 (en) * | 2006-06-19 | 2013-02-19 | Ambx Uk Limited | Game enhancer |
Also Published As
Publication number | Publication date |
---|---|
US20020169012A1 (en) | 2002-11-14 |
EP1401547B1 (en) | 2012-01-25 |
US7740531B2 (en) | 2010-06-22 |
EP1401547A1 (en) | 2004-03-31 |
ATE542577T1 (en) | 2012-02-15 |
JP2004520918A (en) | 2004-07-15 |
JP4601255B2 (en) | 2010-12-22 |
CN1232322C (en) | 2005-12-21 |
WO2002092182A1 (en) | 2002-11-21 |
CN1462201A (en) | 2003-12-17 |
KR20030021250A (en) | 2003-03-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20100062860A1 (en) | Operation of a set of devices | |
US8176115B2 (en) | Real-world representation system and language | |
US8548613B2 (en) | System and method for an interactive device for use with a media device | |
JP4097651B2 (en) | Game machine | |
US20100077293A1 (en) | Enabled device and a method of operating a set of devices | |
US20110218039A1 (en) | Method for generating an effect script corresponding to a game play event | |
US10625153B2 (en) | Lighting for video games | |
CN109683839B (en) | Method, equipment and system for split screen display and multi-terminal interaction | |
US8271945B2 (en) | Assets and effects | |
US20080305713A1 (en) | Shadow Generation Apparatus and Method | |
CN114189743B (en) | Data transmission method, device, electronic equipment and storage medium | |
JP2007241949A (en) | Image generation device, game machine, image generation method, and program | |
CN110632898B (en) | Deduction system | |
KR20050113225A (en) | Selectable real-world representation system descriptions | |
CN117440184B (en) | Live broadcast equipment and control method thereof | |
KR101833272B1 (en) | Song Accompaniment System for Enhancing Vibrancy | |
CN116828248A (en) | Scene adjustment method, device, equipment and storage medium | |
CN116543459A (en) | Intelligent landscape lighting method and system and electronic equipment | |
CN114011067A (en) | Game fighting method and device, electronic equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |