[go: nahoru, domu]

US20030040361A1 - Method and apparatus for generating tactile feedback via relatively low-burden and/or zero burden telemetry - Google Patents

Method and apparatus for generating tactile feedback via relatively low-burden and/or zero burden telemetry Download PDF

Info

Publication number
US20030040361A1
US20030040361A1 US10/201,584 US20158402A US2003040361A1 US 20030040361 A1 US20030040361 A1 US 20030040361A1 US 20158402 A US20158402 A US 20158402A US 2003040361 A1 US2003040361 A1 US 2003040361A1
Authority
US
United States
Prior art keywords
telemetry
tactile feedback
generated
application
tactile
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/201,584
Inventor
Craig Thorner
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=46280912&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=US20030040361(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Priority claimed from US08/309,763 external-priority patent/US5684722A/en
Priority claimed from US08/409,327 external-priority patent/US5669818A/en
Priority claimed from US08/935,762 external-priority patent/US6422941B1/en
Application filed by Individual filed Critical Individual
Priority to US10/201,584 priority Critical patent/US20030040361A1/en
Publication of US20030040361A1 publication Critical patent/US20030040361A1/en
Priority to US11/929,426 priority patent/US8328638B2/en
Priority to US12/105,007 priority patent/US8308558B2/en
Priority to US13/711,585 priority patent/US8834266B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/24Constructional details thereof, e.g. game controllers with detachable joystick handles
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/25Output arrangements for video game devices
    • A63F13/28Output arrangements for video game devices responding to control signals received from the game device for affecting ambient conditions, e.g. for vibrating players' seats, activating scent dispensers or affecting temperature or light
    • A63F13/285Generating tactile feedback signals via the game input device, e.g. force feedback
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/50Controlling the output signals based on the game progress
    • A63F13/54Controlling the output signals based on the game progress involving acoustic signals, e.g. for simulating revolutions per minute [RPM] dependent engine sounds in a driving game or reverberation against a virtual wall
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/016Input arrangements with force or tactile feedback as computer generated output to the user
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B9/00Simulators for teaching or training purposes
    • G09B9/02Simulators for teaching or training purposes for teaching control of vehicles or other craft
    • G09B9/04Simulators for teaching or training purposes for teaching control of vehicles or other craft for teaching control of land vehicles
    • G09B9/05Simulators for teaching or training purposes for teaching control of vehicles or other craft for teaching control of land vehicles the view from a vehicle being simulated
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/22Setup operations, e.g. calibration, key configuration or button assignment
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/10Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
    • A63F2300/1037Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals being specially adapted for converting control signals received from the game device into a haptic signal, e.g. using force feedback
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/30Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by output arrangements for receiving control signals generated by the game device
    • A63F2300/302Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by output arrangements for receiving control signals generated by the game device specially adapted for receiving control signals not targeted to a display device or game input means, e.g. vibrating driver's seat, scent dispenser
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/80Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game specially adapted for executing a specific type of game
    • A63F2300/8082Virtual reality

Definitions

  • This invention relates to tactile feedback systems for computer and video game, simulation, and training systems, and, more particularly, to a method for using telemetry to generate and/or shape the tactile feedback generated by said systems, such that real-time tactile feedback is provided to enhance a user's or spectator's experience while interacting with said systems, and at a zero burden or a heretofore impossibly low burden to software developers, yet producing very effective results.
  • AudioSense® is a sound analysis technology that generates tactile feedback in real time without developer support of any kind.
  • AudioSense® has inherent limitations that are overcome and/or minimized by intelliVIBE®, which is the type of tactile feedback that has typically required developer support—and therefore, intelliVIBE® suffers from the very limitations described previously herein.
  • AudioSense® audio analysis technology allows tactile feedback hardware to function with no support whatsoever from any given developer of any given software title, application, or hardware system
  • this telemetry based method allows intelliVIBE® (and other similar interfaces) to self-generate effective and thoughtful tactile feedback with minimal, simple, and one-time-only support by any given developer.
  • the teachings herein can be applied to other areas, such as remotely piloted vehicles, remotely controlled vehicles, UAVs (Unmated Aerial Vehicles), spacecraft, and other types of tele-operated or tele-presense vehicles, where telemetry is available or can be provided, thereby producing effective tactile feedback for the operators or spectators of such vehicles.
  • the present invention provides a method and corresponding apparatus for providing tactile feedback for computer and video game, training, and simulation systems. More specifically, the tactile feedback system comprises a software application, a tactile feedback controller, and a plurality of tactile sensation generators.
  • the innovative method herein provides that any given software application can provide simple telemetry about its current state to a shared memory data structure, and subsequently leave all the burden of producing tactile feedback to an externally operating thread that employs general purpose algorithms and graphically adjustable scaling factors and configuration options to yield effective and appropriate tactile feedback without further effort by the given software application. It is a simple one-time only operation for any given software developer to implement the telemetry support. Thusly, the tactile feedback so produced is achieved with very low burden to the software developer.
  • the innovative method herein has the means to self-generate its own telemetry by examining the state of the host computer's control input devices (i.e., keyboard, mouse, joystick, etc.), and to use said self-generated telemetry as the basis for applying real-time directional scaling to tactile feedback simultaneously produced via real-time audio analysis.
  • the present invention can produce tactile feedback via the combination of real-time audio analysis with simultaneous directional scaling produced by real-time self-generated telemetry in combination with pre-defined general-purpose algorithms. This capability is achieved at a zero burden, that is, with no effort whatsoever nor any acknowledgement of its existence from the perspective of the developers of any given software application, yet yields results that resemble very carefully and thoughtfully crafted tactile feedback.
  • the tactile feedback method and apparatus is provided in a preferred embodiment comprised by (1) a software application (currently known as the Tactile Feedback System Control Panel or TFS Control Panel), which configures the Tactile Feedback System (TFS) and handles the intelliVIBE® modules, each of which is adapted to handle the telemetry provided by any given software title, application, or system, and provides a graphical user interface for adjusting the available tactile feedback parameters that descend from the provided telemetry, (2) a tactile feedback controller (the presently preferred embodiment is currently known as the TFS2 (Tactile Feedback System version 2.0) which incorporates cooperative AudioSense® and intelliVIBE® capabilities, and (3) a plurality of tactile sensation generators, the currently preferred embodiment of which is a tactile feedback seat known as the TFSU (Tactile Feedback Seating Unit) and a tactile feedback vest (U.S. Pat. No. 5,565,840) known by the trademark name FLAKjaketTM.
  • a software application currently known as the Tactile Feedback System Control Panel or TFS Control Panel
  • FIG. 1 depicts a block diagram of an illustrative configuration of the present invention
  • FIG. 2 is a flow chart for using illustrative application generated telemetry to generate tactile feedback for a driving simulation
  • FIG. 3 is a flow chart for using illustrative application generated telemetry to generate tactile feedback for a flight simulation
  • FIG. 4 is a flow chart for using illustrative self-generated telemetry to generate scaling factors for directionally scaling tactile feedback simultaneously produced via audio analysis;
  • FIG. 5 shows an illustrative group of graphical controls for adjusting various telemetry scaling factors
  • FIG. 5A depicts the graphical controls of FIG. 5, specifically illustrative RPM telemetry scaling factors
  • FIG. 5B depicts the graphical controls of FIG. 5, specifically illustrative SPEED telemetry scaling factors
  • FIG. 5C depicts the graphical controls of FIG. 5, specifically illustrative TAKE OFF telemetry scaling factors
  • FIG. 5D depicts the graphical controls of FIG. 5, specifically illustrative ROLL telemetry scaling factors
  • FIG. 5E depicts the graphical controls of FIG. 5, specifically illustrative G-FORCE telemetry scaling factors
  • FIG. 5F depicts the graphical controls of FIG. 5, specifically illustrative WEAPONS telemetry scaling factors
  • FIG. 5G depicts the graphical controls of FIG. 5, specifically illustrative HOTKEYS telemetry scaling factors
  • FIG. 5H depicts the graphical controls of FIG. 5, specifically illustrative AFTERBURNER telemetry scaling factors
  • FIG. 6 depicts an aircraft and the 4 quadrants used as the basis for an innovative aircraft roll algorithm
  • FIG. 7 depicts the actual mathematical algorithm used to generate an aircraft's roll scaling factor based upon the aircraft's current quadrant as in FIG. 6, and how to utilize said scaling factor once derived from the roll telemetry data;
  • FIG. 1 depicts a block diagram of a universal tactile feedback system 100 interacting with a host computer 102 .
  • the host computer 102 is connected to the universal tactile feedback controller 200 by an audio connection 112 and/or digital connection 116 .
  • the universal tactile feedback controller 200 is connected to multiple independent tactile sensation generators 500 by connection 460 .
  • the tactile sensation generators 500 collectively refer to hardware that produces physical sensations that one can feel while interacting with the software application 104 .
  • no specific limitation is intended for any or all of the connections 112 , 116 , 460 with regard to their particular form, whether analog, digital, wired, wireless, and so on.
  • the host computer 102 is presumed to be the host for some desired software application 104 .
  • the software application 104 is the application for which the present invention will generate tactile feedback. If the developers of the given software application 104 have supported the simple telemetry requirement, there will exist application generated telemetry 600 .
  • This application-generated telemetry 600 is a simple data structure that contains information about the state of operation of said software application 104 .
  • typical telemetry would include the car's engine RPM, the car's MPH, the car's throttle position, lateral G-forces, acceleration, deceleration, brake pedal position, steering wheel deflection, current gear, surface type (i.e., is the car currently on asphalt, dirt, gravel, sand, wet asphalt, icy asphalt, snow, mud, etc.), and damage flags (i.e., are there any flat tires, deformations of the vehicle's body panels, or equipment failures, etc.).
  • This illustrative telemetry is by no means comprehensive, but should be sufficient to illustrate the type of data common to car racing games and/or simulations.
  • the telemetry is entered into a shared data structure by a common memory mapping technique readily understood by those skilled in the art.
  • the computer host is running on a Microsoft Windows platform wherein a predetermined shared memory area name is agreed and known to both the software application 104 and the telemetry processor 700 .
  • the software application 104 creates the telemetry area by opening a shared memory area with the predetermined shared memory are name.
  • the telemetry processor detects the shared memory area by said predetermined name, and uses the standard Windows coding “MapViewOfFile” to map the application-generated telemetry 600 into the memory of the telemetry processor 700 .
  • This simple operation affords the telemetry processor 700 access to the real-time state data (telemetry) of the software application 104 .
  • it allows the telemetry processor 700 to know what is going on inside the software application 104 , to the extent and detail inherent in the data structure provided by the software application 104 for the benefit of the application generated telemetry 600 shared memory structure, and ultimately for the benefit of the owners of the universal tactile feedback system 100 and the users of the software application 104 .
  • Providing the application generated telemetry 600 is a very simple operation for any reasonably skilled software developer, and even if new to the entire concept, most often it is a process of only several tens of minutes to effectuate the shared memory area once the data structure that is to be mapped there is known.
  • the telemetry processor 700 can examine the state of the host computer 102 in real time and generate useful telemetry by eavesdropping on the state of the host computer's 102 keyboard, mouse, joystick, and the like.
  • This raw self-generated telemetry 800 is an effective substitute for application-generated telemetry; however, the application-generated telemetry is much preferred if it is available.
  • the universal tactile feedback system runs in its host-independent audio analysis mode of operation, which is known by the registered trademark name AudioSense® (U.S. Trademark Registration Number 2261526, registered Jul. 14, 1999).
  • AudioSense® autonomously generates useful “base level” tactile feedback, which is then “directionally scaled” by the self-generated telemetry 800 that is culled by the telemetry processor 700 from the host computer 102 .
  • AudioSense® has no capabilities as of this writing to apply left, and/or right, and/or front, and/or back directional scaling based upon its audio analysis. While an effective technology, it is greatly enhanced by directional scaling from self-generated telemetry 800 .
  • the presumption is that if the software application 104 provides application-generated telemetry 600 for the benefit of the telemetry processor, most often AudioSense® will ideally be disabled while such telemetry is available. However, AudioSense® can work simultaneously and cooperatively with application generated telemetry 600 as well, sometimes lending an “analog flavor” to the tactile feedback produced by the universal tactile feedback system 100 .
  • the telemetry processor 700 provides a graphical user interface 900 for configuring and adjusting the way the available telemetry is utilized.
  • the telemetry processor 700 is itself a software application that manages the available telemetry and performs the processing that converts said telemetry into a useful data stream for generating and/or shaping tactile feedback. More detail will be provided on this later in this disclosure.
  • the telemetry processor 700 will process telemetry from a typical car based game or simulation.
  • the telemetry processor 700 begins at the START 702 .
  • the telemetry processor 700 accepts the telemetry via a shared memory mapped data structure.
  • the telemetry processor 700 is given the following illustrative values: the simulated car's engine RPM, vehicle MPH, throttle position, lateral g-force, forward acceleration and deceleration, the car's brake position, the car's steering wheel deflection, the current transmission gear, the surface type that the vehicle is currently on, and damage flags that might indicate various damage typical of cars, such as flat tires, body damage, engine failure, transmission failure, and the like. All of the received telemetry is in a “raw” state, that is, in a format native to the software application 104 that generated it. In step 706 , scaling factors are applied to “normalize” the telemetry.
  • the engine might idle at 700 RPM and might redline at 5500 RPM. However, in another car simulation, the engine might idle at 1500 RPM and redline at 9000 RPM.
  • a specific constant RPM scaling factor might be used to always scale the RPM range for any given simulated car to a range of 0 to 100, and then, a baseline constant may be added, for example, 20. This would mean as long as the engine is started, the lowest scaled RPM value would be 0+20, or 20, where the highest RPM value would be 100+20, or 120. In this way, the “raw” telemetry will have been scaled into a useful range of 20 to 120.
  • the baseline RPM constant +20 would be skipped, thereby leaving the RPM at 0. In this way, a useful RPM value is created.
  • another scaling layer is applied from a graphical user interface 900 that is provided by the telemetry processor 700 .
  • the graphical user interface 900 allows the user to make the RPM value more or less prevalent in the tactile feedback that is perceived.
  • the RPM scaling factor would be a floating-point decimal value from 0.00 up to 2.00, including all interim values. Let's say the user leaves the RPM scaling factor in the graphical interface 900 at 1.00.
  • the final scaled RPM range would be 20-120, or zero if the simulated car's engine was off or disabled. If the user set the graphical RPM scaling value to 1.50, the final scaled RPM range would be 20 times 1.5, or 30, up to 120 times 1.5, or 180, giving a final range of 30 to 180. In this way, by adjusting a simple control, the original unadulterated raw RPM value as generated by the simulation is converted to a useful and preferred strength. A similar process is used for all the telemetry variables, such that the relative strength of each can be tweaked to produce useful results. However, all the final results descend directly from the raw telemetry as calculated by the original software application 104 as mapped into the application generated telemetry 600 shared memory area. More detail will be examined later in this disclosure.
  • step 708 we determine the baseline feedback level from the fundamental telemetry variables engine RPM, vehicle MPH, and throttle position. These are non-directionally scaled forces, which is why they together comprise what is known as the “baseline” tactile feedback level.
  • the final scaled engine RPM value is calculated.
  • MPH With regard to MPH, lets assume in one game the given car has a speed range of 0 to 130 MPH, and in another game or simulation, the car has a speed range of 0 to 236 MPH. In the case where the top speed is 130 MPH, a constant scaling factor would be the current speed always divided by 1.3 (or rather multiplied by 0.769) such that the speed range was reduced to 0 to 100.
  • the constant scaling factor would be the current speed divided by 2.36 (or rather multiplied by 0.424), thereby yielding the same range 0 to 100.
  • a MPH scaling factor is provided where the user can increase or decrease the importance of the MPH by adjusting a multiplier with a graphical scroll bar, where the scaled MPH range now from 0 to 100 would be multiplied by any selected number 0.00 to 2.00, in the same way was the RPM was previously.
  • the next step is to compare the final RPM result and MPH result, and to accept the highest value as the baseline tactile feedback value for that cycle through the telemetry processor 700 , which typically runs at hundreds of cycles second.
  • This way if you are going 0 MPH but your simulated engine has produced a final RPM vibration value of 88, you'll feel tactile feedback at a “strength” of 88. Again, this is illustrative only.
  • the telemetry processor factors in the throttle position, because RPM or MPH should both feel stronger if you are full on the throttle, or not on the throttle at all. For example, if your throttle position runs from 0.00 at no throttle to 1.00 at full throttle, we may set a first scaling factor of “plus 0.60”, that is, adding 0.60 to each end of the throttle range, which changes the throttle scaling factor from 0.00-1.00 to 0.60-1.60. Now, the graphical scaling factor is utilized to change that range yet again according to the user's preference.
  • the final throttle scaling factor becomes 0.82 times 0.60 (or 0.492) at the low end, and 0.82 times 1.60 (or 1.312) at the high end, so the final range is 0.492 to 1.312, depending on the position of the throttle.
  • the raw telemetry calculated by the software application 104 has produced tactile feedback that is indicative of the engine RPM, vehicle MPH, and driver's throttle position. This is the “baseline” tactile feedback.
  • the AudioSense® audio analysis mode of operation of the universal tactile feedback controller 200 would have provided this baseline tactile feedback level. The importance of this cannot be stressed enough, because directional scaling is completely worthless if it is not based on an effective “baseline”.
  • the baseline tactile feedback values are now scaled according to the lateral gforce, acceleration, and deceleration telemetry.
  • the unique nature of the preferred embodiment is that the tactile feedback seat in the parent cases has multiple zones, that is, a back zone, a left leg zone, and a right leg zone.
  • Much like stereo speakers can create the illusion of a sound originating from any point between only two speakers by biasing the amplitude of the broadcast sound toward one speaker, and reducing it at the other, the tactile feedback seat functions in much the same way. If we are in a left turn, simple math is applied to increase the tactile feedback felt in the right leg zone, and to decrease the tactile feedback in the left leg zone.
  • the tactile feedback for the right leg is be multiplied by 2.3, while the left leg is divided by 2.3.
  • the scaling factors would change the numerical range of the G's to bring them into a useful range. In this way, the driver of the simulated car feels tactile cues that indicate he or she is being thrown to the outside of the turn, in direct proportion to the lateral g-forces involved.
  • tactile feedback in the back zone is increased (i.e., multiplied by the scaled telemetry acceleration value) while tactile feedback in the legs is simultaneously reduced (i.e., divided by the scaled telemetry acceleration value), such that the illusion is provided that the driver is being thrown backward into the seat.
  • tactile feedback is increased at the legs and simultaneously reduced at the back, such that the illusion is provided that the driver is being thrown forward in the seat, as previously in direct proportion to the g-force values involved.
  • the first layer of scaling factors work in the same way to bring the raw telemetry into a useful range, while the graphical interface is used to set the user's preference as to how strong this directional scaling is to be perceived.
  • tactile feedback texturing is applied to the now directionally scaled tactile feedback values.
  • Texturing adds choppiness and other characteristics to the tactile feedback. For example, if a “damage flag” indicates that there is a flat tire on the left side of the car, there may be a “flat tire scaling factor” provided in the graphical user interface 900 within the telemetry processor 700 that sets the flat tire scaling factor as 1.5 times (1.5 ⁇ ).
  • the texturing to the left leg zone in the tactile feedback seat (the side of the vehicle with the simulated flat tire) would now be periodically multiplied by 1.5 ⁇ in some time period that is directly proportional to the current MPH of the car.
  • step 714 we transmit the current tactile feedback results to the universal tactile feedback controller 200 with the standard intelliVIBE® protocol, but no limitation should be inferred about which types of tactile feedback devices can benefit from this telemetry based approach.
  • Steps 722 , 724 , 726 , 728 are analogous to steps 702 , 704 , 706 , 708 of FIG. 2. However, for step 730 , we refer now to FIG. 6.
  • the major difference here is the way the telemetry processor handles g-forces, pitch, and roll, which are data typical of flight simulations.
  • the innovative approach here allows two possibilities for rendering aircraft rolls, that is, when an aircraft such as a modem military jet rolls on its center axis in a continuous right or left turn. In reality, when a jet enters a continuous roll to the right, the passengers get pushed to the left side and stay there due to the g-forces involved that are stable for the duration of the rightward roll. This is calculated just like the lateral g-forces of the car of FIG. 2. However, an aircraft's roll can be divided into four quadrants, each of which affects the rendering of the roll in a different way. Rather than a continuous push to one side, which is a realistic and readily available but perhaps a dull representation, we can artificially induce additional aircraft orientation data in the tactile feedback that is produced.
  • FIG. 6 an aircraft 740 is depicted with its tail 742 pointing straight up, that is, the aircraft is perfectly level.
  • the current quadrant is determined by the orientation of the aircraft's tail 742 .
  • FIG. 7 the quadrants 1, 2, 3, and 4 are analogous to the same quadrants of FIG. 6.
  • FIG. 7 explains the simple algorithm by which the right leg tactile feedback value and the left leg tactile feedback value are scaled depending in which quadrant the aircraft's tail is oriented.
  • the users have additional scaling factors to augment of reduce the prevalence of the g-forces.
  • negative g-forces when experiencing negative g-forces, one gets the feeling of being more weightless than is normal. Therefore, when negative g-forces are present, the negative g-force value is multiplied by the back value to increase the back zone tactile feedback, while simultaneously used as a divisor to mathematically reduce the tactile feedback to the legs, making one feel like they are rising out of the tactile feedback seat.
  • the telemetry processor 700 does not have the benefit of application-generated telemetry. Therefore, the universal tactile feedback controller 200 will run in its audio analysis mode of operation (AudioSense®).
  • AudioSense® the baseline tactile feedback is produced without capabilities for directional biasing, regardless as to the forces acting on the user's vehicle or person in any particular application. Simply put, the tactile feedback system has no awareness as to the specific calculations that are necessarily taking place in any particular application.
  • the telemetry processor begins in step 782 and in step 784 examines the telemetry it can derive from the host computer's 102 control input devices.
  • the user of a non-telemetry supportive title i.e., no intelliVIBE® support
  • the telemetry processor can track the movement of the player's joystick and can know its deflection from center on both the X and Y axes, that is, left and right (steering) and forward and back (acceleration and brake) of the simulated vehicle.
  • the telemetry processor 700 will do the best it can with what data it can find.
  • the user will adjust the joystick scaling factors such that the longer the accelerator axis (presumably the Y axis, but configured as such by the user) is pressed forward, the more speed the telemetry processor will assume will have been gained by the simulated vehicle.
  • the simulated vehicle is assumed to be going faster, it will automatically adjust the X-axis scaling factors (the car's steering) to make the lateral g-forces more prevalent.
  • the telemetry processor will indicate this in the stream of scaling factors that are being broadcast to the universal tactile feedback controller 200 such that a deceleration effect is produced by directionally biasing the tactile feedback in the seat to the legs section of the tactile feedback seating unit, and away from the back section.
  • the scaling factors are calculated in real time based upon assumed data, so they do not have the finesse of real-time accurate telemetry that would ideally be available.
  • the scaling factors are 8 bit values, (0-255), wherein whatever value given as a scaling factor is divided by 32 within the universal tactile feedback controller 200 .
  • the scaling factors exist in a data table, where each single scaling factor is applied to only one of eight zones: back, left leg, right leg, chest, left hand, right hand, left foot, right foot. These zones are illustrative and not assumed to match this presently preferred embodiment. Sending a scaling factor of 32 yields no change to a particular tactile feedback zone because of the subsequent divide by 32 that occurs within the universal tactile feedback controller 200 (i.e., 32 divided by 32 equals 1.0).
  • a scaling factor of 16 sent to the left leg will cut the tactile feedback there in half, because 16/32 yields a scaling factor of 0.5 ⁇ .
  • the divide by 32 algorithm is used to yield very fast mathematical performance by using a shortcut rather than full floating point math which would impede the performance of AudioSense® if the microcontroller inside the universal tactile feedback controller 200 does not offer full hardware multiply and divide.
  • an effect will be rendered according to the scaling parameters in effect for the cannon.
  • the intelliVIBE® cannon effect will be disabled until reset by the user.
  • AudioSense® will still do its best to render the cannon by interpreting the sound of it firing if the cannon is not yet depleted of ammunition.
  • the telemetry processor can assume that the aircraft is airborne, and can then do its best to use the joystick position to send scaling factors to the left leg, right leg, and back zones of the tactile feedback seat in order to simulate the G-forces and bank roll effects that would ideally be more accurately prompted by fully supported telemetry.
  • the combination of cooperative AudioSense® with intelliVIBE® scaling is a very effective means of self-generating very appropriate tactile feedback, regardless of the lack of support by any given application.
  • FIG. 5 a group of user adjustable telemetry scaling factors is presented. These controls are illustrative only for the purposes of understanding. Getting a closer look at FIG. 5A, the RPM baseline tactile feedback generator is presented. The baseline setting of 30 establishes the basic level of prevalence of engine RPM induced tactile feedback. The multiplier at the bottom shows 0.7, which is used to scale the 30 baseline setting. If the aircraft is known to be on the ground, the engine RPM setting is multiplied by 0.9, whereas if the aircraft is in the air, the engine RPM is multiplied by 0.6, thereby making the engine RPM feel weaker once in the air.
  • the speed value is multiplied by 0.8 if the aircraft is on the ground, while it is multiplied by 0.08 if the aircraft is in the air.
  • This has the effect of making the speed tactile effect ten times weaker when the aircraft is airborne, but the typical maximum speed when on the ground might be only 180 MPH, but when airborne, a modern military aircraft can travel at 1,300 MPH and higher.
  • These scaling factors allow the tactile feedback to be matched to a useful range regardless as to the performance envelope of any particular aircraft, simply by adjusting the slider knob of the corresponding scaling factor control.
  • ground texturing is set to begin at 30 MPH and end at 250 MPH, which is well past a conservative take-off speed. Ultimately as the aircraft goes faster, the ground will create more and more texturing, that is, bumpiness in the tactile feedback. Upon takeoff, the ground texturing will be reduced to 25% (multiplied by 0.25 ⁇ ) until the take off speed is surpassed, which then nullifies the ground texturing, until the aircraft comes in for a landing, at which time the ground texturing will automatically take effect yet again.
  • FIG. 5D presents the bank scaling value.
  • the 1.0 as shown lets the bank scaling take place just as the calculations in FIG. 7 indicate. However, if the user wants the bank scaling to be stronger, they only need to increase the corresponding 1.0 scale factor setting. To reduce bank scaling, setting bank scaling to toward 0 will nullify the effect accordingly.
  • FIG. 5E presents the positive and negative G-force multipliers. In positive g-force conditions, these settings multiply the legs by 1.3 times the g-force value, and multiply the back by 2.5 times the g-force value. In negative G-force conditions, the back is reduced by half and the legs are scaled to 80% for the duration of the negative G-force condition.
  • the scaling factors for the main gun are available. During the firing of the main gun, 100 pulses per second will occur in the tactile feedback, alternating between a high pulse percentage of 100% and a low pulse percentage of 12% as determined by the scaling factors.
  • the “use high intensity” check box allows the main gun to temporarily over-ride the current intensity setting for the duration of the main gun being fired. Knowing that the main gun is being fired can be given to us by application generated telemetry 600 or self-generated telemetry 800 , same as with all the effects and scaling factors of FIG. 5A through FIG. 5H.
  • the telemetry processor 700 is told what key press equals what known effect. This is configured by the user. For example, we can see in the upper left corner that the “G” key activates the landing gear. This way, if the user presses the G key to deploy the landing gear, the telemetry processor can run a simple effect representing the landing gear. If telemetry is not provided that indicates the true state of the landing gear, the telemetry processor 700 will fall back on its self-generating capability and keep track of the G key presses on its own and will do its best to stay in sync with the simulation in question.
  • FIG. 5H an illustrative afterburner effect is shown.
  • Many modem jets have multiple stages of afterburner.
  • FIG. 5H there are 3 stages of afterburner, where stages one, two, and three yield tactile “thumps” at 60%, 80%, and 100% strength, respectively, each lasting only 0.01 seconds, but also incorporating an RPM scaling factor of 1.33 ⁇ , 1.5 ⁇ , or 1.66 ⁇ times respectively, which will increase the prevalence of the engine RPM according to what level of afterburner the aircraft is in, for as long as the afterburner is active.
  • the innovative methods and apparatus of the present invention allows for the most versatile tactile feedback results to be produced, both with and without support of any kind by the host computer 102 and the executing software application 104 .
  • the best condition is for fully detailed low-burden telemetry to be provided for the benefit of the universal tactile feedback controller 200 by a given developer.
  • the telemetry processor 700 of the present invention is able to generate data on its own to supplement the AudioSense® audio analysis mode of operation with real time directional biasing of all tactile feedback, in the end producing effective zero-burden feedback from the point of the game developer . . . in the end eliminating the fundamental problem of lack of quality developer support!

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Educational Administration (AREA)
  • Educational Technology (AREA)
  • Business, Economics & Management (AREA)
  • Acoustics & Sound (AREA)
  • User Interface Of Digital Computer (AREA)
  • Toys (AREA)

Abstract

A method and apparatus for providing tactile feedback for computer and video game systems is disclosed which provides real time tactile feedback to enhance a user's experience while interacting with a computer/video game or simulation. The method relies on application-generated telemetry that is generated by an executing application for the benefit of a telemetry processing method, and/or telemetry that is self-generated by said telemetry processor when no telemetry has been provided by said executing application. The application-generated telemetry is used to produce tactile feedback that is derived directly from the internal calculations of said application, where said telemetry is processed by a telemetry processor that calculates said derivative tactile feedback in real time by applying scaling values and algorithms to the raw telemetry. In the event that application-generated telemetry is not available or relatively limited, self-generated telemetry is utilized to supplement the provided telemetry. This telemetry can be used to scale in real time the results of an audio analysis tactile feedback generation method. The system's tactile sensation generators include independent groups of one or more actuators, e.g., electric motors or solenoids, that are embedded within or attached to various devices that can be connected to the system.

Description

    RELATED APPLICATION DATA
  • This application is a continuation-in-part of U.S. patent application Ser. No. 08/935,762 (assigned U.S. Pat. No. 6,422,941 set to issue on Jul. 23, 2002) which itself is a continuation-in-part of U.S. Pat. No. 5,669,818 (filed Mar. 23, 1995, issued Sep. 23, 1997) and U.S. Pat. No. 5,684,722 (filed Sep. 21, 1994, issued Nov. 4, 1997). All teachings of the parent applications and/or patents are incorporated herein by reference.[0001]
  • FIELD OF THE INVENTION
  • This invention relates to tactile feedback systems for computer and video game, simulation, and training systems, and, more particularly, to a method for using telemetry to generate and/or shape the tactile feedback generated by said systems, such that real-time tactile feedback is provided to enhance a user's or spectator's experience while interacting with said systems, and at a zero burden or a heretofore impossibly low burden to software developers, yet producing very effective results. [0002]
  • BACKGROUND OF THE DISCLOSURE
  • The implementation of tactile feedback in computer gaming, simulation, and training systems, where “tactile feedback” refers to physical sensations that one feels while interacting with said systems, has heretofore been plagued by the fundamental limitation of poor developer support. Typically, software developers are pressed for both time and money, and they are under constant pressure to release their software into the marketplace as soon as is practicable. In this competitive field, it is difficult for any given developer to spend the time necessary to create thoughtful and artistic tactile feedback. With regard to tactile feedback, proper implementation is more of an art than a science. No right or wrong answer necessarily exists for how a particular tactile event should feel, such as driving a car on gravel or firing a fictional laser cannon in a fictional spacecraft. Therefore, much time can be spent (and from the developer's point of view, even wasted) in tweaking the subjective feel of the desired effects, ad infinitum. In the end, there is a vast difference between the quality of tactile feedback that was merely implemented into a given software application, and tactile feedback that was thoughtfully and artistically crafted to render effective results (given of course the limitations of the particular tactile feedback hardware in use). In cases where the code that renders tactile feedback is incorporated by any given developer into that developer's wares, it is difficult for that code to be upgraded at a later time once a given product is released, because developers are usually fighting the time constraints of the post-release cycle where updates to some given software are periodically released to fix bugs or add features that were left out of the original release (or are functioning with substantial deficits) due to time constraints in effect during the prerelease period. Ultimately, servicing the tactile feedback support is less than a priority to the developer, and the effectiveness of said tactile feedback suffers as a result. Furthermore, given the subjective nature of tactile feedback, its effectiveness and quality varies greatly in each title, application, and/or implementation. [0003]
  • With regard to the patent applications and patents for which this application is a continuation-in-part, much of the difficulty presented by lack of developer support has been overcome by AudioSense®, which is a sound analysis technology that generates tactile feedback in real time without developer support of any kind. However, AudioSense® has inherent limitations that are overcome and/or minimized by intelliVIBE®, which is the type of tactile feedback that has typically required developer support—and therefore, intelliVIBE® suffers from the very limitations described previously herein. [0004]
  • Until the innovation that is the subject matter of this patent application, software developers were relied upon to decide what events in their software required what sensations, and what the triggering mechanisms were for initiating those sensations. For example, Immersion Corporation of 801 Fox Lane, San Jose, Calif. 95131 USA (NASDAQ: IMMR) markets and licenses technologies collectively falling under the trademark name TouchSense™. In using TouchSense™, developers can utilize a toolkit known as “Immersion Studio” (as of this writing, currently in version 4.1.0). When using Immersion Studio, developers can more easily create tactile effects, but are relied on to determine the subjective artistic tactile feel of those effects, and when such effects are to be rendered. This is precisely the burden that this patent application is intended to relieve. [0005]
  • Therefore, a need exists in the art for a relatively quick and simple method by which any given software developer is required only to activate a simple shared data structure, which will provided real-time telemetry to an external executable, application, function, or code segment (known to the present invention as an “intelliVIBE® module”), which in turn will itself generate the necessary tactile feedback control signals via general purpose algorithms that are shaped by said telemetry, thereby relieving the developer of the most substantial burdens of supporting an interface such as intelliVIBE®, and/or Immersion Corporation's TouchSense™, and/or any other proprietary methodology for generating tactile feedback. For reference, throughout this application, the terms “tactile feedback” should be understood to include and be synonymous with the terms “force feedback”. In much the same way that AudioSense® audio analysis technology allows tactile feedback hardware to function with no support whatsoever from any given developer of any given software title, application, or hardware system, this telemetry based method allows intelliVIBE® (and other similar interfaces) to self-generate effective and thoughtful tactile feedback with minimal, simple, and one-time-only support by any given developer. Furthermore, due to the telemetry basis for the tactile feedback, the teachings herein can be applied to other areas, such as remotely piloted vehicles, remotely controlled vehicles, UAVs (Unmated Aerial Vehicles), spacecraft, and other types of tele-operated or tele-presense vehicles, where telemetry is available or can be provided, thereby producing effective tactile feedback for the operators or spectators of such vehicles. [0006]
  • SUMMARY OF THE INVENTION
  • The present invention provides a method and corresponding apparatus for providing tactile feedback for computer and video game, training, and simulation systems. More specifically, the tactile feedback system comprises a software application, a tactile feedback controller, and a plurality of tactile sensation generators. The innovative method herein provides that any given software application can provide simple telemetry about its current state to a shared memory data structure, and subsequently leave all the burden of producing tactile feedback to an externally operating thread that employs general purpose algorithms and graphically adjustable scaling factors and configuration options to yield effective and appropriate tactile feedback without further effort by the given software application. It is a simple one-time only operation for any given software developer to implement the telemetry support. Thusly, the tactile feedback so produced is achieved with very low burden to the software developer. Additionally, should no telemetry be provided by a given software application, the innovative method herein has the means to self-generate its own telemetry by examining the state of the host computer's control input devices (i.e., keyboard, mouse, joystick, etc.), and to use said self-generated telemetry as the basis for applying real-time directional scaling to tactile feedback simultaneously produced via real-time audio analysis. The present invention can produce tactile feedback via the combination of real-time audio analysis with simultaneous directional scaling produced by real-time self-generated telemetry in combination with pre-defined general-purpose algorithms. This capability is achieved at a zero burden, that is, with no effort whatsoever nor any acknowledgement of its existence from the perspective of the developers of any given software application, yet yields results that resemble very carefully and thoughtfully crafted tactile feedback. [0007]
  • More specifically, the tactile feedback method and apparatus is provided in a preferred embodiment comprised by (1) a software application (currently known as the Tactile Feedback System Control Panel or TFS Control Panel), which configures the Tactile Feedback System (TFS) and handles the intelliVIBE® modules, each of which is adapted to handle the telemetry provided by any given software title, application, or system, and provides a graphical user interface for adjusting the available tactile feedback parameters that descend from the provided telemetry, (2) a tactile feedback controller (the presently preferred embodiment is currently known as the TFS2 (Tactile Feedback System version 2.0) which incorporates cooperative AudioSense® and intelliVIBE® capabilities, and (3) a plurality of tactile sensation generators, the currently preferred embodiment of which is a tactile feedback seat known as the TFSU (Tactile Feedback Seating Unit) and a tactile feedback vest (U.S. Pat. No. 5,565,840) known by the trademark name FLAKjaket™.[0008]
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The teachings of the present invention can be readily understood by considering the following detailed description in conjunction with the accompanying drawings, in which: [0009]
  • FIG. 1 depicts a block diagram of an illustrative configuration of the present invention; [0010]
  • FIG. 2 is a flow chart for using illustrative application generated telemetry to generate tactile feedback for a driving simulation; [0011]
  • FIG. 3 is a flow chart for using illustrative application generated telemetry to generate tactile feedback for a flight simulation; [0012]
  • FIG. 4 is a flow chart for using illustrative self-generated telemetry to generate scaling factors for directionally scaling tactile feedback simultaneously produced via audio analysis; [0013]
  • FIG. 5 shows an illustrative group of graphical controls for adjusting various telemetry scaling factors; [0014]
  • FIG. 5A depicts the graphical controls of FIG. 5, specifically illustrative RPM telemetry scaling factors; [0015]
  • FIG. 5B depicts the graphical controls of FIG. 5, specifically illustrative SPEED telemetry scaling factors; [0016]
  • FIG. 5C depicts the graphical controls of FIG. 5, specifically illustrative TAKE OFF telemetry scaling factors; [0017]
  • FIG. 5D depicts the graphical controls of FIG. 5, specifically illustrative ROLL telemetry scaling factors; [0018]
  • FIG. 5E depicts the graphical controls of FIG. 5, specifically illustrative G-FORCE telemetry scaling factors; [0019]
  • FIG. 5F depicts the graphical controls of FIG. 5, specifically illustrative WEAPONS telemetry scaling factors; [0020]
  • FIG. 5G depicts the graphical controls of FIG. 5, specifically illustrative HOTKEYS telemetry scaling factors; [0021]
  • FIG. 5H depicts the graphical controls of FIG. 5, specifically illustrative AFTERBURNER telemetry scaling factors; [0022]
  • FIG. 6 depicts an aircraft and the 4 quadrants used as the basis for an innovative aircraft roll algorithm; [0023]
  • FIG. 7 depicts the actual mathematical algorithm used to generate an aircraft's roll scaling factor based upon the aircraft's current quadrant as in FIG. 6, and how to utilize said scaling factor once derived from the roll telemetry data;[0024]
  • To facilitate understanding, identical reference numerals have been used, where possible, to designate identical elements that are common to the figures. [0025]
  • DETAILED DESCRIPTION OF THE DRAWINGS
  • FIG. 1 depicts a block diagram of a universal [0026] tactile feedback system 100 interacting with a host computer 102. The host computer 102 is connected to the universal tactile feedback controller 200 by an audio connection 112 and/or digital connection 116. The universal tactile feedback controller 200 is connected to multiple independent tactile sensation generators 500 by connection 460. The tactile sensation generators 500 collectively refer to hardware that produces physical sensations that one can feel while interacting with the software application 104. In this illustrative embodiment, no specific limitation is intended for any or all of the connections 112, 116, 460 with regard to their particular form, whether analog, digital, wired, wireless, and so on. The host computer 102 is presumed to be the host for some desired software application 104. The software application 104 is the application for which the present invention will generate tactile feedback. If the developers of the given software application 104 have supported the simple telemetry requirement, there will exist application generated telemetry 600. This application-generated telemetry 600 is a simple data structure that contains information about the state of operation of said software application 104. For example, if the software application 104 is a car racing simulation or car racing game, typical telemetry would include the car's engine RPM, the car's MPH, the car's throttle position, lateral G-forces, acceleration, deceleration, brake pedal position, steering wheel deflection, current gear, surface type (i.e., is the car currently on asphalt, dirt, gravel, sand, wet asphalt, icy asphalt, snow, mud, etc.), and damage flags (i.e., are there any flat tires, deformations of the vehicle's body panels, or equipment failures, etc.). This illustrative telemetry is by no means comprehensive, but should be sufficient to illustrate the type of data common to car racing games and/or simulations. It is easy for developers to provide this type of telemetry because they already calculate these variables (and most often many more variables than these) to make their software function properly. Each different game or simulation will have a unique set of telemetry data. The telemetry is entered into a shared data structure by a common memory mapping technique readily understood by those skilled in the art. In the illustrative embodiment, the computer host is running on a Microsoft Windows platform wherein a predetermined shared memory area name is agreed and known to both the software application 104 and the telemetry processor 700. The software application 104 creates the telemetry area by opening a shared memory area with the predetermined shared memory are name. The telemetry processor detects the shared memory area by said predetermined name, and uses the standard Windows coding “MapViewOfFile” to map the application-generated telemetry 600 into the memory of the telemetry processor 700. This simple operation affords the telemetry processor 700 access to the real-time state data (telemetry) of the software application 104. In short, it allows the telemetry processor 700 to know what is going on inside the software application 104, to the extent and detail inherent in the data structure provided by the software application 104 for the benefit of the application generated telemetry 600 shared memory structure, and ultimately for the benefit of the owners of the universal tactile feedback system 100 and the users of the software application 104. Providing the application generated telemetry 600 is a very simple operation for any reasonably skilled software developer, and even if new to the entire concept, most often it is a process of only several tens of minutes to effectuate the shared memory area once the data structure that is to be mapped there is known.
  • Continuing with FIG. 1, if a given software developer does not provide the necessary application generated [0027] telemetry 600, the telemetry processor 700 can examine the state of the host computer 102 in real time and generate useful telemetry by eavesdropping on the state of the host computer's 102 keyboard, mouse, joystick, and the like. This raw self-generated telemetry 800 is an effective substitute for application-generated telemetry; however, the application-generated telemetry is much preferred if it is available. When there is no application generated telemetry available, the universal tactile feedback system runs in its host-independent audio analysis mode of operation, which is known by the registered trademark name AudioSense® (U.S. Trademark Registration Number 2261526, registered Jul. 14, 1999). In this case, AudioSense® autonomously generates useful “base level” tactile feedback, which is then “directionally scaled” by the self-generated telemetry 800 that is culled by the telemetry processor 700 from the host computer 102. AudioSense® has no capabilities as of this writing to apply left, and/or right, and/or front, and/or back directional scaling based upon its audio analysis. While an effective technology, it is greatly enhanced by directional scaling from self-generated telemetry 800. The presumption is that if the software application 104 provides application-generated telemetry 600 for the benefit of the telemetry processor, most often AudioSense® will ideally be disabled while such telemetry is available. However, AudioSense® can work simultaneously and cooperatively with application generated telemetry 600 as well, sometimes lending an “analog flavor” to the tactile feedback produced by the universal tactile feedback system 100.
  • Further in FIG. 1, the [0028] telemetry processor 700 provides a graphical user interface 900 for configuring and adjusting the way the available telemetry is utilized. The telemetry processor 700 is itself a software application that manages the available telemetry and performs the processing that converts said telemetry into a useful data stream for generating and/or shaping tactile feedback. More detail will be provided on this later in this disclosure.
  • Continuing on to FIG. 2, an illustrative flow chart is shown wherein the [0029] telemetry processor 700 will process telemetry from a typical car based game or simulation. The telemetry processor 700 begins at the START 702. In step 704, the telemetry processor 700 accepts the telemetry via a shared memory mapped data structure. In the illustrative data set, the telemetry processor 700 is given the following illustrative values: the simulated car's engine RPM, vehicle MPH, throttle position, lateral g-force, forward acceleration and deceleration, the car's brake position, the car's steering wheel deflection, the current transmission gear, the surface type that the vehicle is currently on, and damage flags that might indicate various damage typical of cars, such as flat tires, body damage, engine failure, transmission failure, and the like. All of the received telemetry is in a “raw” state, that is, in a format native to the software application 104 that generated it. In step 706, scaling factors are applied to “normalize” the telemetry. For example, in one car simulation, the engine might idle at 700 RPM and might redline at 5500 RPM. However, in another car simulation, the engine might idle at 1500 RPM and redline at 9000 RPM. A specific constant RPM scaling factor might be used to always scale the RPM range for any given simulated car to a range of 0 to 100, and then, a baseline constant may be added, for example, 20. This would mean as long as the engine is started, the lowest scaled RPM value would be 0+20, or 20, where the highest RPM value would be 100+20, or 120. In this way, the “raw” telemetry will have been scaled into a useful range of 20 to 120. If the raw RPM value is 0, however, the baseline RPM constant +20 would be skipped, thereby leaving the RPM at 0. In this way, a useful RPM value is created. Now, once the initial scaled RPM value is rendered, another scaling layer is applied from a graphical user interface 900 that is provided by the telemetry processor 700. The graphical user interface 900 allows the user to make the RPM value more or less prevalent in the tactile feedback that is perceived. Typically, the RPM scaling factor would be a floating-point decimal value from 0.00 up to 2.00, including all interim values. Let's say the user leaves the RPM scaling factor in the graphical interface 900 at 1.00. In that case, the final scaled RPM range would be 20-120, or zero if the simulated car's engine was off or disabled. If the user set the graphical RPM scaling value to 1.50, the final scaled RPM range would be 20 times 1.5, or 30, up to 120 times 1.5, or 180, giving a final range of 30 to 180. In this way, by adjusting a simple control, the original unadulterated raw RPM value as generated by the simulation is converted to a useful and preferred strength. A similar process is used for all the telemetry variables, such that the relative strength of each can be tweaked to produce useful results. However, all the final results descend directly from the raw telemetry as calculated by the original software application 104 as mapped into the application generated telemetry 600 shared memory area. More detail will be examined later in this disclosure.
  • Continuing with FIG. 2, in [0030] step 708, we determine the baseline feedback level from the fundamental telemetry variables engine RPM, vehicle MPH, and throttle position. These are non-directionally scaled forces, which is why they together comprise what is known as the “baseline” tactile feedback level. In much the same process as above, the final scaled engine RPM value is calculated. With regard to MPH, lets assume in one game the given car has a speed range of 0 to 130 MPH, and in another game or simulation, the car has a speed range of 0 to 236 MPH. In the case where the top speed is 130 MPH, a constant scaling factor would be the current speed always divided by 1.3 (or rather multiplied by 0.769) such that the speed range was reduced to 0 to 100. Likewise, in the case where the car's top speed was 236 MPH, the constant scaling factor would be the current speed divided by 2.36 (or rather multiplied by 0.424), thereby yielding the same range 0 to 100. This makes the scaled MPH range consistent from game to game and simulation to simulation. Then, in the graphical interface 900 within the telemetry processor 700 application, a MPH scaling factor is provided where the user can increase or decrease the importance of the MPH by adjusting a multiplier with a graphical scroll bar, where the scaled MPH range now from 0 to 100 would be multiplied by any selected number 0.00 to 2.00, in the same way was the RPM was previously.
  • The next step is to compare the final RPM result and MPH result, and to accept the highest value as the baseline tactile feedback value for that cycle through the [0031] telemetry processor 700, which typically runs at hundreds of cycles second. This way, if you are going 0 MPH but your simulated engine has produced a final RPM vibration value of 88, you'll feel tactile feedback at a “strength” of 88. Again, this is illustrative only. Likewise, if you are traveling at a velocity of 150 MPH, and your MPH scaling factor is 0.70, but you've just blown your engine up and are at 0 (zero) engine RPM, you'll still feel 150 MPH times 0.70, for a vibration “strength” of 105. In this way, RPM and MPH are constantly fighting each other depending on which is the higher value at each and every single pass through the code in the telemetry processor 700 (which may be running at several hundred cycles per second).
  • Now the telemetry processor factors in the throttle position, because RPM or MPH should both feel stronger if you are full on the throttle, or not on the throttle at all. For example, if your throttle position runs from 0.00 at no throttle to 1.00 at full throttle, we may set a first scaling factor of “plus 0.60”, that is, adding 0.60 to each end of the throttle range, which changes the throttle scaling factor from 0.00-1.00 to 0.60-1.60. Now, the graphical scaling factor is utilized to change that range yet again according to the user's preference. If the user sets the graphical scaling factor for the throttle to 0.82, the final throttle scaling factor becomes 0.82 times 0.60 (or 0.492) at the low end, and 0.82 times 1.60 (or 1.312) at the high end, so the final range is 0.492 to 1.312, depending on the position of the throttle. [0032]
  • From this very simple basis, the raw telemetry calculated by the [0033] software application 104 has produced tactile feedback that is indicative of the engine RPM, vehicle MPH, and driver's throttle position. This is the “baseline” tactile feedback. Were the application generated telemetry 600 not provided, the AudioSense® audio analysis mode of operation of the universal tactile feedback controller 200 would have provided this baseline tactile feedback level. The importance of this cannot be stressed enough, because directional scaling is completely worthless if it is not based on an effective “baseline”.
  • Continuing on to step [0034] 710, the baseline tactile feedback values are now scaled according to the lateral gforce, acceleration, and deceleration telemetry. The unique nature of the preferred embodiment is that the tactile feedback seat in the parent cases has multiple zones, that is, a back zone, a left leg zone, and a right leg zone. Much like stereo speakers can create the illusion of a sound originating from any point between only two speakers by biasing the amplitude of the broadcast sound toward one speaker, and reducing it at the other, the tactile feedback seat functions in much the same way. If we are in a left turn, simple math is applied to increase the tactile feedback felt in the right leg zone, and to decrease the tactile feedback in the left leg zone. For example, if the lateral g-force is 2.3 g's, the tactile feedback for the right leg is be multiplied by 2.3, while the left leg is divided by 2.3. The scaling factors would change the numerical range of the G's to bring them into a useful range. In this way, the driver of the simulated car feels tactile cues that indicate he or she is being thrown to the outside of the turn, in direct proportion to the lateral g-forces involved. Likewise, upon acceleration, tactile feedback in the back zone is increased (i.e., multiplied by the scaled telemetry acceleration value) while tactile feedback in the legs is simultaneously reduced (i.e., divided by the scaled telemetry acceleration value), such that the illusion is provided that the driver is being thrown backward into the seat. Likewise, on deceleration, tactile feedback is increased at the legs and simultaneously reduced at the back, such that the illusion is provided that the driver is being thrown forward in the seat, as previously in direct proportion to the g-force values involved. The first layer of scaling factors work in the same way to bring the raw telemetry into a useful range, while the graphical interface is used to set the user's preference as to how strong this directional scaling is to be perceived. By changing any given scaling factor, the math that is applied to the raw scaled data is change accordingly. In this way it is very easy to dial in excellent tactile feedback, because all the calculations are based on the mathematics that are already built into the simulation due to the nature of the shared telemetry.
  • Continuing on to step [0035] 712, tactile feedback texturing is applied to the now directionally scaled tactile feedback values. Texturing adds choppiness and other characteristics to the tactile feedback. For example, if a “damage flag” indicates that there is a flat tire on the left side of the car, there may be a “flat tire scaling factor” provided in the graphical user interface 900 within the telemetry processor 700 that sets the flat tire scaling factor as 1.5 times (1.5×). The texturing to the left leg zone in the tactile feedback seat (the side of the vehicle with the simulated flat tire) would now be periodically multiplied by 1.5× in some time period that is directly proportional to the current MPH of the car. Again, very simple math is used to create dramatic effects, all directly descending from the raw telemetry and the same general-purpose algorithms in place for most games or simulation of the car type genre. In other tactile feedback technologies, the effects for flat tires are canned periodic cycles of bumping that has no correspondence with the speed of the given vehicle because there is no telemetry shared between the application and the tactile feedback algorithms.
  • In the case of the car's current gear (the vehicle's transmission) telemetry data element, simple math can yield a brief skip in the tactile feedback throughout the tactile feedback seat any time the gear changes. For example, if the gear has just changed from 2 to 3, the telemetry processor may have two factors associated with gear changes, that is, a percent reduction value, and a time value in milliseconds. If the gear change scaling factors indicate that the tactile feedback throughout the seat should be multiplied by 0.25 and that condition should last for 100 milliseconds every time a gear changes, then each gear change will cause a direct self-generating effect of a 75% reduction in perceived tactile feedback for 0.10 seconds at every gear change. For the brake pedal position telemetry data, hard braking at fast speeds can make other very brief scaling factors take effect. Surface type can layer additional scaling factors upon the baseline vibration. The end result is that the telemetry data itself carries the information about what is happening, and all perceived tactile feedback descend directly from the actual state of the simulation or game. [0036]
  • In [0037] step 714, we transmit the current tactile feedback results to the universal tactile feedback controller 200 with the standard intelliVIBE® protocol, but no limitation should be inferred about which types of tactile feedback devices can benefit from this telemetry based approach.
  • Continuing on to FIG. 3, it is nearly identical to FIG. 2, but the telemetry carries data relevant to flight simulations instead of car simulations. [0038] Steps 722, 724, 726, 728 are analogous to steps 702,704,706,708 of FIG. 2. However, for step 730, we refer now to FIG. 6.
  • The major difference here is the way the telemetry processor handles g-forces, pitch, and roll, which are data typical of flight simulations. The innovative approach here allows two possibilities for rendering aircraft rolls, that is, when an aircraft such as a modem military jet rolls on its center axis in a continuous right or left turn. In reality, when a jet enters a continuous roll to the right, the passengers get pushed to the left side and stay there due to the g-forces involved that are stable for the duration of the rightward roll. This is calculated just like the lateral g-forces of the car of FIG. 2. However, an aircraft's roll can be divided into four quadrants, each of which affects the rendering of the roll in a different way. Rather than a continuous push to one side, which is a realistic and readily available but perhaps a dull representation, we can artificially induce additional aircraft orientation data in the tactile feedback that is produced. [0039]
  • Referring now to FIG. 6, an [0040] aircraft 740 is depicted with its tail 742 pointing straight up, that is, the aircraft is perfectly level. The current quadrant is determined by the orientation of the aircraft's tail 742. Whichever quadrant 1, 2, 3 or 4 the aircraft's tail 742 points to, that is the quadrant the aircraft is considered to be in for the purposes of rendering the “roll” effect upon the baseline vibration. Referring now to FIG. 7, the quadrants 1, 2, 3, and 4 are analogous to the same quadrants of FIG. 6. However, FIG. 7 explains the simple algorithm by which the right leg tactile feedback value and the left leg tactile feedback value are scaled depending in which quadrant the aircraft's tail is oriented. With this simple math applied to the telemetry derived “baseline” vibration values, the user gets a excellent feel of the aircraft's orientation while interacting with any given aircraft simulation or game. With the innovative algorithm of FIG. 7, a continual roll to the right in the aircraft does not produce a static leftward bias as would be the case if driving a car, but rather, the left-right leg bias actually rolls around the seat in direct proportion to the simulated aircraft's orientation. When fully inverted, that is, upside-down, the bias disappears and the legs tactile feedback scaling is nullified. This is a completely artificial effect, having no basis in reality; however, the effect is remarkable and highly effective.
  • When in any vehicle, positive and negative g-forces are experienced. Some vehicles induce more positive and negative g-forces than others, but nonetheless, because of the telemetry offered by a given simulation or game, the same mathematics apply and are effective across the board. When experiencing positive g-forces, one gets the feeling of weighing more than they actually weigh, in direct proportion to the positive g-forces that are experienced. Therefore, with the telemetry based system, positive g-forces are mathematically multiplied against the leg section (both left and right legs), thereby making the user feel “heavier”, while simultaneously used as a divisor to mathematically reduce the tactile feedback to the back zone. Most often the raw g-forces are scaled into the range of 1.000 to 2.000. Of course, as with all scaling factors, the users have additional scaling factors to augment of reduce the prevalence of the g-forces. Likewise, when experiencing negative g-forces, one gets the feeling of being more weightless than is normal. Therefore, when negative g-forces are present, the negative g-force value is multiplied by the back value to increase the back zone tactile feedback, while simultaneously used as a divisor to mathematically reduce the tactile feedback to the legs, making one feel like they are rising out of the tactile feedback seat. [0041]
  • When the roll of an aircraft is simultaneously layered with the positive and negative g-forces, one gets to experience a very tactile and informative representation of the aircraft's flight envelope. Similarly, pitch and yaw, and any and all other telemetry, can be layered into the calculations as well. The forces so calculated are not arbitrary, yet are based upon the actual math occurring in real time in the shared telemetry memory area. The only subjective element to all these forces are the user's adjustment of the graphical scaling factors in order to make any particular effect more or less prevalent in comparison to all the other forces for which telemetry data is available. [0042]
  • Referring now to FIG. 4, the [0043] telemetry processor 700 does not have the benefit of application-generated telemetry. Therefore, the universal tactile feedback controller 200 will run in its audio analysis mode of operation (AudioSense®). When in AudioSense®, the baseline tactile feedback is produced without capabilities for directional biasing, regardless as to the forces acting on the user's vehicle or person in any particular application. Simply put, the tactile feedback system has no awareness as to the specific calculations that are necessarily taking place in any particular application. When this is the case, the telemetry processor begins in step 782 and in step 784 examines the telemetry it can derive from the host computer's 102 control input devices. If, for example, the user of a non-telemetry supportive title (i.e., no intelliVIBE® support) is running in AudioSense®, they can give specific scaling instructions to the telemetry processor 700 so that it can generate real time tactile feedback zone-specific scaling factors that will introduce real time scaling upon the AudioSense® audio analysis results. For example, if the user is playing a car racing simulation that offers no intelliVIBE® support, the telemetry processor can track the movement of the player's joystick and can know its deflection from center on both the X and Y axes, that is, left and right (steering) and forward and back (acceleration and brake) of the simulated vehicle. Certainly this is not as good as real time telemetry, but the telemetry processor 700 will do the best it can with what data it can find. In this case, the user will adjust the joystick scaling factors such that the longer the accelerator axis (presumably the Y axis, but configured as such by the user) is pressed forward, the more speed the telemetry processor will assume will have been gained by the simulated vehicle. As the simulated vehicle is assumed to be going faster, it will automatically adjust the X-axis scaling factors (the car's steering) to make the lateral g-forces more prevalent. Likewise, if the Y axis indicates that the user has stepped on the brake of the simulated vehicle, the telemetry processor will indicate this in the stream of scaling factors that are being broadcast to the universal tactile feedback controller 200 such that a deceleration effect is produced by directionally biasing the tactile feedback in the seat to the legs section of the tactile feedback seating unit, and away from the back section.
  • The scaling factors are calculated in real time based upon assumed data, so they do not have the finesse of real-time accurate telemetry that would ideally be available. The scaling factors are 8 bit values, (0-255), wherein whatever value given as a scaling factor is divided by 32 within the universal [0044] tactile feedback controller 200. The scaling factors exist in a data table, where each single scaling factor is applied to only one of eight zones: back, left leg, right leg, chest, left hand, right hand, left foot, right foot. These zones are illustrative and not assumed to match this presently preferred embodiment. Sending a scaling factor of 32 yields no change to a particular tactile feedback zone because of the subsequent divide by 32 that occurs within the universal tactile feedback controller 200 (i.e., 32 divided by 32 equals 1.0). However, a scaling factor of 16 sent to the left leg will cut the tactile feedback there in half, because 16/32 yields a scaling factor of 0.5×. The divide by 32 algorithm is used to yield very fast mathematical performance by using a shortcut rather than full floating point math which would impede the performance of AudioSense® if the microcontroller inside the universal tactile feedback controller 200 does not offer full hardware multiply and divide.
  • By monitoring the joystick input over time, fairly accurate assumptions can be made about a particular game or simulation, but this requires more complicated help by way of pre-determined application-[0045] specific telemetry processor 700 configuration files. For example, in the real F16 Falcon military aircraft, the main cannon only has capacity enough for approximately 2.3 seconds of trigger time before it runs out of 30 mm cannon shells. Therefore, the telemetry processor 700 for an accurate F16 simulation with limited ammunition can assume from its configuration data that the main joystick trigger (or some corresponding keyboard key) fires the main cannon. Without proper telemetry being given, the telemetry processor will keep track of the time the cannon trigger is active. During the presumed activation of the cannon, an effect will be rendered according to the scaling parameters in effect for the cannon. After approximately 2.3 seconds of trigger time, the intelliVIBE® cannon effect will be disabled until reset by the user. However, AudioSense® will still do its best to render the cannon by interpreting the sound of it firing if the cannon is not yet depleted of ammunition. Additionally, if the throttle has been active and at a high power, and the joystick has tended to be pulled back, the telemetry processor can assume that the aircraft is airborne, and can then do its best to use the joystick position to send scaling factors to the left leg, right leg, and back zones of the tactile feedback seat in order to simulate the G-forces and bank roll effects that would ideally be more accurately prompted by fully supported telemetry. Even with the slightly less-than-ideal implementation of AudioSense® audio analysis and assumed game state data by way of brute force self-generated telemetry via game-specific control input analysis, the combination of cooperative AudioSense® with intelliVIBE® scaling is a very effective means of self-generating very appropriate tactile feedback, regardless of the lack of support by any given application.
  • Referring now to FIG. 5, a group of user adjustable telemetry scaling factors is presented. These controls are illustrative only for the purposes of understanding. Getting a closer look at FIG. 5A, the RPM baseline tactile feedback generator is presented. The baseline setting of 30 establishes the basic level of prevalence of engine RPM induced tactile feedback. The multiplier at the bottom shows 0.7, which is used to scale the 30 baseline setting. If the aircraft is known to be on the ground, the engine RPM setting is multiplied by 0.9, whereas if the aircraft is in the air, the engine RPM is multiplied by 0.6, thereby making the engine RPM feel weaker once in the air. [0046]
  • Referring now to FIG. 5B, the speed value is multiplied by 0.8 if the aircraft is on the ground, while it is multiplied by 0.08 if the aircraft is in the air. This has the effect of making the speed tactile effect ten times weaker when the aircraft is airborne, but the typical maximum speed when on the ground might be only 180 MPH, but when airborne, a modern military aircraft can travel at 1,300 MPH and higher. These scaling factors allow the tactile feedback to be matched to a useful range regardless as to the performance envelope of any particular aircraft, simply by adjusting the slider knob of the corresponding scaling factor control. [0047]
  • Looking now at FIG. 5C, we have a compound effect called ground texturing. Ground texturing is set to begin at 30 MPH and end at 250 MPH, which is well past a conservative take-off speed. Ultimately as the aircraft goes faster, the ground will create more and more texturing, that is, bumpiness in the tactile feedback. Upon takeoff, the ground texturing will be reduced to 25% (multiplied by 0.25×) until the take off speed is surpassed, which then nullifies the ground texturing, until the aircraft comes in for a landing, at which time the ground texturing will automatically take effect yet again. [0048]
  • FIG. 5D presents the bank scaling value. The 1.0 as shown lets the bank scaling take place just as the calculations in FIG. 7 indicate. However, if the user wants the bank scaling to be stronger, they only need to increase the corresponding 1.0 scale factor setting. To reduce bank scaling, setting bank scaling to toward 0 will nullify the effect accordingly. [0049]
  • FIG. 5E presents the positive and negative G-force multipliers. In positive g-force conditions, these settings multiply the legs by 1.3 times the g-force value, and multiply the back by 2.5 times the g-force value. In negative G-force conditions, the back is reduced by half and the legs are scaled to 80% for the duration of the negative G-force condition. [0050]
  • In FIG. 5F, the scaling factors for the main gun are available. During the firing of the main gun, 100 pulses per second will occur in the tactile feedback, alternating between a high pulse percentage of 100% and a low pulse percentage of 12% as determined by the scaling factors. The “use high intensity” check box allows the main gun to temporarily over-ride the current intensity setting for the duration of the main gun being fired. Knowing that the main gun is being fired can be given to us by application generated [0051] telemetry 600 or self-generated telemetry 800, same as with all the effects and scaling factors of FIG. 5A through FIG. 5H.
  • In FIG. 5G, the [0052] telemetry processor 700 is told what key press equals what known effect. This is configured by the user. For example, we can see in the upper left corner that the “G” key activates the landing gear. This way, if the user presses the G key to deploy the landing gear, the telemetry processor can run a simple effect representing the landing gear. If telemetry is not provided that indicates the true state of the landing gear, the telemetry processor 700 will fall back on its self-generating capability and keep track of the G key presses on its own and will do its best to stay in sync with the simulation in question.
  • In FIG. 5H, an illustrative afterburner effect is shown. Many modem jets have multiple stages of afterburner. In FIG. 5H, there are 3 stages of afterburner, where stages one, two, and three yield tactile “thumps” at 60%, 80%, and 100% strength, respectively, each lasting only 0.01 seconds, but also incorporating an RPM scaling factor of 1.33×, 1.5×, or 1.66× times respectively, which will increase the prevalence of the engine RPM according to what level of afterburner the aircraft is in, for as long as the afterburner is active. [0053]
  • Ultimately, the innovative methods and apparatus of the present invention allows for the most versatile tactile feedback results to be produced, both with and without support of any kind by the [0054] host computer 102 and the executing software application 104. The best condition is for fully detailed low-burden telemetry to be provided for the benefit of the universal tactile feedback controller 200 by a given developer. However, in the absence of given telemetry via a shared data structure by standard windows memory mapping, the telemetry processor 700 of the present invention is able to generate data on its own to supplement the AudioSense® audio analysis mode of operation with real time directional biasing of all tactile feedback, in the end producing effective zero-burden feedback from the point of the game developer . . . in the end eliminating the fundamental problem of lack of quality developer support!
  • Although various embodiments which incorporate the teachings of the present invention have been shown and described in detail herein, those skilled in the art can readily devise many other varied embodiments that still incorporate these teachings. Accordingly, the described embodiments are to be considered in all respects only as illustrative and not restrictive. Therefore, it is to be understood that nothing shall be construed as a limitation of the present invention, other than the structure recited in the appended claims. [0055]

Claims (6)

What is claimed is:
1. In a computer gaming or simulation system, a method for producing tactile sensations in correspondence to the activity occurring on said system, said method comprising:
an executing software application that is providing telemetry indicative of its real-time state;
a real-time telemetry processor for processing said telemetry with various scaling factors and algorithms;
wherein said tactile sensations are derived from said telemetry processor.
2. The method of claim 1, where said tactile sensations are generated by:
a flexible pad;
a plurality of actuators, attached to said pad, for selectively generating tactile sensation.
3. The apparatus of claim 2, wherein said plurality of actuators are comprised by offset weights affixed to DC motors, such that when said motors are energized, said offset weights induce physical vibrations of varying intensities.
4. In a computer gaming or simulation system, a method for producing tactile sensations in correspondence to the activity occurring on said system, said method comprising:
an executing software application that is not providing telemetry, and therefore, providing no information indicative of its real-time state;
a real-time telemetry processor that will self-generate telemetry for its own use by examining the state of said computer gaming or simulation system; and
using said telemetry to directionally bias and/or supplement and otherwise enhance the tactile effects occurring as a result of simultaneous audio analysis.
5. The method of claim 4, where said tactile sensations are generated by:
a flexible pad;
a plurality of actuators, attached to said pad, for selectively generating tactile sensation.
6. The apparatus of claim 5, wherein said plurality of actuators are comprised by offset weights affixed to DC motors, such that when said motors are energized, said offset weights induce physical vibrations of varying intensities.
US10/201,584 1994-09-21 2002-07-23 Method and apparatus for generating tactile feedback via relatively low-burden and/or zero burden telemetry Abandoned US20030040361A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US10/201,584 US20030040361A1 (en) 1994-09-21 2002-07-23 Method and apparatus for generating tactile feedback via relatively low-burden and/or zero burden telemetry
US11/929,426 US8328638B2 (en) 1994-09-21 2007-10-30 Method and apparatus for generating tactile feedback via relatively low-burden and/or zero burden telemetry
US12/105,007 US8308558B2 (en) 1994-09-21 2008-04-17 Universal tactile feedback system for computer video games and simulations
US13/711,585 US8834266B2 (en) 1994-09-21 2012-12-11 Methods and apparatuses for generating tactile feedback for games and simulations

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US08/309,763 US5684722A (en) 1994-09-21 1994-09-21 Apparatus and method for generating a control signal for a tactile sensation generator
US08/409,327 US5669818A (en) 1995-03-23 1995-03-23 Seat-based tactile sensation generator
US08/935,762 US6422941B1 (en) 1994-09-21 1997-09-23 Universal tactile feedback system for computer video games and simulations
US10/201,584 US20030040361A1 (en) 1994-09-21 2002-07-23 Method and apparatus for generating tactile feedback via relatively low-burden and/or zero burden telemetry

Related Parent Applications (3)

Application Number Title Priority Date Filing Date
US08/409,327 Continuation-In-Part US5669818A (en) 1994-09-21 1995-03-23 Seat-based tactile sensation generator
US08/935,762 Continuation-In-Part US6422941B1 (en) 1994-09-21 1997-09-23 Universal tactile feedback system for computer video games and simulations
US08/935,762 Continuation US6422941B1 (en) 1994-09-21 1997-09-23 Universal tactile feedback system for computer video games and simulations

Related Child Applications (3)

Application Number Title Priority Date Filing Date
US08/309,763 Continuation US5684722A (en) 1994-09-21 1994-09-21 Apparatus and method for generating a control signal for a tactile sensation generator
US11/929,426 Continuation US8328638B2 (en) 1994-09-21 2007-10-30 Method and apparatus for generating tactile feedback via relatively low-burden and/or zero burden telemetry
US12/105,007 Continuation US8308558B2 (en) 1994-09-21 2008-04-17 Universal tactile feedback system for computer video games and simulations

Publications (1)

Publication Number Publication Date
US20030040361A1 true US20030040361A1 (en) 2003-02-27

Family

ID=46280912

Family Applications (4)

Application Number Title Priority Date Filing Date
US10/201,584 Abandoned US20030040361A1 (en) 1994-09-21 2002-07-23 Method and apparatus for generating tactile feedback via relatively low-burden and/or zero burden telemetry
US11/929,426 Expired - Fee Related US8328638B2 (en) 1994-09-21 2007-10-30 Method and apparatus for generating tactile feedback via relatively low-burden and/or zero burden telemetry
US12/105,007 Expired - Fee Related US8308558B2 (en) 1994-09-21 2008-04-17 Universal tactile feedback system for computer video games and simulations
US13/711,585 Expired - Fee Related US8834266B2 (en) 1994-09-21 2012-12-11 Methods and apparatuses for generating tactile feedback for games and simulations

Family Applications After (3)

Application Number Title Priority Date Filing Date
US11/929,426 Expired - Fee Related US8328638B2 (en) 1994-09-21 2007-10-30 Method and apparatus for generating tactile feedback via relatively low-burden and/or zero burden telemetry
US12/105,007 Expired - Fee Related US8308558B2 (en) 1994-09-21 2008-04-17 Universal tactile feedback system for computer video games and simulations
US13/711,585 Expired - Fee Related US8834266B2 (en) 1994-09-21 2012-12-11 Methods and apparatuses for generating tactile feedback for games and simulations

Country Status (1)

Country Link
US (4) US20030040361A1 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030114226A1 (en) * 2001-12-14 2003-06-19 Square Co., Ltd. Network game system, game server system, client system, network game processing method, and recording medium
US20060119576A1 (en) * 2004-12-06 2006-06-08 Naturalpoint, Inc. Systems and methods for using a movable object to control a computer
US20060119575A1 (en) * 2004-12-06 2006-06-08 Naturalpoint, Inc. Systems and methods for using a movable object to control a computer
EP2067508A1 (en) * 2007-11-29 2009-06-10 AMBX UK Limited A method for providing a sensory effect to augment an experience provided by a video game
US20120262404A1 (en) * 2009-12-21 2012-10-18 Kyocera Corporation Tactile sensation providing apparatus and control method for tactile sensation providing apparatus
US8660673B2 (en) 2011-03-23 2014-02-25 Electronic Entertainment Design And Research In-game contextual telemetry systems and methods
US20150169058A1 (en) * 2012-03-30 2015-06-18 Nvf Tech Ltd Touch and Haptics Device
US20170108995A1 (en) * 2015-10-16 2017-04-20 Microsoft Technology Licensing, Llc Customizing Program Features on a Per-User Basis
US20180068537A1 (en) * 2013-09-06 2018-03-08 Immersion Corporation Spatialized haptic feedback based on dynamically scaled values
US11392203B2 (en) * 2018-03-27 2022-07-19 Sony Corporation Information processing apparatus, information processing method, and program

Families Citing this family (94)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7419425B1 (en) * 2001-02-15 2008-09-02 Bally Gaming, Inc. Shared secondary game station and system
JP4380563B2 (en) * 2005-03-02 2009-12-09 ソニー株式会社 Editing device and editing processing program
TW200946189A (en) * 2008-05-12 2009-11-16 Princeton Technology Corp Game controller with audio signal outputting device
US20090319694A1 (en) * 2008-06-20 2009-12-24 Microsoft Corporation Association of an input and output of a peripheral device in a computing system
US20100031808A1 (en) * 2008-08-05 2010-02-11 Honeywell International Inc. Method, apparatus, and system of providing sensor-based tactile feedback
US8306934B2 (en) * 2009-03-30 2012-11-06 International Business Machines Corporation Demo verification provisioning
US20150134946A1 (en) * 2009-04-22 2015-05-14 Harold Lee Peterson System, method and computer program product for selecting and offering computational functionalities to a user
US8246467B2 (en) * 2009-04-29 2012-08-21 Apple Inc. Interactive gaming with co-located, networked direction and location aware devices
WO2010135472A2 (en) * 2009-05-19 2010-11-25 Ev-Ip, Llc Methods and apparatus for utilizing electrically powered vehicles
US20110093092A1 (en) * 2009-10-19 2011-04-21 Bum Suk Choi Method and apparatus for creating and reproducing of motion effect
US20110183752A1 (en) * 2010-01-27 2011-07-28 Bey Jamaal N H Video game body suit
US20130144942A1 (en) * 2010-03-19 2013-06-06 Gopi Krishnan Nambiar Session persistence for accessing textsites
TWI426789B (en) * 2010-08-19 2014-02-11 Hon Hai Prec Ind Co Ltd Electronic device and method for providing scene feeling thereof
KR20120028003A (en) * 2010-09-14 2012-03-22 삼성전자주식회사 Apparatus and method for 3-dimensional tactile display
US8337305B2 (en) 2010-11-17 2012-12-25 Steelseries Aps Apparatus and method for managing user inputs in video games
CA2825846A1 (en) 2011-02-01 2012-08-09 Timeplay Entertainment Corporation Systems and methods for interactive experiences and controllers therefor
US20120289312A1 (en) * 2011-05-11 2012-11-15 Hamlin Vernon W Controlling a motion capable chair in a wagering game system based on environments and ecologies
EP2767041A4 (en) * 2011-10-11 2015-05-06 Timeplay Entertainment Corp Systems and methods for interactive experiences and controllers therefor
KR102024006B1 (en) * 2012-02-10 2019-09-24 삼성전자주식회사 Apparatus and method for controlling vibration flow between vibration devices
US8902159B1 (en) 2012-07-24 2014-12-02 John Matthews Ergonomic support apparatus having situational sensory augmentation
US8976043B2 (en) * 2012-08-20 2015-03-10 Textron Innovations, Inc. Illuminated sidestick controller, such as an illuminated sidestick controller for use in aircraft
JP6071372B2 (en) * 2012-09-21 2017-02-01 キヤノン株式会社 Electronic device and control method of electronic device
US9716327B2 (en) * 2012-09-28 2017-07-25 Intel Corporation System, circuit module, and circuit module connector
US9565742B2 (en) 2012-10-26 2017-02-07 Lutron Electronics Co., Inc. Battery-powered retrofit remote control device
US9101297B2 (en) 2012-12-11 2015-08-11 Elwha Llc Time-based unobtrusive active eye interrogation
US9039180B2 (en) 2012-12-11 2015-05-26 Elwah LLC Self-aligning unobtrusive active eye interrogation
US9039179B2 (en) 2012-12-11 2015-05-26 Elwha Llc Unobtrusive active eye interrogation
US20140349762A1 (en) * 2013-03-15 2014-11-27 Alfred M. Haas Gtg
US9690594B2 (en) * 2013-04-29 2017-06-27 Honeywell International Inc. System for configuring a controller
WO2014201151A1 (en) 2013-06-11 2014-12-18 Immersion Corporation Systems and methods for pressure-based haptic effects
CN105723735B (en) 2013-08-28 2020-02-07 低音炮公司 Multi-level tactile sound device
US9619980B2 (en) 2013-09-06 2017-04-11 Immersion Corporation Systems and methods for generating haptic effects associated with audio signals
US9652945B2 (en) * 2013-09-06 2017-05-16 Immersion Corporation Method and system for providing haptic effects based on information complementary to multimedia content
US9576445B2 (en) 2013-09-06 2017-02-21 Immersion Corp. Systems and methods for generating haptic effects associated with an envelope in audio signals
WO2015100482A1 (en) * 2013-12-31 2015-07-09 Iftech Inventing Future Technology Inc. Wearable devices, systems, methods and architectures for sensory stimulation and manipulation, and physiological data acquisition
US9138651B2 (en) * 2014-01-03 2015-09-22 Jonathan B Gaug Universal game pedal controller and method therefor
US9866986B2 (en) 2014-01-24 2018-01-09 Sony Corporation Audio speaker system with virtual music performance
US9704491B2 (en) 2014-02-11 2017-07-11 Disney Enterprises, Inc. Storytelling environment: distributed immersive audio soundscape
CN111399646A (en) 2014-03-21 2020-07-10 意美森公司 Systems, methods, and computer-readable media for force-based object manipulation and haptic detection
US9633557B2 (en) 2014-06-24 2017-04-25 Lutron Electronics Co., Inc. Battery-powered retrofit remote control device
KR102373337B1 (en) 2014-09-02 2022-03-11 애플 인크. Semantic framework for variable haptic output
US20170011602A1 (en) * 2014-12-11 2017-01-12 Elwha Llc Wearable haptic feedback devices and methods of fabricating wearable haptic feedback devices
US10166466B2 (en) 2014-12-11 2019-01-01 Elwha Llc Feedback for enhanced situational awareness
US9922518B2 (en) 2014-12-11 2018-03-20 Elwha Llc Notification of incoming projectiles
US10449445B2 (en) 2014-12-11 2019-10-22 Elwha Llc Feedback for enhanced situational awareness
US9741215B2 (en) 2014-12-11 2017-08-22 Elwha Llc Wearable haptic feedback devices and methods of fabricating wearable haptic feedback devices
US9579542B2 (en) 2014-12-29 2017-02-28 Robert Quinn Networked game system having multiple full-body exercise apparatus controllers
US9649530B2 (en) 2014-12-29 2017-05-16 Robert Quinn Full-body exercise system including a plurality of independently operable appendicular members
US9861852B2 (en) 2014-12-29 2018-01-09 Robert Quinn Game system having full-body exercise apparatus controller with independently operable appendicular members
EP3263195B1 (en) * 2015-02-27 2024-08-14 Sony Group Corporation Information processing device, information processing method, information processing system
US10446344B2 (en) 2015-05-27 2019-10-15 Microsoft Technology Licensing, Llc Hair trigger travel stop with on-demand switching
JP6606355B2 (en) * 2015-05-29 2019-11-13 キヤノン株式会社 Information processing apparatus, information processing method, and program
CN104998409A (en) * 2015-06-11 2015-10-28 将军兄弟有限公司 Vibration feedback system and vibration feedback method
CN105404607B (en) * 2015-11-20 2018-02-13 英业达科技有限公司 The data transmission method of general serial input and output
CN105611443B (en) 2015-12-29 2019-07-19 歌尔股份有限公司 A kind of control method of earphone, control system and earphone
CN105630021B (en) * 2015-12-31 2018-07-31 歌尔股份有限公司 A kind of the tactile vibrations control system and method for intelligent terminal
CN105511514B (en) 2015-12-31 2019-03-15 歌尔股份有限公司 A kind of the tactile vibrations control system and method for intelligent terminal
US9468844B1 (en) * 2016-01-20 2016-10-18 Chun Hung Yu Method for transmitting signals between wearable motion capture units and a video game engine
US9616333B1 (en) * 2016-01-20 2017-04-11 Chun Hung Yu Method for capturing and implementing body movement data through a video game engine
US9981182B2 (en) 2016-02-12 2018-05-29 Disney Enterprises, Inc. Systems and methods for providing immersive game feedback using haptic effects
US9924291B2 (en) 2016-02-16 2018-03-20 Sony Corporation Distributed wireless speaker system
US9826330B2 (en) * 2016-03-14 2017-11-21 Sony Corporation Gimbal-mounted linear ultrasonic speaker assembly
US20170348594A1 (en) * 2016-06-02 2017-12-07 Wind River Systems, Inc. Device, System, and Method for Motion Feedback Controller
DK179823B1 (en) 2016-06-12 2019-07-12 Apple Inc. Devices, methods, and graphical user interfaces for providing haptic feedback
DK201670737A1 (en) 2016-06-12 2018-01-22 Apple Inc Devices, Methods, and Graphical User Interfaces for Providing Haptic Feedback
US9794724B1 (en) 2016-07-20 2017-10-17 Sony Corporation Ultrasonic speaker assembly using variable carrier frequency to establish third dimension sound locating
US10556176B2 (en) * 2016-07-26 2020-02-11 Nintendo Co., Ltd. Vibration control system, vibration control method, and non-transitory computer-readable storage medium with executable vibration control program stored thereon
US10328345B2 (en) * 2016-07-26 2019-06-25 Nintendo Co., Ltd. Vibration control system, vibration control method, and non-transitory computer-readable storage medium with executable vibration control program stored thereon
JP2018030107A (en) * 2016-08-26 2018-03-01 レノボ・シンガポール・プライベート・リミテッド Haptic feedback system, electronic equipment and method for generating haptic feedback
DK201670720A1 (en) 2016-09-06 2018-03-26 Apple Inc Devices, Methods, and Graphical User Interfaces for Generating Tactile Outputs
DK179278B1 (en) * 2016-09-06 2018-03-26 Apple Inc Devices, methods and graphical user interfaces for haptic mixing
DK201670728A1 (en) 2016-09-06 2018-03-19 Apple Inc Devices, Methods, and Graphical User Interfaces for Providing Feedback During Interaction with an Intensity-Sensitive Button
DK180150B1 (en) * 2016-09-06 2020-06-26 Apple Inc Devices, methods and graphical user interfaces for providing feedback during interaction with an intensity sensitive button
EP3343324B1 (en) 2016-12-27 2021-06-09 Nintendo Co., Ltd. Information processing program, information processing system, and information processing method
US10350491B2 (en) 2016-12-27 2019-07-16 Nintendo Co., Ltd. Techniques for variable vibration waveform generation based on number of controllers
US10152296B2 (en) 2016-12-28 2018-12-11 Harman International Industries, Incorporated Apparatus and method for providing a personalized bass tactile output associated with an audio signal
DK201770372A1 (en) 2017-05-16 2019-01-08 Apple Inc. Tactile feedback for locked device user interfaces
JP7244985B2 (en) * 2017-05-19 2023-03-23 川崎重工業株式会社 Operating device and operating system
CA3092689A1 (en) 2017-10-23 2019-05-02 Patent Holding Company 001, Llc Communication devices, methods, and systems
US11148046B2 (en) * 2018-01-16 2021-10-19 Vr Leo Usa, Inc. Chip structure of VR self-service game joy stick
JP6604669B1 (en) * 2018-09-06 2019-11-13 neten株式会社 Bone vibration sensation device and method of using the same
US10800433B2 (en) * 2018-09-14 2020-10-13 Honda Motor Co., Ltd. Seat haptic system and method of equalizing haptic output
US20220413614A1 (en) * 2018-11-20 2022-12-29 Whirlwind VR, Inc. System and Method for a Surface-Optimized Tactile Transducer
US11076460B2 (en) * 2019-01-18 2021-07-27 Alliance Sports Group, L.P. Lighting system
US11543110B2 (en) * 2019-01-18 2023-01-03 Alliance Sports Group, L.P. Lighting system
US10951951B2 (en) * 2019-07-30 2021-03-16 Sony Interactive Entertainment Inc. Haptics metadata in a spectating stream
JP7072550B2 (en) * 2019-11-19 2022-05-20 株式会社ソニー・インタラクティブエンタテインメント Operation device
US11443737B2 (en) 2020-01-14 2022-09-13 Sony Corporation Audio video translation into multiple languages for respective listeners
US11610459B2 (en) * 2020-04-13 2023-03-21 Google Llc Factory and user calibration of haptic systems
US20220111290A1 (en) * 2020-10-09 2022-04-14 Contact Control Interfaces, LLC Haptic engine for spatial computing
US11934583B2 (en) 2020-10-30 2024-03-19 Datafeel Inc. Wearable data communication apparatus, kits, methods, and systems
US11606636B2 (en) * 2021-06-03 2023-03-14 Feng-Chou Lin Somatosensory vibration generating device and method for forming somatosensory vibration
TWI831081B (en) * 2021-11-11 2024-02-01 國立陽明交通大學 Modular pneumatic somatosensory device
NL2032166B1 (en) * 2022-06-14 2023-12-21 Heusinkveld Eng B V Input device for a simulator environment, simulator provided therewith, and method for operating such device and simulator

Citations (91)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3270440A (en) * 1963-02-08 1966-09-06 Goodyear Aerospace Corp Seat for flight motion simulator
US3397952A (en) * 1965-09-29 1968-08-20 Aluminum Co Of America Production of fused alumina
US3736551A (en) * 1970-11-16 1973-05-29 Us Navy Swimmers tactile command navigation apparatus
US4030208A (en) * 1976-01-15 1977-06-21 The Singer Company Seat vibration system for simulating aircraft buffeting
US4081829A (en) * 1976-08-23 1978-03-28 Atari, Inc. Audio activated video display
US4218702A (en) * 1977-05-18 1980-08-19 Societe Nationale Industrielle Means for remote control of an aircraft video system for surveying ground activity
US4244120A (en) * 1979-06-11 1981-01-13 The United States Of America As Represented By The Secretary Of The Navy Acceleration cueing simulation device
US4257062A (en) * 1978-12-29 1981-03-17 Meredith Russell W Personalized audio-visual system
US4310315A (en) * 1978-08-21 1982-01-12 Therapeople, Inc. Tactile learning device for the handicapped
US4321044A (en) * 1978-11-14 1982-03-23 The Singer Company Advanced G cueing system
US4527980A (en) * 1983-04-07 1985-07-09 Miller Robert W Flight simulating video game
US4574391A (en) * 1983-08-22 1986-03-04 Funai Electric Company Limited Stereophonic sound producing apparatus for a game machine
US4581491A (en) * 1984-05-04 1986-04-08 Research Corporation Wearable tactile sensory aid providing information on voice pitch and intonation patterns
US4795296A (en) * 1986-11-17 1989-01-03 California Institute Of Technology Hand-held robot end effector controller having movement and force control
US4800721A (en) * 1987-02-13 1989-01-31 Caterpillar Inc. Force feedback lever
US4823634A (en) * 1987-11-03 1989-04-25 Culver Craig F Multifunction tactile manipulatable control
US4856078A (en) * 1988-03-23 1989-08-08 Zenith Electronics Corporation DC fan speed control
US4855822A (en) * 1988-01-26 1989-08-08 Honeywell, Inc. Human engineered remote driving system
US4949119A (en) * 1989-01-12 1990-08-14 Atari Games Corporation Gearshift for a vehicle simulator using computer controlled realistic real world forces
US4983901A (en) * 1989-04-21 1991-01-08 Allergan, Inc. Digital electronic foot control for medical apparatus and the like
US4988981A (en) * 1987-03-17 1991-01-29 Vpl Research, Inc. Computer data entry and manipulation apparatus and method
US5035242A (en) * 1990-04-16 1991-07-30 David Franklin Method and apparatus for sound responsive tactile stimulation of deaf individuals
US5103404A (en) * 1985-12-06 1992-04-07 Tensor Development, Inc. Feedback for a manipulator
US5105183A (en) * 1989-04-27 1992-04-14 Digital Equipment Corporation System for displaying video from a plurality of sources on a display
US5107262A (en) * 1988-10-13 1992-04-21 Ministere De La Culture, De La Communication, Des Grands Travaux Et Du Bicentenaire Modular retroactive keyboard and a flat modular actuator
US5184319A (en) * 1990-02-02 1993-02-02 Kramer James F Force feedback and textures simulating interface device
US5185561A (en) * 1991-07-23 1993-02-09 Digital Equipment Corporation Torque motor as a tactile feedback device in a computer system
US5184956A (en) * 1990-02-20 1993-02-09 Codes Rousseau Method and device for training in the driving of vehicles
US5186629A (en) * 1991-08-22 1993-02-16 International Business Machines Corporation Virtual graphics display capable of presenting icons and windows to the blind computer user and method
US5193963A (en) * 1990-10-31 1993-03-16 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Force reflecting hand controller
US5203563A (en) * 1991-03-21 1993-04-20 Atari Games Corporation Shaker control device
US5209661A (en) * 1990-10-29 1993-05-11 Systems Control Technology, Inc. Motor control desired dynamic load of a simulating system and method
US5220260A (en) * 1991-10-24 1993-06-15 Lex Computer And Management Corporation Actuator having electronically controllable tactile responsiveness
US5223776A (en) * 1990-12-31 1993-06-29 Honeywell Inc. Six-degree virtual pivot controller
US5240419A (en) * 1991-07-15 1993-08-31 Hughes Training, Inc. Condition monitor request processing system
US5275565A (en) * 1991-05-23 1994-01-04 Atari Games Corporation Modular display simulator and method
US5277584A (en) * 1991-09-06 1994-01-11 Occusym Limited Liability Company Vehicle vibration simulator and method for programming and using same
US5286203A (en) * 1992-10-07 1994-02-15 Aai Microflite Simulation International Simulating horizontal stabilizer trimming in an aircraft
US5296871A (en) * 1992-07-27 1994-03-22 Paley W Bradford Three-dimensional mouse with tactile feedback
US5299810A (en) * 1991-03-21 1994-04-05 Atari Games Corporation Vehicle simulator including cross-network feedback
US5307456A (en) * 1990-12-04 1994-04-26 Sony Electronics, Inc. Integrated multi-media production and authoring system
US5320538A (en) * 1992-09-23 1994-06-14 Hughes Training, Inc. Interactive aircraft training system and method
US5327120A (en) * 1992-07-06 1994-07-05 Motorola, Inc. Stabilized electromagnetic resonant armature tactile vibrator
US5381080A (en) * 1992-02-26 1995-01-10 Vdo Adolf Schindling Ag Control device
US5381347A (en) * 1992-12-21 1995-01-10 Microsoft Corporation Method and system for displaying images on a display device using an offscreen video memory
US5389865A (en) * 1992-12-02 1995-02-14 Cybernet Systems Corporation Method and system for providing a tactile virtual reality and manipulator defining an interface device therefor
US5388992A (en) * 1991-06-19 1995-02-14 Audiological Engineering Corporation Method and apparatus for tactile transduction of acoustic signals from television receivers
US5396266A (en) * 1993-06-08 1995-03-07 Technical Research Associates, Inc. Kinesthetic feedback apparatus and method
US5399091A (en) * 1992-04-27 1995-03-21 Tomy Company, Ltd. Drive simulation apparatus
US5405152A (en) * 1993-06-08 1995-04-11 The Walt Disney Company Method and apparatus for an interactive video game with physical feedback
US5428395A (en) * 1992-06-18 1995-06-27 Samsung Electronics Co., Ltd. Encoding and decoding method and apparatus thereof using a variable picture partitioning technique
US5482051A (en) * 1994-03-10 1996-01-09 The University Of Akron Electromyographic virtual reality system
US5512919A (en) * 1992-03-31 1996-04-30 Pioneer Electronic Corporation Three-dimensional coordinates input apparatus
US5513130A (en) * 1990-02-22 1996-04-30 Redmond Productions, Inc. Methods and apparatus for generating and processing synthetic and absolute real time environments
US5513129A (en) * 1993-07-14 1996-04-30 Fakespace, Inc. Method and system for controlling computer-generated virtual environment in response to audio signals
US5513100A (en) * 1993-06-10 1996-04-30 The University Of British Columbia Velocity controller with force feedback stiffness control
US5542672A (en) * 1995-03-17 1996-08-06 Meredith; Chris Fishing rod and reel electronic game controller
US5629594A (en) * 1992-12-02 1997-05-13 Cybernet Systems Corporation Force feedback system
US5631861A (en) * 1990-02-02 1997-05-20 Virtual Technologies, Inc. Force feedback and texture simulating interface device
US5642469A (en) * 1994-11-03 1997-06-24 University Of Washington Direct-drive manipulator for pen-based force display
US5643087A (en) * 1994-05-19 1997-07-01 Microsoft Corporation Input device including digital force feedback apparatus
US5707237A (en) * 1993-04-20 1998-01-13 Kabushiki Kaisha Ace Denken Driving simulation system
US5709219A (en) * 1994-01-27 1998-01-20 Microsoft Corporation Method and apparatus to create a complex tactile sensation
US5714978A (en) * 1994-12-05 1998-02-03 Nec Corporation Adjacent cursor system with tactile feedback for the blind
US5734373A (en) * 1993-07-16 1998-03-31 Immersion Human Interface Corporation Method and apparatus for controlling force feedback interface systems utilizing a host computer
US5736978A (en) * 1995-05-26 1998-04-07 The United States Of America As Represented By The Secretary Of The Air Force Tactile graphics display
US5739811A (en) * 1993-07-16 1998-04-14 Immersion Human Interface Corporation Method and apparatus for controlling human-computer interface systems providing force feedback
US5754023A (en) * 1995-10-26 1998-05-19 Cybernet Systems Corporation Gyro-stabilized platforms for force-feedback applications
US5755577A (en) * 1995-03-29 1998-05-26 Gillio; Robert G. Apparatus and method for recording data of a surgical procedure
US5767839A (en) * 1995-01-18 1998-06-16 Immersion Human Interface Corporation Method and apparatus for providing passive force feedback to human-computer interface systems
US5781172A (en) * 1990-12-05 1998-07-14 U.S. Philips Corporation Data input device for use with a data processing apparatus and a data processing apparatus provided with such a device
US5784052A (en) * 1995-03-13 1998-07-21 U.S. Philips Corporation Vertical translation of mouse or trackball enables truly 3D input
US5790108A (en) * 1992-10-23 1998-08-04 University Of British Columbia Controller
US5857986A (en) * 1996-05-24 1999-01-12 Moriyasu; Hiro Interactive vibrator for multimedia
US5865624A (en) * 1995-11-09 1999-02-02 Hayashigawa; Larry Reactive ride simulator apparatus and method
US5889672A (en) * 1991-10-24 1999-03-30 Immersion Corporation Tactiley responsive user interface device and method therefor
US5897437A (en) * 1995-10-09 1999-04-27 Nintendo Co., Ltd. Controller pack
US5904724A (en) * 1996-01-19 1999-05-18 Margolin; Jed Method and apparatus for remotely piloting an aircraft
US5921780A (en) * 1996-06-28 1999-07-13 Myers; Nicole J. Racecar simulator and driver training system and method
US5944151A (en) * 1995-08-03 1999-08-31 Vdo Adolf Schindling Ag Operating device
US6046709A (en) * 1997-01-17 2000-04-04 Intergraph Corporation Multiple display synchronization apparatus and method
US6088017A (en) * 1995-11-30 2000-07-11 Virtual Technologies, Inc. Tactile feedback man-machine interface device
US6184868B1 (en) * 1998-09-17 2001-02-06 Immersion Corp. Haptic feedback control devices
US6198206B1 (en) * 1998-03-20 2001-03-06 Active Control Experts, Inc. Inertial/audio unit and construction
US6211861B1 (en) * 1998-06-23 2001-04-03 Immersion Corporation Tactile mouse device
US6232932B1 (en) * 1998-07-16 2001-05-15 Craig A. Thorner Apparatus and method for providing modular reconfigurable multi-function displays for computer simulations
US6394904B1 (en) * 2000-05-12 2002-05-28 Twentieth Century Fox Film Simulation system
US6550565B2 (en) * 2000-02-18 2003-04-22 Delphi Technologies, Inc. Variable road feedback device for steer-by-wire systems
US6714045B2 (en) * 2001-07-02 2004-03-30 Intrinsity, Inc. Static transmission of FAST14 logic 1-of-N signals
US6752716B1 (en) * 1997-11-07 2004-06-22 Kabushiki Kaisha Sega Enterprises Game machine for simulating vibration
US6864877B2 (en) * 2000-09-28 2005-03-08 Immersion Corporation Directional tactile feedback for haptic feedback interface devices

Family Cites Families (72)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US1738289A (en) 1926-05-19 1929-12-03 Western Electric Co Tactile reception of sound
US3397352A (en) 1965-03-19 1968-08-13 American Standard Inc Fan speed control means employing a vibrating contact
US3919691A (en) 1971-05-26 1975-11-11 Bell Telephone Labor Inc Tactile man-machine communication system
US3902687A (en) 1973-06-25 1975-09-02 Robert E Hightower Aircraft indicator system
US3983640A (en) 1974-11-06 1976-10-05 The Singer Company Advanced G seat for aircraft simulation
FR2411603A2 (en) 1977-12-19 1979-07-13 Zarudiansky Alain DEVICE AND METHOD FOR RECORDING OF RESTITUTION AND SYNTHESIS OF TACTILE SENSATIONS
US4237449A (en) 1978-06-16 1980-12-02 Zibell J Scott Signalling device for hard of hearing persons
EP0085518B1 (en) 1982-01-22 1989-08-16 British Aerospace Public Limited Company Control apparatus
US4484191A (en) 1982-06-14 1984-11-20 Vavra George S Tactile signaling systems for aircraft
US4560983A (en) 1982-09-17 1985-12-24 Ampex Corporation Dynamically interactive responsive control device and system
US4477043A (en) 1982-12-15 1984-10-16 The United States Of America As Represented By The Secretary Of The Air Force Biodynamic resistant control stick
US4553748A (en) 1983-07-26 1985-11-19 Allen Ray L Electrostatically enhanced game
US4604016A (en) 1983-08-03 1986-08-05 Joyce Stephen A Multi-dimensional force-torque hand controller having force feedback
US4791416A (en) 1985-02-05 1988-12-13 Zenith Electronics Corporation Touch control system for controllable apparatus
IL74762A (en) 1985-03-29 1992-11-15 Morag Meir Helmet information display system
JPH0537531Y2 (en) 1985-06-11 1993-09-22
GB2179775B (en) 1985-08-30 1989-01-25 Unisys Corp Tactile alarm system for gaining the attention of an individual
NL8602624A (en) 1986-10-20 1988-05-16 Oce Nederland Bv INPUT DEVICE WITH TAKTILE FEEDBACK.
US4771344A (en) 1986-11-13 1988-09-13 James Fallacaro System for enhancing audio and/or visual presentation
US5986643A (en) 1987-03-24 1999-11-16 Sun Microsystems, Inc. Tactile feedback mechanism for a data processing system
US4779615A (en) 1987-05-13 1988-10-25 Frazier Richard K Tactile stimulator
US4868549A (en) 1987-05-18 1989-09-19 International Business Machines Corporation Feedback mouse
US4965559A (en) 1988-05-31 1990-10-23 Motorola, Inc. Multi-channel graphics controller
US4885565A (en) 1988-06-01 1989-12-05 General Motors Corporation Touchscreen CRT with tactile feedback
NL8801653A (en) 1988-06-29 1990-01-16 Stork Kwant Bv OPERATING SYSTEM.
US5044956A (en) 1989-01-12 1991-09-03 Atari Games Corporation Control device such as a steering wheel for video vehicle simulator with realistic feedback forces
US5076517A (en) 1989-08-14 1991-12-31 United Technologies Corporation Programmable, linear collective control system for a helicopter
CA2075996A1 (en) 1990-02-16 1991-08-17 Lester S. Caine Display system
US5269687A (en) 1990-08-01 1993-12-14 Atari Games Corporation System and method for recursive driver training
US5165897A (en) 1990-08-10 1992-11-24 Tini Alloy Company Programmable tactile stimulator array system and method of operation
CA2060361A1 (en) 1991-01-31 1992-08-01 Masao Fukuda Display control device
US5274363A (en) 1991-02-01 1993-12-28 Ibm Interactive display system
US5354162A (en) 1991-02-26 1994-10-11 Rutgers University Actuator system for providing force feedback to portable master support
GB9108497D0 (en) 1991-04-20 1991-06-05 Ind Limited W Human/computer interface
US5146566A (en) 1991-05-29 1992-09-08 Ibm Corporation Input/output system for computer user interface using magnetic levitation
CA2068476C (en) 1991-08-19 1996-07-23 Frank A. Mckiel, Jr. Audio user interface with stereo and filtered sound effects
US5559432A (en) 1992-02-27 1996-09-24 Logue; Delmar L. Joystick generating a polar coordinates signal utilizing a rotating magnetic field within a hollow toroid core
US5589828A (en) 1992-03-05 1996-12-31 Armstrong; Brad A. 6 Degrees of freedom controller with capability of tactile feedback
US5368484A (en) 1992-05-22 1994-11-29 Atari Games Corp. Vehicle simulator with realistic operating feedback
US5366376A (en) 1992-05-22 1994-11-22 Atari Games Corporation Driver training system and method with performance data feedback
WO1994000206A1 (en) 1992-06-29 1994-01-06 Kabushiki Kaisha Ace Denken Chair for game machine
US5551701A (en) 1992-08-19 1996-09-03 Thrustmaster, Inc. Reconfigurable video game controller with graphical reconfiguration display
US5666473A (en) 1992-10-08 1997-09-09 Science & Technology Corporation & Unm Tactile computer aided sculpting device
US5451924A (en) 1993-01-14 1995-09-19 Massachusetts Institute Of Technology Apparatus for providing sensory substitution of force feedback
EP0607580A1 (en) 1993-01-21 1994-07-27 International Business Machines Corporation Tactile feedback mechanism for cursor control
JP3686686B2 (en) 1993-05-11 2005-08-24 松下電器産業株式会社 Haptic device, data input device, and data input device device
US5805140A (en) 1993-07-16 1998-09-08 Immersion Corporation High bandwidth force feedback interface using voice coils and flexures
US5701140A (en) 1993-07-16 1997-12-23 Immersion Human Interface Corp. Method and apparatus for providing a cursor control interface with force feedback
US5625576A (en) 1993-10-01 1997-04-29 Massachusetts Institute Of Technology Force reflecting haptic interface
DE4400790A1 (en) 1993-11-10 1995-05-18 Johannes Dipl Ing Geisen GUI computer mouse with force feed back and position data
US5553148A (en) 1994-06-20 1996-09-03 Werle; Ben Apparatus and method for producing vibratory sensations to accompany audible sounds in a properly phased relationship
US5803738A (en) 1994-06-24 1998-09-08 Cgsd Corporation Apparatus for robotic force simulation
ATE189539T1 (en) 1994-08-18 2000-02-15 Interval Research Corp INPUT DEVICE FOR VIDEO WITH TACTILE FEEDBACK DEPENDING ON THE CONTENT OF THE VIDEO
US5669818A (en) 1995-03-23 1997-09-23 Thorner; Craig Seat-based tactile sensation generator
US5684722A (en) 1994-09-21 1997-11-04 Thorner; Craig Apparatus and method for generating a control signal for a tactile sensation generator
US6422941B1 (en) 1994-09-21 2002-07-23 Craig Thorner Universal tactile feedback system for computer video games and simulations
US5565840A (en) 1994-09-21 1996-10-15 Thorner; Craig Tactile sensation generator
US5666138A (en) 1994-11-22 1997-09-09 Culver; Craig F. Interface control
US5583478A (en) 1995-03-01 1996-12-10 Renzi; Ronald Virtual environment tactile system
US5691898A (en) * 1995-09-27 1997-11-25 Immersion Human Interface Corp. Safe and low cost computer peripherals with force feedback for consumer applications
US5589854A (en) 1995-06-22 1996-12-31 Tsai; Ming-Chang Touching feedback device
US5802318A (en) 1995-07-25 1998-09-01 Compaq Computer Corporation Universal serial bus keyboard system
US5959613A (en) 1995-12-01 1999-09-28 Immersion Corporation Method and apparatus for shaping force signals for a force feedback device
US5825308A (en) 1996-11-26 1998-10-20 Immersion Human Interface Corporation Force feedback interface having isotonic and isometric functionality
US5973670A (en) 1996-12-31 1999-10-26 International Business Machines Corporation Tactile feedback controller for computer cursor control device
US5956046A (en) 1997-12-17 1999-09-21 Sun Microsystems, Inc. Scene synchronization of multiple computer displays
US6659773B2 (en) 1998-03-04 2003-12-09 D-Box Technology Inc. Motion transducer system
US6139324A (en) 1998-03-04 2000-10-31 D-Box Audio Inc. Multi-sense home entertainment chair transducer system
US5980255A (en) 1998-03-16 1999-11-09 Cae Electronics Ltd. Seat for motion simulator and method of motion simulation
US6693622B1 (en) 1999-07-01 2004-02-17 Immersion Corporation Vibrotactile haptic feedback devices
US6827579B2 (en) 2000-11-16 2004-12-07 Rutgers, The State University Of Nj Method and apparatus for rehabilitation of neuromotor disorders
JP4008381B2 (en) 2003-05-09 2007-11-14 リンナイ株式会社 Stove top

Patent Citations (100)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3270440A (en) * 1963-02-08 1966-09-06 Goodyear Aerospace Corp Seat for flight motion simulator
US3397952A (en) * 1965-09-29 1968-08-20 Aluminum Co Of America Production of fused alumina
US3736551A (en) * 1970-11-16 1973-05-29 Us Navy Swimmers tactile command navigation apparatus
US4030208A (en) * 1976-01-15 1977-06-21 The Singer Company Seat vibration system for simulating aircraft buffeting
US4081829A (en) * 1976-08-23 1978-03-28 Atari, Inc. Audio activated video display
US4218702A (en) * 1977-05-18 1980-08-19 Societe Nationale Industrielle Means for remote control of an aircraft video system for surveying ground activity
US4310315A (en) * 1978-08-21 1982-01-12 Therapeople, Inc. Tactile learning device for the handicapped
US4321044A (en) * 1978-11-14 1982-03-23 The Singer Company Advanced G cueing system
US4257062A (en) * 1978-12-29 1981-03-17 Meredith Russell W Personalized audio-visual system
US4244120A (en) * 1979-06-11 1981-01-13 The United States Of America As Represented By The Secretary Of The Navy Acceleration cueing simulation device
US4527980A (en) * 1983-04-07 1985-07-09 Miller Robert W Flight simulating video game
US4574391A (en) * 1983-08-22 1986-03-04 Funai Electric Company Limited Stereophonic sound producing apparatus for a game machine
US4581491A (en) * 1984-05-04 1986-04-08 Research Corporation Wearable tactile sensory aid providing information on voice pitch and intonation patterns
US5103404A (en) * 1985-12-06 1992-04-07 Tensor Development, Inc. Feedback for a manipulator
US4795296A (en) * 1986-11-17 1989-01-03 California Institute Of Technology Hand-held robot end effector controller having movement and force control
US4800721A (en) * 1987-02-13 1989-01-31 Caterpillar Inc. Force feedback lever
US4988981A (en) * 1987-03-17 1991-01-29 Vpl Research, Inc. Computer data entry and manipulation apparatus and method
US4988981B1 (en) * 1987-03-17 1999-05-18 Vpl Newco Inc Computer data entry and manipulation apparatus and method
US4823634A (en) * 1987-11-03 1989-04-25 Culver Craig F Multifunction tactile manipulatable control
US4855822A (en) * 1988-01-26 1989-08-08 Honeywell, Inc. Human engineered remote driving system
US4856078A (en) * 1988-03-23 1989-08-08 Zenith Electronics Corporation DC fan speed control
US5107262A (en) * 1988-10-13 1992-04-21 Ministere De La Culture, De La Communication, Des Grands Travaux Et Du Bicentenaire Modular retroactive keyboard and a flat modular actuator
US4949119A (en) * 1989-01-12 1990-08-14 Atari Games Corporation Gearshift for a vehicle simulator using computer controlled realistic real world forces
US4983901A (en) * 1989-04-21 1991-01-08 Allergan, Inc. Digital electronic foot control for medical apparatus and the like
US5105183A (en) * 1989-04-27 1992-04-14 Digital Equipment Corporation System for displaying video from a plurality of sources on a display
US5184319A (en) * 1990-02-02 1993-02-02 Kramer James F Force feedback and textures simulating interface device
US5631861A (en) * 1990-02-02 1997-05-20 Virtual Technologies, Inc. Force feedback and texture simulating interface device
US5184956A (en) * 1990-02-20 1993-02-09 Codes Rousseau Method and device for training in the driving of vehicles
US5513130A (en) * 1990-02-22 1996-04-30 Redmond Productions, Inc. Methods and apparatus for generating and processing synthetic and absolute real time environments
US5035242A (en) * 1990-04-16 1991-07-30 David Franklin Method and apparatus for sound responsive tactile stimulation of deaf individuals
US5209661A (en) * 1990-10-29 1993-05-11 Systems Control Technology, Inc. Motor control desired dynamic load of a simulating system and method
US5634794A (en) * 1990-10-29 1997-06-03 Systems Control Technology Inc. Aircraft simulator and method
US5193963A (en) * 1990-10-31 1993-03-16 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Force reflecting hand controller
US5307456A (en) * 1990-12-04 1994-04-26 Sony Electronics, Inc. Integrated multi-media production and authoring system
US5781172A (en) * 1990-12-05 1998-07-14 U.S. Philips Corporation Data input device for use with a data processing apparatus and a data processing apparatus provided with such a device
US5223776A (en) * 1990-12-31 1993-06-29 Honeywell Inc. Six-degree virtual pivot controller
US5203563A (en) * 1991-03-21 1993-04-20 Atari Games Corporation Shaker control device
US5299810A (en) * 1991-03-21 1994-04-05 Atari Games Corporation Vehicle simulator including cross-network feedback
US5275565A (en) * 1991-05-23 1994-01-04 Atari Games Corporation Modular display simulator and method
US5388992A (en) * 1991-06-19 1995-02-14 Audiological Engineering Corporation Method and apparatus for tactile transduction of acoustic signals from television receivers
US5240419A (en) * 1991-07-15 1993-08-31 Hughes Training, Inc. Condition monitor request processing system
US5185561A (en) * 1991-07-23 1993-02-09 Digital Equipment Corporation Torque motor as a tactile feedback device in a computer system
US5186629A (en) * 1991-08-22 1993-02-16 International Business Machines Corporation Virtual graphics display capable of presenting icons and windows to the blind computer user and method
US5277584A (en) * 1991-09-06 1994-01-11 Occusym Limited Liability Company Vehicle vibration simulator and method for programming and using same
US5889672A (en) * 1991-10-24 1999-03-30 Immersion Corporation Tactiley responsive user interface device and method therefor
US5889670A (en) * 1991-10-24 1999-03-30 Immersion Corporation Method and apparatus for tactilely responsive user interface
US5220260A (en) * 1991-10-24 1993-06-15 Lex Computer And Management Corporation Actuator having electronically controllable tactile responsiveness
US5414337A (en) * 1991-10-24 1995-05-09 Lex Computer And Management Corporation Actuator having electronically controllable tactile responsiveness
US5381080A (en) * 1992-02-26 1995-01-10 Vdo Adolf Schindling Ag Control device
US5512919A (en) * 1992-03-31 1996-04-30 Pioneer Electronic Corporation Three-dimensional coordinates input apparatus
US5399091A (en) * 1992-04-27 1995-03-21 Tomy Company, Ltd. Drive simulation apparatus
US5428395A (en) * 1992-06-18 1995-06-27 Samsung Electronics Co., Ltd. Encoding and decoding method and apparatus thereof using a variable picture partitioning technique
US5327120A (en) * 1992-07-06 1994-07-05 Motorola, Inc. Stabilized electromagnetic resonant armature tactile vibrator
US5296871A (en) * 1992-07-27 1994-03-22 Paley W Bradford Three-dimensional mouse with tactile feedback
US5320538A (en) * 1992-09-23 1994-06-14 Hughes Training, Inc. Interactive aircraft training system and method
US5286203A (en) * 1992-10-07 1994-02-15 Aai Microflite Simulation International Simulating horizontal stabilizer trimming in an aircraft
US5790108A (en) * 1992-10-23 1998-08-04 University Of British Columbia Controller
US6104158A (en) * 1992-12-02 2000-08-15 Immersion Corporation Force feedback system
US5629594A (en) * 1992-12-02 1997-05-13 Cybernet Systems Corporation Force feedback system
US5389865A (en) * 1992-12-02 1995-02-14 Cybernet Systems Corporation Method and system for providing a tactile virtual reality and manipulator defining an interface device therefor
US5381347A (en) * 1992-12-21 1995-01-10 Microsoft Corporation Method and system for displaying images on a display device using an offscreen video memory
US5707237A (en) * 1993-04-20 1998-01-13 Kabushiki Kaisha Ace Denken Driving simulation system
US5396266A (en) * 1993-06-08 1995-03-07 Technical Research Associates, Inc. Kinesthetic feedback apparatus and method
US5405152A (en) * 1993-06-08 1995-04-11 The Walt Disney Company Method and apparatus for an interactive video game with physical feedback
US5513100A (en) * 1993-06-10 1996-04-30 The University Of British Columbia Velocity controller with force feedback stiffness control
US5513129A (en) * 1993-07-14 1996-04-30 Fakespace, Inc. Method and system for controlling computer-generated virtual environment in response to audio signals
US5739811A (en) * 1993-07-16 1998-04-14 Immersion Human Interface Corporation Method and apparatus for controlling human-computer interface systems providing force feedback
US5734373A (en) * 1993-07-16 1998-03-31 Immersion Human Interface Corporation Method and apparatus for controlling force feedback interface systems utilizing a host computer
US5709219A (en) * 1994-01-27 1998-01-20 Microsoft Corporation Method and apparatus to create a complex tactile sensation
US5742278A (en) * 1994-01-27 1998-04-21 Microsoft Corporation Force feedback joystick with digital signal processor controlled by host processor
US5482051A (en) * 1994-03-10 1996-01-09 The University Of Akron Electromyographic virtual reality system
US6036495A (en) * 1994-05-19 2000-03-14 Microsoft Corporation Interactive simulation including force feedback
US5643087A (en) * 1994-05-19 1997-07-01 Microsoft Corporation Input device including digital force feedback apparatus
US5642469A (en) * 1994-11-03 1997-06-24 University Of Washington Direct-drive manipulator for pen-based force display
US5714978A (en) * 1994-12-05 1998-02-03 Nec Corporation Adjacent cursor system with tactile feedback for the blind
US5767839A (en) * 1995-01-18 1998-06-16 Immersion Human Interface Corporation Method and apparatus for providing passive force feedback to human-computer interface systems
US5784052A (en) * 1995-03-13 1998-07-21 U.S. Philips Corporation Vertical translation of mouse or trackball enables truly 3D input
US5542672A (en) * 1995-03-17 1996-08-06 Meredith; Chris Fishing rod and reel electronic game controller
US5755577A (en) * 1995-03-29 1998-05-26 Gillio; Robert G. Apparatus and method for recording data of a surgical procedure
US5736978A (en) * 1995-05-26 1998-04-07 The United States Of America As Represented By The Secretary Of The Air Force Tactile graphics display
US5944151A (en) * 1995-08-03 1999-08-31 Vdo Adolf Schindling Ag Operating device
US5897437A (en) * 1995-10-09 1999-04-27 Nintendo Co., Ltd. Controller pack
US5754023A (en) * 1995-10-26 1998-05-19 Cybernet Systems Corporation Gyro-stabilized platforms for force-feedback applications
US5865624A (en) * 1995-11-09 1999-02-02 Hayashigawa; Larry Reactive ride simulator apparatus and method
US6275213B1 (en) * 1995-11-30 2001-08-14 Virtual Technologies, Inc. Tactile feedback man-machine interface device
US6424333B1 (en) * 1995-11-30 2002-07-23 Immersion Corporation Tactile feedback man-machine interface device
US6088017A (en) * 1995-11-30 2000-07-11 Virtual Technologies, Inc. Tactile feedback man-machine interface device
US5904724A (en) * 1996-01-19 1999-05-18 Margolin; Jed Method and apparatus for remotely piloting an aircraft
US5857986A (en) * 1996-05-24 1999-01-12 Moriyasu; Hiro Interactive vibrator for multimedia
US5921780A (en) * 1996-06-28 1999-07-13 Myers; Nicole J. Racecar simulator and driver training system and method
US6046709A (en) * 1997-01-17 2000-04-04 Intergraph Corporation Multiple display synchronization apparatus and method
US6752716B1 (en) * 1997-11-07 2004-06-22 Kabushiki Kaisha Sega Enterprises Game machine for simulating vibration
US6198206B1 (en) * 1998-03-20 2001-03-06 Active Control Experts, Inc. Inertial/audio unit and construction
US6211861B1 (en) * 1998-06-23 2001-04-03 Immersion Corporation Tactile mouse device
US6232932B1 (en) * 1998-07-16 2001-05-15 Craig A. Thorner Apparatus and method for providing modular reconfigurable multi-function displays for computer simulations
US6184868B1 (en) * 1998-09-17 2001-02-06 Immersion Corp. Haptic feedback control devices
US6550565B2 (en) * 2000-02-18 2003-04-22 Delphi Technologies, Inc. Variable road feedback device for steer-by-wire systems
US6394904B1 (en) * 2000-05-12 2002-05-28 Twentieth Century Fox Film Simulation system
US6864877B2 (en) * 2000-09-28 2005-03-08 Immersion Corporation Directional tactile feedback for haptic feedback interface devices
US6714045B2 (en) * 2001-07-02 2004-03-30 Intrinsity, Inc. Static transmission of FAST14 logic 1-of-N signals

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7927217B2 (en) * 2001-12-14 2011-04-19 Kabushiki Kaisha Square Enix Network game system, game server system, client system, network game processing method, and recording medium
US20030114226A1 (en) * 2001-12-14 2003-06-19 Square Co., Ltd. Network game system, game server system, client system, network game processing method, and recording medium
US20060119576A1 (en) * 2004-12-06 2006-06-08 Naturalpoint, Inc. Systems and methods for using a movable object to control a computer
US20060119575A1 (en) * 2004-12-06 2006-06-08 Naturalpoint, Inc. Systems and methods for using a movable object to control a computer
US8179366B2 (en) * 2004-12-06 2012-05-15 Naturalpoint, Inc. Systems and methods for using a movable object to control a computer
EP2067508A1 (en) * 2007-11-29 2009-06-10 AMBX UK Limited A method for providing a sensory effect to augment an experience provided by a video game
US9880622B2 (en) * 2009-12-21 2018-01-30 Kyocera Corporation Tactile sensation providing apparatus and control method for tactile sensation providing apparatus when using an application that does not support operation of tactile sensation
US20120262404A1 (en) * 2009-12-21 2012-10-18 Kyocera Corporation Tactile sensation providing apparatus and control method for tactile sensation providing apparatus
US8660673B2 (en) 2011-03-23 2014-02-25 Electronic Entertainment Design And Research In-game contextual telemetry systems and methods
US20150169058A1 (en) * 2012-03-30 2015-06-18 Nvf Tech Ltd Touch and Haptics Device
US10216275B2 (en) * 2012-03-30 2019-02-26 Nvf Tech Ltd Touch and haptics device
US20180068537A1 (en) * 2013-09-06 2018-03-08 Immersion Corporation Spatialized haptic feedback based on dynamically scaled values
US10553083B2 (en) * 2013-09-06 2020-02-04 Immersion Corporation Spatialized haptic feedback based on dynamically scaled values
US10909821B2 (en) 2013-09-06 2021-02-02 Immersion Corporation Spatialized haptic feedback based on dynamically scaled values
US20170108995A1 (en) * 2015-10-16 2017-04-20 Microsoft Technology Licensing, Llc Customizing Program Features on a Per-User Basis
US10101870B2 (en) * 2015-10-16 2018-10-16 Microsoft Technology Licensing, Llc Customizing program features on a per-user basis
US11392203B2 (en) * 2018-03-27 2022-07-19 Sony Corporation Information processing apparatus, information processing method, and program

Also Published As

Publication number Publication date
US20130165226A1 (en) 2013-06-27
US8328638B2 (en) 2012-12-11
US20080280678A1 (en) 2008-11-13
US20090069081A1 (en) 2009-03-12
US8308558B2 (en) 2012-11-13
US8834266B2 (en) 2014-09-16

Similar Documents

Publication Publication Date Title
US8328638B2 (en) Method and apparatus for generating tactile feedback via relatively low-burden and/or zero burden telemetry
EP0951714B1 (en) Method and apparatus for shaping force signals for a force feedback device
US6752716B1 (en) Game machine for simulating vibration
Israr et al. Surround haptics: Tactile feedback for immersive gaming experiences
EP3491632B1 (en) Apparatus to simulate driving and corresponding method
US9558676B2 (en) Method for simulating specific movements by haptic feedback, and device implementing the method
JP2010123128A (en) System and method for providing augmented reality experience
KR20090122875A (en) Real-time virtual realrity sports platform apparutus using feedback motion base and power feedback health equipments
WO2005050427A1 (en) Tactile force sense information display system and method
US20110124413A1 (en) Haptic-simulation home-video game
US20220049957A1 (en) Gyroscope devices with control rotors and reaction wheels
JPWO2005066918A1 (en) Simulation device
CN113593358A (en) Two-degree-of-freedom VR airship driving simulation system
JPH08276070A (en) Game machine
Yoo et al. A framework for a multi-sensory VR effect system with motional display
KR20140101455A (en) Control device of combat simulation game
JP2001066981A (en) Experiencing device for video and sound direction
JP3078355U (en) Desktop vibration joystick
KR101975941B1 (en) Apparatus for controlling operation of virtual reality flying motion simulator and method thereof
JP2002085855A (en) System of utilizing game software and method of delivery for the same
CN106110660A (en) Three 360 degree of axles comprehensive high speed sensation simulation emulator
ITMI20071318A1 (en) SYSTEM, METHOD AND VIDEO-LUDIC PROGRAM TO DISTRIBUTE CONTENTS OF A SPORTS EVENT THROUGH A COMMUNICATION NETWORK.
KR20020063998A (en) Simulation-Playing Machine
KR20160132301A (en) Frame-based shoes that has vibration motor for software(game) event reaction
KR20240122164A (en) Spherical device for motion simulation and system for providing experiential content including the same

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION