[go: nahoru, domu]

Jump to content

Automotive head-up display

From Wikipedia, the free encyclopedia

An automotive head-up display or automotive heads-up display — also known as an auto-HUD — is any transparent display that presents data in the automobile without requiring users to look away from their usual viewpoints. The origin of the name stems from a pilot being able to view information with the head positioned "up" and looking forward, instead of angled down looking at lower instruments. At this time, there are three different approaches to OEM HUDs in automobiles. The first is to treat the back of the windshield in such a way that an image projected onto it will reflect to the driver. The second is to have a small combiner that is separate from the windshield. Combiners can be retracted. The third is to laminate a transparent display in between layers of the windshield glass.[1]

HUD in a BMW E60
HUD in a Pontiac Bonneville showing a speed of 47 mph (76 km/h)
The green arrow on the windshield near the top of this picture is a Head Up Display on a 2013 Toyota Prius. It toggles between the GPS navigation instruction arrow and the speedometer. The arrow is animated to appear scrolling forward as the car approaches the turn. The image is projected without any kind of glass combiner.
HUD in a Mazda using a retractable combiner rather than being reflected from the windshield.
Mazda CX-9 active driving display with traffic sign recognition

Timeline

[edit]
  • 1988: Nissan was the first manufacturer to offer a HUD in the JDM market with the 1988 Nissan Silvia S13.
  • 1988: General Motors began using head-up displays. Their first HUD units were installed on Oldsmobile Cutlass Supreme Indy Pace Cars and replicas. Optional HUD units were subsequently offered on the Cutlass Supreme and Pontiac Grand Prix before being more widely available.
  • 1989–1994: Nissan offered a head-up display in the Nissan 240SX.[2]
  • 1991: Toyota, for the Japanese market only, released a HUD system for the Toyota Crown Majesta.
  • 1998: The first High Content Reconfigurable display appeared on the Chevrolet Corvette (C5). (1999 Model Year)
  • 1999: Cadillac DTS with night-vision via Head-up Display. (Model Year 2000)
  • 2003: Cadillac introduced a HUD system for the Cadillac XLR.
  • 2003: BMW was involved in large developments for automotive HUD systems for the 2003 E60 5 Series.
  • 2012: Pioneer Corporation introduced a navigation system that projects a HUD in place of the driver's visor that presents animations of conditions ahead, a form of augmented reality (AR).[3][4]

These displays are becoming increasingly available in production cars, and usually offer speedometer, tachometer, and navigation system displays.

Night vision information is also displayed via HUD on certain General Motors, Honda, Toyota and Lexus vehicles. Other manufactures such as Audi, BMW, Citroën, Nissan, Mazda, Kia, Mercedes and Volvo currently offer some form of HUD system.

Motorcycle helmet HUDs are also commercially available.[5]

Add-on HUD systems also exist, projecting the display onto a glass combiner mounted on the windshield. These systems have been marketed to police agencies for use with in-vehicle computers.[citation needed]

Eyes-on-the-Road-Benefit

[edit]

The Eyes-on-the-Road-Benefit (ERB), also known as the Head-Up-Display-Advantage, is the term given to the purported advantages provided to motorists when driving using a head-up display (HUD).[6] This can also be referred to as a heads-up-device or heads-up design, as compared to traditional dashboard designs, which are referred to as Head-Down-Design (HDD). The benefit of Eyes-on-the-Road systems stems from increased situational awareness and elimination of the need to look away from the road whilst driving, thereby increasing reaction time to external hazards, such as pedestrians.[7] There is some evidence to suggest that the scope of the ERB is limited to low cognitive load situations in which the driving task is not particularly complex.[6]

Aetiology

[edit]

Research into the ERB primarily utilizes virtual reality driving simulators to mimic real life driving scenarios while eliminating situational variability. In order to examine HUDs and HDDs, studies often compare hazard reaction time, situational awareness, and quality of driving (such as speed consistency) using both systems. The extent of the ERB on different demographics, particularly those of age and experience level, are of particular interest.[citation needed] The interaction between work-load and the influence of ERB are also frequently examined for research.

Exogenous saccadic gaze

[edit]

Saccadic gaze is the perceptual mechanism through which the eye is inadvertently drawn to external stimulus without the individual's conscious action.[8] An involuntary gaze is most easily drawn by movement or distinct changes in illumination in an individual's visual field.[9] These external stimuli can be beneficial in such situations as the movement of a pedestrian about to walk out onto the road, in turn allowing the driver to take evasive action. Exogenous cues can also be irrelevant, and often dangerous, leading to distraction from goal behaviours, such as the flashing of a cellphone taking one's eyes off the road. By superimposing vital driving information onto the horizon in a driver's direct line of sight, HUDS allow important exogenous cues, like the movements of other vehicles to draw the gaze of a driver whilst they monitor vital vehicle feedback such as speed or revolution count.[10] It is theorized that this can facilitate faster reaction times to hazards and improve situational awareness. A collaborative project between Faurecia Groupe and Indian Institute of Science developed an eye gaze and finger controlled head up display[11] for cars that can also automatically estimate drivers’ cognitive load and distraction.[12]

Ideal visual field

[edit]

The ideal visual field is the area in which stimuli are most accurately, rapidly, and efficiently processed by the eye. In humans, this field is thought to be within 20 degrees above or below the vertical meridian of an individual's gaze and 60 degrees either side of the horizontal meridian.[13] If an object is beyond these boundaries it will require eye movement to bring the stimuli out of periphery. By including feedback instruments in the primary field of vision, HUDs allow for the horizon and all associated stimuli to stay in the primary field vision where the information may still be processed and acknowledged by a motorist.[14]

Manifestation

[edit]

Reaction time

[edit]

Reaction time, and more specifically delayed reaction, is widely cited as a key contributor to vehicular accidents.[15] Reaction time in relation to the ERB is defined as the time it takes for a motorist to react to an external hazard or stimuli and then carry out the appropriate reaction, or evasive maneuver such as braking when a vehicle in front stops. The feedback offered by a HUD is projected onto the windshield of a vehicle with the aim of integrating outside stimuli and the instrumental feedback; thus removing the need to remove a driver's eyes from the road. Studies of reaction time to hazards in HUD vs HDD designs have found that the average reaction times for HUD are faster.[7] This trend appears to continue across demographics, including both categories of experience level and age.[16][17]

Speed maintenance and driving quality

[edit]

Speed maintenance is the extent to which a driver maintains a speed and adjusts their speed to suit traffic laws and environmental conditions. The use of HUDs appears to produce better speed maintenance in drivers under experimental conditions when compared to HDDs.[6] It is theorized that this is because having the speedometer at the eye level of the vehicle operator allows for continuous monitoring of the vehicle's speed. HUD use also appears to increase general driving quality, including staying within road markings, and increased smoothness of driving and navigation abilities.[18] Drivers’ capacity to focus on external cues, such as road texture, road demarcations and street signs is increased by using a seamless interface where focus on the road isn't interrupted to assess speed and other information.

Limitations

[edit]

Work load

[edit]

The influence of ERB on drivers is not universal. There is evidence that as the complexity of driving tasks increases, the benefits of using a HUD are decreased, and in some circumstances, they are no longer statistically significant. The ERB is diminished, for example, when individuals are driving cognitively demanding vehicles, such as industrial vehicles, or when they are asked to multitask while driving.[6] One study has shown that when placed in a cognitively demanding condition, individuals shift their focus from the road alone to focus on other tasks such as shifting gears or talking to others. Subsequently, a driver's ability to process HUD feedback requires diversion of attention, much akin to that which occurs whilst using a HDD.[6]

Placement

[edit]

There are limitations for where a HUD can be placed or projected in a vehicle before it begins to diminish the ERB and becomes more of a distraction. HUDs can be constructed so that the instrumental feedback appears to be projected out into the horizon, rather than displayed directly on the windshield.[19] In test situations, a projected HUD which appears near the nose of the vehicle is said to result in the most rapid response times and best situational awareness on the part of the driver, as well as facilitating better driving quality.[19] For in-glass laminated HUD, the display glass part is integrated in the windshield while the electronics shall be placed and hidden inside the vehicle body. The information is displayed directly on the windshield.

See also

[edit]
  • Augmented reality – View of the real world with computer-generated supplementary features
  • B-pillar – Vertical or near vertical support of a car's window area or greenhouse
  • Overtaking – Vehicle-passing maneuver
  • Smartwatch – Wearable computer in the form of a watch

References

[edit]
  1. ^ LUMINEQ. "How is LUMINEQ transparent display laminated in glass". www.lumineq.com. Retrieved 17 March 2022.
  2. ^ "Nissan 240 Review". Edmunds.com. Retrieved 4 December 2012.
  3. ^ Pioneer launches car navigation with augmented reality, heads-up displays. System also uses dash cams to share images of street conditions across Japan. Alabaster, Jay | Computerworld | Pioneer launches car navigation with augmented reality, heads-up displays June 28, 2013
  4. ^ Ulanoff, Lance | Mashable | Pioneer AR Heads Up Display Augments Your Driving Reality January 11, 2012
  5. ^ "Mike, Werner. "Test Driving the SportVue Motorcycle HUD". Motorcycles in the Fast Lane. 8 November 2005. Accessed 14 February 2007". News.motorbiker.org. Archived from the original on 30 March 2010. Retrieved 2 October 2009.
  6. ^ a b c d e Liu, Yung-Ching; Wen, Ming-Hui (1 November 2004). "Comparison of head-up display (HUD) vs. head-down display (HDD): driving performance of commercial vehicle operators in Taiwan". International Journal of Human-Computer Studies. 61 (5): 679–697. doi:10.1016/j.ijhcs.2004.06.002.
  7. ^ a b Vlachos, George; Papanastasiou, Stylianos; Charissis, Vassilis (14 April 2008). "Comparative Study of Prototype Automotive HUD vs. HDD: Collision Avoidance Simulation and Results". SAE Technical Paper Series. Vol. 1. doi:10.4271/2008-01-0203. Retrieved 6 June 2016.
  8. ^ Murray, David W.; Bradshaw, Kevin J.; McLauchlan, Philip F.; Reid, Ian D.; Sharkey, Paul M. (1 November 1995). "Driving saccade to pursuit using image motion". International Journal of Computer Vision. 16 (3): 205–228. doi:10.1007/BF01539627. ISSN 0920-5691. S2CID 1296467.
  9. ^ Trappenberg, Thomas P.; Dorris, Michael C.; Munoz, Douglas P.; Klein, Raymond M. (1 February 2001). "A Model of Saccade Initiation Based on the Competitive Integration of Exogenous and Endogenous Signals in the Superior Colliculus". Journal of Cognitive Neuroscience. 13 (2): 256–271. doi:10.1162/089892901564306. ISSN 0898-929X. PMID 11244550. S2CID 2291169.
  10. ^ Charissis, V.; Naef, M. (1 June 2007). "Evaluation of Prototype Automotive Head-Up Display Interface: Testing Driver's Focusing Ability through a VR Simulation". 2007 IEEE Intelligent Vehicles Symposium. pp. 560–565. doi:10.1109/IVS.2007.4290174. ISBN 978-1-4244-1067-5. S2CID 545446.
  11. ^ Prabhakar, Gowdham; Ramakrishnan, Aparna; Madan, Modiksha; Murthy, L. R. D.; Sharma, Vinay Krishna; Deshmukh, Sachin; Biswas, Pradipta (2020). "Interactive gaze and finger controlled HUD for cars". Journal on Multimodal User Interfaces. 14: 101–121. doi:10.1007/s12193-019-00316-9. ISSN 1783-8738. S2CID 208261516.
  12. ^ Biswas, Pradipta; Prabhakar, Gowdham (1 April 2018). "Detecting drivers' cognitive load from saccadic intrusion". Transportation Research Part F: Traffic Psychology and Behaviour. 54: 63–78. doi:10.1016/j.trf.2018.01.017. ISSN 1369-8478.
  13. ^ Curcio, Christine A.; Allen, Kimberly A. (1 October 1990). "Topography of ganglion cells in human retina". The Journal of Comparative Neurology. 300 (1): 5–25. doi:10.1002/cne.903000103. ISSN 1096-9861. PMID 2229487. S2CID 20259581.
  14. ^ Underwood, Geoffrey; Chapman, Peter; Brocklehurst, Neil; Underwood, Jean; Crundall, David (1 January 2003). "Visual attention while driving: sequences of eye fixations made by experienced and novice drivers". Ergonomics. 46 (6): 629–646. doi:10.1080/0014013031000090116. ISSN 0014-0139. PMID 12745692. S2CID 15597397.
  15. ^ Harbluk, Joanne L.; Noy, Y. Ian; Trbovich, Patricia L.; Eizenman, Moshe (1 March 2007). "An on-road assessment of cognitive distraction: Impacts on drivers' visual behavior and braking performance". Accident Analysis & Prevention. 39 (2): 372–379. doi:10.1016/j.aap.2006.08.013. PMID 17054894.
  16. ^ Travis, David (1 July 1993). "The human factors perspective". Displays. 14 (3): 178–181. doi:10.1016/0141-9382(93)90040-C.
  17. ^ Wolffsohn, J. S; McBrien, N. A; Edgar, G. K; Stout, T (1 May 1998). "The influence of cognition and age on accommodation, detection rate and response times when using a car head-up display (HUD)". Ophthalmic and Physiological Optics. 18 (3): 243–253. doi:10.1016/S0275-5408(97)00094-X. PMID 9829111.
  18. ^ Harrison, Beverly L.; Ishii, Hiroshi; Vicente, Kim J.; Buxton, William A. S. (1 January 1995). "Transparent layered user interfaces: An evaluation of a display design to enhance focused and divided attention". Proceedings of the SIGCHI conference on Human factors in computing systems - CHI '95. New York, NY, USA: ACM Press/Addison-Wesley Publishing Co. pp. 317–324. doi:10.1145/223904.223945. ISBN 978-0201847055. S2CID 13361860.
  19. ^ a b Smith, Shana; Fu, Shih-Hang (1 April 2011). "The relationships between automobile head-up display presentation images and drivers' Kansei" (PDF). Displays. 32 (2): 58–68. doi:10.1016/j.displa.2010.12.001.
[edit]