patents.google.com

CN112470060A - Head-up display - Google Patents

  • ️Tue Mar 09 2021

CN112470060A - Head-up display - Google Patents

Head-up display Download PDF

Info

Publication number
CN112470060A
CN112470060A CN201980049245.4A CN201980049245A CN112470060A CN 112470060 A CN112470060 A CN 112470060A CN 201980049245 A CN201980049245 A CN 201980049245A CN 112470060 A CN112470060 A CN 112470060A Authority
CN
China
Prior art keywords
image
light
camera
display
driver
Prior art date
2018-07-23
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201980049245.4A
Other languages
Chinese (zh)
Inventor
小野长平
三泽昭央
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Maxell Ltd
Original Assignee
Maxell Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
2018-07-23
Filing date
2019-05-16
Publication date
2021-03-09
2019-05-16 Application filed by Maxell Ltd filed Critical Maxell Ltd
2021-03-09 Publication of CN112470060A publication Critical patent/CN112470060A/en
Status Pending legal-status Critical Current

Links

  • 230000003287 optical effect Effects 0.000 claims abstract description 53
  • 238000001514 detection method Methods 0.000 claims abstract description 31
  • 210000001508 eye Anatomy 0.000 claims abstract description 29
  • 230000002265 prevention Effects 0.000 claims description 11
  • 230000004313 glare Effects 0.000 claims description 9
  • 238000003384 imaging method Methods 0.000 claims description 8
  • 230000000903 blocking effect Effects 0.000 claims description 7
  • 230000007246 mechanism Effects 0.000 claims description 2
  • 238000010586 diagram Methods 0.000 description 24
  • 238000000034 method Methods 0.000 description 20
  • 230000008569 process Effects 0.000 description 18
  • 238000012545 processing Methods 0.000 description 12
  • 210000003128 head Anatomy 0.000 description 9
  • 238000004891 communication Methods 0.000 description 7
  • 238000012544 monitoring process Methods 0.000 description 5
  • 210000005252 bulbus oculi Anatomy 0.000 description 4
  • 238000012937 correction Methods 0.000 description 4
  • 230000000007 visual effect Effects 0.000 description 3
  • 230000006870 function Effects 0.000 description 2
  • 238000009434 installation Methods 0.000 description 2
  • 238000005259 measurement Methods 0.000 description 2
  • XAGFODPZIPBFFR-UHFFFAOYSA-N aluminium Chemical compound [Al] XAGFODPZIPBFFR-UHFFFAOYSA-N 0.000 description 1
  • 229910052782 aluminium Inorganic materials 0.000 description 1
  • 230000008859 change Effects 0.000 description 1
  • 239000003086 colorant Substances 0.000 description 1
  • 239000000428 dust Substances 0.000 description 1
  • 230000000694 effects Effects 0.000 description 1
  • 238000005286 illumination Methods 0.000 description 1
  • 230000006872 improvement Effects 0.000 description 1
  • 239000004973 liquid crystal related substance Substances 0.000 description 1
  • 238000012986 modification Methods 0.000 description 1
  • 230000004048 modification Effects 0.000 description 1
  • 238000005070 sampling Methods 0.000 description 1

Images

Classifications

    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K35/00Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K35/00Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
    • B60K35/10Input arrangements, i.e. from user to vehicle, associated with vehicle functions or specially adapted therefor
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K35/00Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
    • B60K35/20Output arrangements, i.e. from vehicle to user, associated with vehicle functions or specially adapted therefor
    • B60K35/21Output arrangements, i.e. from vehicle to user, associated with vehicle functions or specially adapted therefor using visual output, e.g. blinking lights or matrix displays
    • B60K35/22Display screens
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K35/00Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
    • B60K35/20Output arrangements, i.e. from vehicle to user, associated with vehicle functions or specially adapted therefor
    • B60K35/21Output arrangements, i.e. from vehicle to user, associated with vehicle functions or specially adapted therefor using visual output, e.g. blinking lights or matrix displays
    • B60K35/23Head-up displays [HUD]
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K35/00Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
    • B60K35/50Instruments characterised by their means of attachment to or integration in the vehicle
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/0093Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B5/00Optical elements other than lenses
    • G02B5/08Mirrors
    • G02B5/10Mirrors with curved faces
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B7/00Mountings, adjusting means, or light-tight connections, for optical elements
    • G02B7/18Mountings, adjusting means, or light-tight connections, for optical elements for prisms; for mirrors
    • G02B7/182Mountings, adjusting means, or light-tight connections, for optical elements for prisms; for mirrors for mirrors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/10Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
    • H04N23/11Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths for generating image signals from visible and infrared light wavelengths
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K2360/00Indexing scheme associated with groups B60K35/00 or B60K37/00 relating to details of instruments or dashboards
    • B60K2360/149Instrument input by detecting viewing direction not otherwise provided for
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K2360/00Indexing scheme associated with groups B60K35/00 or B60K37/00 relating to details of instruments or dashboards
    • B60K2360/20Optical features of instruments
    • B60K2360/21Optical features of instruments using cameras
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K2360/00Indexing scheme associated with groups B60K35/00 or B60K37/00 relating to details of instruments or dashboards
    • B60K2360/20Optical features of instruments
    • B60K2360/23Optical features of instruments using reflectors
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K35/00Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
    • B60K35/60Instruments characterised by their location or relative disposition in or on vehicles
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0118Head-up displays characterised by optical features comprising devices for improving the contrast of the display / brillance control visibility
    • G02B2027/012Head-up displays characterised by optical features comprising devices for improving the contrast of the display / brillance control visibility comprising devices for attenuating parasitic image effects
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0138Head-up displays characterised by optical features comprising image capture systems, e.g. camera
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/014Head-up displays characterised by optical features comprising information/image processing systems
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0179Display position adjusting means not related to the information to be displayed
    • G02B2027/0181Adaptation to the pilot/driver
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0179Display position adjusting means not related to the information to be displayed
    • G02B2027/0187Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Optics & Photonics (AREA)
  • Chemical & Material Sciences (AREA)
  • Transportation (AREA)
  • Mechanical Engineering (AREA)
  • Combustion & Propulsion (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Instrument Panels (AREA)

Abstract

抬头显示器具备影像显示装置、虚像光学系统、收容影像显示装置以及所述虚像光学系统的框体、框体开口部、覆盖框体开口部并透射影像光的防眩板、对包括交通工具的驾驶员的眼睛的脸部区域进行摄像的摄像机、根据摄像机摄像到的图像来检测驾驶员的视点的视点检测装置以及主控制装置。摄像机处于框体的外侧,且按照在防眩板中映入包括驾驶员的眼睛的脸部映入的区域进行摄像的朝向设置于避开影像光从框体开口部射出的光路的位置。主控制装置根据视点检测装置检测到的驾驶员的视点位置,对虚像光学系统进行调整影像光的出射方向的控制。

Figure 201980049245

The head-up display includes an image display device, a virtual image optical system, a frame that accommodates the image display device and the virtual image optical system, an opening of the frame, an anti-glare plate that covers the opening of the frame and transmits image light, and is useful for driving including vehicles. A camera that captures an image of the face area of the driver's eyes, a viewpoint detection device that detects the driver's viewpoint from an image captured by the camera, and a main control device. The camera is located on the outside of the housing, and is installed at a position avoiding the optical path of the video light emitted from the opening of the housing so as to capture the image in the anti-glare plate in the area where the face of the driver including the driver's eyes is projected. The main control device controls the virtual image optical system to adjust the output direction of the image light based on the position of the driver's viewpoint detected by the viewpoint detection device.

Figure 201980049245

Description

Head-up display

Technical Field

The present invention relates to a head-up display.

Background

For example, in a vehicle such as an automobile, information such as a vehicle speed and an engine speed is generally displayed on an Instrument panel (Instrument panel) in an Instrument panel. Further, a screen for car navigation or the like is embedded in the instrument panel or displayed on a display provided on the instrument panel. In order to visually recognize these information, the driver needs to move the line of sight greatly, and therefore, as a technique for reducing the amount of movement of the line of sight, a Head Up Display (hereinafter, sometimes referred to as "HUD") is known in which information such as vehicle speed and information such as instructions related to car navigation are projected onto a windshield, a combiner, or the like to be displayed.

As a technique related to HUD, patent document 1 discloses the following HUD: "A head-up display device includes: an eyeball position detection unit that detects an eyeball position of a user; a control unit that calculates an imaging position of a display image from the first eyeball position detected by the eyeball position detection unit; an optical path adjusting unit for adjusting the emitting direction of the visible light according to the calculation result of the control unit; and a storage unit that stores the first eye position as a reference eye position, wherein the control unit calculates an imaging position of the display image from the second eye position when it is determined that the second eye position detected by the eye position detection unit after the time when the first eye position is detected has a difference of a predetermined distance or a predetermined angle or more from the reference eye position, and drives the optical path adjustment unit (abstract) based on the calculation result.

Documents of the prior art

Patent document

Patent document 1: japanese laid-open patent publication No. 2008-155720

Disclosure of Invention

The viewpoint detecting camera is provided near the ceiling near the instrument panel and the windshield. However, when the viewpoint detection camera is disposed in the instrument panel, imaging is performed at a position looking down at the driver, and when the viewpoint detection camera is disposed near the ceiling, imaging is performed at a position looking down at the driver. The face image used for the viewpoint detection is preferably a front image of the face of the driver, but if the viewpoint detection camera is disposed in front of the driver, the view is blocked, which is not preferable.

In this regard, in patent document 1, it is not considered that the face image of the viewpoint detecting camera is brought close to the frontal image, and there is a room for further study on improvement of the accuracy of viewpoint detection.

The present invention has been made in view of the above circumstances, and an object thereof is to provide a HUD capable of controlling a virtual image display position with improved viewpoint detection accuracy.

In order to solve the above problem, the present invention has the structure described in the claim. For example, the head-up display is a head-up display for displaying a virtual image in front of a vehicle, and includes: an image display device including a light source and a display element, for displaying a display object on the display element and emitting image light including the display object; a virtual image optical system that enlarges and projects the image light toward the projection target member while changing an emission direction; a frame body which accommodates the image display device and the virtual image optical system; a frame opening portion provided in the frame so that the image light is emitted from the frame; an antiglare shield covering the frame opening and transmitting the image light; a camera that captures an image of a face area including eyes of a driver of the vehicle; a viewpoint detection device that detects a viewpoint of the driver from an image captured by the camera; and a main control device connected to the image display device, the virtual image optical system, and the viewpoint detection device, wherein the camera is provided outside the housing at a position avoiding an optical path through which the image light is emitted from the housing opening, in a direction in which an area of the antiglare shield in which a face including the eyes of the driver is projected is imaged, and the main control device controls the virtual image optical system to adjust the emission direction of the image light based on the viewpoint position of the driver detected by the viewpoint detection device.

According to the present invention, it is possible to provide a HUD capable of controlling the virtual image display position with improved viewpoint detection accuracy. Problems, structures, and effects other than those described above will become apparent from the following description of the embodiments.

Drawings

Fig. 1 is a schematic configuration diagram of a HUD.

Fig. 2 is an enlarged configuration diagram of the mirror driving unit.

Fig. 3 is a diagram showing the angle of view of the camera.

Fig. 4 is a system configuration diagram of the HUD.

Fig. 5 is a diagram showing a relationship between a viewpoint, a reflection position of the image light L on the windshield, and a virtual image position.

Fig. 6 is a flowchart showing the flow of processing performed with the HUD.

Fig. 7 is a diagram showing an image captured in a state where illuminance is different and a histogram showing a distribution of luminance values of each image.

Fig. 8 is a view showing a state in which the concave mirror is rotated at the light shielding angle.

Fig. 9 is a diagram showing a 3-dimensional spatial coordinate system of the vehicle interior.

Fig. 10A is an explanatory diagram of the amount of viewpoint movement accompanying the movement of the head of the driver.

Fig. 10B is an explanatory diagram of the viewpoint movement amount according to the movement of the head of the driver.

Fig. 10C is an explanatory diagram of the viewpoint movement amount according to the movement of the head of the driver.

Fig. 11 is a schematic diagram showing the installation position of the shutter in the HUD.

Fig. 12 is an enlarged explanatory view of the shutter member.

Fig. 13 is a diagram showing a state where the shutter is in the light shielding position.

Fig. 14 is a diagram showing a state where the shutter is in the escape position.

Fig. 15 is a view showing another embodiment in which a reflection member is provided to the dazzle prevention plate.

Description of the symbols

1: HUD; 2: a vehicle; 3: a windshield; 4: an instrument panel; 5: a driver; 7: an instrument panel opening; 10: a light-shielding cover; 20: a HUD control device; 30: an image display device; 40: a virtual image optical system; 43: a concave mirror; 44: a mirror driving section; 50: a frame body; 51: a frame opening; 52: an antiglare shield; 70: a camera; 100: a virtual image; 700: a viewpoint detecting device.

Detailed Description

Hereinafter, embodiments of the present invention will be described in detail with reference to the drawings. In all the drawings for explaining the embodiments, the same members are denoted by the same reference numerals in principle, and redundant explanations thereof are omitted. In each of the embodiments described below, a case where a Head Up Display (HUD) is provided in a vehicle as a vehicle will be described as an example, but the embodiments can also be applied to other vehicles such as an electric train.

Referring to fig. 1, the configuration of the HUD1 according to the present embodiment will be described. Fig. 1 is a schematic configuration diagram of the HUD 1. In the HUD1 of fig. 1, a

concave mirror

43 included in the virtual image optical system and a

mirror driving unit

44 for rotating the

concave mirror

43 are used as a light blocking member for opening and closing the optical path of the image light L.

As shown in fig. 1, the HUD1 is provided in the

dashboard

4 of the

vehicle

2. The

instrument panel

4 includes an instrument panel opening 7 through which the image light L emitted from the HUD1 passes. The image light L emitted from the HUD1 reaches the

windshield

3 as a projection target member through the instrument panel opening 7, is reflected by the

windshield

3, and enters the eyes of the

driver

5. The

driver

5 visually recognizes the

virtual image

100 based on the image light L at a position forward of the

windshield

3. The member to be projected is not limited to the

windshield

3, and other members such as a combiner may be used as long as the member to be projected with the image light L is used.

The HUD1 includes a

housing

50, a HUD control device 20 (corresponding to a main control device) and an

image display device

30 housed or attached to the

housing

50, and a virtual image

optical system

40 configured to enlarge and project image light L emitted from the

image display device

30.

A housing opening 51, which is an exit port for the image light L, is formed in the upper surface of the

housing

50. The housing opening 51 is covered with an antiglare shield (glare trap) 52 for preventing dust and dirt from entering the

housing

50. The

dazzle prevention plate

52 is formed of a member that transmits visible light.

The

image Display device

30 is configured using an LCD (Liquid Crystal Display), and operates in accordance with a control signal from the

HUD control device

20. More specifically, the

image display device

30 includes a

light source

31, an illumination

optical system

32, and a

transmissive display element

33 that displays a display object. Light emitted from the

light source

31 passes through the

display element

33, and image light L (see fig. 4) including a display object is emitted from the

image display device

30.

The virtual image

optical system

40 is configured such that the

lens member

41, the

folding mirror

42, and the

concave mirror

43 are arranged in this order from a position close to the

image display device

30 along the emission direction of the image light L. Further, the virtual image

optical system

40 includes a

mirror driving unit

44 configured to rotate the

concave mirror

43.

The

lens unit

41 is a unit of at least one lens for adjusting an optical distance from the

image display device

30 to the

folding mirror

42.

The reflecting

mirror

42 is a mirror that reflects the optical path of the image light L emitted from the

lens member

41 toward the

concave mirror

43.

The

concave mirror

43 reflects the image light L reflected by the reflecting

mirror

42 toward the housing opening 51. The

concave mirror

43 is rotated by a

mirror driving unit

44.

Fig. 2 is an enlarged configuration diagram of the

mirror driving unit

44.

The

mirror driving unit

44 includes: a

mirror motor

44a and a

rotary shaft

44b of the

mirror motor

44 a; a

mirror cover

45 connected to the front end of the

rotating shaft

44b and accommodating the

concave mirror

43; the

mirror holder

46 is a bearing member for the

rotary shaft

44 b.

As shown in fig. 1, the

mirror motor

44a is connected to the

HUD control device

20 via a

bus

61. The

mirror motor

44a is driven in accordance with a control signal from the

HUD control device

20. When the

mirror motor

44a is driven, the

rotation shaft

44b rotates, and the

mirror cover

45 coupled to the

rotation shaft

44b rotates. As a result, the

concave mirror

43 rotates integrally with the

mirror cover

45.

Since the

concave mirror

43 rotates, the image light L can be reflected toward the

windshield

3 with its reflection angle changed. When the reflection angle of the image light L is changed, the reflection position of the image light L at the

windshield

3 is changed, and the optical path length of the image light L can be changed. Since the distance of the virtual image plane displaying the

virtual image

100 changes due to the change in the optical path length, and in addition, the reflection position changes, the display height of the virtual image plane changes.

A

light shielding cover

10 for shielding direct sunlight from entering the HUD1 is provided on the windshield-side end portion of the

instrument panel

4 at the instrument panel opening 7. At the

shade cover

10, the

camera

70 is disposed in a position and orientation that includes the

dazzle prevention plate

52 in the angle of view. The

camera

70 is connected to the

HUD control device

20 by a

communication line

62, and an image captured by the

camera

70 is output to a viewpoint detection device 700 (see fig. 4) described later. When the

HUD control device

20 has the function of the

viewpoint detecting device

700, an image is output to the

HUD control device

20.

The image of the

camera

70 is used for both the viewpoint detection processing of the

driver

5 and the incident amount monitoring processing of the direct sunlight incident on the HUD 1. Therefore, both the eyes of the

driver

5 and the

glare shield

52 need to be imaged in the image. Therefore, in the present embodiment, as shown in fig. 3, the lens of the

camera

70 is directed toward the

glare shield

52, and the light reflected by the surface of the face including the eyes of the

driver

5 is reflected by the

windshield

3, reaches the

glare shield

52, is reflected by the surface of the

glare shield

52, and enters the attachment position of the lens of the

camera

70 and the orientation (angle of view) of the lens, and is provided in the

shade cover

10. The attachment position of the

camera

70 to the

dazzle prevention plate

52 and the orientation of the lens are one of the features of the HUD1 of the present embodiment.

The

camera

70 is equipped with a near-infrared

light source

70a such as an LED that emits near-infrared light and a CMOS70b as an imaging element that detects near-infrared light to visible light. In the viewpoint detection processing, near-infrared light is used. That is, the near-infrared light generated from the near-infrared

light source

70a mounted on the

camera

70 reaches the face surface (face area) including the eyes of the

driver

5 and is reflected, and the reflected near-infrared light reaches the

antiglare plate

52 and is reflected again toward the CMOS70b, and enters the CMOS70 b. The

camera

70 generates and outputs a face image of the

driver

5 based on the near-infrared light image from the output from the CMOS70 b. This enables the

driver

5 to detect the viewpoint even when the outside is dark, for example, at night, in a tunnel, or in an underground parking lot.

On the other hand, visible light is used for monitoring the amount of incident sunlight. The visible light reaching the

dazzle prevention plate

52 is incident from the

dazzle prevention plate

52 to the CMOS70 b. The

camera

70 generates and outputs an outer surface image of the

dazzle prevention plate

52 based on a visible light image from the output from the CMOS70 b.

Fig. 3 is a diagram showing the angle of view of the

camera

70.

Generally, the line of sight direction of the

driver

5 who visually recognizes the

virtual image

100 has a depression angle (for example, 1) with respect to the horizontal direction°~5°). Therefore, when the

driver

5 is imaged while matching the optical axis of the

camera

70 with the optical path of the image light L, the

driver

5 is imaged in a head-down state, and the accuracy of detecting the coordinates of the eyes is lower than in the case where the line of sight is directed in the horizontal direction.

Therefore, in the present embodiment, as shown in fig. 3, the

camera

70 is provided so as to match the orientation and position of the optical axis of the

camera

70 in such a manner that the surface of the face including the eyes is imaged at an angle close to the horizontal (an angle closer to the horizontal than the case where the eyes are imaged along the optical axis of the image light L). For example, the

camera

70 is set to an orientation in which the optical axis of the

camera

70 is incident toward the center of the eyepoint of the

driver

5 at an angle within an allowable range that can be regarded as horizontal with respect to the horizontal axis. This makes it possible to capture the eyes of the

driver

5 in a direction close to the horizontal direction and output an image close to the frontal image. Therefore, the accuracy of the viewpoint detection processing in step S20, which will be described later, can be expected to be improved.

Fig. 4 is a system configuration diagram of the HUD 1.

The

HUD Control device

20 includes a 1 st ECU (electronic Control Unit) 21, a 1 st

nonvolatile memory

22 connected to the 1 st ECU via a system bus, a 1

st memory

23, a light

source adjustment Unit

24, a

distortion correction Unit

25, a display

element drive Unit

26, a 1 st CAN communication Unit 27 (CAN: Controller Area Network), a mirror adjustment Unit 28, and a

timer

29. The 1 st CAN

communication unit

27 is connected to the

viewpoint detecting device

700 via CAN. The light

source adjustment unit

24 is connected to the

light source

31, the display

element driving unit

26 is connected to the

display element

33, and the mirror adjustment unit 28 is connected to the

mirror motor

44 a. The 1 st ECU21 acquires the vehicle information 9 such as speed and engine speed from the 1 st CAN

communication unit

27 via the CAN of the

vehicle

2, and displays the display object including the vehicle information 9 as a

virtual image

100.

The

viewpoint detecting apparatus

700 includes a 2 nd ECU701, a

camera control unit

702, a 2 nd CAN

communication unit

703, a 2 nd

nonvolatile memory

704, and a 2 nd memory 705. The input stage of the 2 nd ECU701 is connected to the output stage of the

camera

70, and the image of the

camera

70 is transmitted from the 2 nd CAN

communication unit

703 to the

HUD control device

20 via the 2

nd ECU

701. Further, the 2 nd ECU701 is connected to the

camera control unit

702, the 2 nd

nonvolatile memory

704, and the 2 nd memory 705, respectively.

Next, adjustment of the display position of the

virtual image

100 in the HUD1 having the configuration described above will be described with reference to fig. 5. Fig. 5 is a diagram showing a relationship between the viewpoint, the reflection position of the image light L on the

windshield

3, and the virtual image position.

In the HUD1, the following control is performed: the

viewpoint detecting device

700 detects the viewpoint using the image of the

camera

70, and adjusts the display position of the

virtual image

100 projected onto the

windshield

3 of the

vehicle

2 by controlling the rotation angle (inclination angle) of the

concave mirror

43 in accordance with the detected viewpoint of the

driver

5. In fig. 5, the rotational (tilting) position of the

concave mirror

43, the projection position of the

virtual image

100 on the

windshield

3, and the position of the

virtual image

100 as viewed from the viewpoint of the

driver

5 are indicated by A, B, C, respectively, corresponding to the position A, B, C of the viewpoint of the

driver

5.

The above control is realized, for example, by: the 2 nd CAN

communication unit

703 of the

viewpoint detecting apparatus

700 notifies the

HUD control apparatus

20 of the viewpoint information, and the

HUD control apparatus

20 rotates the

concave mirror

43. In this

viewpoint detecting apparatus

700, acquisition of information from the

camera

70 is performed under the control of the 2

nd ECU

701. In the

HUD control device

20, the acquisition of the vehicle information 9, the control of the

video display device

30, and the driving of the

concave mirror

43 are performed under the control of the 1

st ECU

21. However, the present invention is not necessarily limited to these examples, and other types of devices may be provided.

Fig. 6 is a flowchart showing the flow of processing performed by the HUD 1.

The HUD1 is also characterized by using the image of the

camera

70 in combination with the process of adjusting the virtual image position in accordance with the viewpoint of the

driver

5 and the process of monitoring the amount of sunlight entering the HUD 1.

When the power of the HUD1 is turned on (S10/yes), the

camera

70 starts imaging (S11), and outputs the imaged image to the

viewpoint detecting device

700. This image reflects the

emission region

52a (see fig. 15) of the image light L in the

glare shield

52, and includes, as an object, the entire face including the eyes of the

driver

5 or a part of the face reflected on the

emission region

52a of the image light L. When the near-infrared light reflected by the surface of the face of the

driver

5 reaches a position different from the

emission region

52a in the

dazzle prevention plate

52, the region where the near-infrared light reaches is included in the angle of view of the

camera

70, and the face image is captured.

The 2 nd ECU701 of the visual

point detection apparatus

700 acquires an image from the

camera

70 and executes the visual point detection process (S20). The details of which will be described later.

The 2 nd ECU701 calculates a viewpoint movement amount (S21) which is a difference between the viewpoint position obtained in step S20 and the viewpoint position used when the display position of the

virtual image

100 is determined from the current HUD1, and transmits the viewpoint movement amount to the

HUD control device

20. The details of which will be described later.

In parallel with the viewpoint detection process (S20), the 2 nd ECU701 calculates the luminance value of the image of the camera 70 (S30). In the present embodiment, the case where the 2 nd ECU701 calculates the luminance value of the entire image is described as an example with reference to fig. 7, but the luminance value of a partial region such as the central region of the image may be detected instead of the entire image. Fig. 7 is a diagram showing an image captured in a state where illuminance is different and a histogram showing a distribution of luminance values of each image.

In the 2

nd image

212 captured at a standard brightness, for example, a brightness at which the

driver

5 can drive without lighting the headlights without excessively dazzling, the peak value of the 2

nd histogram

222 indicating the distribution of the brightness values of the image is not large, and the brightness values are dispersed. The area of the 2 nd histogram 222 (a portion to which colors are added in the figure) indicates the total number of pixels of the image, and the areas of the 1

st histogram

221 and the 3

rd histogram

223, which will be described later, are the same as the area of the 2

nd histogram

222.

On the other hand, the 1

st histogram

221 showing the distribution of the luminance values of the 1

st image

211 captured in a dark place has a large number of dark luminance values. The 3

rd histogram

223 showing the distribution of the luminance values of the 3

rd image

213 captured in a location where the luminance is too bright, that is, a location where the amount of solar light irradiated to the

antiglare plate

52 is large, has a large number of pixels of bright luminance values.

Therefore, the 2 nd ECU701 obtains the luminance value of each image as an integrated value of (luminance value × number of degrees), and outputs the luminance value to the

HUD control device

20.

When the 1 st ECU21 of the

HUD control device

20 determines that the brightness value of the image obtained in step S30 is equal to or greater than the predetermined light-shielding threshold value (S40/yes), the

timer

29 is used to start measurement of the elapsed time t from when the brightness value of the image becomes equal to or greater than the light-shielding threshold value (S41). When the elapsed time t is equal to or greater than the predetermined time threshold (S42/yes), the mirror adjustment unit 28 rotates the

concave mirror

43 so that the sunlight does not reach the image display device 30 (hereinafter referred to as "light shielding angle") (S43). At this point in time, the counting by the

timer

29 may be ended.

Fig. 8 is a diagram showing a state in which the

concave mirror

43 is rotated to the light shielding angle.

A part of the sunlight Sb that has reached the

antiglare plate

52 is reflected and enters the

camera

70, and a part thereof passes through the

antiglare plate

52 and reaches the

concave mirror

43. When the

concave mirror

43 is rotated at the light shielding angle, the sunlight Sb that has reached the

concave mirror

43 is reflected by the mirror surface of the

concave mirror

43 and is reflected in the lower direction of the

housing

50, that is, in a direction different from the direction toward the

folding mirror

42.

Accordingly, the image light L is not irradiated toward the

antiglare plate

52, and therefore the display of the

virtual image

100 disappears, and the optical path through which the sunlight Sb incident from the

antiglare plate

52 reaches the

image display device

30 via the

concave mirror

43 and the

folding mirror

42 can also be cut off. Further, the sunlight Sb incident from the

antiglare plate

52 can be prevented from being reflected by the

concave mirror

43 and re-irradiated toward the

antiglare plate

52 and irradiated to the

windshield

3.

On the other hand, in the case where the luminance value of the image is smaller than the light-shielding threshold (S40/no), the 1 st ECU21 causes the mirror adjustment section 28 to return (restore) the mirror surface of the

concave mirror

43 to the optical path of the video light L (S45) during the rotation of the

concave mirror

43 by the light-shielding angle, that is, in the case where the luminance value of the image is already equal to or greater than the light-shielding threshold and the mirror surface of the

concave mirror

43 is made to avoid the optical path of the video light L (S44/yes).

After step S45 is completed, or when the

concave mirror

43 is not rotating at the shading angle (S44/no), that is, when the mirror surface of the

concave mirror

43 is on the optical path of the video light L, or when the elapsed time t is less than the time threshold (S42/no), the 1 st ECU21 compares whether or not the movement amount of the viewpoint is equal to or greater than a predetermined movement amount threshold (S46).

When the movement amount of the viewpoint is equal to or greater than the movement amount threshold (S46/yes), 1 st ECU21 performs mirror adjustment processing for matching the virtual image position with the viewpoint with respect to mirror adjustment unit 28 (S47). More specifically, a signal for rotating the

concave mirror

43 in accordance with the viewpoint movement amount calculated in S21 is output from the mirror adjustment unit 28 to the

mirror motor

44 a. This allows the virtual image to be displayed in accordance with the position of the viewpoint detected by

viewpoint detecting apparatus

700.

After the mirror adjustment process (S47) or when the movement amount of the viewpoint is smaller than the movement amount threshold (S46/no), the 1 st ECU21 performs distortion correction of the display object displayed as the

virtual image

100 in the

distortion correction unit

25, and the display

element driving unit

26 displays the display object after the distortion correction on the

display element

33. At the same time, ECU 1 21 causes light

source adjustment unit

24 to execute the lighting process of

light source

31 and causes display

element

33 to emit video light L including the display object. The image light L is enlarged in the virtual image

optical system

40, projected to the

windshield

3, and incident to the eyes of the

driver

5, thereby displaying the virtual image 100 (S48). If the power of the HUD1 is not off (S49/NO), the process returns to step S11, and the subsequent processes are repeated. When the power of the HUD1 becomes off (S49/yes), the virtual image display processing by the HUD1 is ended.

< S20: viewpoint detection processing >

Fig. 9 is a diagram showing a 3-dimensional spatial coordinate system inside the

vehicle

2.

Inside the

vehicle

2, the

camera

70 is fixedly disposed at a position where the face (viewpoint) of the

driver

5 can be imaged, and inside the

light shielding cover

10 in the present embodiment (see fig. 3). At this time, the

camera

70 is disposed at the center in the 3-dimensional space coordinates inside the

vehicle

2, and the x-axis is set to the longitudinal direction of the vehicle 2 (the traveling direction of the vehicle), the y-axis is set to the width direction of the vehicle, and the z-axis is set to the height (vertical) direction. Fig. 9 shows a state when this state is viewed from above. The

camera

70 captures an image including the face (viewpoint) of the

driver

5 from the front of the

driver

5 in a 3-dimensional space inside the vehicle centered thereon. In fig. 9,

reference numeral

71 denotes an imaging range of the

camera

70.

The 2 nd ECU701 detects the eyes (viewpoint) from the 2-dimensional camera image captured by the

camera

70. Next, the 2 nd ECU701 obtains a straight line from the camera 70 (the origin of the 3-dimensional space) toward the detected eye (viewpoint) 5 of the driver 5 (i.e., connects the

camera

70 and the viewpoint). This makes it possible to specify a straight line in the direction toward the eyes (viewpoint) of the

driver

5 when viewed from the

camera

70, i.e., the origin of the 3-dimensional space. In other words, when the 1 point (the position of the eye of the

driver

5 in the above description) on the camera image (2 d) is specified from the relationship between the installation direction of the

camera

70 and the angle of view, the straight line on the 3 d space connecting the

camera

70 and the 1 point on the image can be obtained. However, it is not possible to define where the eye is located on a straight line.

Next, an intersection point between the straight line in 3 dimensions obtained in the above and a vertical plane (z-x plane) in 3 dimensions passing through the center of the seat (or the central axis of the steering wheel 8) is obtained, and the obtained intersection point is a viewpoint position of the eyes of the

driver

5. A vertical plane (z-x plane) in 3 dimensions passing through the center of the seat (center axis of the steering wheel) is a z-x plane that is distant from the camera 70 (origin) to the center of the seat in the width direction (y axis) of the

vehicle

2, and this distance can be set in advance when the

camera

70 is disposed inside the

vehicle

2 or can be determined by measurement.

That is, the 2 nd ECU701 obtains the coordinates (x) of the viewpoint of the

driver

5 from the face of the

driver

5 captured by the

camera

70 by the viewpoint detection processing described aboved,yd,zd) And stored in the 2 nd

nonvolatile memory

704. Then, the images of the

camera

70 captured at the down-sampling cycle as described above are used for the viewpoint moving amount arithmetic processing (S21).

< S21: viewpoint movement amount calculation processing >

The

driver

5 may tilt forward or move his head, and the viewpoint position may be displaced. Therefore, the 2 nd ECU701 estimates the orientation of the face of the

driver

5 from the image captured by the

camera

70, determines whether the face of the

driver

5 is located at the center of the seat from the estimated orientation of the face, and performs position adjustment of the

concave mirror

43. Fig. 10A, 10B, and 10C are explanatory views of the amount of viewpoint movement accompanying the movement of the head of the driver.

As shown in fig. 10A, the orientation of the face of the driver 5 (the 3 directions of roll, pitch, and yaw, as indicated by the arrows in the figure) is estimated from the image captured by the

camera

70, and when it is determined that the face is oriented toward the front of the vehicle, the viewpoint coordinates (x) of the

driver

5 obtained as described above are usedd,yd,zd) The display position adjustment of the

virtual image

100 based on the angle adjustment of the

concave mirror

43 is performed. As shown in fig. 10B and 10C, when the face of the

driver

5 is oriented away from the front of the vehicle, the display position of the

virtual image

100 is not adjusted by the angle adjustment of the

concave mirror

43. In the present embodiment, particularly, in order to suppress the display position of the

virtual image

100 from frequently matching the fine movement of the head of the driver 5Instead, the 2 nd ECU701 obtains the previous viewpoint detection position (x)dt,ydt,zdt) The viewpoint detecting position (x) at this timedt+1,ydt+1,zdt+1) difference (Δ xd,Δyd,Δzd) When Δ xd、Δyd、ΔzdIs a movement amount threshold value Δ x predetermined to determine that the head of the

driver

5 is moveddth、Δydth、ΔzdthWhen the above, will contain (Δ x)d,Δyd,Δzd) The viewpoint moving amount information of (2) is outputted to the

HUD control device

20.

According to the present embodiment, since the viewpoint of the

driver

5 is detected and the display position of the

virtual image

100 is adjusted in accordance with the detection result, the visibility of the

virtual image

100 can be improved.

Further, the

camera

70 for viewpoint detection is provided in the visual field generated by the

shade cover

10 when viewed from the

driver

5, so that the aesthetic appearance in the

vehicle

2 is improved.

Further, since the

camera

70 captures the face of the

driver

5 reflected on the

antiglare plate

52, the face image can be captured substantially from the front, and the viewpoint detection can be performed with high accuracy.

The

camera

70 can function as a monitoring camera for the amount of sunlight incident on the HUD1, and the number of components can be reduced as compared with a case where a monitoring member is separately prepared.

Further, since the

camera

70 is disposed at a position avoiding the optical path of the video light L of the HUD1, even after the incident amount of sunlight becomes too large and the optical path of the video light L is shielded, the incident amount of sunlight can be monitored, and when the incident amount is smaller than the light shielding threshold, the optical path of the video light L can be restored again.

The present invention is not limited to the above embodiments, and various modifications within a range not departing from the gist of the present invention belong to the technical scope of the present invention. For example, the calculation formula used in the description of each process is only one embodiment of the process, and other calculation formulas that bring about calculation results necessary for the process may be applied.

In addition, as the structure for blocking the optical path of the image light L in the above description, a structure is used in which the mirror surface of the

concave mirror

43 is made to avoid the optical path of the image light L (the mirror surface of the

concave mirror

43 is rotated at a blocking angle from the optical path of the image light L), but the structure is not limited thereto. Fig. 11 to 14 show a shutter structure as a light blocking structure for the image light L.

Fig. 11 is a schematic diagram showing the mounting position of the

shutter

306 in the HUD1 a. Fig. 12 is an enlarged explanatory view of the

shutter member

300. Fig. 13 is a diagram showing a state where the

shutter

306 is in the light shielding position. Fig. 14 is a diagram showing a state where the

shutter

306 is at the escape position.

The HUD1a shown in fig. 11 includes a

shutter member

300 for disposing or avoiding the

shutter

306 in front of the exit surface (surface on the

lens member

41 side) of the image light L in the

image display device

30. The

shutter member

300 is connected to the

HUD control device

20 by a

bus

61. The

HUD control device

20 controls the driving of the

shutter member

300 so that the

shutter

306 is positioned at the light shielding position (see fig. 13) in step S43 and the

shutter

306 is positioned at the escape position (see fig. 14) in step S45.

As shown in fig. 12, the

shutter member

300 includes a

shutter

306 and a mechanism for moving the

shutter

306, specifically, a

shutter motor

301, a

gear train

302, a

lead screw

303, a

shutter holder

304, a

lead screw holder

305, and a

slide guide

307. The

shutter motor

301 is communicatively connected to the

HUD control device

20 via the

bus

61. The

lead screw holder

305 and the

slide guide

307 are attached to the

frame

50.

The

shutter motor

301 has a rotating shaft coupled to the

gear train

302, and one end of the

lead screw

303 is coupled to the

gear train

302.

The other end of the

lead screw

303 is held by a

lead screw holder

305. A thread is cut on the surface of the

lead screw

303.

Shutter retainer

304 retains

shutter

306. At a position opposite to the

lead screw

303 in the

shutter holder

304, a screw groove is cut. The thread of the

lead screw

303 engages with the thread groove of the

shutter holder

304.

The

slide guide

307 is provided opposite to the

lead screw

303 with the

shutter

306 interposed therebetween. A groove shape is provided in the

slide guide

307 on the surface opposite to the

shutter holder

304. The end of the

shutter holder

304 facing the

slide guide

307 is inserted into the groove of the

slide guide

307, and the

shutter holder

304 and the

shutter

306 are prevented from wobbling when the

shutter

306 moves from the off position to the light blocking position or from the light blocking position to the off position, and can move stably.

When a forward rotation signal or a reverse rotation signal is sent from the

HUD control device

20 to the

shutter motor

301, the

shutter motor

301 rotates forward or in reverse in accordance therewith. When the rotation shaft of the

shutter motor

301 rotates in the forward direction or in the reverse direction, the

lead screw

303 rotates in the forward direction or in the reverse direction via the

gear train

302. Then, the

shutter holder

304 moves from the escape position to the light shielding position or from the light shielding position to the escape position along the thread of the

lead screw

303. The

shutter

306 moves from the escape position to the light-shielding position or from the light-shielding position to the escape position integrally with the

shutter holder

304. In this example, the

shutter member

300 is disposed in front of the

image display device

30, but the

shutter member

300 may be provided on the optical path of the image light L in the HUD1a, and is not limited to the front of the

image display device

30.

In the above description, although the mirror adjustment unit 28 has been described as driving the

mirror driving unit

44 in step S45, the process may be performed only by storing a flag indicating that the mirror surface is rotated at the shading angle. Further, when the virtual image position is controlled in step S47, the

concave mirror

43 may be restored at the same time. This can reduce the number of times of driving the

concave mirror

43.

In the above, when the virtual image display is performed in the immediately preceding process (S48) without turning off the virtual image display after rotating the

concave mirror

43 at the light-shielding angle in step S43, the

concave mirror

43 is rotated at the light-shielding angle while the virtual image display process is kept being performed. However, since the generation of the image light L after the

concave mirror

43 is rotated at the light-shielding angle is unnecessary regardless of the display of the

virtual image

100, a process of turning off the virtual image display, for example, a process of turning off the

light source

31 or a process of making the display object non-display in the

display element

33 may be added after step S43. In this case, when the virtual image is displayed in the next step S48, the process of turning on the

light source

31 or displaying the display object on the

display element

33 is started again.

Fig. 15 is a view showing another embodiment in which the

reflection member

52b is provided in the

dazzle prevention plate

52. In fig. 14, a

reflection member

52b that reflects natural light (including near infrared light to visible light) is provided in a region of the

antiglare plate

52 that is different from the

emission region

52a of the image light L. Then, the reflecting

member

52b reflects the natural light, which is reflected by the face surface of the

driver

5 and reaches the

dazzle prevention plate

52 via the

windshield

3, toward the

camera

70. The reflecting

member

52b may be an aluminum plate or a mirror, for example. This enables the

camera

70 to more clearly capture the face image of the

driver

5.

Claims (8)

1. A head-up display that displays a virtual image in front of a vehicle, the head-up display comprising:

an image display device including a light source and a display element, for displaying a display object on the display element and emitting image light including the display object;

a virtual image optical system that enlarges and projects the image light toward the projection target member while changing an emission direction;

a frame body which accommodates the image display device and the virtual image optical system;

a frame opening portion provided in the frame so that the image light is emitted from the frame;

an antiglare shield covering the frame opening and transmitting the image light;

a camera that captures an image of a face area including eyes of a driver of the vehicle;

a viewpoint detection device that detects a viewpoint of the driver from an image captured by the camera; and

a main control device connected to the image display device, the virtual image optical system, and the viewpoint detection device,

the camera is provided outside the housing at a position avoiding an optical path of the image light emitted from the housing opening in a direction of imaging a region of the antiglare plate in which a face including the eyes of the driver is projected,

the main control device controls the virtual image optical system to adjust the emitting direction of the image light according to the viewpoint position of the driver detected by the viewpoint detection device.

2. The head-up display of claim 1,

the camera includes an image pickup element that detects near-infrared light,

an image including a near-infrared light image is output.

3. The head-up display of claim 1,

the camera includes an image pickup element that detects near infrared light to visible light,

the glare shield further includes a reflecting member that reflects natural light that is reflected by the surface of the driver's face and reaches the glare shield via the projected member, toward the camera, in a region of the glare shield that is different from the region in which the image light is emitted,

the camera detects the natural light reflected from the antiglare plate and takes a picture of the image.

4. The head-up display of claim 1,

the camera is disposed in such a direction that an optical axis of the camera is incident toward the center of the driver's eyepoint at an angle within an allowable range that can be regarded as horizontal with respect to a horizontal axis.

5. The head-up display of claim 1,

the head-up display further includes a light shielding member that opens and closes an optical path of the image light from the dazzle prevention plate toward the projected member,

the main control device is also connected to the shade member,

the camera is arranged outside the frame body according to a direction for shooting a region of at least a partial region of the emitting region of the image light in the anti-dazzle plate,

the main control device controls the light shielding member to close the optical path of the image light when the brightness value of the image captured by the camera is equal to or greater than a predetermined light shielding threshold value.

6. The heads-up display of claim 5,

the main control device controls the light shielding member to open the optical path of the image light when a brightness value of an image newly captured by the camera after the light shielding member is controlled to close the optical path of the image light is smaller than the light shielding threshold.

7. The heads-up display of claim 5,

the virtual image optical system includes a concave mirror that reflects the image light toward the projection member and a mirror driving unit that changes an orientation of a mirror surface of the concave mirror,

the main control device outputs a control signal for rotating the concave mirror to the mirror driving unit based on the viewpoint position of the driver detected by the viewpoint detecting device, and outputs a control signal for rotating the mirror surface of the concave mirror at an angle avoiding the optical path of the image light to the mirror driving unit when the luminance value of the image captured by the camera is equal to or greater than a predetermined light shielding threshold value.

8. The heads-up display of claim 5,

the light-shielding member is a shutter member,

the shutter member includes a shutter and a moving mechanism that moves the shutter to a retracted position and a light-shielding position,

the avoiding position is a position where the shutter is disposed so as to be offset from the optical path of the image light, and the light blocking position is a position where the shutter is disposed on the optical path of the image light.

CN201980049245.4A 2018-07-23 2019-05-16 Head-up display Pending CN112470060A (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
JP2018137944 2018-07-23
JP2018-137944 2018-07-23
PCT/JP2019/019533 WO2020021823A1 (en) 2018-07-23 2019-05-16 Head-up display

Publications (1)

Publication Number Publication Date
CN112470060A true CN112470060A (en) 2021-03-09

Family

ID=69180624

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201980049245.4A Pending CN112470060A (en) 2018-07-23 2019-05-16 Head-up display

Country Status (4)

Country Link
US (1) US11280999B2 (en)
JP (1) JP7024088B2 (en)
CN (1) CN112470060A (en)
WO (1) WO2020021823A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114594607A (en) * 2022-03-23 2022-06-07 业成科技(成都)有限公司 Optical film, preparation method thereof, head-up display, and vehicle

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7504387B2 (en) * 2020-03-31 2024-06-24 パナソニックオートモーティブシステムズ株式会社 Imaging and display system
CN112564616A (en) * 2020-11-06 2021-03-26 江苏泽景汽车电子股份有限公司 HUD's LCD screen prevents hot protection device of sunlight width of cloth
JP7521459B2 (en) * 2021-03-05 2024-07-24 株式会社デンソー Display device
JP7647223B2 (en) 2021-03-25 2025-03-18 日本精機株式会社 Instrument panel, vehicle display device, and bezel body
DE102021113893A1 (en) * 2021-05-28 2022-12-01 Bayerische Motoren Werke Aktiengesellschaft Monitoring system for a vehicle
JP7106022B1 (en) * 2021-06-16 2022-07-25 三菱電機株式会社 Aerial image display device and aerial image display method
JP7421535B2 (en) * 2021-12-14 2024-01-24 矢崎総業株式会社 heads up display device
CN116088175A (en) * 2022-11-30 2023-05-09 业成科技(成都)有限公司 Display system and vehicle head-up display system thereof
WO2024189893A1 (en) * 2023-03-16 2024-09-19 日産自動車株式会社 Driver monitoring device and automobile provided with same

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU2003214595A1 (en) * 2002-03-07 2003-09-16 Yechezkal Evan Spero Enhanced vision for driving
JP2008155720A (en) 2006-12-22 2008-07-10 Nippon Seiki Co Ltd Head-up display device
JP2011133508A (en) 2009-12-22 2011-07-07 Topcon Corp Scanned type display-device optical system, three-dimensional display device and head-up display device
JP6004706B2 (en) * 2012-04-04 2016-10-12 三菱電機株式会社 Display device and head-up display system provided with the same
JP2015152746A (en) 2014-02-14 2015-08-24 日本精機株式会社 display device
JP6540988B2 (en) * 2014-06-09 2019-07-10 日本精機株式会社 Head-up display device
WO2016067574A1 (en) 2014-10-29 2016-05-06 パナソニックIpマネジメント株式会社 Display control device and display control program
US20200159013A1 (en) 2016-08-08 2020-05-21 Maxell, Ltd. Head up display apparatus
JP2018097252A (en) 2016-12-15 2018-06-21 株式会社Jvcケンウッド Head-up display

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114594607A (en) * 2022-03-23 2022-06-07 业成科技(成都)有限公司 Optical film, preparation method thereof, head-up display, and vehicle

Also Published As

Publication number Publication date
JPWO2020021823A1 (en) 2021-07-01
US11280999B2 (en) 2022-03-22
JP7024088B2 (en) 2022-02-22
WO2020021823A1 (en) 2020-01-30
US20210271076A1 (en) 2021-09-02

Similar Documents

Publication Publication Date Title
JP7024088B2 (en) 2022-02-22 Head-up display
JP2005184225A (en) 2005-07-07 Vehicular display
CN106029416B (en) 2018-03-30 Sunshading board
US9638919B1 (en) 2017-05-02 Method for reducing reflection when operating a head-up display of a motor vehicle
JP2008242134A (en) 2008-10-09 Display device
US20200233216A1 (en) 2020-07-23 Vehicle arrangement, vehicle and method for providing a display
JP2015063190A (en) 2015-04-09 Automatic anti-glare device
JP6232691B2 (en) 2017-11-22 VEHICLE DISPLAY CONTROL DEVICE, VEHICLE DISPLAY DEVICE, AND VEHICLE DISPLAY CONTROL METHOD
JP2008044603A (en) 2008-02-28 Glare-proof device for vehicle
JP2010130647A (en) 2010-06-10 Vehicle periphery checking system
JP4798015B2 (en) 2011-10-19 Anti-glare control device
JP2010179817A (en) 2010-08-19 Anti-dazzling device for vehicle
WO2016018320A1 (en) 2016-02-04 System for projecting an image within a vehicle interior
US20250053005A1 (en) 2025-02-13 Head-up display apparatus
JP2009035162A (en) 2009-02-19 Rear view monitoring device
KR20170126149A (en) 2017-11-17 System for Providing Front Pillar View for a Vehicle
JP5352266B2 (en) 2013-11-27 Anti-glare device
JP2006137209A (en) 2006-06-01 Anti-glare device for vehicle
JP7182368B2 (en) 2022-12-02 VEHICLE DISPLAY DEVICE, METHOD AND COMPUTER PROGRAM FOR CONTROLLING VEHICLE DISPLAY DEVICE
CN117647889A (en) 2024-03-05 Display device for vehicle
JP2015085879A (en) 2015-05-07 Vehicular display device
JP4033170B2 (en) 2008-01-16 Vehicle display device
KR101826542B1 (en) 2018-02-07 Side mirror assist system and control method thereof
JP2004350228A (en) 2004-12-09 Nighttime rear visual field support system for vehicle
JP2008137595A (en) 2008-06-19 Vehicular external display device

Legal Events

Date Code Title Description
2021-03-09 PB01 Publication
2021-03-09 PB01 Publication
2021-03-26 SE01 Entry into force of request for substantive examination
2021-03-26 SE01 Entry into force of request for substantive examination
2021-12-03 CB02 Change of applicant information

Address after: Kyoto, Japan

Applicant after: MAXELL, Ltd.

Address before: Kyoto, Japan

Applicant before: MAXELL HOLDINGS, Ltd.

2021-12-03 CB02 Change of applicant information
2021-12-03 TA01 Transfer of patent application right

Effective date of registration: 20211119

Address after: Kyoto, Japan

Applicant after: MAXELL HOLDINGS, Ltd.

Address before: Kyoto, Japan

Applicant before: MAXELL, Ltd.

2021-12-03 TA01 Transfer of patent application right