WO2018180579A1 - 撮像制御装置、および撮像制御装置の制御方法、並びに移動体 - Google Patents
撮像制御装置、および撮像制御装置の制御方法、並びに移動体 Download PDFInfo
- Publication number
- WO2018180579A1 WO2018180579A1 PCT/JP2018/010397 JP2018010397W WO2018180579A1 WO 2018180579 A1 WO2018180579 A1 WO 2018180579A1 JP 2018010397 W JP2018010397 W JP 2018010397W WO 2018180579 A1 WO2018180579 A1 WO 2018180579A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- cameras
- vehicle
- moving body
- vertical direction
- imaging control
- Prior art date
Links
- 238000003384 imaging method Methods 0.000 title claims abstract description 92
- 238000000034 method Methods 0.000 title claims abstract description 52
- 238000001514 detection method Methods 0.000 claims description 95
- 238000005259 measurement Methods 0.000 abstract description 22
- 238000004891 communication Methods 0.000 description 45
- 230000033001 locomotion Effects 0.000 description 34
- 230000002093 peripheral effect Effects 0.000 description 30
- 230000008569 process Effects 0.000 description 29
- 238000012545 processing Methods 0.000 description 27
- 238000012544 monitoring process Methods 0.000 description 13
- 230000006870 function Effects 0.000 description 11
- 230000010365 information processing Effects 0.000 description 10
- 230000010354 integration Effects 0.000 description 9
- 238000010586 diagram Methods 0.000 description 8
- 238000006073 displacement reaction Methods 0.000 description 7
- 238000003702 image correction Methods 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 6
- 230000000007 visual effect Effects 0.000 description 5
- 238000004364 calculation method Methods 0.000 description 4
- 238000009434 installation Methods 0.000 description 4
- 230000007246 mechanism Effects 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 4
- 230000004075 alteration Effects 0.000 description 3
- 238000004590 computer program Methods 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 2
- 238000013459 approach Methods 0.000 description 2
- 230000010267 cellular communication Effects 0.000 description 2
- 238000002485 combustion reaction Methods 0.000 description 2
- 238000012937 correction Methods 0.000 description 2
- 230000007613 environmental effect Effects 0.000 description 2
- 230000005236 sound signal Effects 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000001816 cooling Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 230000000116 mitigating effect Effects 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
Images
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/20—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/22—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle
- B60R1/23—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view
- B60R1/27—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view providing all-round vision, e.g. using omnidirectional cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/239—Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/20—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/31—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles providing stereoscopic vision
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C3/00—Measuring distances in line of sight; Optical rangefinders
- G01C3/02—Details
- G01C3/06—Use of electric means to obtain final indication
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/271—Image signal generators wherein the generated image signals comprise depth maps or disparity maps
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/10—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used
- B60R2300/105—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used using multiple cameras
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/10—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used
- B60R2300/107—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used using stereoscopic cameras
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/30—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing
- B60R2300/301—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing combining image information with other obstacle sensor information, e.g. using RADAR/LIDAR/SONAR sensors for estimating risk of collision
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0081—Depth or disparity estimation from stereoscopic image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
Definitions
- the present disclosure relates to an imaging control device, a control method for the imaging control device, and a moving body, and in particular, an imaging control device capable of realizing distance measurement with a set of cameras arranged in a vertical direction with high accuracy,
- the present invention also relates to a control method for an imaging control apparatus, and a moving object.
- a stereo camera system configured by arranging a set of cameras in a horizontal direction on the left and right in a vehicle is mounted, and the difference between the same subject in a set of images caused by parallax between the cameras is used to A technique for measuring a distance has been proposed (see Patent Document 1).
- the stereo camera system of Patent Document 1 is configured to monitor a front side by arranging a set of cameras in the horizontal direction at the front end of the vehicle, but monitors a side by arranging a set of cameras in the vertical direction.
- the side monitoring from the front end of the vehicle can be advantageously configured.
- the readout timing of the pixel signal is shifted in units of rows, so that the subject that should exist at the same position in the horizontal direction is shifted, so-called focal plane. Distortion occurs.
- the present disclosure has been made in view of such a situation, and in particular, achieves distance measurement with a pair of cameras arranged in the vertical direction with high accuracy.
- An imaging control device is mounted on a moving body, and a distance between observation points in a detection range based on a pair of cameras constituting a stereo camera system and images captured by the pair of cameras.
- the pair of cameras are arranged on the side surface of the moving body and arranged in a direction perpendicular to the road surface, and pixel signals to be imaged are arranged in an array. It is an imaging control device that sequentially reads out in the vertical direction in units.
- the one set of cameras can be made to read out pixel signals in order from the column of pixels on the front side of the moving body, and in each column in the vertical direction.
- pixel signals are sequentially read out from the pixel columns on the front side of the moving body, and in each column sequentially from the bottom to the top in the vertical direction or from the top to the bottom. Can be made.
- the pair of cameras may be arranged on each of the left and right side surfaces of the moving body and arranged side by side in a direction perpendicular to the road surface.
- the pair of cameras may be configured to sequentially read out pixel signals in order from the pixel row on the front side of the moving body, and for each row in the vertical direction from bottom to top or from top to bottom. it can.
- the one set of cameras can be arranged on each of the left and right side surfaces of the moving body and arranged side by side in a direction perpendicular to the road surface.
- a set of cameras sequentially reads out pixel signals from the bottom in the vertical direction to the top in the vertical direction for each column from the front pixel row of the moving body. It is possible to sequentially read out pixel signals in the vertical direction from the top to the bottom for each row in order from the row of pixels on the front side of the moving body.
- the set of cameras is provided on at least one of the left and right side surfaces in front of the moving body, the left and right side surfaces in the center of the moving body, and the left and right side surfaces in the rear of the moving body, respectively. It can be arranged side by side so that parallax occurs in a direction perpendicular to the road surface.
- each of the cameras is a camera having a wider angle than a predetermined angle. can do.
- the camera pixels can be arranged in an array so that the horizontal width is longer than the vertical width.
- the camera pixels may be arranged in an array so that the ratio of the horizontal width to the vertical width is approximately 16: 9 or 4: 3. .
- the control information of the imaging control device is detected by detecting a distance of an observation point in a detection range based on an image captured by a pair of cameras that are mounted on a moving body and configure a stereo camera system.
- a moving object is configured to determine a distance between observation points in a detection range based on a set of cameras constituting a stereo camera system mounted on the moving object and images captured by the one set of cameras.
- a pair of cameras arranged on the side surface of the moving body and arranged in a direction perpendicular to the road surface, and the pixel signals to be imaged are arranged in an array. It is a moving body provided with an imaging control device that sequentially reads out pixels in a vertical direction.
- the distance between observation points in a detection range is detected based on an image captured by a set of cameras constituting a stereo camera system mounted on a moving body, and the set of cameras is The pixel signals arranged on the side surface of the moving body and arranged in the vertical direction with respect to the road surface are sequentially read out in the vertical direction in units of pixels arranged in an array.
- FIG. 1 is a block diagram illustrating an embodiment of an in-vehicle system to which the present technology is applied.
- the in-vehicle system 1 is a system that is mounted on a vehicle and provides driving support.
- the in-vehicle system 1 monitors the periphery of the vehicle and performs a process for preventing a collision or contact with a surrounding vehicle, bicycle, person or the like. More specifically, the in-vehicle system 1 performs notification of the danger of collision or contact, and control of a braking device such as a brake system for avoiding the collision or contact.
- the vehicle on which the in-vehicle system 1 is mounted is not particularly limited and includes, for example, a three-wheel truck, a small truck, a small passenger car, a large passenger car, a large bus, a large truck, a large special vehicle, a small special vehicle, and the like.
- a vehicle in which the in-vehicle system 1 is mounted is also referred to as a host vehicle, and a vehicle other than the host vehicle is also referred to as another vehicle.
- the in-vehicle system 1 includes a camera system 10, an imaging control unit 11, a vehicle information sensor 12, an information processing unit 13, a display unit 14, a braking control unit 15, and a braking device 16.
- the camera system 10 is a stereo camera composed of a pair of cameras arranged on the left and right side surfaces in front of the vehicle body in a direction perpendicular to the road surface, and captures an image obtained by imaging the side portion of the vehicle. 11 is output. Details of the camera system 10 will be described later with reference to FIG.
- peripheral image data an image captured by the camera system 10 is referred to as a peripheral image
- peripheral image data data indicating the peripheral image
- the imaging control unit 11 measures the distance to the subject (object) in the peripheral image as distance data based on the peripheral image captured by the camera system 10 and outputs the distance data together with the peripheral image data to the information processing unit 13.
- the vehicle information sensor 12 includes various sensors used for detecting the movement of the host vehicle.
- the vehicle information sensor 12 includes a speed sensor, a steering angle sensor, a GPS (global positioning system) receiver, and the like.
- the vehicle information sensor 12 supplies data indicating the detection result (hereinafter referred to as vehicle sensor data) to the motion prediction unit 32.
- the information processing unit 13 is configured by, for example, an ECU (Electronic Control Unit).
- the information processing unit 13 includes a surrounding situation detection unit 31, a motion prediction unit 32, a collision prediction unit 33, and an HMI (Human (Machine Interface) control unit 34.
- ECU Electronic Control Unit
- HMI Human (Machine Interface) control unit 34.
- the surrounding situation detection unit 31 detects the surrounding situation of the host vehicle based on the surrounding image data and the distance data.
- the surrounding situation detection unit 31 includes a space generation unit 41 and a moving object detection unit 42.
- the space generation unit 41 generates a three-dimensional space map indicating the shape, position, and the like of objects around the host vehicle based on the surrounding image data and the distance data.
- the space generation unit 41 supplies the three-dimensional space map to the motion prediction unit 32 and the HMI control unit 34.
- the space generation unit 41 supplies the peripheral image data to the HMI control unit 34.
- the moving object detection unit 42 detects moving objects around the host vehicle based on the surrounding image data and the three-dimensional space map.
- the moving object detection unit 42 supplies the detection result of the moving object to the motion prediction unit 32 and the HMI control unit 34.
- the motion prediction unit 32 performs motion prediction of the host vehicle based on the vehicle sensor data. In addition, the motion prediction unit 32 performs motion prediction of a moving object around the host vehicle based on the three-dimensional space map and the detection result of the moving object. The motion prediction unit 32 supplies the prediction result of the motion of the host vehicle and the moving objects around the host vehicle to the collision prediction unit 33 and the HMI control unit 34.
- the collision prediction unit 33 performs collision prediction of the moving body around the host vehicle based on the prediction result of the movement of the host vehicle and the moving body around the host vehicle.
- the collision prediction unit 33 supplies the result of the collision prediction to the HMI control unit 34 and the braking control unit 15.
- the HMI control unit 34 controls the HMI of the host vehicle. For example, the HMI control unit 34 determines the situation around the own vehicle based on the three-dimensional space map, the detection result of the moving object around the own vehicle, and the motion prediction and the collision prediction result around the own vehicle. Peripheral monitoring image data for displaying the peripheral monitoring image shown is generated. The HMI control unit 34 supplies the periphery monitoring image data to the display unit 14 to display the periphery monitoring image. In this case, the HMI control unit 34 functions as an image processing unit.
- the display unit 14 includes, for example, various displays.
- the display unit 14 displays various images such as a peripheral monitoring image under the control of the HMI control unit 34.
- the braking control unit 15 is configured by, for example, an ECU (Electronic Control Unit).
- the braking control unit 15 controls the braking device 16 based on the collision prediction result by the collision prediction unit 33, and performs an emergency stop or the like of the host vehicle.
- the braking device 16 is constituted by, for example, a brake system of the own vehicle.
- FIG. 2 shows an arrangement example of a stereo camera system constituting the camera system 10.
- Stereo camera system 111A is arranged, for example, on the front left side of the front bumper of vehicle 100.
- the stereo camera system 111A captures an image of a detection range 112A on the left side of the vehicle 100 including an area that is a driver's blind spot, and performs imaging control of image data indicating an image obtained as a result of the imaging (hereinafter referred to as a left image).
- a left image an image obtained as a result of the imaging
- the stereo camera system 111B is disposed, for example, on the front right side of the front bumper of the vehicle 100.
- the stereo camera system 111B images the detection range 112B in the right direction of the vehicle 100 including the area that becomes the driver's blind spot, and performs imaging control of image data indicating an image obtained as a result of imaging (hereinafter referred to as a right image).
- a right image an image obtained as a result of imaging
- stereo camera systems 111A and 111B when it is not necessary to particularly distinguish the stereo camera systems 111A and 111B, they are simply referred to as the stereo camera system 111, and other configurations are also referred to.
- the stereo camera system 111 ⁇ / b> A has a configuration in which two cameras 131 ⁇ / b> A and 132 ⁇ / b> A are set as one set, and is arranged side by side in the vertical direction (that is, the vertical direction) on the front left side of the vehicle 100. That is, the cameras 131A and 132A are arranged in a plane perpendicular to the reference plane (road surface 101) so that the parallax is generated in the height (vertical) direction.
- the stereo camera system 111A is installed only on the left side of the vehicle 100, but in reality, the stereo camera system 111B including the cameras 131B and 132B is also installed on the right side.
- the optical axis of the camera 131A and the optical axis of the camera 132A are directed in a direction (front side of the page) perpendicular to the dotted line in FIG. 3 that passes through the camera 131A and the camera 132A.
- a dotted line in FIG. 3 that is a perpendicular to the road surface 101 passing through the center point of the stereo camera system 111 intersects with a point T on the road surface 101. That is, the point T is a point on the road surface 101 immediately below the stereo camera system 111A (that is, directly below the vehicle 100).
- the camera system 10 includes stereo camera systems 111A and 111B.
- the stereo camera system 111A includes a camera 131A and a camera 132A.
- the stereo camera system 111B includes cameras 131B and 132B.
- Image data of images captured by the cameras 131A and 132A is supplied to the stereo distance measuring unit 151A, and image data of images captured by the imaging units 131B and 132B is supplied to the stereo distance measuring unit 151B.
- the imaging control unit 11 includes stereo distance measurement units 151A and 151B and an integration unit 152.
- Stereo distance measuring section 151 ⁇ / b> A measures the distance to the subject (object) in detection range 112 ⁇ / b> A (FIG. 2) on the left side of vehicle 100 and outputs the distance to integration section 152.
- Stereo distance measuring section 151 ⁇ / b> B measures the distance to the subject (object) in detection range 112 ⁇ / b> B on the right side of vehicle 100 and outputs the distance to integration section 152.
- the integration unit 152 acquires and integrates the outputs of the stereo distance measurement units 151A and 151B, grasps the state of the entire periphery of the vehicle 100, and outputs it to the surrounding state detection unit 31 as surrounding image data and distance data.
- the stereo distance measuring unit 151 includes image correcting units 171 and 172 and a stereo image processing unit 173.
- the outputs of the cameras 131 and 132 of the stereo camera system 111 are supplied to an image correction unit 171 and an image correction unit 172, respectively, and lens aberrations and the like are corrected as preprocessing.
- the captured images include distortion.
- the image correction units 171 and 172 perform processing for correcting this distortion and projecting the image onto a plane to obtain a plane image for distance calculation.
- the stereo image processing unit 173 detects the distance to the object from the outputs of the image correction unit 171 and the image correction unit 172. That is, the object shown in one image of the cameras 131 and 132 is detected from the other image, and the distance is calculated from the displacement of the position.
- the wide-angle camera is, for example, a camera provided with a lens of 35 mm or less, particularly a lens of 28 mm or less in terms of 35 mm.
- the wide-angle camera may be a camera that can capture an image with a viewing angle of 60 degrees or more, particularly 120 degrees or more, or 150 degrees or more.
- the viewing angle can be 180 degrees or more.
- the wide-angle lens having a wide viewing angle or the camera may be a fish-eye lens (f ⁇ lens) or a fish-eye camera, or a camera called a super-wide-angle lens or a super-wide-angle camera.
- Stereo Camera System Coordinate System Next, the coordinate system of the stereo camera system 111 will be described with reference to FIG.
- the coordinate system of the stereo camera system 111 is defined as shown in FIG.
- the center point of one camera 131 is Oa
- the center point of the other camera 132 is Ob
- the center point Oa and the center point Ob are Let O be the midpoint (ie, the center point of the stereo camera system 111).
- the target point of imaging is P
- the distance (baseline length) between the center point Oa and the center point Ob is L.
- the lower angle in the figure the lower side from the central point O of the straight line 141 and the straight line 144)
- the angle formed by the line segment is defined as ⁇ .
- the angle ⁇ is an angle formed between the imaging target point P and the stereo camera system 111.
- the target point represents an object (subject) to be imaged, that is, monitored.
- a person around the vehicle 100, an obstacle, another vehicle, and the like are schematically represented as points.
- the lower angle in the figure (the angle between the straight line 142 and the lower line segment in the figure from the central point Oa of the straight line 144) ) Is ⁇ a.
- the lower angle in the figure (the angle between the straight line 143 and the lower line segment in the figure from the central point Ob of the straight line 144) ) Is ⁇ b.
- An angle formed by the straight line 142 and the straight line 143 is ⁇ .
- the distance between the center point O and the target point P is ⁇
- the distance between the target point P and the center point Oa is ⁇ a
- the distance between the target point P and the center point Ob is ⁇ b.
- Equation (3) is obtained from Equation (1) and Equation (2).
- the distance (base line length) L between the center point Oa and the center point Ob is about several centimeters to several tens of centimeters, whereas the distance ⁇ from the center point O to the target point P is about several meters.
- ⁇ a and ⁇ b hold.
- ⁇ b ⁇ ⁇ a always holds. From these conditions, the following approximate expression (4) holds.
- the difference ⁇ a ⁇ b is the angle difference of the incident light.
- the distance to the object is determined by ⁇ a ⁇ b.
- the baseline length L is a constant, it can be seen from equation (5) that ⁇ a ⁇ b is inversely proportional to the distance ⁇ to the object. Therefore, if the distance between the object and the stereo camera system 111 increases, the distance measurement accuracy decreases.
- One of the main reasons for distance measurement in the stereo camera system 111 mounted on the vehicle 100 is to detect obstacles around the vehicle 100 and prevent contact between the vehicle 100 and the obstacles. Therefore, it is reasonable that the distance measurement accuracy increases as the distance between the vehicle 100 and the obstacle is shorter.
- the subject H visually recognizes the subject H with each of the left and right eyes 221 and 222 as shown in Example Ex1.
- the left eye 221 visually recognizes the subject H in the visual field L in the figure
- the right eye 222 in the visual field R in the figure.
- the subject H is visually recognized.
- a human recognizes the shift of the subject H in the visual fields L and R as parallax, and spatially recognizes the subject according to the parallax.
- the visual fields L and R have different distances from the eyes 221 and 222, but are different for convenience of explanation. In reality, it is assumed that they are the same distance.
- a stereo camera system When a stereo camera system is configured by arranging cameras in the horizontal direction so as to correspond to the eyes 221 and 222 based on the same principle, images 231 and 232 are captured as shown in Example Ex2 in each camera.
- Example Ex2 the subject H is captured as the subjects H1 and H2, respectively. Note that the images 231 and 232 in the example Ex2 correspond to the visual fields L and R of the example Ex1, respectively.
- the horizontal displacement of the subjects H1 and H2 in the images 231 and 232 is caused by the parallax of the camera that captures each image, and is caused by the parallax corresponding to the displacement in the respective images 231 and 232 of the subjects H1 and H2.
- the distance to the subject can be measured.
- the camera sequentially reads out from the top to the bottom in units of rows, and in each row, pixel signals are sequentially read from pixels adjacent in the horizontal direction from left to right. Therefore, in order to recognize the subjects H1 and H2 that are the same subject H, the pixel signals of the pixels in the same row are compared.
- the horizontal right arrow across the images 231 and 232 represents that pixels in the same row are compared.
- the stereo camera system 111 of the present disclosure since the cameras 131 and 132 are arranged in the vertical direction, the image captured by each of them is a case where the subject H is stationary, for example, Images are captured as images 233 and 234 as indicated by Ex3.
- the pixels belonging to the subjects H1 and H2 and having the same horizontal direction are If they belong to the same subject, the pixels shifted in the vertical direction by the amount corresponding to the parallax become the same pixel signal, so that the subject H can be recognized.
- the horizontal coordinate X of the images 233 and 234 is positive in the right direction in the figure
- the vertical coordinate Y is positive in the downward direction in the figure
- the pixel signal in each camera is, for example, in row units
- the pixel signals belonging to the subjects H1 and H2 are shifted to the right in the horizontal direction by a distance corresponding to the moving speed of the subject H as the readout row advances downward.
- the lower the row in the image the larger the displacement to the right side of the pixel belonging to H is, the focal plane distortion occurs. Therefore, the left of the subjects H1 and H2 as indicated by the subjects H1 ′ and H2 ′. Inclined distortion occurs.
- Example Ex4 the pixel belonging to the subject H2 'in the image 234 belongs to the upper row in the image than the pixel belonging to the subject H1' in the image 233. Accordingly, the pixel signal of the pixel belonging to the subject H2 'in the image 234 is read out earlier than the pixel signal of the pixel belonging to the subject H1' in the image 233.
- the subject H2 'in the image 234 is closer to the left side in the image than the subject H1' in the image 233.
- the pixel signals are sequentially read out in the vertical direction in units of rows, and for each row When reading in the horizontal direction, the same common subject cannot be recognized in an image captured by a pair of cameras, and distance measurement itself may not be possible.
- the horizontal direction is sequentially performed in units of columns.
- the pixel signals are generated so as to be sequentially read in the vertical direction, for example, from the top to the bottom.
- the deviation in the readout timing of the pixel signals in the same column is only the number of pixels corresponding to the parallax, so that the influence of the focal plane distortion can be reduced.
- the shift of the subjects H1 and H2 can also be reduced.
- the subjects H1 and H2 are the same subject H by comparing pixel signals at the same position.
- a general camera sequentially reads out pixel signals from top to bottom in units of rows and from left to right in each row, so that, for example, as shown in Example Ex12 in FIG.
- the correct images 233 and 234 are read out. Therefore, as shown by an example Ex13 in FIG. 8, it may be realized by rotating a general camera by 90 degrees (rotating rightward in the drawing).
- an image captured by the cameras 131 and 132 constituting the stereo camera system 111 of the present disclosure has a ratio A: B between the length in the horizontal direction and the length in the vertical direction, for example, 4: 3 or 16: 9, that is, the horizontal direction is wider, and pixel signals are sequentially read out in units of columns in the horizontal direction, and in each column, in units of pixels in the vertical direction. Then, the pixel signals are sequentially read out.
- the stereo camera system 111 of the present disclosure suppresses the horizontal shift of the subject due to the shift in the readout timing while ensuring the horizontal viewing angle, and performs the distance measurement with high accuracy. Make it happen.
- the pixel signal of the pixel in the column SC1 on the front side of the vehicle 100 is the uppermost stage in the figure.
- the pixel signal of the lowermost pixel 251 is read out sequentially from the first pixel
- the pixel signal of the lowermost pixel 251 is read out sequentially from the uppermost pixel of the adjacent column SC2.
- pixel signals of pixels adjacent in the downward direction are sequentially read out from the pixel signal of the uppermost pixel for each column.
- the left part in the figure is the front (front) of the vehicle 100
- the upper part in the figure is the upper part (up) of the vehicle 100
- the lower part in the figure is the lower part of the vehicle 100 ( down).
- the pixel signal of the pixel in the column SC1 ′ on the front side of the vehicle 100 is the uppermost pixel in the drawing.
- the pixel signal of the lowermost pixel 251 is read out sequentially from the lowermost pixel, the uppermost pixel in the adjacent column SC2 ′ is sequentially read downward.
- the pixel signals of the uppermost pixel are read out for each column, and the pixel signals of the pixels adjacent in the downward direction are sequentially read out.
- the right part in the figure is the front (front) of the vehicle 100
- the upper part in the figure is the upper part (up) of the vehicle 100
- the lower part in the figure is the lower part of the vehicle 100 ( down).
- pixel signals of each column are described as an example in which pixel signals of pixels adjacent in the downward direction are sequentially read from the uppermost pixel. You may make it read a pixel signal sequentially upward from the pixel of the lowest stage.
- the camera 131 may be reversed and provided on the left side part.
- the display unit 14 equipment provided in advance in the host vehicle may be used, or a dedicated display or the like may be provided.
- the display 271 of the car navigation system of the own vehicle or the instrument panel 272 can be used as the display unit 14.
- the display unit 14 can be configured by a transmissive display that is provided so as to overlap the area P1 in front of the driver's seat of the windshield 273 of the host vehicle.
- step S1 the information processing unit 13 controls the imaging control unit 11 to execute the peripheral image data and distance data acquisition process of the vehicle 100, and acquires the peripheral image data and the distance data.
- the surrounding state detection unit 31 controls the imaging control unit 11 to acquire the surrounding image data and distance data of the vehicle 100 based on the image captured by the camera system 10.
- the imaging control unit 11 controls the camera system 10 to execute the peripheral image data and distance data acquisition process, and acquires the peripheral image data of the vehicle 100 and the distance data that is the stereo distance measurement result.
- the motion prediction unit 32 acquires vehicle sensor data from each sensor of the vehicle information sensor 12. Details of the peripheral image data and distance data acquisition processing will be described later with reference to the flowchart of FIG.
- step S2 the space generation unit 41 performs a space generation process. That is, the space generation unit 41 generates (or updates) a three-dimensional space map indicating the shape, position, and the like of objects around the host vehicle based on the surrounding image data and the distance data. Note that the objects in the vicinity of the host vehicle include not only moving objects but also stationary objects (for example, buildings, road surfaces, etc.). The space generation unit 41 supplies the generated three-dimensional space map to the motion prediction unit 32 and the HMI control unit 34.
- any method can be used as a method of generating the three-dimensional space map.
- a technique such as SLAM (Simultaneous Localization and Mapping) is used.
- the moving object detection unit 42 performs moving object detection. Specifically, the moving object detection unit 42 detects moving objects around the host vehicle based on the surrounding sensor data and the three-dimensional space map. For example, the moving object detection unit 42 detects the presence or absence of a moving object around the host vehicle, the type, size, shape, position, and the like of the moving object. The moving object detection unit 42 supplies the detection result of the moving object to the motion prediction unit 32 and the HMI control unit 34.
- the moving object to be detected includes not only a moving object that is actually moving, but also a moving object that is temporarily stationary such as a stopped vehicle, a bicycle, or a stopped pedestrian.
- the moving object detection unit 42 can also detect a moving object around the host vehicle based only on the surrounding sensor data without using a three-dimensional space map, for example. In this case, the processes of step S2 and step S3 can be interchanged.
- step S4 the moving object detection unit 42 determines whether there is a moving object in the vicinity based on the result of the process in step S3. If it is determined that there are no moving objects in the vicinity, the process returns to step S1.
- step S1 to step S4 is repeatedly executed until it is determined in step S4 that a moving object exists in the vicinity.
- step S4 if it is determined in step S4 that a moving object is present in the vicinity, the process proceeds to step S5.
- step S5 the motion prediction unit 32 performs motion prediction. Specifically, the motion prediction unit 32 predicts the moving speed and the moving direction of the host vehicle based on the vehicle sensor data. In addition, the motion prediction unit 32 predicts the moving speed and moving direction of moving objects around the host vehicle based on the three-dimensional space map and the detection results of moving objects around the host vehicle. The motion prediction unit 32 supplies the prediction result to the collision prediction unit 33 and the HMI control unit 34.
- the collision prediction unit 33 performs collision prediction. Specifically, the collision prediction unit 33 determines whether or not there is a possibility that the moving body around the own vehicle collides with or comes into contact with the own vehicle based on the prediction result of the movement of the own vehicle and the surrounding moving body. In addition, the time required for collision or contact of a moving object that may possibly collide or come into contact (hereinafter referred to as a collision prediction time) is predicted.
- the collision prediction unit 33 predicts the risk of each moving object colliding with or coming into contact with the host vehicle, and sets the rank based on a predetermined definition. For example, a moving object that is stationary and a moving object that is moving in a direction away from the host vehicle is set to a risk level 1. Among moving objects that are moving in the direction approaching the host vehicle, moving objects that have a predicted collision time exceeding T1 seconds (for example, 5 seconds) are set to a risk level 2. Of the moving objects that are moving in the direction approaching the host vehicle, the moving object that has a predicted collision time within T1 seconds and exceeds T2 seconds (for example, 1 second) is set to a degree of risk 3. Of the moving objects that are moving in the direction approaching the host vehicle, a moving object that has a predicted collision time of T2 seconds or less is set to a degree of risk 4.
- T1 seconds for example, 5 seconds
- T2 seconds for example, 1 second
- a stationary moving body and a moving body moving away from the host vehicle may also be set to any one of the risk levels 2 to 4 based on the collision prediction time.
- the collision prediction unit 33 supplies the result of the collision prediction to the HMI control unit 34 and the braking control unit 15.
- step S7 the collision prediction unit 33 determines whether there is a risk of collision or contact. For example, the collision prediction unit 33 determines that there is no risk of collision and contact when there is no moving object having a risk level of 3 or more in the vicinity of the host vehicle, and the process returns to step S1.
- steps S1 to S7 are repeatedly executed until it is determined in step S7 that there is a risk of collision or contact.
- step S7 for example, the collision prediction unit 33 determines that there is a risk of collision or contact when there is a moving body having a risk level of 3 or more around the host vehicle, and the process proceeds to step S8.
- step S8 the HMI control unit 34 sets the moving object type.
- the HMI control unit 34 classifies the moving objects detected by the moving object detection unit 42 into a vehicle, a motorbike, a bicycle, a pedestrian, and other five types.
- a motorbike is also a kind of vehicle, but here, a motorbike is distinguished from other vehicles.
- the HMI control unit 34 calculates the display position of the superimposed image.
- the superimposed image includes, for example, a frame indicating the position of each moving object (hereinafter referred to as a moving object frame) and a bar indicating a predicted motion of each moving object (hereinafter referred to as a motion prediction bar).
- the HMI control unit 34 determines the three-dimensional of the moving object frame corresponding to each moving object based on the position of each moving object in the three-dimensional space map and the height and width of each moving object as viewed from the traveling direction of each moving object.
- the display position in the space map is calculated.
- the HMI control unit 34 calculates the position of each moving object after x seconds (for example, after one second) in the three-dimensional space map based on the result of motion prediction of each moving object. Next, the HMI control unit 34 calculates the display position of the motion prediction bar corresponding to each moving object on the three-dimensional space map based on the current position of each moving object on the three-dimensional space map and the position after x seconds. To do. For example, the HMI control unit 34 calculates the length and direction of the motion prediction bar by using the current moving object's leading end in the moving direction as the starting point and the moving object's moving direction's leading end x seconds later as the ending point. .
- step S10 the in-vehicle system 10 presents the surrounding situation.
- the HMI control unit 34 converts the display position of the superimposed image (moving object frame, motion prediction bar, etc.) in the three-dimensional space map into the display position in the peripheral image presented to the driver. Further, the HMI control unit 34 converts the position of the road surface in the three-dimensional space map into the display position in the peripheral image, and calculates the display position of the grid indicating the position of the road surface in the peripheral image. Then, the HMI control unit 34 generates surrounding monitoring image data indicating the surrounding monitoring image, supplies it to the display unit 14, and displays the surrounding monitoring image.
- step S11 the braking control unit 15 determines whether or not an emergency stop is necessary. For example, the brake control unit 15 determines that there is no need for an emergency stop when there is no moving object of the risk level 4 in moving objects around the host vehicle, and the process returns to step S1.
- step S11 Thereafter, the processing from step S1 to step S11 is repeatedly executed until it is determined in step S11 that an emergency stop is necessary.
- step S11 for example, the braking control unit 15 determines that an emergency stop is necessary when there is a moving body of risk 4 in the moving body around the host vehicle, and the process proceeds to step S12.
- step S12 the braking control unit 15 controls the braking device 16 to urgently stop the host vehicle. Thereby, the collision or contact with the moving body around the host vehicle is prevented.
- the danger of collision or contact with moving objects around the host vehicle can be notified in an easy-to-understand manner, and the driver can reliably recognize the danger of collision or contact.
- an emergency stop is performed, so that an accident is prevented.
- peripheral image data and distance data acquisition processing will be described with reference to the flowchart of FIG.
- step S31 the camera 131 and the camera 132 constituting the stereo camera system 111 capture an observation point.
- step S32 the image correction unit 171 corrects lens aberration, camera image distortion, and the like for the image captured by the camera 131.
- the image correction unit 172 corrects lens aberration, camera image distortion, and the like for the image captured by the camera 132. That is, image distortion is corrected for distance calculation, and the image is projected onto a virtual plane to obtain a plane image.
- step S33 the stereo image processing unit 173 calculates the distance to the observation point, and outputs a captured image including a planar image together with the distance data to the integration unit 152 as peripheral image data. That is, the camera 131 and the camera 132 are arranged at positions separated by a distance L in the vertical direction. Accordingly, the image captured by the camera 131 and the image captured by the camera 132 have a phase difference, and the distance to the observation point can be calculated based on the phase difference. That is, an object corresponding to an object shown in one of the images of the cameras 131 and 132 is detected from the other image, and the distance is calculated from the displacement of the position of the object in the two images.
- the above process is executed by each of the stereo camera systems 111A and 111B provided on the left and right front side surfaces of the vehicle 100.
- step S ⁇ b> 34 the integration unit 93 integrates the respective measurement data of the stereo camera systems 111 ⁇ / b> A and 111 ⁇ / b> B and the peripheral image data, and outputs the data to the imaging control unit 11 together with the peripheral image data as distance data in all directions of the vehicle 100. .
- step S35 the stereo image processing unit 173 determines whether to end the process. If the user has not yet instructed the end of the process, the process returns to step S31, and the subsequent processes are repeated. If the end of the process is instructed, the process ends.
- the distance data as the distance measurement results in the detection ranges 112A and 112B by the stereo distance measuring units 151A and 151B are output to the integrating unit 152 together with the peripheral image data, and are sequentially and repeatedly output to the imaging control unit 11. Is done.
- step S31 when the cameras 131 and 132 image the detection range, as described above, in the order from the front pixel row of the vehicle 100 to the rear pixel row adjacent in the horizontal direction, and In each pixel column, pixel signals of pixels adjacent in the downward direction are sequentially read from the uppermost pixel, or pixel signals of pixels adjacent in the upward direction are sequentially read from the lowermost pixel.
- the focal plane distortion of the subject in the images picked up by the cameras 131 and 132, and the displacement of the subject in the horizontal direction due to the deviation of the readout timing, can be obtained. Occurrence can be suppressed, and the subjects in the respective images can be recognized as the same subject, so that it is possible to realize distance measurement by deviation based on parallax with high accuracy.
- the imaging control unit 11 measures the distance to the subject (target object) in the peripheral image as distance data.
- the measurement of the distance data itself is another configuration.
- the space generation unit 41 may simultaneously generate distance data and a three-dimensional space map.
- the detection process of the type of moving object has been described with respect to the example implemented by the moving object detection unit 42, but other configurations may be implemented. For example, you may make it the imaging control part 11 implement.
- the stereo camera system 111 has been described with respect to an example in which the cameras 131 and 132 are arranged side by side so that parallax is generated in the vertical direction on the front side surface of the vehicle 100.
- the stereo camera system 111 may be arranged so that the parallax of the cameras 131 and 132 is generated in the vertical direction.
- the stereo camera system 111 may be provided at a position in the vicinity of the central side surface portion of the vehicle 100 so that the cameras 131 and 132 image the side surface direction so that parallax occurs in the vertical direction.
- the stereo camera system 111 is attached to the door mirror 301 side by side in the vertical direction so that the parallax of the cameras 131 and 132 is generated in the vertical direction. May be.
- the stereo camera system 111 may be at a position other than the door mirror 301 as long as it is in the vicinity of the central side surface of the vehicle 100.
- the stereo camera system 111 may be attached to a pillar (front pillar, center pillar, rear pillar, etc.), door, roof rail, etc. Also good.
- the stereo camera system 111 may be configured such that the cameras 131 and 132 are attached to the rear side surface portion of the vehicle 100 so that parallax occurs in the vertical direction.
- the stereo camera system 111 when the cameras 131 and 132 are attached to the rear side surface portion of the vehicle 100 so as to generate a parallax in the vertical direction, the stereo camera system 111 is used when moving backward. Since the number of uses increases, the pixel signal in the detection range on the rear side of the side surface may be read out at higher speed.
- pixel signals of pixels adjacent in the downward direction are sequentially read out from the pixel signal of the uppermost pixel for each pixel column from the pixel columns on the rear side of the pixel array of the cameras 131 and 132, or
- the pixel signals of pixels adjacent in the upward direction may be sequentially read out from the pixel signals of the pixels.
- the pixel signal of the pixel in the column SC8 on the rear side of the vehicle 100 is When the pixel signal of the lowermost pixel 251 is read out sequentially from the lower pixel, the uppermost pixel of the adjacent column SC7 is sequentially read downward. Similarly, pixel signals of pixels adjacent in the downward direction are sequentially read out from the pixel signal of the uppermost pixel in each column in the order of columns SC6 to SC1.
- the right part in the figure is the rear (Rear) of the vehicle 100
- the upper part in the figure is the upper part (up) of the vehicle 100
- the lower part in the figure is the lower part of the vehicle 100 ( down).
- the pixel signal of the pixel in the column SC8 ′ on the rear side of the vehicle 100 is the uppermost pixel in the drawing.
- the pixel signal of the lowermost pixel 251 is read out sequentially from the bottom, the uppermost pixel in the adjacent column SC7 ′ is sequentially read downward.
- the pixel signals of the uppermost pixel are read out, and the pixel signals of the pixels adjacent in the downward direction are sequentially read out.
- the left part in the figure is the rear (Rear) of the vehicle 100
- the upper part in the figure is the upper part (up) of the vehicle 100
- the lower part in the figure is the lower part of the vehicle 100 ( down).
- the pixel signal of each column is described as an example in which pixel signals of pixels adjacent in the downward direction are sequentially read from the uppermost pixel. You may make it read a pixel signal sequentially upward from the pixel of the lowest stage.
- the camera 131 in the upper left part of FIG. 15 is installed upside down while maintaining the front-rear direction with respect to the vehicle 100 so that it is provided on the rear left side part as shown in the lower right part of FIG. Also good.
- the camera 131 it is not necessary to separately configure the camera 131 to be provided on the rear right side portion and the camera 131 to be provided on the left side portion, and to be provided on both the left and right side portions with the same configuration. It becomes possible to make it.
- the stereo camera system 111 is a side part of the vehicle 100, it may be installed in any of the front, the center, and the back, and may be installed in combination of at least one of them.
- the stereo camera system 111 when the stereo camera system 111 is installed on the front side surface portion and the rear side surface portion of the vehicle 100, cameras that can capture a distant place with a narrow angle with a slightly narrow viewing angle are adopted as the cameras 131 and 132. You may do it. That is, when the stereo camera system 111 is installed on the front side surface portion and the rear side surface portion of the vehicle 100, it is possible to realize a quicker situation in the intersection by adopting a narrow-angle camera. .
- the stereo camera system 111 when the stereo camera system 111 is installed on the central side surface portion of the vehicle 100, a camera that can take a wide angle with a slightly wider viewing angle and can capture the vicinity well is adopted as the cameras 131 and 132. Good. That is, when the stereo camera system 111 is installed on the central side surface of the vehicle 100, it is possible to realize the situation in a wide range including the front and rear by adopting a wide-angle camera. Become.
- the technology according to the present disclosure can be applied to various products.
- the technology according to the present disclosure may be any type of movement such as an automobile, an electric vehicle, a hybrid electric vehicle, a motorcycle, a bicycle, personal mobility, an airplane, a drone, a ship, a robot, a construction machine, and an agricultural machine (tractor). You may implement
- FIG. 16 is a block diagram illustrating a schematic configuration example of a vehicle control system 7000 that is an example of a mobile control system to which the technology according to the present disclosure can be applied.
- the vehicle control system 7000 includes a plurality of electronic control units connected via a communication network 7010.
- the vehicle control system 7000 includes a drive system control unit 7100, a body system control unit 7200, a battery control unit 7300, a vehicle exterior information detection unit 7400, a vehicle interior information detection unit 7500, and an integrated control unit 7600. .
- the communication network 7010 for connecting the plurality of control units conforms to an arbitrary standard such as CAN (Controller Area Network), LIN (Local Interconnect Network), LAN (Local Area Network), or FlexRay (registered trademark). It may be an in-vehicle communication network.
- Each control unit includes a microcomputer that performs arithmetic processing according to various programs, a storage unit that stores programs executed by the microcomputer or parameters used for various calculations, and a drive circuit that drives various devices to be controlled. Is provided.
- Each control unit includes a network I / F for communicating with other control units via a communication network 7010, and is connected to devices or sensors inside and outside the vehicle by wired communication or wireless communication. A communication I / F for performing communication is provided. In FIG.
- a microcomputer 7610 As a functional configuration of the integrated control unit 7600, a microcomputer 7610, a general-purpose communication I / F 7620, a dedicated communication I / F 7630, a positioning unit 7640, a beacon receiving unit 7650, an in-vehicle device I / F 7660, an audio image output unit 7670, An in-vehicle network I / F 7680 and a storage unit 7690 are illustrated.
- other control units include a microcomputer, a communication I / F, a storage unit, and the like.
- the drive system control unit 7100 controls the operation of the device related to the drive system of the vehicle according to various programs.
- the drive system control unit 7100 includes a driving force generator for generating a driving force of a vehicle such as an internal combustion engine or a driving motor, a driving force transmission mechanism for transmitting the driving force to wheels, and a steering angle of the vehicle. It functions as a control device such as a steering mechanism that adjusts and a braking device that generates a braking force of the vehicle.
- the drive system control unit 7100 may have a function as a control device such as ABS (Antilock Brake System) or ESC (Electronic Stability Control).
- a vehicle state detection unit 7110 is connected to the drive system control unit 7100.
- the vehicle state detection unit 7110 includes, for example, a gyro sensor that detects the angular velocity of the rotational movement of the vehicle body, an acceleration sensor that detects the acceleration of the vehicle, an operation amount of an accelerator pedal, an operation amount of a brake pedal, and steering of a steering wheel. At least one of sensors for detecting an angle, an engine speed, a rotational speed of a wheel, or the like is included.
- the drive system control unit 7100 performs arithmetic processing using a signal input from the vehicle state detection unit 7110, and controls an internal combustion engine, a drive motor, an electric power steering device, a brake device, or the like.
- the body system control unit 7200 controls the operation of various devices mounted on the vehicle body according to various programs.
- the body system control unit 7200 functions as a keyless entry system, a smart key system, a power window device, or a control device for various lamps such as a headlamp, a back lamp, a brake lamp, a blinker, or a fog lamp.
- the body control unit 7200 can be input with radio waves or various switch signals transmitted from a portable device that substitutes for a key.
- the body system control unit 7200 receives input of these radio waves or signals, and controls a door lock device, a power window device, a lamp, and the like of the vehicle.
- the battery control unit 7300 controls the secondary battery 7310 that is a power supply source of the drive motor according to various programs. For example, information such as battery temperature, battery output voltage, or remaining battery capacity is input to the battery control unit 7300 from a battery device including the secondary battery 7310. The battery control unit 7300 performs arithmetic processing using these signals, and controls the temperature adjustment of the secondary battery 7310 or the cooling device provided in the battery device.
- the outside information detection unit 7400 detects information outside the vehicle on which the vehicle control system 7000 is mounted.
- the outside information detection unit 7400 is connected to at least one of the imaging unit 7410 and the outside information detection unit 7420.
- the imaging unit 7410 includes at least one of a ToF (Time Of Flight) camera, a stereo camera, a monocular camera, an infrared camera, and other cameras.
- the outside information detection unit 7420 detects, for example, current weather or an environmental sensor for detecting weather, or other vehicles, obstacles, pedestrians, etc. around the vehicle equipped with the vehicle control system 7000. At least one of the surrounding information detection sensors.
- the environmental sensor may be, for example, at least one of a raindrop sensor that detects rainy weather, a fog sensor that detects fog, a sunshine sensor that detects sunlight intensity, and a snow sensor that detects snowfall.
- the ambient information detection sensor may be at least one of an ultrasonic sensor, a radar device, and a LIDAR (Light Detection and Ranging, Laser Imaging Detection and Ranging) device.
- the imaging unit 7410 and the outside information detection unit 7420 may be provided as independent sensors or devices, or may be provided as a device in which a plurality of sensors or devices are integrated.
- FIG. 17 shows an example of installation positions of the imaging unit 7410 and the vehicle outside information detection unit 7420.
- the imaging units 7910, 7912, 7914, 7916, and 7918 are provided at, for example, at least one of the front nose, the side mirror, the rear bumper, the back door, and the upper part of the windshield in the vehicle interior of the vehicle 7900.
- An imaging unit 7910 provided in the front nose and an imaging unit 7918 provided in the upper part of the windshield in the vehicle interior mainly acquire an image in front of the vehicle 7900.
- Imaging units 7912 and 7914 provided in the side mirror mainly acquire an image of the side of the vehicle 7900.
- An imaging unit 7916 provided in the rear bumper or the back door mainly acquires an image behind the vehicle 7900.
- the imaging unit 7918 provided on the upper part of the windshield in the passenger compartment is mainly used for detecting a preceding vehicle or a pedestrian, an obstacle, a traffic light, a traffic sign, a lane, or
- FIG. 17 shows an example of shooting ranges of the respective imaging units 7910, 7912, 7914, and 7916.
- the imaging range a indicates the imaging range of the imaging unit 7910 provided in the front nose
- the imaging ranges b and c indicate the imaging ranges of the imaging units 7912 and 7914 provided in the side mirrors, respectively
- the imaging range d The imaging range of the imaging part 7916 provided in the rear bumper or the back door is shown. For example, by superimposing the image data captured by the imaging units 7910, 7912, 7914, and 7916, an overhead image when the vehicle 7900 is viewed from above is obtained.
- the vehicle outside information detection units 7920, 7922, 7924, 7926, 7928, and 7930 provided on the front, rear, sides, corners of the vehicle 7900 and the upper part of the windshield in the vehicle interior may be, for example, an ultrasonic sensor or a radar device.
- the vehicle outside information detection units 7920, 7926, and 7930 provided on the front nose, the rear bumper, the back door, and the windshield in the vehicle interior of the vehicle 7900 may be, for example, LIDAR devices.
- These outside information detection units 7920 to 7930 are mainly used for detecting a preceding vehicle, a pedestrian, an obstacle, and the like.
- the vehicle exterior information detection unit 7400 causes the imaging unit 7410 to capture an image outside the vehicle and receives the captured image data. Further, the vehicle exterior information detection unit 7400 receives detection information from the vehicle exterior information detection unit 7420 connected thereto.
- the vehicle exterior information detection unit 7420 is an ultrasonic sensor, a radar device, or a LIDAR device
- the vehicle exterior information detection unit 7400 transmits ultrasonic waves, electromagnetic waves, or the like, and receives received reflected wave information.
- the outside information detection unit 7400 may perform an object detection process or a distance detection process such as a person, a car, an obstacle, a sign, or a character on a road surface based on the received information.
- the vehicle exterior information detection unit 7400 may perform environment recognition processing for recognizing rainfall, fog, road surface conditions, or the like based on the received information.
- the vehicle outside information detection unit 7400 may calculate a distance to an object outside the vehicle based on the received information.
- the outside information detection unit 7400 may perform image recognition processing or distance detection processing for recognizing a person, a car, an obstacle, a sign, a character on a road surface, or the like based on the received image data.
- the vehicle exterior information detection unit 7400 performs processing such as distortion correction or alignment on the received image data, and combines the image data captured by the different imaging units 7410 to generate an overhead image or a panoramic image. Also good.
- the vehicle exterior information detection unit 7400 may perform viewpoint conversion processing using image data captured by different imaging units 7410.
- the vehicle interior information detection unit 7500 detects vehicle interior information.
- a driver state detection unit 7510 that detects the driver's state is connected to the in-vehicle information detection unit 7500.
- Driver state detection unit 7510 may include a camera that captures an image of the driver, a biosensor that detects biometric information of the driver, a microphone that collects sound in the passenger compartment, and the like.
- the biometric sensor is provided, for example, on a seat surface or a steering wheel, and detects biometric information of an occupant sitting on the seat or a driver holding the steering wheel.
- the vehicle interior information detection unit 7500 may calculate the degree of fatigue or concentration of the driver based on the detection information input from the driver state detection unit 7510, and determines whether the driver is asleep. May be.
- the vehicle interior information detection unit 7500 may perform a process such as a noise canceling process on the collected audio signal.
- the integrated control unit 7600 controls the overall operation in the vehicle control system 7000 according to various programs.
- An input unit 7800 is connected to the integrated control unit 7600.
- the input unit 7800 is realized by a device that can be input by a passenger, such as a touch panel, a button, a microphone, a switch, or a lever.
- the integrated control unit 7600 may be input with data obtained by recognizing voice input through a microphone.
- the input unit 7800 may be, for example, a remote control device using infrared rays or other radio waves, or may be an external connection device such as a mobile phone or a PDA (Personal Digital Assistant) that supports the operation of the vehicle control system 7000. May be.
- the input unit 7800 may be, for example, a camera.
- the passenger can input information using a gesture.
- data obtained by detecting the movement of the wearable device worn by the passenger may be input.
- the input unit 7800 may include, for example, an input control circuit that generates an input signal based on information input by a passenger or the like using the input unit 7800 and outputs the input signal to the integrated control unit 7600.
- a passenger or the like operates the input unit 7800 to input various data or instruct a processing operation to the vehicle control system 7000.
- the storage unit 7690 may include a ROM (Read Only Memory) that stores various programs executed by the microcomputer, and a RAM (Random Access Memory) that stores various parameters, calculation results, sensor values, and the like.
- the storage unit 7690 may be realized by a magnetic storage device such as an HDD (Hard Disc Drive), a semiconductor storage device, an optical storage device, a magneto-optical storage device, or the like.
- General-purpose communication I / F 7620 is a general-purpose communication I / F that mediates communication with various devices existing in the external environment 7750.
- General-purpose communication I / F7620 is a cellular communication protocol such as GSM (Global System of Mobile communications), WiMAX, LTE (Long Term Evolution) or LTE-A (LTE-Advanced), or wireless LAN (Wi-Fi (registered trademark)). Other wireless communication protocols such as Bluetooth (registered trademark) may also be implemented.
- the general-purpose communication I / F 7620 is connected to a device (for example, an application server or a control server) existing on an external network (for example, the Internet, a cloud network, or an operator-specific network) via, for example, a base station or an access point.
- the general-purpose communication I / F 7620 is a terminal (for example, a driver, a pedestrian or a store terminal, or an MTC (Machine Type Communication) terminal) that exists in the vicinity of the vehicle using, for example, P2P (Peer To Peer) technology. You may connect with.
- a terminal for example, a driver, a pedestrian or a store terminal, or an MTC (Machine Type Communication) terminal
- P2P Peer To Peer
- the dedicated communication I / F 7630 is a communication I / F that supports a communication protocol formulated for use in vehicles.
- the dedicated communication I / F 7630 is a standard protocol such as WAVE (Wireless Access in Vehicle Environment), DSRC (Dedicated Short Range Communications), or cellular communication protocol, which is a combination of the lower layer IEEE 802.11p and the upper layer IEEE 1609. May be implemented.
- the dedicated communication I / F 7630 typically includes vehicle-to-vehicle communication, vehicle-to-infrastructure communication, vehicle-to-home communication, and vehicle-to-pedestrian communication. ) Perform V2X communication, which is a concept that includes one or more of the communications.
- the positioning unit 7640 receives, for example, a GNSS signal from a GNSS (Global Navigation Satellite System) satellite (for example, a GPS signal from a GPS (Global Positioning System) satellite), performs positioning, and performs latitude, longitude, and altitude of the vehicle.
- the position information including is generated.
- the positioning unit 7640 may specify the current position by exchanging signals with the wireless access point, or may acquire position information from a terminal such as a mobile phone, PHS, or smartphone having a positioning function.
- the beacon receiving unit 7650 receives, for example, radio waves or electromagnetic waves transmitted from a radio station installed on the road, and acquires information such as the current position, traffic jam, closed road, or required time. Note that the function of the beacon receiving unit 7650 may be included in the dedicated communication I / F 7630 described above.
- the in-vehicle device I / F 7660 is a communication interface that mediates the connection between the microcomputer 7610 and various in-vehicle devices 7760 present in the vehicle.
- the in-vehicle device I / F 7660 may establish a wireless connection using a wireless communication protocol such as a wireless LAN, Bluetooth (registered trademark), NFC (Near Field Communication), or WUSB (Wireless USB).
- the in-vehicle device I / F 7660 is connected to a USB (Universal Serial Bus), HDMI (High-Definition Multimedia Interface), or MHL (Mobile High-definition Link) via a connection terminal (and a cable if necessary). ) Etc. may be established.
- the in-vehicle device 7760 may include, for example, at least one of a mobile device or a wearable device that a passenger has, or an information device that is carried into or attached to the vehicle.
- In-vehicle device 7760 may include a navigation device that searches for a route to an arbitrary destination.
- In-vehicle device I / F 7660 exchanges control signals or data signals with these in-vehicle devices 7760.
- the in-vehicle network I / F 7680 is an interface that mediates communication between the microcomputer 7610 and the communication network 7010.
- the in-vehicle network I / F 7680 transmits and receives signals and the like in accordance with a predetermined protocol supported by the communication network 7010.
- the microcomputer 7610 of the integrated control unit 7600 is connected via at least one of a general-purpose communication I / F 7620, a dedicated communication I / F 7630, a positioning unit 7640, a beacon receiving unit 7650, an in-vehicle device I / F 7660, and an in-vehicle network I / F 7680.
- the vehicle control system 7000 is controlled according to various programs based on the acquired information. For example, the microcomputer 7610 calculates a control target value of the driving force generation device, the steering mechanism, or the braking device based on the acquired information inside and outside the vehicle, and outputs a control command to the drive system control unit 7100. Also good.
- the microcomputer 7610 realizes ADAS (Advanced Driver Assistance System) functions including vehicle collision avoidance or impact mitigation, following traveling based on inter-vehicle distance, vehicle speed maintaining traveling, vehicle collision warning, or vehicle lane departure warning. You may perform the cooperative control for the purpose. Further, the microcomputer 7610 controls the driving force generator, the steering mechanism, the braking device, or the like based on the acquired information on the surroundings of the vehicle, so that the microcomputer 7610 automatically travels independently of the driver's operation. You may perform the cooperative control for the purpose of driving.
- ADAS Advanced Driver Assistance System
- the microcomputer 7610 is information acquired via at least one of the general-purpose communication I / F 7620, the dedicated communication I / F 7630, the positioning unit 7640, the beacon receiving unit 7650, the in-vehicle device I / F 7660, and the in-vehicle network I / F 7680.
- the three-dimensional distance information between the vehicle and the surrounding structure or an object such as a person may be generated based on the above and local map information including the peripheral information of the current position of the vehicle may be created.
- the microcomputer 7610 may generate a warning signal by predicting a danger such as a collision of a vehicle, approach of a pedestrian or the like or an approach to a closed road based on the acquired information.
- the warning signal may be, for example, a signal for generating a warning sound or lighting a warning lamp.
- the audio image output unit 7670 transmits an output signal of at least one of audio and image to an output device capable of visually or audibly notifying information to a vehicle occupant or the outside of the vehicle.
- an audio speaker 7710, a display unit 7720, and an instrument panel 7730 are illustrated as output devices.
- Display unit 7720 may include at least one of an on-board display and a head-up display, for example.
- the display portion 7720 may have an AR (Augmented Reality) display function.
- the output device may be other devices such as headphones, wearable devices such as glasses-type displays worn by passengers, projectors, and lamps.
- the display device can display the results obtained by various processes performed by the microcomputer 7610 or information received from other control units in various formats such as text, images, tables, and graphs. Display visually. Further, when the output device is an audio output device, the audio output device converts an audio signal made up of reproduced audio data or acoustic data into an analog signal and outputs it aurally.
- At least two control units connected via the communication network 7010 may be integrated as one control unit.
- each control unit may be configured by a plurality of control units.
- the vehicle control system 7000 may include another control unit not shown.
- some or all of the functions of any of the control units may be given to other control units. That is, as long as information is transmitted and received via the communication network 7010, the predetermined arithmetic processing may be performed by any one of the control units.
- a sensor or device connected to one of the control units may be connected to another control unit, and a plurality of control units may transmit / receive detection information to / from each other via the communication network 7010. .
- a computer program for realizing the functions of the imaging control unit 11 and the information processing unit 13 according to the present embodiment described with reference to FIG. 1 can be installed in any control unit or the like. It is also possible to provide a computer-readable recording medium in which such a computer program is stored.
- the recording medium is, for example, a magnetic disk, an optical disk, a magneto-optical disk, a flash memory, or the like. Further, the above computer program may be distributed via a network, for example, without using a recording medium.
- the imaging control unit 11 and the information processing unit 13 according to the present embodiment described using FIG. 1 can be applied to the integrated control unit 7600 of the application example illustrated in FIG. .
- the stereo distance measuring units 151 ⁇ / b> A and 151 ⁇ / b> B and the integration unit 152 of the imaging control unit 11 correspond to the microcomputer 7610 of the integration control unit 7600.
- the integrated control unit 7600 performs distance data and peripheral image data acquisition processing based on an image captured by the camera system 10 to acquire distance data and peripheral image data with a peripheral object. it can.
- the imaging control unit 11 and the information processing unit 13 described with reference to FIG. 1 are configured with a module (for example, one die) for the integrated control unit 7600 shown in FIG. Integrated circuit module).
- the imaging control unit 11 and the information processing unit 13 described with reference to FIG. 1 may be realized by a plurality of control units of the vehicle control system 7000 illustrated in FIG.
- this indication can also take the following structures.
- a set of cameras mounted on a moving body and constituting a stereo camera system;
- a detection unit that detects a distance of an observation point in a detection range based on an image captured by the set of cameras;
- the one set of cameras is arranged on the side surface of the moving body and arranged in the vertical direction with respect to the road surface, and sequentially reads out pixel signals to be picked up in the vertical direction in units of pixels arranged in an array.
- Control device ⁇ 2> The imaging control device according to ⁇ 1>, wherein each of the one set of cameras reads out pixel signals in order from a column of pixels on the front side of the moving body, and in each column in the vertical direction.
- the one set of cameras sequentially reads out pixel signals in order from the column of pixels on the front side of the moving body, and sequentially from bottom to top in the vertical direction or from top to bottom in each column.
- the imaging control apparatus according to ⁇ 2>.
- ⁇ 4> The one set of cameras is arranged on each of the left and right side surfaces of the moving body and arranged in a direction perpendicular to the road surface, and the one set of cameras on the left and right of the moving body
- the imaging control device according to ⁇ 3>, in which pixel signals are sequentially read from a pixel row on the front side of the moving body in order, and for each row in a vertical direction from bottom to top or from top to bottom.
- the one set of cameras is arranged on each of the left and right side surfaces of the movable body and arranged side by side in a direction perpendicular to the road surface.
- the pixel signals are sequentially read out from the pixel columns on the front side of the moving body in order from the bottom to the top in the vertical direction, and the other set of left and right cameras is connected to the front of the moving body.
- the imaging control apparatus according to ⁇ 3> in which pixel signals are sequentially read from the top side in the vertical direction and from the top in the vertical direction to the down direction in each column.
- the set of cameras includes at least one of left and right side surfaces in front of the moving body, left and right side surfaces in the center of the moving body, and left and right side surfaces in the rear of the moving body.
- the imaging control device according to any one of ⁇ 1> to ⁇ 5>, wherein the imaging control devices are arranged side by side so as to generate parallax in a direction perpendicular to the road surface.
- ⁇ 7> When the one set of cameras is arranged in a vertical direction with respect to the road surface on each of the left and right side surfaces in front of the moving body or the left and right side surfaces behind the moving body.
- each of the cameras has a wider angle than a predetermined angle.
- the imaging control device according to any one of ⁇ 6>.
- ⁇ 9> The imaging control device according to any one of ⁇ 1> to ⁇ 8>, wherein the pixels of the camera are arranged in an array so that a horizontal width is longer than a vertical width. . ⁇ 10> The pixels of the camera are arranged in an array so that the ratio of the horizontal width to the vertical width is approximately 16: 9 or 4: 3. ⁇ 9>
- a control method for an imaging control apparatus including a detection unit that detects a distance of an observation point in a detection range based on an image captured by a pair of cameras that are mounted on a moving body and constitute a stereo camera system. And The set of cameras is arranged on the side surface of the moving body and arranged side by side in the vertical direction with respect to the road surface, and sequentially reads out pixel signals to be imaged in the vertical direction in units of pixels arranged in an array.
- a method for controlling an imaging control apparatus including: ⁇ 12> a set of cameras constituting a stereo camera system mounted on a moving body; A detection unit that detects a distance of an observation point in a detection range based on an image captured by the set of cameras; The one set of cameras is arranged on the side surface of the moving body and arranged in the vertical direction with respect to the road surface, and sequentially reads out the imaged pixel signals in the vertical direction in units of pixels arranged in an array.
- 1 in-vehicle system 10 camera system, 11 imaging control unit, 12 vehicle information sensor, 13 information processing unit, 14 display unit, 15 braking control unit, 16 braking device, 31 peripheral situation detection unit, 32 motion prediction unit, 33 collision prediction Unit, 34 HMI control unit, 100 vehicle, 111, 111A, 111B stereo camera system, 131, 131A, 131B, 132, 132A, 132B camera, 151, 151A, 151B stereo distance measurement unit, 152 integration unit, 171, 172 images Correction unit, 173 Stereo image processing unit
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Mechanical Engineering (AREA)
- Physics & Mathematics (AREA)
- Signal Processing (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- Electromagnetism (AREA)
- Measurement Of Optical Distance (AREA)
- Image Analysis (AREA)
- Studio Devices (AREA)
- Transforming Light Signals Into Electric Signals (AREA)
- Length Measuring Devices By Optical Means (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
1.実施の形態
2.変形例
3.応用例
<1-1.車載システムの構成例>
図1は、本技術を適用した車載システムの一実施の形態を示すブロック図である。
図2は、カメラシステム10を構成するステレオカメラシステムの配置例を示している。
ステレオカメラシステム111Aは、図3に示されるように、2台のカメラ131A,132Aを1組とした構成であり、車両100の前方左側面に垂直方向(すなわち縦方向)に並べて配置される。すなわち、カメラ131A,132Aは、その視差が高さ(垂直)方向に生じるように、基準面(路面101)と垂直な面内に配置される。
次に、図4のブロック図を参照して、撮像制御部11の詳細な構成について説明する。
次に、図5のブロック図を参照して、図4のステレオ距離計測部151の構成例について説明する。
次に、図6を参照して、ステレオカメラシステム111の座標系について説明する。ステレオカメラシステム111の座標系は図6に示されるように定義される。
ただし、α=θa-θbである。
・・・(1)
ρ・sinθ=ρa・sin(π-θa)=ρa・sinθa
・・・(2)
sin(θa-θb)=L/ρ・sinθa・sinθb/sinθ
・・・(3)
sinθa・sinθb≒sin2θ
・・・(4)
sin(θa-θb)≒L/ρ・sinθ
・・・(5)
次に、本開示のステレオカメラシステム111の画素信号の読み出し方法について説明するにあたって、行単位で順次垂直方向に、かつ、各行において水平方向に画素信号を読み出すカメラを用いたステレオカメラシステムについて説明する。
そこで、本開示のステレオカメラシステム111に属するカメラ131,132は、列単位で水平方向に順次読み出し、かつ、各列に対して垂直方向に画素信号を読み出す。
次に、図9を参照して、カメラ131,132の画素信号の具体的な読み出し方法について説明する。尚、図9においては、カメラ131における画素信号の読み出し方法を説明するものとするが、カメラ132においても同様である。また、カメラ131を構成する画素アレイは、水平方向×垂直方向が8個×6個の画素251により構成されるものとするが、これ以外の画素数で構成されていてもよい。
次に、図10を参照して、表示部14の配置例について説明する。
次に、図11のフローチャートを参照して、車載システム10により実行される周辺監視処理について説明する。この処理は、例えば、自車両を起動し、運転を開始するための操作が行われたとき、例えば、自車両のイグニッションスイッチ、パワースイッチ、又は、スタートスイッチ等がオンされたとき開始される。また、この処理は、例えば、運転を終了するための操作が行われたとき、例えば、自車両のイグニッションスイッチ、パワースイッチ、又は、スタートスイッチ等がオフされたとき終了する。
次に、図12のフローチャートを参照して、周辺画像データおよび距離データ取得処理について説明する。
以上においては、ステレオカメラシステム111が、車両100の前方側面部にカメラ131,132が垂直方向に視差が生じるように並べて配置される例について説明してきたが、車両100の車体の前方以外の位置に、カメラ131,132の視差が垂直方向に生じるようにステレオカメラシステム111が配置されてもよい。
本開示に係る技術は、様々な製品へ応用することができる。例えば、本開示に係る技術は、自動車、電気自動車、ハイブリッド電気自動車、自動二輪車、自転車、パーソナルモビリティ、飛行機、ドローン、船舶、ロボット、建設機械、農業機械(トラクター)などのいずれかの種類の移動体に搭載される装置として実現されてもよい。
<1> 移動体に搭載され、ステレオカメラシステムを構成する1組のカメラと、
前記1組のカメラにより撮像された画像に基づいて、検出範囲における観測点の距離を検出する検出部を含み、
前記1組のカメラは、前記移動体の側面で、かつ、路面に対して垂直方向に並べて配設され、撮像される画素信号を、アレイ状に配置された画素単位で垂直方向に順次読み出す
撮像制御装置。
<2> 前記1組のカメラは、それぞれ前記移動体の前方側の画素の列から順に、かつ、各列について、垂直方向に順に画素信号を読み出す
<1>に記載の撮像制御装置。
<3> 前記1組のカメラは、それぞれ前記移動体の前方側の画素の列から順に、かつ、各列について、垂直方向に下から上方向、または、上から下方向に順次画素信号を読み出す
<2>に記載の撮像制御装置。
<4> 前記1組のカメラは、前記移動体の左右の側面のそれぞれに、かつ、前記路面に対して垂直方向に並べて配設され、前記移動体の左右のそれぞれの前記1組のカメラが、前記移動体の前方側の画素の列から順に、かつ、各列について、垂直方向に下から上方向、または、上から下方向に順次画素信号を読み出す
<3>に記載の撮像制御装置。
<5> 前記1組のカメラは、前記移動体の左右の側面のそれぞれに、かつ、前記路面に対して垂直方向に並べて配設され、前記移動体の左右の一方の前記1組のカメラが、前記移動体の前方側の画素の列から順に、かつ、各列について、垂直方向に下から上方向に順次画素信号を読み出し、左右の他方の前記1組のカメラが、前記移動体の前方側の画素の列から順に、かつ、各列について、垂直方向に上から下方向に順次画素信号を読み出す
<3>に記載の撮像制御装置。
<6> 前記1組のカメラは、前記移動体の前方の左右のそれぞれの側面、前記移動体の中央の左右のそれぞれの側面、および前記移動体の後方の左右のそれぞれの側面の少なくともいずれかに、それぞれ前記路面に対して垂直方向に視差が生じるように並べて配設される
<1>乃至<5>のいずれかに記載の撮像制御装置。
<7> 前記1組のカメラが、前記移動体の前方の左右のそれぞれ側面、または、前記移動体の後方の左右のそれぞれの側面に、前記路面に対して垂直方向に並べて配設される場合、前記カメラは、それぞれ所定の角度よりも狭角度のカメラである
<6>に記載の撮像制御装置。
<8> 前記1組のカメラが、前記移動体の中央の左右のそれぞれの側面に、前記路面に対して垂直方向に並べて配設される場合、前記カメラは、それぞれ所定の角度よりも広角度のカメラである
<6>のいずれかに記載の撮像制御装置。
<9> 前記カメラの画素の配置は、水平方向の幅が、垂直方向の幅よりも長くなるように、アレイ状に配置される
<1>乃至<8>のいずれかに記載の撮像制御装置。
<10> 前記カメラの画素の配置は、前記水平方向の幅と、前記垂直方向の幅との比が、略16:9または略4:3となるようにアレイ状に配置される
<9>に記載の撮像制御装置。
<11> 移動体に搭載され、ステレオカメラシステムを構成する1組のカメラにより撮像された画像に基づいて、検出範囲における観測点の距離を検出する検出部を含む撮像制御装置の制御方法であって、
前記1組のカメラが、前記移動体の側面で、かつ、路面に対して垂直方向に並べて配設され、撮像される画素信号を、アレイ状に配置された画素単位で垂直方向に順次読み出す
ステップを含む撮像制御装置の制御方法。
<12> 移動体に搭載されたステレオカメラシステムを構成する1組のカメラと、
前記1組のカメラにより撮像された画像に基づいて、検出範囲における観測点の距離を検出する検出部とを有し、
前記1組のカメラは、前記移動体の側面で、かつ、路面に対して垂直方向に並べて配設され、撮像される画素信号を、アレイ状に配置された画素単位で垂直方向に順次読み出す撮像制御装置
を備えた移動体。
Claims (12)
- 移動体に搭載され、ステレオカメラシステムを構成する1組のカメラと、
前記1組のカメラにより撮像された画像に基づいて、検出範囲における観測点の距離を検出する検出部を含み、
前記1組のカメラは、前記移動体の側面で、かつ、路面に対して垂直方向に並べて配設され、撮像される画素信号を、アレイ状に配置された画素単位で垂直方向に順次読み出す
撮像制御装置。 - 前記1組のカメラは、それぞれ前記移動体の前方側の画素の列から順に、かつ、各列について、垂直方向に順に画素信号を読み出す
請求項1に記載の撮像制御装置。 - 前記1組のカメラは、それぞれ前記移動体の前方側の画素の列から順に、かつ、各列について、垂直方向に下から上方向、または、上から下方向に順次画素信号を読み出す
請求項2に記載の撮像制御装置。 - 前記1組のカメラは、前記移動体の左右の側面のそれぞれに、かつ、前記路面に対して垂直方向に並べて配設され、前記移動体の左右のそれぞれの前記1組のカメラが、前記移動体の前方側の画素の列から順に、かつ、各列について、垂直方向に下から上方向、または、上から下方向に順次画素信号を読み出す
請求項3に記載の撮像制御装置。 - 前記1組のカメラは、前記移動体の左右の側面のそれぞれに、かつ、前記路面に対して垂直方向に並べて配設され、前記移動体の左右の一方の前記1組のカメラが、前記移動体の前方側の画素の列から順に、かつ、各列について、垂直方向に下から上方向に順次画素信号を読み出し、左右の他方の前記1組のカメラが、前記移動体の前方側の画素の列から順に、かつ、各列について、垂直方向に上から下方向に順次画素信号を読み出す
請求項3に記載の撮像制御装置。 - 前記1組のカメラは、前記移動体の前方の左右のそれぞれの側面、前記移動体の中央の左右のそれぞれの側面、および前記移動体の後方の左右のそれぞれの側面の少なくともいずれかに、それぞれ前記路面に対して垂直方向に視差が生じるように並べて配設される
請求項1に記載の撮像制御装置。 - 前記1組のカメラが、前記移動体の前方の左右のそれぞれ側面、または、前記移動体の後方の左右のそれぞれの側面に、前記路面に対して垂直方向に並べて配設される場合、前記カメラは、それぞれ所定の角度よりも狭角度のカメラである
請求項6に記載の撮像制御装置。 - 前記1組のカメラが、前記移動体の中央の左右のそれぞれの側面に、前記路面に対して垂直方向に並べて配設される場合、前記カメラは、それぞれ所定の角度よりも広角度のカメラである
請求項6に記載の撮像制御装置。 - 前記カメラの画素の配置は、水平方向の幅が、垂直方向の幅よりも長くなるように、アレイ状に配置される
請求項1に記載の撮像制御装置。 - 前記カメラの画素の配置は、前記水平方向の幅と、前記垂直方向の幅との比が、略16:9または略4:3となるようにアレイ状に配置される
請求項9に記載の撮像制御装置。 - 移動体に搭載され、ステレオカメラシステムを構成する1組のカメラにより撮像された画像に基づいて、検出範囲における観測点の距離を検出する検出部を含む撮像制御装置の制御方法であって、
前記1組のカメラが、前記移動体の側面で、かつ、路面に対して垂直方向に並べて配設され、撮像される画素信号を、アレイ状に配置された画素単位で垂直方向に順次読み出す
ステップを含む撮像制御装置の制御方法。 - 移動体に搭載されたステレオカメラシステムを構成する1組のカメラと、
前記1組のカメラにより撮像された画像に基づいて、検出範囲における観測点の距離を検出する検出部とを有し、
前記1組のカメラは、前記移動体の側面で、かつ、路面に対して垂直方向に並べて配設され、撮像される画素信号を、アレイ状に配置された画素単位で垂直方向に順次読み出す撮像制御装置
を備えた移動体。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/496,647 US10864855B2 (en) | 2017-03-31 | 2018-03-16 | Imaging control apparatus, method for controlling imaging control apparatus, and mobile body |
JP2019509275A JP7196063B2 (ja) | 2017-03-31 | 2018-03-16 | 撮像制御装置、および撮像制御装置の制御方法、並びに移動体 |
CN201880020151.XA CN110574357B (zh) | 2017-03-31 | 2018-03-16 | 成像控制设备、用于控制成像控制设备的方法以及移动体 |
EP18777554.9A EP3606063A1 (en) | 2017-03-31 | 2018-03-16 | Imaging control device, control method for imaging control device, and mobile object |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2017070079 | 2017-03-31 | ||
JP2017-070079 | 2017-03-31 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2018180579A1 true WO2018180579A1 (ja) | 2018-10-04 |
Family
ID=63676954
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2018/010397 WO2018180579A1 (ja) | 2017-03-31 | 2018-03-16 | 撮像制御装置、および撮像制御装置の制御方法、並びに移動体 |
Country Status (5)
Country | Link |
---|---|
US (1) | US10864855B2 (ja) |
EP (1) | EP3606063A1 (ja) |
JP (1) | JP7196063B2 (ja) |
CN (1) | CN110574357B (ja) |
WO (1) | WO2018180579A1 (ja) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20200401819A1 (en) * | 2019-06-22 | 2020-12-24 | Erik Volkerink | Image based locationing |
WO2022163307A1 (ja) * | 2021-01-29 | 2022-08-04 | ソニーセミコンダクタソリューションズ株式会社 | 情報処理システム、情報処理方法、および情報処理装置 |
DE102020214591B4 (de) | 2019-11-27 | 2024-07-18 | Mitsubishi Electric Corporation | Fahrzeugnetzwerksystem |
DE112022005080T5 (de) | 2021-10-18 | 2024-09-05 | Hitachi Astemo, Ltd. | Fahrzeugmontiertes bildgebungssystem |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11025848B2 (en) * | 2018-08-31 | 2021-06-01 | Canon Kabushiki Kaisha | Photoelectric conversion device, imaging system, moving body, and stackable semiconductor device |
JP2020051942A (ja) * | 2018-09-27 | 2020-04-02 | 株式会社Subaru | 車両の走行環境検出装置及び走行制御システム |
US11910092B2 (en) * | 2020-10-01 | 2024-02-20 | Black Sesame Technologies Inc. | Panoramic look-around view generation method, in-vehicle device and in-vehicle system |
US12055632B2 (en) | 2020-10-13 | 2024-08-06 | Waymo Llc | LIDAR based stereo camera correction |
JP7554105B2 (ja) * | 2020-12-02 | 2024-09-19 | ヤマハ発動機株式会社 | 距離認識システムおよびその制御方法、船舶 |
CN113689552A (zh) * | 2021-08-27 | 2021-11-23 | 北京百度网讯科技有限公司 | 车载环视模型的调整方法、装置、电子设备和存储介质 |
CN114440821B (zh) * | 2022-02-08 | 2023-12-12 | 三一智矿科技有限公司 | 基于单目相机的测距方法及装置、介质、设备 |
JP2024125077A (ja) * | 2023-03-03 | 2024-09-13 | キヤノン株式会社 | 測距装置、撮像装置、移動体、測距方法およびプログラム |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS59201114A (ja) * | 1983-04-28 | 1984-11-14 | Nippon Denso Co Ltd | 車両用自動走行装置 |
JP2008009906A (ja) * | 2006-06-30 | 2008-01-17 | Matsushita Electric Ind Co Ltd | 固体撮像装置及び撮像システム |
JP2011205215A (ja) * | 2010-03-24 | 2011-10-13 | Fujifilm Corp | 撮像装置およびその制御方法 |
JP2013053962A (ja) * | 2011-09-05 | 2013-03-21 | Sharp Corp | カメラシステム、衝突予測システム、駐車ガイドシステム |
JP2013070177A (ja) | 2011-09-21 | 2013-04-18 | Toshiba Alpine Automotive Technology Corp | 車載用カメラ |
JP2016217944A (ja) * | 2015-05-22 | 2016-12-22 | シャープ株式会社 | 計測装置、および計測方法 |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4299036B2 (ja) * | 2003-03-31 | 2009-07-22 | シャープ株式会社 | 車両用監視装置 |
US8199975B2 (en) * | 2006-12-12 | 2012-06-12 | Cognex Corporation | System and method for side vision detection of obstacles for vehicles |
US9225965B2 (en) * | 2008-11-07 | 2015-12-29 | Telecom Italia S.P.A. | Method and system for producing multi-view 3D visual contents |
WO2012103193A1 (en) * | 2011-01-26 | 2012-08-02 | Magna Electronics Inc. | Rear vision system with trailer angle detection |
EP2523163B1 (en) | 2011-05-10 | 2019-10-16 | Harman Becker Automotive Systems GmbH | Method and program for calibrating a multicamera system |
ITTO20110414A1 (it) * | 2011-05-11 | 2012-11-12 | St Microelectronics Pvt Ltd | Procedimento ed apparecchiatura per elaborare segnali video, prodotto informatico e segnale codificato relativi |
US20160232415A1 (en) * | 2011-08-16 | 2016-08-11 | Israel L'Heureux | Detection detection of cell phone or mobile device use in motor vehicle |
CN102490673A (zh) * | 2011-12-13 | 2012-06-13 | 中科院微电子研究所昆山分所 | 基于车联网的汽车主动安全控制系统及其控制方法 |
US9446713B2 (en) * | 2012-09-26 | 2016-09-20 | Magna Electronics Inc. | Trailer angle detection system |
DE102016110903A1 (de) * | 2015-06-14 | 2016-12-15 | Facense Ltd. | Head-Mounted-Devices zur Messung physiologischer Reaktionen |
US20180283851A1 (en) * | 2015-09-01 | 2018-10-04 | The University Of Tokyo | Motion detection device and three-dimensional shape measurement device using same |
US20170150176A1 (en) * | 2015-11-25 | 2017-05-25 | Qualcomm Incorporated | Linear-model prediction with non-square prediction units in video coding |
-
2018
- 2018-03-16 EP EP18777554.9A patent/EP3606063A1/en not_active Ceased
- 2018-03-16 JP JP2019509275A patent/JP7196063B2/ja active Active
- 2018-03-16 US US16/496,647 patent/US10864855B2/en active Active
- 2018-03-16 CN CN201880020151.XA patent/CN110574357B/zh active Active
- 2018-03-16 WO PCT/JP2018/010397 patent/WO2018180579A1/ja active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS59201114A (ja) * | 1983-04-28 | 1984-11-14 | Nippon Denso Co Ltd | 車両用自動走行装置 |
JP2008009906A (ja) * | 2006-06-30 | 2008-01-17 | Matsushita Electric Ind Co Ltd | 固体撮像装置及び撮像システム |
JP2011205215A (ja) * | 2010-03-24 | 2011-10-13 | Fujifilm Corp | 撮像装置およびその制御方法 |
JP2013053962A (ja) * | 2011-09-05 | 2013-03-21 | Sharp Corp | カメラシステム、衝突予測システム、駐車ガイドシステム |
JP2013070177A (ja) | 2011-09-21 | 2013-04-18 | Toshiba Alpine Automotive Technology Corp | 車載用カメラ |
JP2016217944A (ja) * | 2015-05-22 | 2016-12-22 | シャープ株式会社 | 計測装置、および計測方法 |
Non-Patent Citations (1)
Title |
---|
See also references of EP3606063A4 |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20200401819A1 (en) * | 2019-06-22 | 2020-12-24 | Erik Volkerink | Image based locationing |
US11620832B2 (en) * | 2019-06-22 | 2023-04-04 | Hendrik J. Volkerink | Image based locationing |
DE102020214591B4 (de) | 2019-11-27 | 2024-07-18 | Mitsubishi Electric Corporation | Fahrzeugnetzwerksystem |
WO2022163307A1 (ja) * | 2021-01-29 | 2022-08-04 | ソニーセミコンダクタソリューションズ株式会社 | 情報処理システム、情報処理方法、および情報処理装置 |
DE112022005080T5 (de) | 2021-10-18 | 2024-09-05 | Hitachi Astemo, Ltd. | Fahrzeugmontiertes bildgebungssystem |
Also Published As
Publication number | Publication date |
---|---|
EP3606063A4 (en) | 2020-02-05 |
CN110574357A (zh) | 2019-12-13 |
EP3606063A1 (en) | 2020-02-05 |
US20200108774A1 (en) | 2020-04-09 |
JP7196063B2 (ja) | 2022-12-26 |
JPWO2018180579A1 (ja) | 2020-02-13 |
US10864855B2 (en) | 2020-12-15 |
CN110574357B (zh) | 2021-10-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7196063B2 (ja) | 撮像制御装置、および撮像制御装置の制御方法、並びに移動体 | |
US11076141B2 (en) | Image processing device, image processing method, and vehicle | |
JP6819681B2 (ja) | 撮像制御装置および方法、並びに車両 | |
JP6819680B2 (ja) | 撮像制御装置および方法、並びに車両 | |
JP6834964B2 (ja) | 画像処理装置、画像処理方法、およびプログラム | |
WO2017057044A1 (ja) | 情報処理装置及び情報処理方法 | |
JP6764573B2 (ja) | 画像処理装置、画像処理方法、およびプログラム | |
US11585898B2 (en) | Signal processing device, signal processing method, and program | |
WO2019163315A1 (ja) | 情報処理装置、撮像装置、及び撮像システム | |
US20200349367A1 (en) | Image processing device, image processing method, and program | |
JP6922169B2 (ja) | 情報処理装置および方法、車両、並びに情報処理システム | |
WO2019035391A1 (en) | SERVER, METHOD, NON-TRANSIENT COMPUTER-READABLE MEDIUM, AND SYSTEM | |
US20230186651A1 (en) | Control device, projection system, control method, and program | |
WO2022097470A1 (ja) | 信号処理装置、撮像装置及び信号処理方法 | |
WO2020255589A1 (ja) | 情報処理装置、情報処理方法及びプログラム | |
WO2020116204A1 (ja) | 情報処理装置、情報処理方法、プログラム、移動体制御装置、及び、移動体 | |
WO2020195969A1 (ja) | 情報処理装置、情報処理方法及びプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18777554 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2019509275 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2018777554 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2018777554 Country of ref document: EP Effective date: 20191031 |