WO2020001223A1 - 交通信号灯检测及智能驾驶方法和装置、车辆、电子设备 - Google Patents
交通信号灯检测及智能驾驶方法和装置、车辆、电子设备 Download PDFInfo
- Publication number
- WO2020001223A1 WO2020001223A1 PCT/CN2019/089062 CN2019089062W WO2020001223A1 WO 2020001223 A1 WO2020001223 A1 WO 2020001223A1 CN 2019089062 W CN2019089062 W CN 2019089062W WO 2020001223 A1 WO2020001223 A1 WO 2020001223A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- traffic
- traffic signal
- state
- light
- attributes
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 103
- 238000001514 detection method Methods 0.000 title claims abstract description 74
- 238000012549 training Methods 0.000 claims description 70
- 230000004044 response Effects 0.000 claims description 42
- 238000000605 extraction Methods 0.000 claims description 30
- 239000003086 colorant Substances 0.000 claims description 19
- 230000008569 process Effects 0.000 claims description 15
- 238000004590 computer program Methods 0.000 claims description 9
- 238000012545 processing Methods 0.000 claims description 9
- 238000004891 communication Methods 0.000 description 12
- 238000005516 engineering process Methods 0.000 description 8
- 238000013528 artificial neural network Methods 0.000 description 6
- 238000004422 calculation algorithm Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 6
- 230000003068 static effect Effects 0.000 description 6
- 238000003672 processing method Methods 0.000 description 4
- 230000035945 sensitivity Effects 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 238000013527 convolutional neural network Methods 0.000 description 2
- 230000010485 coping Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000007704 transition Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000000739 chaotic effect Effects 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000011664 signaling Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/09—Arrangements for giving variable traffic instructions
- G08G1/0962—Arrangements for giving variable traffic instructions having an indicator mounted inside the vehicle, e.g. giving voice messages
- G08G1/09623—Systems involving the acquisition of information from passive traffic signs by means mounted on the vehicle
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W30/00—Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
- B60W30/18—Propelling the vehicle
- B60W30/18009—Propelling the vehicle related to particular drive situations
- B60W30/18154—Approaching an intersection
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W30/00—Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
- B60W30/18—Propelling the vehicle
- B60W30/18009—Propelling the vehicle related to particular drive situations
- B60W30/18159—Traversing an intersection
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0212—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory
- G05D1/0221—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory involving a learning process
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/56—Extraction of image or video features relating to colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
- G06V20/584—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of vehicle lights or traffic lights
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/01—Detecting movement of traffic to be counted or controlled
- G08G1/04—Detecting movement of traffic to be counted or controlled using optical or ultrasonic detectors
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/09—Arrangements for giving variable traffic instructions
- G08G1/0962—Arrangements for giving variable traffic instructions having an indicator mounted inside the vehicle, e.g. giving voice messages
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2552/00—Input parameters relating to infrastructure
- B60W2552/53—Road markings, e.g. lane marker or crosswalk
Definitions
- the present disclosure relates to computer vision technology, and more particularly to a method and device for detecting traffic lights and intelligent driving, vehicles, and electronic equipment.
- Traffic light detection and status determination are important issues in the field of intelligent driving. Traffic lights are important traffic signals and play an irreplaceable role in modern transportation systems. Traffic light detection and status determination can instruct the vehicle to stop and advance during automatic driving to ensure the safe driving of the vehicle.
- the embodiments of the present disclosure provide a technology for detecting traffic lights and intelligent driving.
- the detection network includes a region-based full convolution network and a multi-task recognition network, including:
- a smart driving method including:
- a traffic light detection device including:
- a video stream acquiring unit configured to acquire a video stream including a traffic signal light
- An area determining unit configured to determine a candidate area of a traffic signal light in at least one frame of the video stream
- An attribute recognition unit is configured to determine at least two attributes of a traffic light in the image based on the candidate area.
- an intelligent driving device including:
- a video stream acquisition unit configured to acquire a video stream including a traffic signal based on an image acquisition device provided on a vehicle
- An area determining unit configured to determine a candidate area of a traffic signal light in at least one frame of the video stream
- An attribute recognition unit configured to determine at least two attributes of a traffic light in the image based on the candidate area
- a state determining unit configured to determine a state of the traffic signal light based on at least two attributes of the traffic signal light in the image
- An intelligent control unit is configured to intelligently control the vehicle according to a state of the traffic signal light.
- a vehicle including the traffic light detection device according to any one of the above or the intelligent driving device according to any one of the above.
- an electronic device including a processor, and the processor includes the traffic light detection device according to any one of the above or the intelligent driving device according to any one of the above.
- an electronic device including: a memory for storing executable instructions;
- a processor configured to communicate with the memory to execute the executable instructions to complete the operation of the traffic light detection method according to any one of the above, or complete the operation of the intelligent driving method according to any one of the above.
- a computer-readable storage medium for storing computer-readable instructions, and when the instructions are executed, the traffic signal detection method according to any one of the foregoing or any one of the above is performed.
- a computer program product including computer-readable code.
- the computer-readable code runs on a device, a processor in the device executes to implement any of the foregoing.
- a traffic signal detection and intelligent driving method and device, vehicle, and electronic device are provided to obtain a video stream including the traffic signal; determine a candidate area of the traffic signal in at least one frame of the video stream; based on The candidate area determines at least two attributes of the traffic signal in the image.
- FIG. 1 is a schematic flowchart of a traffic signal detection method provided by the present disclosure.
- FIG. 2 is a schematic structural diagram of a traffic light detection device provided by the present disclosure.
- FIG. 3 is a schematic flowchart of a smart driving method provided by the present disclosure.
- FIG. 4 is a schematic structural diagram of an intelligent driving device provided by the present disclosure.
- FIG. 5 is a schematic structural diagram of an electronic device suitable for implementing a terminal device or a server according to an embodiment of the present disclosure.
- Embodiments of the invention can be applied to a computer system / server, which can operate with many other general or special purpose computing system environments or configurations.
- Examples of well-known computing systems, environments, and / or configurations suitable for use with computer systems / servers include, but are not limited to: personal computer systems, server computer systems, thin clients, thick clients, handheld or laptop devices, based on Microprocessor systems, set-top boxes, programmable consumer electronics, network personal computers, small computer systems, mainframe computer systems, and distributed cloud computing technology environments including any of the above, and so on.
- a computer system / server may be described in the general context of computer system executable instructions, such as program modules, executed by a computer system.
- program modules may include routines, programs, target programs, components, logic, data structures, and so on, which perform specific tasks or implement specific abstract data types.
- the computer system / server can be implemented in a distributed cloud computing environment. In a distributed cloud computing environment, tasks are performed by remote processing devices linked through a communication network. In a distributed cloud computing environment, program modules may be located on a local or remote computing system storage medium including a storage device.
- FIG. 1 is a schematic flowchart of a traffic signal detection method provided by the present disclosure.
- the method can be executed by any electronic device, such as a terminal device, a server, a mobile device, a vehicle-mounted device, and so on.
- the method in this embodiment includes:
- Step 110 Obtain a video stream including a traffic signal.
- the identification of traffic lights is usually based on the on-board video recorded during the vehicle's travel, and the on-board video is parsed to obtain a video stream including at least one frame of image.
- the video stream can be installed on the vehicle.
- the camera device captures the video of the vehicle's forward direction or surrounding environment. If there are traffic lights in the vehicle's forward direction or surrounding environment, it will be captured by the camera device.
- the captured video stream is the video stream including the traffic signal light.
- the images in the video stream may include traffic signals in each frame of images, or at least one frame of images may include traffic signals.
- the step 110 may be executed by the processor calling a corresponding instruction stored in the memory, or may be executed by the video stream obtaining unit 21 executed by the processor.
- Step 120 Determine a candidate area of a traffic light in at least one frame of the video stream.
- a candidate region is determined from an image including a traffic signal in the video stream, and the candidate region refers to a region that may include a traffic signal in the image.
- the detection of the area of the traffic signal can be based on neural networks or other types of detection models.
- a region-based full convolutional network is used to determine candidate regions of traffic lights in at least one frame of image of the video stream.
- Region-based, fully convolutional networks are used to detect signal images to obtain candidate regions that may include traffic lights.
- R-FCN can be regarded as a fast convolutional neural network (Faster Regions (with CNN, Faster RCNN), the detection speed is faster than Faster RCNN.
- the step 120 may be executed by the processor calling a corresponding instruction stored in the memory, or may be executed by the area determining unit 22 executed by the processor.
- Step 130 Determine at least two attributes of traffic lights in the image based on the candidate area.
- the attributes of traffic lights are used to describe traffic lights, which can be defined according to actual needs.
- the attributes of traffic lights can be used to describe the location or location of traffic lights, such as red, green, and yellow.
- Etc. attributes, which are used to describe the shape of traffic lights (such as circles, straight arrows, polyline arrows, etc.), and other attributes used to describe other aspects of traffic lights.
- At least two attributes of the traffic signal light include any two or more of the following: location area, color, and shape.
- the color of the traffic signal includes three colors of red, yellow, and green
- the shape includes an arrow shape, a circle, or other shapes.
- the signals may not be accurately identified. Therefore, in this embodiment, by identifying at least two of a location area, a color, and a shape, for example, when determining a location area and a color of a traffic signal, it is possible to determine a current traffic signal position in the image (corresponding to which direction of the vehicle ),
- the state of the traffic light display can be determined by color (red, green or yellow corresponding to different states), and the assisted driving or automatic driving can be realized by identifying the different states of the traffic light; when determining the location area and shape of the traffic light, You can determine where the current traffic light is in the image (corresponding to which direction of the vehicle), and determine the status of the traffic light by its shape (for example, arrows pointing in different directions indicate different states, or human figures in different shapes indicate different states );
- the color of traffic light includes three colors of red, yellow, and green
- the step 130 may be executed by the processor calling a corresponding instruction stored in the memory, or may be executed by the attribute recognition unit 23 executed by the processor.
- a method for detecting a traffic signal based on the above embodiments of the present disclosure is to obtain a video stream including a traffic signal; determine a candidate region of the traffic signal in at least one frame of the video stream; and determine at least two of the traffic signal in the image based on the candidate region.
- This kind of attribute can realize the recognition of various kinds of information of the traffic signal by obtaining at least two attributes of the traffic signal, reduce the recognition time, and improve the accuracy of the traffic signal recognition.
- operation 130 may include:
- a multi-task recognition network is used to determine at least two attributes of traffic lights in an image based on candidate regions.
- At least two attributes of traffic lights are identified through a network. Compared with the case where at least two attributes are identified based on at least two networks, respectively, the size of the network is reduced and the efficiency of attribute recognition of traffic lights is improved. .
- a multi-task recognition network is used to identify candidate areas that may include traffic lights.
- the recognition process may include feature extraction and attribute recognition.
- the multi-task recognition network may include feature extraction branches, and The feature extraction branch is connected to at least two task branches, and different task branches are used to determine different kinds of attributes of the traffic light.
- Each attribute recognition task requires feature extraction for candidate regions.
- the feature extraction branches are connected to at least two task branches, so that the feature extraction operations of at least two task branches are combined in the same feature extraction branch.
- Feature extraction is required for at least two task branches, which reduces the structure of the multi-task recognition network and accelerates the speed of attribute recognition.
- the process of obtaining at least two attributes may include:
- the candidate features are processed based on at least two task branches respectively to obtain at least two attributes of traffic lights in the image.
- the feature extraction branch may include at least one convolution layer, using the candidate region as an input image, and performing feature extraction on the candidate region through the feature extraction branch to obtain candidate features (feature map or feature vector) of the candidate region.
- candidate features can be obtained by at least two task branches, the position and color of the traffic signal, or the position and shape of the traffic signal, or the color and shape of the traffic signal.
- the multi-task branch is used to obtain The color, position and shape of the signal light; while checking the position of the signal light, the current state of the signal light can be identified by the color of the signal light, which can be well applied in the field of automatic driving. The accuracy of the signal light recognition can be improved by identifying the shape of the signal light .
- At least two task branches include, but are not limited to, a detection branch, an identification branch, and a classification branch;
- the candidate features are processed based on at least two task branches respectively to obtain at least two attributes of traffic lights in the image, including:
- the position of the candidate feature is detected by the detection branch to determine the location area of the traffic signal;
- Color classification of candidate features through classification branches, determining the color of the area where the traffic signal is located, and determining the color of the traffic signal;
- the shape of the candidate feature is identified through the recognition branch, the shape of the area where the traffic signal is located is determined, and the shape of the traffic signal is determined.
- This embodiment can simultaneously identify any two or three attributes of the location area, color, and shape of the traffic signal through different branches, saving time for multi-task recognition, reducing the size of the detection network, and enabling multi-task recognition.
- the network is faster in training and application process, and if the location area of the traffic signal is obtained first, the color and shape of the traffic signal can be obtained faster; because the color of the traffic signal is usually only three (red, green and yellow), therefore, for the Color recognition can be implemented using trained classification branches (other network layers except convolution layers in ordinary multi-task recognition networks).
- the method may further include:
- the difference between consecutive frames of the video stream may be small, and the location of traffic lights is identified based only on the candidate area of traffic lights in at least one frame of image. It is possible to identify the location area in consecutive frames as the same location area. As a result, the identified location area is inaccurate.
- key points are identified in the image, the location area of the traffic signal in the image is determined based on the key point, and the traffic signal obtained by the multi-task recognition network is adjusted based on the location area of the key point. Location, improving the accuracy of location area recognition.
- the key point identification and / or tracking may be implemented based on any one of the existing technologies that can realize key point identification and / or tracking.
- the tracking of the key points of the traffic lights in the video stream is performed by using a static key point tracking technology to obtain an area where the key points of the traffic lights in the video stream may exist.
- the location area of the traffic signal lights obtained through the detection branch can easily cause some frames to be missed due to the small gap between consecutive images and the selection of the threshold. Based on the static keypoint tracking technology, the detection effect of the vehicle network on the vehicle video is improved.
- the characteristic points of an image can be simply understood as the more prominent points in the image, such as corner points, bright points in darker areas, and so on.
- feature points in the video image are detected and described (Oriented, FAST, Rotated, Brief, ORB) feature points:
- the definition of the ORB feature points is based on the gray value of the image around the feature points. During detection, a circle around the candidate feature points is considered. If there are enough pixels in the area around the candidate point and the difference between the gray value of the candidate feature point reaches a preset value, the candidate point is considered as a key feature point. In this embodiment, the key points of the traffic signal are identified. Therefore, the key points are the key points of the traffic signal.
- the key points of the traffic signal can be used to implement static tracking of the traffic signal in the video stream. Not only does it occupy one pixel, that is, the key point of the traffic signal obtained in this embodiment includes at least one pixel, it can be understood that the key point of the traffic signal corresponds to a location area.
- tracking the key points of the traffic lights in the video stream includes:
- the two consecutive frames referred to in this embodiment can be two consecutive acquisition frames in the video stream, or two consecutive detection frames in the video stream (because the video stream can be detected frame by frame or can be sampled and detected, so The meanings of both detection frames and acquisition frames are not exactly the same); by associating the key points of traffic lights in multiple consecutive frames of video in the video stream, the key points of traffic lights can be tracked in the video stream, Based on the tracking result, the position and area of at least one frame of image in the video stream can be adjusted.
- the tracking of the key points of the traffic signal in the video stream can be achieved based on the Hamming distance, the Euclidean distance, the joint Bayesian distance, or the cosine distance between the key points of the traffic signal. Limitation is based on the distance between key points of a traffic light.
- the Hamming distance is used in data transmission error control coding.
- the Hamming distance is a concept that represents the number of different bits corresponding to two (same length) words, and performs an exclusive OR operation on the two character strings, and counts them. The result is a number of 1, then this number is the Hamming distance, and the Hamming distance between two images is the number of different data bits between the two images. Based on the Hamming distance between the key points of at least one traffic signal in two frames of signal images, we can know the distance that the signal lights move between the two signal images, and the key points of traffic signals can be tracked.
- tracking the key points of the traffic signal in the video stream based on the distance between the key points of the traffic signal includes:
- the key points of the traffic signal are tracked in the video stream.
- Traffic lights usually do not appear individually, and because traffic lights cannot be represented by a key point in the image, at least one key point of the traffic light is included in the image. For different traffic lights (for example, the Traffic lights, left turn traffic lights) need to be tracked separately. This embodiment overcomes the problem of chaotic tracking of different traffic lights by tracking in consecutive frames based on the key points of the same traffic lights.
- determining a location area of a key point of the same traffic signal in two consecutive frames of images may be determined based on a smaller value (for example, a minimum value) of a Hamming distance between the key points of at least one traffic signal.
- the idea of the Brute Force algorithm is to match the first character of the target string S with the first character of the pattern string T. If they are equal, continue to compare the first character of S. Two characters and the second character of T; if they are not equal, then compare the second character of S and the first character of T, and compare them in turn until the final match is obtained.
- the BruteForce algorithm is a brute force Force algorithm.
- adjusting the position area of the signal light based on the tracking result includes:
- the position area of the signal light After adjusting the position area of the signal light based on the tracking result, the position area of the signal light is more stable and more suitable for application in a video scene.
- the position area corresponding to the key point of the traffic signal light in at least one frame of the video stream can be determined based on the tracking result.
- the overlapping part between the position area in the tracking result and the position area of the signal light is in the position area of the signal light If the ratio exceeds the set ratio, it can be determined that the tracking result coincides with the location area of the signal light; otherwise, it is determined that the tracking result does not coincide with the location area of the signal light.
- adjusting the position area of the signal light based on the comparison result includes:
- the location area corresponding to the key point of the traffic signal is used to replace the location area of the signal light.
- the comparison result of whether the location area corresponding to the key point of the traffic signal and the location area of the signal light in the signal image are compared can include the following three situations:
- the location area corresponding to the key point of the traffic signal and the location area of the signal light match (that is, coincidence), that is, the key point location area of the matching traffic signal in the two frames before and after the movement is the same as the detected location area of the signal light, no correction is required; If the location area of the key point of the traffic signal and the location area of the detected signal light are roughly matched, then based on the offset of the location area of the key point of the traffic signal in the previous and subsequent frames, the width and height of the detected signal light remain unchanged.
- the method may further include:
- the yellow light in a traffic signal is only a transition state between red and green lights, so the duration of its existence is shorter than that of red and green lights.
- the detection framework based on R-FCN only inputs a limited number of images at a time, and the number of yellow lights in the image is very small compared to red and green lights. Therefore, it is impossible to effectively train the detection network and improve the sensitivity of the model to yellow lights. Therefore, the present disclosure obtains simultaneous recognition of the position, color, and / or shape of a signal light by training a region-based full convolutional network and a multi-task recognition network.
- the detection network may further include:
- the classification network is trained based on the new training image set; the classification network is used to classify the training images based on the color of the traffic lights.
- the classification network is obtained by removing a candidate regional network (RPN) and a proposal layer from a detection network in the prior art.
- the classification network may include a multi-task identification network. Feature extraction branch and classification branch; training the classification network based on a new training image set based on a preset scale alone can improve the accuracy of the classification network's color classification of traffic lights.
- the training image set of the training network is acquired through acquisition, and the R-FCN region-based full convolution network is trained with the acquired training image set; the number of traffic lights and yellow lights in the acquired training image set is adjusted, optionally, preset The number of traffic lights of different colors in the ratio is the same or the number difference is less than the allowable threshold;
- the color of traffic lights includes red, yellow and green.
- the ratio of the three colors of red, yellow, and green is preset to the same ratio (such as: red: yellow: green is 1: 1: 1), or the number of red, yellow, and green colors is controlled to be less than the allowable threshold, so that the three The color ratio is close to 1: 1: 1.
- a new training image set can be constructed by extracting the training signals with corresponding colors from the training image set; or the yellow light images in the training image set can be repeatedly called so that the number of yellow light images is the same as that of the red and green light images. The number is in accordance with the preset ratio), and the classification network is trained with the adjusted new training image set, which overcomes the shortcoming that the number of yellow light images is much smaller than the traffic light image data, and improves the classification network's recognition accuracy of yellow lights.
- the method before adjusting the parameters of the region-based full convolutional network and the multi-task recognition network based on the training image set, the method further includes:
- some or all parameters in the multi-task recognition network may be initialized based on the parameters of the trained classification network, for example, the feature extraction branch and classification branch in the multi-task recognition network are initialized with the parameters of the trained classification network;
- the parameters may include, for example, the size of the convolution kernel, the weight of the convolution connection, and the like.
- the region-based full convolutional network and multi-task recognition network are trained with the initial training image set.
- the parameter pairs in the trained classification network are used for detection. Some parameters in the network are initialized.
- the feature extraction branch and classification branch obtained at this time have a good effect on the color classification of traffic lights and improve the accuracy of yellow light classification.
- the disclosed traffic signal detection method can be applied in the fields of intelligent driving, high-precision maps, and the like;
- Car video can be used as input to output the position and status of traffic lights to assist the vehicle's safe driving.
- It can also be used to build high-precision maps to detect traffic light locations.
- the method further includes:
- This embodiment automatically recognizes at least two attributes of the traffic signal and obtains the state of the traffic signal in the video stream, eliminating the need for the driver to distractively observe the traffic signal during the driving process, improving the safety of the vehicle and reducing the risk of human error. Traffic danger caused by mistake.
- the intelligent driving control includes: sending prompt information or warning information, and / or controlling a driving state of the vehicle according to a state of a traffic signal light.
- Identifying at least two attributes of traffic lights can provide a basis for intelligent driving.
- Intelligent driving includes autonomous driving and assisted driving.
- autonomous driving the driving state of the vehicle is controlled according to the state of the traffic lights (such as: parking, deceleration, steering, etc.)
- prompt information or warning information to inform the driver of the current status of traffic lights; in the case of assisted driving, usually only the prompt information or warning information is issued, the authority to control the vehicle still belongs to the driver, and the driver according to the prompt The information or warning information controls the vehicle accordingly.
- the method further includes: storing attributes, states and corresponding images of traffic lights.
- This embodiment acquires more information (attributes, states, and corresponding images) of the traffic signals by storing the attributes, states, and corresponding images of the traffic signals, and provides more operation basis for intelligent driving. It is also possible to establish a high-precision map based on the time and location corresponding to the stored traffic signal, and determine the location of the traffic lights in the high-precision map based on the image corresponding to the stored traffic signal.
- the state of the traffic signal light includes, but is not limited to, a traffic permission state, a traffic prohibition state, or a waiting state;
- the state of the traffic signal is a traffic-permitting state.
- the state of the traffic signal is a no-traffic state.
- the state of the traffic signal is a waiting state.
- the colors of traffic lights include red, green, and yellow, and different colors correspond to different traffic conditions.
- Red means that vehicles and / or pedestrians are prohibited
- green means that vehicles and / or pedestrians are allowed
- yellow means that vehicles and / Or the pedestrian pass needs to pause and wait
- the auxiliary color can also include shapes such as traffic signals, for example: a plus sign shape (an optional first preset shape) indicates that traffic is allowed, and a fork shape (an optional first The second preset shape) indicates that traffic is prohibited, the minus shape (an optional third preset shape) indicates a waiting state, and the like.
- performing intelligent driving control on the vehicle according to the state of the traffic signal light includes:
- control the vehicle In response to the state of the traffic light being a traffic permitted state, control the vehicle to perform one or more operations such as starting, maintaining the driving state, decelerating, turning, turning on the turn signal, turning on the brake light, and other controls needed to control the traffic of the vehicle. ;
- the color of the traffic signal is green and the shape is a left-pointing arrow
- the state of operation can achieve safer intelligent driving, improve driving safety, and reduce potential safety hazards caused by human error.
- the foregoing program may be stored in a computer-readable storage medium.
- the program is executed, the program is executed.
- the method includes the steps of the foregoing method embodiment.
- the foregoing storage medium includes: a ROM, a RAM, a magnetic disk, or an optical disk, and other media that can store program codes.
- FIG. 2 is a schematic structural diagram of an embodiment of a traffic light detection device according to the present disclosure.
- the traffic signal detection device of this embodiment may be used to implement the foregoing embodiments of the traffic signal detection methods of the present disclosure.
- the apparatus of this embodiment includes:
- the video stream acquiring unit 21 is configured to acquire a video stream including a traffic signal.
- the identification of traffic lights is usually based on the on-board video recorded during the vehicle's travel, and the on-board video is parsed to obtain a video stream including at least one frame of image.
- the video stream can be installed on the vehicle.
- the camera device captures the video of the vehicle's forward direction or surrounding environment. If there are traffic lights in the vehicle's forward direction or surrounding environment, it will be captured by the camera device.
- the captured video stream is the video stream including the traffic signal light.
- the images in the video stream may include traffic signals in each frame of images, or at least one frame of images may include traffic signals.
- the area determining unit 22 is configured to determine a candidate area of a traffic light in at least one frame of an image of a video stream.
- a candidate region is determined from an image including a traffic signal in the video stream, and the candidate region refers to a region that may include a traffic signal in the image.
- the detection of the area of the traffic signal can be based on neural networks or other types of detection models.
- a region-based full convolutional network is used to determine candidate regions of traffic lights in at least one frame of image of the video stream.
- the region-based full convolutional network (R-FCN) is used to detect the signal image to obtain candidate regions that may include traffic lights.
- R-FCN can be regarded as an improved version of Faster RCNN, and the detection speed is faster than Faster RCNN.
- the attribute recognition unit 23 is configured to determine at least two attributes of the traffic signal light in the image based on the candidate region.
- the attributes of traffic lights are used to describe traffic lights, which can be defined according to actual needs.
- the attributes of traffic lights can be used to describe the location or location of traffic lights, such as red, green, and yellow.
- Etc. attributes, which are used to describe the shape of traffic lights (such as circles, straight arrows, polyline arrows, etc.), and other attributes used to describe other aspects of traffic lights.
- various types of information of a traffic signal are recognized by obtaining at least two attributes of the traffic signal, which reduces the recognition time and improves the accuracy of traffic signal recognition.
- At least two attributes of the traffic signal light include any two or more of the following: location area, color, and shape.
- the determination of at least two attributes of a traffic light may be based on a neural network or other type of recognition model.
- the attribute recognition unit 23 is configured to use a multi-task recognition network to determine at least two attributes of a traffic light in an image based on a candidate region.
- At least two attributes of traffic lights are identified through a network. Compared with the case where at least two attributes are identified based on at least two networks, respectively, the size of the network is reduced and the efficiency of attribute recognition of traffic lights is improved. .
- the multi-task recognition network includes a feature extraction branch and at least two task branches respectively connected to the feature extraction branch, and different task branches are used to determine different types of attributes of traffic lights;
- the attribute recognition unit 23 includes:
- a feature extraction module configured to perform feature extraction on the candidate region based on the feature extraction branch to obtain candidate features
- a branch attribute module is configured to process candidate features based on at least two task branches, respectively, to obtain at least two attributes of traffic lights in an image.
- At least two task branches include, but are not limited to, a detection branch, an identification branch, and a classification branch;
- the branch attribute module is used to detect the position of candidate features through the detection branch to determine the location area of the traffic signal; to classify the candidate features by color classification to determine the color of the location area of the traffic signal and determine the color of the traffic signal;
- the branch performs shape recognition on candidate features, determines the shape of the area where the traffic signal is located, and determines the shape of the traffic signal.
- the method further includes:
- a key point determining unit configured to identify key points of at least one frame of an image in a video stream, and determine key points of a traffic signal light in the image
- Key point tracking unit which is used to track the key points of the traffic lights in the video stream to obtain the tracking results
- a position adjusting unit is configured to adjust a position area of a traffic signal based on a tracking result.
- the differences between consecutive frames of the video stream may be small.
- the traffic signal position recognition is based on the candidate signal traffic region in each frame of the image. It is possible to identify the location regions in consecutive frames as the same location region. As a result, the identified location area is inaccurate.
- keypoints are identified in the image, and the location area of the traffic signal in the image is determined based on the keypoint. Based on the location area of the keypoint, the Location, improving the accuracy of location area recognition.
- the key point identification and / or tracking may be implemented based on any one of the existing technologies that can realize key point identification and / or tracking.
- the tracking of the key points of the traffic lights in the video stream is performed by using a static key point tracking technology to obtain an area where the key points of the traffic lights in the video stream may exist.
- the key point tracking unit is configured to track the key points of the traffic signal in the video stream based on the distance between the key points of the traffic signal in two consecutive frames of images;
- the two consecutive frames referred to in this embodiment may be two consecutive acquisition frames in the video stream, or two consecutive detection frames in the video stream.
- the meanings of both detection frames and acquisition frames are not exactly the same); by associating the key points of traffic lights in multiple consecutive frames of video in the video stream, the key points of traffic lights can be tracked in the video stream, Based on the tracking results, the position and area of each frame of the video stream can be adjusted.
- the tracking of the key points of the traffic signal in the video stream can be achieved based on the Hamming distance, the Euclidean distance, the joint Bayesian distance, or the cosine distance between the key points of the traffic signal. Limitation is based on the distance between key points of a traffic light.
- the key point tracking unit tracks the key points of the traffic signal in the video stream based on the distance between the key points of the traffic signal, it is used to determine two consecutive frames of images based on the distance between the key points of the traffic signal.
- the location area of the key points of the same traffic signal in the video; according to the location area of the key points of the same traffic signal in two consecutive frames, the key points of the traffic signal are tracked in the video stream.
- the position adjustment unit is configured to compare whether the tracking result coincides with the position area of the signal light to obtain a comparison result; and adjust the position area of the signal light based on the comparison result.
- the position area of the signal light After adjusting the position area of the signal light based on the tracking result, the position area of the signal light is more stable and more suitable for application in a video scene.
- the position area corresponding to the key point of the traffic signal light in at least one frame of the video stream can be determined based on the tracking result.
- the overlapping part between the position area in the tracking result and the position area of the signal light is in the position area of the signal light If the ratio exceeds the set ratio, it can be determined that the tracking result coincides with the location area of the signal light; otherwise, it is determined that the tracking result does not coincide with the location area of the signal light.
- the position adjustment unit adjusts the position area of the signal light based on the comparison result, the position area corresponding to the key point of the traffic signal does not coincide with the position area of the signal light, and the position area corresponding to the key point of the traffic signal is not coincident Replace the location area of the semaphore.
- it may further include:
- a pre-training unit configured to train a region-based full convolutional network based on the acquired training image set, where the training image set includes multiple training images with labeled attributes;
- a training unit for adjusting parameters in a region-based full convolutional network and a multi-task recognition network based on a training image set.
- the yellow light in a traffic light is only a transition state between red and green lights, so it exists for a shorter period of time than red and green lights.
- the detection framework based on R-FCN only inputs a limited number of images at a time, and the number of yellow lights in the image is very small compared to red and green lights. Therefore, it is impossible to effectively train the detection network and improve the sensitivity of the model to yellow lights. Therefore, the present disclosure obtains simultaneous recognition of the position, color, and / or shape of a signal light by training a region-based full convolutional network and a multi-task recognition network.
- the pre-training unit and the training unit may further include:
- a classification training unit is used to obtain a new training image set whose color proportion of traffic lights is in accordance with a preset ratio based on the training image set; to train a classification network based on the new training image set; the classification network is used to classify the training images based on the color of the traffic signal.
- the number of traffic lights of different colors in the preset ratio is the same or the number difference is less than the allowable threshold
- the color of traffic lights includes red, yellow and green.
- the ratio of the three colors of red, yellow, and green is preset to the same ratio (such as: red: yellow: green is 1: 1: 1), or the number of red, yellow, and green colors is controlled to be less than the allowable threshold, so that the three The color ratio is close to 1: 1: 1.
- a new training image set can be constructed by extracting the training signals with corresponding colors from the training image set; or the yellow light images in the training image set can be repeatedly called so that the number of yellow light images is the same as that of the red and green light images. The number is in accordance with the preset ratio), and the classification network is trained with the adjusted new training image set, which overcomes the shortcoming that the number of yellow light images is much smaller than the traffic light image data, and improves the classification network's recognition accuracy of yellow lights.
- the method may further include:
- An initialization unit is configured to initialize at least some parameters in the multi-task recognition network based on the parameters of the trained classification network.
- the apparatus in this embodiment may further include:
- a state determining unit configured to determine a state of a traffic signal based on at least two attributes of the traffic signal in the image
- the intelligent control unit is used for intelligent driving control of the vehicle according to the state of the traffic signal light.
- This embodiment automatically recognizes at least two attributes of the traffic signal and obtains the state of the traffic signal in the video stream, eliminating the need for the driver to distractively observe the traffic signal during the driving process, improving the safety of the vehicle and reducing the risk of human error. Traffic danger caused by mistake.
- the intelligent driving control includes: sending prompt information or warning information, and / or controlling a driving state of the vehicle according to a state of a traffic signal light.
- it further includes:
- the storage unit is configured to store attributes, states and corresponding images of traffic lights.
- the state of the traffic signal light includes, but is not limited to, a traffic permission state, a traffic prohibition state, or a waiting state;
- a state determining unit configured to determine that the state of the traffic signal is a traffic permitted state in response to the color of the traffic signal being green and / or the shape being a first preset shape
- the state of the traffic signal is a waiting state.
- the intelligent control unit is configured to control the vehicle to perform one or more operations of starting, maintaining a driving state, decelerating, turning, turning on a turn signal, and turning on a brake light in response to a state of a traffic signal light being an allowed traffic state;
- one or more operations of controlling the vehicle to stop, decelerate, and turn on the brake light are controlled.
- FIG. 3 is a flowchart of an embodiment of a smart driving method according to the present disclosure. As shown in FIG. 3, the method in this embodiment includes:
- Step 310 Obtain a video stream including a traffic signal based on an image acquisition device provided on the vehicle.
- the on-board video is parsed to obtain a video stream including at least one frame of the image, for example, the forward or surrounding of the vehicle can be captured by a camera device installed on the vehicle
- the video of the environment if there are traffic lights in the forward direction of the vehicle or in the surrounding environment, will be captured by the camera device, and the captured video stream is a video stream including traffic lights.
- the images in the video stream may include traffic signals in each frame of images, or at least one frame of images may include traffic signals.
- this step 310 may be executed by the processor calling a corresponding instruction stored in the memory, or may be executed by the video stream obtaining unit 21 executed by the processor.
- Step 320 Determine a candidate area of a traffic signal in at least one frame of the video stream.
- step 320 may be executed by the processor by calling a corresponding instruction stored in the memory, or may be executed by the area determining unit 22 executed by the processor.
- Step 330 Determine at least two attributes of a traffic light in the image based on the candidate area.
- the attributes of traffic lights are used to describe traffic lights, which can be defined according to actual needs.
- the attributes of traffic lights can be used to describe the location or location of traffic lights, such as red, green, and yellow.
- Etc. attributes, used to describe the shape of traffic lights (such as circles, straight arrows, polyline arrows, etc.) attributes, and other attributes used to describe other aspects of traffic lights.
- At least two attributes of the traffic signal light include any two or more of the following: location area, color, and shape.
- the color of the traffic signal includes three colors of red, yellow, and green
- the shape includes an arrow shape, a circle, or other shapes.
- the signals may not be accurately identified. Therefore, in this embodiment, by identifying at least two of a location area, a color, and a shape, for example, when determining a location area and a color of a traffic signal, it is possible to determine a current traffic signal position in the image (corresponding to which direction of the vehicle ),
- the state of the traffic light display can be determined by color (red, green or yellow corresponding to different states), and the assisted driving or automatic driving can be realized by identifying the different states of the traffic light; when determining the location area and shape of the traffic light, You can determine where the current traffic light is in the image (corresponding to which direction of the vehicle), and determine the status of the traffic light by its shape (for example, arrows pointing in different directions indicate different states, or human figures in different shapes indicate different states );
- the color of traffic light includes three colors of red, yellow, and green
- step 330 may be executed by the processor by calling a corresponding instruction stored in the memory, or may be executed by the attribute recognition unit 23 executed by the processor.
- Step 340 Determine the state of the traffic signal based on at least two attributes of the traffic signal in the image.
- Existing image processing methods usually can only deal with one task (for example, one of location recognition or color classification), and the traffic light includes information such as location area, color, and shape.
- the status of the traffic light needs to be determined It is not only necessary to determine the location area of traffic signals, but also at least the color or shape. Therefore, if the general image processing method is applied, at least two neural networks are required to process the video stream, and the processing results need to be synthesized in order to Determine the current state of the traffic signal; this embodiment simultaneously obtains at least two attributes of the traffic signal, determines the state of the traffic signal with at least two attributes, and quickly and accurately identifies the state of the traffic signal.
- this step 340 may be executed by the processor calling a corresponding instruction stored in the memory, or may be executed by the state determination unit 44 executed by the processor.
- Step 350 Perform intelligent driving control on the vehicle according to the state of the traffic signal light.
- step 350 may be executed by the processor by calling corresponding instructions stored in the memory, or may be executed by the intelligent control unit 45 executed by the processor.
- an image acquisition device on a vehicle can obtain a video stream in real time, and realize the real-time identification of the attributes of traffic lights to determine the status of the traffic lights. Based on the status of the traffic lights, intelligent driving is achieved without the need for the driver to distract from observation during the driving process. Traffic lights reduce the hidden dangers of traffic safety, and to a certain extent reduce the danger of traffic caused by human error. Intelligent driving can include assisted driving and autonomous driving. Generally, assisted driving uses warning lights for warning and automatic driving uses signals for driving control.
- the intelligent driving control includes: sending prompt information or warning information, and / or controlling a driving state of the vehicle according to a state of a traffic signal light.
- Identifying at least two attributes of traffic lights can provide a basis for intelligent driving.
- Intelligent driving includes autonomous driving and assisted driving.
- autonomous driving the driving state of the vehicle is controlled according to the state of the traffic lights (such as: parking, deceleration, steering, etc.)
- prompt information or warning information to inform the driver of the current status of traffic lights; in the case of assisted driving, usually only the prompt information or warning information is issued, the authority to control the vehicle still belongs to the driver, and the driver according to the prompt The information or warning information controls the vehicle accordingly.
- the intelligent driving method provided in the embodiment of the present application further includes:
- This embodiment acquires more information (attributes, states, and corresponding images) of the traffic signals by storing the attributes, states, and corresponding images of the traffic signals, and provides more operation basis for intelligent driving. It is also possible to establish a high-precision map based on the time and location corresponding to the stored traffic signal, and determine the location of the traffic lights in the high-precision map based on the image corresponding to the stored traffic signal.
- the states of the traffic signal light include, but are not limited to, a traffic permitted state, a traffic prohibited state, and a waiting state;
- Step 340 may include:
- determining that the state of the traffic signal is a state of allowing traffic In response to that the color of the traffic signal is green and / or the shape is a first preset shape, determining that the state of the traffic signal is a state of allowing traffic;
- the state of the traffic signal is a waiting state.
- the colors of traffic lights include red, green, and yellow, and different colors correspond to different traffic states.
- Red means that vehicles and / or pedestrians are not allowed to pass
- green means that vehicles and / or pedestrians are allowed to pass
- yellow means that vehicles and / Or pedestrians need to pause and wait
- auxiliary colors can also include shapes such as traffic signals, such as: plus shape (an optional first preset shape) indicates that traffic is allowed, fork shape (an optional The second preset shape) indicates that traffic is prohibited, the minus shape (an optional third preset shape) indicates a waiting state, and the like.
- step 350 may include:
- control the vehicle In response to the state of the traffic light being a traffic permitted state, control the vehicle to perform one or more operations of starting, keeping driving, decelerating, turning, turning on the turn signal, turning on the brake light, and controlling other controls required during the passage of the vehicle ;
- the foregoing program may be stored in a computer-readable storage medium.
- the program is executed, the program is executed.
- the method includes the steps of the foregoing method embodiment.
- the foregoing storage medium includes: a ROM, a RAM, a magnetic disk, or an optical disk, and other media that can store program codes.
- FIG. 4 is a schematic structural diagram of an embodiment of an intelligent driving device according to the present disclosure.
- the intelligent driving device of this embodiment may be used to implement the foregoing intelligent driving method embodiments of the present disclosure.
- the apparatus of this embodiment includes:
- the video stream acquiring unit 21 is configured to acquire a video stream including a traffic signal based on an image acquisition device provided on a vehicle.
- the on-board video is parsed to obtain a video stream including at least one frame of the image, for example, the forward or surrounding of the vehicle can be captured by a camera device installed on the vehicle
- the video of the environment if there are traffic lights in the forward direction of the vehicle or in the surrounding environment, will be captured by the camera device, and the captured video stream is a video stream including traffic lights.
- the images in the video stream may include traffic signals in each frame of images, or at least one frame of images may include traffic signals.
- the area determining unit 22 is configured to determine a candidate area of a traffic light in at least one frame of an image of a video stream.
- the attribute recognition unit 23 is configured to determine at least two attributes of the traffic signal light in the image based on the candidate region.
- the attributes of traffic lights are used to describe traffic lights, which can be defined according to actual needs.
- the attributes of traffic lights can be used to describe the location or location of traffic lights, such as red, green, and yellow.
- Etc. attributes, used to describe the shape of traffic lights (such as circles, straight arrows, polyline arrows, etc.) attributes, and other attributes used to describe other aspects of traffic lights.
- the state determining unit 44 is configured to determine a state of the traffic signal light based on at least two attributes of the traffic signal light in the image.
- Existing image processing methods usually can only deal with one task (for example, one of location recognition or color classification), and the traffic light includes information such as location area, color, and shape.
- the status of the traffic light needs to be determined It is not only necessary to determine the location area of traffic signals, but also at least the color or shape. Therefore, if the general image processing method is applied, at least two neural networks are required to process the video stream, and the processing results need to be synthesized in order to Determine the current state of the traffic signal; this embodiment simultaneously obtains at least two attributes of the traffic signal, determines the state of the traffic signal with at least two attributes, and quickly and accurately identifies the state of the traffic signal.
- the intelligent control unit 45 is configured to perform intelligent driving control on the vehicle according to the state of the traffic signal light.
- an image acquisition device on a vehicle can obtain a video stream in real time, and realize the real-time identification of the attributes of traffic lights to determine the status of the traffic lights. Based on the status of the traffic lights, intelligent driving is achieved without the need for the driver to distract and observe Traffic lights reduce the hidden dangers of traffic safety, and to a certain extent reduce the danger of traffic caused by human error.
- Intelligent driving can include assisted driving and autonomous driving. Generally, assisted driving uses warning lights for warning and alerting, and automatic driving uses signaling lights for driving control.
- the intelligent driving control includes: sending prompt information or warning information, and / or controlling a driving state of the vehicle according to a state of a traffic signal light.
- it further includes:
- the storage unit is configured to store attributes, states, and corresponding images of traffic lights.
- At least two attributes of the traffic signal light include any two or more of the following: location area, color, and shape.
- the states of the traffic signal light include, but are not limited to, a traffic permitted state, a traffic prohibited state, and a waiting state;
- the state determining unit 44 is configured to determine that the state of the traffic signal is a traffic permitted state in response to the color of the traffic signal being green and / or the shape being a first preset shape;
- the state of the traffic signal is a waiting state.
- the intelligent control unit 45 is configured to control the vehicle to perform one or more operations of starting, maintaining a driving state, decelerating, turning, turning on a turn signal, and turning on a brake light in response to a state of a traffic signal being an allowed traffic state. ;
- one or more operations of controlling the vehicle to stop, decelerate, and turn on the brake light are controlled.
- a vehicle including the traffic light detection device according to any one of the above embodiments or the intelligent driving device according to any one of the above embodiments.
- an electronic device including a processor, where the processor includes the traffic light detection device according to any one of the above or the intelligent driving device according to any one of the above embodiments.
- an electronic device including: a memory for storing executable instructions;
- a processor configured to communicate with the memory to execute the executable instructions to complete the operation of the traffic signal detection method according to any one of the above embodiments, or to complete the operation of the intelligent driving method according to any one of the above embodiments.
- An embodiment of the present disclosure further provides an electronic device, which may be, for example, a mobile terminal, a personal computer (PC), a tablet computer, a server, and the like.
- an electronic device which may be, for example, a mobile terminal, a personal computer (PC), a tablet computer, a server, and the like.
- FIG. 5 illustrates a schematic structural diagram of an electronic device 500 suitable for implementing a terminal device or a server of an embodiment of the present disclosure.
- the electronic device 500 includes one or more processors and a communication unit.
- the one or more processors are, for example, one or more central processing units (CPUs) 501, and / or one or more image processors (GPUs) 513, etc.
- CPUs central processing units
- GPUs image processors
- the processors may be stored in a read-only memory (ROM) 502 or executable instructions loaded from the storage section 508 into the random access memory (RAM) 503 to perform various appropriate actions and processes.
- the communication unit 512 may include, but is not limited to, a network card, and the network card may include, but is not limited to, an IB (Infiniband) network card.
- the processor may communicate with the read-only memory 502 and / or the random access memory 503 to execute executable instructions, connect to the communication unit 512 through the bus 504, and communicate with other target devices via the communication unit 512, thereby completing the embodiments of the present disclosure.
- An operation corresponding to any of the methods is, for example, acquiring a video stream including a traffic signal; determining a candidate region of a traffic signal in at least one frame of an image of the video stream; and determining at least two attributes of the traffic signal in the image based on the candidate region.
- RAM 503 can also store various programs and data required for the operation of the device.
- the CPU 501, the ROM 502, and the RAM 503 are connected to each other through a bus 504.
- ROM502 is an optional module.
- the RAM 503 stores executable instructions, or writes executable instructions to the ROM 502 at runtime, and the executable instructions cause the central processing unit 501 to perform operations corresponding to the foregoing communication method.
- An input / output (I / O) interface 505 is also connected to the bus 504.
- the communication unit 512 may be provided in an integrated manner, or may be provided with a plurality of sub-modules (for example, a plurality of IB network cards) and connected on a bus link.
- the following components are connected to the I / O interface 505: an input portion 506 including a keyboard, a mouse, and the like; an output portion 507 including a cathode ray tube (CRT), a liquid crystal display (LCD), and the speaker; a storage portion 508 including a hard disk and the like ; And a communication section 509 including a network interface card such as a LAN card, a modem, and the like.
- the communication section 509 performs communication processing via a network such as the Internet.
- the driver 510 is also connected to the I / O interface 505 as necessary.
- a removable medium 511 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory, etc., is installed on the drive 510 as needed, so that a computer program read therefrom is installed into the storage section 508 as needed.
- FIG. 5 is only an optional implementation manner. In the specific practice process, the number and types of the components in FIG. 5 can be selected, deleted, added or replaced according to actual needs. Different functional component settings can also be implemented in separate settings or integrated settings. For example, GPU513 and CPU501 can be set separately or GPU513 can be integrated on CPU501. Communication unit 512 can be set separately or integrated on CPU501 or GPU513. ,and many more. These alternative embodiments all fall within the protection scope of the present disclosure.
- embodiments of the present disclosure include a computer program product including a computer program tangibly embodied on a machine-readable medium, the computer program including program code for performing a method shown in a flowchart, and the program code may include a corresponding Executing instructions corresponding to the method steps provided in the embodiments of the present disclosure, for example, acquiring a video stream including traffic lights; determining candidate areas of traffic lights in at least one frame of video of the video stream; and determining at least two traffic lights in the image based on the candidate areas Kinds of attributes.
- the computer program may be downloaded and installed from a network through the communication section 509, and / or installed from a removable medium 511.
- a central processing unit (CPU) 501 When the computer program is executed by a central processing unit (CPU) 501, operations of the above-mentioned functions defined in the method of the present disclosure are performed.
- a computer-readable storage medium for storing computer-readable instructions, and when the instructions are executed, the traffic signal detection method according to any one of the foregoing or any one of the foregoing is performed.
- a computer program product including computer-readable code.
- the computer-readable code runs on a device
- a processor in the device executes the program to implement The instructions of the traffic signal detection method or the intelligent driving method according to any one of the above.
- the method and apparatus of the present invention may be implemented by software, hardware, firmware or any combination of software, hardware, firmware.
- the above-mentioned order of the steps of the method is for illustration only, and the steps of the method of the present invention are not limited to the order specifically described above, unless otherwise specifically stated.
- the present invention can also be implemented as programs recorded in a recording medium, which programs include machine-readable instructions for implementing the method according to the present invention.
- the present invention also covers a recording medium storing a program for executing the method according to the present invention.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Evolutionary Computation (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- General Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Physics (AREA)
- General Engineering & Computer Science (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Automation & Control Theory (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- Aviation & Aerospace Engineering (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Traffic Control Systems (AREA)
- Image Analysis (AREA)
- Lighting Device Outwards From Vehicle And Optical Signal (AREA)
Abstract
Description
Claims (57)
- 一种交通信号灯检测方法,其特征在于,包括:获取包括有交通信号灯的视频流;确定所述视频流的至少一帧图像中交通信号灯的候选区域;基于所述候选区域确定所述图像中交通信号灯的至少两种属性。
- 根据权利要求1所述的方法,其特征在于,所述交通信号灯的至少两种属性包括以下任意两种或两种以上:位置区域、颜色和形状。
- 根据权利要求1或2所述的方法,其特征在于,所述确定所述视频流的至少一帧图像中交通信号灯的候选区域,包括:利用基于区域的全卷积网络,确定所述视频流的至少一帧图像中交通信号灯的候选区域。
- 根据权利要求1-3任一所述的方法,其特征在于,所述基于所述候选区域确定所述图像中交通信号灯的至少两种属性,包括:利用多任务识别网络,基于所述候选区域确定所述图像中交通信号灯的至少两种属性。
- 根据权利要求4所述的方法,其特征在于,所述多任务识别网络包括特征提取分支、以及分别与所述特征提取分支连接的至少两个任务分支,不同的任务分支用于确定所述交通信号灯的不同种类属性;所述利用多任务识别网络,基于所述候选区域确定所述图像中交通信号灯的至少两种属性,包括:基于所述特征提取分支对所述候选区域进行特征提取,得到候选特征;分别基于所述至少两个任务分支对所述候选特征进行处理,获得所述图像中交通信号灯的至少两种属性。
- 根据权利要求5所述的方法,其特征在于,所述至少两个任务分支包括:检测分支、识别分支和分类分支;所述分别基于所述至少两个任务分支对所述候选特征进行处理,获得所述图像中交通信号灯的至少两种属性,包括:经所述检测分支对所述候选特征进行位置检测,确定交通信号灯的位置区域;经所述分类分支对所述候选特征进行颜色分类,确定交通信号灯所在位置区域的颜色,确定所述交通信号灯的颜色;经所述识别分支对所述候选特征进行形状识别,确定所述交通信号灯所在位置区域的形状,确定所述交通信号灯的形状。
- 根据权利要求1-6任一所述的方法,其特征在于,所述确定所述视频流的至少一帧图像中交通信号灯的候选区域之前,还包括:对所述视频流中的至少一帧图像进行关键点识别,确定所述图像中的交通信号灯的关键点;对所述视频流中的交通信号灯的关键点进行跟踪,得到跟踪结果;基于所述跟踪结果对所述交通信号灯的位置区域进行调整。
- 根据权利要求7所述的方法,其特征在于,所述对所述视频流中的交通信号灯的关键点进行跟踪,包括:基于连续两帧所述图像中所述交通信号灯的关键点之间的距离;基于所述交通信号灯的关键点之间的距离对所述视频流中的交通信号灯的关键点进行跟踪。
- 根据权利要求8所述的方法,其特征在于,所述基于所述交通信号灯的关键点之间的距离对所 述视频流中的交通信号灯的关键点进行跟踪,包括:基于所述交通信号灯的关键点之间的距离,确定连续两帧图像中同一交通信号灯的关键点的位置区域;根据所述同一交通信号灯的关键点在连续两帧所述图像中的位置区域,在所述视频流中对交通信号灯的关键点进行跟踪。
- 根据权利要求7-9任一所述的方法,其特征在于,所述基于所述跟踪结果对所述信号灯的位置区域进行调整,包括:对比所述跟踪结果与所述信号灯的位置区域是否重合,得到对比结果;基于所述对比结果对所述信号灯的位置区域进行调整。
- 根据权利要求10所述的方法,其特征在于,所述基于所述对比结果对所述信号灯的位置区域进行调整,包括:响应于所述交通信号灯的关键点对应的位置区域和所述信号灯的位置区域不重合,以所述交通信号灯的关键点对应的位置区域替换所述信号灯的位置区域。
- 根据权利要求4-11任一所述的方法,其特征在于,所述确定所述视频流的至少一帧图像中交通信号灯的候选区域之前,还包括:基于采集的训练图像集训练所述基于区域的全卷积网络,所述训练图像集包括多个具有标注属性的训练图像;基于所述训练图像集调整所述基于区域的全卷积网络和所述多任务识别网络中的参数。
- 根据权利要求12所述的方法,其特征在于,所述基于所述训练图像集调整所述基于区域的全卷积网络和所述多任务识别网络中的参数之前,还包括:基于所述训练图像集获取交通信号灯的颜色比例符合预设比例的新训练图像集;基于所述新训练图像集训练分类网络;所述分类网络用于基于所述交通信号灯的颜色对所述训练图像进行分类。
- 根据权利要求13所述的方法,其特征在于,所述预设比例中不同颜色的交通信号灯的数量相同或者数量差异小于容许阈值;所述交通信号灯的颜色包括红色、黄色和绿色。
- 根据权利要求14所述的方法,其特征在于,所述基于所述训练图像集调整所述基于区域的全卷积网络和所述多任务识别网络的参数之前,还包括:基于所述训练后的分类网络的参数初始化所述多任务识别网络中的至少部分参数。
- 根据权利要求1-15任一所述的方法,其特征在于,还包括:基于所述图像中交通信号灯的至少两种属性确定所述交通信号灯的状态;根据所述交通信号灯的状态对所述车辆进行智能驾驶控制。
- 根据权利要求16所述的方法,其特征在于,所述智能驾驶控制包括:发出提示信息或告警信息,和/或,根据所述交通信号灯的状态控制所述车辆的行驶状态。
- 根据权利要求16或17所述的方法,其特征在于,还包括:存储所述交通信号灯的属性、状态及其对应的所述图像。
- 根据权利要求16-18任一所述的方法,其特征在于,所述交通信号灯的状态包括:允许通行状态、禁止通行状态或等待状态;所述基于所述图像中交通信号灯的至少两种属性确定所述交通信号灯的状态,包括以下至少之一:响应于所述交通信号灯的颜色为绿色和/或形状为第一预设形状,确定所述交通信号灯的状态为允 许通行状态;响应于所述交通信号灯的颜色为红色和/或形状为第二预设形状,确定所述交通信号灯的状态为禁止通行状态;响应于所述交通信号灯的颜色为黄色和/或形状为第三预设形状,确定所述交通信号灯的状态为等待状态。
- 根据权利要求19所述的方法,其特征在于,所述根据所述交通信号灯的状态对所述车辆进行智能驾驶控制,包括:响应于所述交通信号灯的状态为允许通行状态,控制所述车辆执行启动、保持行驶状态、减速、转向、开启转向灯、开启刹车灯中的一种或多种操作;响应于所述交通信号灯的状态为禁止通行状态或等待状态,控制所述车辆停车、减速、开启刹车灯中的一种或多种操作。
- 一种智能驾驶方法,其特征在于,包括:基于设置在车辆上的图像采集装置获取包括有交通信号灯的视频流;确定所述视频流的至少一帧图像中交通信号灯的候选区域;基于所述候选区域确定所述图像中交通信号灯的至少两种属性;基于所述图像中交通信号灯的至少两种属性确定所述交通信号灯的状态;根据所述交通信号灯的状态对所述车辆进行智能驾驶控制。
- 根据权利要求21所述的方法,其特征在于,所述智能驾驶控制包括:发出提示信息或告警信息,和/或,根据所述交通信号灯的状态控制所述车辆的行驶状态。
- 根据权利要求21或22所述的方法,其特征在于,还包括:存储所述交通信号灯的属性、状态及其对应的所述图像。
- 根据权利要求21-23任一所述的方法,其特征在于,所述交通信号灯的至少两种属性包括以下任意两种或两种以上:位置区域、颜色和形状。
- 根据权利要求24所述的方法,其特征在于,所述交通信号灯的状态包括:允许通行状态、禁止通行状态和等待状态;所述基于所述图像中交通信号灯的至少两种属性确定所述交通信号灯的状态,包括:响应于所述交通信号灯的颜色为绿色和/或形状为第一预设形状,确定所述交通信号灯的状态为允许通行状态;响应于所述交通信号灯的颜色为红色和/或形状为第二预设形状,确定所述交通信号灯的状态为禁止通行状态;响应于所述交通信号灯的颜色为黄色和/或形状为第三预设形状,确定所述交通信号灯的状态为等待状态。
- 根据权利要求25所述的方法,其特征在于,所述根据所述交通信号灯的状态对所述车辆进行智能驾驶控制,包括:响应于所述交通信号灯的状态为允许通行状态,控制所述车辆执行启动、保持行驶状态、减速、转向、开启转向灯、开启刹车灯中的一种或多种操作;响应于所述交通信号灯的状态为禁止通行状态或等待状态,控制所述车辆停车、减速、开启刹车灯中的一种或多种操作。
- 一种交通信号灯检测装置,其特征在于,包括:视频流获取单元,用于获取包括有交通信号灯的视频流;区域确定单元,用于确定所述视频流的至少一帧图像中交通信号灯的候选区域;属性识别单元,用于基于所述候选区域确定所述图像中交通信号灯的至少两种属性。
- 根据权利要求27所述的装置,其特征在于,所述交通信号灯的至少两种属性包括以下任意两种或两种以上:位置区域、颜色和形状。
- 根据权利要求27或28所述的装置,其特征在于,所述区域确定单元,用于利用基于区域的全卷积网络,确定所述视频流的至少一帧图像中交通信号灯的候选区域。
- 根据权利要求27-29任一所述的装置,其特征在于,所述属性识别单元,用于利用多任务识别网络,基于所述候选区域确定所述图像中交通信号灯的至少两种属性。
- 根据权利要求30所述的装置,其特征在于,所述多任务识别网络包括特征提取分支、以及分别与所述特征提取分支连接的至少两个任务分支,不同的任务分支用于确定所述交通信号灯的不同种类属性;所述属性识别单元,包括:特征提取模块,用于基于所述特征提取分支对所述候选区域进行特征提取,得到候选特征;分支属性模块,用于分别基于所述至少两个任务分支对所述候选特征进行处理,获得所述图像中交通信号灯的至少两种属性。
- 根据权利要求31所述的装置,其特征在于,所述至少两个任务分支包括:检测分支、识别分支和分类分支;所述分支属性模块,用于经所述检测分支对所述候选特征进行位置检测,确定交通信号灯的位置区域;经所述分类分支对所述候选特征进行颜色分类,确定交通信号灯所在位置区域的颜色,确定所述交通信号灯的颜色;经所述识别分支对所述候选特征进行形状识别,确定所述交通信号灯所在位置区域的形状,确定所述交通信号灯的形状。
- 根据权利要求27-32任一所述的装置,其特征在于,还包括:关键点确定单元,用于对所述视频流中的至少一帧图像进行关键点识别,确定所述图像中的交通信号灯的关键点;关键点跟踪单元,用于对所述视频流中的交通信号灯的关键点进行跟踪,得到跟踪结果;位置调整单元,用于基于所述跟踪结果对所述交通信号灯的位置区域进行调整。
- 根据权利要求33所述的装置,其特征在于,所述关键点跟踪单元,用于基于连续两帧所述图像中所述交通信号灯的关键点之间的距离;基于所述交通信号灯的关键点之间的距离对所述视频流中的交通信号灯的关键点进行跟踪。
- 根据权利要求34所述的装置,其特征在于,所述关键点跟踪单元基于所述交通信号灯的关键点之间的距离对所述视频流中的交通信号灯的关键点进行跟踪时,用于基于所述交通信号灯的关键点之间的距离,确定连续两帧图像中同一交通信号灯的关键点的位置区域;根据所述同一交通信号灯的关键点在连续两帧所述图像中的位置区域,在所述视频流中对交通信号灯的关键点进行跟踪。
- 根据权利要求33-35任一所述的装置,其特征在于,所述位置调整单元,用于对比所述跟踪结果与所述信号灯的位置区域是否重合,得到对比结果;基于所述对比结果对所述信号灯的位置区域进行调整。
- 根据权利要求36所述的装置,其特征在于,所述位置调整单元基于所述对比结果对所述信号灯的位置区域进行调整时,用于响应于所述交通信号灯的关键点对应的位置区域和所述信号灯的位置区域不重合,以所述交通信号灯的关键点对应的位置区域替换所述信号灯的位置区域。
- 根据权利要求30-37任一所述的装置,其特征在于,还包括:预训练单元,用于基于采集的训练图像集训练所述基于区域的全卷积网络,所述训练图像集包括多个具有标注属性的训练图像;训练单元,用于基于所述训练图像集调整所述基于区域的全卷积网络和所述多任务识别网络中的参数。
- 根据权利要求38所述的装置,其特征在于,还包括:分类训练单元,用于基于所述训练图像集获取交通信号灯的颜色比例符合预设比例的新训练图像集;基于所述新训练图像集训练分类网络;所述分类网络用于基于所述交通信号灯的颜色对所述训练图像进行分类。
- 根据权利要求39所述的装置,其特征在于,所述预设比例中不同颜色的交通信号灯的数量相同或者数量差异小于容许阈值;所述交通信号灯的颜色包括红色、黄色和绿色。
- 根据权利要求40所述的装置,其特征在于,还包括:初始化单元,用于基于所述训练后的分类网络的参数初始化所述多任务识别网络中的至少部分参数。
- 根据权利要求27-41任一所述的装置,其特征在于,还包括:状态确定单元,用于基于所述图像中交通信号灯的至少两种属性确定所述交通信号灯的状态;智能控制单元,用于根据所述交通信号灯的状态对所述车辆进行智能驾驶控制。
- 根据权利要求42所述的装置,其特征在于,所述智能驾驶控制包括:发出提示信息或告警信息,和/或,根据所述交通信号灯的状态控制所述车辆的行驶状态。
- 根据权利要求42或43所述的装置,其特征在于,还包括:存储单元,用于存储所述交通信号灯的属性、状态及其对应的所述图像。
- 根据权利要求42-44任一所述的装置,其特征在于,所述交通信号灯的状态包括:允许通行状态、禁止通行状态或等待状态;所述状态确定单元,用于响应于所述交通信号灯的颜色为绿色和/或形状为第一预设形状,确定所述交通信号灯的状态为允许通行状态;响应于所述交通信号灯的颜色为红色和/或形状为第二预设形状,确定所述交通信号灯的状态为禁止通行状态;响应于所述交通信号灯的颜色为黄色和/或形状为第三预设形状,确定所述交通信号灯的状态为等待状态。
- 根据权利要求45所述的装置,其特征在于,所述智能控制单元,用于响应于所述交通信号灯的状态为允许通行状态,控制所述车辆执行启动、保持行驶状态、减速、转向、开启转向灯、开启刹车灯中的一种或多种操作;响应于所述交通信号灯的状态为禁止通行状态或等待状态,控制所述车辆停车、减速、开启刹车灯中的一种或多种操作。
- 一种智能驾驶装置,其特征在于,包括:视频流获取单元,用于基于设置在车辆上的图像采集装置获取包括有交通信号灯的视频流;区域确定单元,用于确定所述视频流的至少一帧图像中交通信号灯的候选区域;属性识别单元,用于基于所述候选区域确定所述图像中交通信号灯的至少两种属性;状态确定单元,用于基于所述图像中交通信号灯的至少两种属性确定所述交通信号灯的状态;智能控制单元,用于根据所述交通信号灯的状态对所述车辆进行智能驾驶控制。
- 根据权利要求47所述的装置,其特征在于,所述智能驾驶控制包括:发出提示信息或告警信 息,和/或,根据所述交通信号灯的状态控制所述车辆的行驶状态。
- 根据权利要求47或48所述的装置,其特征在于,还包括:存储单元,用于存储所述交通信号灯的属性、状态及其对应的所述图像。
- 根据权利要求47-49任一所述的装置,其特征在于,所述交通信号灯的至少两种属性包括以下任意两种或两种以上:位置区域、颜色和形状。
- 根据权利要求50所述的装置,其特征在于,所述交通信号灯的状态包括:允许通行状态、禁止通行状态和等待状态;所述状态确定单元,用于响应于所述交通信号灯的颜色为绿色和/或形状为第一预设形状,确定所述交通信号灯的状态为允许通行状态;响应于所述交通信号灯的颜色为红色和/或形状为第二预设形状,确定所述交通信号灯的状态为禁止通行状态;响应于所述交通信号灯的颜色为黄色和/或形状为第三预设形状,确定所述交通信号灯的状态为等待状态。
- 根据权利要求51所述的装置,其特征在于,所述智能控制单元,用于响应于所述交通信号灯的状态为允许通行状态,控制所述车辆执行启动、保持行驶状态、减速、转向、开启转向灯、开启刹车灯中的一种或多种操作;响应于所述交通信号灯的状态为禁止通行状态或等待状态,控制所述车辆停车、减速、开启刹车灯中的一种或多种操作。
- 一种车辆,其特征在于,包括权利要求27至46任意一项所述的交通信号灯检测装置或权利要求47至52任意一项所述的智能驾驶装置。
- 一种电子设备,其特征在于,包括处理器,所述处理器包括权利要求27至46任意一项所述的交通信号灯检测装置或权利要求47至52任意一项所述的智能驾驶装置。
- 一种电子设备,其特征在于,包括:存储器,用于存储可执行指令;以及处理器,用于与所述存储器通信以执行所述可执行指令从而完成权利要求1至20任意一项所述交通信号灯检测方法的操作,或完成权利要求21至26任意一项所述的智能驾驶方法的操作。
- 一种计算机可读存储介质,用于存储计算机可读取的指令,其特征在于,所述指令被执行时执行权利要求1至20任意一项所述交通信号灯检测方法或权利要求21至26任意一项所述的智能驾驶方法的操作。
- 一种计算机程序产品,包括计算机可读代码,其特征在于,当所述计算机可读代码在设备上运行时,所述设备中的处理器执行用于实现权利要求1至20任意一项所述交通信号灯检测方法或权利要求21至26任意一项所述的智能驾驶方法的指令。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
SG11202007333PA SG11202007333PA (en) | 2018-06-29 | 2019-05-29 | Traffic light detection method and apparatus, intelligent driving method and apparatus, vehicle, and electronic device |
JP2020550090A JP7111827B2 (ja) | 2018-06-29 | 2019-05-29 | 交通信号灯検出方法、インテリジェントドライブ方法及び装置、車両並びに電子機器 |
KR1020207029615A KR102447352B1 (ko) | 2018-06-29 | 2019-05-29 | 교통 신호등 검출 및 지능형 주행을 위한 방법 및 디바이스, 차량, 및 전자 디바이스 |
US16/944,234 US20200353932A1 (en) | 2018-06-29 | 2020-07-31 | Traffic light detection method and apparatus, intelligent driving method and apparatus, vehicle, and electronic device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810697683.9A CN110660254B (zh) | 2018-06-29 | 2018-06-29 | 交通信号灯检测及智能驾驶方法和装置、车辆、电子设备 |
CN201810697683.9 | 2018-06-29 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/944,234 Continuation US20200353932A1 (en) | 2018-06-29 | 2020-07-31 | Traffic light detection method and apparatus, intelligent driving method and apparatus, vehicle, and electronic device |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020001223A1 true WO2020001223A1 (zh) | 2020-01-02 |
Family
ID=68985543
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/089062 WO2020001223A1 (zh) | 2018-06-29 | 2019-05-29 | 交通信号灯检测及智能驾驶方法和装置、车辆、电子设备 |
Country Status (6)
Country | Link |
---|---|
US (1) | US20200353932A1 (zh) |
JP (1) | JP7111827B2 (zh) |
KR (1) | KR102447352B1 (zh) |
CN (1) | CN110660254B (zh) |
SG (1) | SG11202007333PA (zh) |
WO (1) | WO2020001223A1 (zh) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113674544A (zh) * | 2021-07-21 | 2021-11-19 | 阿波罗智联(北京)科技有限公司 | 信号机的报警方法、装置、电子设备和可读存储介质 |
WO2022020709A1 (en) * | 2020-07-23 | 2022-01-27 | Waymo Llc | Detecting traffic signaling states with neural networks |
CN116681935A (zh) * | 2023-05-31 | 2023-09-01 | 国家深海基地管理中心 | 一种深海热液喷口自主识别定位方法及系统 |
Families Citing this family (29)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11866046B2 (en) * | 2017-12-10 | 2024-01-09 | Anatoly S. Weiser | Smart traffic control devices and beacons, methods of their operation, and use by vehicles of information provided by the devices and beacons |
US10836393B2 (en) * | 2017-12-10 | 2020-11-17 | Anatoly S. Weiser | Smart traffic control devices and beacons, methods of their operation, and use by vehicles of information provided by the devices and beacons |
CN110688500B (zh) * | 2018-06-20 | 2021-09-14 | 华为技术有限公司 | 一种数据库构建方法、一种定位方法及其相关设备 |
DE112020002604T5 (de) * | 2019-05-29 | 2022-03-03 | Mobileye Vision Technologies Ltd. | Systeme und verfahren für die fahrzeugnavigation |
KR20210052525A (ko) * | 2019-06-27 | 2021-05-10 | 센스타임 그룹 리미티드 | 표시등의 지시 정보 인식 방법 및 장치, 전자 기기 및 기억 매체 |
CN111292531B (zh) * | 2020-02-06 | 2022-07-29 | 阿波罗智能技术(北京)有限公司 | 交通信号灯的跟踪方法、装置、设备及存储介质 |
CN111400531B (zh) * | 2020-03-13 | 2024-04-05 | 广州文远知行科技有限公司 | 目标标注方法、装置、设备与计算机可读存储介质 |
JP7115502B2 (ja) | 2020-03-23 | 2022-08-09 | トヨタ自動車株式会社 | 物体状態識別装置、物体状態識別方法及び物体状態識別用コンピュータプログラムならびに制御装置 |
JP7359735B2 (ja) * | 2020-04-06 | 2023-10-11 | トヨタ自動車株式会社 | 物体状態識別装置、物体状態識別方法及び物体状態識別用コンピュータプログラムならびに制御装置 |
JP7388971B2 (ja) | 2020-04-06 | 2023-11-29 | トヨタ自動車株式会社 | 車両制御装置、車両制御方法及び車両制御用コンピュータプログラム |
CN112289021A (zh) * | 2020-09-24 | 2021-01-29 | 深圳一清创新科技有限公司 | 一种交通信号灯的检测方法、装置及自动驾驶汽车 |
KR20220083945A (ko) * | 2020-12-11 | 2022-06-21 | 현대자동차주식회사 | 신호등 정보 제공 장치, 그를 포함한 차량 시스템 및 그 방법 |
CN112507951B (zh) * | 2020-12-21 | 2023-12-12 | 阿波罗智联(北京)科技有限公司 | 指示灯识别方法、装置、设备、路侧设备和云控平台 |
US11776281B2 (en) * | 2020-12-22 | 2023-10-03 | Toyota Research Institute, Inc. | Systems and methods for traffic light detection and classification |
CN112699773B (zh) * | 2020-12-28 | 2023-09-01 | 阿波罗智联(北京)科技有限公司 | 交通灯识别方法、装置及电子设备 |
CN112349101B (zh) * | 2021-01-08 | 2021-04-27 | 深圳裹动智驾科技有限公司 | 高精地图生成方法、以及识别交通灯的方法及系统 |
CN113011251B (zh) * | 2021-02-03 | 2024-06-04 | 深圳大学 | 一种基于交通灯几何属性的行人交通灯识别方法 |
US12014549B2 (en) | 2021-03-04 | 2024-06-18 | Toyota Research Institute, Inc. | Systems and methods for vehicle light signal classification |
CN113033464B (zh) * | 2021-04-10 | 2023-11-21 | 阿波罗智联(北京)科技有限公司 | 信号灯检测方法、装置、设备以及存储介质 |
CN113077630B (zh) * | 2021-04-30 | 2022-06-28 | 安徽江淮汽车集团股份有限公司 | 基于深度学习的红绿灯检测方法、装置、设备及存储介质 |
CN113450588B (zh) * | 2021-06-28 | 2023-04-07 | 通视(天津)信息技术有限公司 | 交通信号灯时信息的处理方法、装置以及电子设备 |
CN113469109B (zh) * | 2021-07-16 | 2022-09-13 | 阿波罗智联(北京)科技有限公司 | 交通灯识别结果处理方法、装置、路侧设备及云控平台 |
CN113989774B (zh) * | 2021-10-27 | 2024-11-08 | 广州小鹏自动驾驶科技有限公司 | 一种交通灯检测方法、装置、车辆和可读存储介质 |
CN114067291A (zh) * | 2021-11-24 | 2022-02-18 | 智道网联科技(北京)有限公司 | 信号灯检测方法和装置 |
KR102472649B1 (ko) * | 2021-12-28 | 2022-11-30 | 포티투닷 주식회사 | 객체를 트래킹하기 위한 방법 및 장치 |
CN114782924B (zh) * | 2022-05-10 | 2024-07-12 | 智道网联科技(北京)有限公司 | 用于自动驾驶的交通灯检测方法、装置及电子设备 |
CN114973205B (zh) * | 2022-06-28 | 2024-08-13 | 深圳一清创新科技有限公司 | 一种红绿灯跟踪方法、装置、及无人驾驶汽车 |
CN115984823B (zh) * | 2023-02-27 | 2023-06-09 | 安徽蔚来智驾科技有限公司 | 交通信号灯感知方法、车辆控制方法、设备、介质及车辆 |
WO2024197629A1 (zh) * | 2023-03-29 | 2024-10-03 | 华为技术有限公司 | 检测方法、装置以及智能驾驶设备 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106650641A (zh) * | 2016-12-05 | 2017-05-10 | 北京文安智能技术股份有限公司 | 一种交通信号灯定位识别方法、装置及系统 |
CN106909937A (zh) * | 2017-02-09 | 2017-06-30 | 北京汽车集团有限公司 | 交通信号灯识别方法、车辆控制方法、装置及车辆 |
CN107527511A (zh) * | 2016-06-22 | 2017-12-29 | 杭州海康威视数字技术股份有限公司 | 一种智能车辆行车提醒方法及装置 |
CN107978165A (zh) * | 2017-12-12 | 2018-05-01 | 南京理工大学 | 基于计算机视觉的交叉口标志标线与信号灯智能感知方法 |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4357137B2 (ja) * | 2001-05-11 | 2009-11-04 | 富士通マイクロエレクトロニクス株式会社 | 移動体追跡方法及びシステム |
US6999004B2 (en) * | 2002-06-17 | 2006-02-14 | Siemens Corporate Research, Inc. | System and method for vehicle detection and tracking |
CN103489324B (zh) * | 2013-09-22 | 2015-09-09 | 北京联合大学 | 一种基于无人驾驶的实时动态红绿灯检测识别方法 |
KR20150047214A (ko) * | 2013-10-24 | 2015-05-04 | 현대모비스 주식회사 | 카메라 센서를 이용한 scc 시스템 및 방법 |
WO2015077514A1 (en) * | 2013-11-20 | 2015-05-28 | Certusview Technologies, Llc | Systems, methods, and apparatus for tracking an object |
CN103729863B (zh) * | 2013-12-06 | 2016-05-25 | 南京金智视讯技术有限公司 | 基于自主学习的交通信号灯全自动定位识别的方法 |
CN106462727B (zh) * | 2014-01-30 | 2020-02-11 | 移动眼视力科技有限公司 | 车辆、车道终止检测系统和方法 |
WO2015175826A1 (en) * | 2014-05-14 | 2015-11-19 | Kapach Yair | Systems and methods for detecting traffic signs |
US9779314B1 (en) | 2014-08-21 | 2017-10-03 | Waymo Llc | Vision-based detection and classification of traffic lights |
CN105893971A (zh) * | 2016-04-01 | 2016-08-24 | 上海理工大学 | 一种基于Gabor和稀疏表示的交通信号灯识别方法 |
CN106023623A (zh) * | 2016-07-28 | 2016-10-12 | 南京理工大学 | 基于机器视觉的车载交通信号与标志的识别及预警方法 |
CN106570494A (zh) * | 2016-11-21 | 2017-04-19 | 北京智芯原动科技有限公司 | 基于卷积神经网络的交通信号灯识别方法及装置 |
CN106897742B (zh) * | 2017-02-21 | 2020-10-27 | 北京市商汤科技开发有限公司 | 用于检测视频中物体的方法、装置和电子设备 |
CN106837649B (zh) * | 2017-03-03 | 2018-06-22 | 吉林大学 | 基于信号灯倒计时识别的自学习智能起停系统 |
WO2018175441A1 (en) * | 2017-03-20 | 2018-09-27 | Mobileye Vision Technologies Ltd. | Navigation by augmented path prediction |
US10699142B2 (en) * | 2017-04-20 | 2020-06-30 | GM Global Technology Operations LLC | Systems and methods for traffic signal light detection |
CN108108761B (zh) * | 2017-12-21 | 2020-05-01 | 西北工业大学 | 一种基于深度特征学习的快速交通信号灯检测方法 |
-
2018
- 2018-06-29 CN CN201810697683.9A patent/CN110660254B/zh active Active
-
2019
- 2019-05-29 JP JP2020550090A patent/JP7111827B2/ja active Active
- 2019-05-29 WO PCT/CN2019/089062 patent/WO2020001223A1/zh active Application Filing
- 2019-05-29 SG SG11202007333PA patent/SG11202007333PA/en unknown
- 2019-05-29 KR KR1020207029615A patent/KR102447352B1/ko active IP Right Grant
-
2020
- 2020-07-31 US US16/944,234 patent/US20200353932A1/en not_active Abandoned
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107527511A (zh) * | 2016-06-22 | 2017-12-29 | 杭州海康威视数字技术股份有限公司 | 一种智能车辆行车提醒方法及装置 |
CN106650641A (zh) * | 2016-12-05 | 2017-05-10 | 北京文安智能技术股份有限公司 | 一种交通信号灯定位识别方法、装置及系统 |
CN106909937A (zh) * | 2017-02-09 | 2017-06-30 | 北京汽车集团有限公司 | 交通信号灯识别方法、车辆控制方法、装置及车辆 |
CN107978165A (zh) * | 2017-12-12 | 2018-05-01 | 南京理工大学 | 基于计算机视觉的交叉口标志标线与信号灯智能感知方法 |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2022020709A1 (en) * | 2020-07-23 | 2022-01-27 | Waymo Llc | Detecting traffic signaling states with neural networks |
US11328519B2 (en) | 2020-07-23 | 2022-05-10 | Waymo Llc | Detecting traffic signaling states with neural networks |
US11783596B2 (en) | 2020-07-23 | 2023-10-10 | Waymo Llc | Detecting traffic signaling states with neural networks |
CN113674544A (zh) * | 2021-07-21 | 2021-11-19 | 阿波罗智联(北京)科技有限公司 | 信号机的报警方法、装置、电子设备和可读存储介质 |
US11955004B2 (en) | 2021-07-21 | 2024-04-09 | Apollo Intelligent Connectivity (Beijing) Technology Co., Ltd. | Signal machine alarm method and apparatus, electronic device and readable storage medium |
CN116681935A (zh) * | 2023-05-31 | 2023-09-01 | 国家深海基地管理中心 | 一种深海热液喷口自主识别定位方法及系统 |
CN116681935B (zh) * | 2023-05-31 | 2024-01-23 | 国家深海基地管理中心 | 一种深海热液喷口自主识别定位方法及系统 |
Also Published As
Publication number | Publication date |
---|---|
US20200353932A1 (en) | 2020-11-12 |
JP7111827B2 (ja) | 2022-08-02 |
KR102447352B1 (ko) | 2022-09-23 |
SG11202007333PA (en) | 2020-08-28 |
JP2021519968A (ja) | 2021-08-12 |
CN110660254B (zh) | 2022-04-08 |
CN110660254A (zh) | 2020-01-07 |
KR20200128145A (ko) | 2020-11-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020001223A1 (zh) | 交通信号灯检测及智能驾驶方法和装置、车辆、电子设备 | |
JP7106664B2 (ja) | 知的運転制御方法および装置、電子機器、プログラムならびに媒体 | |
US11840239B2 (en) | Multiple exposure event determination | |
WO2020048265A1 (zh) | 多级目标分类及交通标志检测方法和装置、设备、介质 | |
Wu et al. | Applying a functional neurofuzzy network to real-time lane detection and front-vehicle distance measurement | |
CN108335489A (zh) | 高速公路车辆行为语义分析及异常行为监控系统及方法 | |
CN105205785B (zh) | 一种可定位的大型车辆运行管理系统及其运行方法 | |
US9965677B2 (en) | Method and system for OCR-free vehicle identification number localization | |
Li et al. | Lane detection based on connection of various feature extraction methods | |
KR20210080459A (ko) | 차선 검출방법, 장치, 전자장치 및 가독 저장 매체 | |
Widyantoro et al. | Traffic lights detection and recognition based on color segmentation and circle hough transform | |
Hatolkar et al. | A survey on road traffic sign recognition system using convolution neural network | |
Liu et al. | Real-time traffic light recognition based on smartphone platforms | |
KR20230048434A (ko) | 자기 차량의 자율 및 지원 운전에서 환경 특징을 연속적이고 적응적으로 검출하는 컴퓨터 구현 방법 | |
Luo et al. | Detection and recognition of obscured traffic signs during vehicle movement | |
Bhupathi et al. | Sharp curve detection of autonomous vehicles using dbscan and augmented sliding window techniques | |
CA3196453A1 (en) | Lane line detection method and apparatus | |
Kalms et al. | Robust lane recognition for autonomous driving | |
Oza et al. | Object Detection using IoT and Machine Learning to Avoid Accident and Improve Road Safety | |
Li et al. | A multiple-lane vehicle tracking method for forward collision warning system applications | |
Erwin et al. | Detection of highway lane using color filtering and line determination | |
Yang | An enhanced three-frame-differencing approach for vehicle detection under challenging environmental conditions | |
Du et al. | Traffic light recognition based on prior knowledge and optimized threshold segmentation | |
Khalfin et al. | Developing, Analyzing, and Evaluating Vehicular Lane Keeping Algorithms Under Dynamic Lighting and Weather Conditions Using Electric Vehicles | |
Kenk et al. | Driving Perception in Challenging Road Scenarios: An Empirical Study |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19827213 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2020550090 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20207029615 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19827213 Country of ref document: EP Kind code of ref document: A1 |