Meng et al., 2023 - Google Patents
HYDRO-3D: Hybrid object detection and tracking for cooperative perception using 3D LiDARMeng et al., 2023
View PDF- Document ID
- 15514217328943471581
- Author
- Meng Z
- Xia X
- Xu R
- Liu W
- Ma J
- Publication year
- Publication venue
- IEEE Transactions on Intelligent Vehicles
External Links
Snippet
3D-LiDAR-based cooperative perception has been generating significant interest for its ability to tackle challenges such as occlusion, sparse point clouds, and out-of-range issues that can be problematic for single-vehicle perception. Despite its effectiveness in …
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/00624—Recognising scenes, i.e. recognition of a whole field of perception; recognising scene-specific objects
- G06K9/00791—Recognising scenes perceived from the perspective of a land vehicle, e.g. recognising lanes, obstacles or traffic signs on road scenes
- G06K9/00805—Detecting potential obstacles
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/00624—Recognising scenes, i.e. recognition of a whole field of perception; recognising scene-specific objects
- G06K9/00791—Recognising scenes perceived from the perspective of a land vehicle, e.g. recognising lanes, obstacles or traffic signs on road scenes
- G06K9/00798—Recognition of lanes or road borders, e.g. of lane markings, or recognition of driver's driving pattern in relation to lanes perceived from the vehicle; Analysis of car trajectory relative to detected road
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/16—Anti-collision systems
- G08G1/167—Driving aids for lane monitoring, lane changing, e.g. blind spot detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/36—Image preprocessing, i.e. processing the image information without deciding about the identity of the image
- G06K9/46—Extraction of features or characteristics of the image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/20—Image acquisition
- G06K9/32—Aligning or centering of the image pick-up or image-field
- G06K9/3233—Determination of region of interest
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/01—Detecting movement of traffic to be counted or controlled
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/66—Radar-tracking systems; Analogous systems where the wavelength or the kind of wave is irrelevant
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/86—Combinations of radar systems with non-radar systems, e.g. sonar, direction finder
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30248—Vehicle exterior or interior
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in preceding groups
- G01C21/26—Navigation; Navigational instruments not provided for in preceding groups specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Meng et al. | HYDRO-3D: Hybrid object detection and tracking for cooperative perception using 3D LiDAR | |
Muhammad et al. | Deep learning for safe autonomous driving: Current challenges and future directions | |
Nidamanuri et al. | A progressive review: Emerging technologies for ADAS driven solutions | |
Meyer et al. | Graph convolutional networks for 3d object detection on radar data | |
Datondji et al. | A survey of vision-based traffic monitoring of road intersections | |
Laugier et al. | Probabilistic analysis of dynamic scenes and collision risks assessment to improve driving safety | |
Nguyen et al. | Stereo-camera-based urban environment perception using occupancy grid and object tracking | |
US11755917B2 (en) | Generating depth from camera images and known depth data using neural networks | |
Shi et al. | Grid-centric traffic scenario perception for autonomous driving: A comprehensive review | |
Revilloud et al. | An improved approach for robust road marking detection and tracking applied to multi-lane estimation | |
Singh | Vision-radar fusion for robotics bev detections: A survey | |
Bai et al. | Cyber mobility mirror: A deep learning-based real-world object perception platform using roadside LiDAR | |
Iqbal et al. | Modeling perception in autonomous vehicles via 3d convolutional representations on lidar | |
Gao et al. | A survey of collaborative perception in intelligent vehicles at intersections | |
Barbosa et al. | Camera-radar perception for autonomous vehicles and ADAS: Concepts, datasets and metrics | |
Stäcker et al. | RC-BEVFusion: A plug-in module for radar-camera bird’s eye view feature fusion | |
Xu et al. | M2da: multi-modal fusion transformer incorporating driver attention for autonomous driving | |
Yan et al. | Int2: Interactive trajectory prediction at intersections | |
Wu | Fusion-based modeling of an intelligent algorithm for enhanced object detection using a Deep Learning Approach on radar and camera data | |
CN118823308A (en) | A target detection and tracking network model and method based on multi-source heterogeneous sensor information fusion | |
Pravallika et al. | Deep Learning Frontiers in 3D Object Detection: A Comprehensive Review for Autonomous Driving | |
Lee et al. | Map matching-based driving lane recognition for low-cost precise vehicle positioning on highways | |
Zhu et al. | A survey on deep learning approaches for data integration in autonomous driving system | |
Lee et al. | Ego‐lane index‐aware vehicular localisation using the DeepRoad Network for urban environments | |
Dong et al. | TS-BEV: BEV object detection algorithm based on temporal-spatial feature fusion |