Beveridge, 2021 - Google Patents
Consistent Depth Estimation in Data-Driven Simulation for Autonomous DrivingBeveridge, 2021
View PDF- Document ID
- 5591480972403833127
- Author
- Beveridge M
- Publication year
External Links
Snippet
In this work we propose consistent depth estimation for viewpoint reconstruction in data- driven simulation, combining aspects of learning-based monocular depth prediction and structure-from-motion to increase temporal video depth accuracy. We demonstrate efficacy …
- 238000004088 simulation 0 title abstract description 42
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/00624—Recognising scenes, i.e. recognition of a whole field of perception; recognising scene-specific objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T13/00—Animation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06Q—DATA PROCESSING SYSTEMS OR METHODS, SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce, e.g. shopping or e-commerce
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Liao et al. | Kitti-360: A novel dataset and benchmarks for urban scene understanding in 2d and 3d | |
Bonatti et al. | Towards a robust aerial cinematography platform: Localizing and tracking moving targets in unstructured environments | |
Luo et al. | Every pixel counts++: Joint learning of geometry and motion with 3d holistic understanding | |
Shamwell et al. | Unsupervised deep visual-inertial odometry with online error correction for RGB-D imagery | |
JP7178396B2 (en) | Method and computer system for generating data for estimating 3D pose of object included in input image | |
Yang et al. | Every pixel counts: Unsupervised geometry learning with holistic 3d motion understanding | |
Rhinehart et al. | R2p2: A reparameterized pushforward policy for diverse, precise generative path forecasting | |
Zhou et al. | Unsupervised learning of depth and ego-motion from video | |
US20210049371A1 (en) | Localisation, mapping and network training | |
US11948310B2 (en) | Systems and methods for jointly training a machine-learning-based monocular optical flow, depth, and scene flow estimator | |
Riegler et al. | Connecting the dots: Learning representations for active monocular depth estimation | |
Mo et al. | Terra: A smart and sensible digital twin framework for robust robot deployment in challenging environments | |
Murali et al. | Utilizing semantic visual landmarks for precise vehicle navigation | |
Chen et al. | Deep learning for visual localization and mapping: A survey | |
CN104040593B (en) | Method and apparatus for 3D model deformation | |
Li et al. | MannequinChallenge: Learning the depths of moving people by watching frozen people | |
Wang et al. | Unsupervised learning of 3d scene flow from monocular camera | |
Luvizon et al. | Scene‐Aware 3D Multi‐Human Motion Capture from a Single Camera | |
Yang et al. | Recovering and simulating pedestrians in the wild | |
Singh et al. | Fast semantic-aware motion state detection for visual slam in dynamic environment | |
Park et al. | Test-Time Adaptation for Depth Completion | |
Rukhovich et al. | Estimation of absolute scale in monocular SLAM using synthetic data | |
Zhao et al. | How Challenging is a Challenge? CEMS: a Challenge Evaluation Module for SLAM Visual Perception | |
Yang et al. | Mapping technology in visual slam: A review | |
Doğan et al. | An augmented crowd simulation system using automatic determination of navigable areas |