Mehl et al., 2023 - Google Patents
M-fuse: Multi-frame fusion for scene flow estimationMehl et al., 2023
View PDF- Document ID
- 13113916594957209352
- Author
- Mehl L
- Jahedi A
- Schmalfuss J
- Bruhn A
- Publication year
- Publication venue
- Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision
External Links
Snippet
Recently, neural network for scene flow estimation show impressive results on automotive data such as the KITTI benchmark. However, despite of using sophisticated rigidity assumptions and parametrizations, such networks are typically limited to only two frame …
- 230000004927 fusion 0 title abstract description 52
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/00221—Acquiring or recognising human faces, facial parts, facial sketches, facial expressions
- G06K9/00268—Feature extraction; Face representation
- G06K9/00281—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/62—Methods or arrangements for recognition using electronic means
- G06K9/6201—Matching; Proximity measures
- G06K9/6202—Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20112—Image segmentation details
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/36—Image preprocessing, i.e. processing the image information without deciding about the identity of the image
- G06K9/46—Extraction of features or characteristics of the image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/10—Geometric effects
- G06T15/20—Perspective computation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/30—Information retrieval; Database structures therefor; File system structures therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7106665B2 (en) | MONOCULAR DEPTH ESTIMATION METHOD AND DEVICE, DEVICE AND STORAGE MEDIUM THEREOF | |
Shivakumar et al. | Dfusenet: Deep fusion of rgb and sparse depth information for image guided dense depth completion | |
CN110832501B (en) | System and method for pose invariant facial alignment | |
US11049270B2 (en) | Method and apparatus for calculating depth map based on reliability | |
US12039657B2 (en) | View synthesis of a dynamic scene | |
Liu et al. | Sift flow: Dense correspondence across different scenes | |
Luo et al. | Real-time dense monocular SLAM with online adapted depth prediction network | |
CN112750133A (en) | Computer vision training system and method for training a computer vision system | |
Mehl et al. | M-fuse: Multi-frame fusion for scene flow estimation | |
Hung et al. | Consistent binocular depth and scene flow with chained temporal profiles | |
Xiong et al. | Self-supervised monocular depth and visual odometry learning with scale-consistent geometric constraints | |
Liu et al. | A survey on deep learning methods for scene flow estimation | |
Qiao et al. | SF-Net: Learning scene flow from RGB-D images with CNNs | |
Li et al. | Two-stage adaptive object scene flow using hybrid cnn-crf model | |
Huang et al. | Learning optical flow with R-CNN for visual odometry | |
Lee et al. | Self-supervised monocular depth and motion learning in dynamic scenes: Semantic prior to rescue | |
Da Silveira et al. | Indoor depth estimation from single spherical images | |
Long et al. | Detail preserving residual feature pyramid modules for optical flow | |
EP4350632A2 (en) | Method and appratus with neural rendering based on view augmentation | |
Zhang et al. | Unsupervised learning of monocular depth and ego-motion with space–temporal-centroid loss | |
Haji-Esmaeili et al. | Large-scale monocular depth estimation in the wild | |
Meng et al. | Un-VDNet: unsupervised network for visual odometry and depth estimation | |
Zhang et al. | Dynamic fusion network for light field depth estimation | |
CN113962846A (en) | Image alignment method and device, computer readable storage medium and electronic device | |
Chi et al. | Subspace-PNP: A geometric constraint loss for mutual assistance of depth and optical flow estimation |