CN107993260A - A kind of light field image depth estimation method based on mixed type convolutional neural networks - Google Patents
A kind of light field image depth estimation method based on mixed type convolutional neural networks Download PDFInfo
- Publication number
- CN107993260A CN107993260A CN201711337965.XA CN201711337965A CN107993260A CN 107993260 A CN107993260 A CN 107993260A CN 201711337965 A CN201711337965 A CN 201711337965A CN 107993260 A CN107993260 A CN 107993260A
- Authority
- CN
- China
- Prior art keywords
- mrow
- light field
- neural networks
- convolutional neural
- msub
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/13—Edge detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10052—Images from lightfield camera
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention discloses a kind of light field image depth estimation method based on mixed type convolutional neural networks.The present invention includes the training of the structure, convolutional neural networks model of training dataset, generates the estimation of Depth figure of light field image.Light field depth calculation problem is converted to classification problem by the present invention, efficiently uses the contact between regional area pixel depth;Light field data is represented by four-dimensional parameter;The present invention utilizes the EPI image cathetus slopes relation proportional to scene depth of light field image, and using EPI images as a medium, four-dimension light field image is mapped in two dimensional image.By extracting the EPI blocks region in light field image centre visual angle image corresponding to pixel, by the way of polar curve segment region pair, a kind of training dataset of new light field image estimation of Depth is constructed.Present invention utilizes advantage of the deep learning in terms of feature abstraction, and the Stability and veracity of its estimation of Depth is than general conventional method advantageously.
Description
Technical field
The present invention relates to light field (Light Field) image domains and deep learning (Deep Learning) field, specifically
It is related to a kind of light field image depth estimation method based on mixed type convolutional neural networks.
Background technology
In recent years, as the fast development of photoelectric technology, new imaging device continue to bring out, optical field imaging is as a kind of new
Emerging 3 Dimension Image Technique enjoys researcher to pay close attention to its unique imaging process.Traditional camera can only record one it is two-dimentional flat
Face, can not obtain depth information of scene, and optical field imaging can record four-dimensional position and direction of the light radiation in communication process
Information, thus in image reconstruction process, more abundant image information can be obtained.Optical field imaging can realize digital reunion
Jiao, synthetic aperture, obtain big depth image, and three-dimensional reconstruction.As light field technology constantly moves to maturity, light field estimation of Depth
Receive more and more attention, research both domestic and external at present focuses primarily upon the following aspects:
Light field data is carried out 4D parametrizations to represent, direction and the angle information of capture is made full use of, improves depth calculation
As a result accuracy;Light field data is utilized into the lax expression of polar plane image, improves the time efficiency during depth calculation;Profit
With the 4D architectural features of light field, model is built, makes full use of the abundant information of various visual angles, realizes that the exact depth for blocking scene is estimated
Meter;Pass through interpolation and local restriction, the resolution ratio of raising scene depth result of calculation.
At present, the estimation of Depth in relation to light field has been achieved for bigger progress, and the method for mainstream can substantially divide at present
For three classes:
(1) estimation of Depth based on polar plane image.According to the polar curve slope of polar figure (Epipolar Image, EPI)
Proportional relation, Goldluecke ask texture oblique using the structure tensor of EPI image gradients with light field image Scene depth
Rate, using the energy-saving global restriction of energy function is minimized, realizes the estimation of depth information.
(2) estimation of Depth based on various visual angles.According to its constraints, various visual angles depth estimation method can be divided into three
Kind:Estimated based on matched various visual angles estimation of Depth, the various visual angles estimation of Depth based on model, colour consistency various visual angles depth
Meter.Various visual angles depth estimation method utilizes the matching relationship estimation scene depth between different visual angles, can effectively eliminate visual angle string
Disturb and the influence of light field aliasing effect.
(3) estimation of Depth based on refocusing.Using digital refocusing technology, a series of focus stack diagrams can be obtained
Picture, by being detected to its being focused property, can accurately estimate scene depth.
Due to illumination variation and different scenes complexity, the estimation of Depth performance of existing algorithm to varying degrees by
To limitation, the accuracy of algorithm has much room for improvement, and is easily influenced be subject to noise in image, easily in depth locus of discontinuity
There is erroneous estimation in domain, texture weaker area, cause accuracy relatively low.
The content of the invention
The present invention is directed to the deficiency of above-mentioned light field image depth computing method, and deep learning innovatively is applied to light field
Picture depth calculating field, proposes a kind of light field image depth estimation method based on mixed type convolutional neural networks.The present invention
Light field depth calculation problem is converted into classification problem, efficiently uses the contact between regional area pixel depth.Light field data
Represented by four-dimensional parameter, therefore be difficult to directly apply to two-dimensional convolution neural network structure.The present invention utilizes the EPI of light field image
Image cathetus slope relation proportional to scene depth, using EPI images as a medium, is mapped to four-dimension light field image
In two dimensional image.By extracting the EPI blocks region in light field image centre visual angle image corresponding to pixel, using polar curve segment
Region constructs a kind of training dataset of new light field image estimation of Depth to the mode of (EPI Patch-pairs).
The technical solution adopted by the present invention to solve the technical problems includes the following steps:
Step 1. extracts polar figure:
For the light field picture in light field image data set, with centre visual angle image hanging down per a line and corresponding to each row
Nogata carries out polar figure extraction to horizontal direction;
Step 2. generation polar curve segment region pair:
Extract the polar curve segment horizontally and vertically corresponding to each pixel of polar figure, and with horizontal direction and
Vertical direction is one group, generation polar curve segment region pair;
Step 3. builds training data:
Edge detection is carried out to center multi-view image, by the polar curve block region corresponding to the unsharp pixel region of texture
It is right, be considered as invalid data and removed from data set, in order to obtain more preferable training effect, by the characteristic in data set into
Row equilibrating is handled, and is made the quantity of each feature identical in error range, is built final training dataset;
Step 4. neural network model is trained:
Mixed type convolutional neural networks are built, mixed type convolutional neural networks are carried out using the data set having had been built up
Training, generates model file.
Extraction polar figure described in step 1, specifically:
Use LF(x, y, s, t) represents 4D light field datas, wherein (x, y) representation space coordinate, (s, t) represents microlens array
Middle sub-lens coordinate, centre visual angle image refer to the image (s=s that light passes through main lens to be formed0) for centre visual angle figure
Pixel p (the x of picturei,yi), the EPI of horizontal direction can use LF(x, y, s, t) represents that size is Ns×Nx, each pixel level side
To EPI blocks be exactly with (xi,s0) centered on a block size be (Ns×Wx) region;It can similarly obtain, row coordinate is xiHang down
Nogata to EPI L can be expressed as in light field imageF(x, y, s, t), size Nt×Ny, the EPI blocks region of vertical direction is just
It is with (yi,t0) centered on size be (Nt×Wy) region.
Generation polar curve segment region pair described in step 2, specifically:
The light-field camera of use uses 9 × 9 lens array, and the resolution ratio of each sub-lens imaging is 512 × 512,
So the size of polar figure is 512 × 9, the corresponding polar curve block region obtained of each centre visual angle graphical pixel point is with current
Centered on location of pixels, size is 13 × 9 region;Pixel is corresponded to the Liang Ge areas horizontally and vertically got
Domain can extract multiple polar curves pair as one group of polar curve pair, each illumination field patterns.
Structure training data described in step 3, specifically:
Judge pixel whether be invalid data process, the method for use is:Centering is carried out using Canny operators
Heart multi-view image carries out edge detection, is to be effective by this polar curve segment region if more clearly edge can be detected
Data, if cannot, it is considered as invalid data and weeds out.
Structure training data described in step 3, the method that the process of data set balance uses is to use the side of repeat replication
Formula, makes the training set quantity of different characteristic tend to balance.
Mixed type convolutional neural networks described in step 4, specific structure are as follows:
Connected using two relatively independent convolutional neural networks, first order neutral net is first trained, by the first order
Training data of the prediction result of neutral net as second level convolutional neural networks;First order convolutional neural networks are specially:
The input of first order convolutional neural networks is the polar curve segment region pair of extraction, in order to enable the network to learning level
Different characteristic on direction and vertical direction polar curve segment region, horizontal direction is handled using two independent convolution sub-networks
With the polar curve segment region of vertical direction, each sub-network has 7 layers of convolutional layer, and the size of convolution kernel is 2 × 2, first layer volume
The quantity of product core is 16, generates 16 characteristic patterns, and each layer below will double;Behind convolutional layer with a full articulamentum phase
Even, the output of layer 7 convolutional layer is integrated into one-dimensional characteristic vector;Output layer is the classification layer of convolutional neural networks, receives and
From the output of two sub-networks, the scope of output is exactly the scope classified;The estimation of Depth scope of light field image in database
For -4~4, in the range of this, 229 classes, precision 0.035 are classified as, Softmax functions are received to come from two sons
The feature vector of network;In the training stage, using cross entropy as loss function, it is specifically defined as:
Wherein, y is the output valve of mixed type convolutional neural networks, ygtIt is label value, wherein, λ1=0.5, λ2=0.25.
The second level convolutional neural networks are specially:
The training data input of second level convolutional neural networks is the prediction result of first order convolutional neural networks, by light field
Light field image in database carries out regression forecasting with trained first order convolutional neural networks, generates primary light field figure
The estimation of Depth figure of picture, as the training data of second level convolutional neural networks;Second level convolutional neural networks are by one
Convolutional layer and a full articulamentum composition, convolutional layer use 7 × 7 convolution kernel, and convolution kernel number is 32, the characteristic pattern number of generation
Measure as 32, be connected behind convolutional layer with a full articulamentum, the final result of output light field picture depth figure.
The beneficial effects of the invention are as follows:
The present invention light field image scene depth is estimated using the method for deep learning, contain data set structure with
In two stages of convolutional neural networks model training, build the stage, it is proposed that a kind of new light field image data set structure in data set
Construction method, dimensionality reduction is carried out by light field data, and training dataset is reached enough quantity, meets network training to data set
It is required that.In the convolutional neural networks training stage, it is proposed that a kind of convolutional neural networks structure of mixed type, employs two relatively
Independent sub-network structure, first order neutral net are trained using the training set built.Second level neutral net is by
The prediction result of level-one neutral net continues the training of second level network as training set.Being experimentally confirmed this method has
Obvious performance boost.Compared to the scene depth computational methods of traditional light field image, the present invention asks light field depth calculation
Topic is converted to classification problem, efficiently uses the contact between regional area pixel depth.The present invention is to the scene in light field image
Estimation of Depth is carried out, improves scene depth is estimated in light field image accuracy and stability.
Brief description of the drawings
Fig. 1 is the method flow diagram of the light field image estimation of Depth based on mixed type convolutional neural networks
Fig. 2 is the primary network station concrete structure diagram of the mixed type convolutional neural networks of the present invention.
Fig. 3 is the two grade network concrete structure diagram of the mixed type convolutional neural networks of the present invention.
Embodiment
The present invention is made further instructions below in conjunction with the accompanying drawings:
As shown in Figs. 1-3, a kind of light field image depth estimation method based on mixed type convolutional neural networks, specific implementation
It is as follows:
Step 1. extraction polar figure (Epipolar Image, EPI):
For the light field picture in light field image data set, with centre visual angle image hanging down per a line and corresponding to each row
Nogata carries out polar figure extraction (EPI Extraction) to horizontal direction.
Step 2. generates polar curve segment region to (EPI-Patch Pairs):
The polar curve segment (EPI-Patch) horizontally and vertically corresponding to each pixel of polar figure is extracted, and
With horizontally and vertically for one group, to (EPI-Patch Pairs), each light field image can in generation polar curve segment region
The a large amount of this data of generation.
Step 3. training data (Training Data) is built:
Edge detection is carried out to center multi-view image, by the polar curve block region corresponding to the unsharp pixel region of texture
It is right, be considered as invalid data and removed from data set, in order to obtain more preferable training effect, by the characteristic in data set into
Row equilibrating is handled, and is made the quantity of each feature essentially identical, is built final training dataset.
Step 4. neural network model is trained:
Mixed type convolutional neural networks are built, mixed type convolutional neural networks are carried out using the data set having had been built up
Training, generates model file.
Extraction polar figure described in step 1, specifically:
Use LF(x, y, s, t) represents 4D light field datas, wherein (x, y) representation space coordinate, (s, t) represents microlens array
Middle sub-lens coordinate, centre visual angle image refer to the image (s=s that light passes through main lens to be formed0) for centre visual angle figure
Pixel p (the x of picturei,yi), the EPI of horizontal direction can use LF(x, y, s, t) represents that size is Ns×Nx, each pixel level side
To EPI blocks be exactly with (xi,s0) centered on a block size be (Ns×Wx) region, can similarly obtain, row coordinate is xiHang down
Nogata to EPI L can be expressed as in light field imageF(x, y, s, t), its size are Nt×Ny, the EPI blocks area of vertical direction
Domain is exactly with (yi,t0) centered on size be (Nt×Wy) region.
Generation polar curve segment region pair described in step 2, specifically:
The light-field camera of use uses 9 × 9 lens array, and the resolution ratio of each sub-lens imaging is 512 × 512,
So the size of polar figure (EPI) is 512 × 9, the polar curve block region of the corresponding acquirement of each centre visual angle graphical pixel point is
Centered on current pixel position, size is 13 × 9 region.Pixel is corresponded to get horizontally and vertically
As one group of polar curve pair, each illumination field patterns can extract many this polar curves pair in two regions.
The structure of training data described in step 3, specifically:
First by the polar curve corresponding to the unsharp pixel region of texture to being considered as invalid data and being picked from data set
Remove, in order to obtain more preferable training effect, the characteristic in data set is balanced processing, makes the quantity of each feature
It is essentially identical, data set is finally converted to the input format of convolutional neural networks requirement, builds training dataset.Judging picture
Whether vegetarian refreshments is invalid process, and the method for use is:Carry out carrying out edge inspection to center multi-view image using Canny operators
Survey, if more clearly edge can be detected, by this polar curve segment region be for valid data, if cannot, be considered as nothing
Effect data simultaneously weed out.The process of data set balance, using method be to use repeat replication by the way of, make the instruction of different characteristic
Practice collection quantity to tend to balance.
Neural network model training described in step 4, specifically:
The data set built is trained by hybrid neural network, generates model file.It is of the invention innovative
Mixed type convolutional neural networks are employed, it is specific as follows:
Present networks employ two relatively independent convolutional neural networks and connect, and first train first order neutral net,
Training data using the prediction result of first order neutral net as second level convolutional neural networks.
First order convolutional neural networks are specially:
As shown in Fig. 2, the input of this network be previous step extraction polar curve segment region to (EPI Patch-
Pairs), in order to make network to employ two with the different characteristic on learning level direction and vertical direction polar curve segment region
A independent convolution sub-network handles polar curve segment region horizontally and vertically, and each sub-network has 7 layers of volume
Lamination, the size of convolution kernel is 2 × 2, and the quantity of first layer convolution kernel is 16, generates 16 characteristic patterns
(featuremap) each layer behind will double, and be connected behind convolutional layer with a full articulamentum, by the 7th
The output of layer convolutional layer is integrated into one-dimensional characteristic vector.Output layer is the classification layer of convolutional neural networks, and reception comes from two sons
The output of network, the scope of output are exactly the scope that we classify, and the estimation of Depth scope of the light field image in database is -4
~4, in the range of this, 229 classes, precision 0.035 are classified as, Softmax functions are received to come from two sub-networks
Feature vector.It can be obtained using two sub-networks and more accurately classified.
In the training stage, using cross entropy as loss function, it is specifically defined as:
Wherein, y is the output valve of network model defined in us, ygtIt is label value, in our experiment, λ1=
0.5, λ2=0.25.
Second level convolutional neural networks are specially:
As shown in figure 3, prediction of the training data input of second level convolutional neural networks for first order convolutional neural networks
As a result, the light field image in light field data storehouse is subjected to regression forecasting with trained first order convolutional neural networks, it is raw
Into the estimation of Depth figure of primary light field image, as the training data of second level convolutional neural networks.Second level convolution god
It is made of through network a convolutional layer and a full articulamentum, convolutional layer uses 7 × 7 convolution kernel, and convolution kernel number is 32, raw
Into characteristic pattern quantity be 32, be connected with a full articulamentum behind convolutional layer, the final result of output light field picture depth figure.
Claims (7)
1. a kind of light field image depth estimation method based on mixed type convolutional neural networks, it is characterised in that including following step
Suddenly:
Step 1. extracts polar figure:
For the light field picture in light field image data set, with Vertical Square of the centre visual angle image per a line and corresponding to each row
Polar figure extraction is carried out to horizontal direction;
Step 2. generation polar curve segment region pair:
The polar curve segment horizontally and vertically corresponding to each pixel of polar figure is extracted, and with horizontal direction and vertically
Direction is one group, generation polar curve segment region pair;
Step 3. builds training data:
Edge detection is carried out to center multi-view image, by the polar curve block region pair corresponding to the unsharp pixel region of texture,
It is considered as invalid data and is removed from data set, in order to obtain more preferable training effect, the characteristic in data set is carried out
Equilibrating is handled, and is made the quantity of each feature identical in error range, is built final training dataset;
Step 4. neural network model is trained:
Mixed type convolutional neural networks are built, mixed type convolutional neural networks are instructed using the data set having had been built up
Practice, generate model file.
2. a kind of light field image depth estimation method based on mixed type convolutional neural networks according to claim 1, its
It is characterized in that the extraction polar figure described in step 1, specifically:
Use LF(x, y, s, t) represents 4D light field datas, wherein (x, y) representation space coordinate, (s, t) represents microlens array neutron
Lens coordinate, centre visual angle image refer to the image (s=s that light passes through main lens to be formed0) for centre visual angle image
Pixel p (xi,yi), the EPI of horizontal direction can use LF(x, y, s, t) represents that size is Ns×Nx, each pixel level direction
EPI blocks are exactly with (xi,s0) centered on a block size be (Ns×Wx) region;It can similarly obtain, row coordinate is xiVertical Square
To EPI L can be expressed as in light field imageF(x, y, s, t), size Nt×Ny, the EPI blocks region of vertical direction be exactly with
(yi,t0) centered on size be (Nt×Wy) region.
3. a kind of light field image depth estimation method based on mixed type convolutional neural networks according to claim 2, its
It is characterized in that the generation polar curve segment region pair described in step 2, specifically:
The light-field camera of use uses 9 × 9 lens array, and the resolution ratio of each sub-lens imaging is 512 × 512, so
The size of polar figure is 512 × 9, and the corresponding polar curve block region obtained of each centre visual angle graphical pixel point is with current pixel
Centered on position, size is 13 × 9 region;Pixel is corresponded to horizontally and vertically two regions got to make
For one group of polar curve pair, each illumination field patterns can extract multiple polar curves pair.
4. a kind of light field image depth estimation method based on mixed type convolutional neural networks according to claim 3, its
It is characterized in that the structure training data described in step 3, specifically:
Judge pixel whether be invalid data process, the method for use is:Using Canny operators regard center
Angle image carries out edge detection, if more clearly edge can be detected, by this polar curve segment region be for valid data,
If cannot, it is considered as invalid data and weeds out.
5. a kind of light field image depth estimation method based on mixed type convolutional neural networks according to claim 4, its
It is characterized in that the structure training data described in step 3, specifically:
The method that the process of data set balance uses is using by the way of repeat replication, makes the training set quantity of different characteristic tend to
Balance.
6. a kind of light field image depth estimation method based on mixed type convolutional neural networks according to claim 4, its
It is characterized in that the mixed type convolutional neural networks described in step 4, specific structure is as follows:
Connected using two relatively independent convolutional neural networks, first train first order neutral net, by first order nerve
Training data of the prediction result of network as second level convolutional neural networks;First order convolutional neural networks are specially:
The input of first order convolutional neural networks is the polar curve segment region pair of extraction, in order to enable the network to learning level direction
With the different characteristic on vertical direction polar curve segment region, handle horizontal direction using two independent convolution sub-networks and hang down
Nogata to polar curve segment region, each sub-network has 7 layers of a convolutional layer, and the size of convolution kernel is 2 × 2, first layer convolution kernel
Quantity be 16, generate 16 characteristic patterns, each layer below will double;It is connected behind convolutional layer with a full articulamentum,
The output of layer 7 convolutional layer is integrated into one-dimensional characteristic vector;Output layer is the classification layer of convolutional neural networks, and reception comes from
The output of two sub-networks, the scope of output are exactly the scope classified;The estimation of Depth scope of light field image in database for-
4~4, in the range of this, 229 classes, precision 0.035 are classified as, Softmax functions are received to come from two sub-networks
Feature vector;In the training stage, using cross entropy as loss function, it is specifically defined as:
<mrow>
<mi>&epsiv;</mi>
<mrow>
<mo>(</mo>
<mi>y</mi>
<mo>,</mo>
<msup>
<mi>y</mi>
<mrow>
<mi>g</mi>
<mi>t</mi>
</mrow>
</msup>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mo>-</mo>
<munder>
<mo>&Sigma;</mo>
<msup>
<mi>y</mi>
<mi>i</mi>
</msup>
</munder>
<mi>p</mi>
<mrow>
<mo>(</mo>
<msub>
<mi>y</mi>
<mi>i</mi>
</msub>
<mo>,</mo>
<msup>
<mi>y</mi>
<mrow>
<mi>g</mi>
<mi>t</mi>
</mrow>
</msup>
<mo>)</mo>
</mrow>
<mo>*</mo>
<mi>l</mi>
<mi>o</mi>
<mi>g</mi>
<mrow>
<mo>(</mo>
<mfrac>
<msup>
<mi>e</mi>
<mrow>
<mo>-</mo>
<msub>
<mi>y</mi>
<mi>i</mi>
</msub>
</mrow>
</msup>
<mrow>
<munder>
<mo>&Sigma;</mo>
<mi>j</mi>
</munder>
<msup>
<mi>e</mi>
<msub>
<mi>y</mi>
<mi>j</mi>
</msub>
</msup>
</mrow>
</mfrac>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<mi>p</mi>
<mrow>
<mo>(</mo>
<msub>
<mi>y</mi>
<mi>i</mi>
</msub>
<mo>,</mo>
<msup>
<mi>y</mi>
<mrow>
<mi>g</mi>
<mi>t</mi>
</mrow>
</msup>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<msub>
<mi>&lambda;</mi>
<mn>1</mn>
</msub>
</mtd>
<mtd>
<mrow>
<msub>
<mi>y</mi>
<mi>i</mi>
</msub>
<mo>=</mo>
<msup>
<msub>
<mi>y</mi>
<mi>i</mi>
</msub>
<mrow>
<mi>g</mi>
<mi>t</mi>
</mrow>
</msup>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<msub>
<mi>&lambda;</mi>
<mn>2</mn>
</msub>
</mtd>
<mtd>
<mrow>
<mo>|</mo>
<mrow>
<msub>
<mi>y</mi>
<mi>i</mi>
</msub>
<mo>-</mo>
<msup>
<mi>y</mi>
<mrow>
<mi>g</mi>
<mi>t</mi>
</mrow>
</msup>
</mrow>
<mo>|</mo>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mn>0</mn>
</mtd>
<mtd>
<mrow>
<mi>o</mi>
<mi>t</mi>
<mi>h</mi>
<mi>e</mi>
<mi>r</mi>
<mi>w</mi>
<mi>i</mi>
<mi>s</mi>
<mi>e</mi>
</mrow>
</mtd>
</mtr>
</mtable>
</mfenced>
</mrow>
Wherein, y is the output valve of mixed type convolutional neural networks, ygtIt is label value, wherein, λ1=0.5, λ2=0.25.
7. a kind of light field image depth estimation method based on mixed type convolutional neural networks according to claim 6, its
It is characterized in that the second level convolutional neural networks are specially:
The training data input of second level convolutional neural networks is the prediction result of first order convolutional neural networks, by light field data
Light field image in storehouse carries out regression forecasting with trained first order convolutional neural networks, generates primary light field image
Estimation of Depth figure, as the training data of second level convolutional neural networks;Second level convolutional neural networks are by a convolution
Layer and a full articulamentum composition, convolutional layer use 7 × 7 convolution kernel, and convolution kernel number is 32, and the characteristic pattern quantity of generation is
32, it is connected with a full articulamentum behind convolutional layer, the final result of output light field picture depth figure.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711337965.XA CN107993260A (en) | 2017-12-14 | 2017-12-14 | A kind of light field image depth estimation method based on mixed type convolutional neural networks |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711337965.XA CN107993260A (en) | 2017-12-14 | 2017-12-14 | A kind of light field image depth estimation method based on mixed type convolutional neural networks |
Publications (1)
Publication Number | Publication Date |
---|---|
CN107993260A true CN107993260A (en) | 2018-05-04 |
Family
ID=62038479
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201711337965.XA Pending CN107993260A (en) | 2017-12-14 | 2017-12-14 | A kind of light field image depth estimation method based on mixed type convolutional neural networks |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107993260A (en) |
Cited By (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109031668A (en) * | 2018-09-18 | 2018-12-18 | 京东方科技集团股份有限公司 | A kind of virtual reality device and its focal length intelligent adjusting method |
CN109299656A (en) * | 2018-08-13 | 2019-02-01 | 浙江零跑科技有限公司 | A kind of deeply determining method of vehicle-mounted vision system scene visual |
CN109344818A (en) * | 2018-09-28 | 2019-02-15 | 合肥工业大学 | A kind of light field well-marked target detection method based on depth convolutional network |
CN109496316A (en) * | 2018-07-28 | 2019-03-19 | 合刃科技(深圳)有限公司 | Image identification system |
CN109544621A (en) * | 2018-11-21 | 2019-03-29 | 马浩鑫 | Light field depth estimation method, system and medium based on convolutional neural networks |
CN109584340A (en) * | 2018-12-11 | 2019-04-05 | 苏州中科广视文化科技有限公司 | New Century Planned Textbook synthetic method based on depth convolutional neural networks |
CN109934863A (en) * | 2019-03-13 | 2019-06-25 | 北京信息科技大学 | A kind of light field depth information estimation method based on intensive connecting-type convolutional neural networks |
CN109949354A (en) * | 2019-03-13 | 2019-06-28 | 北京信息科技大学 | A kind of light field depth information estimation method based on full convolutional neural networks |
CN110060303A (en) * | 2019-03-18 | 2019-07-26 | 英特科利(江苏)医用内窥影像技术有限公司 | A kind of two step scaling methods of light-field camera |
CN110163246A (en) * | 2019-04-08 | 2019-08-23 | 杭州电子科技大学 | The unsupervised depth estimation method of monocular light field image based on convolutional neural networks |
CN110823094A (en) * | 2019-11-08 | 2020-02-21 | 北京理工大学 | Point light source three-dimensional coordinate measuring method and device |
CN111028273A (en) * | 2019-11-27 | 2020-04-17 | 山东大学 | Light field depth estimation method based on multi-stream convolution neural network and implementation system thereof |
CN111445465A (en) * | 2020-03-31 | 2020-07-24 | 江南大学 | Light field image snowflake or rain strip detection and removal method and device based on deep learning |
CN111489407A (en) * | 2020-04-09 | 2020-08-04 | 中国科学技术大学先进技术研究院 | Light field image editing method, device, equipment and storage medium |
CN112116646A (en) * | 2020-09-23 | 2020-12-22 | 南京工程学院 | Light field image depth estimation method based on depth convolution neural network |
CN112150526A (en) * | 2020-07-27 | 2020-12-29 | 浙江大学 | Light field image depth estimation method based on depth learning |
CN112508171A (en) * | 2020-11-23 | 2021-03-16 | 中国辐射防护研究院 | Image depth estimation method and device based on multilayer convolutional neural network |
CN112633248A (en) * | 2021-01-05 | 2021-04-09 | 清华大学深圳国际研究生院 | Deep learning all-in-focus microscopic image acquisition method |
CN112767466A (en) * | 2021-01-20 | 2021-05-07 | 大连理工大学 | Light field depth estimation method based on multi-mode information |
CN113661428A (en) * | 2019-03-29 | 2021-11-16 | 脸谱科技有限责任公司 | Trajectory estimation of MEMS reflectors |
CN118262901A (en) * | 2024-04-07 | 2024-06-28 | 中国人民解放军总医院第六医学中心 | Deep learning-based lung cancer type prediction system |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105357515A (en) * | 2015-12-18 | 2016-02-24 | 天津中科智能识别产业技术研究院有限公司 | Color and depth imaging method and device based on structured light and light-field imaging |
-
2017
- 2017-12-14 CN CN201711337965.XA patent/CN107993260A/en active Pending
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105357515A (en) * | 2015-12-18 | 2016-02-24 | 天津中科智能识别产业技术研究院有限公司 | Color and depth imaging method and device based on structured light and light-field imaging |
Non-Patent Citations (3)
Title |
---|
KATRIN HONAUER 等: ""A Dataset and Evaluation Methodology for Depth Estimation on 4D Light Fields"", 《ACCV 2016: COMPUTER VISION – ACCV 2016 》 * |
YAOXIANG LUO 等: ""EPI-Patch Based Convolutional Neural Network for Depth Estimation on 4D Light Field"", 《ICONIP 2017: NEURAL INFORMATION PROCESSING》 * |
潘志伟: ""基于卷积神经网络的光场图像深度估计"", 《中国优秀硕士学位论文全文数据库 信息科技辑》 * |
Cited By (36)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020024079A1 (en) * | 2018-07-28 | 2020-02-06 | 合刃科技(深圳)有限公司 | Image recognition system |
CN109496316A (en) * | 2018-07-28 | 2019-03-19 | 合刃科技(深圳)有限公司 | Image identification system |
CN109496316B (en) * | 2018-07-28 | 2022-04-01 | 合刃科技(深圳)有限公司 | Image recognition system |
CN109299656A (en) * | 2018-08-13 | 2019-02-01 | 浙江零跑科技有限公司 | A kind of deeply determining method of vehicle-mounted vision system scene visual |
CN109031668A (en) * | 2018-09-18 | 2018-12-18 | 京东方科技集团股份有限公司 | A kind of virtual reality device and its focal length intelligent adjusting method |
CN109031668B (en) * | 2018-09-18 | 2020-12-01 | 京东方科技集团股份有限公司 | Virtual reality device and intelligent focal length adjusting method thereof |
US10812703B2 (en) | 2018-09-18 | 2020-10-20 | Beijing Boe Display Technology Co., Ltd. | Virtual reality device, method for adjusting focal lengths automatically, method for producing virtual reality device and computer readable medium |
CN109344818A (en) * | 2018-09-28 | 2019-02-15 | 合肥工业大学 | A kind of light field well-marked target detection method based on depth convolutional network |
CN109344818B (en) * | 2018-09-28 | 2020-04-14 | 合肥工业大学 | Light field significant target detection method based on deep convolutional network |
CN109544621A (en) * | 2018-11-21 | 2019-03-29 | 马浩鑫 | Light field depth estimation method, system and medium based on convolutional neural networks |
CN109584340B (en) * | 2018-12-11 | 2023-04-18 | 苏州中科广视文化科技有限公司 | New visual angle synthesis method based on deep convolutional neural network |
CN109584340A (en) * | 2018-12-11 | 2019-04-05 | 苏州中科广视文化科技有限公司 | New Century Planned Textbook synthetic method based on depth convolutional neural networks |
CN109949354A (en) * | 2019-03-13 | 2019-06-28 | 北京信息科技大学 | A kind of light field depth information estimation method based on full convolutional neural networks |
CN109949354B (en) * | 2019-03-13 | 2023-11-14 | 北京信息科技大学 | Light field depth information estimation method based on full convolution neural network |
CN109934863B (en) * | 2019-03-13 | 2023-11-14 | 北京信息科技大学 | Light field depth information estimation method based on dense connection type convolutional neural network |
CN109934863A (en) * | 2019-03-13 | 2019-06-25 | 北京信息科技大学 | A kind of light field depth information estimation method based on intensive connecting-type convolutional neural networks |
CN110060303A (en) * | 2019-03-18 | 2019-07-26 | 英特科利(江苏)医用内窥影像技术有限公司 | A kind of two step scaling methods of light-field camera |
CN113661428A (en) * | 2019-03-29 | 2021-11-16 | 脸谱科技有限责任公司 | Trajectory estimation of MEMS reflectors |
CN110163246B (en) * | 2019-04-08 | 2021-03-30 | 杭州电子科技大学 | Monocular light field image unsupervised depth estimation method based on convolutional neural network |
CN110163246A (en) * | 2019-04-08 | 2019-08-23 | 杭州电子科技大学 | The unsupervised depth estimation method of monocular light field image based on convolutional neural networks |
CN110823094B (en) * | 2019-11-08 | 2021-03-30 | 北京理工大学 | Point light source three-dimensional coordinate measuring method and device |
CN110823094A (en) * | 2019-11-08 | 2020-02-21 | 北京理工大学 | Point light source three-dimensional coordinate measuring method and device |
CN111028273B (en) * | 2019-11-27 | 2023-04-07 | 山东大学 | Light field depth estimation method based on multi-stream convolution neural network and implementation system thereof |
CN111028273A (en) * | 2019-11-27 | 2020-04-17 | 山东大学 | Light field depth estimation method based on multi-stream convolution neural network and implementation system thereof |
CN111445465A (en) * | 2020-03-31 | 2020-07-24 | 江南大学 | Light field image snowflake or rain strip detection and removal method and device based on deep learning |
CN111489407A (en) * | 2020-04-09 | 2020-08-04 | 中国科学技术大学先进技术研究院 | Light field image editing method, device, equipment and storage medium |
CN111489407B (en) * | 2020-04-09 | 2023-06-02 | 中国科学技术大学先进技术研究院 | Light field image editing method, device, equipment and storage medium |
CN112150526A (en) * | 2020-07-27 | 2020-12-29 | 浙江大学 | Light field image depth estimation method based on depth learning |
CN112116646A (en) * | 2020-09-23 | 2020-12-22 | 南京工程学院 | Light field image depth estimation method based on depth convolution neural network |
CN112116646B (en) * | 2020-09-23 | 2023-11-24 | 南京工程学院 | Depth estimation method for light field image based on depth convolution neural network |
CN112508171A (en) * | 2020-11-23 | 2021-03-16 | 中国辐射防护研究院 | Image depth estimation method and device based on multilayer convolutional neural network |
CN112633248B (en) * | 2021-01-05 | 2023-08-18 | 清华大学深圳国际研究生院 | Deep learning full-in-focus microscopic image acquisition method |
CN112633248A (en) * | 2021-01-05 | 2021-04-09 | 清华大学深圳国际研究生院 | Deep learning all-in-focus microscopic image acquisition method |
CN112767466B (en) * | 2021-01-20 | 2022-10-11 | 大连理工大学 | Light field depth estimation method based on multi-mode information |
CN112767466A (en) * | 2021-01-20 | 2021-05-07 | 大连理工大学 | Light field depth estimation method based on multi-mode information |
CN118262901A (en) * | 2024-04-07 | 2024-06-28 | 中国人民解放军总医院第六医学中心 | Deep learning-based lung cancer type prediction system |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107993260A (en) | A kind of light field image depth estimation method based on mixed type convolutional neural networks | |
Zhao et al. | Pyramid global context network for image dehazing | |
CN110188835B (en) | Data-enhanced pedestrian re-identification method based on generative confrontation network model | |
CN104061907B (en) | The most variable gait recognition method in visual angle based on the coupling synthesis of gait three-D profile | |
WO2018076212A1 (en) | De-convolutional neural network-based scene semantic segmentation method | |
CN107066916B (en) | Scene semantic segmentation method based on deconvolution neural network | |
CN108090960A (en) | A kind of Object reconstruction method based on geometrical constraint | |
CN109558806A (en) | The detection method and system of high score Remote Sensing Imagery Change | |
CN106372648A (en) | Multi-feature-fusion-convolutional-neural-network-based plankton image classification method | |
CN108960404B (en) | Image-based crowd counting method and device | |
CN104156957B (en) | Stable and high-efficiency high-resolution stereo matching method | |
CN108648161A (en) | The binocular vision obstacle detection system and method for asymmetric nuclear convolutional neural networks | |
CN111639587B (en) | Hyperspectral image classification method based on multi-scale spectrum space convolution neural network | |
CN105528785A (en) | Binocular visual image stereo matching method | |
CN105894513B (en) | Take the remote sensing image variation detection method and system of imaged object change in time and space into account | |
CN102034267A (en) | Three-dimensional reconstruction method of target based on attention | |
CN114092697A (en) | Building facade semantic segmentation method with attention fused with global and local depth features | |
CN116206133A (en) | RGB-D significance target detection method | |
CN103971354A (en) | Method for reconstructing low-resolution infrared image into high-resolution infrared image | |
Gao et al. | Joint optimization of depth and ego-motion for intelligent autonomous vehicles | |
Song et al. | Accurate 3D reconstruction from circular light field using CNN-LSTM | |
Yuan et al. | Structure flow-guided network for real depth super-resolution | |
Chu et al. | DBFGAN: Dual branch feature guided aggregation network for remote sensing image | |
Song et al. | A single image dehazing method based on end-to-end cpad-net network in deep learning environment | |
Li et al. | Automatic Road Extraction from High-Resolution Remote Sensing Image Based on Bat Model and Mutual Information Matching. |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20180504 |